Apr 21 00:03:06.060913 ip-10-0-132-206 systemd[1]: Starting Kubernetes Kubelet... Apr 21 00:03:06.495843 ip-10-0-132-206 kubenswrapper[2568]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 00:03:06.495843 ip-10-0-132-206 kubenswrapper[2568]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 21 00:03:06.495843 ip-10-0-132-206 kubenswrapper[2568]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 00:03:06.495843 ip-10-0-132-206 kubenswrapper[2568]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 21 00:03:06.495843 ip-10-0-132-206 kubenswrapper[2568]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 00:03:06.497555 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.497482 2568 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 21 00:03:06.500720 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500706 2568 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 00:03:06.500720 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500720 2568 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500724 2568 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500727 2568 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500730 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500733 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500736 2568 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500739 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500742 2568 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500745 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500748 2568 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500750 2568 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500758 2568 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500762 2568 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500765 2568 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500767 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500770 2568 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500772 2568 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500775 2568 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500778 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500780 2568 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 00:03:06.500785 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500783 2568 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500786 2568 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500789 2568 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500791 2568 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500794 2568 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500797 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500800 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500802 2568 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500805 2568 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500808 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500810 2568 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500813 2568 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500815 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500818 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500820 2568 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500823 2568 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500825 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500828 2568 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500830 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 00:03:06.501278 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500834 2568 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500837 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500840 2568 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500843 2568 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500845 2568 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500848 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500850 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500853 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500856 2568 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500858 2568 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500860 2568 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500863 2568 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500866 2568 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500869 2568 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500871 2568 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500874 2568 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500877 2568 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500879 2568 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500882 2568 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 00:03:06.501759 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500885 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500888 2568 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500890 2568 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500893 2568 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500896 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500898 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500901 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500903 2568 feature_gate.go:328] unrecognized feature gate: Example Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500906 2568 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500910 2568 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500913 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500917 2568 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500920 2568 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500923 2568 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500926 2568 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500928 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500932 2568 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500936 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500938 2568 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500941 2568 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 00:03:06.502238 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500944 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500947 2568 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500950 2568 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500952 2568 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500955 2568 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500957 2568 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.500960 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501288 2568 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501293 2568 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501296 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501298 2568 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501301 2568 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501303 2568 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501306 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501309 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501312 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501314 2568 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501317 2568 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501320 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501322 2568 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 00:03:06.502735 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501325 2568 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501328 2568 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501331 2568 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501334 2568 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501337 2568 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501339 2568 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501342 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501344 2568 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501347 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501350 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501352 2568 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501355 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501358 2568 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501360 2568 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501363 2568 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501365 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501368 2568 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501370 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501373 2568 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 00:03:06.503243 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501375 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501379 2568 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501382 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501384 2568 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501387 2568 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501390 2568 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501392 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501395 2568 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501397 2568 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501400 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501403 2568 feature_gate.go:328] unrecognized feature gate: Example Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501405 2568 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501408 2568 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501411 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501414 2568 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501416 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501419 2568 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501421 2568 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501424 2568 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501426 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 00:03:06.503733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501429 2568 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501431 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501434 2568 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501437 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501439 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501441 2568 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501444 2568 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501447 2568 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501449 2568 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501451 2568 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501455 2568 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501467 2568 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501471 2568 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501475 2568 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501478 2568 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501481 2568 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501484 2568 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501487 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501489 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 00:03:06.504250 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501492 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501495 2568 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501498 2568 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501501 2568 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501503 2568 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501506 2568 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501508 2568 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501511 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501513 2568 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501515 2568 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501518 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501521 2568 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501523 2568 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501526 2568 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.501528 2568 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502800 2568 flags.go:64] FLAG: --address="0.0.0.0" Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502809 2568 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502816 2568 flags.go:64] FLAG: --anonymous-auth="true" Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502825 2568 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502830 2568 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502833 2568 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 21 00:03:06.504745 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502838 2568 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502842 2568 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502845 2568 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502848 2568 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502852 2568 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502855 2568 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502858 2568 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502861 2568 flags.go:64] FLAG: --cgroup-root="" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502863 2568 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502866 2568 flags.go:64] FLAG: --client-ca-file="" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502869 2568 flags.go:64] FLAG: --cloud-config="" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502872 2568 flags.go:64] FLAG: --cloud-provider="external" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502875 2568 flags.go:64] FLAG: --cluster-dns="[]" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502879 2568 flags.go:64] FLAG: --cluster-domain="" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502881 2568 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502885 2568 flags.go:64] FLAG: --config-dir="" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502887 2568 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502891 2568 flags.go:64] FLAG: --container-log-max-files="5" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502895 2568 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502898 2568 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502901 2568 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502904 2568 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502907 2568 flags.go:64] FLAG: --contention-profiling="false" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502910 2568 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 21 00:03:06.505258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502913 2568 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502916 2568 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502919 2568 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502923 2568 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502926 2568 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502929 2568 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502933 2568 flags.go:64] FLAG: --enable-load-reader="false" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502936 2568 flags.go:64] FLAG: --enable-server="true" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502938 2568 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502942 2568 flags.go:64] FLAG: --event-burst="100" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502946 2568 flags.go:64] FLAG: --event-qps="50" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502949 2568 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502952 2568 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502954 2568 flags.go:64] FLAG: --eviction-hard="" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502958 2568 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502961 2568 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502964 2568 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502967 2568 flags.go:64] FLAG: --eviction-soft="" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502970 2568 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502973 2568 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502976 2568 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502979 2568 flags.go:64] FLAG: --experimental-mounter-path="" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502981 2568 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502984 2568 flags.go:64] FLAG: --fail-swap-on="true" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502987 2568 flags.go:64] FLAG: --feature-gates="" Apr 21 00:03:06.505856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502991 2568 flags.go:64] FLAG: --file-check-frequency="20s" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502994 2568 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.502997 2568 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503000 2568 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503004 2568 flags.go:64] FLAG: --healthz-port="10248" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503007 2568 flags.go:64] FLAG: --help="false" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503010 2568 flags.go:64] FLAG: --hostname-override="ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503013 2568 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503016 2568 flags.go:64] FLAG: --http-check-frequency="20s" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503018 2568 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503022 2568 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503025 2568 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503028 2568 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503031 2568 flags.go:64] FLAG: --image-service-endpoint="" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503034 2568 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503037 2568 flags.go:64] FLAG: --kube-api-burst="100" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503040 2568 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503043 2568 flags.go:64] FLAG: --kube-api-qps="50" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503045 2568 flags.go:64] FLAG: --kube-reserved="" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503048 2568 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503051 2568 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503054 2568 flags.go:64] FLAG: --kubelet-cgroups="" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503057 2568 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503060 2568 flags.go:64] FLAG: --lock-file="" Apr 21 00:03:06.506457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503063 2568 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503065 2568 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503068 2568 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503074 2568 flags.go:64] FLAG: --log-json-split-stream="false" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503077 2568 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503079 2568 flags.go:64] FLAG: --log-text-split-stream="false" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503082 2568 flags.go:64] FLAG: --logging-format="text" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503085 2568 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503089 2568 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503091 2568 flags.go:64] FLAG: --manifest-url="" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503094 2568 flags.go:64] FLAG: --manifest-url-header="" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503098 2568 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503103 2568 flags.go:64] FLAG: --max-open-files="1000000" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503107 2568 flags.go:64] FLAG: --max-pods="110" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503110 2568 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503113 2568 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503115 2568 flags.go:64] FLAG: --memory-manager-policy="None" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503118 2568 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503121 2568 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503124 2568 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503127 2568 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503133 2568 flags.go:64] FLAG: --node-status-max-images="50" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503139 2568 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503142 2568 flags.go:64] FLAG: --oom-score-adj="-999" Apr 21 00:03:06.507051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503146 2568 flags.go:64] FLAG: --pod-cidr="" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503149 2568 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503154 2568 flags.go:64] FLAG: --pod-manifest-path="" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503157 2568 flags.go:64] FLAG: --pod-max-pids="-1" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503160 2568 flags.go:64] FLAG: --pods-per-core="0" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503163 2568 flags.go:64] FLAG: --port="10250" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503166 2568 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503168 2568 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-0b08cd381714ca0e1" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503172 2568 flags.go:64] FLAG: --qos-reserved="" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503174 2568 flags.go:64] FLAG: --read-only-port="10255" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503177 2568 flags.go:64] FLAG: --register-node="true" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503180 2568 flags.go:64] FLAG: --register-schedulable="true" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503183 2568 flags.go:64] FLAG: --register-with-taints="" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503187 2568 flags.go:64] FLAG: --registry-burst="10" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503190 2568 flags.go:64] FLAG: --registry-qps="5" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503193 2568 flags.go:64] FLAG: --reserved-cpus="" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503196 2568 flags.go:64] FLAG: --reserved-memory="" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503200 2568 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503203 2568 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503206 2568 flags.go:64] FLAG: --rotate-certificates="false" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503209 2568 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503213 2568 flags.go:64] FLAG: --runonce="false" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503216 2568 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503219 2568 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503221 2568 flags.go:64] FLAG: --seccomp-default="false" Apr 21 00:03:06.507637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503224 2568 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503227 2568 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503230 2568 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503233 2568 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503236 2568 flags.go:64] FLAG: --storage-driver-password="root" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503239 2568 flags.go:64] FLAG: --storage-driver-secure="false" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503242 2568 flags.go:64] FLAG: --storage-driver-table="stats" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503245 2568 flags.go:64] FLAG: --storage-driver-user="root" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503248 2568 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503251 2568 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503254 2568 flags.go:64] FLAG: --system-cgroups="" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503258 2568 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503263 2568 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503265 2568 flags.go:64] FLAG: --tls-cert-file="" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503268 2568 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503272 2568 flags.go:64] FLAG: --tls-min-version="" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503275 2568 flags.go:64] FLAG: --tls-private-key-file="" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503277 2568 flags.go:64] FLAG: --topology-manager-policy="none" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503280 2568 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503284 2568 flags.go:64] FLAG: --topology-manager-scope="container" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503287 2568 flags.go:64] FLAG: --v="2" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503291 2568 flags.go:64] FLAG: --version="false" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503295 2568 flags.go:64] FLAG: --vmodule="" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503298 2568 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.503302 2568 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 21 00:03:06.508249 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503387 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503390 2568 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503393 2568 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503397 2568 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503400 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503402 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503405 2568 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503408 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503410 2568 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503414 2568 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503418 2568 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503421 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503424 2568 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503426 2568 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503429 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503432 2568 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503434 2568 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503437 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503439 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503442 2568 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 00:03:06.508998 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503445 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503447 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503450 2568 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503452 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503454 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503458 2568 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503460 2568 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503463 2568 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503465 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503468 2568 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503470 2568 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503473 2568 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503475 2568 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503478 2568 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503480 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503483 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503486 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503488 2568 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503491 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503494 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 00:03:06.509624 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503496 2568 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503499 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503501 2568 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503504 2568 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503507 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503509 2568 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503512 2568 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503515 2568 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503518 2568 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503520 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503522 2568 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503525 2568 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503528 2568 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503532 2568 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503535 2568 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503537 2568 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503540 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503544 2568 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503547 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503550 2568 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 00:03:06.510325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503553 2568 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503555 2568 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503558 2568 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503561 2568 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503563 2568 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503566 2568 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503568 2568 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503571 2568 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503573 2568 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503576 2568 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503578 2568 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503581 2568 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503583 2568 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503586 2568 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503588 2568 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503591 2568 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503593 2568 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503595 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503598 2568 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 00:03:06.511074 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503601 2568 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 00:03:06.511544 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503614 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 00:03:06.511544 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503617 2568 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 00:03:06.511544 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503620 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 00:03:06.511544 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503622 2568 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 00:03:06.511544 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503625 2568 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 00:03:06.511544 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.503627 2568 feature_gate.go:328] unrecognized feature gate: Example Apr 21 00:03:06.511544 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.504440 2568 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 00:03:06.511857 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.511840 2568 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 21 00:03:06.511891 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.511858 2568 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 21 00:03:06.511918 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511906 2568 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 00:03:06.511918 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511912 2568 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 00:03:06.511918 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511917 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511921 2568 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511924 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511928 2568 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511932 2568 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511935 2568 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511939 2568 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511942 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511944 2568 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511947 2568 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511950 2568 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511952 2568 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511955 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511957 2568 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511960 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511963 2568 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511966 2568 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511968 2568 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511971 2568 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 00:03:06.512000 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511974 2568 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511977 2568 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511979 2568 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511982 2568 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511984 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511987 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511989 2568 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511992 2568 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511995 2568 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.511997 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512000 2568 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512002 2568 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512005 2568 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512008 2568 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512011 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512014 2568 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512016 2568 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512019 2568 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512021 2568 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512024 2568 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 00:03:06.512465 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512026 2568 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512029 2568 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512031 2568 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512034 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512037 2568 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512040 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512042 2568 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512045 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512047 2568 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512050 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512052 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512055 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512057 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512060 2568 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512063 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512066 2568 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512068 2568 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512071 2568 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512073 2568 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 00:03:06.512962 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512076 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512079 2568 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512081 2568 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512084 2568 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512086 2568 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512088 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512091 2568 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512093 2568 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512096 2568 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512098 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512101 2568 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512103 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512106 2568 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512108 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512111 2568 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512113 2568 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512116 2568 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512119 2568 feature_gate.go:328] unrecognized feature gate: Example Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512122 2568 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512124 2568 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 00:03:06.513423 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512127 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512129 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512132 2568 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512134 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512137 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512139 2568 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.512144 2568 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512233 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512238 2568 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512242 2568 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512244 2568 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512248 2568 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512250 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512253 2568 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512256 2568 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 00:03:06.513974 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512259 2568 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512262 2568 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512264 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512267 2568 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512269 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512272 2568 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512274 2568 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512277 2568 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512280 2568 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512282 2568 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512285 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512288 2568 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512290 2568 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512293 2568 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512296 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512298 2568 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512300 2568 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512303 2568 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512305 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512308 2568 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 00:03:06.514349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512310 2568 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512313 2568 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512316 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512319 2568 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512321 2568 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512325 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512327 2568 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512330 2568 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512332 2568 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512335 2568 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512337 2568 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512340 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512342 2568 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512345 2568 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512347 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512350 2568 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512352 2568 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512355 2568 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512357 2568 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512360 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 00:03:06.514872 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512362 2568 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512365 2568 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512368 2568 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512370 2568 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512373 2568 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512375 2568 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512378 2568 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512381 2568 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512383 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512385 2568 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512388 2568 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512390 2568 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512393 2568 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512395 2568 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512397 2568 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512400 2568 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512402 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512405 2568 feature_gate.go:328] unrecognized feature gate: Example Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512409 2568 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 00:03:06.515349 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512412 2568 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512415 2568 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512417 2568 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512420 2568 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512424 2568 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512427 2568 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512430 2568 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512432 2568 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512435 2568 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512438 2568 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512440 2568 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512443 2568 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512445 2568 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512448 2568 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512450 2568 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512453 2568 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512455 2568 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512458 2568 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 00:03:06.515916 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:06.512460 2568 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 00:03:06.516348 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.512465 2568 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 00:03:06.516348 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.513142 2568 server.go:962] "Client rotation is on, will bootstrap in background" Apr 21 00:03:06.516348 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.515647 2568 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 21 00:03:06.516658 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.516647 2568 server.go:1019] "Starting client certificate rotation" Apr 21 00:03:06.516753 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.516737 2568 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 21 00:03:06.516786 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.516776 2568 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 21 00:03:06.542294 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.542274 2568 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 21 00:03:06.544778 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.544760 2568 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 21 00:03:06.558921 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.558903 2568 log.go:25] "Validated CRI v1 runtime API" Apr 21 00:03:06.564111 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.564096 2568 log.go:25] "Validated CRI v1 image API" Apr 21 00:03:06.565320 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.565306 2568 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 21 00:03:06.570489 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.570470 2568 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 21 00:03:06.574194 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.574174 2568 fs.go:135] Filesystem UUIDs: map[104dff1a-a71f-457f-8bf3-027cc0fceeac:/dev/nvme0n1p3 2f20e71d-0dc2-4f08-8f12-001bb83ac7dc:/dev/nvme0n1p4 7B77-95E7:/dev/nvme0n1p2] Apr 21 00:03:06.574260 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.574193 2568 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 21 00:03:06.579443 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.579346 2568 manager.go:217] Machine: {Timestamp:2026-04-21 00:03:06.577654283 +0000 UTC m=+0.402633239 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3099761 MemoryCapacity:33164488704 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec28c4e1966eb13ada873f4c694fd2d6 SystemUUID:ec28c4e1-966e-b13a-da87-3f4c694fd2d6 BootID:ba5274aa-a26a-4375-a71e-2940c5a6e8cb Filesystems:[{Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16582246400 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16582242304 Type:vfs Inodes:4048399 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6632898560 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:3e:16:4f:26:d5 Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:3e:16:4f:26:d5 Speed:0 Mtu:9001} {Name:ovs-system MacAddress:0e:7c:05:33:c1:e7 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33164488704 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 21 00:03:06.579443 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.579438 2568 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 21 00:03:06.579553 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.579502 2568 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 21 00:03:06.582266 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.582241 2568 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 21 00:03:06.582420 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.582267 2568 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-132-206.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 21 00:03:06.582465 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.582429 2568 topology_manager.go:138] "Creating topology manager with none policy" Apr 21 00:03:06.582465 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.582437 2568 container_manager_linux.go:306] "Creating device plugin manager" Apr 21 00:03:06.582465 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.582450 2568 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 21 00:03:06.583170 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.583160 2568 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 21 00:03:06.584367 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.584357 2568 state_mem.go:36] "Initialized new in-memory state store" Apr 21 00:03:06.584588 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.584579 2568 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 21 00:03:06.586228 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.586211 2568 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-p7fh9" Apr 21 00:03:06.586883 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.586873 2568 kubelet.go:491] "Attempting to sync node with API server" Apr 21 00:03:06.586922 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.586886 2568 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 21 00:03:06.586922 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.586897 2568 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 21 00:03:06.586922 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.586905 2568 kubelet.go:397] "Adding apiserver pod source" Apr 21 00:03:06.586922 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.586912 2568 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 21 00:03:06.587965 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.587954 2568 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 21 00:03:06.588006 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.587971 2568 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 21 00:03:06.592885 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.592863 2568 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 21 00:03:06.593311 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.593296 2568 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-p7fh9" Apr 21 00:03:06.594060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.594047 2568 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 21 00:03:06.595805 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595791 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595809 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595816 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595823 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595832 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595840 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595845 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595851 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595857 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595865 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595873 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 21 00:03:06.595887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.595881 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 21 00:03:06.596947 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.596937 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 21 00:03:06.596947 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.596947 2568 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 21 00:03:06.600223 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.600209 2568 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 21 00:03:06.600307 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.600243 2568 server.go:1295] "Started kubelet" Apr 21 00:03:06.600366 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.600330 2568 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 21 00:03:06.600414 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.600376 2568 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 21 00:03:06.600985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.600339 2568 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 21 00:03:06.601326 ip-10-0-132-206 systemd[1]: Started Kubernetes Kubelet. Apr 21 00:03:06.601953 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.601928 2568 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 21 00:03:06.603415 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.603398 2568 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 00:03:06.603503 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.603470 2568 server.go:317] "Adding debug handlers to kubelet server" Apr 21 00:03:06.606959 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.606935 2568 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-132-206.ec2.internal" not found Apr 21 00:03:06.607783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.607761 2568 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 00:03:06.608672 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.608643 2568 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 21 00:03:06.609026 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.609010 2568 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 21 00:03:06.609488 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.609473 2568 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 21 00:03:06.609488 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.609477 2568 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 21 00:03:06.609641 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.609502 2568 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 21 00:03:06.609641 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.609622 2568 reconstruct.go:97] "Volume reconstruction finished" Apr 21 00:03:06.609641 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.609631 2568 reconciler.go:26] "Reconciler: start to sync state" Apr 21 00:03:06.610014 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:06.609796 2568 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-132-206.ec2.internal\" not found" Apr 21 00:03:06.611800 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.611164 2568 factory.go:153] Registering CRI-O factory Apr 21 00:03:06.611800 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.611185 2568 factory.go:223] Registration of the crio container factory successfully Apr 21 00:03:06.611800 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.611448 2568 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 00:03:06.611947 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.611849 2568 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 21 00:03:06.611947 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.611863 2568 factory.go:55] Registering systemd factory Apr 21 00:03:06.611947 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.611872 2568 factory.go:223] Registration of the systemd container factory successfully Apr 21 00:03:06.611947 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.611894 2568 factory.go:103] Registering Raw factory Apr 21 00:03:06.611947 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.611906 2568 manager.go:1196] Started watching for new ooms in manager Apr 21 00:03:06.612292 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.612275 2568 manager.go:319] Starting recovery of all containers Apr 21 00:03:06.613657 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:06.613634 2568 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-10-0-132-206.ec2.internal\" not found" node="ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.615133 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:06.615111 2568 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 21 00:03:06.622105 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.621973 2568 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-132-206.ec2.internal" not found Apr 21 00:03:06.622361 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.622349 2568 manager.go:324] Recovery completed Apr 21 00:03:06.626224 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.626212 2568 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 00:03:06.627892 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.627876 2568 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-132-206.ec2.internal" event="NodeHasSufficientMemory" Apr 21 00:03:06.627964 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.627904 2568 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-132-206.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 00:03:06.627964 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.627914 2568 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-132-206.ec2.internal" event="NodeHasSufficientPID" Apr 21 00:03:06.628300 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.628286 2568 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 21 00:03:06.628300 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.628298 2568 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 21 00:03:06.628377 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.628313 2568 state_mem.go:36] "Initialized new in-memory state store" Apr 21 00:03:06.630149 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.630139 2568 policy_none.go:49] "None policy: Start" Apr 21 00:03:06.630184 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.630152 2568 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 21 00:03:06.630184 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.630162 2568 state_mem.go:35] "Initializing new in-memory state store" Apr 21 00:03:06.665172 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.665157 2568 manager.go:341] "Starting Device Plugin manager" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:06.665190 2568 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.665200 2568 server.go:85] "Starting device plugin registration server" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.665418 2568 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.665431 2568 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.665501 2568 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.665577 2568 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.665586 2568 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:06.665989 2568 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:06.666027 2568 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-132-206.ec2.internal\" not found" Apr 21 00:03:06.684060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.677705 2568 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-132-206.ec2.internal" not found Apr 21 00:03:06.713092 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.713056 2568 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 21 00:03:06.714329 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.714308 2568 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 21 00:03:06.714410 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.714334 2568 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 21 00:03:06.714410 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.714351 2568 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 21 00:03:06.714410 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.714356 2568 kubelet.go:2451] "Starting kubelet main sync loop" Apr 21 00:03:06.714410 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:06.714388 2568 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 21 00:03:06.717053 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.717031 2568 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 00:03:06.765867 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.765824 2568 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 00:03:06.766663 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.766649 2568 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-132-206.ec2.internal" event="NodeHasSufficientMemory" Apr 21 00:03:06.766742 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.766680 2568 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-132-206.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 00:03:06.766742 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.766697 2568 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-132-206.ec2.internal" event="NodeHasSufficientPID" Apr 21 00:03:06.766742 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.766724 2568 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.773131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.773117 2568 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.773195 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:06.773138 2568 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-132-206.ec2.internal\": node \"ip-10-0-132-206.ec2.internal\" not found" Apr 21 00:03:06.815102 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.815051 2568 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal"] Apr 21 00:03:06.819326 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.819309 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.819410 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.819312 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.844108 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.844093 2568 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.848475 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.848462 2568 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.855031 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.855019 2568 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 21 00:03:06.861777 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.861758 2568 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 21 00:03:06.911873 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.911854 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/7b08454bff9ca2a8282774b411985429-config\") pod \"kube-apiserver-proxy-ip-10-0-132-206.ec2.internal\" (UID: \"7b08454bff9ca2a8282774b411985429\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.911873 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.911879 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/fd14846dcd7d8f37d2ba1fd40c2a4cd6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal\" (UID: \"fd14846dcd7d8f37d2ba1fd40c2a4cd6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:06.911996 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:06.911901 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fd14846dcd7d8f37d2ba1fd40c2a4cd6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal\" (UID: \"fd14846dcd7d8f37d2ba1fd40c2a4cd6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.012764 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.012739 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/7b08454bff9ca2a8282774b411985429-config\") pod \"kube-apiserver-proxy-ip-10-0-132-206.ec2.internal\" (UID: \"7b08454bff9ca2a8282774b411985429\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.012764 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.012765 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/fd14846dcd7d8f37d2ba1fd40c2a4cd6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal\" (UID: \"fd14846dcd7d8f37d2ba1fd40c2a4cd6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.012908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.012780 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fd14846dcd7d8f37d2ba1fd40c2a4cd6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal\" (UID: \"fd14846dcd7d8f37d2ba1fd40c2a4cd6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.012908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.012839 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/7b08454bff9ca2a8282774b411985429-config\") pod \"kube-apiserver-proxy-ip-10-0-132-206.ec2.internal\" (UID: \"7b08454bff9ca2a8282774b411985429\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.012908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.012894 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fd14846dcd7d8f37d2ba1fd40c2a4cd6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal\" (UID: \"fd14846dcd7d8f37d2ba1fd40c2a4cd6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.012999 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.012926 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/fd14846dcd7d8f37d2ba1fd40c2a4cd6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal\" (UID: \"fd14846dcd7d8f37d2ba1fd40c2a4cd6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.157317 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.157252 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.164674 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.164658 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" Apr 21 00:03:07.516510 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.516491 2568 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 21 00:03:07.517100 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.516647 2568 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 00:03:07.517100 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.516654 2568 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 00:03:07.517100 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.516682 2568 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 00:03:07.587768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.587740 2568 apiserver.go:52] "Watching apiserver" Apr 21 00:03:07.595475 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.595447 2568 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-19 23:58:06 +0000 UTC" deadline="2027-11-25 12:12:23.884209435 +0000 UTC" Apr 21 00:03:07.595475 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.595473 2568 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="14004h9m16.288739851s" Apr 21 00:03:07.596854 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.596836 2568 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 21 00:03:07.597196 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.597174 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7","openshift-cluster-node-tuning-operator/tuned-7vg48","openshift-dns/node-resolver-wx87b","openshift-image-registry/node-ca-bscx2","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal","openshift-multus/multus-additional-cni-plugins-rm7p5","openshift-multus/network-metrics-daemon-pf4ls","kube-system/konnectivity-agent-lmssf","kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal","openshift-multus/multus-g6sk6","openshift-network-diagnostics/network-check-target-4tj2p","openshift-network-operator/iptables-alerter-8cccm","openshift-ovn-kubernetes/ovnkube-node-z7lpw"] Apr 21 00:03:07.602204 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.602182 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.602408 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.602375 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.604451 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.604432 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.604984 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.604967 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 21 00:03:07.605049 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.604982 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 21 00:03:07.605049 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.604995 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 21 00:03:07.605049 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.605018 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-c7dbj\"" Apr 21 00:03:07.605175 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.605087 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-zt6jn\"" Apr 21 00:03:07.605175 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.605146 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 21 00:03:07.606140 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.606125 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 21 00:03:07.606557 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.606542 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.606882 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.606866 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 21 00:03:07.606882 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.606875 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 21 00:03:07.607129 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.607115 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-qfhfx\"" Apr 21 00:03:07.608699 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.608668 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-jsxbx\"" Apr 21 00:03:07.608781 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.608714 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 21 00:03:07.608868 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.608852 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 21 00:03:07.608922 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.608872 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 21 00:03:07.608922 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.608896 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.609083 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.609067 2568 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 21 00:03:07.610927 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.610909 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:07.611205 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:07.611180 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:07.611797 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.611469 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 21 00:03:07.611797 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.611478 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 21 00:03:07.612931 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.612171 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-dhzhp\"" Apr 21 00:03:07.612931 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.612390 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 21 00:03:07.612931 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.612577 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 21 00:03:07.612931 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.612786 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 21 00:03:07.614029 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.614010 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:07.615572 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615553 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-device-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.615673 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615581 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-modprobe-d\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.615673 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615596 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-kubernetes\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.615673 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615630 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdzpr\" (UniqueName: \"kubernetes.io/projected/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-kube-api-access-jdzpr\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.615673 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615652 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cnibin\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.615673 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615667 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-run\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615693 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-host\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615712 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cni-binary-copy\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615742 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615759 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7qpv\" (UniqueName: \"kubernetes.io/projected/d439efe8-c633-4c7c-ad01-515f448307fd-kube-api-access-l7qpv\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615774 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-registration-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615789 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-systemd\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615801 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-lib-modules\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615815 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-serviceca\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615848 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-etc-selinux\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615887 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-sys-fs\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615946 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysconfig\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.615985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.615985 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysctl-d\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616010 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysctl-conf\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616079 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-sys\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616101 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm6gl\" (UniqueName: \"kubernetes.io/projected/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-kube-api-access-nm6gl\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616124 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-hosts-file\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616152 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-tmp-dir\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616177 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-system-cni-dir\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616201 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616212 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-sprp8\"" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616231 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-kubelet-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616257 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616310 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616256 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-socket-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616353 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616364 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlfmd\" (UniqueName: \"kubernetes.io/projected/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-kube-api-access-hlfmd\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616392 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-var-lib-kubelet\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616419 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-os-release\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616448 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-tuned\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616471 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-tmp\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.616578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616494 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-host\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.617256 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616519 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr8kc\" (UniqueName: \"kubernetes.io/projected/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-kube-api-access-sr8kc\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.617256 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616542 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.617256 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616557 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:07.617256 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.616572 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qghgd\" (UniqueName: \"kubernetes.io/projected/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-kube-api-access-qghgd\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.618341 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.618325 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 21 00:03:07.618419 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.618408 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-sgq7g\"" Apr 21 00:03:07.618504 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.618489 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:07.618574 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:07.618557 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:07.620616 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.620586 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.622097 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.622080 2568 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 21 00:03:07.622554 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.622540 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 21 00:03:07.622619 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.622594 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 21 00:03:07.622811 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.622793 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.622882 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.622867 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 21 00:03:07.623009 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.622992 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-p6d6z\"" Apr 21 00:03:07.626768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.626541 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 21 00:03:07.626768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.626594 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-tljr2\"" Apr 21 00:03:07.626768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.626623 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 21 00:03:07.626768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.626635 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 21 00:03:07.626768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.626692 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 21 00:03:07.626768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.626635 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 21 00:03:07.626768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.626735 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 21 00:03:07.646571 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.646542 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd14846dcd7d8f37d2ba1fd40c2a4cd6.slice/crio-b967863036e68388e7c3f60cbb381a60da5eefa5a2953ad7a6707ba0fbb2d83b WatchSource:0}: Error finding container b967863036e68388e7c3f60cbb381a60da5eefa5a2953ad7a6707ba0fbb2d83b: Status 404 returned error can't find the container with id b967863036e68388e7c3f60cbb381a60da5eefa5a2953ad7a6707ba0fbb2d83b Apr 21 00:03:07.646826 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.646813 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b08454bff9ca2a8282774b411985429.slice/crio-5d6dfef158a824c08b2eda55461d9b38dee9bcfb37715af42db43816f5dfc8ea WatchSource:0}: Error finding container 5d6dfef158a824c08b2eda55461d9b38dee9bcfb37715af42db43816f5dfc8ea: Status 404 returned error can't find the container with id 5d6dfef158a824c08b2eda55461d9b38dee9bcfb37715af42db43816f5dfc8ea Apr 21 00:03:07.647971 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.647953 2568 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-t9cq4" Apr 21 00:03:07.651259 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.651245 2568 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 00:03:07.656506 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.656489 2568 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-t9cq4" Apr 21 00:03:07.710796 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.710779 2568 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 21 00:03:07.716708 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716692 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-registration-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.716781 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716715 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-lib-modules\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.716781 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716731 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-serviceca\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.716781 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716748 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-etc-kubernetes\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.716902 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716817 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-registration-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.716902 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716852 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:07.716902 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716881 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-etc-selinux\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.716902 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716882 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-lib-modules\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716879 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" event={"ID":"7b08454bff9ca2a8282774b411985429","Type":"ContainerStarted","Data":"5d6dfef158a824c08b2eda55461d9b38dee9bcfb37715af42db43816f5dfc8ea"} Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716906 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysconfig\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716949 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysconfig\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716963 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysctl-d\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716971 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-etc-selinux\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.716997 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysctl-conf\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717022 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-sys\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717045 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nm6gl\" (UniqueName: \"kubernetes.io/projected/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-kube-api-access-nm6gl\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717069 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-hosts-file\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717095 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-cni-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717117 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysctl-d\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717130 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-hosts-file\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717122 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-tmp-dir\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717149 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-serviceca\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717227 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-sys\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717263 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/b470fb01-818e-4ce0-8ff7-35c94e72e940-agent-certs\") pod \"konnectivity-agent-lmssf\" (UID: \"b470fb01-818e-4ce0-8ff7-35c94e72e940\") " pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717289 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-os-release\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717317 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-sysctl-conf\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717321 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-cni-multus\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717360 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6wnj\" (UniqueName: \"kubernetes.io/projected/18a5a55e-ea26-46ec-a17f-1cac467f786c-kube-api-access-t6wnj\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717386 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/1c09a22d-4059-455d-908f-0b608d24e6bb-iptables-alerter-script\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717410 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-slash\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717429 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-etc-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717452 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hlfmd\" (UniqueName: \"kubernetes.io/projected/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-kube-api-access-hlfmd\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717487 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-var-lib-kubelet\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717519 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-os-release\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717535 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-tmp-dir\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717544 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovnkube-config\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.717681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717582 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-var-lib-kubelet\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717598 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-tmp\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717639 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sr8kc\" (UniqueName: \"kubernetes.io/projected/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-kube-api-access-sr8kc\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717641 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-os-release\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717664 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-kubelet\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717688 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c09a22d-4059-455d-908f-0b608d24e6bb-host-slash\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717713 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-kubelet\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717735 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-node-log\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717762 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qghgd\" (UniqueName: \"kubernetes.io/projected/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-kube-api-access-qghgd\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717788 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-system-cni-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717811 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-cnibin\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717839 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-k8s-cni-cncf-io\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717862 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovnkube-script-lib\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717862 2568 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717887 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-kubernetes\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717911 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jdzpr\" (UniqueName: \"kubernetes.io/projected/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-kube-api-access-jdzpr\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717934 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cnibin\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.718324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717957 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/b470fb01-818e-4ce0-8ff7-35c94e72e940-konnectivity-ca\") pod \"konnectivity-agent-lmssf\" (UID: \"b470fb01-818e-4ce0-8ff7-35c94e72e940\") " pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.717976 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-ovn\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718000 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-run\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718020 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-host\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718066 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-kubernetes\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718073 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-host\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718068 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/18a5a55e-ea26-46ec-a17f-1cac467f786c-cni-binary-copy\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718066 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-run\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718111 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-hostroot\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718143 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-var-lib-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718113 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cnibin\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718170 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718198 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovn-node-metrics-cert\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718222 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcdbj\" (UniqueName: \"kubernetes.io/projected/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-kube-api-access-hcdbj\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718249 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-systemd\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718283 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-netns\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718306 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-cni-bin\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.718923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718313 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-systemd\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718329 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-run-ovn-kubernetes\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718353 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" event={"ID":"fd14846dcd7d8f37d2ba1fd40c2a4cd6","Type":"ContainerStarted","Data":"b967863036e68388e7c3f60cbb381a60da5eefa5a2953ad7a6707ba0fbb2d83b"} Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718363 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-sys-fs\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718398 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-sys-fs\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718446 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-systemd-units\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718476 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-system-cni-dir\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718503 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718506 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-system-cni-dir\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718523 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-conf-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718547 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqgnt\" (UniqueName: \"kubernetes.io/projected/1c09a22d-4059-455d-908f-0b608d24e6bb-kube-api-access-cqgnt\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718567 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-cni-bin\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718583 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-kubelet-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718598 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-socket-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718624 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718634 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-socket-dir-parent\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.719634 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718660 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-systemd\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718683 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718695 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-socket-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718659 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-kubelet-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718699 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-env-overrides\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718758 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-tuned\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718787 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-host\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718812 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718840 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718863 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-host\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718939 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-multus-certs\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718967 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-log-socket\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.718991 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-cni-netd\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:07.719011 2568 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719019 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-device-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719043 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-modprobe-d\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:07.719088 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs podName:d439efe8-c633-4c7c-ad01-515f448307fd nodeName:}" failed. No retries permitted until 2026-04-21 00:03:08.219058577 +0000 UTC m=+2.044037543 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs") pod "network-metrics-daemon-pf4ls" (UID: "d439efe8-c633-4c7c-ad01-515f448307fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:07.720262 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719108 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-device-dir\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719135 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-modprobe-d\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719169 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-run-netns\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719250 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cni-binary-copy\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719295 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719306 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719353 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7qpv\" (UniqueName: \"kubernetes.io/projected/d439efe8-c633-4c7c-ad01-515f448307fd-kube-api-access-l7qpv\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719380 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-daemon-config\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719783 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-cni-binary-copy\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.720751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.719892 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.720991 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.720862 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-tmp\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.720991 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.720887 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-etc-tuned\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.725292 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.725266 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qghgd\" (UniqueName: \"kubernetes.io/projected/dcfdee3f-a46f-46a3-aaf5-68b7638cc717-kube-api-access-qghgd\") pod \"multus-additional-cni-plugins-rm7p5\" (UID: \"dcfdee3f-a46f-46a3-aaf5-68b7638cc717\") " pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.725379 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.725299 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdzpr\" (UniqueName: \"kubernetes.io/projected/ef105ba8-c06b-4056-a230-9a44e0e3f1b9-kube-api-access-jdzpr\") pod \"node-resolver-wx87b\" (UID: \"ef105ba8-c06b-4056-a230-9a44e0e3f1b9\") " pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.725818 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.725799 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr8kc\" (UniqueName: \"kubernetes.io/projected/cde5a5a1-f134-48c1-be68-a8c2a968e9b5-kube-api-access-sr8kc\") pod \"node-ca-bscx2\" (UID: \"cde5a5a1-f134-48c1-be68-a8c2a968e9b5\") " pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.726139 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.726121 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm6gl\" (UniqueName: \"kubernetes.io/projected/e40ce47d-6c1e-40cc-aa93-65c7a44b897b-kube-api-access-nm6gl\") pod \"tuned-7vg48\" (UID: \"e40ce47d-6c1e-40cc-aa93-65c7a44b897b\") " pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.726949 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.726930 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlfmd\" (UniqueName: \"kubernetes.io/projected/b6e1f3f0-1a93-4f93-b178-3cfbddb84192-kube-api-access-hlfmd\") pod \"aws-ebs-csi-driver-node-c58z7\" (UID: \"b6e1f3f0-1a93-4f93-b178-3cfbddb84192\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.726995 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.726932 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7qpv\" (UniqueName: \"kubernetes.io/projected/d439efe8-c633-4c7c-ad01-515f448307fd-kube-api-access-l7qpv\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:07.820217 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820196 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-etc-kubernetes\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820217 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820165 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-etc-kubernetes\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820322 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820236 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:07.820322 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820258 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-cni-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820322 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820283 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/b470fb01-818e-4ce0-8ff7-35c94e72e940-agent-certs\") pod \"konnectivity-agent-lmssf\" (UID: \"b470fb01-818e-4ce0-8ff7-35c94e72e940\") " pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:07.820322 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820306 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-os-release\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820464 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820328 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-cni-multus\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820464 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820353 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-t6wnj\" (UniqueName: \"kubernetes.io/projected/18a5a55e-ea26-46ec-a17f-1cac467f786c-kube-api-access-t6wnj\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820464 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820377 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/1c09a22d-4059-455d-908f-0b608d24e6bb-iptables-alerter-script\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.820464 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820399 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-cni-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820464 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820414 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-cni-multus\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820464 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820407 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-slash\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820464 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820446 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-slash\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820472 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-etc-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820482 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-os-release\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820500 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovnkube-config\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820525 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-kubelet\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820526 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-etc-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820567 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c09a22d-4059-455d-908f-0b608d24e6bb-host-slash\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820587 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-kubelet\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820596 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-kubelet\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820639 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-kubelet\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820652 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-node-log\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820655 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c09a22d-4059-455d-908f-0b608d24e6bb-host-slash\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820680 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-system-cni-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820704 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-cnibin\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820729 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-node-log\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820730 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-k8s-cni-cncf-io\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820738 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-system-cni-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820762 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovnkube-script-lib\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.820783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820764 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-k8s-cni-cncf-io\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820785 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-cnibin\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820808 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/b470fb01-818e-4ce0-8ff7-35c94e72e940-konnectivity-ca\") pod \"konnectivity-agent-lmssf\" (UID: \"b470fb01-818e-4ce0-8ff7-35c94e72e940\") " pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820826 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-ovn\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820841 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/18a5a55e-ea26-46ec-a17f-1cac467f786c-cni-binary-copy\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820856 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-hostroot\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820873 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/1c09a22d-4059-455d-908f-0b608d24e6bb-iptables-alerter-script\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820893 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-hostroot\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820927 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-ovn\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820954 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-var-lib-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.820977 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821000 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovn-node-metrics-cert\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821025 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hcdbj\" (UniqueName: \"kubernetes.io/projected/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-kube-api-access-hcdbj\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821070 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-var-lib-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821119 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-netns\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821145 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-cni-bin\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821174 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-run-ovn-kubernetes\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821201 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-systemd-units\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.821598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821225 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-conf-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821228 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovnkube-config\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821249 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cqgnt\" (UniqueName: \"kubernetes.io/projected/1c09a22d-4059-455d-908f-0b608d24e6bb-kube-api-access-cqgnt\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821274 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-cni-bin\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821291 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-netns\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821293 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovnkube-script-lib\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821302 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-socket-dir-parent\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821328 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-systemd\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821333 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-conf-dir\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821334 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-var-lib-cni-bin\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821355 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821368 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-cni-bin\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821382 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-env-overrides\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821391 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-systemd-units\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821406 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/b470fb01-818e-4ce0-8ff7-35c94e72e940-konnectivity-ca\") pod \"konnectivity-agent-lmssf\" (UID: \"b470fb01-818e-4ce0-8ff7-35c94e72e940\") " pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821435 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-multus-certs\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821461 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-log-socket\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821463 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-systemd\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821498 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/18a5a55e-ea26-46ec-a17f-1cac467f786c-cni-binary-copy\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821515 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-log-socket\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821528 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-cni-netd\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821552 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-host-run-multus-certs\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821572 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-cni-netd\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821584 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-run-openvswitch\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821597 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821633 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-run-netns\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821644 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-socket-dir-parent\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821691 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-run-ovn-kubernetes\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821693 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-host-run-netns\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821758 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-daemon-config\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.821787 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-env-overrides\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.822068 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/18a5a55e-ea26-46ec-a17f-1cac467f786c-multus-daemon-config\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.822719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.822625 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/b470fb01-818e-4ce0-8ff7-35c94e72e940-agent-certs\") pod \"konnectivity-agent-lmssf\" (UID: \"b470fb01-818e-4ce0-8ff7-35c94e72e940\") " pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:07.823201 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.823187 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-ovn-node-metrics-cert\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.826653 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:07.826639 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 00:03:07.826699 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:07.826655 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 00:03:07.826699 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:07.826670 2568 projected.go:194] Error preparing data for projected volume kube-api-access-dnd8l for pod openshift-network-diagnostics/network-check-target-4tj2p: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:07.826761 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:07.826736 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l podName:0b745e1d-cf68-43fd-b452-36e45a05eb93 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:08.326723803 +0000 UTC m=+2.151702760 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-dnd8l" (UniqueName: "kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l") pod "network-check-target-4tj2p" (UID: "0b745e1d-cf68-43fd-b452-36e45a05eb93") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:07.828219 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.828202 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6wnj\" (UniqueName: \"kubernetes.io/projected/18a5a55e-ea26-46ec-a17f-1cac467f786c-kube-api-access-t6wnj\") pod \"multus-g6sk6\" (UID: \"18a5a55e-ea26-46ec-a17f-1cac467f786c\") " pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.829440 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.829422 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqgnt\" (UniqueName: \"kubernetes.io/projected/1c09a22d-4059-455d-908f-0b608d24e6bb-kube-api-access-cqgnt\") pod \"iptables-alerter-8cccm\" (UID: \"1c09a22d-4059-455d-908f-0b608d24e6bb\") " pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.829874 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.829860 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcdbj\" (UniqueName: \"kubernetes.io/projected/89e3cc31-318f-4b7b-9926-eebe7c5e2e88-kube-api-access-hcdbj\") pod \"ovnkube-node-z7lpw\" (UID: \"89e3cc31-318f-4b7b-9926-eebe7c5e2e88\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.932042 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.932022 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" Apr 21 00:03:07.937861 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.937840 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6e1f3f0_1a93_4f93_b178_3cfbddb84192.slice/crio-0e1f6f483cd6fafafe0c5f3ad4cee74731c8df2d7cc5bc8bee686d1faf1c0f6e WatchSource:0}: Error finding container 0e1f6f483cd6fafafe0c5f3ad4cee74731c8df2d7cc5bc8bee686d1faf1c0f6e: Status 404 returned error can't find the container with id 0e1f6f483cd6fafafe0c5f3ad4cee74731c8df2d7cc5bc8bee686d1faf1c0f6e Apr 21 00:03:07.947757 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.947742 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-7vg48" Apr 21 00:03:07.953325 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.953305 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode40ce47d_6c1e_40cc_aa93_65c7a44b897b.slice/crio-034ae1e95275f95bdf7337a699e9ac7ecfd36cb258376ecce288b0e91dbc0d6c WatchSource:0}: Error finding container 034ae1e95275f95bdf7337a699e9ac7ecfd36cb258376ecce288b0e91dbc0d6c: Status 404 returned error can't find the container with id 034ae1e95275f95bdf7337a699e9ac7ecfd36cb258376ecce288b0e91dbc0d6c Apr 21 00:03:07.953767 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.953744 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wx87b" Apr 21 00:03:07.958671 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.958653 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bscx2" Apr 21 00:03:07.959724 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.959705 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef105ba8_c06b_4056_a230_9a44e0e3f1b9.slice/crio-735854c30bd13ea9aaf6be026d2bafdc5a59be198e665c7ff30be68440809be9 WatchSource:0}: Error finding container 735854c30bd13ea9aaf6be026d2bafdc5a59be198e665c7ff30be68440809be9: Status 404 returned error can't find the container with id 735854c30bd13ea9aaf6be026d2bafdc5a59be198e665c7ff30be68440809be9 Apr 21 00:03:07.963774 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.963756 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" Apr 21 00:03:07.966711 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.966695 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcde5a5a1_f134_48c1_be68_a8c2a968e9b5.slice/crio-95b9209b9d28d6b45bbeae4b8a7210a58bbc1b66e382a652dee4e6cfccddc58d WatchSource:0}: Error finding container 95b9209b9d28d6b45bbeae4b8a7210a58bbc1b66e382a652dee4e6cfccddc58d: Status 404 returned error can't find the container with id 95b9209b9d28d6b45bbeae4b8a7210a58bbc1b66e382a652dee4e6cfccddc58d Apr 21 00:03:07.970408 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.970358 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:07.970600 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.970579 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcfdee3f_a46f_46a3_aaf5_68b7638cc717.slice/crio-23ddd71a13f059be03062754d4e5634fbd6d4f33afd5870873bebee4a99fe0a3 WatchSource:0}: Error finding container 23ddd71a13f059be03062754d4e5634fbd6d4f33afd5870873bebee4a99fe0a3: Status 404 returned error can't find the container with id 23ddd71a13f059be03062754d4e5634fbd6d4f33afd5870873bebee4a99fe0a3 Apr 21 00:03:07.976110 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.976093 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-g6sk6" Apr 21 00:03:07.976839 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.976821 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb470fb01_818e_4ce0_8ff7_35c94e72e940.slice/crio-3502b2498ab6416ee1d9ce99e0aabeffce6777ff125318b6c5d12d42fb7a6e41 WatchSource:0}: Error finding container 3502b2498ab6416ee1d9ce99e0aabeffce6777ff125318b6c5d12d42fb7a6e41: Status 404 returned error can't find the container with id 3502b2498ab6416ee1d9ce99e0aabeffce6777ff125318b6c5d12d42fb7a6e41 Apr 21 00:03:07.981701 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.981670 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18a5a55e_ea26_46ec_a17f_1cac467f786c.slice/crio-5b3f96ab4179e81614cc3e46e910b826486401d8ed3265a836adb189c56d06b6 WatchSource:0}: Error finding container 5b3f96ab4179e81614cc3e46e910b826486401d8ed3265a836adb189c56d06b6: Status 404 returned error can't find the container with id 5b3f96ab4179e81614cc3e46e910b826486401d8ed3265a836adb189c56d06b6 Apr 21 00:03:07.982414 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.982394 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-8cccm" Apr 21 00:03:07.986965 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:07.986948 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:07.991137 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.991115 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c09a22d_4059_455d_908f_0b608d24e6bb.slice/crio-5b84488057973af2274faa937a2a8324f4dcea0552721550ca609a9acf1d2819 WatchSource:0}: Error finding container 5b84488057973af2274faa937a2a8324f4dcea0552721550ca609a9acf1d2819: Status 404 returned error can't find the container with id 5b84488057973af2274faa937a2a8324f4dcea0552721550ca609a9acf1d2819 Apr 21 00:03:07.993532 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:07.993513 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89e3cc31_318f_4b7b_9926_eebe7c5e2e88.slice/crio-c29c97169835b2a5f72a89a9d564690e6647f37da5e1b946a22eb67b20e3189f WatchSource:0}: Error finding container c29c97169835b2a5f72a89a9d564690e6647f37da5e1b946a22eb67b20e3189f: Status 404 returned error can't find the container with id c29c97169835b2a5f72a89a9d564690e6647f37da5e1b946a22eb67b20e3189f Apr 21 00:03:08.224735 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.224660 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:08.224890 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:08.224818 2568 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:08.224890 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:08.224879 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs podName:d439efe8-c633-4c7c-ad01-515f448307fd nodeName:}" failed. No retries permitted until 2026-04-21 00:03:09.224860029 +0000 UTC m=+3.049838975 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs") pod "network-metrics-daemon-pf4ls" (UID: "d439efe8-c633-4c7c-ad01-515f448307fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:08.381341 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.381308 2568 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 00:03:08.397848 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.397825 2568 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 00:03:08.427129 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.426584 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:08.427129 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:08.426768 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 00:03:08.427129 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:08.426786 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 00:03:08.427129 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:08.426798 2568 projected.go:194] Error preparing data for projected volume kube-api-access-dnd8l for pod openshift-network-diagnostics/network-check-target-4tj2p: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:08.427129 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:08.426847 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l podName:0b745e1d-cf68-43fd-b452-36e45a05eb93 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:09.426829876 +0000 UTC m=+3.251808821 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-dnd8l" (UniqueName: "kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l") pod "network-check-target-4tj2p" (UID: "0b745e1d-cf68-43fd-b452-36e45a05eb93") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:08.658437 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.658358 2568 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-19 23:58:07 +0000 UTC" deadline="2027-11-28 13:50:39.71877787 +0000 UTC" Apr 21 00:03:08.658437 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.658393 2568 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="14077h47m31.060388559s" Apr 21 00:03:08.716015 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.715989 2568 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 00:03:08.736424 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.736394 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"c29c97169835b2a5f72a89a9d564690e6647f37da5e1b946a22eb67b20e3189f"} Apr 21 00:03:08.750884 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.750547 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-8cccm" event={"ID":"1c09a22d-4059-455d-908f-0b608d24e6bb","Type":"ContainerStarted","Data":"5b84488057973af2274faa937a2a8324f4dcea0552721550ca609a9acf1d2819"} Apr 21 00:03:08.778756 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.778732 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6sk6" event={"ID":"18a5a55e-ea26-46ec-a17f-1cac467f786c","Type":"ContainerStarted","Data":"5b3f96ab4179e81614cc3e46e910b826486401d8ed3265a836adb189c56d06b6"} Apr 21 00:03:08.791254 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.791228 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerStarted","Data":"23ddd71a13f059be03062754d4e5634fbd6d4f33afd5870873bebee4a99fe0a3"} Apr 21 00:03:08.813493 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.813458 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bscx2" event={"ID":"cde5a5a1-f134-48c1-be68-a8c2a968e9b5","Type":"ContainerStarted","Data":"95b9209b9d28d6b45bbeae4b8a7210a58bbc1b66e382a652dee4e6cfccddc58d"} Apr 21 00:03:08.824176 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.824146 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-lmssf" event={"ID":"b470fb01-818e-4ce0-8ff7-35c94e72e940","Type":"ContainerStarted","Data":"3502b2498ab6416ee1d9ce99e0aabeffce6777ff125318b6c5d12d42fb7a6e41"} Apr 21 00:03:08.843311 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.843285 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wx87b" event={"ID":"ef105ba8-c06b-4056-a230-9a44e0e3f1b9","Type":"ContainerStarted","Data":"735854c30bd13ea9aaf6be026d2bafdc5a59be198e665c7ff30be68440809be9"} Apr 21 00:03:08.861709 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.861680 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-7vg48" event={"ID":"e40ce47d-6c1e-40cc-aa93-65c7a44b897b","Type":"ContainerStarted","Data":"034ae1e95275f95bdf7337a699e9ac7ecfd36cb258376ecce288b0e91dbc0d6c"} Apr 21 00:03:08.868747 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:08.868725 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" event={"ID":"b6e1f3f0-1a93-4f93-b178-3cfbddb84192","Type":"ContainerStarted","Data":"0e1f6f483cd6fafafe0c5f3ad4cee74731c8df2d7cc5bc8bee686d1faf1c0f6e"} Apr 21 00:03:09.232398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:09.232362 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:09.232562 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:09.232538 2568 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:09.232659 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:09.232597 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs podName:d439efe8-c633-4c7c-ad01-515f448307fd nodeName:}" failed. No retries permitted until 2026-04-21 00:03:11.232579472 +0000 UTC m=+5.057558423 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs") pod "network-metrics-daemon-pf4ls" (UID: "d439efe8-c633-4c7c-ad01-515f448307fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:09.434089 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:09.433467 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:09.434089 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:09.433651 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 00:03:09.434089 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:09.433670 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 00:03:09.434089 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:09.433682 2568 projected.go:194] Error preparing data for projected volume kube-api-access-dnd8l for pod openshift-network-diagnostics/network-check-target-4tj2p: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:09.434089 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:09.433736 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l podName:0b745e1d-cf68-43fd-b452-36e45a05eb93 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:11.433718878 +0000 UTC m=+5.258697835 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-dnd8l" (UniqueName: "kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l") pod "network-check-target-4tj2p" (UID: "0b745e1d-cf68-43fd-b452-36e45a05eb93") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:09.659592 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:09.659505 2568 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-19 23:58:07 +0000 UTC" deadline="2027-11-09 05:30:02.944313103 +0000 UTC" Apr 21 00:03:09.659592 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:09.659538 2568 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="13613h26m53.284779901s" Apr 21 00:03:09.715425 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:09.715399 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:09.715564 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:09.715511 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:09.715663 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:09.715636 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:09.715778 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:09.715749 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:11.250079 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:11.249782 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:11.250079 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:11.250052 2568 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:11.250573 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:11.250115 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs podName:d439efe8-c633-4c7c-ad01-515f448307fd nodeName:}" failed. No retries permitted until 2026-04-21 00:03:15.25009998 +0000 UTC m=+9.075078929 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs") pod "network-metrics-daemon-pf4ls" (UID: "d439efe8-c633-4c7c-ad01-515f448307fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:11.451482 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:11.451448 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:11.451669 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:11.451631 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 00:03:11.451669 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:11.451652 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 00:03:11.451669 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:11.451665 2568 projected.go:194] Error preparing data for projected volume kube-api-access-dnd8l for pod openshift-network-diagnostics/network-check-target-4tj2p: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:11.451831 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:11.451726 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l podName:0b745e1d-cf68-43fd-b452-36e45a05eb93 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:15.451707341 +0000 UTC m=+9.276686291 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-dnd8l" (UniqueName: "kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l") pod "network-check-target-4tj2p" (UID: "0b745e1d-cf68-43fd-b452-36e45a05eb93") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:11.715253 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:11.714755 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:11.715253 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:11.714883 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:11.715253 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:11.714952 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:11.715253 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:11.715059 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:13.715448 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:13.715097 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:13.715448 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:13.715226 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:13.715448 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:13.715250 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:13.715448 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:13.715364 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:15.283741 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:15.283704 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:15.284181 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:15.283895 2568 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:15.284181 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:15.283965 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs podName:d439efe8-c633-4c7c-ad01-515f448307fd nodeName:}" failed. No retries permitted until 2026-04-21 00:03:23.283944342 +0000 UTC m=+17.108923294 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs") pod "network-metrics-daemon-pf4ls" (UID: "d439efe8-c633-4c7c-ad01-515f448307fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:15.485908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:15.485847 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:15.486109 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:15.486064 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 00:03:15.486109 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:15.486085 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 00:03:15.486109 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:15.486099 2568 projected.go:194] Error preparing data for projected volume kube-api-access-dnd8l for pod openshift-network-diagnostics/network-check-target-4tj2p: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:15.486307 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:15.486160 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l podName:0b745e1d-cf68-43fd-b452-36e45a05eb93 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:23.486138769 +0000 UTC m=+17.311117718 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-dnd8l" (UniqueName: "kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l") pod "network-check-target-4tj2p" (UID: "0b745e1d-cf68-43fd-b452-36e45a05eb93") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:15.715088 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:15.715022 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:15.715293 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:15.715037 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:15.715293 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:15.715146 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:15.715293 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:15.715235 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:17.715112 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:17.715082 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:17.715489 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:17.715092 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:17.715489 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:17.715211 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:17.715489 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:17.715304 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:19.715460 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:19.715426 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:19.715973 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:19.715430 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:19.715973 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:19.715671 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:19.715973 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:19.715535 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:21.714883 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:21.714855 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:21.715256 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:21.714884 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:21.715256 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:21.714957 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:21.715256 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:21.715086 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:23.351077 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:23.351039 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:23.351574 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:23.351204 2568 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:23.351574 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:23.351272 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs podName:d439efe8-c633-4c7c-ad01-515f448307fd nodeName:}" failed. No retries permitted until 2026-04-21 00:03:39.351251368 +0000 UTC m=+33.176230334 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs") pod "network-metrics-daemon-pf4ls" (UID: "d439efe8-c633-4c7c-ad01-515f448307fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:23.552468 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:23.552432 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:23.552659 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:23.552558 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 00:03:23.552659 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:23.552573 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 00:03:23.552659 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:23.552582 2568 projected.go:194] Error preparing data for projected volume kube-api-access-dnd8l for pod openshift-network-diagnostics/network-check-target-4tj2p: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:23.552659 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:23.552644 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l podName:0b745e1d-cf68-43fd-b452-36e45a05eb93 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:39.55262768 +0000 UTC m=+33.377606643 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-dnd8l" (UniqueName: "kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l") pod "network-check-target-4tj2p" (UID: "0b745e1d-cf68-43fd-b452-36e45a05eb93") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:23.715420 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:23.715345 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:23.715420 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:23.715368 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:23.715646 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:23.715466 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:23.715713 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:23.715635 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:25.715242 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:25.715216 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:25.715500 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:25.715216 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:25.715500 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:25.715314 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:25.715500 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:25.715390 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:26.911161 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.910797 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wx87b" event={"ID":"ef105ba8-c06b-4056-a230-9a44e0e3f1b9","Type":"ContainerStarted","Data":"b32cb5f0438a5b28233d66e755e3f4315c512dad7d963ce95ba817b63e6732b8"} Apr 21 00:03:26.912092 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.912071 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-7vg48" event={"ID":"e40ce47d-6c1e-40cc-aa93-65c7a44b897b","Type":"ContainerStarted","Data":"cbef59da5d55540008e441b78859bdbdd257a3baabcbfd98df2a6d5aa66bc230"} Apr 21 00:03:26.913258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.913237 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" event={"ID":"b6e1f3f0-1a93-4f93-b178-3cfbddb84192","Type":"ContainerStarted","Data":"ddc01f8bf28ab5d9cee3358e7c162b8699573bdd8ac9e3d796f6669364504b97"} Apr 21 00:03:26.914357 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.914338 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" event={"ID":"7b08454bff9ca2a8282774b411985429","Type":"ContainerStarted","Data":"f3b4e386c8a6877dab707a244697dc9082c26578386e7403a299d159ba902c29"} Apr 21 00:03:26.916488 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.916472 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:03:26.916759 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.916742 2568 generic.go:358] "Generic (PLEG): container finished" podID="89e3cc31-318f-4b7b-9926-eebe7c5e2e88" containerID="ea83206f1e719fb62539ceb257d5488e3a95559d1350b74c697ea7bb322ccaae" exitCode=1 Apr 21 00:03:26.916817 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.916792 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"d65530bc441470c914779e968da9a3c9ea3ed72d0ac2fbe2ec7332592f5b5d77"} Apr 21 00:03:26.916817 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.916811 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"71780ef60a6c6ff1dd4a073277b4d8bc78f76d64c08c25ea4e109848181ca64a"} Apr 21 00:03:26.916874 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.916822 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"c1693ebf02b1739666c719d58fc991581ed70f62b14cc340cd64375d179ecb1d"} Apr 21 00:03:26.916874 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.916830 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"cb8e2d5bf801a874d1a2bcdf7faeff9dab4e513611ebd781ac9ed3c71090f03e"} Apr 21 00:03:26.916874 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.916839 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerDied","Data":"ea83206f1e719fb62539ceb257d5488e3a95559d1350b74c697ea7bb322ccaae"} Apr 21 00:03:26.916874 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.916849 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"40f190256e293a47f2915d4b9dff90273d4bbe9ad1ca44940ec28e94bad740fa"} Apr 21 00:03:26.918000 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.917981 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6sk6" event={"ID":"18a5a55e-ea26-46ec-a17f-1cac467f786c","Type":"ContainerStarted","Data":"7797c4b183e16ec9b1d10b6fa6db87aecbefd506c03507a454768b2d2b286b5e"} Apr 21 00:03:26.919195 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.919177 2568 generic.go:358] "Generic (PLEG): container finished" podID="dcfdee3f-a46f-46a3-aaf5-68b7638cc717" containerID="9c618a0631a6ea02f02716f48535dc01018432ba3536ca3bc0edb3c61f0f1941" exitCode=0 Apr 21 00:03:26.919298 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.919226 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerDied","Data":"9c618a0631a6ea02f02716f48535dc01018432ba3536ca3bc0edb3c61f0f1941"} Apr 21 00:03:26.920444 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.920320 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bscx2" event={"ID":"cde5a5a1-f134-48c1-be68-a8c2a968e9b5","Type":"ContainerStarted","Data":"91e3b12a1cfcff7e7739d791a4d46b2a724adb726d493c77038eae6f5c93461d"} Apr 21 00:03:26.921540 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.921523 2568 generic.go:358] "Generic (PLEG): container finished" podID="fd14846dcd7d8f37d2ba1fd40c2a4cd6" containerID="a9f425181e3bf81d740cc641c1786a69b4dd7bc42c090fd0115e8450fd98913f" exitCode=0 Apr 21 00:03:26.921637 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.921571 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" event={"ID":"fd14846dcd7d8f37d2ba1fd40c2a4cd6","Type":"ContainerDied","Data":"a9f425181e3bf81d740cc641c1786a69b4dd7bc42c090fd0115e8450fd98913f"} Apr 21 00:03:26.922858 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.922841 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-lmssf" event={"ID":"b470fb01-818e-4ce0-8ff7-35c94e72e940","Type":"ContainerStarted","Data":"61d0f9f26362d57996b650bf83f4983c0812ec7cb8e2f8cd267fa0729ba4dbaa"} Apr 21 00:03:26.927206 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.927168 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-wx87b" podStartSLOduration=3.212491014 podStartE2EDuration="20.927157802s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.963541054 +0000 UTC m=+1.788519998" lastFinishedPulling="2026-04-21 00:03:25.678207829 +0000 UTC m=+19.503186786" observedRunningTime="2026-04-21 00:03:26.926653417 +0000 UTC m=+20.751632382" watchObservedRunningTime="2026-04-21 00:03:26.927157802 +0000 UTC m=+20.752136766" Apr 21 00:03:26.984589 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:26.984545 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-132-206.ec2.internal" podStartSLOduration=20.984531023 podStartE2EDuration="20.984531023s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:03:26.984529909 +0000 UTC m=+20.809508875" watchObservedRunningTime="2026-04-21 00:03:26.984531023 +0000 UTC m=+20.809509989" Apr 21 00:03:27.017682 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.017647 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-bscx2" podStartSLOduration=3.306978928 podStartE2EDuration="21.017636798s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.968371977 +0000 UTC m=+1.793350924" lastFinishedPulling="2026-04-21 00:03:25.679029844 +0000 UTC m=+19.504008794" observedRunningTime="2026-04-21 00:03:27.001390395 +0000 UTC m=+20.826369378" watchObservedRunningTime="2026-04-21 00:03:27.017636798 +0000 UTC m=+20.842615762" Apr 21 00:03:27.017819 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.017798 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-lmssf" podStartSLOduration=3.350839384 podStartE2EDuration="21.017794535s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.978980452 +0000 UTC m=+1.803959395" lastFinishedPulling="2026-04-21 00:03:25.645935597 +0000 UTC m=+19.470914546" observedRunningTime="2026-04-21 00:03:27.01748564 +0000 UTC m=+20.842464617" watchObservedRunningTime="2026-04-21 00:03:27.017794535 +0000 UTC m=+20.842773500" Apr 21 00:03:27.036900 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.036865 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-g6sk6" podStartSLOduration=3.207457599 podStartE2EDuration="21.036854242s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.985586686 +0000 UTC m=+1.810565633" lastFinishedPulling="2026-04-21 00:03:25.814983334 +0000 UTC m=+19.639962276" observedRunningTime="2026-04-21 00:03:27.036526257 +0000 UTC m=+20.861505223" watchObservedRunningTime="2026-04-21 00:03:27.036854242 +0000 UTC m=+20.861833206" Apr 21 00:03:27.053327 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.053295 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-7vg48" podStartSLOduration=3.352530191 podStartE2EDuration="21.053285001s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.954659249 +0000 UTC m=+1.779638192" lastFinishedPulling="2026-04-21 00:03:25.655414055 +0000 UTC m=+19.480393002" observedRunningTime="2026-04-21 00:03:27.052980438 +0000 UTC m=+20.877959405" watchObservedRunningTime="2026-04-21 00:03:27.053285001 +0000 UTC m=+20.878263963" Apr 21 00:03:27.715048 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.715017 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:27.715219 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:27.715139 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:27.715219 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.715206 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:27.715321 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:27.715303 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:27.926024 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.926002 2568 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 21 00:03:27.926855 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.926832 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" event={"ID":"fd14846dcd7d8f37d2ba1fd40c2a4cd6","Type":"ContainerStarted","Data":"f5d7cd037a5465f72839a9853a80b7da47f36781ec96bce0d21e7455adc7902e"} Apr 21 00:03:27.928793 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.928770 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" event={"ID":"b6e1f3f0-1a93-4f93-b178-3cfbddb84192","Type":"ContainerStarted","Data":"67740d4a26e0cda8044c78f2381545a4b67e460b2921c937ef6b2a067e5deabf"} Apr 21 00:03:27.930472 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.930424 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-8cccm" event={"ID":"1c09a22d-4059-455d-908f-0b608d24e6bb","Type":"ContainerStarted","Data":"a1e9fad931ade9746c902b158e754b3ecc3791a6f6f74cd802d1b8cfcaa98482"} Apr 21 00:03:27.946397 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.946359 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-132-206.ec2.internal" podStartSLOduration=21.946343991 podStartE2EDuration="21.946343991s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:03:27.945838947 +0000 UTC m=+21.770817915" watchObservedRunningTime="2026-04-21 00:03:27.946343991 +0000 UTC m=+21.771322958" Apr 21 00:03:27.962803 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:27.962754 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-8cccm" podStartSLOduration=4.287171357 podStartE2EDuration="21.962743054s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.992648131 +0000 UTC m=+1.817627074" lastFinishedPulling="2026-04-21 00:03:25.668219817 +0000 UTC m=+19.493198771" observedRunningTime="2026-04-21 00:03:27.962410341 +0000 UTC m=+21.787389305" watchObservedRunningTime="2026-04-21 00:03:27.962743054 +0000 UTC m=+21.787722022" Apr 21 00:03:28.421277 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.421244 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-kt2pp"] Apr 21 00:03:28.424836 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.424813 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.424965 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:28.424891 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-kt2pp" podUID="2f5feffc-f6d3-4684-8a39-a8fa87b1feda" Apr 21 00:03:28.492829 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.492791 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.492978 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.492843 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-kubelet-config\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.492978 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.492890 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-dbus\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.587999 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.587953 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:28.588640 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.588602 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:28.593599 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.593572 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-dbus\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.593727 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.593653 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.593727 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.593690 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-kubelet-config\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.593833 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.593780 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-kubelet-config\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.593985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.593965 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-dbus\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:28.593985 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:28.593975 2568 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:28.594139 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:28.594060 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret podName:2f5feffc-f6d3-4684-8a39-a8fa87b1feda nodeName:}" failed. No retries permitted until 2026-04-21 00:03:29.094041508 +0000 UTC m=+22.919020456 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret") pod "global-pull-secret-syncer-kt2pp" (UID: "2f5feffc-f6d3-4684-8a39-a8fa87b1feda") : object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:28.676078 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.675956 2568 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-21T00:03:27.92602066Z","UUID":"6db3e61a-ba0b-4e7a-a4a4-39ddaf4b6314","Handler":null,"Name":"","Endpoint":""} Apr 21 00:03:28.677632 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.677596 2568 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 21 00:03:28.677735 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.677642 2568 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 21 00:03:28.934721 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.934653 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:03:28.935086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.935046 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"8220e8266e72382c06061d8d88b368964c543bef02d75baafb7e00eb56b13b98"} Apr 21 00:03:28.936863 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.936804 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" event={"ID":"b6e1f3f0-1a93-4f93-b178-3cfbddb84192","Type":"ContainerStarted","Data":"53bc1e5b396702297334b82ce9ef59f34bb3e360047bf3c95c6083a2f5113c07"} Apr 21 00:03:28.937447 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.937422 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:28.937862 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.937840 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-lmssf" Apr 21 00:03:28.953288 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:28.953239 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-c58z7" podStartSLOduration=2.203454706 podStartE2EDuration="22.953222354s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.939335007 +0000 UTC m=+1.764313951" lastFinishedPulling="2026-04-21 00:03:28.689102641 +0000 UTC m=+22.514081599" observedRunningTime="2026-04-21 00:03:28.952922938 +0000 UTC m=+22.777901906" watchObservedRunningTime="2026-04-21 00:03:28.953222354 +0000 UTC m=+22.778201321" Apr 21 00:03:29.096851 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:29.096820 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:29.097025 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:29.096931 2568 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:29.097025 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:29.096986 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret podName:2f5feffc-f6d3-4684-8a39-a8fa87b1feda nodeName:}" failed. No retries permitted until 2026-04-21 00:03:30.096967995 +0000 UTC m=+23.921946941 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret") pod "global-pull-secret-syncer-kt2pp" (UID: "2f5feffc-f6d3-4684-8a39-a8fa87b1feda") : object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:29.714912 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:29.714879 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:29.715071 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:29.714903 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:29.715071 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:29.714989 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:29.715164 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:29.715122 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:30.105838 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:30.105747 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:30.106290 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:30.105982 2568 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:30.106290 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:30.106080 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret podName:2f5feffc-f6d3-4684-8a39-a8fa87b1feda nodeName:}" failed. No retries permitted until 2026-04-21 00:03:32.106058017 +0000 UTC m=+25.931036960 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret") pod "global-pull-secret-syncer-kt2pp" (UID: "2f5feffc-f6d3-4684-8a39-a8fa87b1feda") : object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:30.718125 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:30.717914 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:30.718312 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:30.718226 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-kt2pp" podUID="2f5feffc-f6d3-4684-8a39-a8fa87b1feda" Apr 21 00:03:30.944202 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:30.944018 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:03:30.944725 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:30.944524 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"49024a043af79e71ade609f84f0e77266499853fdc51b58106cff8c01a55b21b"} Apr 21 00:03:30.945222 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:30.945202 2568 scope.go:117] "RemoveContainer" containerID="ea83206f1e719fb62539ceb257d5488e3a95559d1350b74c697ea7bb322ccaae" Apr 21 00:03:31.409212 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.408998 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:31.715368 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.715292 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:31.715530 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.715297 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:31.715530 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:31.715401 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:31.715530 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:31.715457 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:31.947567 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.947536 2568 generic.go:358] "Generic (PLEG): container finished" podID="dcfdee3f-a46f-46a3-aaf5-68b7638cc717" containerID="bfaf12627838dfb99142250ee427c237ee39fc7b79030db0122b24106bc28267" exitCode=0 Apr 21 00:03:31.947718 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.947634 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerDied","Data":"bfaf12627838dfb99142250ee427c237ee39fc7b79030db0122b24106bc28267"} Apr 21 00:03:31.951220 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.951201 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:03:31.951518 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.951497 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" event={"ID":"89e3cc31-318f-4b7b-9926-eebe7c5e2e88","Type":"ContainerStarted","Data":"cb9eb8a6f731500ff8cc8aed88a55db2441ae3cd5c4ae9b0ca308112ce432405"} Apr 21 00:03:31.951777 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.951763 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:31.951856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.951787 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:31.966282 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.966216 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:31.966282 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:31.966274 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:03:32.119997 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.119961 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:32.120463 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:32.120442 2568 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:32.120669 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:32.120656 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret podName:2f5feffc-f6d3-4684-8a39-a8fa87b1feda nodeName:}" failed. No retries permitted until 2026-04-21 00:03:36.12063438 +0000 UTC m=+29.945613324 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret") pod "global-pull-secret-syncer-kt2pp" (UID: "2f5feffc-f6d3-4684-8a39-a8fa87b1feda") : object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:32.715580 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.715559 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:32.715880 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:32.715670 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-kt2pp" podUID="2f5feffc-f6d3-4684-8a39-a8fa87b1feda" Apr 21 00:03:32.806686 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.806643 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" podStartSLOduration=8.662176537 podStartE2EDuration="26.806627938s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.994957583 +0000 UTC m=+1.819936529" lastFinishedPulling="2026-04-21 00:03:26.139408974 +0000 UTC m=+19.964387930" observedRunningTime="2026-04-21 00:03:32.011200089 +0000 UTC m=+25.836179053" watchObservedRunningTime="2026-04-21 00:03:32.806627938 +0000 UTC m=+26.631606903" Apr 21 00:03:32.807100 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.807086 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-4tj2p"] Apr 21 00:03:32.807197 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.807179 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:32.807273 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:32.807257 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:32.810532 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.810512 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-kt2pp"] Apr 21 00:03:32.811162 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.811142 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-pf4ls"] Apr 21 00:03:32.811272 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.811260 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:32.811390 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:32.811369 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:32.954641 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.954573 2568 generic.go:358] "Generic (PLEG): container finished" podID="dcfdee3f-a46f-46a3-aaf5-68b7638cc717" containerID="0ed5832496c6fff1232d46fc45c1b6339058746a392c0e0e3d4badfb4649f451" exitCode=0 Apr 21 00:03:32.954732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.954641 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerDied","Data":"0ed5832496c6fff1232d46fc45c1b6339058746a392c0e0e3d4badfb4649f451"} Apr 21 00:03:32.954768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:32.954729 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:32.954857 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:32.954834 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-kt2pp" podUID="2f5feffc-f6d3-4684-8a39-a8fa87b1feda" Apr 21 00:03:33.958812 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:33.958744 2568 generic.go:358] "Generic (PLEG): container finished" podID="dcfdee3f-a46f-46a3-aaf5-68b7638cc717" containerID="06e9b6d1e9a5b964a33c27c362a70ec70f76ad4f8c6c38f5e6f17b4e75bc5c15" exitCode=0 Apr 21 00:03:33.959151 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:33.958835 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerDied","Data":"06e9b6d1e9a5b964a33c27c362a70ec70f76ad4f8c6c38f5e6f17b4e75bc5c15"} Apr 21 00:03:34.715435 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:34.715401 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:34.715590 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:34.715401 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:34.715590 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:34.715506 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:34.715716 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:34.715401 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:34.715716 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:34.715653 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:34.715716 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:34.715691 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-kt2pp" podUID="2f5feffc-f6d3-4684-8a39-a8fa87b1feda" Apr 21 00:03:36.154540 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:36.154326 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:36.154936 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:36.154471 2568 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:36.154936 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:36.154628 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret podName:2f5feffc-f6d3-4684-8a39-a8fa87b1feda nodeName:}" failed. No retries permitted until 2026-04-21 00:03:44.154593162 +0000 UTC m=+37.979572109 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret") pod "global-pull-secret-syncer-kt2pp" (UID: "2f5feffc-f6d3-4684-8a39-a8fa87b1feda") : object "kube-system"/"original-pull-secret" not registered Apr 21 00:03:36.715768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:36.715735 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:36.715768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:36.715769 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:36.715985 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:36.715812 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:36.715985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:36.715868 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:36.715985 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:36.715908 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:36.715985 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:36.715936 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-kt2pp" podUID="2f5feffc-f6d3-4684-8a39-a8fa87b1feda" Apr 21 00:03:38.714539 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:38.714509 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:38.714983 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:38.714515 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:38.714983 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:38.714655 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pf4ls" podUID="d439efe8-c633-4c7c-ad01-515f448307fd" Apr 21 00:03:38.714983 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:38.714526 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:38.714983 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:38.714719 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-4tj2p" podUID="0b745e1d-cf68-43fd-b452-36e45a05eb93" Apr 21 00:03:38.714983 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:38.714784 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-kt2pp" podUID="2f5feffc-f6d3-4684-8a39-a8fa87b1feda" Apr 21 00:03:39.014306 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.014281 2568 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-132-206.ec2.internal" event="NodeReady" Apr 21 00:03:39.014473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.014408 2568 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 21 00:03:39.047330 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.047298 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-65b58b5547-wd45g"] Apr 21 00:03:39.072004 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.071934 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-t4sw5"] Apr 21 00:03:39.072195 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.072051 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.075138 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.074856 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 21 00:03:39.075138 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.074893 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-pctqm\"" Apr 21 00:03:39.075138 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.074906 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 21 00:03:39.075138 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.074913 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 21 00:03:39.081868 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.081535 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 21 00:03:39.087542 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.087521 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pvnc8"] Apr 21 00:03:39.087687 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.087671 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.095788 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.095769 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-79cql\"" Apr 21 00:03:39.096655 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.096632 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 21 00:03:39.097345 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.097328 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 21 00:03:39.099461 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.099440 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-65b58b5547-wd45g"] Apr 21 00:03:39.099548 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.099513 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pvnc8"] Apr 21 00:03:39.099548 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.099545 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:39.099666 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.099546 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-t4sw5"] Apr 21 00:03:39.118859 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.118836 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 21 00:03:39.118954 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.118914 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 21 00:03:39.119033 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.119008 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-2glpj\"" Apr 21 00:03:39.119134 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.119091 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 21 00:03:39.178569 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.178544 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.178696 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.178573 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-certificates\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.178696 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.178593 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-bound-sa-token\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.178696 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.178632 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-installation-pull-secrets\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.178867 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.178730 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lk87\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-kube-api-access-5lk87\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.178867 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.178788 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-image-registry-private-configuration\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.178867 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.178816 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/37aff1f7-e5cb-4278-ab90-899ff0513a20-ca-trust-extracted\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.178867 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.178839 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-trusted-ca\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.279908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.279837 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.279908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.279895 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280108 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.279936 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-certificates\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280108 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.279956 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-bound-sa-token\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280108 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.279974 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7g9c\" (UniqueName: \"kubernetes.io/projected/fbbce771-9a06-4a5a-9dbf-526282060d89-kube-api-access-g7g9c\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:39.280108 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.279999 2568 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 00:03:39.280108 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.280020 2568 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-65b58b5547-wd45g: secret "image-registry-tls" not found Apr 21 00:03:39.280108 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.280079 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls podName:37aff1f7-e5cb-4278-ab90-899ff0513a20 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:39.780058819 +0000 UTC m=+33.605037777 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls") pod "image-registry-65b58b5547-wd45g" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20") : secret "image-registry-tls" not found Apr 21 00:03:39.280108 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280102 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-installation-pull-secrets\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280133 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/169c80c3-e9e6-400e-9412-bab767e5cb98-tmp-dir\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.280452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280160 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psnvx\" (UniqueName: \"kubernetes.io/projected/169c80c3-e9e6-400e-9412-bab767e5cb98-kube-api-access-psnvx\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.280452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280200 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5lk87\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-kube-api-access-5lk87\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280246 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-image-registry-private-configuration\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280274 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/37aff1f7-e5cb-4278-ab90-899ff0513a20-ca-trust-extracted\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280300 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-trusted-ca\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280325 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:39.280452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280355 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/169c80c3-e9e6-400e-9412-bab767e5cb98-config-volume\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.280903 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280567 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-certificates\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.280903 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.280746 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/37aff1f7-e5cb-4278-ab90-899ff0513a20-ca-trust-extracted\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.281474 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.281455 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-trusted-ca\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.284623 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.284582 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-image-registry-private-configuration\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.284623 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.284598 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-installation-pull-secrets\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.288428 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.288381 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-bound-sa-token\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.288559 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.288539 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lk87\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-kube-api-access-5lk87\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.381639 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.381587 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:39.381639 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.381644 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/169c80c3-e9e6-400e-9412-bab767e5cb98-config-volume\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.381835 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.381741 2568 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 00:03:39.381835 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.381792 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.381835 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.381806 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert podName:fbbce771-9a06-4a5a-9dbf-526282060d89 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:39.881789795 +0000 UTC m=+33.706768737 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert") pod "ingress-canary-pvnc8" (UID: "fbbce771-9a06-4a5a-9dbf-526282060d89") : secret "canary-serving-cert" not found Apr 21 00:03:39.381956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.381855 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g7g9c\" (UniqueName: \"kubernetes.io/projected/fbbce771-9a06-4a5a-9dbf-526282060d89-kube-api-access-g7g9c\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:39.381956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.381885 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/169c80c3-e9e6-400e-9412-bab767e5cb98-tmp-dir\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.381956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.381901 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-psnvx\" (UniqueName: \"kubernetes.io/projected/169c80c3-e9e6-400e-9412-bab767e5cb98-kube-api-access-psnvx\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.381956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.381932 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:39.382272 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.381857 2568 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 00:03:39.382272 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.382008 2568 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:39.382272 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.382036 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls podName:169c80c3-e9e6-400e-9412-bab767e5cb98 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:39.882014946 +0000 UTC m=+33.706993906 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls") pod "dns-default-t4sw5" (UID: "169c80c3-e9e6-400e-9412-bab767e5cb98") : secret "dns-default-metrics-tls" not found Apr 21 00:03:39.382272 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.382055 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs podName:d439efe8-c633-4c7c-ad01-515f448307fd nodeName:}" failed. No retries permitted until 2026-04-21 00:04:11.382045492 +0000 UTC m=+65.207024453 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs") pod "network-metrics-daemon-pf4ls" (UID: "d439efe8-c633-4c7c-ad01-515f448307fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 00:03:39.382272 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.382132 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/169c80c3-e9e6-400e-9412-bab767e5cb98-config-volume\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.382272 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.382201 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/169c80c3-e9e6-400e-9412-bab767e5cb98-tmp-dir\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.391870 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.391851 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-psnvx\" (UniqueName: \"kubernetes.io/projected/169c80c3-e9e6-400e-9412-bab767e5cb98-kube-api-access-psnvx\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.392165 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.392144 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7g9c\" (UniqueName: \"kubernetes.io/projected/fbbce771-9a06-4a5a-9dbf-526282060d89-kube-api-access-g7g9c\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:39.583573 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.583517 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:39.583702 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.583630 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 00:03:39.583702 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.583644 2568 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 00:03:39.583702 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.583653 2568 projected.go:194] Error preparing data for projected volume kube-api-access-dnd8l for pod openshift-network-diagnostics/network-check-target-4tj2p: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:39.583817 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.583705 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l podName:0b745e1d-cf68-43fd-b452-36e45a05eb93 nodeName:}" failed. No retries permitted until 2026-04-21 00:04:11.583691828 +0000 UTC m=+65.408670772 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-dnd8l" (UniqueName: "kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l") pod "network-check-target-4tj2p" (UID: "0b745e1d-cf68-43fd-b452-36e45a05eb93") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 00:03:39.784802 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.784767 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:39.785266 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.784882 2568 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 00:03:39.785266 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.784896 2568 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-65b58b5547-wd45g: secret "image-registry-tls" not found Apr 21 00:03:39.785266 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.784954 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls podName:37aff1f7-e5cb-4278-ab90-899ff0513a20 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:40.784940482 +0000 UTC m=+34.609919425 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls") pod "image-registry-65b58b5547-wd45g" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20") : secret "image-registry-tls" not found Apr 21 00:03:39.885568 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.885506 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:39.885568 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.885553 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:39.885812 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.885652 2568 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 00:03:39.885812 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.885668 2568 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 00:03:39.885812 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.885706 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert podName:fbbce771-9a06-4a5a-9dbf-526282060d89 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:40.885689969 +0000 UTC m=+34.710668914 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert") pod "ingress-canary-pvnc8" (UID: "fbbce771-9a06-4a5a-9dbf-526282060d89") : secret "canary-serving-cert" not found Apr 21 00:03:39.885812 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:39.885719 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls podName:169c80c3-e9e6-400e-9412-bab767e5cb98 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:40.885713512 +0000 UTC m=+34.710692455 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls") pod "dns-default-t4sw5" (UID: "169c80c3-e9e6-400e-9412-bab767e5cb98") : secret "dns-default-metrics-tls" not found Apr 21 00:03:39.973927 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:39.973899 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerStarted","Data":"077fb45660324b8dfe9943b693a252eb946bb9d49afc7ce8cb89820b9533c268"} Apr 21 00:03:40.715066 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.714970 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:03:40.715066 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.715002 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:40.715066 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.715026 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:03:40.717711 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.717679 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 21 00:03:40.717711 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.717681 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 21 00:03:40.718848 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.718825 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-24wfg\"" Apr 21 00:03:40.718975 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.718827 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 21 00:03:40.718975 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.718831 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 21 00:03:40.719093 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.719006 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-n52qc\"" Apr 21 00:03:40.792580 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.792556 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:40.792958 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:40.792698 2568 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 00:03:40.792958 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:40.792714 2568 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-65b58b5547-wd45g: secret "image-registry-tls" not found Apr 21 00:03:40.792958 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:40.792766 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls podName:37aff1f7-e5cb-4278-ab90-899ff0513a20 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:42.792748327 +0000 UTC m=+36.617727277 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls") pod "image-registry-65b58b5547-wd45g" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20") : secret "image-registry-tls" not found Apr 21 00:03:40.893443 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.893420 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:40.893537 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.893484 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:40.893574 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:40.893559 2568 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 00:03:40.893638 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:40.893559 2568 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 00:03:40.893679 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:40.893628 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls podName:169c80c3-e9e6-400e-9412-bab767e5cb98 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:42.893597846 +0000 UTC m=+36.718576789 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls") pod "dns-default-t4sw5" (UID: "169c80c3-e9e6-400e-9412-bab767e5cb98") : secret "dns-default-metrics-tls" not found Apr 21 00:03:40.893719 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:40.893690 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert podName:fbbce771-9a06-4a5a-9dbf-526282060d89 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:42.893668159 +0000 UTC m=+36.718647106 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert") pod "ingress-canary-pvnc8" (UID: "fbbce771-9a06-4a5a-9dbf-526282060d89") : secret "canary-serving-cert" not found Apr 21 00:03:40.977829 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.977809 2568 generic.go:358] "Generic (PLEG): container finished" podID="dcfdee3f-a46f-46a3-aaf5-68b7638cc717" containerID="077fb45660324b8dfe9943b693a252eb946bb9d49afc7ce8cb89820b9533c268" exitCode=0 Apr 21 00:03:40.977924 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:40.977862 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerDied","Data":"077fb45660324b8dfe9943b693a252eb946bb9d49afc7ce8cb89820b9533c268"} Apr 21 00:03:41.981912 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:41.981880 2568 generic.go:358] "Generic (PLEG): container finished" podID="dcfdee3f-a46f-46a3-aaf5-68b7638cc717" containerID="a88712ec1ab39aac418cfc3eadeffee5b076a3d9df845c96b26c303d4f46758d" exitCode=0 Apr 21 00:03:41.982336 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:41.981917 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerDied","Data":"a88712ec1ab39aac418cfc3eadeffee5b076a3d9df845c96b26c303d4f46758d"} Apr 21 00:03:42.807060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:42.806854 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:42.807199 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:42.806994 2568 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 00:03:42.807199 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:42.807117 2568 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-65b58b5547-wd45g: secret "image-registry-tls" not found Apr 21 00:03:42.807199 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:42.807174 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls podName:37aff1f7-e5cb-4278-ab90-899ff0513a20 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:46.807157794 +0000 UTC m=+40.632136741 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls") pod "image-registry-65b58b5547-wd45g" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20") : secret "image-registry-tls" not found Apr 21 00:03:42.908308 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:42.908283 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:42.908419 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:42.908323 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:42.908458 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:42.908416 2568 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 00:03:42.908491 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:42.908465 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert podName:fbbce771-9a06-4a5a-9dbf-526282060d89 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:46.908452857 +0000 UTC m=+40.733431800 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert") pod "ingress-canary-pvnc8" (UID: "fbbce771-9a06-4a5a-9dbf-526282060d89") : secret "canary-serving-cert" not found Apr 21 00:03:42.908528 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:42.908422 2568 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 00:03:42.908559 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:42.908544 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls podName:169c80c3-e9e6-400e-9412-bab767e5cb98 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:46.908533673 +0000 UTC m=+40.733512616 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls") pod "dns-default-t4sw5" (UID: "169c80c3-e9e6-400e-9412-bab767e5cb98") : secret "dns-default-metrics-tls" not found Apr 21 00:03:42.986492 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:42.986451 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" event={"ID":"dcfdee3f-a46f-46a3-aaf5-68b7638cc717","Type":"ContainerStarted","Data":"3d1510686717cc2183652d8610b75f1464cfa1948f5c53486de7b78e12ae8fca"} Apr 21 00:03:43.008450 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:43.008408 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rm7p5" podStartSLOduration=5.18218844 podStartE2EDuration="37.008395691s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:03:07.973686835 +0000 UTC m=+1.798665793" lastFinishedPulling="2026-04-21 00:03:39.799894101 +0000 UTC m=+33.624873044" observedRunningTime="2026-04-21 00:03:43.006653037 +0000 UTC m=+36.831632003" watchObservedRunningTime="2026-04-21 00:03:43.008395691 +0000 UTC m=+36.833374656" Apr 21 00:03:44.218056 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:44.218023 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:44.221109 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:44.221079 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/2f5feffc-f6d3-4684-8a39-a8fa87b1feda-original-pull-secret\") pod \"global-pull-secret-syncer-kt2pp\" (UID: \"2f5feffc-f6d3-4684-8a39-a8fa87b1feda\") " pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:44.325728 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:44.325698 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-kt2pp" Apr 21 00:03:44.484823 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:44.484764 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-kt2pp"] Apr 21 00:03:44.493812 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:44.493776 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f5feffc_f6d3_4684_8a39_a8fa87b1feda.slice/crio-011c9a6b10f09d46121039baf54e0dd3915e4d9230c7a88ca9b66d9195b4b4c7 WatchSource:0}: Error finding container 011c9a6b10f09d46121039baf54e0dd3915e4d9230c7a88ca9b66d9195b4b4c7: Status 404 returned error can't find the container with id 011c9a6b10f09d46121039baf54e0dd3915e4d9230c7a88ca9b66d9195b4b4c7 Apr 21 00:03:44.991713 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:44.991685 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-kt2pp" event={"ID":"2f5feffc-f6d3-4684-8a39-a8fa87b1feda","Type":"ContainerStarted","Data":"011c9a6b10f09d46121039baf54e0dd3915e4d9230c7a88ca9b66d9195b4b4c7"} Apr 21 00:03:46.839454 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:46.839248 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:46.839913 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:46.839409 2568 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 00:03:46.839913 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:46.839505 2568 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-65b58b5547-wd45g: secret "image-registry-tls" not found Apr 21 00:03:46.839913 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:46.839562 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls podName:37aff1f7-e5cb-4278-ab90-899ff0513a20 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:54.839542715 +0000 UTC m=+48.664521678 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls") pod "image-registry-65b58b5547-wd45g" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20") : secret "image-registry-tls" not found Apr 21 00:03:46.940621 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:46.940590 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:46.940797 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:46.940672 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:46.940797 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:46.940712 2568 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 00:03:46.940797 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:46.940749 2568 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 00:03:46.940797 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:46.940783 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert podName:fbbce771-9a06-4a5a-9dbf-526282060d89 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:54.940765144 +0000 UTC m=+48.765744100 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert") pod "ingress-canary-pvnc8" (UID: "fbbce771-9a06-4a5a-9dbf-526282060d89") : secret "canary-serving-cert" not found Apr 21 00:03:46.940954 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:46.940811 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls podName:169c80c3-e9e6-400e-9412-bab767e5cb98 nodeName:}" failed. No retries permitted until 2026-04-21 00:03:54.940799087 +0000 UTC m=+48.765778044 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls") pod "dns-default-t4sw5" (UID: "169c80c3-e9e6-400e-9412-bab767e5cb98") : secret "dns-default-metrics-tls" not found Apr 21 00:03:47.859184 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:47.859146 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg"] Apr 21 00:03:47.897285 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:47.897247 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg"] Apr 21 00:03:47.897458 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:47.897387 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" Apr 21 00:03:47.900056 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:47.899922 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Apr 21 00:03:47.900056 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:47.899949 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Apr 21 00:03:47.900056 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:47.899981 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-lljwn\"" Apr 21 00:03:48.048855 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:48.048823 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbs7k\" (UniqueName: \"kubernetes.io/projected/848545eb-5390-420a-8f8e-3b327e38e776-kube-api-access-lbs7k\") pod \"migrator-74bb7799d9-mfrlg\" (UID: \"848545eb-5390-420a-8f8e-3b327e38e776\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" Apr 21 00:03:48.149735 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:48.149658 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lbs7k\" (UniqueName: \"kubernetes.io/projected/848545eb-5390-420a-8f8e-3b327e38e776-kube-api-access-lbs7k\") pod \"migrator-74bb7799d9-mfrlg\" (UID: \"848545eb-5390-420a-8f8e-3b327e38e776\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" Apr 21 00:03:48.159203 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:48.159177 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbs7k\" (UniqueName: \"kubernetes.io/projected/848545eb-5390-420a-8f8e-3b327e38e776-kube-api-access-lbs7k\") pod \"migrator-74bb7799d9-mfrlg\" (UID: \"848545eb-5390-420a-8f8e-3b327e38e776\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" Apr 21 00:03:48.208658 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:48.208600 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" Apr 21 00:03:49.226048 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:49.226021 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg"] Apr 21 00:03:49.228508 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:49.228480 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod848545eb_5390_420a_8f8e_3b327e38e776.slice/crio-5ff9931b93170645aca0a37c21eb52c6bddf112460048e7703552e2464b47149 WatchSource:0}: Error finding container 5ff9931b93170645aca0a37c21eb52c6bddf112460048e7703552e2464b47149: Status 404 returned error can't find the container with id 5ff9931b93170645aca0a37c21eb52c6bddf112460048e7703552e2464b47149 Apr 21 00:03:50.002551 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.002516 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-kt2pp" event={"ID":"2f5feffc-f6d3-4684-8a39-a8fa87b1feda","Type":"ContainerStarted","Data":"96c3caf1f32338ffe290fc52e37de6ea1dd66bd2a77b4f6b88eddfdc35553b8f"} Apr 21 00:03:50.003441 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.003413 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" event={"ID":"848545eb-5390-420a-8f8e-3b327e38e776","Type":"ContainerStarted","Data":"5ff9931b93170645aca0a37c21eb52c6bddf112460048e7703552e2464b47149"} Apr 21 00:03:50.017568 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.017527 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-kt2pp" podStartSLOduration=16.970185008 podStartE2EDuration="22.0175177s" podCreationTimestamp="2026-04-21 00:03:28 +0000 UTC" firstStartedPulling="2026-04-21 00:03:44.495684097 +0000 UTC m=+38.320663040" lastFinishedPulling="2026-04-21 00:03:49.543016786 +0000 UTC m=+43.367995732" observedRunningTime="2026-04-21 00:03:50.017175564 +0000 UTC m=+43.842154529" watchObservedRunningTime="2026-04-21 00:03:50.0175177 +0000 UTC m=+43.842496662" Apr 21 00:03:50.534210 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.534183 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-wx87b_ef105ba8-c06b-4056-a230-9a44e0e3f1b9/dns-node-resolver/0.log" Apr 21 00:03:50.844587 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.844520 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-865cb79987-slq6f"] Apr 21 00:03:50.863486 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.863464 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-slq6f"] Apr 21 00:03:50.863628 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.863588 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:50.866342 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.866314 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Apr 21 00:03:50.866342 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.866332 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Apr 21 00:03:50.866505 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.866332 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Apr 21 00:03:50.867684 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.867667 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Apr 21 00:03:50.867684 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.867676 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-bqvw7\"" Apr 21 00:03:50.974210 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.974180 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dxsg\" (UniqueName: \"kubernetes.io/projected/1f4121b7-6318-482b-91e3-ed7dc979bd59-kube-api-access-9dxsg\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:50.974329 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.974247 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f4121b7-6318-482b-91e3-ed7dc979bd59-signing-cabundle\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:50.974329 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:50.974306 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f4121b7-6318-482b-91e3-ed7dc979bd59-signing-key\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:51.075216 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.075193 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9dxsg\" (UniqueName: \"kubernetes.io/projected/1f4121b7-6318-482b-91e3-ed7dc979bd59-kube-api-access-9dxsg\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:51.075323 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.075245 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f4121b7-6318-482b-91e3-ed7dc979bd59-signing-cabundle\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:51.075323 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.075276 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f4121b7-6318-482b-91e3-ed7dc979bd59-signing-key\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:51.075846 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.075828 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f4121b7-6318-482b-91e3-ed7dc979bd59-signing-cabundle\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:51.087407 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.087377 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f4121b7-6318-482b-91e3-ed7dc979bd59-signing-key\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:51.087493 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.087443 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dxsg\" (UniqueName: \"kubernetes.io/projected/1f4121b7-6318-482b-91e3-ed7dc979bd59-kube-api-access-9dxsg\") pod \"service-ca-865cb79987-slq6f\" (UID: \"1f4121b7-6318-482b-91e3-ed7dc979bd59\") " pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:51.171641 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.171619 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-slq6f" Apr 21 00:03:51.321471 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.321446 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-slq6f"] Apr 21 00:03:51.324180 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:03:51.324154 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f4121b7_6318_482b_91e3_ed7dc979bd59.slice/crio-df2d5138f828fba54656325036d607083b842627530c3a308d36959a4ab5c636 WatchSource:0}: Error finding container df2d5138f828fba54656325036d607083b842627530c3a308d36959a4ab5c636: Status 404 returned error can't find the container with id df2d5138f828fba54656325036d607083b842627530c3a308d36959a4ab5c636 Apr 21 00:03:51.735305 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:51.735278 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-bscx2_cde5a5a1-f134-48c1-be68-a8c2a968e9b5/node-ca/0.log" Apr 21 00:03:52.008841 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:52.008758 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" event={"ID":"848545eb-5390-420a-8f8e-3b327e38e776","Type":"ContainerStarted","Data":"76cf593c411cb906c3c4afca7bfdd675c34fe6f2cf1b64860f86b27707595b48"} Apr 21 00:03:52.008841 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:52.008799 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" event={"ID":"848545eb-5390-420a-8f8e-3b327e38e776","Type":"ContainerStarted","Data":"10b45e573bca470c7317a70229ec4c113d19e9f68a47db6b56b6197e17ac9f1a"} Apr 21 00:03:52.009854 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:52.009827 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-slq6f" event={"ID":"1f4121b7-6318-482b-91e3-ed7dc979bd59","Type":"ContainerStarted","Data":"df2d5138f828fba54656325036d607083b842627530c3a308d36959a4ab5c636"} Apr 21 00:03:52.025431 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:52.025383 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-mfrlg" podStartSLOduration=3.186059402 podStartE2EDuration="5.025368837s" podCreationTimestamp="2026-04-21 00:03:47 +0000 UTC" firstStartedPulling="2026-04-21 00:03:49.230412889 +0000 UTC m=+43.055391832" lastFinishedPulling="2026-04-21 00:03:51.069722308 +0000 UTC m=+44.894701267" observedRunningTime="2026-04-21 00:03:52.025165524 +0000 UTC m=+45.850144490" watchObservedRunningTime="2026-04-21 00:03:52.025368837 +0000 UTC m=+45.850347802" Apr 21 00:03:52.535099 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:52.535077 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-mfrlg_848545eb-5390-420a-8f8e-3b327e38e776/migrator/0.log" Apr 21 00:03:52.735184 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:52.735156 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-mfrlg_848545eb-5390-420a-8f8e-3b327e38e776/graceful-termination/0.log" Apr 21 00:03:54.015476 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:54.015399 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-slq6f" event={"ID":"1f4121b7-6318-482b-91e3-ed7dc979bd59","Type":"ContainerStarted","Data":"52f8ca4533bae8d1125e4c173cfc5673a895d51774147a8b5bcbd7bfc3711301"} Apr 21 00:03:54.031515 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:54.031464 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-865cb79987-slq6f" podStartSLOduration=1.617339619 podStartE2EDuration="4.031450261s" podCreationTimestamp="2026-04-21 00:03:50 +0000 UTC" firstStartedPulling="2026-04-21 00:03:51.325991793 +0000 UTC m=+45.150970736" lastFinishedPulling="2026-04-21 00:03:53.740102432 +0000 UTC m=+47.565081378" observedRunningTime="2026-04-21 00:03:54.030630933 +0000 UTC m=+47.855609896" watchObservedRunningTime="2026-04-21 00:03:54.031450261 +0000 UTC m=+47.856429225" Apr 21 00:03:54.901799 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:54.901764 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls\") pod \"image-registry-65b58b5547-wd45g\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:03:54.901953 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:54.901903 2568 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 00:03:54.901953 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:54.901922 2568 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-65b58b5547-wd45g: secret "image-registry-tls" not found Apr 21 00:03:54.902032 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:54.901972 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls podName:37aff1f7-e5cb-4278-ab90-899ff0513a20 nodeName:}" failed. No retries permitted until 2026-04-21 00:04:10.901957247 +0000 UTC m=+64.726936190 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls") pod "image-registry-65b58b5547-wd45g" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20") : secret "image-registry-tls" not found Apr 21 00:03:55.002556 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:55.002508 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:03:55.002716 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:03:55.002632 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:03:55.002716 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:55.002664 2568 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 00:03:55.002790 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:55.002725 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert podName:fbbce771-9a06-4a5a-9dbf-526282060d89 nodeName:}" failed. No retries permitted until 2026-04-21 00:04:11.002710144 +0000 UTC m=+64.827689087 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert") pod "ingress-canary-pvnc8" (UID: "fbbce771-9a06-4a5a-9dbf-526282060d89") : secret "canary-serving-cert" not found Apr 21 00:03:55.002790 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:55.002750 2568 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 00:03:55.002858 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:03:55.002799 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls podName:169c80c3-e9e6-400e-9412-bab767e5cb98 nodeName:}" failed. No retries permitted until 2026-04-21 00:04:11.002785812 +0000 UTC m=+64.827764755 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls") pod "dns-default-t4sw5" (UID: "169c80c3-e9e6-400e-9412-bab767e5cb98") : secret "dns-default-metrics-tls" not found Apr 21 00:04:03.968835 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:03.968808 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7lpw" Apr 21 00:04:08.633930 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.633886 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-kdsh6"] Apr 21 00:04:08.639522 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.639499 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.643762 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.643736 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-pjwht\"" Apr 21 00:04:08.643904 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.643829 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 21 00:04:08.643904 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.643861 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 21 00:04:08.643904 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.643881 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 21 00:04:08.644056 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.643916 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 21 00:04:08.650603 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.650580 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-kdsh6"] Apr 21 00:04:08.686404 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.686381 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-65b58b5547-wd45g"] Apr 21 00:04:08.686519 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:08.686503 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[registry-tls], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-image-registry/image-registry-65b58b5547-wd45g" podUID="37aff1f7-e5cb-4278-ab90-899ff0513a20" Apr 21 00:04:08.708121 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.708098 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/27d810d8-53e4-460c-8437-9151e193bcfe-data-volume\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.708217 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.708157 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/27d810d8-53e4-460c-8437-9151e193bcfe-crio-socket\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.708217 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.708174 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/27d810d8-53e4-460c-8437-9151e193bcfe-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.708313 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.708221 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/27d810d8-53e4-460c-8437-9151e193bcfe-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.708313 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.708243 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh9f8\" (UniqueName: \"kubernetes.io/projected/27d810d8-53e4-460c-8437-9151e193bcfe-kube-api-access-lh9f8\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.721702 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.721675 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-575d785c49-gbr8m"] Apr 21 00:04:08.724339 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.724323 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.738557 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.738539 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-575d785c49-gbr8m"] Apr 21 00:04:08.809050 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809023 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-bound-sa-token\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.809145 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809054 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/27d810d8-53e4-460c-8437-9151e193bcfe-data-volume\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.809145 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809082 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-registry-certificates\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.809145 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809118 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/27d810d8-53e4-460c-8437-9151e193bcfe-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.809145 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809137 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-ca-trust-extracted\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.809354 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809179 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-registry-tls\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.809354 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809210 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmzhz\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-kube-api-access-vmzhz\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.809354 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809254 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-image-registry-private-configuration\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.809354 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809287 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-trusted-ca\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.809354 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809336 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-installation-pull-secrets\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.809554 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809394 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/27d810d8-53e4-460c-8437-9151e193bcfe-crio-socket\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.809554 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809403 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/27d810d8-53e4-460c-8437-9151e193bcfe-data-volume\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.809554 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809429 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/27d810d8-53e4-460c-8437-9151e193bcfe-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.809554 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809461 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lh9f8\" (UniqueName: \"kubernetes.io/projected/27d810d8-53e4-460c-8437-9151e193bcfe-kube-api-access-lh9f8\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.809554 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809521 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/27d810d8-53e4-460c-8437-9151e193bcfe-crio-socket\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.809883 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.809864 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/27d810d8-53e4-460c-8437-9151e193bcfe-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.811406 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.811390 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/27d810d8-53e4-460c-8437-9151e193bcfe-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.822792 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.822774 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh9f8\" (UniqueName: \"kubernetes.io/projected/27d810d8-53e4-460c-8437-9151e193bcfe-kube-api-access-lh9f8\") pod \"insights-runtime-extractor-kdsh6\" (UID: \"27d810d8-53e4-460c-8437-9151e193bcfe\") " pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:08.910090 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910034 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-bound-sa-token\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910090 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910066 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-registry-certificates\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910090 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910086 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-ca-trust-extracted\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910260 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910137 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-registry-tls\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910260 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910167 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vmzhz\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-kube-api-access-vmzhz\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910260 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910195 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-image-registry-private-configuration\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910385 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910318 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-trusted-ca\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910393 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-installation-pull-secrets\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910669 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910642 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-ca-trust-extracted\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.910901 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.910870 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-registry-certificates\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.911241 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.911221 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-trusted-ca\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.912484 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.912465 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-image-registry-private-configuration\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.912647 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.912628 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-installation-pull-secrets\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.912703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.912659 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-registry-tls\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.917323 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.917292 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmzhz\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-kube-api-access-vmzhz\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.918034 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.918015 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1-bound-sa-token\") pod \"image-registry-575d785c49-gbr8m\" (UID: \"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1\") " pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:08.948807 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:08.948791 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-kdsh6" Apr 21 00:04:09.034468 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.034445 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-pctqm\"" Apr 21 00:04:09.042382 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.042361 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:09.045386 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.045363 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:04:09.050173 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.050154 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:04:09.059815 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.059771 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-kdsh6"] Apr 21 00:04:09.063160 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:09.063139 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27d810d8_53e4_460c_8437_9151e193bcfe.slice/crio-bfdecbfce12725abefcb3b86e93e0990ee3137fdeeb6895bc93587ff44d8ddaa WatchSource:0}: Error finding container bfdecbfce12725abefcb3b86e93e0990ee3137fdeeb6895bc93587ff44d8ddaa: Status 404 returned error can't find the container with id bfdecbfce12725abefcb3b86e93e0990ee3137fdeeb6895bc93587ff44d8ddaa Apr 21 00:04:09.112204 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112184 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lk87\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-kube-api-access-5lk87\") pod \"37aff1f7-e5cb-4278-ab90-899ff0513a20\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " Apr 21 00:04:09.112295 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112212 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/37aff1f7-e5cb-4278-ab90-899ff0513a20-ca-trust-extracted\") pod \"37aff1f7-e5cb-4278-ab90-899ff0513a20\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " Apr 21 00:04:09.112295 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112264 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-certificates\") pod \"37aff1f7-e5cb-4278-ab90-899ff0513a20\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " Apr 21 00:04:09.112398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112296 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-installation-pull-secrets\") pod \"37aff1f7-e5cb-4278-ab90-899ff0513a20\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " Apr 21 00:04:09.112398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112337 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-image-registry-private-configuration\") pod \"37aff1f7-e5cb-4278-ab90-899ff0513a20\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " Apr 21 00:04:09.112398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112375 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-bound-sa-token\") pod \"37aff1f7-e5cb-4278-ab90-899ff0513a20\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " Apr 21 00:04:09.112554 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112403 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-trusted-ca\") pod \"37aff1f7-e5cb-4278-ab90-899ff0513a20\" (UID: \"37aff1f7-e5cb-4278-ab90-899ff0513a20\") " Apr 21 00:04:09.112643 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112595 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37aff1f7-e5cb-4278-ab90-899ff0513a20-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "37aff1f7-e5cb-4278-ab90-899ff0513a20" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:04:09.112707 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.112682 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "37aff1f7-e5cb-4278-ab90-899ff0513a20" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:04:09.113039 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.113000 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "37aff1f7-e5cb-4278-ab90-899ff0513a20" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:04:09.114490 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.114417 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-kube-api-access-5lk87" (OuterVolumeSpecName: "kube-api-access-5lk87") pod "37aff1f7-e5cb-4278-ab90-899ff0513a20" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20"). InnerVolumeSpecName "kube-api-access-5lk87". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:04:09.114724 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.114699 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "37aff1f7-e5cb-4278-ab90-899ff0513a20" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:04:09.114828 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.114801 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "37aff1f7-e5cb-4278-ab90-899ff0513a20" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:04:09.114893 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.114880 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "37aff1f7-e5cb-4278-ab90-899ff0513a20" (UID: "37aff1f7-e5cb-4278-ab90-899ff0513a20"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:04:09.159272 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.159254 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-575d785c49-gbr8m"] Apr 21 00:04:09.161959 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:09.161898 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d4145a7_09e7_45ef_bfaa_d03f1f3c92f1.slice/crio-889a82f8dc2bad975f6232aa5190db89a2181c6bd1c8cacfa883d62a88a768d4 WatchSource:0}: Error finding container 889a82f8dc2bad975f6232aa5190db89a2181c6bd1c8cacfa883d62a88a768d4: Status 404 returned error can't find the container with id 889a82f8dc2bad975f6232aa5190db89a2181c6bd1c8cacfa883d62a88a768d4 Apr 21 00:04:09.213821 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.213799 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-5lk87\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-kube-api-access-5lk87\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:09.213895 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.213827 2568 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/37aff1f7-e5cb-4278-ab90-899ff0513a20-ca-trust-extracted\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:09.213895 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.213846 2568 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-certificates\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:09.213895 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.213860 2568 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-installation-pull-secrets\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:09.213895 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.213876 2568 reconciler_common.go:299] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/37aff1f7-e5cb-4278-ab90-899ff0513a20-image-registry-private-configuration\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:09.213895 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.213892 2568 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-bound-sa-token\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:09.214042 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.213905 2568 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37aff1f7-e5cb-4278-ab90-899ff0513a20-trusted-ca\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:09.968138 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.968104 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-6bcc868b7-g8wql"] Apr 21 00:04:09.994219 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.994200 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-6bcc868b7-g8wql"] Apr 21 00:04:09.994309 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.994299 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-6bcc868b7-g8wql" Apr 21 00:04:09.996986 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.996962 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 21 00:04:09.997084 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.997015 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-7vx2j\"" Apr 21 00:04:09.997084 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:09.996974 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 21 00:04:10.049044 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.049015 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-kdsh6" event={"ID":"27d810d8-53e4-460c-8437-9151e193bcfe","Type":"ContainerStarted","Data":"5e0ebd334bd9d587dd3db7a51328d02c6c024db3349354484c9520f33432c408"} Apr 21 00:04:10.049160 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.049053 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-kdsh6" event={"ID":"27d810d8-53e4-460c-8437-9151e193bcfe","Type":"ContainerStarted","Data":"8ddd67658bad74bb585995e80224215fbd2ce4e39580b9dc2426c69fe9b2aaa2"} Apr 21 00:04:10.049160 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.049068 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-kdsh6" event={"ID":"27d810d8-53e4-460c-8437-9151e193bcfe","Type":"ContainerStarted","Data":"bfdecbfce12725abefcb3b86e93e0990ee3137fdeeb6895bc93587ff44d8ddaa"} Apr 21 00:04:10.050332 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.050304 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-575d785c49-gbr8m" event={"ID":"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1","Type":"ContainerStarted","Data":"57374dc9fc118af5484c2cd0d10cf53181804ac706be62362b54b86b9db054b4"} Apr 21 00:04:10.050332 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.050321 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-65b58b5547-wd45g" Apr 21 00:04:10.050332 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.050335 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-575d785c49-gbr8m" event={"ID":"1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1","Type":"ContainerStarted","Data":"889a82f8dc2bad975f6232aa5190db89a2181c6bd1c8cacfa883d62a88a768d4"} Apr 21 00:04:10.050586 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.050503 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:10.067022 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.067000 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd"] Apr 21 00:04:10.077178 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.077111 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-575d785c49-gbr8m" podStartSLOduration=2.077097155 podStartE2EDuration="2.077097155s" podCreationTimestamp="2026-04-21 00:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:04:10.076237126 +0000 UTC m=+63.901216090" watchObservedRunningTime="2026-04-21 00:04:10.077097155 +0000 UTC m=+63.902076121" Apr 21 00:04:10.085662 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.085645 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd"] Apr 21 00:04:10.085738 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.085703 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" Apr 21 00:04:10.087819 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.087797 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-admission-webhook-tls\"" Apr 21 00:04:10.087969 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.087946 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-admission-webhook-dockercfg-7mf9g\"" Apr 21 00:04:10.102699 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.102676 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-65b58b5547-wd45g"] Apr 21 00:04:10.105799 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.105779 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-65b58b5547-wd45g"] Apr 21 00:04:10.120907 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.120533 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/791b8e39-beab-44fd-9b37-14969c3d1b15-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-hmmhd\" (UID: \"791b8e39-beab-44fd-9b37-14969c3d1b15\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" Apr 21 00:04:10.120907 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.120584 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvzkr\" (UniqueName: \"kubernetes.io/projected/92eb6528-431d-47f2-b2d9-729c971eb0c5-kube-api-access-kvzkr\") pod \"downloads-6bcc868b7-g8wql\" (UID: \"92eb6528-431d-47f2-b2d9-729c971eb0c5\") " pod="openshift-console/downloads-6bcc868b7-g8wql" Apr 21 00:04:10.221854 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.221828 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kvzkr\" (UniqueName: \"kubernetes.io/projected/92eb6528-431d-47f2-b2d9-729c971eb0c5-kube-api-access-kvzkr\") pod \"downloads-6bcc868b7-g8wql\" (UID: \"92eb6528-431d-47f2-b2d9-729c971eb0c5\") " pod="openshift-console/downloads-6bcc868b7-g8wql" Apr 21 00:04:10.221973 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.221955 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/791b8e39-beab-44fd-9b37-14969c3d1b15-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-hmmhd\" (UID: \"791b8e39-beab-44fd-9b37-14969c3d1b15\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" Apr 21 00:04:10.222026 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.222002 2568 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/37aff1f7-e5cb-4278-ab90-899ff0513a20-registry-tls\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:10.222086 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:10.222073 2568 secret.go:189] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Apr 21 00:04:10.222146 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:10.222135 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791b8e39-beab-44fd-9b37-14969c3d1b15-tls-certificates podName:791b8e39-beab-44fd-9b37-14969c3d1b15 nodeName:}" failed. No retries permitted until 2026-04-21 00:04:10.722115766 +0000 UTC m=+64.547094714 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/791b8e39-beab-44fd-9b37-14969c3d1b15-tls-certificates") pod "prometheus-operator-admission-webhook-57cf98b594-hmmhd" (UID: "791b8e39-beab-44fd-9b37-14969c3d1b15") : secret "prometheus-operator-admission-webhook-tls" not found Apr 21 00:04:10.230246 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.230226 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvzkr\" (UniqueName: \"kubernetes.io/projected/92eb6528-431d-47f2-b2d9-729c971eb0c5-kube-api-access-kvzkr\") pod \"downloads-6bcc868b7-g8wql\" (UID: \"92eb6528-431d-47f2-b2d9-729c971eb0c5\") " pod="openshift-console/downloads-6bcc868b7-g8wql" Apr 21 00:04:10.306519 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.306497 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-6bcc868b7-g8wql" Apr 21 00:04:10.419470 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.419445 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-6bcc868b7-g8wql"] Apr 21 00:04:10.435355 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:10.435321 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92eb6528_431d_47f2_b2d9_729c971eb0c5.slice/crio-d1bfca87b1456cdb67ed15c16fd2d447088379c051cf8182de3db702f8fd9981 WatchSource:0}: Error finding container d1bfca87b1456cdb67ed15c16fd2d447088379c051cf8182de3db702f8fd9981: Status 404 returned error can't find the container with id d1bfca87b1456cdb67ed15c16fd2d447088379c051cf8182de3db702f8fd9981 Apr 21 00:04:10.718848 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.718821 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37aff1f7-e5cb-4278-ab90-899ff0513a20" path="/var/lib/kubelet/pods/37aff1f7-e5cb-4278-ab90-899ff0513a20/volumes" Apr 21 00:04:10.725791 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.725765 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/791b8e39-beab-44fd-9b37-14969c3d1b15-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-hmmhd\" (UID: \"791b8e39-beab-44fd-9b37-14969c3d1b15\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" Apr 21 00:04:10.728214 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.728192 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/791b8e39-beab-44fd-9b37-14969c3d1b15-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-hmmhd\" (UID: \"791b8e39-beab-44fd-9b37-14969c3d1b15\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" Apr 21 00:04:10.994272 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:10.994206 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" Apr 21 00:04:11.027240 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.027212 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:04:11.027366 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.027259 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:04:11.030072 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.030037 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/169c80c3-e9e6-400e-9412-bab767e5cb98-metrics-tls\") pod \"dns-default-t4sw5\" (UID: \"169c80c3-e9e6-400e-9412-bab767e5cb98\") " pod="openshift-dns/dns-default-t4sw5" Apr 21 00:04:11.030208 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.030098 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fbbce771-9a06-4a5a-9dbf-526282060d89-cert\") pod \"ingress-canary-pvnc8\" (UID: \"fbbce771-9a06-4a5a-9dbf-526282060d89\") " pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:04:11.056260 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.056227 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-6bcc868b7-g8wql" event={"ID":"92eb6528-431d-47f2-b2d9-729c971eb0c5","Type":"ContainerStarted","Data":"d1bfca87b1456cdb67ed15c16fd2d447088379c051cf8182de3db702f8fd9981"} Apr 21 00:04:11.199325 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.199292 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-79cql\"" Apr 21 00:04:11.207485 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.207451 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-t4sw5" Apr 21 00:04:11.211395 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.211371 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-2glpj\"" Apr 21 00:04:11.219381 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.219360 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pvnc8" Apr 21 00:04:11.430727 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.430370 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:04:11.433328 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.433304 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 21 00:04:11.443046 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.443019 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d439efe8-c633-4c7c-ad01-515f448307fd-metrics-certs\") pod \"network-metrics-daemon-pf4ls\" (UID: \"d439efe8-c633-4c7c-ad01-515f448307fd\") " pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:04:11.632224 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.632195 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:04:11.634892 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.634874 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 21 00:04:11.638020 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.638002 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-n52qc\"" Apr 21 00:04:11.645354 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.645337 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 21 00:04:11.646390 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.646365 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pf4ls" Apr 21 00:04:11.656148 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.656053 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnd8l\" (UniqueName: \"kubernetes.io/projected/0b745e1d-cf68-43fd-b452-36e45a05eb93-kube-api-access-dnd8l\") pod \"network-check-target-4tj2p\" (UID: \"0b745e1d-cf68-43fd-b452-36e45a05eb93\") " pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:04:11.662862 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.662812 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pvnc8"] Apr 21 00:04:11.665285 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:11.665264 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbbce771_9a06_4a5a_9dbf_526282060d89.slice/crio-5e83b005026432f23b84dfe23ffea0b2abf16f7eee3d58684190837ac915e1a3 WatchSource:0}: Error finding container 5e83b005026432f23b84dfe23ffea0b2abf16f7eee3d58684190837ac915e1a3: Status 404 returned error can't find the container with id 5e83b005026432f23b84dfe23ffea0b2abf16f7eee3d58684190837ac915e1a3 Apr 21 00:04:11.775167 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.775139 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-pf4ls"] Apr 21 00:04:11.866088 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.866001 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd"] Apr 21 00:04:11.873175 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.873103 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-t4sw5"] Apr 21 00:04:11.884634 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:11.884588 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd439efe8_c633_4c7c_ad01_515f448307fd.slice/crio-5d04b58ed7d77720880e64666cb7c36724bb5a58ff6a66b5b915df77d758dde3 WatchSource:0}: Error finding container 5d04b58ed7d77720880e64666cb7c36724bb5a58ff6a66b5b915df77d758dde3: Status 404 returned error can't find the container with id 5d04b58ed7d77720880e64666cb7c36724bb5a58ff6a66b5b915df77d758dde3 Apr 21 00:04:11.895256 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:11.895230 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod791b8e39_beab_44fd_9b37_14969c3d1b15.slice/crio-b85d3be5b12a864c683e3db9c863c94340cdb00960c20414fe5400d6f0c4e2d6 WatchSource:0}: Error finding container b85d3be5b12a864c683e3db9c863c94340cdb00960c20414fe5400d6f0c4e2d6: Status 404 returned error can't find the container with id b85d3be5b12a864c683e3db9c863c94340cdb00960c20414fe5400d6f0c4e2d6 Apr 21 00:04:11.896119 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:11.896089 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod169c80c3_e9e6_400e_9412_bab767e5cb98.slice/crio-fed77c1514e8be7107296cb5938ceb6fdeea57a1379b63909a0291c4beb662fd WatchSource:0}: Error finding container fed77c1514e8be7107296cb5938ceb6fdeea57a1379b63909a0291c4beb662fd: Status 404 returned error can't find the container with id fed77c1514e8be7107296cb5938ceb6fdeea57a1379b63909a0291c4beb662fd Apr 21 00:04:11.933380 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.933361 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-24wfg\"" Apr 21 00:04:11.941385 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:11.941364 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:04:12.059768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:12.059729 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" event={"ID":"791b8e39-beab-44fd-9b37-14969c3d1b15","Type":"ContainerStarted","Data":"b85d3be5b12a864c683e3db9c863c94340cdb00960c20414fe5400d6f0c4e2d6"} Apr 21 00:04:12.061564 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:12.061534 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-t4sw5" event={"ID":"169c80c3-e9e6-400e-9412-bab767e5cb98","Type":"ContainerStarted","Data":"fed77c1514e8be7107296cb5938ceb6fdeea57a1379b63909a0291c4beb662fd"} Apr 21 00:04:12.063265 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:12.063222 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pvnc8" event={"ID":"fbbce771-9a06-4a5a-9dbf-526282060d89","Type":"ContainerStarted","Data":"5e83b005026432f23b84dfe23ffea0b2abf16f7eee3d58684190837ac915e1a3"} Apr 21 00:04:12.064553 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:12.064526 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pf4ls" event={"ID":"d439efe8-c633-4c7c-ad01-515f448307fd","Type":"ContainerStarted","Data":"5d04b58ed7d77720880e64666cb7c36724bb5a58ff6a66b5b915df77d758dde3"} Apr 21 00:04:12.066467 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:12.066444 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-kdsh6" event={"ID":"27d810d8-53e4-460c-8437-9151e193bcfe","Type":"ContainerStarted","Data":"07a00bcb0ad52b72200e7dca165c813550f556d3e715cac49f9a23a76c3305a0"} Apr 21 00:04:12.077733 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:12.077699 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-4tj2p"] Apr 21 00:04:12.080544 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:12.080516 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b745e1d_cf68_43fd_b452_36e45a05eb93.slice/crio-90816ea373563abd7272d7c561b118f3fc1d52fabeda3c95fc681af3f7f1a0c5 WatchSource:0}: Error finding container 90816ea373563abd7272d7c561b118f3fc1d52fabeda3c95fc681af3f7f1a0c5: Status 404 returned error can't find the container with id 90816ea373563abd7272d7c561b118f3fc1d52fabeda3c95fc681af3f7f1a0c5 Apr 21 00:04:12.082740 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:12.082686 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-kdsh6" podStartSLOduration=1.292448096 podStartE2EDuration="4.082670912s" podCreationTimestamp="2026-04-21 00:04:08 +0000 UTC" firstStartedPulling="2026-04-21 00:04:09.13406963 +0000 UTC m=+62.959048573" lastFinishedPulling="2026-04-21 00:04:11.924292447 +0000 UTC m=+65.749271389" observedRunningTime="2026-04-21 00:04:12.081868195 +0000 UTC m=+65.906847161" watchObservedRunningTime="2026-04-21 00:04:12.082670912 +0000 UTC m=+65.907649878" Apr 21 00:04:13.071856 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:13.071774 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-4tj2p" event={"ID":"0b745e1d-cf68-43fd-b452-36e45a05eb93","Type":"ContainerStarted","Data":"90816ea373563abd7272d7c561b118f3fc1d52fabeda3c95fc681af3f7f1a0c5"} Apr 21 00:04:17.086901 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.086870 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-t4sw5" event={"ID":"169c80c3-e9e6-400e-9412-bab767e5cb98","Type":"ContainerStarted","Data":"5cb98ab2ddf2b43e43ada15c8bf5995bc9a692a863cf7be1a2bf32fc567b3f15"} Apr 21 00:04:17.087364 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.086907 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-t4sw5" event={"ID":"169c80c3-e9e6-400e-9412-bab767e5cb98","Type":"ContainerStarted","Data":"2e0f377fd300fbba64e97de69adf18000e04b1c209224d81c7e39ba349b2e6b8"} Apr 21 00:04:17.087364 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.087014 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-t4sw5" Apr 21 00:04:17.088220 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.088195 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pvnc8" event={"ID":"fbbce771-9a06-4a5a-9dbf-526282060d89","Type":"ContainerStarted","Data":"ae98e1d290394784d9eeab0340773842035b7132b48c40e019978df576859afe"} Apr 21 00:04:17.089645 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.089620 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-4tj2p" event={"ID":"0b745e1d-cf68-43fd-b452-36e45a05eb93","Type":"ContainerStarted","Data":"da9bfc97964e0dc16a3bda0ee01f84b7bcdb787f50f2bb216289a4f30f3ee4ba"} Apr 21 00:04:17.091465 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.091442 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pf4ls" event={"ID":"d439efe8-c633-4c7c-ad01-515f448307fd","Type":"ContainerStarted","Data":"6ecff4e4a87c86d8428cb52925b3392166fec5ed90a5121534a325d7ae851139"} Apr 21 00:04:17.091558 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.091474 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pf4ls" event={"ID":"d439efe8-c633-4c7c-ad01-515f448307fd","Type":"ContainerStarted","Data":"3911f9d89df4bc833100e125e6a0a8cb9ea16606c67715b1efe8d65235ede3c6"} Apr 21 00:04:17.092711 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.092690 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" event={"ID":"791b8e39-beab-44fd-9b37-14969c3d1b15","Type":"ContainerStarted","Data":"ebd3e0dd44e73b51f2acbe0a6a863c1008b5eefca9aeb82ad0fade66a6c336ec"} Apr 21 00:04:17.092937 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.092919 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" Apr 21 00:04:17.097804 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.097786 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" Apr 21 00:04:17.101810 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.101767 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-t4sw5" podStartSLOduration=33.975456383 podStartE2EDuration="38.101753964s" podCreationTimestamp="2026-04-21 00:03:39 +0000 UTC" firstStartedPulling="2026-04-21 00:04:11.919241894 +0000 UTC m=+65.744220840" lastFinishedPulling="2026-04-21 00:04:16.04553947 +0000 UTC m=+69.870518421" observedRunningTime="2026-04-21 00:04:17.101548491 +0000 UTC m=+70.926527451" watchObservedRunningTime="2026-04-21 00:04:17.101753964 +0000 UTC m=+70.926732930" Apr 21 00:04:17.115942 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.115899 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-pf4ls" podStartSLOduration=66.963687291 podStartE2EDuration="1m11.115885334s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:04:11.886769335 +0000 UTC m=+65.711748279" lastFinishedPulling="2026-04-21 00:04:16.038967372 +0000 UTC m=+69.863946322" observedRunningTime="2026-04-21 00:04:17.11522361 +0000 UTC m=+70.940202574" watchObservedRunningTime="2026-04-21 00:04:17.115885334 +0000 UTC m=+70.940864300" Apr 21 00:04:17.133095 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.133059 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-4tj2p" podStartSLOduration=67.115914861 podStartE2EDuration="1m11.133041934s" podCreationTimestamp="2026-04-21 00:03:06 +0000 UTC" firstStartedPulling="2026-04-21 00:04:12.082892127 +0000 UTC m=+65.907871074" lastFinishedPulling="2026-04-21 00:04:16.100019191 +0000 UTC m=+69.924998147" observedRunningTime="2026-04-21 00:04:17.131908421 +0000 UTC m=+70.956887386" watchObservedRunningTime="2026-04-21 00:04:17.133041934 +0000 UTC m=+70.958020910" Apr 21 00:04:17.144938 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.144878 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-hmmhd" podStartSLOduration=3.025204797 podStartE2EDuration="7.144866136s" podCreationTimestamp="2026-04-21 00:04:10 +0000 UTC" firstStartedPulling="2026-04-21 00:04:11.919310271 +0000 UTC m=+65.744289220" lastFinishedPulling="2026-04-21 00:04:16.038971605 +0000 UTC m=+69.863950559" observedRunningTime="2026-04-21 00:04:17.144136241 +0000 UTC m=+70.969115206" watchObservedRunningTime="2026-04-21 00:04:17.144866136 +0000 UTC m=+70.969845102" Apr 21 00:04:17.161005 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:17.160956 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pvnc8" podStartSLOduration=33.786264562 podStartE2EDuration="38.160947193s" podCreationTimestamp="2026-04-21 00:03:39 +0000 UTC" firstStartedPulling="2026-04-21 00:04:11.667683463 +0000 UTC m=+65.492662409" lastFinishedPulling="2026-04-21 00:04:16.042366092 +0000 UTC m=+69.867345040" observedRunningTime="2026-04-21 00:04:17.160550616 +0000 UTC m=+70.985529585" watchObservedRunningTime="2026-04-21 00:04:17.160947193 +0000 UTC m=+70.985926158" Apr 21 00:04:18.101529 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:18.101497 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:04:23.426235 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.426200 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-wllkx"] Apr 21 00:04:23.433569 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.433077 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.437526 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.437236 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 21 00:04:23.437526 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.437280 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 21 00:04:23.437526 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.437467 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 21 00:04:23.437526 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.437484 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 21 00:04:23.438258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.438117 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-9p48h\"" Apr 21 00:04:23.438258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.438125 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 21 00:04:23.438258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.438172 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 21 00:04:23.518557 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518526 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.518557 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518564 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-tls\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.518781 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518595 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqkjj\" (UniqueName: \"kubernetes.io/projected/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-kube-api-access-vqkjj\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.518781 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518633 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-accelerators-collector-config\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.518781 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518688 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-sys\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.518781 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518760 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-textfile\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.518980 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518788 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-root\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.518980 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518872 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-wtmp\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.518980 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.518914 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-metrics-client-ca\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620300 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620254 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620300 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620298 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-tls\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620332 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vqkjj\" (UniqueName: \"kubernetes.io/projected/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-kube-api-access-vqkjj\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620360 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-accelerators-collector-config\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620390 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-sys\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620423 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-textfile\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620447 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-root\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620481 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-wtmp\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620503 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-metrics-client-ca\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620852 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620575 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-root\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620852 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620759 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-wtmp\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620852 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620768 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-sys\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.620852 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.620812 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-textfile\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.621074 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.621058 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-metrics-client-ca\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.621249 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.621202 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-accelerators-collector-config\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.623760 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.623733 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-tls\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.623760 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.623749 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.629119 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.629095 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqkjj\" (UniqueName: \"kubernetes.io/projected/75b150a3-e0e0-4f7d-aee8-713ff97ccf24-kube-api-access-vqkjj\") pod \"node-exporter-wllkx\" (UID: \"75b150a3-e0e0-4f7d-aee8-713ff97ccf24\") " pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:23.748574 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:23.748546 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-wllkx" Apr 21 00:04:26.392516 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.392272 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-68fb649c6d-l6tjh"] Apr 21 00:04:26.395759 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.395732 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.398543 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.398353 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy\"" Apr 21 00:04:26.398543 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.398369 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-web\"" Apr 21 00:04:26.398543 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.398428 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-rules\"" Apr 21 00:04:26.398543 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.398478 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-tls\"" Apr 21 00:04:26.398873 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.398718 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-metrics\"" Apr 21 00:04:26.398920 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.398897 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-grpc-tls-6jtt5p9fjp18u\"" Apr 21 00:04:26.399171 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.399092 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-dockercfg-s6qb2\"" Apr 21 00:04:26.405085 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.405051 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-68fb649c6d-l6tjh"] Apr 21 00:04:26.442429 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.442405 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-tls\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.442546 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.442450 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.442546 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.442480 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-grpc-tls\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.442546 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.442510 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d44788eb-1bb7-419f-be15-c17495046006-metrics-client-ca\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.442546 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.442538 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.442779 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.442565 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.442779 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.442591 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.442779 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.442638 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8lsk\" (UniqueName: \"kubernetes.io/projected/d44788eb-1bb7-419f-be15-c17495046006-kube-api-access-w8lsk\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.543276 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.543250 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-tls\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.543276 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.543285 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.543473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.543310 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-grpc-tls\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.543473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.543343 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d44788eb-1bb7-419f-be15-c17495046006-metrics-client-ca\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.543473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.543376 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.543473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.543409 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.543473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.543438 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.543473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.543470 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-w8lsk\" (UniqueName: \"kubernetes.io/projected/d44788eb-1bb7-419f-be15-c17495046006-kube-api-access-w8lsk\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.544522 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.544475 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d44788eb-1bb7-419f-be15-c17495046006-metrics-client-ca\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.546824 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.546797 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.547280 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.547255 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-tls\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.547460 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.547425 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.547920 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.547902 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.548221 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.548198 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.549194 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.549175 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d44788eb-1bb7-419f-be15-c17495046006-secret-grpc-tls\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.551457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.551433 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8lsk\" (UniqueName: \"kubernetes.io/projected/d44788eb-1bb7-419f-be15-c17495046006-kube-api-access-w8lsk\") pod \"thanos-querier-68fb649c6d-l6tjh\" (UID: \"d44788eb-1bb7-419f-be15-c17495046006\") " pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:26.710512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:26.710429 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:27.103254 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.103215 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-t4sw5" Apr 21 00:04:27.267752 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.267708 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-67c55cd8b4-6q7z5"] Apr 21 00:04:27.270698 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.270668 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.272991 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.272969 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 21 00:04:27.273127 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.273096 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-jxljl\"" Apr 21 00:04:27.273192 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.273129 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 21 00:04:27.273380 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.273360 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 21 00:04:27.273506 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.273393 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 21 00:04:27.273506 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.273404 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 21 00:04:27.279490 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.279367 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 21 00:04:27.281768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.281747 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67c55cd8b4-6q7z5"] Apr 21 00:04:27.348658 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.348633 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-oauth-serving-cert\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.348784 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.348705 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-service-ca\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.348784 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.348733 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6scnr\" (UniqueName: \"kubernetes.io/projected/820598e3-9a95-440f-b460-4157f3766952-kube-api-access-6scnr\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.348784 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.348755 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-oauth-config\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.348926 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.348787 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-serving-cert\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.348926 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.348810 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-trusted-ca-bundle\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.348926 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.348831 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-console-config\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.449988 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.449909 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-oauth-serving-cert\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.449988 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.449982 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-service-ca\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.450473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.450013 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6scnr\" (UniqueName: \"kubernetes.io/projected/820598e3-9a95-440f-b460-4157f3766952-kube-api-access-6scnr\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.450473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.450041 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-oauth-config\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.450473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.450073 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-serving-cert\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.450473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.450373 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-trusted-ca-bundle\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.450473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.450420 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-console-config\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.450719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.450696 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-oauth-serving-cert\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.450815 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.450794 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-service-ca\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.451065 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.451040 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-console-config\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.451271 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.451245 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-trusted-ca-bundle\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.452936 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.452883 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-oauth-config\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.453337 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.453312 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-serving-cert\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.457474 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.457450 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6scnr\" (UniqueName: \"kubernetes.io/projected/820598e3-9a95-440f-b460-4157f3766952-kube-api-access-6scnr\") pod \"console-67c55cd8b4-6q7z5\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:27.583708 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:27.583671 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:29.046881 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:29.046848 2568 patch_prober.go:28] interesting pod/image-registry-575d785c49-gbr8m container/registry namespace/openshift-image-registry: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 21 00:04:29.047305 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:29.046912 2568 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-575d785c49-gbr8m" podUID="1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 21 00:04:29.405060 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:29.405023 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75b150a3_e0e0_4f7d_aee8_713ff97ccf24.slice/crio-852cc0e651dc144d6d4f6d72f7752a24ed50041a86d3c6107259424971a561a7 WatchSource:0}: Error finding container 852cc0e651dc144d6d4f6d72f7752a24ed50041a86d3c6107259424971a561a7: Status 404 returned error can't find the container with id 852cc0e651dc144d6d4f6d72f7752a24ed50041a86d3c6107259424971a561a7 Apr 21 00:04:29.530145 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:29.530118 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-68fb649c6d-l6tjh"] Apr 21 00:04:29.531757 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:29.531729 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd44788eb_1bb7_419f_be15_c17495046006.slice/crio-37b3a285c5da8863beab43fbd63b0c32dd59121ba4c6f3628c2781a1fa120c7c WatchSource:0}: Error finding container 37b3a285c5da8863beab43fbd63b0c32dd59121ba4c6f3628c2781a1fa120c7c: Status 404 returned error can't find the container with id 37b3a285c5da8863beab43fbd63b0c32dd59121ba4c6f3628c2781a1fa120c7c Apr 21 00:04:29.540327 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:29.540303 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67c55cd8b4-6q7z5"] Apr 21 00:04:29.550259 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:04:29.550237 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod820598e3_9a95_440f_b460_4157f3766952.slice/crio-a5af69f07cd47c1d5824b8793303d9ba3fc5d0cb78a1433f169239342169cfdd WatchSource:0}: Error finding container a5af69f07cd47c1d5824b8793303d9ba3fc5d0cb78a1433f169239342169cfdd: Status 404 returned error can't find the container with id a5af69f07cd47c1d5824b8793303d9ba3fc5d0cb78a1433f169239342169cfdd Apr 21 00:04:30.141109 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:30.141052 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-6bcc868b7-g8wql" event={"ID":"92eb6528-431d-47f2-b2d9-729c971eb0c5","Type":"ContainerStarted","Data":"4f26f2892c2db9eddda8cad5028e2ecbec2a349ba72aea26b9b73a4ea9cebbaa"} Apr 21 00:04:30.141828 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:30.141804 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/downloads-6bcc868b7-g8wql" Apr 21 00:04:30.144333 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:30.144269 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-wllkx" event={"ID":"75b150a3-e0e0-4f7d-aee8-713ff97ccf24","Type":"ContainerStarted","Data":"852cc0e651dc144d6d4f6d72f7752a24ed50041a86d3c6107259424971a561a7"} Apr 21 00:04:30.146843 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:30.146814 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" event={"ID":"d44788eb-1bb7-419f-be15-c17495046006","Type":"ContainerStarted","Data":"37b3a285c5da8863beab43fbd63b0c32dd59121ba4c6f3628c2781a1fa120c7c"} Apr 21 00:04:30.149727 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:30.149692 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67c55cd8b4-6q7z5" event={"ID":"820598e3-9a95-440f-b460-4157f3766952","Type":"ContainerStarted","Data":"a5af69f07cd47c1d5824b8793303d9ba3fc5d0cb78a1433f169239342169cfdd"} Apr 21 00:04:30.154721 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:30.154684 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-6bcc868b7-g8wql" Apr 21 00:04:30.159248 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:30.159198 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-6bcc868b7-g8wql" podStartSLOduration=2.131878538 podStartE2EDuration="21.159184247s" podCreationTimestamp="2026-04-21 00:04:09 +0000 UTC" firstStartedPulling="2026-04-21 00:04:10.437336705 +0000 UTC m=+64.262315648" lastFinishedPulling="2026-04-21 00:04:29.464642407 +0000 UTC m=+83.289621357" observedRunningTime="2026-04-21 00:04:30.156689955 +0000 UTC m=+83.981668920" watchObservedRunningTime="2026-04-21 00:04:30.159184247 +0000 UTC m=+83.984163224" Apr 21 00:04:31.063363 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:31.063306 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-575d785c49-gbr8m" Apr 21 00:04:31.158216 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:31.158051 2568 generic.go:358] "Generic (PLEG): container finished" podID="75b150a3-e0e0-4f7d-aee8-713ff97ccf24" containerID="b97b7c4d007771a85b8a14ecf984c2a0bd11a6d1a16c8ec0d037155c1a9c0cbc" exitCode=0 Apr 21 00:04:31.158216 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:31.158196 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-wllkx" event={"ID":"75b150a3-e0e0-4f7d-aee8-713ff97ccf24","Type":"ContainerDied","Data":"b97b7c4d007771a85b8a14ecf984c2a0bd11a6d1a16c8ec0d037155c1a9c0cbc"} Apr 21 00:04:33.683170 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:33.683142 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-67c55cd8b4-6q7z5"] Apr 21 00:04:34.170622 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:34.170523 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" event={"ID":"d44788eb-1bb7-419f-be15-c17495046006","Type":"ContainerStarted","Data":"aa1a7afe4627dddc7cc616ee17e7ab42ce67810eed735f777af57ed028b2efb4"} Apr 21 00:04:34.170622 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:34.170569 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" event={"ID":"d44788eb-1bb7-419f-be15-c17495046006","Type":"ContainerStarted","Data":"459b5ada6c73417957c33ba2467018f9cdfd9fdb088eb74d3410e916d3eb2958"} Apr 21 00:04:34.170622 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:34.170582 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" event={"ID":"d44788eb-1bb7-419f-be15-c17495046006","Type":"ContainerStarted","Data":"d25fe2c7016a428397d1f338d20c9f1b3cf7b057c54d34b764f0f9288e68d678"} Apr 21 00:04:34.172067 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:34.172038 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67c55cd8b4-6q7z5" event={"ID":"820598e3-9a95-440f-b460-4157f3766952","Type":"ContainerStarted","Data":"f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a"} Apr 21 00:04:34.174183 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:34.174154 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-wllkx" event={"ID":"75b150a3-e0e0-4f7d-aee8-713ff97ccf24","Type":"ContainerStarted","Data":"8c7db5509217ede95020d315d626652a4d7e2e04fd2a311560363949d22508f2"} Apr 21 00:04:34.174183 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:34.174184 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-wllkx" event={"ID":"75b150a3-e0e0-4f7d-aee8-713ff97ccf24","Type":"ContainerStarted","Data":"201785c5c2fac4acedcea75efbf6de9298ddbbf1e188d87c118a5faaeb3d6b7c"} Apr 21 00:04:34.188784 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:34.188712 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-67c55cd8b4-6q7z5" podStartSLOduration=3.140574094 podStartE2EDuration="7.188698059s" podCreationTimestamp="2026-04-21 00:04:27 +0000 UTC" firstStartedPulling="2026-04-21 00:04:29.551851282 +0000 UTC m=+83.376830224" lastFinishedPulling="2026-04-21 00:04:33.599975239 +0000 UTC m=+87.424954189" observedRunningTime="2026-04-21 00:04:34.187462417 +0000 UTC m=+88.012441386" watchObservedRunningTime="2026-04-21 00:04:34.188698059 +0000 UTC m=+88.013677025" Apr 21 00:04:34.205421 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:34.205375 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-wllkx" podStartSLOduration=10.355244532 podStartE2EDuration="11.20535959s" podCreationTimestamp="2026-04-21 00:04:23 +0000 UTC" firstStartedPulling="2026-04-21 00:04:29.409173533 +0000 UTC m=+83.234152481" lastFinishedPulling="2026-04-21 00:04:30.259288589 +0000 UTC m=+84.084267539" observedRunningTime="2026-04-21 00:04:34.204065661 +0000 UTC m=+88.029044628" watchObservedRunningTime="2026-04-21 00:04:34.20535959 +0000 UTC m=+88.030338556" Apr 21 00:04:34.822088 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:34.822033 2568 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = unable to pull image or OCI artifact: pull image err: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404: decoding bearer token (last URL \"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_6628ed1260a548e595ee32e9de28abb4&scope=repository%3Aopenshift-release-dev%2Focp-v4.0-art-dev%3Apull&service=quay.io\", body start \"\"): unexpected end of JSON input; artifact err: provided artifact is a container image" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404" Apr 21 00:04:34.822523 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:34.822239 2568 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:prom-label-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404,Command:[],Args:[--insecure-listen-address=127.0.0.1:9095 --upstream=http://127.0.0.1:9090 --label=namespace --enable-label-apis --error-on-replace --rules-with-active-alerts --enable-label-matchers-for-rules-api],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{1 -3} {} 1m DecimalSI},memory: {{15728640 0} {} 15Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w8lsk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000450000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod thanos-querier-68fb649c6d-l6tjh_openshift-monitoring(d44788eb-1bb7-419f-be15-c17495046006): ErrImagePull: unable to pull image or OCI artifact: pull image err: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404: decoding bearer token (last URL \"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_6628ed1260a548e595ee32e9de28abb4&scope=repository%3Aopenshift-release-dev%2Focp-v4.0-art-dev%3Apull&service=quay.io\", body start \"\"): unexpected end of JSON input; artifact err: provided artifact is a container image" logger="UnhandledError" Apr 21 00:04:35.104410 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:35.104379 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prom-label-proxy\" with ErrImagePull: \"unable to pull image or OCI artifact: pull image err: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404: decoding bearer token (last URL \\\"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_6628ed1260a548e595ee32e9de28abb4&scope=repository%3Aopenshift-release-dev%2Focp-v4.0-art-dev%3Apull&service=quay.io\\\", body start \\\"\\\"): unexpected end of JSON input; artifact err: provided artifact is a container image\"" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" podUID="d44788eb-1bb7-419f-be15-c17495046006" Apr 21 00:04:35.179988 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:35.179953 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" event={"ID":"d44788eb-1bb7-419f-be15-c17495046006","Type":"ContainerStarted","Data":"154b2a38592b9e389bebe523d79d64ba73918b5216fba5e9b851e03e53757eba"} Apr 21 00:04:35.180130 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:35.179995 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" event={"ID":"d44788eb-1bb7-419f-be15-c17495046006","Type":"ContainerStarted","Data":"f3cab96c980436ef874144f643c67ac8200cd3e713831f740416918f69fe4b4a"} Apr 21 00:04:35.181562 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:35.181530 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prom-label-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404\\\": ErrImagePull: unable to pull image or OCI artifact: pull image err: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404: decoding bearer token (last URL \\\"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_6628ed1260a548e595ee32e9de28abb4&scope=repository%3Aopenshift-release-dev%2Focp-v4.0-art-dev%3Apull&service=quay.io\\\", body start \\\"\\\"): unexpected end of JSON input; artifact err: provided artifact is a container image\"" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" podUID="d44788eb-1bb7-419f-be15-c17495046006" Apr 21 00:04:36.183722 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:36.183687 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:36.184974 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:36.184944 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prom-label-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404\\\": ErrImagePull: unable to pull image or OCI artifact: pull image err: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404: decoding bearer token (last URL \\\"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_6628ed1260a548e595ee32e9de28abb4&scope=repository%3Aopenshift-release-dev%2Focp-v4.0-art-dev%3Apull&service=quay.io\\\", body start \\\"\\\"): unexpected end of JSON input; artifact err: provided artifact is a container image\"" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" podUID="d44788eb-1bb7-419f-be15-c17495046006" Apr 21 00:04:37.188361 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:37.188314 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prom-label-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404\\\": ErrImagePull: unable to pull image or OCI artifact: pull image err: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404: decoding bearer token (last URL \\\"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_6628ed1260a548e595ee32e9de28abb4&scope=repository%3Aopenshift-release-dev%2Focp-v4.0-art-dev%3Apull&service=quay.io\\\", body start \\\"\\\"): unexpected end of JSON input; artifact err: provided artifact is a container image\"" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" podUID="d44788eb-1bb7-419f-be15-c17495046006" Apr 21 00:04:37.584342 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:37.584306 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:42.193841 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:42.193816 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" Apr 21 00:04:42.195462 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:04:42.195437 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prom-label-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404\\\": ErrImagePull: unable to pull image or OCI artifact: pull image err: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5522f37104f3fac57567fa2e9ec65601f60b8cea3603b12dcda26db8c481f404: decoding bearer token (last URL \\\"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_6628ed1260a548e595ee32e9de28abb4&scope=repository%3Aopenshift-release-dev%2Focp-v4.0-art-dev%3Apull&service=quay.io\\\", body start \\\"\\\"): unexpected end of JSON input; artifact err: provided artifact is a container image\"" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" podUID="d44788eb-1bb7-419f-be15-c17495046006" Apr 21 00:04:49.106404 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:49.106375 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-4tj2p" Apr 21 00:04:53.229326 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:53.229298 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-pvnc8_fbbce771-9a06-4a5a-9dbf-526282060d89/serve-healthcheck-canary/0.log" Apr 21 00:04:58.249879 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:58.249841 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" event={"ID":"d44788eb-1bb7-419f-be15-c17495046006","Type":"ContainerStarted","Data":"c46539516c463b950bad36d44179faf2747b332280088ef60ca10dc789f3a6e8"} Apr 21 00:04:58.271798 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:58.271747 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-68fb649c6d-l6tjh" podStartSLOduration=4.292956108 podStartE2EDuration="32.271730079s" podCreationTimestamp="2026-04-21 00:04:26 +0000 UTC" firstStartedPulling="2026-04-21 00:04:29.533702369 +0000 UTC m=+83.358681311" lastFinishedPulling="2026-04-21 00:04:57.512476336 +0000 UTC m=+111.337455282" observedRunningTime="2026-04-21 00:04:58.269909054 +0000 UTC m=+112.094888019" watchObservedRunningTime="2026-04-21 00:04:58.271730079 +0000 UTC m=+112.096709046" Apr 21 00:04:59.194431 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.194374 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-67c55cd8b4-6q7z5" podUID="820598e3-9a95-440f-b460-4157f3766952" containerName="console" containerID="cri-o://f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a" gracePeriod=15 Apr 21 00:04:59.463677 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.463657 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-67c55cd8b4-6q7z5_820598e3-9a95-440f-b460-4157f3766952/console/0.log" Apr 21 00:04:59.463945 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.463721 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:04:59.598681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.598658 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-console-config\") pod \"820598e3-9a95-440f-b460-4157f3766952\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " Apr 21 00:04:59.598802 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.598688 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-oauth-config\") pod \"820598e3-9a95-440f-b460-4157f3766952\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " Apr 21 00:04:59.598802 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.598708 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-trusted-ca-bundle\") pod \"820598e3-9a95-440f-b460-4157f3766952\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " Apr 21 00:04:59.598802 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.598743 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-oauth-serving-cert\") pod \"820598e3-9a95-440f-b460-4157f3766952\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " Apr 21 00:04:59.598916 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.598879 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-service-ca\") pod \"820598e3-9a95-440f-b460-4157f3766952\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " Apr 21 00:04:59.598965 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.598954 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6scnr\" (UniqueName: \"kubernetes.io/projected/820598e3-9a95-440f-b460-4157f3766952-kube-api-access-6scnr\") pod \"820598e3-9a95-440f-b460-4157f3766952\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " Apr 21 00:04:59.599022 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.598981 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-serving-cert\") pod \"820598e3-9a95-440f-b460-4157f3766952\" (UID: \"820598e3-9a95-440f-b460-4157f3766952\") " Apr 21 00:04:59.599185 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.599158 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "820598e3-9a95-440f-b460-4157f3766952" (UID: "820598e3-9a95-440f-b460-4157f3766952"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:04:59.599291 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.599186 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "820598e3-9a95-440f-b460-4157f3766952" (UID: "820598e3-9a95-440f-b460-4157f3766952"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:04:59.599291 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.599189 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-console-config" (OuterVolumeSpecName: "console-config") pod "820598e3-9a95-440f-b460-4157f3766952" (UID: "820598e3-9a95-440f-b460-4157f3766952"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:04:59.599291 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.599277 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-service-ca" (OuterVolumeSpecName: "service-ca") pod "820598e3-9a95-440f-b460-4157f3766952" (UID: "820598e3-9a95-440f-b460-4157f3766952"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:04:59.600951 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.600923 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "820598e3-9a95-440f-b460-4157f3766952" (UID: "820598e3-9a95-440f-b460-4157f3766952"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:04:59.601053 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.600955 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/820598e3-9a95-440f-b460-4157f3766952-kube-api-access-6scnr" (OuterVolumeSpecName: "kube-api-access-6scnr") pod "820598e3-9a95-440f-b460-4157f3766952" (UID: "820598e3-9a95-440f-b460-4157f3766952"). InnerVolumeSpecName "kube-api-access-6scnr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:04:59.601053 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.601020 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "820598e3-9a95-440f-b460-4157f3766952" (UID: "820598e3-9a95-440f-b460-4157f3766952"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:04:59.699433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.699407 2568 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-service-ca\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:59.699537 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.699433 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6scnr\" (UniqueName: \"kubernetes.io/projected/820598e3-9a95-440f-b460-4157f3766952-kube-api-access-6scnr\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:59.699537 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.699450 2568 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-serving-cert\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:59.699537 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.699463 2568 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-console-config\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:59.699537 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.699476 2568 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/820598e3-9a95-440f-b460-4157f3766952-console-oauth-config\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:59.699537 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.699488 2568 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-trusted-ca-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:04:59.699537 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:04:59.699501 2568 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/820598e3-9a95-440f-b460-4157f3766952-oauth-serving-cert\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:05:00.255742 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.255716 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-67c55cd8b4-6q7z5_820598e3-9a95-440f-b460-4157f3766952/console/0.log" Apr 21 00:05:00.255980 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.255753 2568 generic.go:358] "Generic (PLEG): container finished" podID="820598e3-9a95-440f-b460-4157f3766952" containerID="f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a" exitCode=2 Apr 21 00:05:00.255980 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.255787 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67c55cd8b4-6q7z5" event={"ID":"820598e3-9a95-440f-b460-4157f3766952","Type":"ContainerDied","Data":"f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a"} Apr 21 00:05:00.255980 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.255829 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67c55cd8b4-6q7z5" event={"ID":"820598e3-9a95-440f-b460-4157f3766952","Type":"ContainerDied","Data":"a5af69f07cd47c1d5824b8793303d9ba3fc5d0cb78a1433f169239342169cfdd"} Apr 21 00:05:00.255980 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.255830 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67c55cd8b4-6q7z5" Apr 21 00:05:00.255980 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.255845 2568 scope.go:117] "RemoveContainer" containerID="f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a" Apr 21 00:05:00.263871 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.263854 2568 scope.go:117] "RemoveContainer" containerID="f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a" Apr 21 00:05:00.264143 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:05:00.264125 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a\": container with ID starting with f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a not found: ID does not exist" containerID="f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a" Apr 21 00:05:00.264220 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.264150 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a"} err="failed to get container status \"f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a\": rpc error: code = NotFound desc = could not find container \"f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a\": container with ID starting with f0224ca67e18731f5130dccb7c50324017e22cc076a109c8c78a80e28b95424a not found: ID does not exist" Apr 21 00:05:00.275067 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.275045 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-67c55cd8b4-6q7z5"] Apr 21 00:05:00.278083 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.278063 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-67c55cd8b4-6q7z5"] Apr 21 00:05:00.718569 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:00.718539 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="820598e3-9a95-440f-b460-4157f3766952" path="/var/lib/kubelet/pods/820598e3-9a95-440f-b460-4157f3766952/volumes" Apr 21 00:05:51.609702 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.609670 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-8489bcb85f-tbscm"] Apr 21 00:05:51.610130 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.609922 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="820598e3-9a95-440f-b460-4157f3766952" containerName="console" Apr 21 00:05:51.610130 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.609932 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="820598e3-9a95-440f-b460-4157f3766952" containerName="console" Apr 21 00:05:51.610130 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.609972 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="820598e3-9a95-440f-b460-4157f3766952" containerName="console" Apr 21 00:05:51.613953 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.613938 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.616471 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.616447 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 21 00:05:51.616582 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.616550 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 21 00:05:51.617648 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.617628 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 21 00:05:51.617730 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.617633 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 21 00:05:51.617814 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.617637 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 21 00:05:51.617861 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.617711 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-jxljl\"" Apr 21 00:05:51.622498 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.622479 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 21 00:05:51.622594 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.622495 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8489bcb85f-tbscm"] Apr 21 00:05:51.677702 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.677674 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-serving-cert\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.677819 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.677714 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-oauth-config\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.677819 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.677751 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-service-ca\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.677819 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.677774 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76bcs\" (UniqueName: \"kubernetes.io/projected/5d317df4-fde0-4d56-907f-e21a04729670-kube-api-access-76bcs\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.677819 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.677800 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-oauth-serving-cert\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.678009 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.677893 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-console-config\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.678009 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.677919 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-trusted-ca-bundle\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.779039 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.778999 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-service-ca\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.779039 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.779040 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-76bcs\" (UniqueName: \"kubernetes.io/projected/5d317df4-fde0-4d56-907f-e21a04729670-kube-api-access-76bcs\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.779202 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.779059 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-oauth-serving-cert\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.779202 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.779161 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-console-config\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.779202 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.779183 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-trusted-ca-bundle\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.779322 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.779229 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-serving-cert\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.779322 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.779263 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-oauth-config\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.780073 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.780047 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-service-ca\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.780168 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.780047 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-console-config\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.780168 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.780055 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-oauth-serving-cert\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.780424 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.780402 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-trusted-ca-bundle\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.781857 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.781835 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-oauth-config\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.781971 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.781955 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-serving-cert\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.787872 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.787849 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-76bcs\" (UniqueName: \"kubernetes.io/projected/5d317df4-fde0-4d56-907f-e21a04729670-kube-api-access-76bcs\") pod \"console-8489bcb85f-tbscm\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:51.923566 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:51.923514 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:05:52.036276 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:52.036250 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8489bcb85f-tbscm"] Apr 21 00:05:52.392055 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:52.392024 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8489bcb85f-tbscm" event={"ID":"5d317df4-fde0-4d56-907f-e21a04729670","Type":"ContainerStarted","Data":"e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357"} Apr 21 00:05:52.392055 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:52.392058 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8489bcb85f-tbscm" event={"ID":"5d317df4-fde0-4d56-907f-e21a04729670","Type":"ContainerStarted","Data":"72efdf3b8a760132d2d2da13e96be75e49b5336c1d681e82dc35e8a0f6e86989"} Apr 21 00:05:52.409187 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:05:52.409142 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-8489bcb85f-tbscm" podStartSLOduration=1.409130067 podStartE2EDuration="1.409130067s" podCreationTimestamp="2026-04-21 00:05:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:05:52.407720325 +0000 UTC m=+166.232699289" watchObservedRunningTime="2026-04-21 00:05:52.409130067 +0000 UTC m=+166.234109032" Apr 21 00:06:01.923923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:06:01.923839 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:06:01.923923 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:06:01.923875 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:06:01.928709 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:06:01.928681 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:06:02.420436 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:06:02.420405 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:07:02.058848 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.058818 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-c9fddddfc-kfgvq"] Apr 21 00:07:02.061857 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.061843 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.072259 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.072238 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-c9fddddfc-kfgvq"] Apr 21 00:07:02.148153 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.148125 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-serving-cert\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.148153 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.148152 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-console-config\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.148308 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.148169 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-service-ca\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.148308 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.148192 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-trusted-ca-bundle\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.148308 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.148252 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-oauth-config\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.148308 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.148275 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l75bx\" (UniqueName: \"kubernetes.io/projected/12dee547-67d4-4598-8012-7a098c29cc7c-kube-api-access-l75bx\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.148308 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.148294 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-oauth-serving-cert\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.248996 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.248970 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-console-config\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.248996 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.248999 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-service-ca\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.249191 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249018 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-trusted-ca-bundle\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.249191 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249045 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-oauth-config\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.249191 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249065 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l75bx\" (UniqueName: \"kubernetes.io/projected/12dee547-67d4-4598-8012-7a098c29cc7c-kube-api-access-l75bx\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.249191 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249090 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-oauth-serving-cert\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.249191 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249136 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-serving-cert\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.249835 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249808 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-console-config\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.249954 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249854 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-service-ca\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.249954 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249869 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-oauth-serving-cert\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.250052 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.249972 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-trusted-ca-bundle\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.251528 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.251508 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-serving-cert\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.251584 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.251507 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-oauth-config\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.255875 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.255858 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l75bx\" (UniqueName: \"kubernetes.io/projected/12dee547-67d4-4598-8012-7a098c29cc7c-kube-api-access-l75bx\") pod \"console-c9fddddfc-kfgvq\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.370186 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.370110 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:02.480855 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.480834 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-c9fddddfc-kfgvq"] Apr 21 00:07:02.584342 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.584309 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c9fddddfc-kfgvq" event={"ID":"12dee547-67d4-4598-8012-7a098c29cc7c","Type":"ContainerStarted","Data":"42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b"} Apr 21 00:07:02.584342 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.584344 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c9fddddfc-kfgvq" event={"ID":"12dee547-67d4-4598-8012-7a098c29cc7c","Type":"ContainerStarted","Data":"d95c5a0f82f9902ad04cb0b945e9ed594b9a5d231e08a1bdfd6f2a44c41a6d1e"} Apr 21 00:07:02.600869 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:02.600816 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-c9fddddfc-kfgvq" podStartSLOduration=0.600800343 podStartE2EDuration="600.800343ms" podCreationTimestamp="2026-04-21 00:07:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:07:02.599573209 +0000 UTC m=+236.424552174" watchObservedRunningTime="2026-04-21 00:07:02.600800343 +0000 UTC m=+236.425779310" Apr 21 00:07:12.370447 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:12.370396 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:12.370447 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:12.370446 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:12.375160 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:12.375136 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:12.616898 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:12.616875 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:07:12.656785 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:12.656719 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-8489bcb85f-tbscm"] Apr 21 00:07:37.675821 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:37.675698 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-8489bcb85f-tbscm" podUID="5d317df4-fde0-4d56-907f-e21a04729670" containerName="console" containerID="cri-o://e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357" gracePeriod=15 Apr 21 00:07:37.915583 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:37.915549 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8489bcb85f-tbscm_5d317df4-fde0-4d56-907f-e21a04729670/console/0.log" Apr 21 00:07:37.915702 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:37.915628 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:07:37.999888 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:37.999866 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76bcs\" (UniqueName: \"kubernetes.io/projected/5d317df4-fde0-4d56-907f-e21a04729670-kube-api-access-76bcs\") pod \"5d317df4-fde0-4d56-907f-e21a04729670\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " Apr 21 00:07:37.999989 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:37.999917 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-console-config\") pod \"5d317df4-fde0-4d56-907f-e21a04729670\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " Apr 21 00:07:37.999989 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:37.999936 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-oauth-serving-cert\") pod \"5d317df4-fde0-4d56-907f-e21a04729670\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " Apr 21 00:07:38.000128 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000109 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-trusted-ca-bundle\") pod \"5d317df4-fde0-4d56-907f-e21a04729670\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " Apr 21 00:07:38.000176 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000154 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-oauth-config\") pod \"5d317df4-fde0-4d56-907f-e21a04729670\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " Apr 21 00:07:38.000261 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000235 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-console-config" (OuterVolumeSpecName: "console-config") pod "5d317df4-fde0-4d56-907f-e21a04729670" (UID: "5d317df4-fde0-4d56-907f-e21a04729670"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:07:38.000319 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000245 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "5d317df4-fde0-4d56-907f-e21a04729670" (UID: "5d317df4-fde0-4d56-907f-e21a04729670"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:07:38.000319 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000266 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-service-ca\") pod \"5d317df4-fde0-4d56-907f-e21a04729670\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " Apr 21 00:07:38.000319 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000303 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-serving-cert\") pod \"5d317df4-fde0-4d56-907f-e21a04729670\" (UID: \"5d317df4-fde0-4d56-907f-e21a04729670\") " Apr 21 00:07:38.000535 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000514 2568 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-console-config\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:07:38.000630 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000530 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "5d317df4-fde0-4d56-907f-e21a04729670" (UID: "5d317df4-fde0-4d56-907f-e21a04729670"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:07:38.000630 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000539 2568 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-oauth-serving-cert\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:07:38.000744 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.000649 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-service-ca" (OuterVolumeSpecName: "service-ca") pod "5d317df4-fde0-4d56-907f-e21a04729670" (UID: "5d317df4-fde0-4d56-907f-e21a04729670"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:07:38.002173 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.002146 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "5d317df4-fde0-4d56-907f-e21a04729670" (UID: "5d317df4-fde0-4d56-907f-e21a04729670"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:07:38.002275 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.002214 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d317df4-fde0-4d56-907f-e21a04729670-kube-api-access-76bcs" (OuterVolumeSpecName: "kube-api-access-76bcs") pod "5d317df4-fde0-4d56-907f-e21a04729670" (UID: "5d317df4-fde0-4d56-907f-e21a04729670"). InnerVolumeSpecName "kube-api-access-76bcs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:07:38.002318 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.002269 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "5d317df4-fde0-4d56-907f-e21a04729670" (UID: "5d317df4-fde0-4d56-907f-e21a04729670"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:07:38.100968 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.100939 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-76bcs\" (UniqueName: \"kubernetes.io/projected/5d317df4-fde0-4d56-907f-e21a04729670-kube-api-access-76bcs\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:07:38.100968 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.100963 2568 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-trusted-ca-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:07:38.100968 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.100973 2568 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-oauth-config\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:07:38.101148 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.100982 2568 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5d317df4-fde0-4d56-907f-e21a04729670-service-ca\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:07:38.101148 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.100991 2568 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d317df4-fde0-4d56-907f-e21a04729670-console-serving-cert\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:07:38.684403 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.684375 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8489bcb85f-tbscm_5d317df4-fde0-4d56-907f-e21a04729670/console/0.log" Apr 21 00:07:38.684776 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.684411 2568 generic.go:358] "Generic (PLEG): container finished" podID="5d317df4-fde0-4d56-907f-e21a04729670" containerID="e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357" exitCode=2 Apr 21 00:07:38.684776 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.684445 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8489bcb85f-tbscm" event={"ID":"5d317df4-fde0-4d56-907f-e21a04729670","Type":"ContainerDied","Data":"e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357"} Apr 21 00:07:38.684776 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.684471 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8489bcb85f-tbscm" Apr 21 00:07:38.684776 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.684485 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8489bcb85f-tbscm" event={"ID":"5d317df4-fde0-4d56-907f-e21a04729670","Type":"ContainerDied","Data":"72efdf3b8a760132d2d2da13e96be75e49b5336c1d681e82dc35e8a0f6e86989"} Apr 21 00:07:38.684776 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.684501 2568 scope.go:117] "RemoveContainer" containerID="e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357" Apr 21 00:07:38.693226 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.693209 2568 scope.go:117] "RemoveContainer" containerID="e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357" Apr 21 00:07:38.693444 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:07:38.693426 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357\": container with ID starting with e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357 not found: ID does not exist" containerID="e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357" Apr 21 00:07:38.693489 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.693453 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357"} err="failed to get container status \"e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357\": rpc error: code = NotFound desc = could not find container \"e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357\": container with ID starting with e100c1976bbdc3dabc1911b403ee91a631fdad5855c54f30a20c07dd259c3357 not found: ID does not exist" Apr 21 00:07:38.704287 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.704267 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-8489bcb85f-tbscm"] Apr 21 00:07:38.709632 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.709600 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-8489bcb85f-tbscm"] Apr 21 00:07:38.718193 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:38.718169 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d317df4-fde0-4d56-907f-e21a04729670" path="/var/lib/kubelet/pods/5d317df4-fde0-4d56-907f-e21a04729670/volumes" Apr 21 00:07:39.505183 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.505148 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6"] Apr 21 00:07:39.505439 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.505426 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5d317df4-fde0-4d56-907f-e21a04729670" containerName="console" Apr 21 00:07:39.505439 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.505439 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d317df4-fde0-4d56-907f-e21a04729670" containerName="console" Apr 21 00:07:39.505529 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.505499 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="5d317df4-fde0-4d56-907f-e21a04729670" containerName="console" Apr 21 00:07:39.509906 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.509889 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.512416 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.512392 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 21 00:07:39.512535 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.512412 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 21 00:07:39.512535 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.512392 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-4fqdf\"" Apr 21 00:07:39.515113 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.515088 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6"] Apr 21 00:07:39.612628 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.612582 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-util\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.612768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.612652 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-bundle\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.612768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.612727 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnzpv\" (UniqueName: \"kubernetes.io/projected/4b526576-319c-4f4e-91da-59f9199e3998-kube-api-access-pnzpv\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.713220 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.713199 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pnzpv\" (UniqueName: \"kubernetes.io/projected/4b526576-319c-4f4e-91da-59f9199e3998-kube-api-access-pnzpv\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.713508 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.713229 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-util\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.713508 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.713263 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-bundle\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.713601 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.713564 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-util\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.713662 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.713642 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-bundle\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.720707 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.720675 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnzpv\" (UniqueName: \"kubernetes.io/projected/4b526576-319c-4f4e-91da-59f9199e3998-kube-api-access-pnzpv\") pod \"2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.820346 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.820297 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:39.932171 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:39.932140 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6"] Apr 21 00:07:39.934470 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:07:39.934441 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b526576_319c_4f4e_91da_59f9199e3998.slice/crio-459bd8e3a111d142b97ae42057835a047fb5916727cf7d6b472c072e16bd76d3 WatchSource:0}: Error finding container 459bd8e3a111d142b97ae42057835a047fb5916727cf7d6b472c072e16bd76d3: Status 404 returned error can't find the container with id 459bd8e3a111d142b97ae42057835a047fb5916727cf7d6b472c072e16bd76d3 Apr 21 00:07:40.690868 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:40.690835 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" event={"ID":"4b526576-319c-4f4e-91da-59f9199e3998","Type":"ContainerStarted","Data":"459bd8e3a111d142b97ae42057835a047fb5916727cf7d6b472c072e16bd76d3"} Apr 21 00:07:47.711898 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:47.711862 2568 generic.go:358] "Generic (PLEG): container finished" podID="4b526576-319c-4f4e-91da-59f9199e3998" containerID="bd148255d092e9decaabd94999d8f6a087f51ae9f24c59c058be69005cb4612b" exitCode=0 Apr 21 00:07:47.712288 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:47.711956 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" event={"ID":"4b526576-319c-4f4e-91da-59f9199e3998","Type":"ContainerDied","Data":"bd148255d092e9decaabd94999d8f6a087f51ae9f24c59c058be69005cb4612b"} Apr 21 00:07:50.720942 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:50.720904 2568 generic.go:358] "Generic (PLEG): container finished" podID="4b526576-319c-4f4e-91da-59f9199e3998" containerID="17fe84333b6aaedb5c5b21cb60e40dcdff99186db3f2329569cd28fef3619552" exitCode=0 Apr 21 00:07:50.721316 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:50.721010 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" event={"ID":"4b526576-319c-4f4e-91da-59f9199e3998","Type":"ContainerDied","Data":"17fe84333b6aaedb5c5b21cb60e40dcdff99186db3f2329569cd28fef3619552"} Apr 21 00:07:58.743075 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:58.743034 2568 generic.go:358] "Generic (PLEG): container finished" podID="4b526576-319c-4f4e-91da-59f9199e3998" containerID="20b3c4bdf02250cf11a86e27144aa75fdf164735883b4fe9b49657017dfbbd7d" exitCode=0 Apr 21 00:07:58.743427 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:58.743087 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" event={"ID":"4b526576-319c-4f4e-91da-59f9199e3998","Type":"ContainerDied","Data":"20b3c4bdf02250cf11a86e27144aa75fdf164735883b4fe9b49657017dfbbd7d"} Apr 21 00:07:59.858380 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:59.858359 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:07:59.968062 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:59.968040 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-util\") pod \"4b526576-319c-4f4e-91da-59f9199e3998\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " Apr 21 00:07:59.968162 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:59.968071 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnzpv\" (UniqueName: \"kubernetes.io/projected/4b526576-319c-4f4e-91da-59f9199e3998-kube-api-access-pnzpv\") pod \"4b526576-319c-4f4e-91da-59f9199e3998\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " Apr 21 00:07:59.968162 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:59.968104 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-bundle\") pod \"4b526576-319c-4f4e-91da-59f9199e3998\" (UID: \"4b526576-319c-4f4e-91da-59f9199e3998\") " Apr 21 00:07:59.968773 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:59.968746 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-bundle" (OuterVolumeSpecName: "bundle") pod "4b526576-319c-4f4e-91da-59f9199e3998" (UID: "4b526576-319c-4f4e-91da-59f9199e3998"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:07:59.970211 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:59.970184 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b526576-319c-4f4e-91da-59f9199e3998-kube-api-access-pnzpv" (OuterVolumeSpecName: "kube-api-access-pnzpv") pod "4b526576-319c-4f4e-91da-59f9199e3998" (UID: "4b526576-319c-4f4e-91da-59f9199e3998"). InnerVolumeSpecName "kube-api-access-pnzpv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:07:59.972024 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:07:59.972005 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-util" (OuterVolumeSpecName: "util") pod "4b526576-319c-4f4e-91da-59f9199e3998" (UID: "4b526576-319c-4f4e-91da-59f9199e3998"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:08:00.068934 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:00.068850 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:00.068934 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:00.068881 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pnzpv\" (UniqueName: \"kubernetes.io/projected/4b526576-319c-4f4e-91da-59f9199e3998-kube-api-access-pnzpv\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:00.068934 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:00.068891 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4b526576-319c-4f4e-91da-59f9199e3998-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:00.748802 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:00.748772 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" event={"ID":"4b526576-319c-4f4e-91da-59f9199e3998","Type":"ContainerDied","Data":"459bd8e3a111d142b97ae42057835a047fb5916727cf7d6b472c072e16bd76d3"} Apr 21 00:08:00.748802 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:00.748802 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="459bd8e3a111d142b97ae42057835a047fb5916727cf7d6b472c072e16bd76d3" Apr 21 00:08:00.748952 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:00.748828 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dkwkn6" Apr 21 00:08:06.633558 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:06.633525 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:08:06.634062 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:06.633941 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:08:06.640316 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:06.640155 2568 kubelet.go:1628] "Image garbage collection succeeded" Apr 21 00:08:07.240311 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.240283 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5"] Apr 21 00:08:07.242433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.240535 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4b526576-319c-4f4e-91da-59f9199e3998" containerName="util" Apr 21 00:08:07.242433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.240546 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b526576-319c-4f4e-91da-59f9199e3998" containerName="util" Apr 21 00:08:07.242433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.240556 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4b526576-319c-4f4e-91da-59f9199e3998" containerName="extract" Apr 21 00:08:07.242433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.240561 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b526576-319c-4f4e-91da-59f9199e3998" containerName="extract" Apr 21 00:08:07.242433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.240569 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4b526576-319c-4f4e-91da-59f9199e3998" containerName="pull" Apr 21 00:08:07.242433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.240574 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b526576-319c-4f4e-91da-59f9199e3998" containerName="pull" Apr 21 00:08:07.242433 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.240628 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="4b526576-319c-4f4e-91da-59f9199e3998" containerName="extract" Apr 21 00:08:07.243341 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.243325 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" Apr 21 00:08:07.246657 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.246629 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"kube-root-ca.crt\"" Apr 21 00:08:07.246811 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.246794 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"openshift-service-ca.crt\"" Apr 21 00:08:07.246927 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.246909 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager-operator\"/\"cert-manager-operator-controller-manager-dockercfg-ckl46\"" Apr 21 00:08:07.265166 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.265145 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5"] Apr 21 00:08:07.315208 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.315183 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a84b7721-fefe-4e4f-8881-7a359e489640-tmp\") pod \"cert-manager-operator-controller-manager-54b9655956-9gsr5\" (UID: \"a84b7721-fefe-4e4f-8881-7a359e489640\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" Apr 21 00:08:07.315309 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.315234 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvp8h\" (UniqueName: \"kubernetes.io/projected/a84b7721-fefe-4e4f-8881-7a359e489640-kube-api-access-nvp8h\") pod \"cert-manager-operator-controller-manager-54b9655956-9gsr5\" (UID: \"a84b7721-fefe-4e4f-8881-7a359e489640\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" Apr 21 00:08:07.415624 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.415583 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nvp8h\" (UniqueName: \"kubernetes.io/projected/a84b7721-fefe-4e4f-8881-7a359e489640-kube-api-access-nvp8h\") pod \"cert-manager-operator-controller-manager-54b9655956-9gsr5\" (UID: \"a84b7721-fefe-4e4f-8881-7a359e489640\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" Apr 21 00:08:07.415706 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.415663 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a84b7721-fefe-4e4f-8881-7a359e489640-tmp\") pod \"cert-manager-operator-controller-manager-54b9655956-9gsr5\" (UID: \"a84b7721-fefe-4e4f-8881-7a359e489640\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" Apr 21 00:08:07.415944 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.415930 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a84b7721-fefe-4e4f-8881-7a359e489640-tmp\") pod \"cert-manager-operator-controller-manager-54b9655956-9gsr5\" (UID: \"a84b7721-fefe-4e4f-8881-7a359e489640\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" Apr 21 00:08:07.424049 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.424021 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvp8h\" (UniqueName: \"kubernetes.io/projected/a84b7721-fefe-4e4f-8881-7a359e489640-kube-api-access-nvp8h\") pod \"cert-manager-operator-controller-manager-54b9655956-9gsr5\" (UID: \"a84b7721-fefe-4e4f-8881-7a359e489640\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" Apr 21 00:08:07.552087 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.552032 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" Apr 21 00:08:07.670481 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.670449 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5"] Apr 21 00:08:07.674281 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:08:07.674255 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda84b7721_fefe_4e4f_8881_7a359e489640.slice/crio-5f96727cebe5369371bc71c5dcbd8404e639c578aa911f0091896a66171b1541 WatchSource:0}: Error finding container 5f96727cebe5369371bc71c5dcbd8404e639c578aa911f0091896a66171b1541: Status 404 returned error can't find the container with id 5f96727cebe5369371bc71c5dcbd8404e639c578aa911f0091896a66171b1541 Apr 21 00:08:07.676632 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.676601 2568 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 00:08:07.767310 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:07.767283 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" event={"ID":"a84b7721-fefe-4e4f-8881-7a359e489640","Type":"ContainerStarted","Data":"5f96727cebe5369371bc71c5dcbd8404e639c578aa911f0091896a66171b1541"} Apr 21 00:08:10.779140 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:10.779106 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" event={"ID":"a84b7721-fefe-4e4f-8881-7a359e489640","Type":"ContainerStarted","Data":"ed806a8317df1edd49cb63851a72660095bad064b87709e4d38f769f2f4abb0d"} Apr 21 00:08:10.798747 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:10.798702 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-54b9655956-9gsr5" podStartSLOduration=1.5628988910000001 podStartE2EDuration="3.798687766s" podCreationTimestamp="2026-04-21 00:08:07 +0000 UTC" firstStartedPulling="2026-04-21 00:08:07.676745398 +0000 UTC m=+301.501724342" lastFinishedPulling="2026-04-21 00:08:09.912534273 +0000 UTC m=+303.737513217" observedRunningTime="2026-04-21 00:08:10.796847801 +0000 UTC m=+304.621826767" watchObservedRunningTime="2026-04-21 00:08:10.798687766 +0000 UTC m=+304.623666731" Apr 21 00:08:11.770597 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.770563 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h"] Apr 21 00:08:11.774014 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.773999 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.776515 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.776493 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 21 00:08:11.776515 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.776512 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 21 00:08:11.777620 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.777592 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-4fqdf\"" Apr 21 00:08:11.781666 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.781643 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h"] Apr 21 00:08:11.846387 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.846363 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-util\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.846489 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.846425 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-bundle\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.846489 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.846477 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqtfh\" (UniqueName: \"kubernetes.io/projected/46e010ed-5759-4f78-abbc-50827e53e250-kube-api-access-mqtfh\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.946971 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.946940 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-util\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.947077 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.947006 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-bundle\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.947077 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.947059 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mqtfh\" (UniqueName: \"kubernetes.io/projected/46e010ed-5759-4f78-abbc-50827e53e250-kube-api-access-mqtfh\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.947307 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.947286 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-util\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.947795 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.947779 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-bundle\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:11.961942 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:11.961911 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqtfh\" (UniqueName: \"kubernetes.io/projected/46e010ed-5759-4f78-abbc-50827e53e250-kube-api-access-mqtfh\") pod \"77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:12.083482 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:12.083420 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:12.198571 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:12.198551 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h"] Apr 21 00:08:12.200405 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:08:12.200377 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46e010ed_5759_4f78_abbc_50827e53e250.slice/crio-58a99be76ab9aac7ef8fed68dcfb2a0686a53834349073f35a36bddf73e9179a WatchSource:0}: Error finding container 58a99be76ab9aac7ef8fed68dcfb2a0686a53834349073f35a36bddf73e9179a: Status 404 returned error can't find the container with id 58a99be76ab9aac7ef8fed68dcfb2a0686a53834349073f35a36bddf73e9179a Apr 21 00:08:12.787038 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:12.787004 2568 generic.go:358] "Generic (PLEG): container finished" podID="46e010ed-5759-4f78-abbc-50827e53e250" containerID="ac06ea0fc8b0aa5f022801c25b956441fba225b2170c8b0b98238220e6e31efc" exitCode=0 Apr 21 00:08:12.787404 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:12.787070 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" event={"ID":"46e010ed-5759-4f78-abbc-50827e53e250","Type":"ContainerDied","Data":"ac06ea0fc8b0aa5f022801c25b956441fba225b2170c8b0b98238220e6e31efc"} Apr 21 00:08:12.787404 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:12.787097 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" event={"ID":"46e010ed-5759-4f78-abbc-50827e53e250","Type":"ContainerStarted","Data":"58a99be76ab9aac7ef8fed68dcfb2a0686a53834349073f35a36bddf73e9179a"} Apr 21 00:08:13.277073 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.277045 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-hw9l6"] Apr 21 00:08:13.280174 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.280159 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:13.282397 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.282373 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-webhook-dockercfg-k8tzg\"" Apr 21 00:08:13.282528 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.282412 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"openshift-service-ca.crt\"" Apr 21 00:08:13.282650 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.282636 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"kube-root-ca.crt\"" Apr 21 00:08:13.291354 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.291335 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-hw9l6"] Apr 21 00:08:13.358498 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.358477 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65pzv\" (UniqueName: \"kubernetes.io/projected/6c37eea2-3c8a-424c-8128-0e7307d76af0-kube-api-access-65pzv\") pod \"cert-manager-webhook-587ccfb98-hw9l6\" (UID: \"6c37eea2-3c8a-424c-8128-0e7307d76af0\") " pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:13.358599 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.358516 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6c37eea2-3c8a-424c-8128-0e7307d76af0-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-hw9l6\" (UID: \"6c37eea2-3c8a-424c-8128-0e7307d76af0\") " pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:13.459757 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.459730 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-65pzv\" (UniqueName: \"kubernetes.io/projected/6c37eea2-3c8a-424c-8128-0e7307d76af0-kube-api-access-65pzv\") pod \"cert-manager-webhook-587ccfb98-hw9l6\" (UID: \"6c37eea2-3c8a-424c-8128-0e7307d76af0\") " pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:13.459853 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.459769 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6c37eea2-3c8a-424c-8128-0e7307d76af0-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-hw9l6\" (UID: \"6c37eea2-3c8a-424c-8128-0e7307d76af0\") " pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:13.471633 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.471595 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6c37eea2-3c8a-424c-8128-0e7307d76af0-bound-sa-token\") pod \"cert-manager-webhook-587ccfb98-hw9l6\" (UID: \"6c37eea2-3c8a-424c-8128-0e7307d76af0\") " pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:13.471773 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.471755 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-65pzv\" (UniqueName: \"kubernetes.io/projected/6c37eea2-3c8a-424c-8128-0e7307d76af0-kube-api-access-65pzv\") pod \"cert-manager-webhook-587ccfb98-hw9l6\" (UID: \"6c37eea2-3c8a-424c-8128-0e7307d76af0\") " pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:13.603488 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.603432 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:13.724562 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.724537 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-587ccfb98-hw9l6"] Apr 21 00:08:13.726928 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:08:13.726897 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c37eea2_3c8a_424c_8128_0e7307d76af0.slice/crio-1e2da206ff6114ed04370fb235c3011aef54167e2d1bcbcbac5f609252b0e656 WatchSource:0}: Error finding container 1e2da206ff6114ed04370fb235c3011aef54167e2d1bcbcbac5f609252b0e656: Status 404 returned error can't find the container with id 1e2da206ff6114ed04370fb235c3011aef54167e2d1bcbcbac5f609252b0e656 Apr 21 00:08:13.790677 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.790651 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" event={"ID":"6c37eea2-3c8a-424c-8128-0e7307d76af0","Type":"ContainerStarted","Data":"1e2da206ff6114ed04370fb235c3011aef54167e2d1bcbcbac5f609252b0e656"} Apr 21 00:08:13.985916 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.985890 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-68b757865b-hr6vs"] Apr 21 00:08:13.989013 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.988998 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" Apr 21 00:08:13.991115 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.991089 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-92bjt\"" Apr 21 00:08:13.994287 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:13.994268 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-68b757865b-hr6vs"] Apr 21 00:08:14.064024 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.063988 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fs9g\" (UniqueName: \"kubernetes.io/projected/2e300034-30df-4117-9bfe-3ff1253f285c-kube-api-access-4fs9g\") pod \"cert-manager-cainjector-68b757865b-hr6vs\" (UID: \"2e300034-30df-4117-9bfe-3ff1253f285c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" Apr 21 00:08:14.064024 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.064025 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e300034-30df-4117-9bfe-3ff1253f285c-bound-sa-token\") pod \"cert-manager-cainjector-68b757865b-hr6vs\" (UID: \"2e300034-30df-4117-9bfe-3ff1253f285c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" Apr 21 00:08:14.165227 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.165205 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e300034-30df-4117-9bfe-3ff1253f285c-bound-sa-token\") pod \"cert-manager-cainjector-68b757865b-hr6vs\" (UID: \"2e300034-30df-4117-9bfe-3ff1253f285c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" Apr 21 00:08:14.165325 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.165271 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4fs9g\" (UniqueName: \"kubernetes.io/projected/2e300034-30df-4117-9bfe-3ff1253f285c-kube-api-access-4fs9g\") pod \"cert-manager-cainjector-68b757865b-hr6vs\" (UID: \"2e300034-30df-4117-9bfe-3ff1253f285c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" Apr 21 00:08:14.173626 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.173588 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e300034-30df-4117-9bfe-3ff1253f285c-bound-sa-token\") pod \"cert-manager-cainjector-68b757865b-hr6vs\" (UID: \"2e300034-30df-4117-9bfe-3ff1253f285c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" Apr 21 00:08:14.173804 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.173786 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fs9g\" (UniqueName: \"kubernetes.io/projected/2e300034-30df-4117-9bfe-3ff1253f285c-kube-api-access-4fs9g\") pod \"cert-manager-cainjector-68b757865b-hr6vs\" (UID: \"2e300034-30df-4117-9bfe-3ff1253f285c\") " pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" Apr 21 00:08:14.298866 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.298805 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" Apr 21 00:08:14.411914 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.411782 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-68b757865b-hr6vs"] Apr 21 00:08:14.414395 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:08:14.414374 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e300034_30df_4117_9bfe_3ff1253f285c.slice/crio-57ee29faad6e9064b4efea85db8021b01beafc47ce2b46ebbd0a470f5c2cc115 WatchSource:0}: Error finding container 57ee29faad6e9064b4efea85db8021b01beafc47ce2b46ebbd0a470f5c2cc115: Status 404 returned error can't find the container with id 57ee29faad6e9064b4efea85db8021b01beafc47ce2b46ebbd0a470f5c2cc115 Apr 21 00:08:14.796709 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:14.796668 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" event={"ID":"2e300034-30df-4117-9bfe-3ff1253f285c","Type":"ContainerStarted","Data":"57ee29faad6e9064b4efea85db8021b01beafc47ce2b46ebbd0a470f5c2cc115"} Apr 21 00:08:17.807464 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:17.807433 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" event={"ID":"6c37eea2-3c8a-424c-8128-0e7307d76af0","Type":"ContainerStarted","Data":"ccd21f7fe578c9ba2307f31312c45f4f6809c32d76000a69735acad3f82e2973"} Apr 21 00:08:17.807833 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:17.807504 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:17.808692 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:17.808668 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" event={"ID":"2e300034-30df-4117-9bfe-3ff1253f285c","Type":"ContainerStarted","Data":"034b821afe2601705f51dfa5edf8adfd1ac6c9b170e9ebf8931e8e650a2a1dfc"} Apr 21 00:08:17.810137 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:17.810115 2568 generic.go:358] "Generic (PLEG): container finished" podID="46e010ed-5759-4f78-abbc-50827e53e250" containerID="fdc9089e9ee8d4a9e3ea84fe87aa0de8233599d44e8453cd863814f82ec620d1" exitCode=0 Apr 21 00:08:17.810207 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:17.810156 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" event={"ID":"46e010ed-5759-4f78-abbc-50827e53e250","Type":"ContainerDied","Data":"fdc9089e9ee8d4a9e3ea84fe87aa0de8233599d44e8453cd863814f82ec620d1"} Apr 21 00:08:17.823889 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:17.823849 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" podStartSLOduration=1.342879782 podStartE2EDuration="4.823836034s" podCreationTimestamp="2026-04-21 00:08:13 +0000 UTC" firstStartedPulling="2026-04-21 00:08:13.728682639 +0000 UTC m=+307.553661582" lastFinishedPulling="2026-04-21 00:08:17.209638892 +0000 UTC m=+311.034617834" observedRunningTime="2026-04-21 00:08:17.821292826 +0000 UTC m=+311.646271790" watchObservedRunningTime="2026-04-21 00:08:17.823836034 +0000 UTC m=+311.648814998" Apr 21 00:08:17.851239 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:17.851186 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-68b757865b-hr6vs" podStartSLOduration=1.576693044 podStartE2EDuration="4.851171439s" podCreationTimestamp="2026-04-21 00:08:13 +0000 UTC" firstStartedPulling="2026-04-21 00:08:14.416257672 +0000 UTC m=+308.241236615" lastFinishedPulling="2026-04-21 00:08:17.690736064 +0000 UTC m=+311.515715010" observedRunningTime="2026-04-21 00:08:17.849725273 +0000 UTC m=+311.674704239" watchObservedRunningTime="2026-04-21 00:08:17.851171439 +0000 UTC m=+311.676150404" Apr 21 00:08:18.815513 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:18.815481 2568 generic.go:358] "Generic (PLEG): container finished" podID="46e010ed-5759-4f78-abbc-50827e53e250" containerID="874ac0a7bd2cb39f8c48b1134a777fcf3d851edfa3c8ec7e464a802f9fca42a1" exitCode=0 Apr 21 00:08:18.815892 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:18.815572 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" event={"ID":"46e010ed-5759-4f78-abbc-50827e53e250","Type":"ContainerDied","Data":"874ac0a7bd2cb39f8c48b1134a777fcf3d851edfa3c8ec7e464a802f9fca42a1"} Apr 21 00:08:19.932713 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:19.932693 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:20.011557 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.011530 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-util\") pod \"46e010ed-5759-4f78-abbc-50827e53e250\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " Apr 21 00:08:20.011696 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.011582 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-bundle\") pod \"46e010ed-5759-4f78-abbc-50827e53e250\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " Apr 21 00:08:20.011696 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.011636 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqtfh\" (UniqueName: \"kubernetes.io/projected/46e010ed-5759-4f78-abbc-50827e53e250-kube-api-access-mqtfh\") pod \"46e010ed-5759-4f78-abbc-50827e53e250\" (UID: \"46e010ed-5759-4f78-abbc-50827e53e250\") " Apr 21 00:08:20.012002 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.011973 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-bundle" (OuterVolumeSpecName: "bundle") pod "46e010ed-5759-4f78-abbc-50827e53e250" (UID: "46e010ed-5759-4f78-abbc-50827e53e250"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:08:20.013598 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.013575 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46e010ed-5759-4f78-abbc-50827e53e250-kube-api-access-mqtfh" (OuterVolumeSpecName: "kube-api-access-mqtfh") pod "46e010ed-5759-4f78-abbc-50827e53e250" (UID: "46e010ed-5759-4f78-abbc-50827e53e250"). InnerVolumeSpecName "kube-api-access-mqtfh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:08:20.015194 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.015172 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-util" (OuterVolumeSpecName: "util") pod "46e010ed-5759-4f78-abbc-50827e53e250" (UID: "46e010ed-5759-4f78-abbc-50827e53e250"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:08:20.113055 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.113002 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:20.113055 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.113024 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mqtfh\" (UniqueName: \"kubernetes.io/projected/46e010ed-5759-4f78-abbc-50827e53e250-kube-api-access-mqtfh\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:20.113055 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.113034 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46e010ed-5759-4f78-abbc-50827e53e250-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:20.823311 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.823284 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" event={"ID":"46e010ed-5759-4f78-abbc-50827e53e250","Type":"ContainerDied","Data":"58a99be76ab9aac7ef8fed68dcfb2a0686a53834349073f35a36bddf73e9179a"} Apr 21 00:08:20.823311 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.823308 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/77defbb6647882b321b46d16bdefe62633c3425354d3d93571a1d6a87fjqb2h" Apr 21 00:08:20.823485 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:20.823313 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58a99be76ab9aac7ef8fed68dcfb2a0686a53834349073f35a36bddf73e9179a" Apr 21 00:08:23.818256 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:23.818224 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-587ccfb98-hw9l6" Apr 21 00:08:31.776295 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.776259 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz"] Apr 21 00:08:31.776659 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.776508 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="46e010ed-5759-4f78-abbc-50827e53e250" containerName="util" Apr 21 00:08:31.776659 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.776518 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="46e010ed-5759-4f78-abbc-50827e53e250" containerName="util" Apr 21 00:08:31.776659 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.776533 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="46e010ed-5759-4f78-abbc-50827e53e250" containerName="pull" Apr 21 00:08:31.776659 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.776538 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="46e010ed-5759-4f78-abbc-50827e53e250" containerName="pull" Apr 21 00:08:31.776659 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.776550 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="46e010ed-5759-4f78-abbc-50827e53e250" containerName="extract" Apr 21 00:08:31.776659 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.776555 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="46e010ed-5759-4f78-abbc-50827e53e250" containerName="extract" Apr 21 00:08:31.776659 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.776601 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="46e010ed-5759-4f78-abbc-50827e53e250" containerName="extract" Apr 21 00:08:31.826996 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.826969 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz"] Apr 21 00:08:31.827115 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.827069 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:31.830236 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.830217 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 21 00:08:31.830504 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.830487 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 21 00:08:31.831671 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.831647 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-4fqdf\"" Apr 21 00:08:31.893716 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.893693 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-util\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:31.893804 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.893726 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-bundle\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:31.893804 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.893745 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwpct\" (UniqueName: \"kubernetes.io/projected/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-kube-api-access-kwpct\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:31.994671 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.994643 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-util\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:31.994671 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.994674 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-bundle\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:31.994846 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.994698 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kwpct\" (UniqueName: \"kubernetes.io/projected/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-kube-api-access-kwpct\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:31.994998 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.994980 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-util\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:31.995056 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:31.995038 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-bundle\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:32.001968 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:32.001948 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwpct\" (UniqueName: \"kubernetes.io/projected/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-kube-api-access-kwpct\") pod \"3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:32.136063 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:32.135992 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:32.250772 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:32.250619 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz"] Apr 21 00:08:32.253304 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:08:32.253272 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ca66339_8e85_4d6e_90f1_ea77a9f827ce.slice/crio-3ea602d8719d77a36dc2850f0db59e2456e48eab7ab93aaae08484f0adc651bc WatchSource:0}: Error finding container 3ea602d8719d77a36dc2850f0db59e2456e48eab7ab93aaae08484f0adc651bc: Status 404 returned error can't find the container with id 3ea602d8719d77a36dc2850f0db59e2456e48eab7ab93aaae08484f0adc651bc Apr 21 00:08:32.861740 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:32.861706 2568 generic.go:358] "Generic (PLEG): container finished" podID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerID="f32b82f8e987eaafc3634c249698fb93e58dd2e316bb551bc4696d889468d8bf" exitCode=0 Apr 21 00:08:32.862098 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:32.861795 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" event={"ID":"4ca66339-8e85-4d6e-90f1-ea77a9f827ce","Type":"ContainerDied","Data":"f32b82f8e987eaafc3634c249698fb93e58dd2e316bb551bc4696d889468d8bf"} Apr 21 00:08:32.862098 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:32.861826 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" event={"ID":"4ca66339-8e85-4d6e-90f1-ea77a9f827ce","Type":"ContainerStarted","Data":"3ea602d8719d77a36dc2850f0db59e2456e48eab7ab93aaae08484f0adc651bc"} Apr 21 00:08:33.866505 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:33.866477 2568 generic.go:358] "Generic (PLEG): container finished" podID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerID="0cc7a6360b5cebc5514adddde3fafd7b672aeb44b6dc5664fab32e6730c28fcb" exitCode=0 Apr 21 00:08:33.866872 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:33.866536 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" event={"ID":"4ca66339-8e85-4d6e-90f1-ea77a9f827ce","Type":"ContainerDied","Data":"0cc7a6360b5cebc5514adddde3fafd7b672aeb44b6dc5664fab32e6730c28fcb"} Apr 21 00:08:34.870920 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:34.870885 2568 generic.go:358] "Generic (PLEG): container finished" podID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerID="3ad4ee6fce29352520c40396c4e1ff203dc2b4e75d922139a412fcd5c9d0c8ac" exitCode=0 Apr 21 00:08:34.871290 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:34.870959 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" event={"ID":"4ca66339-8e85-4d6e-90f1-ea77a9f827ce","Type":"ContainerDied","Data":"3ad4ee6fce29352520c40396c4e1ff203dc2b4e75d922139a412fcd5c9d0c8ac"} Apr 21 00:08:35.988134 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:35.988113 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:36.025723 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.025697 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-util\") pod \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " Apr 21 00:08:36.025870 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.025739 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-bundle\") pod \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " Apr 21 00:08:36.025870 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.025768 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwpct\" (UniqueName: \"kubernetes.io/projected/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-kube-api-access-kwpct\") pod \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\" (UID: \"4ca66339-8e85-4d6e-90f1-ea77a9f827ce\") " Apr 21 00:08:36.026578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.026540 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-bundle" (OuterVolumeSpecName: "bundle") pod "4ca66339-8e85-4d6e-90f1-ea77a9f827ce" (UID: "4ca66339-8e85-4d6e-90f1-ea77a9f827ce"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:08:36.027936 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.027909 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-kube-api-access-kwpct" (OuterVolumeSpecName: "kube-api-access-kwpct") pod "4ca66339-8e85-4d6e-90f1-ea77a9f827ce" (UID: "4ca66339-8e85-4d6e-90f1-ea77a9f827ce"). InnerVolumeSpecName "kube-api-access-kwpct". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:08:36.031819 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.031790 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-util" (OuterVolumeSpecName: "util") pod "4ca66339-8e85-4d6e-90f1-ea77a9f827ce" (UID: "4ca66339-8e85-4d6e-90f1-ea77a9f827ce"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:08:36.126503 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.126429 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:36.126503 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.126454 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:36.126503 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.126464 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kwpct\" (UniqueName: \"kubernetes.io/projected/4ca66339-8e85-4d6e-90f1-ea77a9f827ce-kube-api-access-kwpct\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:36.878419 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.878382 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" event={"ID":"4ca66339-8e85-4d6e-90f1-ea77a9f827ce","Type":"ContainerDied","Data":"3ea602d8719d77a36dc2850f0db59e2456e48eab7ab93aaae08484f0adc651bc"} Apr 21 00:08:36.878563 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.878427 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ea602d8719d77a36dc2850f0db59e2456e48eab7ab93aaae08484f0adc651bc" Apr 21 00:08:36.878563 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:36.878399 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3081035efbc3f76b8ae1b663abbd2a6bc5b4896fe94818011b247755c5665dz" Apr 21 00:08:38.384518 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.384486 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-lws-operator/lws-controller-manager-5956879555-jvtz9"] Apr 21 00:08:38.384963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.384800 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerName="extract" Apr 21 00:08:38.384963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.384812 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerName="extract" Apr 21 00:08:38.384963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.384828 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerName="pull" Apr 21 00:08:38.384963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.384833 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerName="pull" Apr 21 00:08:38.384963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.384843 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerName="util" Apr 21 00:08:38.384963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.384848 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerName="util" Apr 21 00:08:38.384963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.384893 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="4ca66339-8e85-4d6e-90f1-ea77a9f827ce" containerName="extract" Apr 21 00:08:38.388878 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.388862 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.391525 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.391495 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"metrics-server-cert\"" Apr 21 00:08:38.391657 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.391558 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"webhook-server-cert\"" Apr 21 00:08:38.392646 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.392586 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"lws-manager-config\"" Apr 21 00:08:38.392646 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.392599 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"openshift-service-ca.crt\"" Apr 21 00:08:38.392814 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.392635 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"kube-root-ca.crt\"" Apr 21 00:08:38.392814 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.392652 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"lws-controller-manager-dockercfg-b9t7h\"" Apr 21 00:08:38.395163 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.395140 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-lws-operator/lws-controller-manager-5956879555-jvtz9"] Apr 21 00:08:38.441601 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.441569 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-cert\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.441762 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.441642 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qljcl\" (UniqueName: \"kubernetes.io/projected/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-kube-api-access-qljcl\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.441762 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.441678 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-metrics-cert\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.441762 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.441715 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-manager-config\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.542895 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.542865 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qljcl\" (UniqueName: \"kubernetes.io/projected/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-kube-api-access-qljcl\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.543024 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.542903 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-metrics-cert\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.543024 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.542928 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-manager-config\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.543024 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.542980 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-cert\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.543664 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.543644 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-manager-config\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.545427 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.545397 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-metrics-cert\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.545525 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.545480 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-cert\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.561534 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.561498 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qljcl\" (UniqueName: \"kubernetes.io/projected/3975eb23-f72b-46ae-82d8-4eb8e6eb58b6-kube-api-access-qljcl\") pod \"lws-controller-manager-5956879555-jvtz9\" (UID: \"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6\") " pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.698740 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.698661 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:38.821844 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.821804 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-lws-operator/lws-controller-manager-5956879555-jvtz9"] Apr 21 00:08:38.823727 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:08:38.823701 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3975eb23_f72b_46ae_82d8_4eb8e6eb58b6.slice/crio-ceef94460d0a84bff33842bc669d26140a0d2c106ed2c31d20ac21cac59efb78 WatchSource:0}: Error finding container ceef94460d0a84bff33842bc669d26140a0d2c106ed2c31d20ac21cac59efb78: Status 404 returned error can't find the container with id ceef94460d0a84bff33842bc669d26140a0d2c106ed2c31d20ac21cac59efb78 Apr 21 00:08:38.885026 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:38.884993 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" event={"ID":"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6","Type":"ContainerStarted","Data":"ceef94460d0a84bff33842bc669d26140a0d2c106ed2c31d20ac21cac59efb78"} Apr 21 00:08:40.892260 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:40.892226 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" event={"ID":"3975eb23-f72b-46ae-82d8-4eb8e6eb58b6","Type":"ContainerStarted","Data":"6dd6d2c0ebecc1e46583c29e07b95ab9f4085c641150e2f98b6e889802062b47"} Apr 21 00:08:40.892703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:40.892317 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:40.908296 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:40.908245 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" podStartSLOduration=1.351889364 podStartE2EDuration="2.90822954s" podCreationTimestamp="2026-04-21 00:08:38 +0000 UTC" firstStartedPulling="2026-04-21 00:08:38.825421263 +0000 UTC m=+332.650400209" lastFinishedPulling="2026-04-21 00:08:40.381761217 +0000 UTC m=+334.206740385" observedRunningTime="2026-04-21 00:08:40.906520564 +0000 UTC m=+334.731499536" watchObservedRunningTime="2026-04-21 00:08:40.90822954 +0000 UTC m=+334.733208507" Apr 21 00:08:48.654294 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.654258 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6"] Apr 21 00:08:48.657626 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.657592 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.660486 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.660468 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"kube-root-ca.crt\"" Apr 21 00:08:48.660596 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.660498 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-webhook-cert\"" Apr 21 00:08:48.660596 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.660574 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-manager-dockercfg-fjxgs\"" Apr 21 00:08:48.660723 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.660574 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-manager-service-cert\"" Apr 21 00:08:48.661023 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.661008 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"openshift-service-ca.crt\"" Apr 21 00:08:48.670053 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.670034 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6"] Apr 21 00:08:48.718442 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.718424 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cfa3adc8-671e-4d7b-b87d-e3eedc861363-webhook-cert\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.718541 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.718451 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6jp5\" (UniqueName: \"kubernetes.io/projected/cfa3adc8-671e-4d7b-b87d-e3eedc861363-kube-api-access-d6jp5\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.718541 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.718477 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cfa3adc8-671e-4d7b-b87d-e3eedc861363-apiservice-cert\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.819471 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.819444 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cfa3adc8-671e-4d7b-b87d-e3eedc861363-webhook-cert\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.819572 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.819479 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d6jp5\" (UniqueName: \"kubernetes.io/projected/cfa3adc8-671e-4d7b-b87d-e3eedc861363-kube-api-access-d6jp5\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.819572 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.819537 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cfa3adc8-671e-4d7b-b87d-e3eedc861363-apiservice-cert\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.821950 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.821930 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cfa3adc8-671e-4d7b-b87d-e3eedc861363-apiservice-cert\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.822031 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.822010 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cfa3adc8-671e-4d7b-b87d-e3eedc861363-webhook-cert\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.828435 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.828415 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6jp5\" (UniqueName: \"kubernetes.io/projected/cfa3adc8-671e-4d7b-b87d-e3eedc861363-kube-api-access-d6jp5\") pod \"opendatahub-operator-controller-manager-587f5698df-hptf6\" (UID: \"cfa3adc8-671e-4d7b-b87d-e3eedc861363\") " pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:48.947166 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.947101 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp"] Apr 21 00:08:48.950490 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.950472 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:48.952926 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.952904 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 21 00:08:48.952926 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.952918 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 21 00:08:48.953079 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.952921 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-4fqdf\"" Apr 21 00:08:48.957394 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.957373 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp"] Apr 21 00:08:48.968174 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:48.968152 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:49.029691 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.024022 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-bundle\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.029691 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.024134 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhjfr\" (UniqueName: \"kubernetes.io/projected/b68577b8-fe7e-43d6-9d49-b8d64ba06165-kube-api-access-vhjfr\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.029691 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.024177 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-util\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.110323 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.108395 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6"] Apr 21 00:08:49.112230 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:08:49.112203 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfa3adc8_671e_4d7b_b87d_e3eedc861363.slice/crio-a5a73057c644392e8a55d2f7666ba4b2766b49e7d73114a8b90e15ecbe7de719 WatchSource:0}: Error finding container a5a73057c644392e8a55d2f7666ba4b2766b49e7d73114a8b90e15ecbe7de719: Status 404 returned error can't find the container with id a5a73057c644392e8a55d2f7666ba4b2766b49e7d73114a8b90e15ecbe7de719 Apr 21 00:08:49.125432 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.125408 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vhjfr\" (UniqueName: \"kubernetes.io/projected/b68577b8-fe7e-43d6-9d49-b8d64ba06165-kube-api-access-vhjfr\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.125525 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.125442 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-util\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.125525 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.125471 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-bundle\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.125794 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.125780 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-bundle\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.125836 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.125815 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-util\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.136571 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.136542 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhjfr\" (UniqueName: \"kubernetes.io/projected/b68577b8-fe7e-43d6-9d49-b8d64ba06165-kube-api-access-vhjfr\") pod \"f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.262137 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.262116 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:49.375031 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.375010 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp"] Apr 21 00:08:49.376969 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:08:49.376939 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb68577b8_fe7e_43d6_9d49_b8d64ba06165.slice/crio-f504b197d1427228d0ceb8599a4fe60cd34211f6fb23cc6dc56c31f2b7c85470 WatchSource:0}: Error finding container f504b197d1427228d0ceb8599a4fe60cd34211f6fb23cc6dc56c31f2b7c85470: Status 404 returned error can't find the container with id f504b197d1427228d0ceb8599a4fe60cd34211f6fb23cc6dc56c31f2b7c85470 Apr 21 00:08:49.923696 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.923652 2568 generic.go:358] "Generic (PLEG): container finished" podID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerID="0bb5c62293aea4975d52d90294450cbcdb9275a5a678a56c91cc38f39b98e33a" exitCode=0 Apr 21 00:08:49.924092 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.923724 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" event={"ID":"b68577b8-fe7e-43d6-9d49-b8d64ba06165","Type":"ContainerDied","Data":"0bb5c62293aea4975d52d90294450cbcdb9275a5a678a56c91cc38f39b98e33a"} Apr 21 00:08:49.924092 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.923766 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" event={"ID":"b68577b8-fe7e-43d6-9d49-b8d64ba06165","Type":"ContainerStarted","Data":"f504b197d1427228d0ceb8599a4fe60cd34211f6fb23cc6dc56c31f2b7c85470"} Apr 21 00:08:49.925311 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:49.925280 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" event={"ID":"cfa3adc8-671e-4d7b-b87d-e3eedc861363","Type":"ContainerStarted","Data":"a5a73057c644392e8a55d2f7666ba4b2766b49e7d73114a8b90e15ecbe7de719"} Apr 21 00:08:51.898475 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:51.898451 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-lws-operator/lws-controller-manager-5956879555-jvtz9" Apr 21 00:08:51.935579 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:51.935547 2568 generic.go:358] "Generic (PLEG): container finished" podID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerID="f12727214c562f19c888888f7ec68963944f33603bfd8135da17c1809799c89c" exitCode=0 Apr 21 00:08:51.935719 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:51.935582 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" event={"ID":"b68577b8-fe7e-43d6-9d49-b8d64ba06165","Type":"ContainerDied","Data":"f12727214c562f19c888888f7ec68963944f33603bfd8135da17c1809799c89c"} Apr 21 00:08:51.937372 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:51.937350 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" event={"ID":"cfa3adc8-671e-4d7b-b87d-e3eedc861363","Type":"ContainerStarted","Data":"9ed9a4492ba3721ea93c7f258830046e357a883082b3969f291f38593be72f8e"} Apr 21 00:08:51.937496 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:51.937399 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:08:51.974053 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:51.973705 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" podStartSLOduration=1.625678469 podStartE2EDuration="3.973684799s" podCreationTimestamp="2026-04-21 00:08:48 +0000 UTC" firstStartedPulling="2026-04-21 00:08:49.113987347 +0000 UTC m=+342.938966290" lastFinishedPulling="2026-04-21 00:08:51.461993674 +0000 UTC m=+345.286972620" observedRunningTime="2026-04-21 00:08:51.971938106 +0000 UTC m=+345.796917072" watchObservedRunningTime="2026-04-21 00:08:51.973684799 +0000 UTC m=+345.798663765" Apr 21 00:08:52.942675 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:52.942639 2568 generic.go:358] "Generic (PLEG): container finished" podID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerID="4ab82f4c070e07ca4d835be0d8c1eabbab98956d6abbd83c5d60300e3492bdff" exitCode=0 Apr 21 00:08:52.943074 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:52.942720 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" event={"ID":"b68577b8-fe7e-43d6-9d49-b8d64ba06165","Type":"ContainerDied","Data":"4ab82f4c070e07ca4d835be0d8c1eabbab98956d6abbd83c5d60300e3492bdff"} Apr 21 00:08:54.063965 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.063945 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:08:54.168018 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.167980 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhjfr\" (UniqueName: \"kubernetes.io/projected/b68577b8-fe7e-43d6-9d49-b8d64ba06165-kube-api-access-vhjfr\") pod \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " Apr 21 00:08:54.168211 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.168027 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-util\") pod \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " Apr 21 00:08:54.168211 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.168085 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-bundle\") pod \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\" (UID: \"b68577b8-fe7e-43d6-9d49-b8d64ba06165\") " Apr 21 00:08:54.168945 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.168907 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-bundle" (OuterVolumeSpecName: "bundle") pod "b68577b8-fe7e-43d6-9d49-b8d64ba06165" (UID: "b68577b8-fe7e-43d6-9d49-b8d64ba06165"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:08:54.170174 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.170153 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b68577b8-fe7e-43d6-9d49-b8d64ba06165-kube-api-access-vhjfr" (OuterVolumeSpecName: "kube-api-access-vhjfr") pod "b68577b8-fe7e-43d6-9d49-b8d64ba06165" (UID: "b68577b8-fe7e-43d6-9d49-b8d64ba06165"). InnerVolumeSpecName "kube-api-access-vhjfr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:08:54.173889 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.173864 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-util" (OuterVolumeSpecName: "util") pod "b68577b8-fe7e-43d6-9d49-b8d64ba06165" (UID: "b68577b8-fe7e-43d6-9d49-b8d64ba06165"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:08:54.269746 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.269712 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vhjfr\" (UniqueName: \"kubernetes.io/projected/b68577b8-fe7e-43d6-9d49-b8d64ba06165-kube-api-access-vhjfr\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:54.269746 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.269743 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:54.269909 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.269759 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b68577b8-fe7e-43d6-9d49-b8d64ba06165-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:08:54.951733 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.951659 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" event={"ID":"b68577b8-fe7e-43d6-9d49-b8d64ba06165","Type":"ContainerDied","Data":"f504b197d1427228d0ceb8599a4fe60cd34211f6fb23cc6dc56c31f2b7c85470"} Apr 21 00:08:54.951733 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.951693 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f504b197d1427228d0ceb8599a4fe60cd34211f6fb23cc6dc56c31f2b7c85470" Apr 21 00:08:54.951942 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:08:54.951757 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f77c1a82ade775f7608969957d57ee0b8db93eeada9825bd6f7f7156c94bgkp" Apr 21 00:09:02.945584 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:02.945555 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/opendatahub-operator-controller-manager-587f5698df-hptf6" Apr 21 00:09:18.383809 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.383768 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5"] Apr 21 00:09:18.384362 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.384164 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerName="util" Apr 21 00:09:18.384362 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.384184 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerName="util" Apr 21 00:09:18.384362 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.384223 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerName="extract" Apr 21 00:09:18.384362 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.384231 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerName="extract" Apr 21 00:09:18.384362 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.384242 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerName="pull" Apr 21 00:09:18.384362 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.384252 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerName="pull" Apr 21 00:09:18.384362 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.384328 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="b68577b8-fe7e-43d6-9d49-b8d64ba06165" containerName="extract" Apr 21 00:09:18.386944 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.386922 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.389199 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.389169 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 21 00:09:18.390199 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.390176 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-4fqdf\"" Apr 21 00:09:18.390331 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.390217 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 21 00:09:18.394046 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.394013 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5"] Apr 21 00:09:18.557902 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.557863 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl96f\" (UniqueName: \"kubernetes.io/projected/a3dc759e-2026-45e2-9002-af697fa66db0-kube-api-access-vl96f\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.558089 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.557909 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-bundle\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.558089 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.558046 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-util\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.659028 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.658942 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-util\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.659028 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.659016 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vl96f\" (UniqueName: \"kubernetes.io/projected/a3dc759e-2026-45e2-9002-af697fa66db0-kube-api-access-vl96f\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.659259 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.659049 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-bundle\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.659438 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.659412 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-util\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.659510 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.659426 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-bundle\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.666439 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.666406 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl96f\" (UniqueName: \"kubernetes.io/projected/a3dc759e-2026-45e2-9002-af697fa66db0-kube-api-access-vl96f\") pod \"4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.696879 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.696859 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:18.815455 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:18.815434 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5"] Apr 21 00:09:18.817486 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:09:18.817456 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3dc759e_2026_45e2_9002_af697fa66db0.slice/crio-dee842c91a11b0d0995996e53a3df8f8aa669e3eaa82c9487ef5e803269f03af WatchSource:0}: Error finding container dee842c91a11b0d0995996e53a3df8f8aa669e3eaa82c9487ef5e803269f03af: Status 404 returned error can't find the container with id dee842c91a11b0d0995996e53a3df8f8aa669e3eaa82c9487ef5e803269f03af Apr 21 00:09:19.034392 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:19.034356 2568 generic.go:358] "Generic (PLEG): container finished" podID="a3dc759e-2026-45e2-9002-af697fa66db0" containerID="0b0ff37981094e4fc4a8b836c4287dbb23b248ecc90cc24425fec45b86cadca6" exitCode=0 Apr 21 00:09:19.034543 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:19.034409 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" event={"ID":"a3dc759e-2026-45e2-9002-af697fa66db0","Type":"ContainerDied","Data":"0b0ff37981094e4fc4a8b836c4287dbb23b248ecc90cc24425fec45b86cadca6"} Apr 21 00:09:19.034543 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:19.034431 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" event={"ID":"a3dc759e-2026-45e2-9002-af697fa66db0","Type":"ContainerStarted","Data":"dee842c91a11b0d0995996e53a3df8f8aa669e3eaa82c9487ef5e803269f03af"} Apr 21 00:09:20.039012 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:20.038980 2568 generic.go:358] "Generic (PLEG): container finished" podID="a3dc759e-2026-45e2-9002-af697fa66db0" containerID="c257c71582f54295805a7fa59ee095f1cfb74eac9bd1e9816950f344575a8251" exitCode=0 Apr 21 00:09:20.039330 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:20.039056 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" event={"ID":"a3dc759e-2026-45e2-9002-af697fa66db0","Type":"ContainerDied","Data":"c257c71582f54295805a7fa59ee095f1cfb74eac9bd1e9816950f344575a8251"} Apr 21 00:09:21.044040 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:21.044006 2568 generic.go:358] "Generic (PLEG): container finished" podID="a3dc759e-2026-45e2-9002-af697fa66db0" containerID="4b04257b8d199154a94d3f3de9fcdfe6adc3a0dd0fbbc19daec574c554d4b492" exitCode=0 Apr 21 00:09:21.044459 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:21.044079 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" event={"ID":"a3dc759e-2026-45e2-9002-af697fa66db0","Type":"ContainerDied","Data":"4b04257b8d199154a94d3f3de9fcdfe6adc3a0dd0fbbc19daec574c554d4b492"} Apr 21 00:09:22.163579 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.163557 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:22.289836 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.289807 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-bundle\") pod \"a3dc759e-2026-45e2-9002-af697fa66db0\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " Apr 21 00:09:22.289996 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.289853 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-util\") pod \"a3dc759e-2026-45e2-9002-af697fa66db0\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " Apr 21 00:09:22.289996 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.289920 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl96f\" (UniqueName: \"kubernetes.io/projected/a3dc759e-2026-45e2-9002-af697fa66db0-kube-api-access-vl96f\") pod \"a3dc759e-2026-45e2-9002-af697fa66db0\" (UID: \"a3dc759e-2026-45e2-9002-af697fa66db0\") " Apr 21 00:09:22.290979 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.290945 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-bundle" (OuterVolumeSpecName: "bundle") pod "a3dc759e-2026-45e2-9002-af697fa66db0" (UID: "a3dc759e-2026-45e2-9002-af697fa66db0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:09:22.292009 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.291982 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3dc759e-2026-45e2-9002-af697fa66db0-kube-api-access-vl96f" (OuterVolumeSpecName: "kube-api-access-vl96f") pod "a3dc759e-2026-45e2-9002-af697fa66db0" (UID: "a3dc759e-2026-45e2-9002-af697fa66db0"). InnerVolumeSpecName "kube-api-access-vl96f". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:09:22.297533 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.297472 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-util" (OuterVolumeSpecName: "util") pod "a3dc759e-2026-45e2-9002-af697fa66db0" (UID: "a3dc759e-2026-45e2-9002-af697fa66db0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:09:22.391452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.391427 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:09:22.391452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.391448 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3dc759e-2026-45e2-9002-af697fa66db0-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:09:22.391589 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:22.391458 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vl96f\" (UniqueName: \"kubernetes.io/projected/a3dc759e-2026-45e2-9002-af697fa66db0-kube-api-access-vl96f\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:09:23.052010 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:23.051935 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" event={"ID":"a3dc759e-2026-45e2-9002-af697fa66db0","Type":"ContainerDied","Data":"dee842c91a11b0d0995996e53a3df8f8aa669e3eaa82c9487ef5e803269f03af"} Apr 21 00:09:23.052010 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:23.051967 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dee842c91a11b0d0995996e53a3df8f8aa669e3eaa82c9487ef5e803269f03af" Apr 21 00:09:23.052010 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:23.051989 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4c892a6d2d8a57acbc49427d3a7d24e253bab25be2c607ef405d2c4835ls2h5" Apr 21 00:09:32.936280 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.936250 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt"] Apr 21 00:09:32.936650 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.936545 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a3dc759e-2026-45e2-9002-af697fa66db0" containerName="pull" Apr 21 00:09:32.936650 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.936557 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3dc759e-2026-45e2-9002-af697fa66db0" containerName="pull" Apr 21 00:09:32.936650 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.936571 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a3dc759e-2026-45e2-9002-af697fa66db0" containerName="util" Apr 21 00:09:32.936650 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.936579 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3dc759e-2026-45e2-9002-af697fa66db0" containerName="util" Apr 21 00:09:32.936650 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.936587 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a3dc759e-2026-45e2-9002-af697fa66db0" containerName="extract" Apr 21 00:09:32.936650 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.936592 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3dc759e-2026-45e2-9002-af697fa66db0" containerName="extract" Apr 21 00:09:32.936650 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.936649 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="a3dc759e-2026-45e2-9002-af697fa66db0" containerName="extract" Apr 21 00:09:32.939878 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.939852 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:32.942465 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.942433 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 21 00:09:32.942732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.942714 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 21 00:09:32.943697 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.943677 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-4fqdf\"" Apr 21 00:09:32.951625 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.951592 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt"] Apr 21 00:09:32.961672 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.961631 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-util\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:32.961950 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.961931 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-bundle\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:32.962045 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:32.962014 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mzkd\" (UniqueName: \"kubernetes.io/projected/5219e693-c91a-47cc-b7e0-bd0f20586e01-kube-api-access-9mzkd\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:33.063278 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:33.063249 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-util\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:33.063405 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:33.063296 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-bundle\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:33.063405 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:33.063344 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9mzkd\" (UniqueName: \"kubernetes.io/projected/5219e693-c91a-47cc-b7e0-bd0f20586e01-kube-api-access-9mzkd\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:33.063659 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:33.063642 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-util\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:33.063762 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:33.063743 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-bundle\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:33.072051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:33.072018 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mzkd\" (UniqueName: \"kubernetes.io/projected/5219e693-c91a-47cc-b7e0-bd0f20586e01-kube-api-access-9mzkd\") pod \"d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:33.250684 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:33.250653 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:33.379362 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:33.379242 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt"] Apr 21 00:09:33.382672 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:09:33.382642 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5219e693_c91a_47cc_b7e0_bd0f20586e01.slice/crio-16e0c8dd9653e5c2e9c4a863a695b12be20395bb039ffbea498e63316068c04c WatchSource:0}: Error finding container 16e0c8dd9653e5c2e9c4a863a695b12be20395bb039ffbea498e63316068c04c: Status 404 returned error can't find the container with id 16e0c8dd9653e5c2e9c4a863a695b12be20395bb039ffbea498e63316068c04c Apr 21 00:09:34.088364 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:34.088339 2568 generic.go:358] "Generic (PLEG): container finished" podID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerID="4ec2008c37043056399ae29aa88f0873fedc7c5f4adc8dd21a65650fdb5588fd" exitCode=0 Apr 21 00:09:34.088689 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:34.088379 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" event={"ID":"5219e693-c91a-47cc-b7e0-bd0f20586e01","Type":"ContainerDied","Data":"4ec2008c37043056399ae29aa88f0873fedc7c5f4adc8dd21a65650fdb5588fd"} Apr 21 00:09:34.088689 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:34.088400 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" event={"ID":"5219e693-c91a-47cc-b7e0-bd0f20586e01","Type":"ContainerStarted","Data":"16e0c8dd9653e5c2e9c4a863a695b12be20395bb039ffbea498e63316068c04c"} Apr 21 00:09:36.096282 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:36.096249 2568 generic.go:358] "Generic (PLEG): container finished" podID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerID="8ed5ae57ea11a76f11c86513fe091b843655755b9e7db271b06286e285b737a3" exitCode=0 Apr 21 00:09:36.096766 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:36.096329 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" event={"ID":"5219e693-c91a-47cc-b7e0-bd0f20586e01","Type":"ContainerDied","Data":"8ed5ae57ea11a76f11c86513fe091b843655755b9e7db271b06286e285b737a3"} Apr 21 00:09:37.100679 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:37.100649 2568 generic.go:358] "Generic (PLEG): container finished" podID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerID="a8764230855ba9c36355481d666346f561f5a052d1cfcc1df68bbc278260be8f" exitCode=0 Apr 21 00:09:37.100998 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:37.100697 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" event={"ID":"5219e693-c91a-47cc-b7e0-bd0f20586e01","Type":"ContainerDied","Data":"a8764230855ba9c36355481d666346f561f5a052d1cfcc1df68bbc278260be8f"} Apr 21 00:09:38.224720 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.224697 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:38.309789 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.309762 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mzkd\" (UniqueName: \"kubernetes.io/projected/5219e693-c91a-47cc-b7e0-bd0f20586e01-kube-api-access-9mzkd\") pod \"5219e693-c91a-47cc-b7e0-bd0f20586e01\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " Apr 21 00:09:38.309911 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.309820 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-util\") pod \"5219e693-c91a-47cc-b7e0-bd0f20586e01\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " Apr 21 00:09:38.309911 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.309855 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-bundle\") pod \"5219e693-c91a-47cc-b7e0-bd0f20586e01\" (UID: \"5219e693-c91a-47cc-b7e0-bd0f20586e01\") " Apr 21 00:09:38.310730 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.310694 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-bundle" (OuterVolumeSpecName: "bundle") pod "5219e693-c91a-47cc-b7e0-bd0f20586e01" (UID: "5219e693-c91a-47cc-b7e0-bd0f20586e01"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:09:38.311938 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.311909 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5219e693-c91a-47cc-b7e0-bd0f20586e01-kube-api-access-9mzkd" (OuterVolumeSpecName: "kube-api-access-9mzkd") pod "5219e693-c91a-47cc-b7e0-bd0f20586e01" (UID: "5219e693-c91a-47cc-b7e0-bd0f20586e01"). InnerVolumeSpecName "kube-api-access-9mzkd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:09:38.314760 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.314730 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-util" (OuterVolumeSpecName: "util") pod "5219e693-c91a-47cc-b7e0-bd0f20586e01" (UID: "5219e693-c91a-47cc-b7e0-bd0f20586e01"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:09:38.410908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.410856 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9mzkd\" (UniqueName: \"kubernetes.io/projected/5219e693-c91a-47cc-b7e0-bd0f20586e01-kube-api-access-9mzkd\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:09:38.410908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.410879 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:09:38.410908 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:38.410889 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5219e693-c91a-47cc-b7e0-bd0f20586e01-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:09:39.108596 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:39.108515 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" event={"ID":"5219e693-c91a-47cc-b7e0-bd0f20586e01","Type":"ContainerDied","Data":"16e0c8dd9653e5c2e9c4a863a695b12be20395bb039ffbea498e63316068c04c"} Apr 21 00:09:39.108596 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:39.108546 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16e0c8dd9653e5c2e9c4a863a695b12be20395bb039ffbea498e63316068c04c" Apr 21 00:09:39.108596 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:39.108571 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d813cbca2f2d4ae5bb9c3e9ca6bc9dc97fa22f4f10cc797dd3b2c1f0c2pt8dt" Apr 21 00:09:50.055318 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.055282 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb"] Apr 21 00:09:50.055805 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.055562 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerName="pull" Apr 21 00:09:50.055805 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.055573 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerName="pull" Apr 21 00:09:50.055805 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.055583 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerName="extract" Apr 21 00:09:50.055805 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.055589 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerName="extract" Apr 21 00:09:50.055805 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.055625 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerName="util" Apr 21 00:09:50.055805 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.055631 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerName="util" Apr 21 00:09:50.055805 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.055676 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="5219e693-c91a-47cc-b7e0-bd0f20586e01" containerName="extract" Apr 21 00:09:50.064400 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.064374 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.067270 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.067232 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Apr 21 00:09:50.067526 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.067271 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"data-science-gateway-data-science-gateway-class-dockercfg-55m7h\"" Apr 21 00:09:50.067779 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.067401 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"istio-ca-root-cert\"" Apr 21 00:09:50.068075 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.068056 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Apr 21 00:09:50.070858 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.070839 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb"] Apr 21 00:09:50.094575 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094550 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-istio-envoy\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.094686 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094582 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/08853a86-b83e-428f-a936-f74a943afee5-istio-podinfo\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.094686 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094629 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/08853a86-b83e-428f-a936-f74a943afee5-istio-token\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.094686 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094648 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/08853a86-b83e-428f-a936-f74a943afee5-istiod-ca-cert\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.094686 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094667 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-credential-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.094686 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094685 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr765\" (UniqueName: \"kubernetes.io/projected/08853a86-b83e-428f-a936-f74a943afee5-kube-api-access-xr765\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.094899 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094703 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-workload-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.094899 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094820 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-workload-certs\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.094899 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.094882 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-istio-data\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196122 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196094 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/08853a86-b83e-428f-a936-f74a943afee5-istio-token\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196233 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196126 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/08853a86-b83e-428f-a936-f74a943afee5-istiod-ca-cert\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196233 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196146 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-credential-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196233 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196163 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xr765\" (UniqueName: \"kubernetes.io/projected/08853a86-b83e-428f-a936-f74a943afee5-kube-api-access-xr765\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196233 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196206 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-workload-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196493 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196456 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-workload-certs\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196493 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196482 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-credential-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196649 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196537 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-istio-data\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196649 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196560 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-istio-envoy\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196649 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196581 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-workload-socket\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196862 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196654 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/08853a86-b83e-428f-a936-f74a943afee5-istio-podinfo\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196862 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196780 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-workload-certs\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196872 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-istio-data\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.196963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.196894 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/08853a86-b83e-428f-a936-f74a943afee5-istiod-ca-cert\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.198544 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.198520 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/08853a86-b83e-428f-a936-f74a943afee5-istio-envoy\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.199004 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.198981 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/08853a86-b83e-428f-a936-f74a943afee5-istio-podinfo\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.218294 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.218268 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/08853a86-b83e-428f-a936-f74a943afee5-istio-token\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.219324 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.219300 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr765\" (UniqueName: \"kubernetes.io/projected/08853a86-b83e-428f-a936-f74a943afee5-kube-api-access-xr765\") pod \"data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb\" (UID: \"08853a86-b83e-428f-a936-f74a943afee5\") " pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.377522 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.377439 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:50.527101 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:50.527068 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb"] Apr 21 00:09:50.531568 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:09:50.531539 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08853a86_b83e_428f_a936_f74a943afee5.slice/crio-ef6745984e6093a6606535411260c073c00b8712b8d3118262ad76770bd78126 WatchSource:0}: Error finding container ef6745984e6093a6606535411260c073c00b8712b8d3118262ad76770bd78126: Status 404 returned error can't find the container with id ef6745984e6093a6606535411260c073c00b8712b8d3118262ad76770bd78126 Apr 21 00:09:51.148063 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:51.148031 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" event={"ID":"08853a86-b83e-428f-a936-f74a943afee5","Type":"ContainerStarted","Data":"ef6745984e6093a6606535411260c073c00b8712b8d3118262ad76770bd78126"} Apr 21 00:09:53.955646 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:53.955590 2568 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"31236220Ki","pods":"250"} Apr 21 00:09:53.955877 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:53.955695 2568 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"31236220Ki","pods":"250"} Apr 21 00:09:53.955877 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:53.955727 2568 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"31236220Ki","pods":"250"} Apr 21 00:09:54.160770 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:54.160736 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" event={"ID":"08853a86-b83e-428f-a936-f74a943afee5","Type":"ContainerStarted","Data":"9d7e86504babc1290fb1c662a26983d439dbe2a37fc28272c4fbc3674d732641"} Apr 21 00:09:54.179565 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:54.179519 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" podStartSLOduration=0.757483152 podStartE2EDuration="4.179506399s" podCreationTimestamp="2026-04-21 00:09:50 +0000 UTC" firstStartedPulling="2026-04-21 00:09:50.53330216 +0000 UTC m=+404.358281103" lastFinishedPulling="2026-04-21 00:09:53.955325389 +0000 UTC m=+407.780304350" observedRunningTime="2026-04-21 00:09:54.178054037 +0000 UTC m=+408.003033003" watchObservedRunningTime="2026-04-21 00:09:54.179506399 +0000 UTC m=+408.004485364" Apr 21 00:09:54.377737 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:54.377704 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:55.382214 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:55.382188 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:56.168053 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:56.168023 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:09:56.168994 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:09:56.168977 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb" Apr 21 00:10:06.537160 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.537125 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-9gvlk"] Apr 21 00:10:06.540286 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.540270 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" Apr 21 00:10:06.542751 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.542720 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"kube-root-ca.crt\"" Apr 21 00:10:06.542864 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.542757 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"kuadrant-operator-catalog-dockercfg-g5nkm\"" Apr 21 00:10:06.542864 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.542791 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"openshift-service-ca.crt\"" Apr 21 00:10:06.546246 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.546225 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-9gvlk"] Apr 21 00:10:06.622125 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.622098 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt6zp\" (UniqueName: \"kubernetes.io/projected/eac6eb27-009d-42d8-928b-36e1fae38174-kube-api-access-rt6zp\") pod \"kuadrant-operator-catalog-9gvlk\" (UID: \"eac6eb27-009d-42d8-928b-36e1fae38174\") " pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" Apr 21 00:10:06.722766 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.722743 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rt6zp\" (UniqueName: \"kubernetes.io/projected/eac6eb27-009d-42d8-928b-36e1fae38174-kube-api-access-rt6zp\") pod \"kuadrant-operator-catalog-9gvlk\" (UID: \"eac6eb27-009d-42d8-928b-36e1fae38174\") " pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" Apr 21 00:10:06.730079 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.730055 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"kube-root-ca.crt\"" Apr 21 00:10:06.740122 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.740097 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"openshift-service-ca.crt\"" Apr 21 00:10:06.750994 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.750965 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt6zp\" (UniqueName: \"kubernetes.io/projected/eac6eb27-009d-42d8-928b-36e1fae38174-kube-api-access-rt6zp\") pod \"kuadrant-operator-catalog-9gvlk\" (UID: \"eac6eb27-009d-42d8-928b-36e1fae38174\") " pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" Apr 21 00:10:06.852953 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.852901 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"kuadrant-operator-catalog-dockercfg-g5nkm\"" Apr 21 00:10:06.860810 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.860793 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" Apr 21 00:10:06.913073 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.913044 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-9gvlk"] Apr 21 00:10:06.979778 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:06.979745 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-9gvlk"] Apr 21 00:10:06.983331 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:10:06.983289 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeac6eb27_009d_42d8_928b_36e1fae38174.slice/crio-38a7327285a66e193b6c139a34ae4580e8ebe78bbbf4012f9d1660d12d0e8367 WatchSource:0}: Error finding container 38a7327285a66e193b6c139a34ae4580e8ebe78bbbf4012f9d1660d12d0e8367: Status 404 returned error can't find the container with id 38a7327285a66e193b6c139a34ae4580e8ebe78bbbf4012f9d1660d12d0e8367 Apr 21 00:10:07.122752 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.121841 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-6f6xn"] Apr 21 00:10:07.127905 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.127883 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:07.129038 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.129017 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-6f6xn"] Apr 21 00:10:07.204086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.204064 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" event={"ID":"eac6eb27-009d-42d8-928b-36e1fae38174","Type":"ContainerStarted","Data":"38a7327285a66e193b6c139a34ae4580e8ebe78bbbf4012f9d1660d12d0e8367"} Apr 21 00:10:07.226893 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.226876 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwzt2\" (UniqueName: \"kubernetes.io/projected/0ea60df0-bfd0-481d-ad91-9bc26c91bafd-kube-api-access-rwzt2\") pod \"kuadrant-operator-catalog-6f6xn\" (UID: \"0ea60df0-bfd0-481d-ad91-9bc26c91bafd\") " pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:07.327452 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.327429 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rwzt2\" (UniqueName: \"kubernetes.io/projected/0ea60df0-bfd0-481d-ad91-9bc26c91bafd-kube-api-access-rwzt2\") pod \"kuadrant-operator-catalog-6f6xn\" (UID: \"0ea60df0-bfd0-481d-ad91-9bc26c91bafd\") " pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:07.334448 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.334427 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwzt2\" (UniqueName: \"kubernetes.io/projected/0ea60df0-bfd0-481d-ad91-9bc26c91bafd-kube-api-access-rwzt2\") pod \"kuadrant-operator-catalog-6f6xn\" (UID: \"0ea60df0-bfd0-481d-ad91-9bc26c91bafd\") " pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:07.437242 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.437174 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:07.555179 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:07.555156 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-6f6xn"] Apr 21 00:10:07.557658 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:10:07.557630 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ea60df0_bfd0_481d_ad91_9bc26c91bafd.slice/crio-696733aaadd0ab143e2f3941309d0a41ffec3781b489b14703bc1b0b55ba349b WatchSource:0}: Error finding container 696733aaadd0ab143e2f3941309d0a41ffec3781b489b14703bc1b0b55ba349b: Status 404 returned error can't find the container with id 696733aaadd0ab143e2f3941309d0a41ffec3781b489b14703bc1b0b55ba349b Apr 21 00:10:08.210675 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:08.210637 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" event={"ID":"0ea60df0-bfd0-481d-ad91-9bc26c91bafd","Type":"ContainerStarted","Data":"696733aaadd0ab143e2f3941309d0a41ffec3781b489b14703bc1b0b55ba349b"} Apr 21 00:10:09.215666 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:09.215629 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" event={"ID":"eac6eb27-009d-42d8-928b-36e1fae38174","Type":"ContainerStarted","Data":"ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341"} Apr 21 00:10:09.215994 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:09.215678 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" podUID="eac6eb27-009d-42d8-928b-36e1fae38174" containerName="registry-server" containerID="cri-o://ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341" gracePeriod=2 Apr 21 00:10:09.232813 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:09.232771 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" podStartSLOduration=1.125612122 podStartE2EDuration="3.232757094s" podCreationTimestamp="2026-04-21 00:10:06 +0000 UTC" firstStartedPulling="2026-04-21 00:10:06.984818794 +0000 UTC m=+420.809797751" lastFinishedPulling="2026-04-21 00:10:09.091963779 +0000 UTC m=+422.916942723" observedRunningTime="2026-04-21 00:10:09.231720125 +0000 UTC m=+423.056699091" watchObservedRunningTime="2026-04-21 00:10:09.232757094 +0000 UTC m=+423.057736059" Apr 21 00:10:09.459135 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:09.459114 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" Apr 21 00:10:09.546546 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:09.546518 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt6zp\" (UniqueName: \"kubernetes.io/projected/eac6eb27-009d-42d8-928b-36e1fae38174-kube-api-access-rt6zp\") pod \"eac6eb27-009d-42d8-928b-36e1fae38174\" (UID: \"eac6eb27-009d-42d8-928b-36e1fae38174\") " Apr 21 00:10:09.548600 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:09.548577 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac6eb27-009d-42d8-928b-36e1fae38174-kube-api-access-rt6zp" (OuterVolumeSpecName: "kube-api-access-rt6zp") pod "eac6eb27-009d-42d8-928b-36e1fae38174" (UID: "eac6eb27-009d-42d8-928b-36e1fae38174"). InnerVolumeSpecName "kube-api-access-rt6zp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:10:09.647760 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:09.647708 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rt6zp\" (UniqueName: \"kubernetes.io/projected/eac6eb27-009d-42d8-928b-36e1fae38174-kube-api-access-rt6zp\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:10.219691 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.219658 2568 generic.go:358] "Generic (PLEG): container finished" podID="eac6eb27-009d-42d8-928b-36e1fae38174" containerID="ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341" exitCode=0 Apr 21 00:10:10.220047 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.219715 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" Apr 21 00:10:10.220047 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.219742 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" event={"ID":"eac6eb27-009d-42d8-928b-36e1fae38174","Type":"ContainerDied","Data":"ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341"} Apr 21 00:10:10.220047 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.219788 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-9gvlk" event={"ID":"eac6eb27-009d-42d8-928b-36e1fae38174","Type":"ContainerDied","Data":"38a7327285a66e193b6c139a34ae4580e8ebe78bbbf4012f9d1660d12d0e8367"} Apr 21 00:10:10.220047 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.219812 2568 scope.go:117] "RemoveContainer" containerID="ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341" Apr 21 00:10:10.221175 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.221156 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" event={"ID":"0ea60df0-bfd0-481d-ad91-9bc26c91bafd","Type":"ContainerStarted","Data":"9c8ce4f4da8d3ad163fe5bf36edeb7e3d93c17821695b1fbc9bab5a7c3d32a29"} Apr 21 00:10:10.228632 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.228600 2568 scope.go:117] "RemoveContainer" containerID="ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341" Apr 21 00:10:10.228885 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:10:10.228863 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341\": container with ID starting with ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341 not found: ID does not exist" containerID="ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341" Apr 21 00:10:10.228940 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.228898 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341"} err="failed to get container status \"ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341\": rpc error: code = NotFound desc = could not find container \"ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341\": container with ID starting with ad3eb88330834b3239067a23c339c20a3aedf1f64d4d54352057d34e28851341 not found: ID does not exist" Apr 21 00:10:10.238035 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.237998 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" podStartSLOduration=1.640512741 podStartE2EDuration="3.237988092s" podCreationTimestamp="2026-04-21 00:10:07 +0000 UTC" firstStartedPulling="2026-04-21 00:10:07.559220134 +0000 UTC m=+421.384199082" lastFinishedPulling="2026-04-21 00:10:09.156695491 +0000 UTC m=+422.981674433" observedRunningTime="2026-04-21 00:10:10.236223922 +0000 UTC m=+424.061202887" watchObservedRunningTime="2026-04-21 00:10:10.237988092 +0000 UTC m=+424.062967104" Apr 21 00:10:10.248260 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.248238 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-9gvlk"] Apr 21 00:10:10.253469 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.253450 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/kuadrant-operator-catalog-9gvlk"] Apr 21 00:10:10.719449 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:10.719420 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eac6eb27-009d-42d8-928b-36e1fae38174" path="/var/lib/kubelet/pods/eac6eb27-009d-42d8-928b-36e1fae38174/volumes" Apr 21 00:10:17.437547 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:17.437510 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:17.437547 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:17.437552 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:17.459064 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:17.459036 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:18.271551 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:18.271525 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/kuadrant-operator-catalog-6f6xn" Apr 21 00:10:21.749416 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.749385 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf"] Apr 21 00:10:21.749938 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.749755 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="eac6eb27-009d-42d8-928b-36e1fae38174" containerName="registry-server" Apr 21 00:10:21.749938 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.749769 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac6eb27-009d-42d8-928b-36e1fae38174" containerName="registry-server" Apr 21 00:10:21.749938 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.749820 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="eac6eb27-009d-42d8-928b-36e1fae38174" containerName="registry-server" Apr 21 00:10:21.754300 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.754285 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.756997 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.756978 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"default-dockercfg-pdvtf\"" Apr 21 00:10:21.760425 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.760403 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf"] Apr 21 00:10:21.837455 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.837430 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg5bj\" (UniqueName: \"kubernetes.io/projected/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-kube-api-access-kg5bj\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.837559 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.837487 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-bundle\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.837600 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.837560 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-util\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.938771 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.938753 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-util\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.938865 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.938794 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kg5bj\" (UniqueName: \"kubernetes.io/projected/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-kube-api-access-kg5bj\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.938865 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.938834 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-bundle\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.939098 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.939082 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-util\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.939174 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.939122 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-bundle\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:21.946403 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:21.946375 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg5bj\" (UniqueName: \"kubernetes.io/projected/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-kube-api-access-kg5bj\") pod \"0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:22.064995 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.064928 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:22.182515 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.182487 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf"] Apr 21 00:10:22.183733 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:10:22.183704 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddac0921f_b18e_4ed9_b1fb_5af4cf784a1e.slice/crio-b70fbdb6fe7f9807b8a299dd729734bc6e0640e1254c6c22fbab1a8c2ce24b5c WatchSource:0}: Error finding container b70fbdb6fe7f9807b8a299dd729734bc6e0640e1254c6c22fbab1a8c2ce24b5c: Status 404 returned error can't find the container with id b70fbdb6fe7f9807b8a299dd729734bc6e0640e1254c6c22fbab1a8c2ce24b5c Apr 21 00:10:22.266333 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.266305 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" event={"ID":"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e","Type":"ContainerStarted","Data":"c1815c02c86d084b3809726f3eb2a8d5f1d6e297e27dae5c1012c955875b8e01"} Apr 21 00:10:22.266462 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.266341 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" event={"ID":"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e","Type":"ContainerStarted","Data":"b70fbdb6fe7f9807b8a299dd729734bc6e0640e1254c6c22fbab1a8c2ce24b5c"} Apr 21 00:10:22.549546 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.549517 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2"] Apr 21 00:10:22.552892 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.552874 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.559112 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.559090 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2"] Apr 21 00:10:22.643506 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.643483 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-bundle\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.643626 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.643538 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-util\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.643681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.643630 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnj6s\" (UniqueName: \"kubernetes.io/projected/2c66175c-c20f-4e2d-a744-485c41e6a54e-kube-api-access-fnj6s\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.744858 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.744822 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-bundle\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.744988 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.744933 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-util\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.745046 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.744988 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fnj6s\" (UniqueName: \"kubernetes.io/projected/2c66175c-c20f-4e2d-a744-485c41e6a54e-kube-api-access-fnj6s\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.745180 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.745160 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-bundle\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.745234 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.745219 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-util\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.752349 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.752329 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnj6s\" (UniqueName: \"kubernetes.io/projected/2c66175c-c20f-4e2d-a744-485c41e6a54e-kube-api-access-fnj6s\") pod \"9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.863109 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.863052 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:22.978585 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:22.978562 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2"] Apr 21 00:10:22.981177 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:10:22.981141 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c66175c_c20f_4e2d_a744_485c41e6a54e.slice/crio-6d74c94dbe41005f4fe93a8a5bff0f34f2765eb52389800e63c5900b82c9aa40 WatchSource:0}: Error finding container 6d74c94dbe41005f4fe93a8a5bff0f34f2765eb52389800e63c5900b82c9aa40: Status 404 returned error can't find the container with id 6d74c94dbe41005f4fe93a8a5bff0f34f2765eb52389800e63c5900b82c9aa40 Apr 21 00:10:23.149292 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.149233 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx"] Apr 21 00:10:23.152717 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.152701 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.161174 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.161148 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx"] Apr 21 00:10:23.250515 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.250494 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mw8t\" (UniqueName: \"kubernetes.io/projected/da67825e-4611-46e0-b960-889d7cd00120-kube-api-access-2mw8t\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.250645 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.250533 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-bundle\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.250645 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.250550 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-util\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.271494 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.271461 2568 generic.go:358] "Generic (PLEG): container finished" podID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerID="44f997f519e8ad227f7ca6fefc872d430ff4a14816a62e7143df6814c30aebd2" exitCode=0 Apr 21 00:10:23.271593 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.271542 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" event={"ID":"2c66175c-c20f-4e2d-a744-485c41e6a54e","Type":"ContainerDied","Data":"44f997f519e8ad227f7ca6fefc872d430ff4a14816a62e7143df6814c30aebd2"} Apr 21 00:10:23.271593 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.271573 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" event={"ID":"2c66175c-c20f-4e2d-a744-485c41e6a54e","Type":"ContainerStarted","Data":"6d74c94dbe41005f4fe93a8a5bff0f34f2765eb52389800e63c5900b82c9aa40"} Apr 21 00:10:23.272992 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.272970 2568 generic.go:358] "Generic (PLEG): container finished" podID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerID="c1815c02c86d084b3809726f3eb2a8d5f1d6e297e27dae5c1012c955875b8e01" exitCode=0 Apr 21 00:10:23.273079 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.272996 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" event={"ID":"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e","Type":"ContainerDied","Data":"c1815c02c86d084b3809726f3eb2a8d5f1d6e297e27dae5c1012c955875b8e01"} Apr 21 00:10:23.351197 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.351168 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2mw8t\" (UniqueName: \"kubernetes.io/projected/da67825e-4611-46e0-b960-889d7cd00120-kube-api-access-2mw8t\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.351342 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.351220 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-bundle\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.351550 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.351531 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-util\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.351790 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.351770 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-bundle\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.351906 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.351846 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-util\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.358918 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.358900 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mw8t\" (UniqueName: \"kubernetes.io/projected/da67825e-4611-46e0-b960-889d7cd00120-kube-api-access-2mw8t\") pod \"19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.467960 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.467939 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:23.552740 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.552712 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9"] Apr 21 00:10:23.557557 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.557534 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.563575 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.563553 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9"] Apr 21 00:10:23.585875 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.585832 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx"] Apr 21 00:10:23.587911 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:10:23.587888 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda67825e_4611_46e0_b960_889d7cd00120.slice/crio-15ea84b9f8dd8590b0467447e35529453054d5c59fc0e84d43e9c9dddc0fbe98 WatchSource:0}: Error finding container 15ea84b9f8dd8590b0467447e35529453054d5c59fc0e84d43e9c9dddc0fbe98: Status 404 returned error can't find the container with id 15ea84b9f8dd8590b0467447e35529453054d5c59fc0e84d43e9c9dddc0fbe98 Apr 21 00:10:23.653303 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.653277 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-util\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.653406 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.653322 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-bundle\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.653406 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.653349 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k89l\" (UniqueName: \"kubernetes.io/projected/b578effd-613c-44c3-9d8d-579f79fa6e6f-kube-api-access-9k89l\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.754726 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.754678 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-util\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.754726 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.754723 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-bundle\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.755020 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.754753 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9k89l\" (UniqueName: \"kubernetes.io/projected/b578effd-613c-44c3-9d8d-579f79fa6e6f-kube-api-access-9k89l\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.755020 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.755008 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-util\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.755096 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.755077 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-bundle\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.763141 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.763124 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k89l\" (UniqueName: \"kubernetes.io/projected/b578effd-613c-44c3-9d8d-579f79fa6e6f-kube-api-access-9k89l\") pod \"5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:23.869530 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:23.869497 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:24.207656 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.207631 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9"] Apr 21 00:10:24.208731 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:10:24.208697 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb578effd_613c_44c3_9d8d_579f79fa6e6f.slice/crio-d1e15bcc76904f390a93e86b9c5364a860b461009eb38f9b9dbb7e35ea708a10 WatchSource:0}: Error finding container d1e15bcc76904f390a93e86b9c5364a860b461009eb38f9b9dbb7e35ea708a10: Status 404 returned error can't find the container with id d1e15bcc76904f390a93e86b9c5364a860b461009eb38f9b9dbb7e35ea708a10 Apr 21 00:10:24.277639 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.277603 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" event={"ID":"b578effd-613c-44c3-9d8d-579f79fa6e6f","Type":"ContainerStarted","Data":"d1e15bcc76904f390a93e86b9c5364a860b461009eb38f9b9dbb7e35ea708a10"} Apr 21 00:10:24.278853 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.278830 2568 generic.go:358] "Generic (PLEG): container finished" podID="da67825e-4611-46e0-b960-889d7cd00120" containerID="8727d17d719680682f65c1eee1356cc477247df44cec035c4b187a0f7e893614" exitCode=0 Apr 21 00:10:24.278941 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.278915 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" event={"ID":"da67825e-4611-46e0-b960-889d7cd00120","Type":"ContainerDied","Data":"8727d17d719680682f65c1eee1356cc477247df44cec035c4b187a0f7e893614"} Apr 21 00:10:24.278997 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.278948 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" event={"ID":"da67825e-4611-46e0-b960-889d7cd00120","Type":"ContainerStarted","Data":"15ea84b9f8dd8590b0467447e35529453054d5c59fc0e84d43e9c9dddc0fbe98"} Apr 21 00:10:24.280500 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.280481 2568 generic.go:358] "Generic (PLEG): container finished" podID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerID="da5b52a55fefa5f26c157f9b45652ebd4041d0b05c89517caea9bceddd408eef" exitCode=0 Apr 21 00:10:24.280583 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.280559 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" event={"ID":"2c66175c-c20f-4e2d-a744-485c41e6a54e","Type":"ContainerDied","Data":"da5b52a55fefa5f26c157f9b45652ebd4041d0b05c89517caea9bceddd408eef"} Apr 21 00:10:24.282301 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.282278 2568 generic.go:358] "Generic (PLEG): container finished" podID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerID="cb3733e10beef1a4f1e630cda2b1c32c715c3dffe332f66fe73522d085160f99" exitCode=0 Apr 21 00:10:24.282405 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:24.282318 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" event={"ID":"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e","Type":"ContainerDied","Data":"cb3733e10beef1a4f1e630cda2b1c32c715c3dffe332f66fe73522d085160f99"} Apr 21 00:10:25.287577 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:25.287545 2568 generic.go:358] "Generic (PLEG): container finished" podID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerID="3b1766e33d8d459de0876be1b2934bd2da8094110cc4b55dc709f9b017f87f65" exitCode=0 Apr 21 00:10:25.288002 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:25.287602 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" event={"ID":"2c66175c-c20f-4e2d-a744-485c41e6a54e","Type":"ContainerDied","Data":"3b1766e33d8d459de0876be1b2934bd2da8094110cc4b55dc709f9b017f87f65"} Apr 21 00:10:25.289483 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:25.289456 2568 generic.go:358] "Generic (PLEG): container finished" podID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerID="d287a597b8c600b4638349dd04858c8f56371e2bcce4616e6f9297da9cc4129d" exitCode=0 Apr 21 00:10:25.289599 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:25.289535 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" event={"ID":"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e","Type":"ContainerDied","Data":"d287a597b8c600b4638349dd04858c8f56371e2bcce4616e6f9297da9cc4129d"} Apr 21 00:10:25.290697 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:25.290671 2568 generic.go:358] "Generic (PLEG): container finished" podID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerID="81119522f870c73cd9c88ca8faac1f6944fa2a21e55ee3355798ef4643457e61" exitCode=0 Apr 21 00:10:25.290774 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:25.290758 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" event={"ID":"b578effd-613c-44c3-9d8d-579f79fa6e6f","Type":"ContainerDied","Data":"81119522f870c73cd9c88ca8faac1f6944fa2a21e55ee3355798ef4643457e61"} Apr 21 00:10:25.292531 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:25.292503 2568 generic.go:358] "Generic (PLEG): container finished" podID="da67825e-4611-46e0-b960-889d7cd00120" containerID="d4424e58d31131fa88d628bae56a90f4b9348c16216e4294badf1b0fca7b04d1" exitCode=0 Apr 21 00:10:25.292636 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:25.292560 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" event={"ID":"da67825e-4611-46e0-b960-889d7cd00120","Type":"ContainerDied","Data":"d4424e58d31131fa88d628bae56a90f4b9348c16216e4294badf1b0fca7b04d1"} Apr 21 00:10:26.298748 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.298716 2568 generic.go:358] "Generic (PLEG): container finished" podID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerID="494f0841d6481ffb75e2348959633029ad3545ae875652ecb5c7c6f481a38d44" exitCode=0 Apr 21 00:10:26.299152 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.298802 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" event={"ID":"b578effd-613c-44c3-9d8d-579f79fa6e6f","Type":"ContainerDied","Data":"494f0841d6481ffb75e2348959633029ad3545ae875652ecb5c7c6f481a38d44"} Apr 21 00:10:26.300784 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.300757 2568 generic.go:358] "Generic (PLEG): container finished" podID="da67825e-4611-46e0-b960-889d7cd00120" containerID="93f27b57f0a38b20e9656d47b9de183bf9ff1955851a710cb58271a431cf86d8" exitCode=0 Apr 21 00:10:26.300872 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.300836 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" event={"ID":"da67825e-4611-46e0-b960-889d7cd00120","Type":"ContainerDied","Data":"93f27b57f0a38b20e9656d47b9de183bf9ff1955851a710cb58271a431cf86d8"} Apr 21 00:10:26.425534 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.425510 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:26.445891 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.445869 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:26.577059 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.577030 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-bundle\") pod \"2c66175c-c20f-4e2d-a744-485c41e6a54e\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " Apr 21 00:10:26.577224 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.577068 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnj6s\" (UniqueName: \"kubernetes.io/projected/2c66175c-c20f-4e2d-a744-485c41e6a54e-kube-api-access-fnj6s\") pod \"2c66175c-c20f-4e2d-a744-485c41e6a54e\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " Apr 21 00:10:26.577224 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.577097 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg5bj\" (UniqueName: \"kubernetes.io/projected/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-kube-api-access-kg5bj\") pod \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " Apr 21 00:10:26.577224 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.577129 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-util\") pod \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " Apr 21 00:10:26.577224 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.577149 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-util\") pod \"2c66175c-c20f-4e2d-a744-485c41e6a54e\" (UID: \"2c66175c-c20f-4e2d-a744-485c41e6a54e\") " Apr 21 00:10:26.577224 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.577195 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-bundle\") pod \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\" (UID: \"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e\") " Apr 21 00:10:26.577801 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.577732 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-bundle" (OuterVolumeSpecName: "bundle") pod "dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" (UID: "dac0921f-b18e-4ed9-b1fb-5af4cf784a1e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:10:26.577801 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.577743 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-bundle" (OuterVolumeSpecName: "bundle") pod "2c66175c-c20f-4e2d-a744-485c41e6a54e" (UID: "2c66175c-c20f-4e2d-a744-485c41e6a54e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:10:26.579218 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.579197 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-kube-api-access-kg5bj" (OuterVolumeSpecName: "kube-api-access-kg5bj") pod "dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" (UID: "dac0921f-b18e-4ed9-b1fb-5af4cf784a1e"). InnerVolumeSpecName "kube-api-access-kg5bj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:10:26.579531 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.579517 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c66175c-c20f-4e2d-a744-485c41e6a54e-kube-api-access-fnj6s" (OuterVolumeSpecName: "kube-api-access-fnj6s") pod "2c66175c-c20f-4e2d-a744-485c41e6a54e" (UID: "2c66175c-c20f-4e2d-a744-485c41e6a54e"). InnerVolumeSpecName "kube-api-access-fnj6s". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:10:26.582783 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.582749 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-util" (OuterVolumeSpecName: "util") pod "2c66175c-c20f-4e2d-a744-485c41e6a54e" (UID: "2c66175c-c20f-4e2d-a744-485c41e6a54e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:10:26.583095 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.583076 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-util" (OuterVolumeSpecName: "util") pod "dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" (UID: "dac0921f-b18e-4ed9-b1fb-5af4cf784a1e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:10:26.678389 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.678295 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:26.678389 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.678339 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:26.678389 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.678349 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-fnj6s\" (UniqueName: \"kubernetes.io/projected/2c66175c-c20f-4e2d-a744-485c41e6a54e-kube-api-access-fnj6s\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:26.678389 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.678361 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kg5bj\" (UniqueName: \"kubernetes.io/projected/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-kube-api-access-kg5bj\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:26.678389 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.678370 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dac0921f-b18e-4ed9-b1fb-5af4cf784a1e-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:26.678389 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:26.678379 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c66175c-c20f-4e2d-a744-485c41e6a54e-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:27.306097 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.306074 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" Apr 21 00:10:27.306097 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.306064 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2" event={"ID":"2c66175c-c20f-4e2d-a744-485c41e6a54e","Type":"ContainerDied","Data":"6d74c94dbe41005f4fe93a8a5bff0f34f2765eb52389800e63c5900b82c9aa40"} Apr 21 00:10:27.306485 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.306105 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d74c94dbe41005f4fe93a8a5bff0f34f2765eb52389800e63c5900b82c9aa40" Apr 21 00:10:27.307778 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.307758 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" Apr 21 00:10:27.307895 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.307757 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf" event={"ID":"dac0921f-b18e-4ed9-b1fb-5af4cf784a1e","Type":"ContainerDied","Data":"b70fbdb6fe7f9807b8a299dd729734bc6e0640e1254c6c22fbab1a8c2ce24b5c"} Apr 21 00:10:27.307963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.307893 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b70fbdb6fe7f9807b8a299dd729734bc6e0640e1254c6c22fbab1a8c2ce24b5c" Apr 21 00:10:27.313297 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.313276 2568 generic.go:358] "Generic (PLEG): container finished" podID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerID="0cfa126100b664a8574cdf734ff425b73995f91aab29556e13ff6207733dd407" exitCode=0 Apr 21 00:10:27.313377 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.313351 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" event={"ID":"b578effd-613c-44c3-9d8d-579f79fa6e6f","Type":"ContainerDied","Data":"0cfa126100b664a8574cdf734ff425b73995f91aab29556e13ff6207733dd407"} Apr 21 00:10:27.428907 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.428886 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:27.584241 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.584218 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-bundle\") pod \"da67825e-4611-46e0-b960-889d7cd00120\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " Apr 21 00:10:27.584395 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.584277 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-util\") pod \"da67825e-4611-46e0-b960-889d7cd00120\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " Apr 21 00:10:27.584395 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.584340 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mw8t\" (UniqueName: \"kubernetes.io/projected/da67825e-4611-46e0-b960-889d7cd00120-kube-api-access-2mw8t\") pod \"da67825e-4611-46e0-b960-889d7cd00120\" (UID: \"da67825e-4611-46e0-b960-889d7cd00120\") " Apr 21 00:10:27.584750 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.584725 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-bundle" (OuterVolumeSpecName: "bundle") pod "da67825e-4611-46e0-b960-889d7cd00120" (UID: "da67825e-4611-46e0-b960-889d7cd00120"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:10:27.586779 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.586754 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da67825e-4611-46e0-b960-889d7cd00120-kube-api-access-2mw8t" (OuterVolumeSpecName: "kube-api-access-2mw8t") pod "da67825e-4611-46e0-b960-889d7cd00120" (UID: "da67825e-4611-46e0-b960-889d7cd00120"). InnerVolumeSpecName "kube-api-access-2mw8t". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:10:27.593135 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.593093 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-util" (OuterVolumeSpecName: "util") pod "da67825e-4611-46e0-b960-889d7cd00120" (UID: "da67825e-4611-46e0-b960-889d7cd00120"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:10:27.685837 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.685773 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:27.685837 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.685803 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2mw8t\" (UniqueName: \"kubernetes.io/projected/da67825e-4611-46e0-b960-889d7cd00120-kube-api-access-2mw8t\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:27.685837 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:27.685813 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da67825e-4611-46e0-b960-889d7cd00120-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:28.318178 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.318152 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" Apr 21 00:10:28.318519 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.318147 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx" event={"ID":"da67825e-4611-46e0-b960-889d7cd00120","Type":"ContainerDied","Data":"15ea84b9f8dd8590b0467447e35529453054d5c59fc0e84d43e9c9dddc0fbe98"} Apr 21 00:10:28.318519 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.318259 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15ea84b9f8dd8590b0467447e35529453054d5c59fc0e84d43e9c9dddc0fbe98" Apr 21 00:10:28.437450 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.437430 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:28.591597 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.591531 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-bundle\") pod \"b578effd-613c-44c3-9d8d-579f79fa6e6f\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " Apr 21 00:10:28.591597 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.591588 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k89l\" (UniqueName: \"kubernetes.io/projected/b578effd-613c-44c3-9d8d-579f79fa6e6f-kube-api-access-9k89l\") pod \"b578effd-613c-44c3-9d8d-579f79fa6e6f\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " Apr 21 00:10:28.591828 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.591657 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-util\") pod \"b578effd-613c-44c3-9d8d-579f79fa6e6f\" (UID: \"b578effd-613c-44c3-9d8d-579f79fa6e6f\") " Apr 21 00:10:28.592179 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.592155 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-bundle" (OuterVolumeSpecName: "bundle") pod "b578effd-613c-44c3-9d8d-579f79fa6e6f" (UID: "b578effd-613c-44c3-9d8d-579f79fa6e6f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:10:28.593723 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.593701 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b578effd-613c-44c3-9d8d-579f79fa6e6f-kube-api-access-9k89l" (OuterVolumeSpecName: "kube-api-access-9k89l") pod "b578effd-613c-44c3-9d8d-579f79fa6e6f" (UID: "b578effd-613c-44c3-9d8d-579f79fa6e6f"). InnerVolumeSpecName "kube-api-access-9k89l". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:10:28.596675 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.596643 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-util" (OuterVolumeSpecName: "util") pod "b578effd-613c-44c3-9d8d-579f79fa6e6f" (UID: "b578effd-613c-44c3-9d8d-579f79fa6e6f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:10:28.692357 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.692324 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:28.692357 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.692351 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b578effd-613c-44c3-9d8d-579f79fa6e6f-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:28.692357 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:28.692360 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9k89l\" (UniqueName: \"kubernetes.io/projected/b578effd-613c-44c3-9d8d-579f79fa6e6f-kube-api-access-9k89l\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:10:29.323451 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:29.323425 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" Apr 21 00:10:29.323831 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:29.323417 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9" event={"ID":"b578effd-613c-44c3-9d8d-579f79fa6e6f","Type":"ContainerDied","Data":"d1e15bcc76904f390a93e86b9c5364a860b461009eb38f9b9dbb7e35ea708a10"} Apr 21 00:10:29.323831 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:29.323525 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1e15bcc76904f390a93e86b9c5364a860b461009eb38f9b9dbb7e35ea708a10" Apr 21 00:10:38.499328 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499294 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-operator-657f44b778-7qwxj"] Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499601 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerName="util" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499625 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerName="util" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499636 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerName="pull" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499641 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerName="pull" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499647 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerName="extract" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499653 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerName="extract" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499659 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerName="extract" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499664 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerName="extract" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499672 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerName="util" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499678 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerName="util" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499683 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="da67825e-4611-46e0-b960-889d7cd00120" containerName="util" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499689 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="da67825e-4611-46e0-b960-889d7cd00120" containerName="util" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499696 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerName="pull" Apr 21 00:10:38.499693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499701 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerName="pull" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499709 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerName="pull" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499714 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerName="pull" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499719 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerName="extract" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499724 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerName="extract" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499735 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="da67825e-4611-46e0-b960-889d7cd00120" containerName="extract" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499740 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="da67825e-4611-46e0-b960-889d7cd00120" containerName="extract" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499748 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerName="util" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499752 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerName="util" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499760 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="da67825e-4611-46e0-b960-889d7cd00120" containerName="pull" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499764 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="da67825e-4611-46e0-b960-889d7cd00120" containerName="pull" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499811 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="dac0921f-b18e-4ed9-b1fb-5af4cf784a1e" containerName="extract" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499819 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="da67825e-4611-46e0-b960-889d7cd00120" containerName="extract" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499826 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="b578effd-613c-44c3-9d8d-579f79fa6e6f" containerName="extract" Apr 21 00:10:38.500086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.499834 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="2c66175c-c20f-4e2d-a744-485c41e6a54e" containerName="extract" Apr 21 00:10:38.505558 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.505537 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" Apr 21 00:10:38.508088 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.508072 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-operator-dockercfg-7bzkp\"" Apr 21 00:10:38.516833 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.516812 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-operator-657f44b778-7qwxj"] Apr 21 00:10:38.671171 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.671137 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rckg\" (UniqueName: \"kubernetes.io/projected/8eac1894-79ad-43d1-aeb1-65ff1d7d538b-kube-api-access-2rckg\") pod \"authorino-operator-657f44b778-7qwxj\" (UID: \"8eac1894-79ad-43d1-aeb1-65ff1d7d538b\") " pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" Apr 21 00:10:38.771699 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.771620 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2rckg\" (UniqueName: \"kubernetes.io/projected/8eac1894-79ad-43d1-aeb1-65ff1d7d538b-kube-api-access-2rckg\") pod \"authorino-operator-657f44b778-7qwxj\" (UID: \"8eac1894-79ad-43d1-aeb1-65ff1d7d538b\") " pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" Apr 21 00:10:38.782210 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.782176 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rckg\" (UniqueName: \"kubernetes.io/projected/8eac1894-79ad-43d1-aeb1-65ff1d7d538b-kube-api-access-2rckg\") pod \"authorino-operator-657f44b778-7qwxj\" (UID: \"8eac1894-79ad-43d1-aeb1-65ff1d7d538b\") " pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" Apr 21 00:10:38.815990 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.815962 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" Apr 21 00:10:38.941858 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:38.941833 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-operator-657f44b778-7qwxj"] Apr 21 00:10:38.943297 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:10:38.943274 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8eac1894_79ad_43d1_aeb1_65ff1d7d538b.slice/crio-2e42e4ae8456373a4749c642bb37688b33575d8a66dcefe752b642b06abdde5b WatchSource:0}: Error finding container 2e42e4ae8456373a4749c642bb37688b33575d8a66dcefe752b642b06abdde5b: Status 404 returned error can't find the container with id 2e42e4ae8456373a4749c642bb37688b33575d8a66dcefe752b642b06abdde5b Apr 21 00:10:39.357292 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:39.357253 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" event={"ID":"8eac1894-79ad-43d1-aeb1-65ff1d7d538b","Type":"ContainerStarted","Data":"2e42e4ae8456373a4749c642bb37688b33575d8a66dcefe752b642b06abdde5b"} Apr 21 00:10:41.367442 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:41.367407 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" event={"ID":"8eac1894-79ad-43d1-aeb1-65ff1d7d538b","Type":"ContainerStarted","Data":"5d52bd94fad6e967d706cb59e076406f8ae41b24a834a827659fd50e65abc887"} Apr 21 00:10:41.367767 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:41.367510 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" Apr 21 00:10:41.386051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:41.386009 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" podStartSLOduration=1.175233821 podStartE2EDuration="3.385997611s" podCreationTimestamp="2026-04-21 00:10:38 +0000 UTC" firstStartedPulling="2026-04-21 00:10:38.945112198 +0000 UTC m=+452.770091144" lastFinishedPulling="2026-04-21 00:10:41.155875979 +0000 UTC m=+454.980854934" observedRunningTime="2026-04-21 00:10:41.383567419 +0000 UTC m=+455.208546385" watchObservedRunningTime="2026-04-21 00:10:41.385997611 +0000 UTC m=+455.210976576" Apr 21 00:10:41.914210 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:41.914175 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g"] Apr 21 00:10:41.917495 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:41.917472 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:10:41.920125 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:41.920107 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"limitador-operator-controller-manager-dockercfg-vqsrk\"" Apr 21 00:10:41.924739 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:41.924683 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g"] Apr 21 00:10:41.993574 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:41.993547 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r472\" (UniqueName: \"kubernetes.io/projected/12396b0e-ca72-4415-9b45-fc33d5345456-kube-api-access-4r472\") pod \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" (UID: \"12396b0e-ca72-4415-9b45-fc33d5345456\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:10:42.094085 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:42.094053 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4r472\" (UniqueName: \"kubernetes.io/projected/12396b0e-ca72-4415-9b45-fc33d5345456-kube-api-access-4r472\") pod \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" (UID: \"12396b0e-ca72-4415-9b45-fc33d5345456\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:10:42.101838 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:42.101818 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r472\" (UniqueName: \"kubernetes.io/projected/12396b0e-ca72-4415-9b45-fc33d5345456-kube-api-access-4r472\") pod \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" (UID: \"12396b0e-ca72-4415-9b45-fc33d5345456\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:10:42.229399 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:42.229373 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:10:42.353082 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:42.353057 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g"] Apr 21 00:10:42.356395 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:10:42.356364 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12396b0e_ca72_4415_9b45_fc33d5345456.slice/crio-edf7697cc7b5a5448d5963caef3f8a807a0d6634658c5605dd26120bcb626bcf WatchSource:0}: Error finding container edf7697cc7b5a5448d5963caef3f8a807a0d6634658c5605dd26120bcb626bcf: Status 404 returned error can't find the container with id edf7697cc7b5a5448d5963caef3f8a807a0d6634658c5605dd26120bcb626bcf Apr 21 00:10:42.372250 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:42.372219 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" event={"ID":"12396b0e-ca72-4415-9b45-fc33d5345456","Type":"ContainerStarted","Data":"edf7697cc7b5a5448d5963caef3f8a807a0d6634658c5605dd26120bcb626bcf"} Apr 21 00:10:44.383256 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:44.383182 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" event={"ID":"12396b0e-ca72-4415-9b45-fc33d5345456","Type":"ContainerStarted","Data":"4dc588e238ae614cd6da0e5f11b9c914c0e7862d6cb3554f74d4f0b3cd1d6aac"} Apr 21 00:10:44.383595 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:44.383296 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:10:44.401369 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:44.401315 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" podStartSLOduration=1.653818681 podStartE2EDuration="3.401299096s" podCreationTimestamp="2026-04-21 00:10:41 +0000 UTC" firstStartedPulling="2026-04-21 00:10:42.358660995 +0000 UTC m=+456.183639938" lastFinishedPulling="2026-04-21 00:10:44.106141396 +0000 UTC m=+457.931120353" observedRunningTime="2026-04-21 00:10:44.398971732 +0000 UTC m=+458.223950737" watchObservedRunningTime="2026-04-21 00:10:44.401299096 +0000 UTC m=+458.226278062" Apr 21 00:10:52.374502 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:52.374470 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/authorino-operator-657f44b778-7qwxj" Apr 21 00:10:55.390323 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:10:55.390296 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:11:00.038440 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.038406 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-5fb9cf77c4-lwm7z"] Apr 21 00:11:00.043291 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.043261 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.054658 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.054637 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5fb9cf77c4-lwm7z"] Apr 21 00:11:00.127727 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.127692 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-oauth-config\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.127727 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.127730 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-service-ca\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.127925 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.127765 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-config\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.127925 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.127814 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-oauth-serving-cert\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.127925 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.127850 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjrq7\" (UniqueName: \"kubernetes.io/projected/90ba90b0-82f6-41e0-9b35-22186054fe5f-kube-api-access-zjrq7\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.127925 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.127875 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-serving-cert\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.127925 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.127901 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-trusted-ca-bundle\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.228224 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.228191 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zjrq7\" (UniqueName: \"kubernetes.io/projected/90ba90b0-82f6-41e0-9b35-22186054fe5f-kube-api-access-zjrq7\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.228368 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.228231 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-serving-cert\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.228368 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.228255 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-trusted-ca-bundle\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.228368 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.228284 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-oauth-config\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.228368 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.228299 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-service-ca\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.228368 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.228322 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-config\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.228368 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.228359 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-oauth-serving-cert\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.229076 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.229046 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-config\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.229209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.229192 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-oauth-serving-cert\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.229279 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.229217 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-service-ca\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.229279 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.229237 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90ba90b0-82f6-41e0-9b35-22186054fe5f-trusted-ca-bundle\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.230692 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.230663 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-oauth-config\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.230800 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.230783 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/90ba90b0-82f6-41e0-9b35-22186054fe5f-console-serving-cert\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.245075 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.245052 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjrq7\" (UniqueName: \"kubernetes.io/projected/90ba90b0-82f6-41e0-9b35-22186054fe5f-kube-api-access-zjrq7\") pod \"console-5fb9cf77c4-lwm7z\" (UID: \"90ba90b0-82f6-41e0-9b35-22186054fe5f\") " pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.352352 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.352296 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:00.471851 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:00.471826 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5fb9cf77c4-lwm7z"] Apr 21 00:11:00.473658 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:11:00.473632 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90ba90b0_82f6_41e0_9b35_22186054fe5f.slice/crio-355f82e2af37903c017a79ecd2c92f3ef4c8536c289c311fc86515d758cdec76 WatchSource:0}: Error finding container 355f82e2af37903c017a79ecd2c92f3ef4c8536c289c311fc86515d758cdec76: Status 404 returned error can't find the container with id 355f82e2af37903c017a79ecd2c92f3ef4c8536c289c311fc86515d758cdec76 Apr 21 00:11:01.452824 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:01.452785 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5fb9cf77c4-lwm7z" event={"ID":"90ba90b0-82f6-41e0-9b35-22186054fe5f","Type":"ContainerStarted","Data":"bf4cced1bde43249ca483a5b144644de33331041113c9235051de18f0260eba0"} Apr 21 00:11:01.452824 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:01.452821 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5fb9cf77c4-lwm7z" event={"ID":"90ba90b0-82f6-41e0-9b35-22186054fe5f","Type":"ContainerStarted","Data":"355f82e2af37903c017a79ecd2c92f3ef4c8536c289c311fc86515d758cdec76"} Apr 21 00:11:01.470029 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:01.469984 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5fb9cf77c4-lwm7z" podStartSLOduration=1.4699706350000001 podStartE2EDuration="1.469970635s" podCreationTimestamp="2026-04-21 00:11:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:11:01.468875182 +0000 UTC m=+475.293854147" watchObservedRunningTime="2026-04-21 00:11:01.469970635 +0000 UTC m=+475.294949599" Apr 21 00:11:09.353410 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.353371 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g"] Apr 21 00:11:09.353866 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.353690 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" containerName="manager" containerID="cri-o://4dc588e238ae614cd6da0e5f11b9c914c0e7862d6cb3554f74d4f0b3cd1d6aac" gracePeriod=2 Apr 21 00:11:09.361597 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.361569 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g"] Apr 21 00:11:09.385830 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.385802 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf"] Apr 21 00:11:09.386124 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.386111 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" containerName="manager" Apr 21 00:11:09.386173 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.386126 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" containerName="manager" Apr 21 00:11:09.386208 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.386187 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" containerName="manager" Apr 21 00:11:09.389085 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.389069 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" Apr 21 00:11:09.391521 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.391493 2568 status_manager.go:895] "Failed to get status for pod" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" err="pods \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" is forbidden: User \"system:node:ip-10-0-132-206.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-132-206.ec2.internal' and this object" Apr 21 00:11:09.401175 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.401149 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf"] Apr 21 00:11:09.487005 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.486980 2568 generic.go:358] "Generic (PLEG): container finished" podID="12396b0e-ca72-4415-9b45-fc33d5345456" containerID="4dc588e238ae614cd6da0e5f11b9c914c0e7862d6cb3554f74d4f0b3cd1d6aac" exitCode=0 Apr 21 00:11:09.504393 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.504369 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77tnx\" (UniqueName: \"kubernetes.io/projected/64f29997-9e48-4455-85ed-1ca939a19be9-kube-api-access-77tnx\") pod \"limitador-operator-controller-manager-85c4996f8c-4vgkf\" (UID: \"64f29997-9e48-4455-85ed-1ca939a19be9\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" Apr 21 00:11:09.604229 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.604164 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:11:09.605407 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.605387 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-77tnx\" (UniqueName: \"kubernetes.io/projected/64f29997-9e48-4455-85ed-1ca939a19be9-kube-api-access-77tnx\") pod \"limitador-operator-controller-manager-85c4996f8c-4vgkf\" (UID: \"64f29997-9e48-4455-85ed-1ca939a19be9\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" Apr 21 00:11:09.606539 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.606513 2568 status_manager.go:895] "Failed to get status for pod" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" err="pods \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" is forbidden: User \"system:node:ip-10-0-132-206.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-132-206.ec2.internal' and this object" Apr 21 00:11:09.613149 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.613119 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-77tnx\" (UniqueName: \"kubernetes.io/projected/64f29997-9e48-4455-85ed-1ca939a19be9-kube-api-access-77tnx\") pod \"limitador-operator-controller-manager-85c4996f8c-4vgkf\" (UID: \"64f29997-9e48-4455-85ed-1ca939a19be9\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" Apr 21 00:11:09.706482 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.706391 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r472\" (UniqueName: \"kubernetes.io/projected/12396b0e-ca72-4415-9b45-fc33d5345456-kube-api-access-4r472\") pod \"12396b0e-ca72-4415-9b45-fc33d5345456\" (UID: \"12396b0e-ca72-4415-9b45-fc33d5345456\") " Apr 21 00:11:09.708317 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.708296 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12396b0e-ca72-4415-9b45-fc33d5345456-kube-api-access-4r472" (OuterVolumeSpecName: "kube-api-access-4r472") pod "12396b0e-ca72-4415-9b45-fc33d5345456" (UID: "12396b0e-ca72-4415-9b45-fc33d5345456"). InnerVolumeSpecName "kube-api-access-4r472". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:11:09.729621 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.729583 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" Apr 21 00:11:09.807275 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.807248 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4r472\" (UniqueName: \"kubernetes.io/projected/12396b0e-ca72-4415-9b45-fc33d5345456-kube-api-access-4r472\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:11:09.849752 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:09.849729 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf"] Apr 21 00:11:09.851417 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:11:09.851389 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64f29997_9e48_4455_85ed_1ca939a19be9.slice/crio-d0a11e994778a4e596ce288a3d95e6ea7bca4d9cc98320b3d5ea9a31498c3ec3 WatchSource:0}: Error finding container d0a11e994778a4e596ce288a3d95e6ea7bca4d9cc98320b3d5ea9a31498c3ec3: Status 404 returned error can't find the container with id d0a11e994778a4e596ce288a3d95e6ea7bca4d9cc98320b3d5ea9a31498c3ec3 Apr 21 00:11:10.353209 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.353178 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:10.353369 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.353247 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:10.358005 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.357982 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:10.360100 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.360076 2568 status_manager.go:895] "Failed to get status for pod" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" err="pods \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" is forbidden: User \"system:node:ip-10-0-132-206.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-132-206.ec2.internal' and this object" Apr 21 00:11:10.492231 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.492201 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" Apr 21 00:11:10.492231 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.492221 2568 scope.go:117] "RemoveContainer" containerID="4dc588e238ae614cd6da0e5f11b9c914c0e7862d6cb3554f74d4f0b3cd1d6aac" Apr 21 00:11:10.493749 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.493684 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" event={"ID":"64f29997-9e48-4455-85ed-1ca939a19be9","Type":"ContainerStarted","Data":"3e1bf89ca15ddd15681990f533787a454bacbec8a514fa4f49888e130e364014"} Apr 21 00:11:10.493749 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.493748 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" event={"ID":"64f29997-9e48-4455-85ed-1ca939a19be9","Type":"ContainerStarted","Data":"d0a11e994778a4e596ce288a3d95e6ea7bca4d9cc98320b3d5ea9a31498c3ec3"} Apr 21 00:11:10.493918 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.493782 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" Apr 21 00:11:10.494457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.494430 2568 status_manager.go:895] "Failed to get status for pod" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" err="pods \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" is forbidden: User \"system:node:ip-10-0-132-206.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-132-206.ec2.internal' and this object" Apr 21 00:11:10.496347 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.496323 2568 status_manager.go:895] "Failed to get status for pod" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" err="pods \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" is forbidden: User \"system:node:ip-10-0-132-206.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-132-206.ec2.internal' and this object" Apr 21 00:11:10.497884 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.497863 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5fb9cf77c4-lwm7z" Apr 21 00:11:10.511455 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.511418 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" podStartSLOduration=1.5114054989999999 podStartE2EDuration="1.511405499s" podCreationTimestamp="2026-04-21 00:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:11:10.51051712 +0000 UTC m=+484.335496181" watchObservedRunningTime="2026-04-21 00:11:10.511405499 +0000 UTC m=+484.336384464" Apr 21 00:11:10.532754 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.532722 2568 status_manager.go:895] "Failed to get status for pod" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" err="pods \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" is forbidden: User \"system:node:ip-10-0-132-206.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-132-206.ec2.internal' and this object" Apr 21 00:11:10.564969 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.564944 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-c9fddddfc-kfgvq"] Apr 21 00:11:10.567226 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.567201 2568 status_manager.go:895] "Failed to get status for pod" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4jn2g" err="pods \"limitador-operator-controller-manager-85c4996f8c-4jn2g\" is forbidden: User \"system:node:ip-10-0-132-206.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-132-206.ec2.internal' and this object" Apr 21 00:11:10.719579 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:10.719552 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12396b0e-ca72-4415-9b45-fc33d5345456" path="/var/lib/kubelet/pods/12396b0e-ca72-4415-9b45-fc33d5345456/volumes" Apr 21 00:11:21.502287 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:21.502256 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-4vgkf" Apr 21 00:11:35.589439 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.589401 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-c9fddddfc-kfgvq" podUID="12dee547-67d4-4598-8012-7a098c29cc7c" containerName="console" containerID="cri-o://42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b" gracePeriod=15 Apr 21 00:11:35.828546 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.828527 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-c9fddddfc-kfgvq_12dee547-67d4-4598-8012-7a098c29cc7c/console/0.log" Apr 21 00:11:35.828658 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.828585 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:11:35.930060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.929978 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-serving-cert\") pod \"12dee547-67d4-4598-8012-7a098c29cc7c\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " Apr 21 00:11:35.930060 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930052 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-service-ca\") pod \"12dee547-67d4-4598-8012-7a098c29cc7c\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " Apr 21 00:11:35.930236 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930077 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l75bx\" (UniqueName: \"kubernetes.io/projected/12dee547-67d4-4598-8012-7a098c29cc7c-kube-api-access-l75bx\") pod \"12dee547-67d4-4598-8012-7a098c29cc7c\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " Apr 21 00:11:35.930236 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930146 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-oauth-config\") pod \"12dee547-67d4-4598-8012-7a098c29cc7c\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " Apr 21 00:11:35.930236 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930184 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-console-config\") pod \"12dee547-67d4-4598-8012-7a098c29cc7c\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " Apr 21 00:11:35.930236 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930207 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-oauth-serving-cert\") pod \"12dee547-67d4-4598-8012-7a098c29cc7c\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " Apr 21 00:11:35.930441 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930242 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-trusted-ca-bundle\") pod \"12dee547-67d4-4598-8012-7a098c29cc7c\" (UID: \"12dee547-67d4-4598-8012-7a098c29cc7c\") " Apr 21 00:11:35.930571 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930545 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-service-ca" (OuterVolumeSpecName: "service-ca") pod "12dee547-67d4-4598-8012-7a098c29cc7c" (UID: "12dee547-67d4-4598-8012-7a098c29cc7c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:11:35.930709 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930676 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-console-config" (OuterVolumeSpecName: "console-config") pod "12dee547-67d4-4598-8012-7a098c29cc7c" (UID: "12dee547-67d4-4598-8012-7a098c29cc7c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:11:35.930786 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930759 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "12dee547-67d4-4598-8012-7a098c29cc7c" (UID: "12dee547-67d4-4598-8012-7a098c29cc7c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:11:35.930832 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.930801 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "12dee547-67d4-4598-8012-7a098c29cc7c" (UID: "12dee547-67d4-4598-8012-7a098c29cc7c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:11:35.932307 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.932282 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "12dee547-67d4-4598-8012-7a098c29cc7c" (UID: "12dee547-67d4-4598-8012-7a098c29cc7c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:11:35.932602 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.932580 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "12dee547-67d4-4598-8012-7a098c29cc7c" (UID: "12dee547-67d4-4598-8012-7a098c29cc7c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:11:35.932602 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:35.932593 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12dee547-67d4-4598-8012-7a098c29cc7c-kube-api-access-l75bx" (OuterVolumeSpecName: "kube-api-access-l75bx") pod "12dee547-67d4-4598-8012-7a098c29cc7c" (UID: "12dee547-67d4-4598-8012-7a098c29cc7c"). InnerVolumeSpecName "kube-api-access-l75bx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:11:36.031432 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.031400 2568 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-console-config\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:11:36.031432 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.031430 2568 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-oauth-serving-cert\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:11:36.031432 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.031439 2568 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-trusted-ca-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:11:36.031627 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.031448 2568 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-serving-cert\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:11:36.031627 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.031457 2568 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12dee547-67d4-4598-8012-7a098c29cc7c-service-ca\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:11:36.031627 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.031467 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l75bx\" (UniqueName: \"kubernetes.io/projected/12dee547-67d4-4598-8012-7a098c29cc7c-kube-api-access-l75bx\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:11:36.031627 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.031476 2568 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12dee547-67d4-4598-8012-7a098c29cc7c-console-oauth-config\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:11:36.599729 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.599700 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-c9fddddfc-kfgvq_12dee547-67d4-4598-8012-7a098c29cc7c/console/0.log" Apr 21 00:11:36.600126 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.599747 2568 generic.go:358] "Generic (PLEG): container finished" podID="12dee547-67d4-4598-8012-7a098c29cc7c" containerID="42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b" exitCode=2 Apr 21 00:11:36.600126 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.599784 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c9fddddfc-kfgvq" event={"ID":"12dee547-67d4-4598-8012-7a098c29cc7c","Type":"ContainerDied","Data":"42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b"} Apr 21 00:11:36.600126 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.599833 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c9fddddfc-kfgvq" Apr 21 00:11:36.600126 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.599845 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c9fddddfc-kfgvq" event={"ID":"12dee547-67d4-4598-8012-7a098c29cc7c","Type":"ContainerDied","Data":"d95c5a0f82f9902ad04cb0b945e9ed594b9a5d231e08a1bdfd6f2a44c41a6d1e"} Apr 21 00:11:36.600126 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.599867 2568 scope.go:117] "RemoveContainer" containerID="42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b" Apr 21 00:11:36.610035 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.610018 2568 scope.go:117] "RemoveContainer" containerID="42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b" Apr 21 00:11:36.610326 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:11:36.610300 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b\": container with ID starting with 42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b not found: ID does not exist" containerID="42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b" Apr 21 00:11:36.610393 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.610332 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b"} err="failed to get container status \"42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b\": rpc error: code = NotFound desc = could not find container \"42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b\": container with ID starting with 42455418fd8cbe3ed1a45dddfa4ec9b8a850863268e81c859e37a74908e96c4b not found: ID does not exist" Apr 21 00:11:36.622413 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.622387 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-c9fddddfc-kfgvq"] Apr 21 00:11:36.626188 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.626164 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-c9fddddfc-kfgvq"] Apr 21 00:11:36.719631 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:36.719586 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12dee547-67d4-4598-8012-7a098c29cc7c" path="/var/lib/kubelet/pods/12dee547-67d4-4598-8012-7a098c29cc7c/volumes" Apr 21 00:11:38.016597 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.016552 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4"] Apr 21 00:11:38.017529 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.017504 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="12dee547-67d4-4598-8012-7a098c29cc7c" containerName="console" Apr 21 00:11:38.017529 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.017527 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="12dee547-67d4-4598-8012-7a098c29cc7c" containerName="console" Apr 21 00:11:38.017741 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.017602 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="12dee547-67d4-4598-8012-7a098c29cc7c" containerName="console" Apr 21 00:11:38.022166 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.022142 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.025095 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.025069 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"maas-default-gateway-openshift-default-dockercfg-flfnm\"" Apr 21 00:11:38.027455 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.027423 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4"] Apr 21 00:11:38.148457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148419 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-envoy\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.148653 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148512 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-podinfo\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.148653 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148572 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-token\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.148653 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148650 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwgzb\" (UniqueName: \"kubernetes.io/projected/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-kube-api-access-lwgzb\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.148794 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148672 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-workload-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.148794 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148693 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-workload-certs\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.148860 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148790 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-data\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.148860 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148834 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-credential-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.148860 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.148851 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istiod-ca-cert\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.249826 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.249801 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-credential-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.249826 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.249831 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istiod-ca-cert\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250055 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.249852 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-envoy\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250055 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.249977 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-podinfo\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250155 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250119 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-token\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250219 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250200 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lwgzb\" (UniqueName: \"kubernetes.io/projected/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-kube-api-access-lwgzb\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250278 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250237 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-workload-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250278 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250251 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"credential-socket\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-credential-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250278 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250263 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-workload-certs\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250421 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250323 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-data\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250517 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250492 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"workload-socket\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-workload-socket\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250672 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-data\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-data\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250787 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250675 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istiod-ca-cert\" (UniqueName: \"kubernetes.io/configmap/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istiod-ca-cert\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.250787 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.250752 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"workload-certs\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-workload-certs\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.252151 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.252129 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-envoy\" (UniqueName: \"kubernetes.io/empty-dir/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-envoy\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.252454 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.252436 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-podinfo\" (UniqueName: \"kubernetes.io/downward-api/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-podinfo\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.257348 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.257323 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"istio-token\" (UniqueName: \"kubernetes.io/projected/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-istio-token\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.257541 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.257520 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwgzb\" (UniqueName: \"kubernetes.io/projected/39ea85a6-6e98-4c3c-a41c-c32ea72a6923-kube-api-access-lwgzb\") pod \"maas-default-gateway-openshift-default-58b6f876-qgzn4\" (UID: \"39ea85a6-6e98-4c3c-a41c-c32ea72a6923\") " pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.334445 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.334386 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:38.456487 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.456458 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4"] Apr 21 00:11:38.457673 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:11:38.457647 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39ea85a6_6e98_4c3c_a41c_c32ea72a6923.slice/crio-4b1b44933b76be6b9dca6722f91a661c45404077e884cabe109d5cbf4e6ad649 WatchSource:0}: Error finding container 4b1b44933b76be6b9dca6722f91a661c45404077e884cabe109d5cbf4e6ad649: Status 404 returned error can't find the container with id 4b1b44933b76be6b9dca6722f91a661c45404077e884cabe109d5cbf4e6ad649 Apr 21 00:11:38.460218 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.460184 2568 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"31236220Ki","pods":"250"} Apr 21 00:11:38.460290 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.460260 2568 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"31236220Ki","pods":"250"} Apr 21 00:11:38.460336 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.460290 2568 kubelet_resources.go:45] "Allocatable" allocatable={"cpu":"7500m","ephemeral-storage":"114345831029","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"31236220Ki","pods":"250"} Apr 21 00:11:38.609576 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.609505 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" event={"ID":"39ea85a6-6e98-4c3c-a41c-c32ea72a6923","Type":"ContainerStarted","Data":"aa2047b01055dfa78a197975e31d3d47ef0f306ecb4c6e430d2ef9f4c3c08d05"} Apr 21 00:11:38.609576 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.609539 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" event={"ID":"39ea85a6-6e98-4c3c-a41c-c32ea72a6923","Type":"ContainerStarted","Data":"4b1b44933b76be6b9dca6722f91a661c45404077e884cabe109d5cbf4e6ad649"} Apr 21 00:11:38.627401 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:38.627340 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" podStartSLOduration=1.6273208270000001 podStartE2EDuration="1.627320827s" podCreationTimestamp="2026-04-21 00:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:11:38.625648641 +0000 UTC m=+512.450627608" watchObservedRunningTime="2026-04-21 00:11:38.627320827 +0000 UTC m=+512.452299793" Apr 21 00:11:39.334539 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:39.334506 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:39.339345 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:39.339315 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:39.613633 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:39.613543 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:39.614635 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:39.614618 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/maas-default-gateway-openshift-default-58b6f876-qgzn4" Apr 21 00:11:56.686355 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.686272 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-ww5jn"] Apr 21 00:11:56.689567 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.689545 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" Apr 21 00:11:56.692121 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.692099 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-authorino-dockercfg-jvk25\"" Apr 21 00:11:56.695754 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.695729 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-ww5jn"] Apr 21 00:11:56.785515 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.785483 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd5fx\" (UniqueName: \"kubernetes.io/projected/f2ae5e1c-d9b3-435f-92d9-b151e15d626a-kube-api-access-bd5fx\") pod \"authorino-f99f4b5cd-ww5jn\" (UID: \"f2ae5e1c-d9b3-435f-92d9-b151e15d626a\") " pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" Apr 21 00:11:56.861693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.861669 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-7498df8756-s4hm2"] Apr 21 00:11:56.864759 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.864745 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-7498df8756-s4hm2" Apr 21 00:11:56.874373 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.874356 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-7498df8756-s4hm2"] Apr 21 00:11:56.886200 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.886178 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bd5fx\" (UniqueName: \"kubernetes.io/projected/f2ae5e1c-d9b3-435f-92d9-b151e15d626a-kube-api-access-bd5fx\") pod \"authorino-f99f4b5cd-ww5jn\" (UID: \"f2ae5e1c-d9b3-435f-92d9-b151e15d626a\") " pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" Apr 21 00:11:56.893152 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.893129 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd5fx\" (UniqueName: \"kubernetes.io/projected/f2ae5e1c-d9b3-435f-92d9-b151e15d626a-kube-api-access-bd5fx\") pod \"authorino-f99f4b5cd-ww5jn\" (UID: \"f2ae5e1c-d9b3-435f-92d9-b151e15d626a\") " pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" Apr 21 00:11:56.987569 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:56.987544 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b6kk\" (UniqueName: \"kubernetes.io/projected/e705177c-f4f3-41ce-9b40-370df2b2a68d-kube-api-access-8b6kk\") pod \"authorino-7498df8756-s4hm2\" (UID: \"e705177c-f4f3-41ce-9b40-370df2b2a68d\") " pod="kuadrant-system/authorino-7498df8756-s4hm2" Apr 21 00:11:57.002554 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:57.002537 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" Apr 21 00:11:57.088232 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:57.088200 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8b6kk\" (UniqueName: \"kubernetes.io/projected/e705177c-f4f3-41ce-9b40-370df2b2a68d-kube-api-access-8b6kk\") pod \"authorino-7498df8756-s4hm2\" (UID: \"e705177c-f4f3-41ce-9b40-370df2b2a68d\") " pod="kuadrant-system/authorino-7498df8756-s4hm2" Apr 21 00:11:57.098469 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:57.098447 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b6kk\" (UniqueName: \"kubernetes.io/projected/e705177c-f4f3-41ce-9b40-370df2b2a68d-kube-api-access-8b6kk\") pod \"authorino-7498df8756-s4hm2\" (UID: \"e705177c-f4f3-41ce-9b40-370df2b2a68d\") " pod="kuadrant-system/authorino-7498df8756-s4hm2" Apr 21 00:11:57.118553 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:57.118527 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-ww5jn"] Apr 21 00:11:57.121334 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:11:57.121314 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2ae5e1c_d9b3_435f_92d9_b151e15d626a.slice/crio-9f0c7c560fb6d7d9ddba5c03675cba4bf5ed6502426913c9917f80ce9355c2e3 WatchSource:0}: Error finding container 9f0c7c560fb6d7d9ddba5c03675cba4bf5ed6502426913c9917f80ce9355c2e3: Status 404 returned error can't find the container with id 9f0c7c560fb6d7d9ddba5c03675cba4bf5ed6502426913c9917f80ce9355c2e3 Apr 21 00:11:57.174140 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:57.174120 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-7498df8756-s4hm2" Apr 21 00:11:57.288977 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:57.288951 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-7498df8756-s4hm2"] Apr 21 00:11:57.290409 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:11:57.290385 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode705177c_f4f3_41ce_9b40_370df2b2a68d.slice/crio-e33fbaffe59ac41c1a095ee59c215c7926681c5b3831af12e925f9c86b430187 WatchSource:0}: Error finding container e33fbaffe59ac41c1a095ee59c215c7926681c5b3831af12e925f9c86b430187: Status 404 returned error can't find the container with id e33fbaffe59ac41c1a095ee59c215c7926681c5b3831af12e925f9c86b430187 Apr 21 00:11:57.685968 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:57.685886 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-7498df8756-s4hm2" event={"ID":"e705177c-f4f3-41ce-9b40-370df2b2a68d","Type":"ContainerStarted","Data":"e33fbaffe59ac41c1a095ee59c215c7926681c5b3831af12e925f9c86b430187"} Apr 21 00:11:57.686971 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:11:57.686946 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" event={"ID":"f2ae5e1c-d9b3-435f-92d9-b151e15d626a","Type":"ContainerStarted","Data":"9f0c7c560fb6d7d9ddba5c03675cba4bf5ed6502426913c9917f80ce9355c2e3"} Apr 21 00:12:01.709107 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:01.709072 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-7498df8756-s4hm2" event={"ID":"e705177c-f4f3-41ce-9b40-370df2b2a68d","Type":"ContainerStarted","Data":"4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b"} Apr 21 00:12:01.710395 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:01.710367 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" event={"ID":"f2ae5e1c-d9b3-435f-92d9-b151e15d626a","Type":"ContainerStarted","Data":"8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc"} Apr 21 00:12:01.724500 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:01.724457 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-7498df8756-s4hm2" podStartSLOduration=2.3549761240000002 podStartE2EDuration="5.724444718s" podCreationTimestamp="2026-04-21 00:11:56 +0000 UTC" firstStartedPulling="2026-04-21 00:11:57.29168909 +0000 UTC m=+531.116668033" lastFinishedPulling="2026-04-21 00:12:00.66115767 +0000 UTC m=+534.486136627" observedRunningTime="2026-04-21 00:12:01.721670267 +0000 UTC m=+535.546649232" watchObservedRunningTime="2026-04-21 00:12:01.724444718 +0000 UTC m=+535.549423682" Apr 21 00:12:01.735361 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:01.735322 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" podStartSLOduration=2.207610774 podStartE2EDuration="5.73531009s" podCreationTimestamp="2026-04-21 00:11:56 +0000 UTC" firstStartedPulling="2026-04-21 00:11:57.122719321 +0000 UTC m=+530.947698264" lastFinishedPulling="2026-04-21 00:12:00.650418629 +0000 UTC m=+534.475397580" observedRunningTime="2026-04-21 00:12:01.73362512 +0000 UTC m=+535.558604089" watchObservedRunningTime="2026-04-21 00:12:01.73531009 +0000 UTC m=+535.560289055" Apr 21 00:12:01.758264 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:01.758242 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-ww5jn"] Apr 21 00:12:03.717669 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:03.717596 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" podUID="f2ae5e1c-d9b3-435f-92d9-b151e15d626a" containerName="authorino" containerID="cri-o://8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc" gracePeriod=30 Apr 21 00:12:03.952455 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:03.952436 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" Apr 21 00:12:04.048997 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.048906 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd5fx\" (UniqueName: \"kubernetes.io/projected/f2ae5e1c-d9b3-435f-92d9-b151e15d626a-kube-api-access-bd5fx\") pod \"f2ae5e1c-d9b3-435f-92d9-b151e15d626a\" (UID: \"f2ae5e1c-d9b3-435f-92d9-b151e15d626a\") " Apr 21 00:12:04.050963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.050937 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2ae5e1c-d9b3-435f-92d9-b151e15d626a-kube-api-access-bd5fx" (OuterVolumeSpecName: "kube-api-access-bd5fx") pod "f2ae5e1c-d9b3-435f-92d9-b151e15d626a" (UID: "f2ae5e1c-d9b3-435f-92d9-b151e15d626a"). InnerVolumeSpecName "kube-api-access-bd5fx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:12:04.150252 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.150228 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-bd5fx\" (UniqueName: \"kubernetes.io/projected/f2ae5e1c-d9b3-435f-92d9-b151e15d626a-kube-api-access-bd5fx\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:12:04.722468 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.722428 2568 generic.go:358] "Generic (PLEG): container finished" podID="f2ae5e1c-d9b3-435f-92d9-b151e15d626a" containerID="8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc" exitCode=0 Apr 21 00:12:04.722915 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.722489 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" Apr 21 00:12:04.722915 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.722513 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" event={"ID":"f2ae5e1c-d9b3-435f-92d9-b151e15d626a","Type":"ContainerDied","Data":"8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc"} Apr 21 00:12:04.722915 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.722556 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-ww5jn" event={"ID":"f2ae5e1c-d9b3-435f-92d9-b151e15d626a","Type":"ContainerDied","Data":"9f0c7c560fb6d7d9ddba5c03675cba4bf5ed6502426913c9917f80ce9355c2e3"} Apr 21 00:12:04.722915 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.722579 2568 scope.go:117] "RemoveContainer" containerID="8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc" Apr 21 00:12:04.732106 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.732087 2568 scope.go:117] "RemoveContainer" containerID="8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc" Apr 21 00:12:04.732354 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:12:04.732336 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc\": container with ID starting with 8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc not found: ID does not exist" containerID="8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc" Apr 21 00:12:04.732412 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.732360 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc"} err="failed to get container status \"8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc\": rpc error: code = NotFound desc = could not find container \"8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc\": container with ID starting with 8c647739d2e3c76c000e039ff25f9a745c8e127c8ea115557ea14cca16b234cc not found: ID does not exist" Apr 21 00:12:04.743414 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.743387 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-ww5jn"] Apr 21 00:12:04.745494 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:04.745465 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-ww5jn"] Apr 21 00:12:06.719111 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:06.719082 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2ae5e1c-d9b3-435f-92d9-b151e15d626a" path="/var/lib/kubelet/pods/f2ae5e1c-d9b3-435f-92d9-b151e15d626a/volumes" Apr 21 00:12:26.616667 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.616633 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg"] Apr 21 00:12:26.617086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.616949 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f2ae5e1c-d9b3-435f-92d9-b151e15d626a" containerName="authorino" Apr 21 00:12:26.617086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.616960 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2ae5e1c-d9b3-435f-92d9-b151e15d626a" containerName="authorino" Apr 21 00:12:26.617086 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.617023 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="f2ae5e1c-d9b3-435f-92d9-b151e15d626a" containerName="authorino" Apr 21 00:12:26.620110 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.620091 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.622593 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.622571 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 21 00:12:26.622593 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.622585 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-4fqdf\"" Apr 21 00:12:26.622733 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.622579 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 21 00:12:26.627782 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.627761 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg"] Apr 21 00:12:26.714630 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.714577 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlmsl\" (UniqueName: \"kubernetes.io/projected/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-kube-api-access-tlmsl\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.714757 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.714688 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-util\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.714757 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.714712 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-bundle\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.815049 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.815027 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-bundle\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.815147 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.815060 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tlmsl\" (UniqueName: \"kubernetes.io/projected/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-kube-api-access-tlmsl\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.815197 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.815165 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-util\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.815424 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.815394 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-bundle\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.815563 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.815547 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-util\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.824627 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.824595 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlmsl\" (UniqueName: \"kubernetes.io/projected/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-kube-api-access-tlmsl\") pod \"7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:26.929757 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:26.929709 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:27.049348 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:27.049322 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg"] Apr 21 00:12:27.050705 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:12:27.050675 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4da9639e_14d3_4c6b_8f4a_831d4c0dfd66.slice/crio-b6faf569b2a23814c3fc16b9107985b482eeb56f0953cb4ef2507d369825273c WatchSource:0}: Error finding container b6faf569b2a23814c3fc16b9107985b482eeb56f0953cb4ef2507d369825273c: Status 404 returned error can't find the container with id b6faf569b2a23814c3fc16b9107985b482eeb56f0953cb4ef2507d369825273c Apr 21 00:12:27.802085 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:27.802052 2568 generic.go:358] "Generic (PLEG): container finished" podID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerID="f61da6f479208a3a4efd6fdea0de7f0f1027c8bec5d4f7f3aeb5fcfbbc04abe3" exitCode=0 Apr 21 00:12:27.802553 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:27.802121 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" event={"ID":"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66","Type":"ContainerDied","Data":"f61da6f479208a3a4efd6fdea0de7f0f1027c8bec5d4f7f3aeb5fcfbbc04abe3"} Apr 21 00:12:27.802553 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:27.802146 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" event={"ID":"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66","Type":"ContainerStarted","Data":"b6faf569b2a23814c3fc16b9107985b482eeb56f0953cb4ef2507d369825273c"} Apr 21 00:12:28.807072 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:28.806992 2568 generic.go:358] "Generic (PLEG): container finished" podID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerID="0ab2afc0c24198c5b69dfb61b58077eb5a712e6fc7ab84b7a5b3fbe1ad59a439" exitCode=0 Apr 21 00:12:28.807379 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:28.807080 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" event={"ID":"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66","Type":"ContainerDied","Data":"0ab2afc0c24198c5b69dfb61b58077eb5a712e6fc7ab84b7a5b3fbe1ad59a439"} Apr 21 00:12:29.812036 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:29.811998 2568 generic.go:358] "Generic (PLEG): container finished" podID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerID="897309bf6e1ebb964f13660ff62153eaa5ae45631a5b9465641e611363ebe77b" exitCode=0 Apr 21 00:12:29.812430 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:29.812078 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" event={"ID":"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66","Type":"ContainerDied","Data":"897309bf6e1ebb964f13660ff62153eaa5ae45631a5b9465641e611363ebe77b"} Apr 21 00:12:30.938969 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:30.938946 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:31.048516 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.048490 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-util\") pod \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " Apr 21 00:12:31.048665 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.048542 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-bundle\") pod \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " Apr 21 00:12:31.048665 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.048631 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlmsl\" (UniqueName: \"kubernetes.io/projected/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-kube-api-access-tlmsl\") pod \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\" (UID: \"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66\") " Apr 21 00:12:31.049051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.049020 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-bundle" (OuterVolumeSpecName: "bundle") pod "4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" (UID: "4da9639e-14d3-4c6b-8f4a-831d4c0dfd66"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:12:31.050573 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.050553 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-kube-api-access-tlmsl" (OuterVolumeSpecName: "kube-api-access-tlmsl") pod "4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" (UID: "4da9639e-14d3-4c6b-8f4a-831d4c0dfd66"). InnerVolumeSpecName "kube-api-access-tlmsl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:12:31.055880 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.055856 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-util" (OuterVolumeSpecName: "util") pod "4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" (UID: "4da9639e-14d3-4c6b-8f4a-831d4c0dfd66"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 00:12:31.149590 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.149502 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tlmsl\" (UniqueName: \"kubernetes.io/projected/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-kube-api-access-tlmsl\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:12:31.149590 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.149542 2568 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-util\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:12:31.149590 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.149555 2568 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4da9639e-14d3-4c6b-8f4a-831d4c0dfd66-bundle\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:12:31.822293 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.822253 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" event={"ID":"4da9639e-14d3-4c6b-8f4a-831d4c0dfd66","Type":"ContainerDied","Data":"b6faf569b2a23814c3fc16b9107985b482eeb56f0953cb4ef2507d369825273c"} Apr 21 00:12:31.822293 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.822282 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7471e62b3b524e5f63095613ed1aa2d2aa2beb9d9bc7600d699dae1350svbcg" Apr 21 00:12:31.822572 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:31.822288 2568 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6faf569b2a23814c3fc16b9107985b482eeb56f0953cb4ef2507d369825273c" Apr 21 00:12:53.875241 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.875209 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:12:53.875703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.875549 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerName="util" Apr 21 00:12:53.875703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.875560 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerName="util" Apr 21 00:12:53.875703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.875567 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerName="pull" Apr 21 00:12:53.875703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.875573 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerName="pull" Apr 21 00:12:53.875703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.875583 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerName="extract" Apr 21 00:12:53.875703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.875588 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerName="extract" Apr 21 00:12:53.875703 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.875656 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="4da9639e-14d3-4c6b-8f4a-831d4c0dfd66" containerName="extract" Apr 21 00:12:53.879979 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.879963 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keycloak-system/maas-keycloak-0" Apr 21 00:12:53.882533 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.882511 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"keycloak-system\"/\"maas-keycloak-initial-admin\"" Apr 21 00:12:53.882672 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.882531 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"keycloak-system\"/\"default-dockercfg-sndxl\"" Apr 21 00:12:53.883864 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.883833 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"keycloak-system\"/\"kube-root-ca.crt\"" Apr 21 00:12:53.884266 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.884239 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"keycloak-system\"/\"openshift-service-ca.crt\"" Apr 21 00:12:53.886701 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:53.886681 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:12:54.020871 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:54.020827 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmsn6\" (UniqueName: \"kubernetes.io/projected/769b108a-869c-4f4d-a678-70f55dfab5eb-kube-api-access-tmsn6\") pod \"maas-keycloak-0\" (UID: \"769b108a-869c-4f4d-a678-70f55dfab5eb\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:12:54.122077 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:54.122048 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-tmsn6\" (UniqueName: \"kubernetes.io/projected/769b108a-869c-4f4d-a678-70f55dfab5eb-kube-api-access-tmsn6\") pod \"maas-keycloak-0\" (UID: \"769b108a-869c-4f4d-a678-70f55dfab5eb\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:12:54.130197 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:54.130140 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmsn6\" (UniqueName: \"kubernetes.io/projected/769b108a-869c-4f4d-a678-70f55dfab5eb-kube-api-access-tmsn6\") pod \"maas-keycloak-0\" (UID: \"769b108a-869c-4f4d-a678-70f55dfab5eb\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:12:54.190334 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:54.190309 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keycloak-system/maas-keycloak-0" Apr 21 00:12:54.313189 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:54.313165 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:12:54.314482 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:12:54.314450 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod769b108a_869c_4f4d_a678_70f55dfab5eb.slice/crio-eabe35b0ec9e6ce7c5b750999489293379c65e67f88fc35362ec5f17e2743d75 WatchSource:0}: Error finding container eabe35b0ec9e6ce7c5b750999489293379c65e67f88fc35362ec5f17e2743d75: Status 404 returned error can't find the container with id eabe35b0ec9e6ce7c5b750999489293379c65e67f88fc35362ec5f17e2743d75 Apr 21 00:12:54.905848 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:54.905802 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="keycloak-system/maas-keycloak-0" event={"ID":"769b108a-869c-4f4d-a678-70f55dfab5eb","Type":"ContainerStarted","Data":"eabe35b0ec9e6ce7c5b750999489293379c65e67f88fc35362ec5f17e2743d75"} Apr 21 00:12:59.929457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:59.929414 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="keycloak-system/maas-keycloak-0" event={"ID":"769b108a-869c-4f4d-a678-70f55dfab5eb","Type":"ContainerStarted","Data":"4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa"} Apr 21 00:12:59.947881 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:12:59.947821 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keycloak-system/maas-keycloak-0" podStartSLOduration=2.012248146 podStartE2EDuration="6.947804277s" podCreationTimestamp="2026-04-21 00:12:53 +0000 UTC" firstStartedPulling="2026-04-21 00:12:54.31568726 +0000 UTC m=+588.140666204" lastFinishedPulling="2026-04-21 00:12:59.251243388 +0000 UTC m=+593.076222335" observedRunningTime="2026-04-21 00:12:59.944240788 +0000 UTC m=+593.769219753" watchObservedRunningTime="2026-04-21 00:12:59.947804277 +0000 UTC m=+593.772783243" Apr 21 00:13:00.190888 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:00.190769 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="keycloak-system/maas-keycloak-0" Apr 21 00:13:00.192438 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:00.192405 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:01.191372 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:01.191322 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:02.191763 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:02.191715 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:03.191665 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:03.191594 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:04.191443 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:04.191381 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:04.191949 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:04.191865 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="keycloak-system/maas-keycloak-0" Apr 21 00:13:05.191445 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:05.191396 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:06.191470 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:06.191425 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:06.697964 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:06.697929 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:13:06.698860 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:06.698832 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:13:07.190716 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:07.190671 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:08.191521 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:08.191477 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:09.191479 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:09.191423 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:10.191128 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:10.191078 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:11.191052 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:11.191008 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.45:9000/health/started\": dial tcp 10.133.0.45:9000: connect: connection refused" Apr 21 00:13:12.287809 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:12.287772 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="keycloak-system/maas-keycloak-0" Apr 21 00:13:12.305310 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:12.305260 2568 prober.go:120] "Probe failed" probeType="Readiness" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 21 00:13:22.294005 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:22.293967 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="keycloak-system/maas-keycloak-0" Apr 21 00:13:23.182045 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.182014 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-8b475cf9f-vd27s"] Apr 21 00:13:23.194648 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.194619 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-vd27s"] Apr 21 00:13:23.194786 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.194734 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-vd27s" Apr 21 00:13:23.284892 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.284869 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9l8h\" (UniqueName: \"kubernetes.io/projected/3467b6c2-2349-4f2c-b5a7-3f6fc46a7993-kube-api-access-j9l8h\") pod \"authorino-8b475cf9f-vd27s\" (UID: \"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993\") " pod="kuadrant-system/authorino-8b475cf9f-vd27s" Apr 21 00:13:23.385423 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.385400 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-j9l8h\" (UniqueName: \"kubernetes.io/projected/3467b6c2-2349-4f2c-b5a7-3f6fc46a7993-kube-api-access-j9l8h\") pod \"authorino-8b475cf9f-vd27s\" (UID: \"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993\") " pod="kuadrant-system/authorino-8b475cf9f-vd27s" Apr 21 00:13:23.392864 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.392843 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9l8h\" (UniqueName: \"kubernetes.io/projected/3467b6c2-2349-4f2c-b5a7-3f6fc46a7993-kube-api-access-j9l8h\") pod \"authorino-8b475cf9f-vd27s\" (UID: \"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993\") " pod="kuadrant-system/authorino-8b475cf9f-vd27s" Apr 21 00:13:23.465397 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.465344 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-vd27s"] Apr 21 00:13:23.465529 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.465517 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-vd27s" Apr 21 00:13:23.490619 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.490579 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-5f85766d6d-6qc8b"] Apr 21 00:13:23.495188 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.495173 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:13:23.498095 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.498075 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-server-cert\"" Apr 21 00:13:23.503131 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.503109 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-5f85766d6d-6qc8b"] Apr 21 00:13:23.587364 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.587191 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/faaaf689-f76b-44cf-b052-656e71b1b42e-tls-cert\") pod \"authorino-5f85766d6d-6qc8b\" (UID: \"faaaf689-f76b-44cf-b052-656e71b1b42e\") " pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:13:23.587364 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.587307 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcctd\" (UniqueName: \"kubernetes.io/projected/faaaf689-f76b-44cf-b052-656e71b1b42e-kube-api-access-qcctd\") pod \"authorino-5f85766d6d-6qc8b\" (UID: \"faaaf689-f76b-44cf-b052-656e71b1b42e\") " pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:13:23.587513 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.587493 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-vd27s"] Apr 21 00:13:23.589777 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:13:23.589754 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3467b6c2_2349_4f2c_b5a7_3f6fc46a7993.slice/crio-d9210c1267866c9a37a4ad420281aa366cd6be0f5635148a6674a785df66a75b WatchSource:0}: Error finding container d9210c1267866c9a37a4ad420281aa366cd6be0f5635148a6674a785df66a75b: Status 404 returned error can't find the container with id d9210c1267866c9a37a4ad420281aa366cd6be0f5635148a6674a785df66a75b Apr 21 00:13:23.591491 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.591472 2568 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 00:13:23.687899 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.687875 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/faaaf689-f76b-44cf-b052-656e71b1b42e-tls-cert\") pod \"authorino-5f85766d6d-6qc8b\" (UID: \"faaaf689-f76b-44cf-b052-656e71b1b42e\") " pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:13:23.688010 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.687965 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qcctd\" (UniqueName: \"kubernetes.io/projected/faaaf689-f76b-44cf-b052-656e71b1b42e-kube-api-access-qcctd\") pod \"authorino-5f85766d6d-6qc8b\" (UID: \"faaaf689-f76b-44cf-b052-656e71b1b42e\") " pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:13:23.690297 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.690275 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/faaaf689-f76b-44cf-b052-656e71b1b42e-tls-cert\") pod \"authorino-5f85766d6d-6qc8b\" (UID: \"faaaf689-f76b-44cf-b052-656e71b1b42e\") " pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:13:23.695022 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.694999 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcctd\" (UniqueName: \"kubernetes.io/projected/faaaf689-f76b-44cf-b052-656e71b1b42e-kube-api-access-qcctd\") pod \"authorino-5f85766d6d-6qc8b\" (UID: \"faaaf689-f76b-44cf-b052-656e71b1b42e\") " pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:13:23.805655 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.805627 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:13:23.921111 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:23.921089 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-5f85766d6d-6qc8b"] Apr 21 00:13:23.922421 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:13:23.922399 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaaaf689_f76b_44cf_b052_656e71b1b42e.slice/crio-f53c9ef8d08ae0db5e1e08815d3f59e4ca67ff42a99e168239f51f408f6aa695 WatchSource:0}: Error finding container f53c9ef8d08ae0db5e1e08815d3f59e4ca67ff42a99e168239f51f408f6aa695: Status 404 returned error can't find the container with id f53c9ef8d08ae0db5e1e08815d3f59e4ca67ff42a99e168239f51f408f6aa695 Apr 21 00:13:24.034672 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.034643 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" event={"ID":"faaaf689-f76b-44cf-b052-656e71b1b42e","Type":"ContainerStarted","Data":"f53c9ef8d08ae0db5e1e08815d3f59e4ca67ff42a99e168239f51f408f6aa695"} Apr 21 00:13:24.036138 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.036118 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-8b475cf9f-vd27s" event={"ID":"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993","Type":"ContainerStarted","Data":"aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e"} Apr 21 00:13:24.036395 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.036142 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-8b475cf9f-vd27s" event={"ID":"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993","Type":"ContainerStarted","Data":"d9210c1267866c9a37a4ad420281aa366cd6be0f5635148a6674a785df66a75b"} Apr 21 00:13:24.036395 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.036222 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-8b475cf9f-vd27s" podUID="3467b6c2-2349-4f2c-b5a7-3f6fc46a7993" containerName="authorino" containerID="cri-o://aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e" gracePeriod=30 Apr 21 00:13:24.051251 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.051213 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-8b475cf9f-vd27s" podStartSLOduration=0.68436421 podStartE2EDuration="1.051199511s" podCreationTimestamp="2026-04-21 00:13:23 +0000 UTC" firstStartedPulling="2026-04-21 00:13:23.591692349 +0000 UTC m=+617.416671308" lastFinishedPulling="2026-04-21 00:13:23.958527666 +0000 UTC m=+617.783506609" observedRunningTime="2026-04-21 00:13:24.04919358 +0000 UTC m=+617.874172546" watchObservedRunningTime="2026-04-21 00:13:24.051199511 +0000 UTC m=+617.876178539" Apr 21 00:13:24.256534 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.256512 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-vd27s" Apr 21 00:13:24.293672 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.293644 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9l8h\" (UniqueName: \"kubernetes.io/projected/3467b6c2-2349-4f2c-b5a7-3f6fc46a7993-kube-api-access-j9l8h\") pod \"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993\" (UID: \"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993\") " Apr 21 00:13:24.295706 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.295676 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3467b6c2-2349-4f2c-b5a7-3f6fc46a7993-kube-api-access-j9l8h" (OuterVolumeSpecName: "kube-api-access-j9l8h") pod "3467b6c2-2349-4f2c-b5a7-3f6fc46a7993" (UID: "3467b6c2-2349-4f2c-b5a7-3f6fc46a7993"). InnerVolumeSpecName "kube-api-access-j9l8h". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:13:24.395145 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:24.395116 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-j9l8h\" (UniqueName: \"kubernetes.io/projected/3467b6c2-2349-4f2c-b5a7-3f6fc46a7993-kube-api-access-j9l8h\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:13:25.041484 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.041399 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" event={"ID":"faaaf689-f76b-44cf-b052-656e71b1b42e","Type":"ContainerStarted","Data":"cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123"} Apr 21 00:13:25.042569 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.042545 2568 generic.go:358] "Generic (PLEG): container finished" podID="3467b6c2-2349-4f2c-b5a7-3f6fc46a7993" containerID="aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e" exitCode=0 Apr 21 00:13:25.042690 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.042601 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-vd27s" Apr 21 00:13:25.042690 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.042642 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-8b475cf9f-vd27s" event={"ID":"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993","Type":"ContainerDied","Data":"aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e"} Apr 21 00:13:25.042690 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.042677 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-8b475cf9f-vd27s" event={"ID":"3467b6c2-2349-4f2c-b5a7-3f6fc46a7993","Type":"ContainerDied","Data":"d9210c1267866c9a37a4ad420281aa366cd6be0f5635148a6674a785df66a75b"} Apr 21 00:13:25.042787 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.042696 2568 scope.go:117] "RemoveContainer" containerID="aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e" Apr 21 00:13:25.051197 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.051177 2568 scope.go:117] "RemoveContainer" containerID="aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e" Apr 21 00:13:25.051468 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:13:25.051452 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e\": container with ID starting with aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e not found: ID does not exist" containerID="aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e" Apr 21 00:13:25.051514 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.051475 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e"} err="failed to get container status \"aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e\": rpc error: code = NotFound desc = could not find container \"aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e\": container with ID starting with aa5647d60ad1758247b1324e9d64fe18f80291600adae2081c9b5a9b18ca867e not found: ID does not exist" Apr 21 00:13:25.057055 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.057011 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" podStartSLOduration=1.68437879 podStartE2EDuration="2.057000384s" podCreationTimestamp="2026-04-21 00:13:23 +0000 UTC" firstStartedPulling="2026-04-21 00:13:23.923748288 +0000 UTC m=+617.748727232" lastFinishedPulling="2026-04-21 00:13:24.296369883 +0000 UTC m=+618.121348826" observedRunningTime="2026-04-21 00:13:25.055011128 +0000 UTC m=+618.879990145" watchObservedRunningTime="2026-04-21 00:13:25.057000384 +0000 UTC m=+618.881979350" Apr 21 00:13:25.071516 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.071493 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-vd27s"] Apr 21 00:13:25.077952 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.077929 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-vd27s"] Apr 21 00:13:25.080835 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.080814 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-7498df8756-s4hm2"] Apr 21 00:13:25.081021 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.080984 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-7498df8756-s4hm2" podUID="e705177c-f4f3-41ce-9b40-370df2b2a68d" containerName="authorino" containerID="cri-o://4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b" gracePeriod=30 Apr 21 00:13:25.319504 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.319482 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-7498df8756-s4hm2" Apr 21 00:13:25.402093 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.402062 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b6kk\" (UniqueName: \"kubernetes.io/projected/e705177c-f4f3-41ce-9b40-370df2b2a68d-kube-api-access-8b6kk\") pod \"e705177c-f4f3-41ce-9b40-370df2b2a68d\" (UID: \"e705177c-f4f3-41ce-9b40-370df2b2a68d\") " Apr 21 00:13:25.404033 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.404006 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e705177c-f4f3-41ce-9b40-370df2b2a68d-kube-api-access-8b6kk" (OuterVolumeSpecName: "kube-api-access-8b6kk") pod "e705177c-f4f3-41ce-9b40-370df2b2a68d" (UID: "e705177c-f4f3-41ce-9b40-370df2b2a68d"). InnerVolumeSpecName "kube-api-access-8b6kk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:13:25.503509 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.503481 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8b6kk\" (UniqueName: \"kubernetes.io/projected/e705177c-f4f3-41ce-9b40-370df2b2a68d-kube-api-access-8b6kk\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:13:25.832390 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.832359 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-d2r2x"] Apr 21 00:13:25.832796 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.832771 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e705177c-f4f3-41ce-9b40-370df2b2a68d" containerName="authorino" Apr 21 00:13:25.832796 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.832799 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="e705177c-f4f3-41ce-9b40-370df2b2a68d" containerName="authorino" Apr 21 00:13:25.832956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.832816 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3467b6c2-2349-4f2c-b5a7-3f6fc46a7993" containerName="authorino" Apr 21 00:13:25.832956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.832824 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="3467b6c2-2349-4f2c-b5a7-3f6fc46a7993" containerName="authorino" Apr 21 00:13:25.832956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.832906 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="e705177c-f4f3-41ce-9b40-370df2b2a68d" containerName="authorino" Apr 21 00:13:25.832956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.832919 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="3467b6c2-2349-4f2c-b5a7-3f6fc46a7993" containerName="authorino" Apr 21 00:13:25.836142 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.836123 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:25.838432 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.838415 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"maas-controller-dockercfg-jrw5v\"" Apr 21 00:13:25.842681 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.842655 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-d2r2x"] Apr 21 00:13:25.905920 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.905894 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dslld\" (UniqueName: \"kubernetes.io/projected/e0591e62-330c-48ac-aab4-1acd891c4abb-kube-api-access-dslld\") pod \"maas-controller-6d4c8f55f9-d2r2x\" (UID: \"e0591e62-330c-48ac-aab4-1acd891c4abb\") " pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:25.970398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.970374 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-6bb55c85cc-zwsx9"] Apr 21 00:13:25.973643 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.973629 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6bb55c85cc-zwsx9" Apr 21 00:13:25.981535 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:25.981513 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-6bb55c85cc-zwsx9"] Apr 21 00:13:26.006503 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.006482 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dslld\" (UniqueName: \"kubernetes.io/projected/e0591e62-330c-48ac-aab4-1acd891c4abb-kube-api-access-dslld\") pod \"maas-controller-6d4c8f55f9-d2r2x\" (UID: \"e0591e62-330c-48ac-aab4-1acd891c4abb\") " pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:26.015178 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.015157 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dslld\" (UniqueName: \"kubernetes.io/projected/e0591e62-330c-48ac-aab4-1acd891c4abb-kube-api-access-dslld\") pod \"maas-controller-6d4c8f55f9-d2r2x\" (UID: \"e0591e62-330c-48ac-aab4-1acd891c4abb\") " pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:26.047792 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.047768 2568 generic.go:358] "Generic (PLEG): container finished" podID="e705177c-f4f3-41ce-9b40-370df2b2a68d" containerID="4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b" exitCode=0 Apr 21 00:13:26.047935 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.047825 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-7498df8756-s4hm2" Apr 21 00:13:26.047935 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.047851 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-7498df8756-s4hm2" event={"ID":"e705177c-f4f3-41ce-9b40-370df2b2a68d","Type":"ContainerDied","Data":"4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b"} Apr 21 00:13:26.047935 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.047882 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-7498df8756-s4hm2" event={"ID":"e705177c-f4f3-41ce-9b40-370df2b2a68d","Type":"ContainerDied","Data":"e33fbaffe59ac41c1a095ee59c215c7926681c5b3831af12e925f9c86b430187"} Apr 21 00:13:26.047935 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.047897 2568 scope.go:117] "RemoveContainer" containerID="4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b" Apr 21 00:13:26.056298 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.056275 2568 scope.go:117] "RemoveContainer" containerID="4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b" Apr 21 00:13:26.056570 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:13:26.056552 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b\": container with ID starting with 4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b not found: ID does not exist" containerID="4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b" Apr 21 00:13:26.056626 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.056580 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b"} err="failed to get container status \"4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b\": rpc error: code = NotFound desc = could not find container \"4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b\": container with ID starting with 4bd2691a2038f74b1949301a01b886d65d2a94a919cdc49ef2d784f4b3304a9b not found: ID does not exist" Apr 21 00:13:26.071919 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.071898 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-7498df8756-s4hm2"] Apr 21 00:13:26.075280 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.075260 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-7498df8756-s4hm2"] Apr 21 00:13:26.084207 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.084159 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6bb55c85cc-zwsx9"] Apr 21 00:13:26.084341 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:13:26.084326 2568 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-8whqs], unattached volumes=[], failed to process volumes=[]: context canceled" pod="opendatahub/maas-controller-6bb55c85cc-zwsx9" podUID="ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d" Apr 21 00:13:26.107423 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.107404 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8whqs\" (UniqueName: \"kubernetes.io/projected/ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d-kube-api-access-8whqs\") pod \"maas-controller-6bb55c85cc-zwsx9\" (UID: \"ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d\") " pod="opendatahub/maas-controller-6bb55c85cc-zwsx9" Apr 21 00:13:26.147413 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.147375 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:26.207888 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.207862 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8whqs\" (UniqueName: \"kubernetes.io/projected/ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d-kube-api-access-8whqs\") pod \"maas-controller-6bb55c85cc-zwsx9\" (UID: \"ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d\") " pod="opendatahub/maas-controller-6bb55c85cc-zwsx9" Apr 21 00:13:26.215266 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.215236 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8whqs\" (UniqueName: \"kubernetes.io/projected/ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d-kube-api-access-8whqs\") pod \"maas-controller-6bb55c85cc-zwsx9\" (UID: \"ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d\") " pod="opendatahub/maas-controller-6bb55c85cc-zwsx9" Apr 21 00:13:26.266204 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.266181 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-d2r2x"] Apr 21 00:13:26.268185 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:13:26.268158 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0591e62_330c_48ac_aab4_1acd891c4abb.slice/crio-6ddec6d71d858f2191ce0b61aba37c12807614ac69b24fd4df9461d28e54ba44 WatchSource:0}: Error finding container 6ddec6d71d858f2191ce0b61aba37c12807614ac69b24fd4df9461d28e54ba44: Status 404 returned error can't find the container with id 6ddec6d71d858f2191ce0b61aba37c12807614ac69b24fd4df9461d28e54ba44 Apr 21 00:13:26.721402 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.721373 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3467b6c2-2349-4f2c-b5a7-3f6fc46a7993" path="/var/lib/kubelet/pods/3467b6c2-2349-4f2c-b5a7-3f6fc46a7993/volumes" Apr 21 00:13:26.721956 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:26.721915 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e705177c-f4f3-41ce-9b40-370df2b2a68d" path="/var/lib/kubelet/pods/e705177c-f4f3-41ce-9b40-370df2b2a68d/volumes" Apr 21 00:13:27.055079 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:27.054991 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" event={"ID":"e0591e62-330c-48ac-aab4-1acd891c4abb","Type":"ContainerStarted","Data":"6ddec6d71d858f2191ce0b61aba37c12807614ac69b24fd4df9461d28e54ba44"} Apr 21 00:13:27.056103 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:27.056072 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6bb55c85cc-zwsx9" Apr 21 00:13:27.062736 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:27.062693 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6bb55c85cc-zwsx9" Apr 21 00:13:27.115902 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:27.115865 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8whqs\" (UniqueName: \"kubernetes.io/projected/ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d-kube-api-access-8whqs\") pod \"ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d\" (UID: \"ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d\") " Apr 21 00:13:27.118398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:27.118372 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d-kube-api-access-8whqs" (OuterVolumeSpecName: "kube-api-access-8whqs") pod "ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d" (UID: "ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d"). InnerVolumeSpecName "kube-api-access-8whqs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:13:27.217211 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:27.217179 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8whqs\" (UniqueName: \"kubernetes.io/projected/ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d-kube-api-access-8whqs\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:13:28.059737 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:28.059706 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6bb55c85cc-zwsx9" Apr 21 00:13:28.091549 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:28.091516 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6bb55c85cc-zwsx9"] Apr 21 00:13:28.095737 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:28.095712 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-6bb55c85cc-zwsx9"] Apr 21 00:13:28.720396 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:28.720361 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d" path="/var/lib/kubelet/pods/ad7d46bb-0b95-4fbd-bc1a-4e0b6294040d/volumes" Apr 21 00:13:29.066866 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:29.066829 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" event={"ID":"e0591e62-330c-48ac-aab4-1acd891c4abb","Type":"ContainerStarted","Data":"d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0"} Apr 21 00:13:29.067265 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:29.066987 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:29.090388 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:29.090339 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" podStartSLOduration=1.8803588850000001 podStartE2EDuration="4.090326111s" podCreationTimestamp="2026-04-21 00:13:25 +0000 UTC" firstStartedPulling="2026-04-21 00:13:26.269446538 +0000 UTC m=+620.094425481" lastFinishedPulling="2026-04-21 00:13:28.47941375 +0000 UTC m=+622.304392707" observedRunningTime="2026-04-21 00:13:29.088908207 +0000 UTC m=+622.913887172" watchObservedRunningTime="2026-04-21 00:13:29.090326111 +0000 UTC m=+622.915305075" Apr 21 00:13:30.777357 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.777324 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-api-7d9df8cf7b-w4vhw"] Apr 21 00:13:30.780871 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.780855 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:30.783163 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.783129 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"maas-api-serving-cert\"" Apr 21 00:13:30.783290 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.783268 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"maas-parameters\"" Apr 21 00:13:30.783349 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.783277 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"maas-api-dockercfg-77wmh\"" Apr 21 00:13:30.790561 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.790541 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-api-7d9df8cf7b-w4vhw"] Apr 21 00:13:30.846361 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.846333 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxrc4\" (UniqueName: \"kubernetes.io/projected/1128d9c8-0dfe-4711-9336-35ea8e96099c-kube-api-access-qxrc4\") pod \"maas-api-7d9df8cf7b-w4vhw\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:30.846495 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.846430 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls\") pod \"maas-api-7d9df8cf7b-w4vhw\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:30.946890 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.946860 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qxrc4\" (UniqueName: \"kubernetes.io/projected/1128d9c8-0dfe-4711-9336-35ea8e96099c-kube-api-access-qxrc4\") pod \"maas-api-7d9df8cf7b-w4vhw\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:30.947017 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.946920 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls\") pod \"maas-api-7d9df8cf7b-w4vhw\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:30.947074 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:13:30.947011 2568 secret.go:189] Couldn't get secret opendatahub/maas-api-serving-cert: secret "maas-api-serving-cert" not found Apr 21 00:13:30.947074 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:13:30.947069 2568 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls podName:1128d9c8-0dfe-4711-9336-35ea8e96099c nodeName:}" failed. No retries permitted until 2026-04-21 00:13:31.44705451 +0000 UTC m=+625.272033452 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "maas-api-tls" (UniqueName: "kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls") pod "maas-api-7d9df8cf7b-w4vhw" (UID: "1128d9c8-0dfe-4711-9336-35ea8e96099c") : secret "maas-api-serving-cert" not found Apr 21 00:13:30.958968 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:30.958948 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxrc4\" (UniqueName: \"kubernetes.io/projected/1128d9c8-0dfe-4711-9336-35ea8e96099c-kube-api-access-qxrc4\") pod \"maas-api-7d9df8cf7b-w4vhw\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:31.452318 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:31.452263 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls\") pod \"maas-api-7d9df8cf7b-w4vhw\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:31.454689 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:31.454662 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls\") pod \"maas-api-7d9df8cf7b-w4vhw\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:31.693056 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:31.693029 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:31.815917 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:31.815894 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-api-7d9df8cf7b-w4vhw"] Apr 21 00:13:31.817662 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:13:31.817634 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1128d9c8_0dfe_4711_9336_35ea8e96099c.slice/crio-1bb4172758ebe6b2f75647e7602a9c0dfacb39fd074b9286ed7fe70258fdfeee WatchSource:0}: Error finding container 1bb4172758ebe6b2f75647e7602a9c0dfacb39fd074b9286ed7fe70258fdfeee: Status 404 returned error can't find the container with id 1bb4172758ebe6b2f75647e7602a9c0dfacb39fd074b9286ed7fe70258fdfeee Apr 21 00:13:32.080920 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:32.080877 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" event={"ID":"1128d9c8-0dfe-4711-9336-35ea8e96099c","Type":"ContainerStarted","Data":"1bb4172758ebe6b2f75647e7602a9c0dfacb39fd074b9286ed7fe70258fdfeee"} Apr 21 00:13:34.089804 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:34.089723 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" event={"ID":"1128d9c8-0dfe-4711-9336-35ea8e96099c","Type":"ContainerStarted","Data":"d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9"} Apr 21 00:13:34.090132 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:34.089885 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:34.106424 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:34.106370 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" podStartSLOduration=2.116275159 podStartE2EDuration="4.106357307s" podCreationTimestamp="2026-04-21 00:13:30 +0000 UTC" firstStartedPulling="2026-04-21 00:13:31.819061197 +0000 UTC m=+625.644040140" lastFinishedPulling="2026-04-21 00:13:33.809143345 +0000 UTC m=+627.634122288" observedRunningTime="2026-04-21 00:13:34.103664974 +0000 UTC m=+627.928643938" watchObservedRunningTime="2026-04-21 00:13:34.106357307 +0000 UTC m=+627.931336271" Apr 21 00:13:40.075482 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:40.075453 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:40.098524 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:40.098496 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:13:40.230074 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:40.230042 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-d2r2x"] Apr 21 00:13:40.230285 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:40.230243 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" podUID="e0591e62-330c-48ac-aab4-1acd891c4abb" containerName="manager" containerID="cri-o://d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0" gracePeriod=10 Apr 21 00:13:40.476931 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:40.476909 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:40.652713 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:40.652629 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dslld\" (UniqueName: \"kubernetes.io/projected/e0591e62-330c-48ac-aab4-1acd891c4abb-kube-api-access-dslld\") pod \"e0591e62-330c-48ac-aab4-1acd891c4abb\" (UID: \"e0591e62-330c-48ac-aab4-1acd891c4abb\") " Apr 21 00:13:40.654732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:40.654707 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0591e62-330c-48ac-aab4-1acd891c4abb-kube-api-access-dslld" (OuterVolumeSpecName: "kube-api-access-dslld") pod "e0591e62-330c-48ac-aab4-1acd891c4abb" (UID: "e0591e62-330c-48ac-aab4-1acd891c4abb"). InnerVolumeSpecName "kube-api-access-dslld". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:13:40.753872 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:40.753849 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dslld\" (UniqueName: \"kubernetes.io/projected/e0591e62-330c-48ac-aab4-1acd891c4abb-kube-api-access-dslld\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:13:41.116852 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.116816 2568 generic.go:358] "Generic (PLEG): container finished" podID="e0591e62-330c-48ac-aab4-1acd891c4abb" containerID="d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0" exitCode=0 Apr 21 00:13:41.117285 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.116871 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" event={"ID":"e0591e62-330c-48ac-aab4-1acd891c4abb","Type":"ContainerDied","Data":"d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0"} Apr 21 00:13:41.117285 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.116877 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" Apr 21 00:13:41.117285 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.116897 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-d2r2x" event={"ID":"e0591e62-330c-48ac-aab4-1acd891c4abb","Type":"ContainerDied","Data":"6ddec6d71d858f2191ce0b61aba37c12807614ac69b24fd4df9461d28e54ba44"} Apr 21 00:13:41.117285 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.116912 2568 scope.go:117] "RemoveContainer" containerID="d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0" Apr 21 00:13:41.125755 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.125737 2568 scope.go:117] "RemoveContainer" containerID="d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0" Apr 21 00:13:41.126004 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:13:41.125984 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0\": container with ID starting with d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0 not found: ID does not exist" containerID="d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0" Apr 21 00:13:41.126059 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.126016 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0"} err="failed to get container status \"d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0\": rpc error: code = NotFound desc = could not find container \"d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0\": container with ID starting with d102f401f556014ab9d795816ab603528ec36512020a8c732289a4447eb7a5e0 not found: ID does not exist" Apr 21 00:13:41.134092 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.134070 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-d2r2x"] Apr 21 00:13:41.136169 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:41.136152 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-d2r2x"] Apr 21 00:13:42.719535 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:42.719505 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0591e62-330c-48ac-aab4-1acd891c4abb" path="/var/lib/kubelet/pods/e0591e62-330c-48ac-aab4-1acd891c4abb/volumes" Apr 21 00:13:58.868265 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:58.868229 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:13:58.868816 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:13:58.868457 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="keycloak-system/maas-keycloak-0" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" containerID="cri-o://4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa" gracePeriod=30 Apr 21 00:14:00.509520 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:00.509488 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:00.595595 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:00.595529 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmsn6\" (UniqueName: \"kubernetes.io/projected/769b108a-869c-4f4d-a678-70f55dfab5eb-kube-api-access-tmsn6\") pod \"769b108a-869c-4f4d-a678-70f55dfab5eb\" (UID: \"769b108a-869c-4f4d-a678-70f55dfab5eb\") " Apr 21 00:14:00.597556 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:00.597532 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/769b108a-869c-4f4d-a678-70f55dfab5eb-kube-api-access-tmsn6" (OuterVolumeSpecName: "kube-api-access-tmsn6") pod "769b108a-869c-4f4d-a678-70f55dfab5eb" (UID: "769b108a-869c-4f4d-a678-70f55dfab5eb"). InnerVolumeSpecName "kube-api-access-tmsn6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:14:00.696263 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:00.696240 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tmsn6\" (UniqueName: \"kubernetes.io/projected/769b108a-869c-4f4d-a678-70f55dfab5eb-kube-api-access-tmsn6\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:14:01.191958 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.191924 2568 generic.go:358] "Generic (PLEG): container finished" podID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerID="4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa" exitCode=143 Apr 21 00:14:01.192122 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.191971 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="keycloak-system/maas-keycloak-0" event={"ID":"769b108a-869c-4f4d-a678-70f55dfab5eb","Type":"ContainerDied","Data":"4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa"} Apr 21 00:14:01.192122 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.191977 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.192122 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.191995 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="keycloak-system/maas-keycloak-0" event={"ID":"769b108a-869c-4f4d-a678-70f55dfab5eb","Type":"ContainerDied","Data":"eabe35b0ec9e6ce7c5b750999489293379c65e67f88fc35362ec5f17e2743d75"} Apr 21 00:14:01.192122 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.192009 2568 scope.go:117] "RemoveContainer" containerID="4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa" Apr 21 00:14:01.201668 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.201648 2568 scope.go:117] "RemoveContainer" containerID="4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa" Apr 21 00:14:01.201909 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:14:01.201890 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa\": container with ID starting with 4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa not found: ID does not exist" containerID="4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa" Apr 21 00:14:01.201965 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.201918 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa"} err="failed to get container status \"4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa\": rpc error: code = NotFound desc = could not find container \"4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa\": container with ID starting with 4b9713df2f0125df2014fd048378e9c51376cb5b0a6d908c3df1c8662e3afdaa not found: ID does not exist" Apr 21 00:14:01.209279 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.209251 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:14:01.212186 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.212169 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:14:01.231820 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.231789 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:14:01.232143 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.232129 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e0591e62-330c-48ac-aab4-1acd891c4abb" containerName="manager" Apr 21 00:14:01.232143 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.232144 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0591e62-330c-48ac-aab4-1acd891c4abb" containerName="manager" Apr 21 00:14:01.232229 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.232154 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" Apr 21 00:14:01.232229 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.232160 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" Apr 21 00:14:01.232229 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.232219 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" containerName="keycloak" Apr 21 00:14:01.232229 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.232229 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="e0591e62-330c-48ac-aab4-1acd891c4abb" containerName="manager" Apr 21 00:14:01.236718 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.236705 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.239472 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.239428 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"keycloak-system\"/\"default-dockercfg-sndxl\"" Apr 21 00:14:01.239594 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.239546 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"keycloak-system\"/\"keycloak-test-realms\"" Apr 21 00:14:01.239687 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.239548 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"keycloak-system\"/\"maas-keycloak-initial-admin\"" Apr 21 00:14:01.239748 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.239689 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"keycloak-system\"/\"openshift-service-ca.crt\"" Apr 21 00:14:01.240380 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.240359 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"keycloak-system\"/\"kube-root-ca.crt\"" Apr 21 00:14:01.242594 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.242572 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:14:01.300164 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.300127 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-realms\" (UniqueName: \"kubernetes.io/configmap/ab489c1f-737b-4a59-8194-df18631a6b8d-test-realms\") pod \"maas-keycloak-0\" (UID: \"ab489c1f-737b-4a59-8194-df18631a6b8d\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.300294 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.300180 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2kkh\" (UniqueName: \"kubernetes.io/projected/ab489c1f-737b-4a59-8194-df18631a6b8d-kube-api-access-h2kkh\") pod \"maas-keycloak-0\" (UID: \"ab489c1f-737b-4a59-8194-df18631a6b8d\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.400920 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.400885 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-h2kkh\" (UniqueName: \"kubernetes.io/projected/ab489c1f-737b-4a59-8194-df18631a6b8d-kube-api-access-h2kkh\") pod \"maas-keycloak-0\" (UID: \"ab489c1f-737b-4a59-8194-df18631a6b8d\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.401048 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.400970 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"test-realms\" (UniqueName: \"kubernetes.io/configmap/ab489c1f-737b-4a59-8194-df18631a6b8d-test-realms\") pod \"maas-keycloak-0\" (UID: \"ab489c1f-737b-4a59-8194-df18631a6b8d\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.401553 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.401537 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"test-realms\" (UniqueName: \"kubernetes.io/configmap/ab489c1f-737b-4a59-8194-df18631a6b8d-test-realms\") pod \"maas-keycloak-0\" (UID: \"ab489c1f-737b-4a59-8194-df18631a6b8d\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.408838 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.408816 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2kkh\" (UniqueName: \"kubernetes.io/projected/ab489c1f-737b-4a59-8194-df18631a6b8d-kube-api-access-h2kkh\") pod \"maas-keycloak-0\" (UID: \"ab489c1f-737b-4a59-8194-df18631a6b8d\") " pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.550730 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.550698 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:01.675967 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.675939 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["keycloak-system/maas-keycloak-0"] Apr 21 00:14:01.677240 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:14:01.677209 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab489c1f_737b_4a59_8194_df18631a6b8d.slice/crio-b1cdf9529ea25ab4779fa9884f50c99ed2998d9cb7a89d7930ea644307c35914 WatchSource:0}: Error finding container b1cdf9529ea25ab4779fa9884f50c99ed2998d9cb7a89d7930ea644307c35914: Status 404 returned error can't find the container with id b1cdf9529ea25ab4779fa9884f50c99ed2998d9cb7a89d7930ea644307c35914 Apr 21 00:14:01.899104 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.899013 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-api-6f98dc884d-4fkhr"] Apr 21 00:14:01.904982 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.904960 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:01.911010 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:01.910989 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-api-6f98dc884d-4fkhr"] Apr 21 00:14:02.004838 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.004807 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlsmh\" (UniqueName: \"kubernetes.io/projected/5ac6af9d-c5be-4343-886c-11c2e5e56222-kube-api-access-hlsmh\") pod \"maas-api-6f98dc884d-4fkhr\" (UID: \"5ac6af9d-c5be-4343-886c-11c2e5e56222\") " pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:02.004990 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.004854 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/5ac6af9d-c5be-4343-886c-11c2e5e56222-maas-api-tls\") pod \"maas-api-6f98dc884d-4fkhr\" (UID: \"5ac6af9d-c5be-4343-886c-11c2e5e56222\") " pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:02.105417 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.105381 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hlsmh\" (UniqueName: \"kubernetes.io/projected/5ac6af9d-c5be-4343-886c-11c2e5e56222-kube-api-access-hlsmh\") pod \"maas-api-6f98dc884d-4fkhr\" (UID: \"5ac6af9d-c5be-4343-886c-11c2e5e56222\") " pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:02.105632 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.105469 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/5ac6af9d-c5be-4343-886c-11c2e5e56222-maas-api-tls\") pod \"maas-api-6f98dc884d-4fkhr\" (UID: \"5ac6af9d-c5be-4343-886c-11c2e5e56222\") " pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:02.108221 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.108196 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/5ac6af9d-c5be-4343-886c-11c2e5e56222-maas-api-tls\") pod \"maas-api-6f98dc884d-4fkhr\" (UID: \"5ac6af9d-c5be-4343-886c-11c2e5e56222\") " pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:02.113137 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.113113 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlsmh\" (UniqueName: \"kubernetes.io/projected/5ac6af9d-c5be-4343-886c-11c2e5e56222-kube-api-access-hlsmh\") pod \"maas-api-6f98dc884d-4fkhr\" (UID: \"5ac6af9d-c5be-4343-886c-11c2e5e56222\") " pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:02.197666 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.197640 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="keycloak-system/maas-keycloak-0" event={"ID":"ab489c1f-737b-4a59-8194-df18631a6b8d","Type":"ContainerStarted","Data":"82c0a74016e4d0c0109697e2dc2c3919669ec2c73ee5911a728818668e2b3bf2"} Apr 21 00:14:02.197766 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.197677 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="keycloak-system/maas-keycloak-0" event={"ID":"ab489c1f-737b-4a59-8194-df18631a6b8d","Type":"ContainerStarted","Data":"b1cdf9529ea25ab4779fa9884f50c99ed2998d9cb7a89d7930ea644307c35914"} Apr 21 00:14:02.215085 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.215031 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="keycloak-system/maas-keycloak-0" podStartSLOduration=0.770588456 podStartE2EDuration="1.21501584s" podCreationTimestamp="2026-04-21 00:14:01 +0000 UTC" firstStartedPulling="2026-04-21 00:14:01.678529627 +0000 UTC m=+655.503508574" lastFinishedPulling="2026-04-21 00:14:02.122957004 +0000 UTC m=+655.947935958" observedRunningTime="2026-04-21 00:14:02.213551317 +0000 UTC m=+656.038530285" watchObservedRunningTime="2026-04-21 00:14:02.21501584 +0000 UTC m=+656.039994806" Apr 21 00:14:02.217170 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.217150 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:02.348199 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.348172 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-api-6f98dc884d-4fkhr"] Apr 21 00:14:02.350036 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:14:02.350001 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ac6af9d_c5be_4343_886c_11c2e5e56222.slice/crio-9bf72ed952aa934134b92a4266a5124da7309e43ed766858341037ae2e18a89d WatchSource:0}: Error finding container 9bf72ed952aa934134b92a4266a5124da7309e43ed766858341037ae2e18a89d: Status 404 returned error can't find the container with id 9bf72ed952aa934134b92a4266a5124da7309e43ed766858341037ae2e18a89d Apr 21 00:14:02.551435 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.551398 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:02.553022 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.552995 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:02.722822 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:02.722784 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="769b108a-869c-4f4d-a678-70f55dfab5eb" path="/var/lib/kubelet/pods/769b108a-869c-4f4d-a678-70f55dfab5eb/volumes" Apr 21 00:14:03.205164 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:03.205114 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-api-6f98dc884d-4fkhr" event={"ID":"5ac6af9d-c5be-4343-886c-11c2e5e56222","Type":"ContainerStarted","Data":"9bf72ed952aa934134b92a4266a5124da7309e43ed766858341037ae2e18a89d"} Apr 21 00:14:03.552064 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:03.552015 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:04.212171 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:04.212009 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-api-6f98dc884d-4fkhr" event={"ID":"5ac6af9d-c5be-4343-886c-11c2e5e56222","Type":"ContainerStarted","Data":"4abffc922428920e523a835abb918f755587337a332181b17f7b98eb94c78f0c"} Apr 21 00:14:04.212171 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:04.212123 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:04.228367 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:04.228306 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-api-6f98dc884d-4fkhr" podStartSLOduration=1.6822547860000001 podStartE2EDuration="3.228288661s" podCreationTimestamp="2026-04-21 00:14:01 +0000 UTC" firstStartedPulling="2026-04-21 00:14:02.351387196 +0000 UTC m=+656.176366140" lastFinishedPulling="2026-04-21 00:14:03.897421057 +0000 UTC m=+657.722400015" observedRunningTime="2026-04-21 00:14:04.227626735 +0000 UTC m=+658.052605703" watchObservedRunningTime="2026-04-21 00:14:04.228288661 +0000 UTC m=+658.053267626" Apr 21 00:14:04.551232 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:04.551183 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:05.551235 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:05.551184 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:06.551587 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:06.551544 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:07.551989 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:07.551947 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:08.551682 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:08.551627 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:09.552246 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:09.552196 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:10.222968 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.222941 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-api-6f98dc884d-4fkhr" Apr 21 00:14:10.264308 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.264274 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-api-7d9df8cf7b-w4vhw"] Apr 21 00:14:10.264812 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.264781 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" podUID="1128d9c8-0dfe-4711-9336-35ea8e96099c" containerName="maas-api" containerID="cri-o://d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9" gracePeriod=30 Apr 21 00:14:10.540763 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.540731 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:14:10.552083 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.552052 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:10.591342 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.591315 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls\") pod \"1128d9c8-0dfe-4711-9336-35ea8e96099c\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " Apr 21 00:14:10.591768 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.591361 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxrc4\" (UniqueName: \"kubernetes.io/projected/1128d9c8-0dfe-4711-9336-35ea8e96099c-kube-api-access-qxrc4\") pod \"1128d9c8-0dfe-4711-9336-35ea8e96099c\" (UID: \"1128d9c8-0dfe-4711-9336-35ea8e96099c\") " Apr 21 00:14:10.593424 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.593389 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls" (OuterVolumeSpecName: "maas-api-tls") pod "1128d9c8-0dfe-4711-9336-35ea8e96099c" (UID: "1128d9c8-0dfe-4711-9336-35ea8e96099c"). InnerVolumeSpecName "maas-api-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:14:10.593529 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.593464 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1128d9c8-0dfe-4711-9336-35ea8e96099c-kube-api-access-qxrc4" (OuterVolumeSpecName: "kube-api-access-qxrc4") pod "1128d9c8-0dfe-4711-9336-35ea8e96099c" (UID: "1128d9c8-0dfe-4711-9336-35ea8e96099c"). InnerVolumeSpecName "kube-api-access-qxrc4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:14:10.692423 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.692359 2568 reconciler_common.go:299] "Volume detached for volume \"maas-api-tls\" (UniqueName: \"kubernetes.io/secret/1128d9c8-0dfe-4711-9336-35ea8e96099c-maas-api-tls\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:14:10.692423 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:10.692397 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qxrc4\" (UniqueName: \"kubernetes.io/projected/1128d9c8-0dfe-4711-9336-35ea8e96099c-kube-api-access-qxrc4\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:14:11.249312 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.249273 2568 generic.go:358] "Generic (PLEG): container finished" podID="1128d9c8-0dfe-4711-9336-35ea8e96099c" containerID="d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9" exitCode=0 Apr 21 00:14:11.249513 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.249339 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" event={"ID":"1128d9c8-0dfe-4711-9336-35ea8e96099c","Type":"ContainerDied","Data":"d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9"} Apr 21 00:14:11.249513 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.249356 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" Apr 21 00:14:11.249513 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.249375 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-api-7d9df8cf7b-w4vhw" event={"ID":"1128d9c8-0dfe-4711-9336-35ea8e96099c","Type":"ContainerDied","Data":"1bb4172758ebe6b2f75647e7602a9c0dfacb39fd074b9286ed7fe70258fdfeee"} Apr 21 00:14:11.249513 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.249396 2568 scope.go:117] "RemoveContainer" containerID="d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9" Apr 21 00:14:11.259739 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.259647 2568 scope.go:117] "RemoveContainer" containerID="d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9" Apr 21 00:14:11.260080 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:14:11.260050 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9\": container with ID starting with d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9 not found: ID does not exist" containerID="d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9" Apr 21 00:14:11.260156 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.260097 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9"} err="failed to get container status \"d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9\": rpc error: code = NotFound desc = could not find container \"d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9\": container with ID starting with d6cb532d3983e7873b48f88b012acef54c57678fd6d5467d2c167e44bb08c2a9 not found: ID does not exist" Apr 21 00:14:11.266732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.266703 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-api-7d9df8cf7b-w4vhw"] Apr 21 00:14:11.272474 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.272446 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-api-7d9df8cf7b-w4vhw"] Apr 21 00:14:11.551535 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.551447 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:11.551742 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:11.551712 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:12.551871 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:12.551814 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:12.723066 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:12.723020 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1128d9c8-0dfe-4711-9336-35ea8e96099c" path="/var/lib/kubelet/pods/1128d9c8-0dfe-4711-9336-35ea8e96099c/volumes" Apr 21 00:14:13.552054 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:13.552006 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:14.551876 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:14.551834 2568 prober.go:120] "Probe failed" probeType="Startup" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="Get \"http://10.133.0.51:9000/health/started\": dial tcp 10.133.0.51:9000: connect: connection refused" Apr 21 00:14:15.677293 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:15.677253 2568 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:15.696663 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:15.696596 2568 prober.go:120] "Probe failed" probeType="Readiness" pod="keycloak-system/maas-keycloak-0" podUID="ab489c1f-737b-4a59-8194-df18631a6b8d" containerName="keycloak" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 21 00:14:25.683424 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:25.683393 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="keycloak-system/maas-keycloak-0" Apr 21 00:14:36.395855 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.395821 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-596f7d8dbd-cwkfd"] Apr 21 00:14:36.396307 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.396160 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1128d9c8-0dfe-4711-9336-35ea8e96099c" containerName="maas-api" Apr 21 00:14:36.396307 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.396171 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="1128d9c8-0dfe-4711-9336-35ea8e96099c" containerName="maas-api" Apr 21 00:14:36.396307 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.396224 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="1128d9c8-0dfe-4711-9336-35ea8e96099c" containerName="maas-api" Apr 21 00:14:36.399270 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.399250 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.401541 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.401516 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"authorino-oidc-ca\"" Apr 21 00:14:36.404831 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.404812 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-596f7d8dbd-cwkfd"] Apr 21 00:14:36.523888 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.523862 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5nfw\" (UniqueName: \"kubernetes.io/projected/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-kube-api-access-x5nfw\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.524015 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.523908 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-tls-cert\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.524015 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.523967 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-oidc-ca\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.624884 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.624848 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x5nfw\" (UniqueName: \"kubernetes.io/projected/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-kube-api-access-x5nfw\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.625022 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.624903 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-tls-cert\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.625022 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.624952 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-oidc-ca\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.625535 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.625518 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-oidc-ca\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.627415 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.627398 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-tls-cert\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.631966 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.631949 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5nfw\" (UniqueName: \"kubernetes.io/projected/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-kube-api-access-x5nfw\") pod \"authorino-596f7d8dbd-cwkfd\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.709480 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.709430 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:14:36.839773 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:36.839744 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-596f7d8dbd-cwkfd"] Apr 21 00:14:37.364377 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.364339 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" event={"ID":"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e","Type":"ContainerStarted","Data":"61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c"} Apr 21 00:14:37.364520 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.364381 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" event={"ID":"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e","Type":"ContainerStarted","Data":"423eb5f82384ab23674b5ccefaf0cfaa8ff404fd595fc6c0964f972ffde73d40"} Apr 21 00:14:37.379821 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.379763 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" podStartSLOduration=0.97626052 podStartE2EDuration="1.379747861s" podCreationTimestamp="2026-04-21 00:14:36 +0000 UTC" firstStartedPulling="2026-04-21 00:14:36.842674834 +0000 UTC m=+690.667653776" lastFinishedPulling="2026-04-21 00:14:37.246162174 +0000 UTC m=+691.071141117" observedRunningTime="2026-04-21 00:14:37.378540758 +0000 UTC m=+691.203519724" watchObservedRunningTime="2026-04-21 00:14:37.379747861 +0000 UTC m=+691.204726826" Apr 21 00:14:37.402837 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.402810 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-5f85766d6d-6qc8b"] Apr 21 00:14:37.403194 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.402998 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" podUID="faaaf689-f76b-44cf-b052-656e71b1b42e" containerName="authorino" containerID="cri-o://cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123" gracePeriod=30 Apr 21 00:14:37.696481 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.696459 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:14:37.835965 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.835875 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcctd\" (UniqueName: \"kubernetes.io/projected/faaaf689-f76b-44cf-b052-656e71b1b42e-kube-api-access-qcctd\") pod \"faaaf689-f76b-44cf-b052-656e71b1b42e\" (UID: \"faaaf689-f76b-44cf-b052-656e71b1b42e\") " Apr 21 00:14:37.835965 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.835916 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/faaaf689-f76b-44cf-b052-656e71b1b42e-tls-cert\") pod \"faaaf689-f76b-44cf-b052-656e71b1b42e\" (UID: \"faaaf689-f76b-44cf-b052-656e71b1b42e\") " Apr 21 00:14:37.838092 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.838066 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faaaf689-f76b-44cf-b052-656e71b1b42e-kube-api-access-qcctd" (OuterVolumeSpecName: "kube-api-access-qcctd") pod "faaaf689-f76b-44cf-b052-656e71b1b42e" (UID: "faaaf689-f76b-44cf-b052-656e71b1b42e"). InnerVolumeSpecName "kube-api-access-qcctd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:14:37.845426 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.845403 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faaaf689-f76b-44cf-b052-656e71b1b42e-tls-cert" (OuterVolumeSpecName: "tls-cert") pod "faaaf689-f76b-44cf-b052-656e71b1b42e" (UID: "faaaf689-f76b-44cf-b052-656e71b1b42e"). InnerVolumeSpecName "tls-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:14:37.936974 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.936944 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qcctd\" (UniqueName: \"kubernetes.io/projected/faaaf689-f76b-44cf-b052-656e71b1b42e-kube-api-access-qcctd\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:14:37.936974 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:37.936972 2568 reconciler_common.go:299] "Volume detached for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/faaaf689-f76b-44cf-b052-656e71b1b42e-tls-cert\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:14:38.369012 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.368977 2568 generic.go:358] "Generic (PLEG): container finished" podID="faaaf689-f76b-44cf-b052-656e71b1b42e" containerID="cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123" exitCode=0 Apr 21 00:14:38.369156 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.369030 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" Apr 21 00:14:38.369156 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.369061 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" event={"ID":"faaaf689-f76b-44cf-b052-656e71b1b42e","Type":"ContainerDied","Data":"cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123"} Apr 21 00:14:38.369156 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.369101 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-5f85766d6d-6qc8b" event={"ID":"faaaf689-f76b-44cf-b052-656e71b1b42e","Type":"ContainerDied","Data":"f53c9ef8d08ae0db5e1e08815d3f59e4ca67ff42a99e168239f51f408f6aa695"} Apr 21 00:14:38.369156 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.369118 2568 scope.go:117] "RemoveContainer" containerID="cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123" Apr 21 00:14:38.377782 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.377766 2568 scope.go:117] "RemoveContainer" containerID="cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123" Apr 21 00:14:38.378050 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:14:38.378032 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123\": container with ID starting with cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123 not found: ID does not exist" containerID="cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123" Apr 21 00:14:38.378093 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.378058 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123"} err="failed to get container status \"cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123\": rpc error: code = NotFound desc = could not find container \"cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123\": container with ID starting with cfff6e9b07a5ba5b63aa2d7d294ae0d27c0ba737c3ea88ca0e49f9f2ccff3123 not found: ID does not exist" Apr 21 00:14:38.389881 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.389858 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-5f85766d6d-6qc8b"] Apr 21 00:14:38.392186 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.392166 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-5f85766d6d-6qc8b"] Apr 21 00:14:38.720120 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:38.720090 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faaaf689-f76b-44cf-b052-656e71b1b42e" path="/var/lib/kubelet/pods/faaaf689-f76b-44cf-b052-656e71b1b42e/volumes" Apr 21 00:14:47.594221 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.594189 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5"] Apr 21 00:14:47.594730 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.594711 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="faaaf689-f76b-44cf-b052-656e71b1b42e" containerName="authorino" Apr 21 00:14:47.594782 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.594734 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="faaaf689-f76b-44cf-b052-656e71b1b42e" containerName="authorino" Apr 21 00:14:47.594861 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.594849 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="faaaf689-f76b-44cf-b052-656e71b1b42e" containerName="authorino" Apr 21 00:14:47.599474 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.599446 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.602025 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.602003 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"llm\"/\"kube-root-ca.crt\"" Apr 21 00:14:47.602025 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.602016 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"llm\"/\"default-dockercfg-lw7xf\"" Apr 21 00:14:47.602203 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.602053 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"llm\"/\"e2e-unab60ef4d3a239b5143b412cab04acac3-kserve-self-signed-certs\"" Apr 21 00:14:47.603746 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.603722 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"llm\"/\"openshift-service-ca.crt\"" Apr 21 00:14:47.608143 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.608105 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5"] Apr 21 00:14:47.617645 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.617600 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-dshm\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.617736 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.617670 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-model-cache\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.617736 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.617690 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-tls-certs\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.617736 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.617720 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-kserve-provision-location\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.617855 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.617806 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-home\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.617855 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.617830 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n97j\" (UniqueName: \"kubernetes.io/projected/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-kube-api-access-2n97j\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.718512 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.718475 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-dshm\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.718732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.718522 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-model-cache\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.718732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.718546 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-tls-certs\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.718732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.718576 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-kserve-provision-location\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.718732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.718627 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-home\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.718732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.718653 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2n97j\" (UniqueName: \"kubernetes.io/projected/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-kube-api-access-2n97j\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.719018 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.718978 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-model-cache\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.719018 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.718993 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-kserve-provision-location\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.719121 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.719038 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-home\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.720736 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.720719 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-dshm\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.720979 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.720963 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-tls-certs\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.725764 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.725739 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n97j\" (UniqueName: \"kubernetes.io/projected/58b39ea6-8e4f-4f11-a2a1-479b6049e95b-kube-api-access-2n97j\") pod \"e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5\" (UID: \"58b39ea6-8e4f-4f11-a2a1-479b6049e95b\") " pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:47.911028 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:47.910943 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:14:48.038293 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:48.036557 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5"] Apr 21 00:14:48.411236 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:48.411198 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" event={"ID":"58b39ea6-8e4f-4f11-a2a1-479b6049e95b","Type":"ContainerStarted","Data":"7c3b7ad98505a66b4eb27a17557678752c2f7193dd988c8144c5943f2b2f696e"} Apr 21 00:14:53.886714 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.886681 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5"] Apr 21 00:14:53.924952 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.924913 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5"] Apr 21 00:14:53.925103 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.925060 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:53.928273 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.928248 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"llm\"/\"premium-simulated-simulated-premium-kserve-self-signed-certs\"" Apr 21 00:14:53.971432 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.971402 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v4qr\" (UniqueName: \"kubernetes.io/projected/73582b91-5337-46b9-9625-c1f9bec6410b-kube-api-access-6v4qr\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:53.971588 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.971458 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-dshm\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:53.971588 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.971530 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/73582b91-5337-46b9-9625-c1f9bec6410b-tls-certs\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:53.971588 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.971571 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-home\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:53.971771 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.971593 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-model-cache\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:53.971771 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:53.971633 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-kserve-provision-location\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072222 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072191 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6v4qr\" (UniqueName: \"kubernetes.io/projected/73582b91-5337-46b9-9625-c1f9bec6410b-kube-api-access-6v4qr\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072391 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072251 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-dshm\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072391 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072296 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/73582b91-5337-46b9-9625-c1f9bec6410b-tls-certs\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072391 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072336 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-home\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072391 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072359 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-model-cache\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072391 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072381 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-kserve-provision-location\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072848 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072821 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-kserve-provision-location\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072952 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072879 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-home\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.072952 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.072913 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-model-cache\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.075067 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.075042 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/73582b91-5337-46b9-9625-c1f9bec6410b-dshm\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.075450 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.075422 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/73582b91-5337-46b9-9625-c1f9bec6410b-tls-certs\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.079999 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.079963 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v4qr\" (UniqueName: \"kubernetes.io/projected/73582b91-5337-46b9-9625-c1f9bec6410b-kube-api-access-6v4qr\") pod \"premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5\" (UID: \"73582b91-5337-46b9-9625-c1f9bec6410b\") " pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.236236 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.236202 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:14:54.715371 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:14:54.715328 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73582b91_5337_46b9_9625_c1f9bec6410b.slice/crio-c97b1818d8781c2693a1b14aad020d1d0267213bc3d1f2ccab47e4fe24b15aa4 WatchSource:0}: Error finding container c97b1818d8781c2693a1b14aad020d1d0267213bc3d1f2ccab47e4fe24b15aa4: Status 404 returned error can't find the container with id c97b1818d8781c2693a1b14aad020d1d0267213bc3d1f2ccab47e4fe24b15aa4 Apr 21 00:14:54.722387 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:54.722361 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5"] Apr 21 00:14:55.445559 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:55.445517 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" event={"ID":"73582b91-5337-46b9-9625-c1f9bec6410b","Type":"ContainerStarted","Data":"b3b691f36cc4dd6b6b5a0decd1bc8d08a28cc073758297ca745aa89d2aceda27"} Apr 21 00:14:55.445559 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:55.445563 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" event={"ID":"73582b91-5337-46b9-9625-c1f9bec6410b","Type":"ContainerStarted","Data":"c97b1818d8781c2693a1b14aad020d1d0267213bc3d1f2ccab47e4fe24b15aa4"} Apr 21 00:14:55.447271 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:14:55.447238 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" event={"ID":"58b39ea6-8e4f-4f11-a2a1-479b6049e95b","Type":"ContainerStarted","Data":"ad192ab53ed278449169c529087f761050954ac708d7b74161ba29247148c2e1"} Apr 21 00:15:00.472620 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:00.472569 2568 generic.go:358] "Generic (PLEG): container finished" podID="58b39ea6-8e4f-4f11-a2a1-479b6049e95b" containerID="ad192ab53ed278449169c529087f761050954ac708d7b74161ba29247148c2e1" exitCode=0 Apr 21 00:15:00.472983 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:00.472647 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" event={"ID":"58b39ea6-8e4f-4f11-a2a1-479b6049e95b","Type":"ContainerDied","Data":"ad192ab53ed278449169c529087f761050954ac708d7b74161ba29247148c2e1"} Apr 21 00:15:02.485577 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:02.485535 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" event={"ID":"58b39ea6-8e4f-4f11-a2a1-479b6049e95b","Type":"ContainerStarted","Data":"b91e740d38214f7186225bc52a6d056f4483b6b0c86dd5003ec1f2c8872d32ba"} Apr 21 00:15:02.485963 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:02.485785 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:15:02.502808 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:02.502746 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" podStartSLOduration=1.951084019 podStartE2EDuration="15.502730508s" podCreationTimestamp="2026-04-21 00:14:47 +0000 UTC" firstStartedPulling="2026-04-21 00:14:48.040029933 +0000 UTC m=+701.865008876" lastFinishedPulling="2026-04-21 00:15:01.591676418 +0000 UTC m=+715.416655365" observedRunningTime="2026-04-21 00:15:02.501038157 +0000 UTC m=+716.326017121" watchObservedRunningTime="2026-04-21 00:15:02.502730508 +0000 UTC m=+716.327709472" Apr 21 00:15:03.491080 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:03.491052 2568 generic.go:358] "Generic (PLEG): container finished" podID="73582b91-5337-46b9-9625-c1f9bec6410b" containerID="b3b691f36cc4dd6b6b5a0decd1bc8d08a28cc073758297ca745aa89d2aceda27" exitCode=0 Apr 21 00:15:03.491459 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:03.491126 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" event={"ID":"73582b91-5337-46b9-9625-c1f9bec6410b","Type":"ContainerDied","Data":"b3b691f36cc4dd6b6b5a0decd1bc8d08a28cc073758297ca745aa89d2aceda27"} Apr 21 00:15:04.497007 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:04.496972 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" event={"ID":"73582b91-5337-46b9-9625-c1f9bec6410b","Type":"ContainerStarted","Data":"a826c29dff6fbe5303990a178298db32eeb8d038fd9c95c0221c0dc0d966ba06"} Apr 21 00:15:04.497368 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:04.497152 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:15:04.515808 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:04.515760 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" podStartSLOduration=11.325321136 podStartE2EDuration="11.515749035s" podCreationTimestamp="2026-04-21 00:14:53 +0000 UTC" firstStartedPulling="2026-04-21 00:15:03.491810778 +0000 UTC m=+717.316789721" lastFinishedPulling="2026-04-21 00:15:03.682238674 +0000 UTC m=+717.507217620" observedRunningTime="2026-04-21 00:15:04.513018555 +0000 UTC m=+718.337997521" watchObservedRunningTime="2026-04-21 00:15:04.515749035 +0000 UTC m=+718.340728004" Apr 21 00:15:13.504438 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:13.504404 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="llm/e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5" Apr 21 00:15:15.513284 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:15.513255 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="llm/premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5" Apr 21 00:15:25.583958 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.583911 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p"] Apr 21 00:15:25.616942 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.616913 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p"] Apr 21 00:15:25.617083 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.617030 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.619647 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.619600 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"llm\"/\"e2e-distinct-2-simulated-kserve-self-signed-certs\"" Apr 21 00:15:25.758243 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.758213 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/030cacec-c448-47ae-ae2d-87c2164c7b80-tls-certs\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.758386 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.758265 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-home\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.758386 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.758312 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-kserve-provision-location\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.758386 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.758343 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-model-cache\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.758386 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.758367 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz62b\" (UniqueName: \"kubernetes.io/projected/030cacec-c448-47ae-ae2d-87c2164c7b80-kube-api-access-rz62b\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.758527 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.758397 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-dshm\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860231 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860157 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-model-cache\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860231 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860201 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rz62b\" (UniqueName: \"kubernetes.io/projected/030cacec-c448-47ae-ae2d-87c2164c7b80-kube-api-access-rz62b\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860430 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860253 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-dshm\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860430 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860321 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/030cacec-c448-47ae-ae2d-87c2164c7b80-tls-certs\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860591 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860376 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-home\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860591 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860491 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-kserve-provision-location\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860591 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860556 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-model-cache\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860796 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860755 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-home\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.860853 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.860830 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-kserve-provision-location\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.863226 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.863202 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/030cacec-c448-47ae-ae2d-87c2164c7b80-dshm\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.863395 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.863375 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/030cacec-c448-47ae-ae2d-87c2164c7b80-tls-certs\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.868051 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.868026 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz62b\" (UniqueName: \"kubernetes.io/projected/030cacec-c448-47ae-ae2d-87c2164c7b80-kube-api-access-rz62b\") pod \"e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p\" (UID: \"030cacec-c448-47ae-ae2d-87c2164c7b80\") " pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:25.927404 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:25.927378 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:26.055291 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:26.055264 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p"] Apr 21 00:15:26.056753 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:15:26.056724 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod030cacec_c448_47ae_ae2d_87c2164c7b80.slice/crio-79a2a71aae409c503b1e8486d7314e1f5a4f44035adf8e1f1bcf75820b353b41 WatchSource:0}: Error finding container 79a2a71aae409c503b1e8486d7314e1f5a4f44035adf8e1f1bcf75820b353b41: Status 404 returned error can't find the container with id 79a2a71aae409c503b1e8486d7314e1f5a4f44035adf8e1f1bcf75820b353b41 Apr 21 00:15:26.592683 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:26.592648 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" event={"ID":"030cacec-c448-47ae-ae2d-87c2164c7b80","Type":"ContainerStarted","Data":"24239edc02c96f41245d237f49a3da854bab7d93bbd2bd9b27c5c1db00be7143"} Apr 21 00:15:26.593041 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:26.592689 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" event={"ID":"030cacec-c448-47ae-ae2d-87c2164c7b80","Type":"ContainerStarted","Data":"79a2a71aae409c503b1e8486d7314e1f5a4f44035adf8e1f1bcf75820b353b41"} Apr 21 00:15:27.289108 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.289079 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s"] Apr 21 00:15:27.292857 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.292838 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.295242 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.295224 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"llm\"/\"e2e-trlp-test-simulated-kserve-self-signed-certs\"" Apr 21 00:15:27.303473 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.303450 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s"] Apr 21 00:15:27.475142 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.475106 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-kserve-provision-location\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.475306 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.475243 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-tls-certs\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.475306 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.475285 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnpbf\" (UniqueName: \"kubernetes.io/projected/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-kube-api-access-qnpbf\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.475398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.475307 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-model-cache\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.475398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.475337 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-home\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.475398 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.475352 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-dshm\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576416 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576341 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-home\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576416 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576373 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-dshm\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576416 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576414 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-kserve-provision-location\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576706 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576458 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-tls-certs\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576706 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576477 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qnpbf\" (UniqueName: \"kubernetes.io/projected/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-kube-api-access-qnpbf\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576706 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576502 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-model-cache\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576865 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576801 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"home\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-home\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576865 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576834 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"model-cache\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-model-cache\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.576958 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.576874 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-kserve-provision-location\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.578619 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.578585 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dshm\" (UniqueName: \"kubernetes.io/empty-dir/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-dshm\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.578962 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.578946 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-tls-certs\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.583919 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.583893 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnpbf\" (UniqueName: \"kubernetes.io/projected/a8d7704c-dc4a-42db-9afb-5c9998f6d0a8-kube-api-access-qnpbf\") pod \"e2e-trlp-test-simulated-kserve-84db68679b-sxf4s\" (UID: \"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8\") " pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.603885 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.603864 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:27.734645 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:27.734619 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s"] Apr 21 00:15:27.735458 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:15:27.735435 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8d7704c_dc4a_42db_9afb_5c9998f6d0a8.slice/crio-b0158a41d9f80c6db2a0ef5294507a297c83f3d53db7a21bd01c2b6c53d54d0e WatchSource:0}: Error finding container b0158a41d9f80c6db2a0ef5294507a297c83f3d53db7a21bd01c2b6c53d54d0e: Status 404 returned error can't find the container with id b0158a41d9f80c6db2a0ef5294507a297c83f3d53db7a21bd01c2b6c53d54d0e Apr 21 00:15:28.604274 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:28.604236 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" event={"ID":"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8","Type":"ContainerStarted","Data":"189fdb9530467e84d748ee5d83272c15e731cc157546a3ff07993fa89843942a"} Apr 21 00:15:28.604655 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:28.604281 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" event={"ID":"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8","Type":"ContainerStarted","Data":"b0158a41d9f80c6db2a0ef5294507a297c83f3d53db7a21bd01c2b6c53d54d0e"} Apr 21 00:15:31.618742 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:31.618709 2568 generic.go:358] "Generic (PLEG): container finished" podID="030cacec-c448-47ae-ae2d-87c2164c7b80" containerID="24239edc02c96f41245d237f49a3da854bab7d93bbd2bd9b27c5c1db00be7143" exitCode=0 Apr 21 00:15:31.619167 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:31.618753 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" event={"ID":"030cacec-c448-47ae-ae2d-87c2164c7b80","Type":"ContainerDied","Data":"24239edc02c96f41245d237f49a3da854bab7d93bbd2bd9b27c5c1db00be7143"} Apr 21 00:15:32.624382 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:32.624350 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" event={"ID":"030cacec-c448-47ae-ae2d-87c2164c7b80","Type":"ContainerStarted","Data":"e62dff04f992a559f786a6c9561dd57a77e0ab7c008cdb08d3e08bd495cb0581"} Apr 21 00:15:32.624760 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:32.624578 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:32.642157 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:32.642114 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" podStartSLOduration=7.429375605 podStartE2EDuration="7.64210366s" podCreationTimestamp="2026-04-21 00:15:25 +0000 UTC" firstStartedPulling="2026-04-21 00:15:31.619525799 +0000 UTC m=+745.444504742" lastFinishedPulling="2026-04-21 00:15:31.832253852 +0000 UTC m=+745.657232797" observedRunningTime="2026-04-21 00:15:32.640973508 +0000 UTC m=+746.465952486" watchObservedRunningTime="2026-04-21 00:15:32.64210366 +0000 UTC m=+746.467082625" Apr 21 00:15:33.629258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:33.629225 2568 generic.go:358] "Generic (PLEG): container finished" podID="a8d7704c-dc4a-42db-9afb-5c9998f6d0a8" containerID="189fdb9530467e84d748ee5d83272c15e731cc157546a3ff07993fa89843942a" exitCode=0 Apr 21 00:15:33.629707 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:33.629297 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" event={"ID":"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8","Type":"ContainerDied","Data":"189fdb9530467e84d748ee5d83272c15e731cc157546a3ff07993fa89843942a"} Apr 21 00:15:34.635417 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:34.635384 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" event={"ID":"a8d7704c-dc4a-42db-9afb-5c9998f6d0a8","Type":"ContainerStarted","Data":"a2feeefe21a610ebed0cd3a4e789dfe7c74be98afddd954df9c3e7413e20ae5c"} Apr 21 00:15:34.635842 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:34.635600 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:15:34.653955 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:34.653913 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" podStartSLOduration=7.477614904 podStartE2EDuration="7.653900956s" podCreationTimestamp="2026-04-21 00:15:27 +0000 UTC" firstStartedPulling="2026-04-21 00:15:33.62995301 +0000 UTC m=+747.454931953" lastFinishedPulling="2026-04-21 00:15:33.806239061 +0000 UTC m=+747.631218005" observedRunningTime="2026-04-21 00:15:34.651862355 +0000 UTC m=+748.476841322" watchObservedRunningTime="2026-04-21 00:15:34.653900956 +0000 UTC m=+748.478879921" Apr 21 00:15:43.642390 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:43.642360 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="llm/e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p" Apr 21 00:15:45.654352 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:15:45.654326 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="llm/e2e-trlp-test-simulated-kserve-84db68679b-sxf4s" Apr 21 00:16:12.763258 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.763222 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-85bd5f48df-qdblp"] Apr 21 00:16:12.766887 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.766869 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.772722 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.772700 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-85bd5f48df-qdblp"] Apr 21 00:16:12.840173 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.840150 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-oidc-ca\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.840298 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.840224 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-tls-cert\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.840298 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.840265 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k4zf\" (UniqueName: \"kubernetes.io/projected/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-kube-api-access-9k4zf\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.940997 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.940971 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-oidc-ca\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.941102 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.941010 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-tls-cert\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.941146 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.941125 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9k4zf\" (UniqueName: \"kubernetes.io/projected/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-kube-api-access-9k4zf\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.941675 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.941657 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-oidc-ca\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.943404 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.943381 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-tls-cert\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:12.948712 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:12.948692 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k4zf\" (UniqueName: \"kubernetes.io/projected/ad0f5f2d-4a40-4038-a23c-210cfa67aef5-kube-api-access-9k4zf\") pod \"authorino-85bd5f48df-qdblp\" (UID: \"ad0f5f2d-4a40-4038-a23c-210cfa67aef5\") " pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:13.079875 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:13.079805 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-85bd5f48df-qdblp" Apr 21 00:16:13.198954 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:13.198920 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-85bd5f48df-qdblp"] Apr 21 00:16:13.200119 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:16:13.200091 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad0f5f2d_4a40_4038_a23c_210cfa67aef5.slice/crio-5321c346f3ebb81f67e285a0916582cf44afbcf26f823444c6aad18360b73152 WatchSource:0}: Error finding container 5321c346f3ebb81f67e285a0916582cf44afbcf26f823444c6aad18360b73152: Status 404 returned error can't find the container with id 5321c346f3ebb81f67e285a0916582cf44afbcf26f823444c6aad18360b73152 Apr 21 00:16:13.785693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:13.784789 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-85bd5f48df-qdblp" event={"ID":"ad0f5f2d-4a40-4038-a23c-210cfa67aef5","Type":"ContainerStarted","Data":"a83c831508a07398740be436eea941902879b38480d184ef9d3dc6e97ea347ab"} Apr 21 00:16:13.785693 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:13.784833 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-85bd5f48df-qdblp" event={"ID":"ad0f5f2d-4a40-4038-a23c-210cfa67aef5","Type":"ContainerStarted","Data":"5321c346f3ebb81f67e285a0916582cf44afbcf26f823444c6aad18360b73152"} Apr 21 00:16:13.803158 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:13.803102 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-85bd5f48df-qdblp" podStartSLOduration=1.437731287 podStartE2EDuration="1.803084819s" podCreationTimestamp="2026-04-21 00:16:12 +0000 UTC" firstStartedPulling="2026-04-21 00:16:13.201282623 +0000 UTC m=+787.026261567" lastFinishedPulling="2026-04-21 00:16:13.566636147 +0000 UTC m=+787.391615099" observedRunningTime="2026-04-21 00:16:13.801915372 +0000 UTC m=+787.626894342" watchObservedRunningTime="2026-04-21 00:16:13.803084819 +0000 UTC m=+787.628063790" Apr 21 00:16:13.836581 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:13.836493 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-596f7d8dbd-cwkfd"] Apr 21 00:16:13.836885 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:13.836832 2568 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" podUID="39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" containerName="authorino" containerID="cri-o://61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c" gracePeriod=30 Apr 21 00:16:14.098954 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.098891 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:16:14.152651 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.152590 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-oidc-ca\") pod \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " Apr 21 00:16:14.152651 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.152645 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-tls-cert\") pod \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " Apr 21 00:16:14.152840 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.152698 2568 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5nfw\" (UniqueName: \"kubernetes.io/projected/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-kube-api-access-x5nfw\") pod \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\" (UID: \"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e\") " Apr 21 00:16:14.154592 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.154569 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-kube-api-access-x5nfw" (OuterVolumeSpecName: "kube-api-access-x5nfw") pod "39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" (UID: "39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e"). InnerVolumeSpecName "kube-api-access-x5nfw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 00:16:14.157309 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.157287 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-oidc-ca" (OuterVolumeSpecName: "oidc-ca") pod "39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" (UID: "39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e"). InnerVolumeSpecName "oidc-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 00:16:14.163557 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.163538 2568 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-tls-cert" (OuterVolumeSpecName: "tls-cert") pod "39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" (UID: "39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e"). InnerVolumeSpecName "tls-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 00:16:14.254317 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.254288 2568 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-x5nfw\" (UniqueName: \"kubernetes.io/projected/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-kube-api-access-x5nfw\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:16:14.254317 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.254316 2568 reconciler_common.go:299] "Volume detached for volume \"oidc-ca\" (UniqueName: \"kubernetes.io/configmap/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-oidc-ca\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:16:14.254317 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.254328 2568 reconciler_common.go:299] "Volume detached for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e-tls-cert\") on node \"ip-10-0-132-206.ec2.internal\" DevicePath \"\"" Apr 21 00:16:14.789824 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.789794 2568 generic.go:358] "Generic (PLEG): container finished" podID="39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" containerID="61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c" exitCode=0 Apr 21 00:16:14.790158 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.789840 2568 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" Apr 21 00:16:14.790158 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.789893 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" event={"ID":"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e","Type":"ContainerDied","Data":"61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c"} Apr 21 00:16:14.790158 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.789927 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-596f7d8dbd-cwkfd" event={"ID":"39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e","Type":"ContainerDied","Data":"423eb5f82384ab23674b5ccefaf0cfaa8ff404fd595fc6c0964f972ffde73d40"} Apr 21 00:16:14.790158 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.789948 2568 scope.go:117] "RemoveContainer" containerID="61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c" Apr 21 00:16:14.798831 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.798815 2568 scope.go:117] "RemoveContainer" containerID="61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c" Apr 21 00:16:14.799076 ip-10-0-132-206 kubenswrapper[2568]: E0421 00:16:14.799058 2568 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c\": container with ID starting with 61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c not found: ID does not exist" containerID="61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c" Apr 21 00:16:14.799123 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.799085 2568 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c"} err="failed to get container status \"61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c\": rpc error: code = NotFound desc = could not find container \"61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c\": container with ID starting with 61ad1bb71f28e46f2ba9a16781ac0c03a97b2c853a4a9049f2f746453f5eeb2c not found: ID does not exist" Apr 21 00:16:14.806601 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.806578 2568 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-596f7d8dbd-cwkfd"] Apr 21 00:16:14.809105 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:14.809083 2568 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-596f7d8dbd-cwkfd"] Apr 21 00:16:16.720673 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:16:16.720638 2568 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" path="/var/lib/kubelet/pods/39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e/volumes" Apr 21 00:18:06.746163 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:18:06.746107 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:18:06.747982 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:18:06.747962 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:23:06.788029 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:23:06.788005 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:23:06.791702 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:23:06.791679 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:28:06.819712 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:28:06.819685 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:28:06.823707 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:28:06.823678 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:33:06.862153 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:33:06.862121 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:33:06.868822 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:33:06.868802 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:38:06.903985 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:06.903959 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:38:06.906423 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:06.905317 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:38:37.095132 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:37.095102 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-85bd5f48df-qdblp_ad0f5f2d-4a40-4038-a23c-210cfa67aef5/authorino/0.log" Apr 21 00:38:41.096323 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:41.096292 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_maas-api-6f98dc884d-4fkhr_5ac6af9d-c5be-4343-886c-11c2e5e56222/maas-api/0.log" Apr 21 00:38:41.696252 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:41.696218 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_opendatahub-operator-controller-manager-587f5698df-hptf6_cfa3adc8-671e-4d7b-b87d-e3eedc861363/manager/0.log" Apr 21 00:38:42.591270 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.591240 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf_dac0921f-b18e-4ed9-b1fb-5af4cf784a1e/util/0.log" Apr 21 00:38:42.597198 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.597163 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf_dac0921f-b18e-4ed9-b1fb-5af4cf784a1e/pull/0.log" Apr 21 00:38:42.602990 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.602957 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf_dac0921f-b18e-4ed9-b1fb-5af4cf784a1e/extract/0.log" Apr 21 00:38:42.719225 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.719200 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx_da67825e-4611-46e0-b960-889d7cd00120/util/0.log" Apr 21 00:38:42.726128 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.726108 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx_da67825e-4611-46e0-b960-889d7cd00120/pull/0.log" Apr 21 00:38:42.732787 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.732765 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx_da67825e-4611-46e0-b960-889d7cd00120/extract/0.log" Apr 21 00:38:42.844995 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.844942 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9_b578effd-613c-44c3-9d8d-579f79fa6e6f/extract/0.log" Apr 21 00:38:42.851700 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.851682 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9_b578effd-613c-44c3-9d8d-579f79fa6e6f/util/0.log" Apr 21 00:38:42.861381 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.861366 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9_b578effd-613c-44c3-9d8d-579f79fa6e6f/pull/0.log" Apr 21 00:38:42.984296 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.984271 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2_2c66175c-c20f-4e2d-a744-485c41e6a54e/util/0.log" Apr 21 00:38:42.990518 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.990497 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2_2c66175c-c20f-4e2d-a744-485c41e6a54e/pull/0.log" Apr 21 00:38:42.996376 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:42.996361 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2_2c66175c-c20f-4e2d-a744-485c41e6a54e/extract/0.log" Apr 21 00:38:43.108392 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:43.108297 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-85bd5f48df-qdblp_ad0f5f2d-4a40-4038-a23c-210cfa67aef5/authorino/0.log" Apr 21 00:38:43.233045 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:43.233022 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-operator-657f44b778-7qwxj_8eac1894-79ad-43d1-aeb1-65ff1d7d538b/manager/0.log" Apr 21 00:38:43.571741 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:43.571712 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_kuadrant-operator-catalog-6f6xn_0ea60df0-bfd0-481d-ad91-9bc26c91bafd/registry-server/0.log" Apr 21 00:38:43.919084 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:43.919005 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_limitador-operator-controller-manager-85c4996f8c-4vgkf_64f29997-9e48-4455-85ed-1ca939a19be9/manager/0.log" Apr 21 00:38:44.260683 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:44.260657 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb_08853a86-b83e-428f-a936-f74a943afee5/istio-proxy/0.log" Apr 21 00:38:44.709578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:44.709494 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_maas-default-gateway-openshift-default-58b6f876-qgzn4_39ea85a6-6e98-4c3c-a41c-c32ea72a6923/istio-proxy/0.log" Apr 21 00:38:45.156674 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:45.156635 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/llm_e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p_030cacec-c448-47ae-ae2d-87c2164c7b80/storage-initializer/0.log" Apr 21 00:38:45.162397 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:45.162375 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/llm_e2e-distinct-2-simulated-kserve-7f849f6b56-k6v5p_030cacec-c448-47ae-ae2d-87c2164c7b80/main/0.log" Apr 21 00:38:45.392081 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:45.392060 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/llm_e2e-trlp-test-simulated-kserve-84db68679b-sxf4s_a8d7704c-dc4a-42db-9afb-5c9998f6d0a8/storage-initializer/0.log" Apr 21 00:38:45.399428 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:45.399412 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/llm_e2e-trlp-test-simulated-kserve-84db68679b-sxf4s_a8d7704c-dc4a-42db-9afb-5c9998f6d0a8/main/0.log" Apr 21 00:38:45.514556 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:45.514523 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/llm_e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5_58b39ea6-8e4f-4f11-a2a1-479b6049e95b/storage-initializer/0.log" Apr 21 00:38:45.521504 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:45.521476 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/llm_e2e-unconfigured-facebook-opt-125m-simulated-kserve-75cdcc6b9s5_58b39ea6-8e4f-4f11-a2a1-479b6049e95b/main/0.log" Apr 21 00:38:45.761054 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:45.761029 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/llm_premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5_73582b91-5337-46b9-9625-c1f9bec6410b/storage-initializer/0.log" Apr 21 00:38:45.771297 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:45.771239 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/llm_premium-simulated-simulated-premium-kserve-6b97b89985-4b9k5_73582b91-5337-46b9-9625-c1f9bec6410b/main/0.log" Apr 21 00:38:52.115782 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:52.115750 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-kt2pp_2f5feffc-f6d3-4684-8a39-a8fa87b1feda/global-pull-secret-syncer/0.log" Apr 21 00:38:52.244489 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:52.244459 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-lmssf_b470fb01-818e-4ce0-8ff7-35c94e72e940/konnectivity-agent/0.log" Apr 21 00:38:52.301997 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:52.301964 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-132-206.ec2.internal_7b08454bff9ca2a8282774b411985429/haproxy/0.log" Apr 21 00:38:55.473563 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.473471 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf_dac0921f-b18e-4ed9-b1fb-5af4cf784a1e/extract/0.log" Apr 21 00:38:55.494388 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.494367 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf_dac0921f-b18e-4ed9-b1fb-5af4cf784a1e/util/0.log" Apr 21 00:38:55.518814 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.518795 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_0acee64185f523d1d1272e9af2e4d9333e0dcde792ba30e1fa9605b759pwrvf_dac0921f-b18e-4ed9-b1fb-5af4cf784a1e/pull/0.log" Apr 21 00:38:55.569538 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.569515 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx_da67825e-4611-46e0-b960-889d7cd00120/extract/0.log" Apr 21 00:38:55.595032 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.595007 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx_da67825e-4611-46e0-b960-889d7cd00120/util/0.log" Apr 21 00:38:55.616663 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.616646 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_19cb86e64775c5699d5aacf881a09c2d51e7f55f9e1ff096f2a667c5e09ndmx_da67825e-4611-46e0-b960-889d7cd00120/pull/0.log" Apr 21 00:38:55.650831 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.650816 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9_b578effd-613c-44c3-9d8d-579f79fa6e6f/extract/0.log" Apr 21 00:38:55.678802 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.678770 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9_b578effd-613c-44c3-9d8d-579f79fa6e6f/util/0.log" Apr 21 00:38:55.699296 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.699279 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_5fc36eb1065777914bfe0ff85f9a202e2a3bafefb563d9e2994474ed73hq7h9_b578effd-613c-44c3-9d8d-579f79fa6e6f/pull/0.log" Apr 21 00:38:55.739772 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.739721 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2_2c66175c-c20f-4e2d-a744-485c41e6a54e/extract/0.log" Apr 21 00:38:55.761493 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.761461 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2_2c66175c-c20f-4e2d-a744-485c41e6a54e/util/0.log" Apr 21 00:38:55.789381 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.789350 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_9438e18bbad664a016701a3153a6b421e2d977be7eb0117f80bcf45ef1c5hv2_2c66175c-c20f-4e2d-a744-485c41e6a54e/pull/0.log" Apr 21 00:38:55.976012 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:55.975978 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-85bd5f48df-qdblp_ad0f5f2d-4a40-4038-a23c-210cfa67aef5/authorino/0.log" Apr 21 00:38:56.028123 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:56.028050 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-operator-657f44b778-7qwxj_8eac1894-79ad-43d1-aeb1-65ff1d7d538b/manager/0.log" Apr 21 00:38:56.114814 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:56.114785 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_kuadrant-operator-catalog-6f6xn_0ea60df0-bfd0-481d-ad91-9bc26c91bafd/registry-server/0.log" Apr 21 00:38:56.325846 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:56.325752 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_limitador-operator-controller-manager-85c4996f8c-4vgkf_64f29997-9e48-4455-85ed-1ca939a19be9/manager/0.log" Apr 21 00:38:58.283138 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.283102 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-wllkx_75b150a3-e0e0-4f7d-aee8-713ff97ccf24/node-exporter/0.log" Apr 21 00:38:58.307704 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.307674 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-wllkx_75b150a3-e0e0-4f7d-aee8-713ff97ccf24/kube-rbac-proxy/0.log" Apr 21 00:38:58.340578 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.340555 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-wllkx_75b150a3-e0e0-4f7d-aee8-713ff97ccf24/init-textfile/0.log" Apr 21 00:38:58.713587 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.713505 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-admission-webhook-57cf98b594-hmmhd_791b8e39-beab-44fd-9b37-14969c3d1b15/prometheus-operator-admission-webhook/0.log" Apr 21 00:38:58.878406 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.878376 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-68fb649c6d-l6tjh_d44788eb-1bb7-419f-be15-c17495046006/thanos-query/0.log" Apr 21 00:38:58.905013 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.902548 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-68fb649c6d-l6tjh_d44788eb-1bb7-419f-be15-c17495046006/kube-rbac-proxy-web/0.log" Apr 21 00:38:58.926546 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.926522 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-68fb649c6d-l6tjh_d44788eb-1bb7-419f-be15-c17495046006/kube-rbac-proxy/0.log" Apr 21 00:38:58.953415 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.953393 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-68fb649c6d-l6tjh_d44788eb-1bb7-419f-be15-c17495046006/prom-label-proxy/0.log" Apr 21 00:38:58.983490 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:58.983459 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-68fb649c6d-l6tjh_d44788eb-1bb7-419f-be15-c17495046006/kube-rbac-proxy-rules/0.log" Apr 21 00:38:59.001559 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:38:59.001544 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-68fb649c6d-l6tjh_d44788eb-1bb7-419f-be15-c17495046006/kube-rbac-proxy-metrics/0.log" Apr 21 00:39:01.324655 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.324625 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5fb9cf77c4-lwm7z_90ba90b0-82f6-41e0-9b35-22186054fe5f/console/0.log" Apr 21 00:39:01.356972 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.356950 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_downloads-6bcc868b7-g8wql_92eb6528-431d-47f2-b2d9-729c971eb0c5/download-server/0.log" Apr 21 00:39:01.610371 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.610281 2568 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q"] Apr 21 00:39:01.610839 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.610815 2568 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" containerName="authorino" Apr 21 00:39:01.610839 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.610839 2568 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" containerName="authorino" Apr 21 00:39:01.610987 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.610921 2568 memory_manager.go:356] "RemoveStaleState removing state" podUID="39bb5d1b-38e1-44d0-961b-f2e4d0b4de8e" containerName="authorino" Apr 21 00:39:01.614045 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.614030 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.616586 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.616562 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-gc8bw\"/\"openshift-service-ca.crt\"" Apr 21 00:39:01.616714 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.616650 2568 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-gc8bw\"/\"default-dockercfg-45ldp\"" Apr 21 00:39:01.617684 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.617670 2568 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-gc8bw\"/\"kube-root-ca.crt\"" Apr 21 00:39:01.623133 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.623113 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q"] Apr 21 00:39:01.640946 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.640919 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-proc\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.641059 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.640954 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-podres\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.641059 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.640985 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-sys\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.641135 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.641066 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgwfs\" (UniqueName: \"kubernetes.io/projected/b793713f-ac74-4bbb-9136-b18b18a8bbd2-kube-api-access-vgwfs\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.641135 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.641115 2568 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-lib-modules\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742096 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742065 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-sys\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742240 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742110 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vgwfs\" (UniqueName: \"kubernetes.io/projected/b793713f-ac74-4bbb-9136-b18b18a8bbd2-kube-api-access-vgwfs\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742240 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742136 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-lib-modules\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742240 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742192 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-proc\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742240 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742192 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-sys\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742240 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742216 2568 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-podres\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742465 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742278 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-proc\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742465 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742331 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-lib-modules\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.742465 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.742333 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/b793713f-ac74-4bbb-9136-b18b18a8bbd2-podres\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.750300 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.750276 2568 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgwfs\" (UniqueName: \"kubernetes.io/projected/b793713f-ac74-4bbb-9136-b18b18a8bbd2-kube-api-access-vgwfs\") pod \"perf-node-gather-daemonset-8cm2q\" (UID: \"b793713f-ac74-4bbb-9136-b18b18a8bbd2\") " pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:01.925732 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:01.925648 2568 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:02.046930 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:02.046900 2568 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q"] Apr 21 00:39:02.048125 ip-10-0-132-206 kubenswrapper[2568]: W0421 00:39:02.048103 2568 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb793713f_ac74_4bbb_9136_b18b18a8bbd2.slice/crio-bb2dc14c471cae7de43cf1df014ecea0b54791a91e91732b52900902dc465c82 WatchSource:0}: Error finding container bb2dc14c471cae7de43cf1df014ecea0b54791a91e91732b52900902dc465c82: Status 404 returned error can't find the container with id bb2dc14c471cae7de43cf1df014ecea0b54791a91e91732b52900902dc465c82 Apr 21 00:39:02.050113 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:02.050094 2568 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 00:39:02.091352 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:02.091325 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" event={"ID":"b793713f-ac74-4bbb-9136-b18b18a8bbd2","Type":"ContainerStarted","Data":"bb2dc14c471cae7de43cf1df014ecea0b54791a91e91732b52900902dc465c82"} Apr 21 00:39:02.856919 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:02.856878 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-t4sw5_169c80c3-e9e6-400e-9412-bab767e5cb98/dns/0.log" Apr 21 00:39:02.877903 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:02.877882 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-t4sw5_169c80c3-e9e6-400e-9412-bab767e5cb98/kube-rbac-proxy/0.log" Apr 21 00:39:02.954803 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:02.954769 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-wx87b_ef105ba8-c06b-4056-a230-9a44e0e3f1b9/dns-node-resolver/0.log" Apr 21 00:39:03.095919 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:03.095886 2568 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" event={"ID":"b793713f-ac74-4bbb-9136-b18b18a8bbd2","Type":"ContainerStarted","Data":"7d44e51efe55c66c5cdab2a25cc87b8f734c1e55459c400c31d482a3431fd2ad"} Apr 21 00:39:03.096099 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:03.096040 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:03.113912 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:03.113826 2568 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" podStartSLOduration=2.113812827 podStartE2EDuration="2.113812827s" podCreationTimestamp="2026-04-21 00:39:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 00:39:03.111663933 +0000 UTC m=+2156.936642894" watchObservedRunningTime="2026-04-21 00:39:03.113812827 +0000 UTC m=+2156.938791792" Apr 21 00:39:03.441264 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:03.441189 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_image-registry-575d785c49-gbr8m_1d4145a7-09e7-45ef-bfaa-d03f1f3c92f1/registry/0.log" Apr 21 00:39:03.510946 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:03.510920 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-bscx2_cde5a5a1-f134-48c1-be68-a8c2a968e9b5/node-ca/0.log" Apr 21 00:39:04.366548 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:04.366518 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_data-science-gateway-data-science-gateway-class-5cb8b776cfg4njb_08853a86-b83e-428f-a936-f74a943afee5/istio-proxy/0.log" Apr 21 00:39:04.632137 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:04.632057 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_maas-default-gateway-openshift-default-58b6f876-qgzn4_39ea85a6-6e98-4c3c-a41c-c32ea72a6923/istio-proxy/0.log" Apr 21 00:39:05.192886 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:05.192859 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-pvnc8_fbbce771-9a06-4a5a-9dbf-526282060d89/serve-healthcheck-canary/0.log" Apr 21 00:39:05.725118 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:05.725089 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-kdsh6_27d810d8-53e4-460c-8437-9151e193bcfe/kube-rbac-proxy/0.log" Apr 21 00:39:05.745044 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:05.745017 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-kdsh6_27d810d8-53e4-460c-8437-9151e193bcfe/exporter/0.log" Apr 21 00:39:05.766536 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:05.766514 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-kdsh6_27d810d8-53e4-460c-8437-9151e193bcfe/extractor/0.log" Apr 21 00:39:07.830773 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:07.830745 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_maas-api-6f98dc884d-4fkhr_5ac6af9d-c5be-4343-886c-11c2e5e56222/maas-api/0.log" Apr 21 00:39:08.039240 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:08.039207 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_opendatahub-operator-controller-manager-587f5698df-hptf6_cfa3adc8-671e-4d7b-b87d-e3eedc861363/manager/0.log" Apr 21 00:39:09.108949 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:09.108919 2568 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-gc8bw/perf-node-gather-daemonset-8cm2q" Apr 21 00:39:09.284457 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:09.284430 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-lws-operator_lws-controller-manager-5956879555-jvtz9_3975eb23-f72b-46ae-82d8-4eb8e6eb58b6/manager/0.log" Apr 21 00:39:13.844989 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:13.844961 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-mfrlg_848545eb-5390-420a-8f8e-3b327e38e776/migrator/0.log" Apr 21 00:39:13.866587 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:13.866563 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-mfrlg_848545eb-5390-420a-8f8e-3b327e38e776/graceful-termination/0.log" Apr 21 00:39:15.498290 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.498261 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rm7p5_dcfdee3f-a46f-46a3-aaf5-68b7638cc717/kube-multus-additional-cni-plugins/0.log" Apr 21 00:39:15.519023 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.518998 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rm7p5_dcfdee3f-a46f-46a3-aaf5-68b7638cc717/egress-router-binary-copy/0.log" Apr 21 00:39:15.540938 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.540916 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rm7p5_dcfdee3f-a46f-46a3-aaf5-68b7638cc717/cni-plugins/0.log" Apr 21 00:39:15.562151 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.562124 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rm7p5_dcfdee3f-a46f-46a3-aaf5-68b7638cc717/bond-cni-plugin/0.log" Apr 21 00:39:15.584474 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.584455 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rm7p5_dcfdee3f-a46f-46a3-aaf5-68b7638cc717/routeoverride-cni/0.log" Apr 21 00:39:15.605552 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.605533 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rm7p5_dcfdee3f-a46f-46a3-aaf5-68b7638cc717/whereabouts-cni-bincopy/0.log" Apr 21 00:39:15.627462 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.627425 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rm7p5_dcfdee3f-a46f-46a3-aaf5-68b7638cc717/whereabouts-cni/0.log" Apr 21 00:39:15.670798 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.670774 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6sk6_18a5a55e-ea26-46ec-a17f-1cac467f786c/kube-multus/0.log" Apr 21 00:39:15.840192 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.840104 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-pf4ls_d439efe8-c633-4c7c-ad01-515f448307fd/network-metrics-daemon/0.log" Apr 21 00:39:15.859734 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:15.859690 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-pf4ls_d439efe8-c633-4c7c-ad01-515f448307fd/kube-rbac-proxy/0.log" Apr 21 00:39:17.343775 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.343743 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-controller/0.log" Apr 21 00:39:17.362522 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.362496 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/0.log" Apr 21 00:39:17.372032 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.372012 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovn-acl-logging/1.log" Apr 21 00:39:17.389372 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.389348 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/kube-rbac-proxy-node/0.log" Apr 21 00:39:17.410100 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.410078 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/kube-rbac-proxy-ovn-metrics/0.log" Apr 21 00:39:17.429652 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.429596 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/northd/0.log" Apr 21 00:39:17.450096 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.450082 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/nbdb/0.log" Apr 21 00:39:17.470977 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.470961 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/sbdb/0.log" Apr 21 00:39:17.567823 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:17.567792 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7lpw_89e3cc31-318f-4b7b-9926-eebe7c5e2e88/ovnkube-controller/0.log" Apr 21 00:39:18.732365 ip-10-0-132-206 kubenswrapper[2568]: I0421 00:39:18.732339 2568 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-4tj2p_0b745e1d-cf68-43fd-b452-36e45a05eb93/network-check-target-container/0.log"