Apr 21 10:01:17.235018 ip-10-0-142-243 systemd[1]: kubelet.service: Failed to load environment files: No such file or directory Apr 21 10:01:17.235031 ip-10-0-142-243 systemd[1]: kubelet.service: Failed to run 'start-pre' task: No such file or directory Apr 21 10:01:17.235041 ip-10-0-142-243 systemd[1]: kubelet.service: Failed with result 'resources'. Apr 21 10:01:17.235338 ip-10-0-142-243 systemd[1]: Failed to start Kubernetes Kubelet. Apr 21 10:01:28.341060 ip-10-0-142-243 systemd[1]: kubelet.service: Failed to schedule restart job: Unit crio.service not found. Apr 21 10:01:28.341078 ip-10-0-142-243 systemd[1]: kubelet.service: Failed with result 'resources'. -- Boot 9607d2c681a5469ab10c0b8d8042b935 -- Apr 21 10:03:53.814550 ip-10-0-142-243 systemd[1]: Starting Kubernetes Kubelet... Apr 21 10:03:54.198591 ip-10-0-142-243 kubenswrapper[2570]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 10:03:54.198591 ip-10-0-142-243 kubenswrapper[2570]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 21 10:03:54.198591 ip-10-0-142-243 kubenswrapper[2570]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 10:03:54.198591 ip-10-0-142-243 kubenswrapper[2570]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 21 10:03:54.198591 ip-10-0-142-243 kubenswrapper[2570]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 10:03:54.200143 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.200055 2570 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 21 10:03:54.202335 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202304 2570 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:54.202335 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202323 2570 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:54.202335 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202328 2570 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:54.202335 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202332 2570 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:54.202335 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202337 2570 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:54.202335 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202341 2570 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202345 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202349 2570 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202355 2570 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202361 2570 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202365 2570 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202368 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202372 2570 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202376 2570 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202379 2570 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202383 2570 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202387 2570 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202391 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202394 2570 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202398 2570 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202402 2570 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202405 2570 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202415 2570 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202419 2570 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202423 2570 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:54.202661 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202426 2570 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202430 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202433 2570 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202439 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202442 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202448 2570 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202454 2570 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202458 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202463 2570 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202467 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202472 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202476 2570 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202489 2570 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202494 2570 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202498 2570 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202503 2570 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202507 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202511 2570 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202515 2570 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:54.203378 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202519 2570 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202524 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202528 2570 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202532 2570 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202536 2570 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202540 2570 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202543 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202547 2570 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202551 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202556 2570 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202560 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202564 2570 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202568 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202572 2570 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202576 2570 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202580 2570 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202584 2570 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202590 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202594 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202599 2570 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:54.204024 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202604 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202608 2570 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202612 2570 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202616 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202620 2570 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202624 2570 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202628 2570 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202632 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202637 2570 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202641 2570 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202644 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202650 2570 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202653 2570 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202657 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202661 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202665 2570 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202669 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202672 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202676 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202680 2570 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:54.204951 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202684 2570 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.202688 2570 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204064 2570 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204075 2570 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204080 2570 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204085 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204089 2570 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204094 2570 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204098 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204103 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204110 2570 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204116 2570 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204121 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204125 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204131 2570 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204136 2570 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204141 2570 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204145 2570 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204149 2570 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204154 2570 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:54.205828 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204158 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204162 2570 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204166 2570 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204171 2570 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204176 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204180 2570 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204184 2570 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204188 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204192 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204196 2570 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204200 2570 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204204 2570 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204209 2570 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204213 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204217 2570 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204221 2570 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204225 2570 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204230 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204234 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204238 2570 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:54.206417 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204242 2570 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204246 2570 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204255 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204259 2570 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204263 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204290 2570 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204295 2570 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204300 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204304 2570 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204308 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204312 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204316 2570 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204320 2570 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204326 2570 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204330 2570 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204334 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204339 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204344 2570 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204348 2570 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204352 2570 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:54.207053 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204356 2570 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204360 2570 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204364 2570 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204369 2570 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204373 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204377 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204381 2570 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204385 2570 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204389 2570 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204393 2570 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204397 2570 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204402 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204406 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204410 2570 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204416 2570 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204421 2570 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204425 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204429 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204433 2570 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:54.207802 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204438 2570 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204441 2570 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204445 2570 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204449 2570 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204458 2570 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204464 2570 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204469 2570 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204474 2570 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.204479 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204574 2570 flags.go:64] FLAG: --address="0.0.0.0" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204586 2570 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204594 2570 flags.go:64] FLAG: --anonymous-auth="true" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204601 2570 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204607 2570 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204612 2570 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204619 2570 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204626 2570 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204630 2570 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204635 2570 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204641 2570 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204646 2570 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 21 10:03:54.208291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204651 2570 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204656 2570 flags.go:64] FLAG: --cgroup-root="" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204661 2570 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204665 2570 flags.go:64] FLAG: --client-ca-file="" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204670 2570 flags.go:64] FLAG: --cloud-config="" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204674 2570 flags.go:64] FLAG: --cloud-provider="external" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204679 2570 flags.go:64] FLAG: --cluster-dns="[]" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204687 2570 flags.go:64] FLAG: --cluster-domain="" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204692 2570 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204697 2570 flags.go:64] FLAG: --config-dir="" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204702 2570 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204708 2570 flags.go:64] FLAG: --container-log-max-files="5" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204713 2570 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204718 2570 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204722 2570 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204727 2570 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204732 2570 flags.go:64] FLAG: --contention-profiling="false" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204736 2570 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204741 2570 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204746 2570 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204750 2570 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204757 2570 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204763 2570 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204767 2570 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204772 2570 flags.go:64] FLAG: --enable-load-reader="false" Apr 21 10:03:54.208895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204776 2570 flags.go:64] FLAG: --enable-server="true" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204781 2570 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204788 2570 flags.go:64] FLAG: --event-burst="100" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204794 2570 flags.go:64] FLAG: --event-qps="50" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204798 2570 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204803 2570 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204808 2570 flags.go:64] FLAG: --eviction-hard="" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204814 2570 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204818 2570 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204823 2570 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204829 2570 flags.go:64] FLAG: --eviction-soft="" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204834 2570 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204839 2570 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204844 2570 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204848 2570 flags.go:64] FLAG: --experimental-mounter-path="" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204853 2570 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204857 2570 flags.go:64] FLAG: --fail-swap-on="true" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204862 2570 flags.go:64] FLAG: --feature-gates="" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204867 2570 flags.go:64] FLAG: --file-check-frequency="20s" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204873 2570 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204877 2570 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204882 2570 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204887 2570 flags.go:64] FLAG: --healthz-port="10248" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204892 2570 flags.go:64] FLAG: --help="false" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204896 2570 flags.go:64] FLAG: --hostname-override="ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.209535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204901 2570 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204905 2570 flags.go:64] FLAG: --http-check-frequency="20s" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204910 2570 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204916 2570 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204922 2570 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204927 2570 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204931 2570 flags.go:64] FLAG: --image-service-endpoint="" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204936 2570 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204940 2570 flags.go:64] FLAG: --kube-api-burst="100" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204945 2570 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204950 2570 flags.go:64] FLAG: --kube-api-qps="50" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204958 2570 flags.go:64] FLAG: --kube-reserved="" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204962 2570 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204966 2570 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204971 2570 flags.go:64] FLAG: --kubelet-cgroups="" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204975 2570 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204980 2570 flags.go:64] FLAG: --lock-file="" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204984 2570 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204989 2570 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.204994 2570 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205002 2570 flags.go:64] FLAG: --log-json-split-stream="false" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205007 2570 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205012 2570 flags.go:64] FLAG: --log-text-split-stream="false" Apr 21 10:03:54.210154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205016 2570 flags.go:64] FLAG: --logging-format="text" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205021 2570 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205026 2570 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205031 2570 flags.go:64] FLAG: --manifest-url="" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205035 2570 flags.go:64] FLAG: --manifest-url-header="" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205042 2570 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205047 2570 flags.go:64] FLAG: --max-open-files="1000000" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205053 2570 flags.go:64] FLAG: --max-pods="110" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205058 2570 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205062 2570 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205067 2570 flags.go:64] FLAG: --memory-manager-policy="None" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205072 2570 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205077 2570 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205081 2570 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205086 2570 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205098 2570 flags.go:64] FLAG: --node-status-max-images="50" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205102 2570 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205107 2570 flags.go:64] FLAG: --oom-score-adj="-999" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205112 2570 flags.go:64] FLAG: --pod-cidr="" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205116 2570 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205125 2570 flags.go:64] FLAG: --pod-manifest-path="" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205130 2570 flags.go:64] FLAG: --pod-max-pids="-1" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205135 2570 flags.go:64] FLAG: --pods-per-core="0" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205140 2570 flags.go:64] FLAG: --port="10250" Apr 21 10:03:54.210727 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205144 2570 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205150 2570 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-0c21b135f70edc55b" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205155 2570 flags.go:64] FLAG: --qos-reserved="" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205159 2570 flags.go:64] FLAG: --read-only-port="10255" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205164 2570 flags.go:64] FLAG: --register-node="true" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205168 2570 flags.go:64] FLAG: --register-schedulable="true" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205173 2570 flags.go:64] FLAG: --register-with-taints="" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205179 2570 flags.go:64] FLAG: --registry-burst="10" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205183 2570 flags.go:64] FLAG: --registry-qps="5" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205187 2570 flags.go:64] FLAG: --reserved-cpus="" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205192 2570 flags.go:64] FLAG: --reserved-memory="" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205198 2570 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205203 2570 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205208 2570 flags.go:64] FLAG: --rotate-certificates="false" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205213 2570 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205217 2570 flags.go:64] FLAG: --runonce="false" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205222 2570 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205227 2570 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205231 2570 flags.go:64] FLAG: --seccomp-default="false" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205238 2570 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205243 2570 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205248 2570 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205253 2570 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205258 2570 flags.go:64] FLAG: --storage-driver-password="root" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205263 2570 flags.go:64] FLAG: --storage-driver-secure="false" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205286 2570 flags.go:64] FLAG: --storage-driver-table="stats" Apr 21 10:03:54.211356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205290 2570 flags.go:64] FLAG: --storage-driver-user="root" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205295 2570 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205300 2570 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205304 2570 flags.go:64] FLAG: --system-cgroups="" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205309 2570 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205318 2570 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205323 2570 flags.go:64] FLAG: --tls-cert-file="" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205327 2570 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205335 2570 flags.go:64] FLAG: --tls-min-version="" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205339 2570 flags.go:64] FLAG: --tls-private-key-file="" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205344 2570 flags.go:64] FLAG: --topology-manager-policy="none" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205348 2570 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205353 2570 flags.go:64] FLAG: --topology-manager-scope="container" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205358 2570 flags.go:64] FLAG: --v="2" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205365 2570 flags.go:64] FLAG: --version="false" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205371 2570 flags.go:64] FLAG: --vmodule="" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205377 2570 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.205382 2570 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205536 2570 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205542 2570 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205548 2570 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205553 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205558 2570 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205563 2570 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:54.212424 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205567 2570 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205573 2570 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205578 2570 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205582 2570 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205587 2570 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205592 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205596 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205601 2570 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205606 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205611 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205615 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205619 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205623 2570 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205627 2570 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205631 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205635 2570 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205639 2570 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205643 2570 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205647 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205651 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:54.213325 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205655 2570 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205659 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205664 2570 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205668 2570 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205672 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205676 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205681 2570 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205685 2570 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205689 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205693 2570 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205697 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205702 2570 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205708 2570 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205719 2570 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205724 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205729 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205733 2570 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205737 2570 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205741 2570 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:54.213865 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205745 2570 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205750 2570 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205754 2570 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205758 2570 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205763 2570 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205767 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205771 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205775 2570 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205779 2570 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205783 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205787 2570 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205791 2570 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205795 2570 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205799 2570 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205803 2570 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205807 2570 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205811 2570 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205815 2570 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205819 2570 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205823 2570 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:54.214481 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205827 2570 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205831 2570 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205835 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205839 2570 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205856 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205860 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205868 2570 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205872 2570 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205878 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205882 2570 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205886 2570 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205890 2570 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205894 2570 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205898 2570 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205911 2570 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205916 2570 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205923 2570 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205928 2570 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205933 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:54.215362 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205937 2570 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:54.215841 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.205941 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:54.215841 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.206490 2570 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 10:03:54.215926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.215901 2570 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 21 10:03:54.215926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.215924 2570 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 21 10:03:54.216032 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216019 2570 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:54.216032 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216031 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216037 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216041 2570 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216046 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216050 2570 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216054 2570 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216058 2570 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216063 2570 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216067 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216071 2570 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216075 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216079 2570 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216083 2570 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216087 2570 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216091 2570 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216096 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216100 2570 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216104 2570 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216108 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216112 2570 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:54.216122 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216116 2570 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216120 2570 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216124 2570 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216128 2570 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216132 2570 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216137 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216140 2570 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216144 2570 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216149 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216152 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216165 2570 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216169 2570 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216173 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216177 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216181 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216184 2570 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216188 2570 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216192 2570 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216196 2570 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216200 2570 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:54.217052 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216205 2570 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216209 2570 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216213 2570 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216217 2570 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216221 2570 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216225 2570 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216229 2570 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216233 2570 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216238 2570 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216242 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216246 2570 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216250 2570 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216291 2570 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216296 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216301 2570 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216305 2570 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216308 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216314 2570 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216319 2570 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216323 2570 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:54.217632 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216327 2570 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216331 2570 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216335 2570 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216347 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216352 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216356 2570 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216361 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216364 2570 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216368 2570 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216372 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216377 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216381 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216385 2570 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216389 2570 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216393 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216397 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216401 2570 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216405 2570 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216412 2570 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216419 2570 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:54.218191 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216423 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216428 2570 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216432 2570 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216436 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216440 2570 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.216448 2570 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216656 2570 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216664 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216669 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216673 2570 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216677 2570 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216682 2570 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216686 2570 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216690 2570 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216693 2570 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:54.218926 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216697 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216709 2570 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216713 2570 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216717 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216721 2570 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216725 2570 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216730 2570 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216734 2570 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216737 2570 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216741 2570 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216745 2570 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216749 2570 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216753 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216757 2570 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216761 2570 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216765 2570 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216769 2570 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216773 2570 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216778 2570 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216781 2570 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:54.219354 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216785 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216790 2570 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216794 2570 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216798 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216803 2570 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216807 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216812 2570 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216816 2570 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216821 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216825 2570 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216829 2570 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216833 2570 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216837 2570 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216841 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216847 2570 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216854 2570 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216861 2570 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216866 2570 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216870 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:54.219852 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216874 2570 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216879 2570 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216883 2570 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216888 2570 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216893 2570 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216897 2570 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216902 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216906 2570 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216910 2570 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216914 2570 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216918 2570 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216923 2570 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216926 2570 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216932 2570 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216938 2570 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216942 2570 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216946 2570 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216950 2570 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216954 2570 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216958 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:54.220391 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216963 2570 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216966 2570 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216970 2570 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216974 2570 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216978 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216983 2570 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216987 2570 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216991 2570 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.216996 2570 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217000 2570 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217004 2570 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217008 2570 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217012 2570 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217016 2570 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217020 2570 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217024 2570 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217028 2570 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:54.221185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:54.217032 2570 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:54.221782 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.217039 2570 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 10:03:54.221782 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.217784 2570 server.go:962] "Client rotation is on, will bootstrap in background" Apr 21 10:03:54.221930 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.221914 2570 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 21 10:03:54.222679 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.222666 2570 server.go:1019] "Starting client certificate rotation" Apr 21 10:03:54.222781 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.222765 2570 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 21 10:03:54.222815 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.222800 2570 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 21 10:03:54.244096 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.244071 2570 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 21 10:03:54.245800 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.245781 2570 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 21 10:03:54.263516 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.263500 2570 log.go:25] "Validated CRI v1 runtime API" Apr 21 10:03:54.268576 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.268561 2570 log.go:25] "Validated CRI v1 image API" Apr 21 10:03:54.269801 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.269782 2570 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 21 10:03:54.271500 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.271483 2570 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 21 10:03:54.273250 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.273228 2570 fs.go:135] Filesystem UUIDs: map[7B77-95E7:/dev/nvme0n1p2 f8f6d347-4bf9-4872-b909-feb24210b8f1:/dev/nvme0n1p4 fa46895a-630f-4abf-88cb-161a93a71881:/dev/nvme0n1p3] Apr 21 10:03:54.273342 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.273250 2570 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 21 10:03:54.280070 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.279874 2570 manager.go:217] Machine: {Timestamp:2026-04-21 10:03:54.276810616 +0000 UTC m=+0.358026450 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3105374 MemoryCapacity:32812179456 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec2032aadd68adced34f4a283090450d SystemUUID:ec2032aa-dd68-adce-d34f-4a283090450d BootID:9607d2c6-81a5-469a-b10c-0b8d8042b935 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6562439168 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16406089728 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16406089728 Type:vfs Inodes:4005393 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:05:28:91:19:43 Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:05:28:91:19:43 Speed:0 Mtu:9001} {Name:ovs-system MacAddress:0a:3e:d5:7d:15:71 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:32812179456 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:34603008 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 21 10:03:54.280070 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.280067 2570 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 21 10:03:54.280170 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.280136 2570 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 21 10:03:54.281154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.281131 2570 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 21 10:03:54.281312 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.281157 2570 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-142-243.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 21 10:03:54.281359 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.281321 2570 topology_manager.go:138] "Creating topology manager with none policy" Apr 21 10:03:54.281359 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.281329 2570 container_manager_linux.go:306] "Creating device plugin manager" Apr 21 10:03:54.281359 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.281341 2570 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 21 10:03:54.281971 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.281954 2570 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 21 10:03:54.283246 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.283236 2570 state_mem.go:36] "Initialized new in-memory state store" Apr 21 10:03:54.283367 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.283358 2570 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 21 10:03:54.285605 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.285595 2570 kubelet.go:491] "Attempting to sync node with API server" Apr 21 10:03:54.285644 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.285616 2570 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 21 10:03:54.285644 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.285628 2570 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 21 10:03:54.285644 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.285637 2570 kubelet.go:397] "Adding apiserver pod source" Apr 21 10:03:54.285736 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.285646 2570 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 21 10:03:54.286675 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.286661 2570 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 21 10:03:54.286710 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.286687 2570 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 21 10:03:54.290586 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.290568 2570 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 21 10:03:54.291832 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.291819 2570 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 21 10:03:54.293308 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293294 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293312 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293318 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293324 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293329 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293335 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293340 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293346 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293351 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293357 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293370 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 21 10:03:54.293374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.293378 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 21 10:03:54.294198 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.294188 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 21 10:03:54.294198 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.294198 2570 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 21 10:03:54.297610 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.297597 2570 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 21 10:03:54.297672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.297632 2570 server.go:1295] "Started kubelet" Apr 21 10:03:54.297753 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.297712 2570 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 21 10:03:54.297814 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.297733 2570 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 21 10:03:54.297857 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.297808 2570 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 21 10:03:54.299115 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.299096 2570 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-142-243.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 21 10:03:54.299249 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.299217 2570 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 21 10:03:54.299367 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.299346 2570 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-142-243.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 21 10:03:54.299368 ip-10-0-142-243 systemd[1]: Started Kubernetes Kubelet. Apr 21 10:03:54.299670 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.299650 2570 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 21 10:03:54.301528 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.301510 2570 server.go:317] "Adding debug handlers to kubelet server" Apr 21 10:03:54.306037 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.306019 2570 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 21 10:03:54.306037 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.306031 2570 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 21 10:03:54.306549 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.306531 2570 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 21 10:03:54.306646 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.306537 2570 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 21 10:03:54.306722 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.306662 2570 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 21 10:03:54.306809 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.306797 2570 reconstruct.go:97] "Volume reconstruction finished" Apr 21 10:03:54.306869 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.305784 2570 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-142-243.ec2.internal.18a8571cd25c0a94 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-142-243.ec2.internal,UID:ip-10-0-142-243.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-10-0-142-243.ec2.internal,},FirstTimestamp:2026-04-21 10:03:54.297608852 +0000 UTC m=+0.378824674,LastTimestamp:2026-04-21 10:03:54.297608852 +0000 UTC m=+0.378824674,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-142-243.ec2.internal,}" Apr 21 10:03:54.306869 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.306831 2570 reconciler.go:26] "Reconciler: start to sync state" Apr 21 10:03:54.307410 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.307392 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:54.307512 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.307450 2570 factory.go:153] Registering CRI-O factory Apr 21 10:03:54.307512 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.307467 2570 factory.go:223] Registration of the crio container factory successfully Apr 21 10:03:54.307610 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.307526 2570 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 21 10:03:54.307610 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.307536 2570 factory.go:55] Registering systemd factory Apr 21 10:03:54.307610 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.307544 2570 factory.go:223] Registration of the systemd container factory successfully Apr 21 10:03:54.307610 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.307565 2570 factory.go:103] Registering Raw factory Apr 21 10:03:54.307610 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.307576 2570 manager.go:1196] Started watching for new ooms in manager Apr 21 10:03:54.308439 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.308420 2570 manager.go:319] Starting recovery of all containers Apr 21 10:03:54.311451 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.311356 2570 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 21 10:03:54.315142 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.315120 2570 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 21 10:03:54.315222 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.315149 2570 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"ip-10-0-142-243.ec2.internal\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Apr 21 10:03:54.320357 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.320340 2570 manager.go:324] Recovery completed Apr 21 10:03:54.323455 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.323436 2570 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-g6wwq" Apr 21 10:03:54.324293 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.324282 2570 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:54.326604 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.326588 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:54.326669 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.326616 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:54.326669 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.326626 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:54.327072 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.327057 2570 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 21 10:03:54.327072 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.327070 2570 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 21 10:03:54.327170 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.327087 2570 state_mem.go:36] "Initialized new in-memory state store" Apr 21 10:03:54.328731 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.328717 2570 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-g6wwq" Apr 21 10:03:54.329984 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.329973 2570 policy_none.go:49] "None policy: Start" Apr 21 10:03:54.330028 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.329988 2570 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 21 10:03:54.330028 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.329999 2570 state_mem.go:35] "Initializing new in-memory state store" Apr 21 10:03:54.369813 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.369799 2570 manager.go:341] "Starting Device Plugin manager" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.369826 2570 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.369835 2570 server.go:85] "Starting device plugin registration server" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.370024 2570 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.370037 2570 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.370136 2570 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.370211 2570 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.370219 2570 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.370806 2570 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 21 10:03:54.384227 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.370832 2570 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:54.456429 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.456379 2570 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 21 10:03:54.457644 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.457627 2570 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 21 10:03:54.457725 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.457654 2570 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 21 10:03:54.457725 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.457670 2570 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 21 10:03:54.457725 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.457676 2570 kubelet.go:2451] "Starting kubelet main sync loop" Apr 21 10:03:54.457725 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.457707 2570 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 21 10:03:54.459820 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.459802 2570 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:54.470756 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.470737 2570 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:54.471499 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.471486 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:54.471575 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.471512 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:54.471575 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.471522 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:54.471575 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.471542 2570 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.477041 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.477026 2570 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.477104 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.477045 2570 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-142-243.ec2.internal\": node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:54.504519 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.504502 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:54.558662 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.558640 2570 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal"] Apr 21 10:03:54.558728 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.558696 2570 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:54.559503 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.559490 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:54.559554 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.559514 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:54.559554 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.559525 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:54.560629 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.560617 2570 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:54.561171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.561156 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.561238 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.561191 2570 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:54.561238 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.561219 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:54.561333 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.561240 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:54.561333 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.561254 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:54.561843 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.561828 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:54.561910 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.561853 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:54.561910 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.561863 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:54.562449 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.562433 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.562522 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.562458 2570 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:54.563083 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.563068 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:54.563161 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.563094 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:54.563161 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.563108 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:54.587814 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.587791 2570 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-142-243.ec2.internal\" not found" node="ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.591812 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.591797 2570 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-142-243.ec2.internal\" not found" node="ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.605586 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.605566 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:54.608095 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.608077 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/b2386108552c5c182289de71222968c1-config\") pod \"kube-apiserver-proxy-ip-10-0-142-243.ec2.internal\" (UID: \"b2386108552c5c182289de71222968c1\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.608176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.608106 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/c1401eb7a080b0c324e5cdb3dfc2f894-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal\" (UID: \"c1401eb7a080b0c324e5cdb3dfc2f894\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.608176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.608132 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1401eb7a080b0c324e5cdb3dfc2f894-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal\" (UID: \"c1401eb7a080b0c324e5cdb3dfc2f894\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.706725 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.706664 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:54.708880 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.708867 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/b2386108552c5c182289de71222968c1-config\") pod \"kube-apiserver-proxy-ip-10-0-142-243.ec2.internal\" (UID: \"b2386108552c5c182289de71222968c1\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.708932 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.708889 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/c1401eb7a080b0c324e5cdb3dfc2f894-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal\" (UID: \"c1401eb7a080b0c324e5cdb3dfc2f894\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.708932 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.708906 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1401eb7a080b0c324e5cdb3dfc2f894-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal\" (UID: \"c1401eb7a080b0c324e5cdb3dfc2f894\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.708992 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.708947 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1401eb7a080b0c324e5cdb3dfc2f894-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal\" (UID: \"c1401eb7a080b0c324e5cdb3dfc2f894\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.709023 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.708981 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/c1401eb7a080b0c324e5cdb3dfc2f894-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal\" (UID: \"c1401eb7a080b0c324e5cdb3dfc2f894\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.709023 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.708981 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/b2386108552c5c182289de71222968c1-config\") pod \"kube-apiserver-proxy-ip-10-0-142-243.ec2.internal\" (UID: \"b2386108552c5c182289de71222968c1\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.807358 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.807332 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:54.889669 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.889655 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.894123 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:54.894103 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" Apr 21 10:03:54.907718 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:54.907696 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:55.008187 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:55.008127 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:55.108648 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:55.108616 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:55.209123 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:55.209100 2570 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-243.ec2.internal\" not found" Apr 21 10:03:55.222404 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.222388 2570 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 21 10:03:55.222530 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.222515 2570 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 10:03:55.296500 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.296453 2570 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:55.306190 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.306168 2570 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 21 10:03:55.306629 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.306609 2570 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" Apr 21 10:03:55.320221 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.320192 2570 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 21 10:03:55.321629 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.321609 2570 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" Apr 21 10:03:55.322619 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.322598 2570 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 21 10:03:55.331155 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.331126 2570 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-20 09:58:54 +0000 UTC" deadline="2028-01-30 03:49:46.567573674 +0000 UTC" Apr 21 10:03:55.331155 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.331154 2570 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 21 10:03:55.331327 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.331154 2570 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="15569h45m51.2364229s" Apr 21 10:03:55.336621 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.336603 2570 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-c2q74" Apr 21 10:03:55.343771 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.343753 2570 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-c2q74" Apr 21 10:03:55.386835 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:55.386808 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1401eb7a080b0c324e5cdb3dfc2f894.slice/crio-42204dcf79158147cfb115eb2748f3c2192877350aaca7528f7c519398e4963f WatchSource:0}: Error finding container 42204dcf79158147cfb115eb2748f3c2192877350aaca7528f7c519398e4963f: Status 404 returned error can't find the container with id 42204dcf79158147cfb115eb2748f3c2192877350aaca7528f7c519398e4963f Apr 21 10:03:55.387347 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:55.387263 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2386108552c5c182289de71222968c1.slice/crio-f2124dcfc994b6b5106b3fe9ac641885bb9b833e281b22eee8590281a5a05f11 WatchSource:0}: Error finding container f2124dcfc994b6b5106b3fe9ac641885bb9b833e281b22eee8590281a5a05f11: Status 404 returned error can't find the container with id f2124dcfc994b6b5106b3fe9ac641885bb9b833e281b22eee8590281a5a05f11 Apr 21 10:03:55.391991 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.391970 2570 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 10:03:55.460321 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.460262 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" event={"ID":"b2386108552c5c182289de71222968c1","Type":"ContainerStarted","Data":"f2124dcfc994b6b5106b3fe9ac641885bb9b833e281b22eee8590281a5a05f11"} Apr 21 10:03:55.461191 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.461173 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" event={"ID":"c1401eb7a080b0c324e5cdb3dfc2f894","Type":"ContainerStarted","Data":"42204dcf79158147cfb115eb2748f3c2192877350aaca7528f7c519398e4963f"} Apr 21 10:03:55.696761 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.696742 2570 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:55.803649 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:55.803623 2570 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:56.286852 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.286824 2570 apiserver.go:52] "Watching apiserver" Apr 21 10:03:56.294800 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.294777 2570 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 21 10:03:56.296090 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.296067 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-lggw9","kube-system/konnectivity-agent-pmfn6","kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q","openshift-dns/node-resolver-rrdxp","openshift-multus/multus-additional-cni-plugins-rczjx","openshift-ovn-kubernetes/ovnkube-node-qbn7w","openshift-cluster-node-tuning-operator/tuned-ngjbx","openshift-image-registry/node-ca-77lqp","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal","openshift-multus/multus-jrkkg","openshift-multus/network-metrics-daemon-ff5dl","openshift-network-diagnostics/network-check-target-2j8xh"] Apr 21 10:03:56.297935 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.297917 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.299180 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.298965 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:03:56.300157 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.300140 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.300669 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.300647 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 21 10:03:56.300779 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.300759 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 21 10:03:56.300938 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.300813 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 21 10:03:56.301477 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.301458 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.301832 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.301814 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 21 10:03:56.301832 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.301831 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 21 10:03:56.301982 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.301874 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 21 10:03:56.302039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.301986 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-ffrvb\"" Apr 21 10:03:56.302039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.302014 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 21 10:03:56.302138 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.302111 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 21 10:03:56.302286 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.302253 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-2hhlj\"" Apr 21 10:03:56.302808 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.302791 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.303955 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.303423 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 21 10:03:56.303955 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.303467 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 21 10:03:56.303955 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.303513 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 21 10:03:56.303955 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.303737 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-z99xn\"" Apr 21 10:03:56.304624 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.304602 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 21 10:03:56.306858 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.306841 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 21 10:03:56.307215 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.307135 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 21 10:03:56.308285 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.307708 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-ttg7n\"" Apr 21 10:03:56.308285 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.307727 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 21 10:03:56.308285 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.307974 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 21 10:03:56.308285 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.308021 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 21 10:03:56.308285 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.308109 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 21 10:03:56.309515 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.308624 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-855nf\"" Apr 21 10:03:56.310509 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.310233 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.311173 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.310319 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.312131 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.312111 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.313323 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.313301 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.314114 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.313927 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 21 10:03:56.314114 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.313968 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-w5c75\"" Apr 21 10:03:56.314114 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.313982 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:03:56.314114 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.313995 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:03:56.314114 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.314093 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-l622j\"" Apr 21 10:03:56.314425 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.314234 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 21 10:03:56.314425 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.314334 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 21 10:03:56.314630 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.314597 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 21 10:03:56.314715 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.314700 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-pgq9t\"" Apr 21 10:03:56.314804 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.314785 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 21 10:03:56.314897 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.314790 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 21 10:03:56.315197 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315180 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:56.315347 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:56.315313 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:03:56.315423 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315387 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 21 10:03:56.315687 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315594 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-systemd\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.315687 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315622 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-registration-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.315687 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315649 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-lib-modules\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.315687 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315682 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-sys-fs\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.315926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315703 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-os-release\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.315926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315719 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-etc-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.315926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315733 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-cni-netd\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.315926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315755 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-kubelet-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.315926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315774 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqrvs\" (UniqueName: \"kubernetes.io/projected/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-kube-api-access-hqrvs\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.315926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315846 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/924092ef-7a63-4db6-afcd-c4fb3be74aec-konnectivity-ca\") pod \"konnectivity-agent-pmfn6\" (UID: \"924092ef-7a63-4db6-afcd-c4fb3be74aec\") " pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:03:56.315926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315878 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-kubernetes\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.315926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315907 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-host\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315940 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-systemd-units\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315967 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysctl-conf\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.315985 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovnkube-script-lib\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316003 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316025 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316064 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x29kw\" (UniqueName: \"kubernetes.io/projected/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-kube-api-access-x29kw\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316093 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-system-cni-dir\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316116 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316138 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-run\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316159 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clglp\" (UniqueName: \"kubernetes.io/projected/6d247fd6-2379-4000-acc6-743c1e73fbf9-kube-api-access-clglp\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316185 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-cni-bin\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316206 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316223 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf5hw\" (UniqueName: \"kubernetes.io/projected/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-kube-api-access-wf5hw\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316246 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-etc-selinux\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316285 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-cnibin\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.316325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316322 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-device-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316347 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-iptables-alerter-script\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316361 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-var-lib-kubelet\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316375 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-systemd\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316496 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316497 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovnkube-config\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:56.316547 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316570 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-env-overrides\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316594 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-cni-binary-copy\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316636 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-modprobe-d\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316703 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-run-netns\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316732 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316747 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316771 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-socket-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316796 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysconfig\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316817 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-kubelet\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316831 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c500689-bb9d-4394-869d-32c31072db85-hosts-file\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.316913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316850 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6d247fd6-2379-4000-acc6-743c1e73fbf9-tmp\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316870 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-ovn\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316886 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovn-node-metrics-cert\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316899 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-var-lib-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316911 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-node-log\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316934 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-log-socket\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.316971 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkfdg\" (UniqueName: \"kubernetes.io/projected/7c500689-bb9d-4394-869d-32c31072db85-kube-api-access-jkfdg\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317016 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-host-slash\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317047 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/924092ef-7a63-4db6-afcd-c4fb3be74aec-agent-certs\") pod \"konnectivity-agent-pmfn6\" (UID: \"924092ef-7a63-4db6-afcd-c4fb3be74aec\") " pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317091 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9c88\" (UniqueName: \"kubernetes.io/projected/a458360c-0ec8-4ac8-9524-f44a7102175d-kube-api-access-l9c88\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317135 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-tuned\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317162 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-slash\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317185 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7c500689-bb9d-4394-869d-32c31072db85-tmp-dir\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317199 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-sys\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317232 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysctl-d\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.317589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.317446 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-wz6ln\"" Apr 21 10:03:56.345065 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.345041 2570 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-20 09:58:55 +0000 UTC" deadline="2027-10-26 02:26:47.676322644 +0000 UTC" Apr 21 10:03:56.345065 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.345062 2570 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="13264h22m51.331262349s" Apr 21 10:03:56.408201 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.408182 2570 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 21 10:03:56.417483 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417460 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-slash\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.417595 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417488 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7c500689-bb9d-4394-869d-32c31072db85-tmp-dir\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.417595 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417504 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-sys\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.417595 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417532 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-netns\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.417595 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417560 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-slash\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.417595 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417560 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69hnp\" (UniqueName: \"kubernetes.io/projected/bd143ad0-d084-434c-87b4-bd355e2b205a-kube-api-access-69hnp\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:56.417595 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417596 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysctl-d\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417616 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-systemd\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417617 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-sys\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417637 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9f83c2fe-affe-48ec-a969-287e374d052c-multus-daemon-config\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417663 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-registration-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417682 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-lib-modules\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417690 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-systemd\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417702 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-system-cni-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417729 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-sys-fs\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417730 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-registration-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417752 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-os-release\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417778 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-socket-dir-parent\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417760 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysctl-d\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417795 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-sys-fs\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417808 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-etc-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417817 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/7c500689-bb9d-4394-869d-32c31072db85-tmp-dir\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417826 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-lib-modules\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.417861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417848 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-etc-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417843 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-cni-netd\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417870 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-os-release\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417876 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-kubelet-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417911 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hqrvs\" (UniqueName: \"kubernetes.io/projected/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-kube-api-access-hqrvs\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417914 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-kubelet-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417935 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/924092ef-7a63-4db6-afcd-c4fb3be74aec-konnectivity-ca\") pod \"konnectivity-agent-pmfn6\" (UID: \"924092ef-7a63-4db6-afcd-c4fb3be74aec\") " pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417968 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-kubernetes\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417979 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-cni-netd\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.417991 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-host\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418018 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9f83c2fe-affe-48ec-a969-287e374d052c-cni-binary-copy\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418035 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-kubernetes\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418044 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-systemd-units\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418082 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-host\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418101 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysctl-conf\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418128 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-systemd-units\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418128 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lngtb\" (UniqueName: \"kubernetes.io/projected/2ae59855-346f-4974-9fe0-7f7658ee240f-kube-api-access-lngtb\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.418635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418183 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovnkube-script-lib\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418233 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418238 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysctl-conf\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418290 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418325 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x29kw\" (UniqueName: \"kubernetes.io/projected/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-kube-api-access-x29kw\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418351 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-system-cni-dir\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418375 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418398 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-run\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418421 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-clglp\" (UniqueName: \"kubernetes.io/projected/6d247fd6-2379-4000-acc6-743c1e73fbf9-kube-api-access-clglp\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418448 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-cni-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418470 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-k8s-cni-cncf-io\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418497 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-cni-bin\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418522 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418541 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/924092ef-7a63-4db6-afcd-c4fb3be74aec-konnectivity-ca\") pod \"konnectivity-agent-pmfn6\" (UID: \"924092ef-7a63-4db6-afcd-c4fb3be74aec\") " pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418547 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wf5hw\" (UniqueName: \"kubernetes.io/projected/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-kube-api-access-wf5hw\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418616 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-etc-selinux\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.419426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418642 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-cnibin\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418669 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-kubelet\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418699 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-device-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418715 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-iptables-alerter-script\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418730 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-var-lib-kubelet\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418753 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-os-release\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418770 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418777 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-cni-bin\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418801 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2ae59855-346f-4974-9fe0-7f7658ee240f-host\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418822 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2ae59855-346f-4974-9fe0-7f7658ee240f-serviceca\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418838 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-device-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418846 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-systemd\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418871 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovnkube-config\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418887 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-etc-selinux\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418898 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-env-overrides\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418924 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-cni-binary-copy\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418925 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-cnibin\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.420167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418959 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-modprobe-d\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418989 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-conf-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.418998 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419015 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419045 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-run-netns\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419071 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419101 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419126 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-socket-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419150 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysconfig\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419177 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419204 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-kubelet\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419228 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c500689-bb9d-4394-869d-32c31072db85-hosts-file\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419251 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6d247fd6-2379-4000-acc6-743c1e73fbf9-tmp\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419298 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-hostroot\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419326 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-multus-certs\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419352 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-etc-kubernetes\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419379 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-ovn\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.420949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419404 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovn-node-metrics-cert\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419432 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-cni-multus\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419440 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-iptables-alerter-script\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419459 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qvf4\" (UniqueName: \"kubernetes.io/projected/9f83c2fe-affe-48ec-a969-287e374d052c-kube-api-access-2qvf4\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419487 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-var-lib-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419522 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-node-log\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419537 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-var-lib-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419550 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-log-socket\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419561 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419577 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jkfdg\" (UniqueName: \"kubernetes.io/projected/7c500689-bb9d-4394-869d-32c31072db85-kube-api-access-jkfdg\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419603 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-host-slash\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419619 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-run\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419629 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/924092ef-7a63-4db6-afcd-c4fb3be74aec-agent-certs\") pod \"konnectivity-agent-pmfn6\" (UID: \"924092ef-7a63-4db6-afcd-c4fb3be74aec\") " pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419621 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-modprobe-d\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419658 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l9c88\" (UniqueName: \"kubernetes.io/projected/a458360c-0ec8-4ac8-9524-f44a7102175d-kube-api-access-l9c88\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419673 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-openvswitch\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419842 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.421676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419885 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-run-netns\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419908 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-tuned\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419939 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-cnibin\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419944 2570 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.419954 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a458360c-0ec8-4ac8-9524-f44a7102175d-system-cni-dir\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420005 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-socket-dir\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420223 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-sysconfig\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420227 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d247fd6-2379-4000-acc6-743c1e73fbf9-var-lib-kubelet\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420288 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-log-socket\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420294 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c500689-bb9d-4394-869d-32c31072db85-hosts-file\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420326 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-systemd\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420329 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-kubelet\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420355 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-node-log\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420383 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-host-slash\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420384 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-run-ovn\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420289 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420421 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-host-cni-bin\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.420845 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a458360c-0ec8-4ac8-9524-f44a7102175d-cni-binary-copy\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.422661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.421350 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-env-overrides\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.423487 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.422150 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovnkube-config\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.423487 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.422187 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovnkube-script-lib\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.423487 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.423348 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6d247fd6-2379-4000-acc6-743c1e73fbf9-tmp\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.423487 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.423422 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-ovn-node-metrics-cert\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.423743 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.423727 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/924092ef-7a63-4db6-afcd-c4fb3be74aec-agent-certs\") pod \"konnectivity-agent-pmfn6\" (UID: \"924092ef-7a63-4db6-afcd-c4fb3be74aec\") " pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:03:56.424515 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.424499 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/6d247fd6-2379-4000-acc6-743c1e73fbf9-etc-tuned\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.427184 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.427159 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqrvs\" (UniqueName: \"kubernetes.io/projected/3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a-kube-api-access-hqrvs\") pod \"iptables-alerter-lggw9\" (UID: \"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a\") " pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.431313 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.431257 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf5hw\" (UniqueName: \"kubernetes.io/projected/e59564ca-d2b4-41c4-a36e-b94f5fc6c40b-kube-api-access-wf5hw\") pod \"ovnkube-node-qbn7w\" (UID: \"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.431612 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.431555 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x29kw\" (UniqueName: \"kubernetes.io/projected/1c55181f-18ce-48a1-89da-e1ee2d4c5aa3-kube-api-access-x29kw\") pod \"aws-ebs-csi-driver-node-ggt8q\" (UID: \"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.431612 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.431555 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9c88\" (UniqueName: \"kubernetes.io/projected/a458360c-0ec8-4ac8-9524-f44a7102175d-kube-api-access-l9c88\") pod \"multus-additional-cni-plugins-rczjx\" (UID: \"a458360c-0ec8-4ac8-9524-f44a7102175d\") " pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.431720 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.431668 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkfdg\" (UniqueName: \"kubernetes.io/projected/7c500689-bb9d-4394-869d-32c31072db85-kube-api-access-jkfdg\") pod \"node-resolver-rrdxp\" (UID: \"7c500689-bb9d-4394-869d-32c31072db85\") " pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.433650 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.433631 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-clglp\" (UniqueName: \"kubernetes.io/projected/6d247fd6-2379-4000-acc6-743c1e73fbf9-kube-api-access-clglp\") pod \"tuned-ngjbx\" (UID: \"6d247fd6-2379-4000-acc6-743c1e73fbf9\") " pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.520439 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520415 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-conf-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520561 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520446 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:03:56.520561 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520470 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:56.520561 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520493 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-hostroot\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520561 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520510 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-conf-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520561 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520517 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-multus-certs\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520561 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520540 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-etc-kubernetes\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520563 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-hostroot\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520564 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-cni-multus\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:56.520596 2570 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520612 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-multus-certs\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520599 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2qvf4\" (UniqueName: \"kubernetes.io/projected/9f83c2fe-affe-48ec-a969-287e374d052c-kube-api-access-2qvf4\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:56.520664 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs podName:bd143ad0-d084-434c-87b4-bd355e2b205a nodeName:}" failed. No retries permitted until 2026-04-21 10:03:57.020640332 +0000 UTC m=+3.101856159 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs") pod "network-metrics-daemon-ff5dl" (UID: "bd143ad0-d084-434c-87b4-bd355e2b205a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520595 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-cni-multus\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520684 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-cnibin\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520623 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-etc-kubernetes\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520712 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-netns\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.520775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520737 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-69hnp\" (UniqueName: \"kubernetes.io/projected/bd143ad0-d084-434c-87b4-bd355e2b205a-kube-api-access-69hnp\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520776 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9f83c2fe-affe-48ec-a969-287e374d052c-multus-daemon-config\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520797 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-netns\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520808 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-system-cni-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520835 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-socket-dir-parent\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520865 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9f83c2fe-affe-48ec-a969-287e374d052c-cni-binary-copy\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520875 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-system-cni-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520778 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-cnibin\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520890 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lngtb\" (UniqueName: \"kubernetes.io/projected/2ae59855-346f-4974-9fe0-7f7658ee240f-kube-api-access-lngtb\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520926 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-cni-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520932 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-socket-dir-parent\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520954 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-k8s-cni-cncf-io\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521010 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-multus-cni-dir\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.520996 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-run-k8s-cni-cncf-io\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521041 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-kubelet\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521078 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-os-release\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521121 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-kubelet\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521153 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-cni-bin\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.521210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521150 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-os-release\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.522049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521183 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f83c2fe-affe-48ec-a969-287e374d052c-host-var-lib-cni-bin\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.522049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521172 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2ae59855-346f-4974-9fe0-7f7658ee240f-host\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.522049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521194 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2ae59855-346f-4974-9fe0-7f7658ee240f-host\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.522049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521220 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2ae59855-346f-4974-9fe0-7f7658ee240f-serviceca\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.522049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521330 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9f83c2fe-affe-48ec-a969-287e374d052c-multus-daemon-config\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.522049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521378 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9f83c2fe-affe-48ec-a969-287e374d052c-cni-binary-copy\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.522049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.521640 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2ae59855-346f-4974-9fe0-7f7658ee240f-serviceca\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.527193 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:56.527171 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:56.527193 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:56.527198 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:56.527378 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:56.527209 2570 projected.go:194] Error preparing data for projected volume kube-api-access-6nrcg for pod openshift-network-diagnostics/network-check-target-2j8xh: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:56.527378 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:56.527251 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg podName:d0ada080-47c9-4134-a37c-b3bd33e4eb38 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:57.027239066 +0000 UTC m=+3.108454876 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6nrcg" (UniqueName: "kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg") pod "network-check-target-2j8xh" (UID: "d0ada080-47c9-4134-a37c-b3bd33e4eb38") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:56.530717 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.530694 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qvf4\" (UniqueName: \"kubernetes.io/projected/9f83c2fe-affe-48ec-a969-287e374d052c-kube-api-access-2qvf4\") pod \"multus-jrkkg\" (UID: \"9f83c2fe-affe-48ec-a969-287e374d052c\") " pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.530969 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.530955 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lngtb\" (UniqueName: \"kubernetes.io/projected/2ae59855-346f-4974-9fe0-7f7658ee240f-kube-api-access-lngtb\") pod \"node-ca-77lqp\" (UID: \"2ae59855-346f-4974-9fe0-7f7658ee240f\") " pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.531033 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.530955 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-69hnp\" (UniqueName: \"kubernetes.io/projected/bd143ad0-d084-434c-87b4-bd355e2b205a-kube-api-access-69hnp\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:56.610413 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.610381 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:03:56.619202 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.619179 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:03:56.627052 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.627015 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" Apr 21 10:03:56.631921 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.631901 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rrdxp" Apr 21 10:03:56.637452 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.637432 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rczjx" Apr 21 10:03:56.644952 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.644934 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-lggw9" Apr 21 10:03:56.651501 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.651483 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" Apr 21 10:03:56.658062 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.658041 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-77lqp" Apr 21 10:03:56.664595 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.664578 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jrkkg" Apr 21 10:03:56.706663 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:56.706639 2570 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:57.001071 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.001030 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c55181f_18ce_48a1_89da_e1ee2d4c5aa3.slice/crio-f8fe744512902ca28059573da56dec556a2b84ac0e6261cfe15ac35afdc27c94 WatchSource:0}: Error finding container f8fe744512902ca28059573da56dec556a2b84ac0e6261cfe15ac35afdc27c94: Status 404 returned error can't find the container with id f8fe744512902ca28059573da56dec556a2b84ac0e6261cfe15ac35afdc27c94 Apr 21 10:03:57.002092 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.002065 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c500689_bb9d_4394_869d_32c31072db85.slice/crio-9fc561b1916c86f635de0e11dad33a806a01003590edfd33d18ca99a5c66640c WatchSource:0}: Error finding container 9fc561b1916c86f635de0e11dad33a806a01003590edfd33d18ca99a5c66640c: Status 404 returned error can't find the container with id 9fc561b1916c86f635de0e11dad33a806a01003590edfd33d18ca99a5c66640c Apr 21 10:03:57.003759 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.003443 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f83c2fe_affe_48ec_a969_287e374d052c.slice/crio-e4172fe661b6452ef82e70e87bc7efeb993472293ca0d9dcd961f0edbb372210 WatchSource:0}: Error finding container e4172fe661b6452ef82e70e87bc7efeb993472293ca0d9dcd961f0edbb372210: Status 404 returned error can't find the container with id e4172fe661b6452ef82e70e87bc7efeb993472293ca0d9dcd961f0edbb372210 Apr 21 10:03:57.006745 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.006714 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3392dd0f_93d6_42ee_a4cb_f9aeb2243a3a.slice/crio-7d94310dbddbb12f4c0dc821edd236972d017cf52143a5e9c60b51c2dbc9d5a7 WatchSource:0}: Error finding container 7d94310dbddbb12f4c0dc821edd236972d017cf52143a5e9c60b51c2dbc9d5a7: Status 404 returned error can't find the container with id 7d94310dbddbb12f4c0dc821edd236972d017cf52143a5e9c60b51c2dbc9d5a7 Apr 21 10:03:57.007369 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.007345 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d247fd6_2379_4000_acc6_743c1e73fbf9.slice/crio-a90e9b87f5749d5f621a0e9d40cb096e3759ff7453cc69b74caf5f05dfcc7544 WatchSource:0}: Error finding container a90e9b87f5749d5f621a0e9d40cb096e3759ff7453cc69b74caf5f05dfcc7544: Status 404 returned error can't find the container with id a90e9b87f5749d5f621a0e9d40cb096e3759ff7453cc69b74caf5f05dfcc7544 Apr 21 10:03:57.009705 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.008725 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda458360c_0ec8_4ac8_9524_f44a7102175d.slice/crio-d43a80fa8ca4a56403b86a2f639118a22fb4a19d19e82e4055f79cb6b2d0a8ef WatchSource:0}: Error finding container d43a80fa8ca4a56403b86a2f639118a22fb4a19d19e82e4055f79cb6b2d0a8ef: Status 404 returned error can't find the container with id d43a80fa8ca4a56403b86a2f639118a22fb4a19d19e82e4055f79cb6b2d0a8ef Apr 21 10:03:57.009705 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.009219 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode59564ca_d2b4_41c4_a36e_b94f5fc6c40b.slice/crio-0a5283a8af1c9c2b0c5bd3fb141282749d709f9e7f613ab6e9d3ef43cc1d149f WatchSource:0}: Error finding container 0a5283a8af1c9c2b0c5bd3fb141282749d709f9e7f613ab6e9d3ef43cc1d149f: Status 404 returned error can't find the container with id 0a5283a8af1c9c2b0c5bd3fb141282749d709f9e7f613ab6e9d3ef43cc1d149f Apr 21 10:03:57.010294 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.010062 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod924092ef_7a63_4db6_afcd_c4fb3be74aec.slice/crio-3ba3333d7e864e51705ad9dae98381435349df984e659cc05a662b240658e68a WatchSource:0}: Error finding container 3ba3333d7e864e51705ad9dae98381435349df984e659cc05a662b240658e68a: Status 404 returned error can't find the container with id 3ba3333d7e864e51705ad9dae98381435349df984e659cc05a662b240658e68a Apr 21 10:03:57.011819 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:03:57.011586 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ae59855_346f_4974_9fe0_7f7658ee240f.slice/crio-5565dbfd40a37485fe970cf37a8c80025ccc231c5946430babd285c19420c2e8 WatchSource:0}: Error finding container 5565dbfd40a37485fe970cf37a8c80025ccc231c5946430babd285c19420c2e8: Status 404 returned error can't find the container with id 5565dbfd40a37485fe970cf37a8c80025ccc231c5946430babd285c19420c2e8 Apr 21 10:03:57.025543 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.025523 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:57.025642 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:57.025624 2570 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:57.025711 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:57.025690 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs podName:bd143ad0-d084-434c-87b4-bd355e2b205a nodeName:}" failed. No retries permitted until 2026-04-21 10:03:58.025670895 +0000 UTC m=+4.106886708 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs") pod "network-metrics-daemon-ff5dl" (UID: "bd143ad0-d084-434c-87b4-bd355e2b205a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:57.126403 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.126378 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:03:57.126517 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:57.126485 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:57.126517 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:57.126498 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:57.126517 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:57.126506 2570 projected.go:194] Error preparing data for projected volume kube-api-access-6nrcg for pod openshift-network-diagnostics/network-check-target-2j8xh: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:57.126606 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:57.126547 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg podName:d0ada080-47c9-4134-a37c-b3bd33e4eb38 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:58.126534514 +0000 UTC m=+4.207750325 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-6nrcg" (UniqueName: "kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg") pod "network-check-target-2j8xh" (UID: "d0ada080-47c9-4134-a37c-b3bd33e4eb38") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:57.345681 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.345608 2570 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-20 09:58:55 +0000 UTC" deadline="2027-11-19 12:24:29.098927543 +0000 UTC" Apr 21 10:03:57.345681 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.345645 2570 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="13850h20m31.753286378s" Apr 21 10:03:57.474708 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.474664 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" event={"ID":"b2386108552c5c182289de71222968c1","Type":"ContainerStarted","Data":"c160eff98638b67dd541507eaf77ad290fd13e2fe31aa28697a2bb886300514a"} Apr 21 10:03:57.485719 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.485085 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-pmfn6" event={"ID":"924092ef-7a63-4db6-afcd-c4fb3be74aec","Type":"ContainerStarted","Data":"3ba3333d7e864e51705ad9dae98381435349df984e659cc05a662b240658e68a"} Apr 21 10:03:57.487840 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.487771 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"0a5283a8af1c9c2b0c5bd3fb141282749d709f9e7f613ab6e9d3ef43cc1d149f"} Apr 21 10:03:57.495174 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.495129 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" event={"ID":"6d247fd6-2379-4000-acc6-743c1e73fbf9","Type":"ContainerStarted","Data":"a90e9b87f5749d5f621a0e9d40cb096e3759ff7453cc69b74caf5f05dfcc7544"} Apr 21 10:03:57.501544 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.501521 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-lggw9" event={"ID":"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a","Type":"ContainerStarted","Data":"7d94310dbddbb12f4c0dc821edd236972d017cf52143a5e9c60b51c2dbc9d5a7"} Apr 21 10:03:57.505203 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.505172 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" event={"ID":"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3","Type":"ContainerStarted","Data":"f8fe744512902ca28059573da56dec556a2b84ac0e6261cfe15ac35afdc27c94"} Apr 21 10:03:57.507679 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.507643 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-77lqp" event={"ID":"2ae59855-346f-4974-9fe0-7f7658ee240f","Type":"ContainerStarted","Data":"5565dbfd40a37485fe970cf37a8c80025ccc231c5946430babd285c19420c2e8"} Apr 21 10:03:57.518926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.518881 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rczjx" event={"ID":"a458360c-0ec8-4ac8-9524-f44a7102175d","Type":"ContainerStarted","Data":"d43a80fa8ca4a56403b86a2f639118a22fb4a19d19e82e4055f79cb6b2d0a8ef"} Apr 21 10:03:57.525133 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.525107 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jrkkg" event={"ID":"9f83c2fe-affe-48ec-a969-287e374d052c","Type":"ContainerStarted","Data":"e4172fe661b6452ef82e70e87bc7efeb993472293ca0d9dcd961f0edbb372210"} Apr 21 10:03:57.530323 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:57.530187 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rrdxp" event={"ID":"7c500689-bb9d-4394-869d-32c31072db85","Type":"ContainerStarted","Data":"9fc561b1916c86f635de0e11dad33a806a01003590edfd33d18ca99a5c66640c"} Apr 21 10:03:58.037490 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:58.036860 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:58.037490 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:58.036993 2570 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:58.037490 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:58.037053 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs podName:bd143ad0-d084-434c-87b4-bd355e2b205a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:00.037035751 +0000 UTC m=+6.118251568 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs") pod "network-metrics-daemon-ff5dl" (UID: "bd143ad0-d084-434c-87b4-bd355e2b205a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:58.138827 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:58.138131 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:03:58.138827 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:58.138346 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:58.138827 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:58.138368 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:58.138827 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:58.138381 2570 projected.go:194] Error preparing data for projected volume kube-api-access-6nrcg for pod openshift-network-diagnostics/network-check-target-2j8xh: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:58.138827 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:58.138442 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg podName:d0ada080-47c9-4134-a37c-b3bd33e4eb38 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:00.138423393 +0000 UTC m=+6.219639209 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-6nrcg" (UniqueName: "kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg") pod "network-check-target-2j8xh" (UID: "d0ada080-47c9-4134-a37c-b3bd33e4eb38") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:58.461361 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:58.460407 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:03:58.461361 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:58.460534 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:03:58.461361 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:58.460979 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:03:58.461361 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:03:58.461087 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:03:58.538940 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:58.538517 2570 generic.go:358] "Generic (PLEG): container finished" podID="c1401eb7a080b0c324e5cdb3dfc2f894" containerID="8ea1f3970e17334fdc9a29145c3501ddb7dbaadec0b437a4cc2b61c95e064345" exitCode=0 Apr 21 10:03:58.543211 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:58.542799 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" event={"ID":"c1401eb7a080b0c324e5cdb3dfc2f894","Type":"ContainerDied","Data":"8ea1f3970e17334fdc9a29145c3501ddb7dbaadec0b437a4cc2b61c95e064345"} Apr 21 10:03:58.556741 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:58.556695 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-243.ec2.internal" podStartSLOduration=3.556681472 podStartE2EDuration="3.556681472s" podCreationTimestamp="2026-04-21 10:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:03:57.491154989 +0000 UTC m=+3.572370822" watchObservedRunningTime="2026-04-21 10:03:58.556681472 +0000 UTC m=+4.637897305" Apr 21 10:03:59.556341 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:59.556220 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" event={"ID":"c1401eb7a080b0c324e5cdb3dfc2f894","Type":"ContainerStarted","Data":"1e21a6fa9b685e1036d53d317ad6fc6b1121303b62b8961eada25367483c0688"} Apr 21 10:03:59.572469 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:03:59.571879 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-243.ec2.internal" podStartSLOduration=4.57186291 podStartE2EDuration="4.57186291s" podCreationTimestamp="2026-04-21 10:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:03:59.571461722 +0000 UTC m=+5.652677569" watchObservedRunningTime="2026-04-21 10:03:59.57186291 +0000 UTC m=+5.653078744" Apr 21 10:04:00.053968 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:00.053920 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:00.054151 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:00.054076 2570 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:00.054151 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:00.054139 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs podName:bd143ad0-d084-434c-87b4-bd355e2b205a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.05412012 +0000 UTC m=+10.135335934 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs") pod "network-metrics-daemon-ff5dl" (UID: "bd143ad0-d084-434c-87b4-bd355e2b205a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:00.155328 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:00.155288 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:00.155503 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:00.155468 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:04:00.155503 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:00.155490 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:04:00.155503 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:00.155502 2570 projected.go:194] Error preparing data for projected volume kube-api-access-6nrcg for pod openshift-network-diagnostics/network-check-target-2j8xh: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:04:00.155647 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:00.155562 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg podName:d0ada080-47c9-4134-a37c-b3bd33e4eb38 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.155544189 +0000 UTC m=+10.236760005 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-6nrcg" (UniqueName: "kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg") pod "network-check-target-2j8xh" (UID: "d0ada080-47c9-4134-a37c-b3bd33e4eb38") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:04:00.458649 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:00.458131 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:00.458649 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:00.458281 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:00.459140 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:00.458988 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:00.459140 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:00.459110 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:02.459088 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:02.458761 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:02.459088 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:02.458901 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:02.459088 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:02.458960 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:02.459608 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:02.459064 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:04.089652 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:04.089611 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:04.090065 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:04.089755 2570 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:04.090065 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:04.089819 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs podName:bd143ad0-d084-434c-87b4-bd355e2b205a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:12.089800726 +0000 UTC m=+18.171016535 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs") pod "network-metrics-daemon-ff5dl" (UID: "bd143ad0-d084-434c-87b4-bd355e2b205a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:04.190638 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:04.190596 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:04.190797 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:04.190775 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:04:04.190797 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:04.190794 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:04:04.190890 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:04.190808 2570 projected.go:194] Error preparing data for projected volume kube-api-access-6nrcg for pod openshift-network-diagnostics/network-check-target-2j8xh: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:04:04.190890 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:04.190862 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg podName:d0ada080-47c9-4134-a37c-b3bd33e4eb38 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:12.190845057 +0000 UTC m=+18.272060869 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-6nrcg" (UniqueName: "kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg") pod "network-check-target-2j8xh" (UID: "d0ada080-47c9-4134-a37c-b3bd33e4eb38") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:04:04.459154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:04.459127 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:04.459344 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:04.459203 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:04.459433 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:04.459354 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:04.459514 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:04.459487 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:06.458927 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:06.458893 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:06.459388 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:06.459020 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:06.459511 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:06.459492 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:06.459625 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:06.459606 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:08.458570 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:08.458532 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:08.458570 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:08.458570 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:08.459079 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:08.458665 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:08.459079 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:08.458797 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:10.458822 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:10.458789 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:10.459206 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:10.458789 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:10.459206 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:10.458932 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:10.459206 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:10.459005 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:12.151392 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:12.151355 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:12.151873 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:12.151543 2570 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:12.151873 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:12.151618 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs podName:bd143ad0-d084-434c-87b4-bd355e2b205a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:28.151591913 +0000 UTC m=+34.232807723 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs") pod "network-metrics-daemon-ff5dl" (UID: "bd143ad0-d084-434c-87b4-bd355e2b205a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:12.251966 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:12.251939 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:12.252678 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:12.252630 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:04:12.252678 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:12.252675 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:04:12.252871 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:12.252695 2570 projected.go:194] Error preparing data for projected volume kube-api-access-6nrcg for pod openshift-network-diagnostics/network-check-target-2j8xh: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:04:12.252871 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:12.252837 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg podName:d0ada080-47c9-4134-a37c-b3bd33e4eb38 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:28.252813158 +0000 UTC m=+34.334029284 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-6nrcg" (UniqueName: "kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg") pod "network-check-target-2j8xh" (UID: "d0ada080-47c9-4134-a37c-b3bd33e4eb38") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:04:12.458753 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:12.458663 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:12.458893 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:12.458663 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:12.458893 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:12.458811 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:12.458893 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:12.458856 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:14.458734 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:14.458698 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:14.459172 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:14.458775 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:14.459172 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:14.458791 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:14.459172 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:14.458890 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:15.580702 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.580120 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-pmfn6" event={"ID":"924092ef-7a63-4db6-afcd-c4fb3be74aec","Type":"ContainerStarted","Data":"6584130fbc830b1f324f0de40ad6171d97be5c87663f8bb98e15ebeb9333f789"} Apr 21 10:04:15.584951 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.584926 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"aaa31b95bc5c292719ee3634b98c0caddfdc9e0469e3e8d73b4e06cf49ab2119"} Apr 21 10:04:15.585082 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.584957 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"926430ab1843acb5d7f5eae7552ec8bf3a7393bb417c425e6a3c77067f1b6036"} Apr 21 10:04:15.585082 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.584970 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"fff5f95fb67b1f87fe32059b3fdbfed2663bae1301d1a8dc56451a97df4cca91"} Apr 21 10:04:15.585082 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.584985 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"b1c450223f0b215b970384f745653c5bec31114cd8d99b2c9c8953566cb4eab5"} Apr 21 10:04:15.585082 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.584996 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"bdcba11ebbb8754ea74850b39e0df20f6baebe5dff1d9ea9dfad9274c44f4df8"} Apr 21 10:04:15.585082 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.585007 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"97956bfd5a91272ed006a88b4d4267ef304bd0ffb436572869d83b561fd3265c"} Apr 21 10:04:15.586547 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.586501 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" event={"ID":"6d247fd6-2379-4000-acc6-743c1e73fbf9","Type":"ContainerStarted","Data":"ecc090fbab55e85594f8464171fcd0d77820ba87a129371bcb2553e40a7b9488"} Apr 21 10:04:15.587965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.587929 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" event={"ID":"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3","Type":"ContainerStarted","Data":"28b54c7f39f07b0362668efc04fe92c5f446a209b0394758bb2d32bd06af7cc3"} Apr 21 10:04:15.589173 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.589151 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-77lqp" event={"ID":"2ae59855-346f-4974-9fe0-7f7658ee240f","Type":"ContainerStarted","Data":"1a1e75a775268c30d57a889bbdae9a34300f0236ba041989e36c96f1016ec694"} Apr 21 10:04:15.590581 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.590555 2570 generic.go:358] "Generic (PLEG): container finished" podID="a458360c-0ec8-4ac8-9524-f44a7102175d" containerID="907be68112bb0dd68254e93f826fadf32daa0428ef2dffa4a4d68462eb73d8bf" exitCode=0 Apr 21 10:04:15.590660 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.590593 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rczjx" event={"ID":"a458360c-0ec8-4ac8-9524-f44a7102175d","Type":"ContainerDied","Data":"907be68112bb0dd68254e93f826fadf32daa0428ef2dffa4a4d68462eb73d8bf"} Apr 21 10:04:15.591836 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.591815 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jrkkg" event={"ID":"9f83c2fe-affe-48ec-a969-287e374d052c","Type":"ContainerStarted","Data":"65feb2eda275a16d679eeac81f06fe20e5b49e08b4b9e3935abaacc6bfd30f7a"} Apr 21 10:04:15.593089 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.593068 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rrdxp" event={"ID":"7c500689-bb9d-4394-869d-32c31072db85","Type":"ContainerStarted","Data":"55b14423ca85c695e156ed87068645a54abb1a5c19beb1fd2b2f56cdb811427e"} Apr 21 10:04:15.595585 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.595524 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-pmfn6" podStartSLOduration=3.923959109 podStartE2EDuration="21.595514112s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.011887358 +0000 UTC m=+3.093103185" lastFinishedPulling="2026-04-21 10:04:14.683442378 +0000 UTC m=+20.764658188" observedRunningTime="2026-04-21 10:04:15.595246938 +0000 UTC m=+21.676462790" watchObservedRunningTime="2026-04-21 10:04:15.595514112 +0000 UTC m=+21.676729943" Apr 21 10:04:15.607918 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.607874 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-rrdxp" podStartSLOduration=3.886471639 podStartE2EDuration="21.607862663s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.005261536 +0000 UTC m=+3.086477349" lastFinishedPulling="2026-04-21 10:04:14.726652556 +0000 UTC m=+20.807868373" observedRunningTime="2026-04-21 10:04:15.60763557 +0000 UTC m=+21.688851414" watchObservedRunningTime="2026-04-21 10:04:15.607862663 +0000 UTC m=+21.689078495" Apr 21 10:04:15.620547 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.620506 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-77lqp" podStartSLOduration=3.90733693 podStartE2EDuration="21.620492708s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.013488317 +0000 UTC m=+3.094704132" lastFinishedPulling="2026-04-21 10:04:14.726644086 +0000 UTC m=+20.807859910" observedRunningTime="2026-04-21 10:04:15.620103417 +0000 UTC m=+21.701319250" watchObservedRunningTime="2026-04-21 10:04:15.620492708 +0000 UTC m=+21.701708540" Apr 21 10:04:15.638560 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.638524 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jrkkg" podStartSLOduration=3.716839385 podStartE2EDuration="21.638513775s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.005283584 +0000 UTC m=+3.086499394" lastFinishedPulling="2026-04-21 10:04:14.926957972 +0000 UTC m=+21.008173784" observedRunningTime="2026-04-21 10:04:15.636492366 +0000 UTC m=+21.717708198" watchObservedRunningTime="2026-04-21 10:04:15.638513775 +0000 UTC m=+21.719729607" Apr 21 10:04:15.675765 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.675721 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-ngjbx" podStartSLOduration=3.957044727 podStartE2EDuration="21.675709031s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.009812958 +0000 UTC m=+3.091028774" lastFinishedPulling="2026-04-21 10:04:14.728477268 +0000 UTC m=+20.809693078" observedRunningTime="2026-04-21 10:04:15.675237867 +0000 UTC m=+21.756453710" watchObservedRunningTime="2026-04-21 10:04:15.675709031 +0000 UTC m=+21.756924897" Apr 21 10:04:15.857883 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:15.857764 2570 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 21 10:04:16.384153 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:16.383996 2570 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-21T10:04:15.857879638Z","UUID":"c38e90fb-9011-4b75-b029-c5f5df4fc4f0","Handler":null,"Name":"","Endpoint":""} Apr 21 10:04:16.386156 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:16.386064 2570 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 21 10:04:16.386156 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:16.386096 2570 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 21 10:04:16.457859 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:16.457835 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:16.457990 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:16.457966 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:16.458186 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:16.458165 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:16.458325 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:16.458303 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:16.596577 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:16.596544 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" event={"ID":"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3","Type":"ContainerStarted","Data":"465326d06e4738b3442ffacc4928bc4f652c4e932f6cb770f47638613a2c5a9a"} Apr 21 10:04:16.599010 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:16.598988 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-lggw9" event={"ID":"3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a","Type":"ContainerStarted","Data":"5b0e330c9fd71eb24337377152a15be93ac32fb84d85e207835c3e8fbaaf9bf2"} Apr 21 10:04:16.624367 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:16.624309 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-lggw9" podStartSLOduration=4.900440499 podStartE2EDuration="22.624291924s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.008209641 +0000 UTC m=+3.089425452" lastFinishedPulling="2026-04-21 10:04:14.732061067 +0000 UTC m=+20.813276877" observedRunningTime="2026-04-21 10:04:16.623601544 +0000 UTC m=+22.704817376" watchObservedRunningTime="2026-04-21 10:04:16.624291924 +0000 UTC m=+22.705507756" Apr 21 10:04:17.603949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:17.603910 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"ccad79462015e522354d549c34b5eb595993e31bca8e895f63aa073dc1d00077"} Apr 21 10:04:17.605973 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:17.605933 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" event={"ID":"1c55181f-18ce-48a1-89da-e1ee2d4c5aa3","Type":"ContainerStarted","Data":"8dd7df307785bb313c50a68eb80ed71843e10f83cfec2a2c17b9de50224491d6"} Apr 21 10:04:17.980334 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:17.980297 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:04:17.980919 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:17.980887 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:04:18.018582 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:18.018534 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ggt8q" podStartSLOduration=4.18305169 podStartE2EDuration="24.0185184s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.002634663 +0000 UTC m=+3.083850473" lastFinishedPulling="2026-04-21 10:04:16.838101362 +0000 UTC m=+22.919317183" observedRunningTime="2026-04-21 10:04:17.634977766 +0000 UTC m=+23.716193598" watchObservedRunningTime="2026-04-21 10:04:18.0185184 +0000 UTC m=+24.099734232" Apr 21 10:04:18.458339 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:18.458303 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:18.458339 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:18.458329 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:18.458573 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:18.458431 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:18.458626 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:18.458583 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:18.607975 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:18.607938 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:04:18.608594 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:18.608571 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-pmfn6" Apr 21 10:04:20.458712 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.458447 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:20.459200 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.458451 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:20.459200 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:20.458805 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:20.459200 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:20.458886 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:20.612521 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.612495 2570 generic.go:358] "Generic (PLEG): container finished" podID="a458360c-0ec8-4ac8-9524-f44a7102175d" containerID="1a3dea7624cb42a6d37b29025b39a66c76bf36e8e790b95538f993acd2826850" exitCode=0 Apr 21 10:04:20.612636 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.612567 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rczjx" event={"ID":"a458360c-0ec8-4ac8-9524-f44a7102175d","Type":"ContainerDied","Data":"1a3dea7624cb42a6d37b29025b39a66c76bf36e8e790b95538f993acd2826850"} Apr 21 10:04:20.615919 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.615898 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" event={"ID":"e59564ca-d2b4-41c4-a36e-b94f5fc6c40b","Type":"ContainerStarted","Data":"5236fe658092434263583b1a6433469bd76ba2dbd2839d25848952b9f9b8a7a1"} Apr 21 10:04:20.616370 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.616352 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:04:20.616497 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.616376 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:04:20.616497 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.616388 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:04:20.631432 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.631412 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:04:20.631529 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.631491 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:04:20.660260 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:20.660224 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" podStartSLOduration=8.774697436 podStartE2EDuration="26.660213094s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.011110526 +0000 UTC m=+3.092326344" lastFinishedPulling="2026-04-21 10:04:14.896626172 +0000 UTC m=+20.977842002" observedRunningTime="2026-04-21 10:04:20.659936102 +0000 UTC m=+26.741151933" watchObservedRunningTime="2026-04-21 10:04:20.660213094 +0000 UTC m=+26.741428926" Apr 21 10:04:21.066623 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:21.066600 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-rrdxp_7c500689-bb9d-4394-869d-32c31072db85/dns-node-resolver/0.log" Apr 21 10:04:21.547318 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:21.547135 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-2j8xh"] Apr 21 10:04:21.547750 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:21.547399 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:21.547750 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:21.547476 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:21.555076 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:21.555052 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ff5dl"] Apr 21 10:04:21.555189 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:21.555175 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:21.555320 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:21.555302 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:21.619111 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:21.619086 2570 generic.go:358] "Generic (PLEG): container finished" podID="a458360c-0ec8-4ac8-9524-f44a7102175d" containerID="cb11a5a83a239bdcb8301361b4b32efdda83ca3405f2bf626c2e9497fee32c2b" exitCode=0 Apr 21 10:04:21.619212 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:21.619180 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rczjx" event={"ID":"a458360c-0ec8-4ac8-9524-f44a7102175d","Type":"ContainerDied","Data":"cb11a5a83a239bdcb8301361b4b32efdda83ca3405f2bf626c2e9497fee32c2b"} Apr 21 10:04:21.646945 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:21.646925 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-77lqp_2ae59855-346f-4974-9fe0-7f7658ee240f/node-ca/0.log" Apr 21 10:04:22.622700 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:22.622670 2570 generic.go:358] "Generic (PLEG): container finished" podID="a458360c-0ec8-4ac8-9524-f44a7102175d" containerID="96683c058c24839fde3905c244e5d7e518f94b97b1b460e68d3515a7a18c7b52" exitCode=0 Apr 21 10:04:22.623123 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:22.622752 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rczjx" event={"ID":"a458360c-0ec8-4ac8-9524-f44a7102175d","Type":"ContainerDied","Data":"96683c058c24839fde3905c244e5d7e518f94b97b1b460e68d3515a7a18c7b52"} Apr 21 10:04:23.458643 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:23.458614 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:23.458643 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:23.458640 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:23.458859 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:23.458743 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:23.458859 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:23.458828 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:25.458525 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:25.458497 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:25.459066 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:25.458501 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:25.459066 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:25.458616 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:25.459066 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:25.458686 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:27.458013 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:27.457971 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:27.458013 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:27.457987 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:27.458472 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:27.458083 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:27.458472 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:27.458241 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:28.175675 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:28.175492 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:28.175884 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:28.175627 2570 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:28.175884 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:28.175750 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs podName:bd143ad0-d084-434c-87b4-bd355e2b205a nodeName:}" failed. No retries permitted until 2026-04-21 10:05:00.175735355 +0000 UTC m=+66.256951165 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs") pod "network-metrics-daemon-ff5dl" (UID: "bd143ad0-d084-434c-87b4-bd355e2b205a") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:28.276225 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:28.276199 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:28.276356 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:28.276330 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:04:28.276356 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:28.276342 2570 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:04:28.276356 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:28.276350 2570 projected.go:194] Error preparing data for projected volume kube-api-access-6nrcg for pod openshift-network-diagnostics/network-check-target-2j8xh: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:04:28.276456 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:28.276391 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg podName:d0ada080-47c9-4134-a37c-b3bd33e4eb38 nodeName:}" failed. No retries permitted until 2026-04-21 10:05:00.276379935 +0000 UTC m=+66.357595744 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-6nrcg" (UniqueName: "kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg") pod "network-check-target-2j8xh" (UID: "d0ada080-47c9-4134-a37c-b3bd33e4eb38") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:04:29.458465 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:29.458433 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:29.458906 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:29.458440 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:29.458906 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:29.458539 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:29.458906 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:29.458598 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:29.639307 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:29.639258 2570 generic.go:358] "Generic (PLEG): container finished" podID="a458360c-0ec8-4ac8-9524-f44a7102175d" containerID="2021c4f7b635081d77ce15eb44be8da2aaf2b24782cc791b4307192f743c83b5" exitCode=0 Apr 21 10:04:29.639451 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:29.639321 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rczjx" event={"ID":"a458360c-0ec8-4ac8-9524-f44a7102175d","Type":"ContainerDied","Data":"2021c4f7b635081d77ce15eb44be8da2aaf2b24782cc791b4307192f743c83b5"} Apr 21 10:04:30.643526 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:30.643495 2570 generic.go:358] "Generic (PLEG): container finished" podID="a458360c-0ec8-4ac8-9524-f44a7102175d" containerID="bc150e42faf80fd7b92c45de7f31136fd87454d1ec822706cfe25f86f3f518db" exitCode=0 Apr 21 10:04:30.643871 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:30.643539 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rczjx" event={"ID":"a458360c-0ec8-4ac8-9524-f44a7102175d","Type":"ContainerDied","Data":"bc150e42faf80fd7b92c45de7f31136fd87454d1ec822706cfe25f86f3f518db"} Apr 21 10:04:31.458330 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:31.458297 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:31.458572 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:31.458393 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:31.458572 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:31.458421 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:31.458572 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:31.458478 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:31.648427 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:31.648389 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rczjx" event={"ID":"a458360c-0ec8-4ac8-9524-f44a7102175d","Type":"ContainerStarted","Data":"1ec9ece0517584850de38b8d1e3fe5a02ffce2810afb9d2824e2fa4692ea5992"} Apr 21 10:04:31.672863 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:31.672815 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rczjx" podStartSLOduration=6.061270235 podStartE2EDuration="37.67280039s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:03:57.010299866 +0000 UTC m=+3.091515689" lastFinishedPulling="2026-04-21 10:04:28.621830033 +0000 UTC m=+34.703045844" observedRunningTime="2026-04-21 10:04:31.671487453 +0000 UTC m=+37.752703284" watchObservedRunningTime="2026-04-21 10:04:31.67280039 +0000 UTC m=+37.754016200" Apr 21 10:04:33.458330 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:33.458304 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:33.458330 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:33.458337 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:33.458712 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:33.458415 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:33.458712 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:33.458552 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:35.458505 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:35.458472 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:35.458958 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:35.458605 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:35.458958 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:35.458657 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:35.458958 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:35.458747 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:37.458400 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:37.458372 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:37.458734 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:37.458434 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:37.458734 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:37.458529 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:37.458734 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:37.458583 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:39.457871 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:39.457839 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:39.457871 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:39.457855 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:39.458319 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:39.457938 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:39.458319 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:39.458071 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:41.458663 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:41.458635 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:41.459057 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:41.458678 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:41.459057 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:41.458749 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:41.459057 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:41.458796 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:43.458420 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.458395 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:43.458709 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.458395 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:43.458709 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:43.458498 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-2j8xh" podUID="d0ada080-47c9-4134-a37c-b3bd33e4eb38" Apr 21 10:04:43.458709 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:43.458587 2570 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ff5dl" podUID="bd143ad0-d084-434c-87b4-bd355e2b205a" Apr 21 10:04:43.728668 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.728605 2570 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-243.ec2.internal" event="NodeReady" Apr 21 10:04:43.728798 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.728699 2570 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 21 10:04:43.796494 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.796464 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-4fwj6"] Apr 21 10:04:43.810534 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.810513 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-ckmzt"] Apr 21 10:04:43.810683 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.810666 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.814813 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.814797 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-nglr4\"" Apr 21 10:04:43.814938 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.814923 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 21 10:04:43.815052 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.815036 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 21 10:04:43.822861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.822843 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.825815 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.825796 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 21 10:04:43.826370 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.826146 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 21 10:04:43.826561 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.826394 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 21 10:04:43.826917 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.826764 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 21 10:04:43.827396 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.827153 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-l8m2h\"" Apr 21 10:04:43.828036 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.828010 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hlz49"] Apr 21 10:04:43.841576 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.841560 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4fwj6"] Apr 21 10:04:43.841576 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.841579 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-ckmzt"] Apr 21 10:04:43.841693 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.841587 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hlz49"] Apr 21 10:04:43.841693 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.841652 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hlz49" Apr 21 10:04:43.845666 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.845650 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 21 10:04:43.845782 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.845767 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-sjbvv\"" Apr 21 10:04:43.846110 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.846085 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 21 10:04:43.846198 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.846112 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 21 10:04:43.889343 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889325 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e51c2a85-13d4-47bb-916a-fb2ecece2c78-crio-socket\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.889451 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889351 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8683a07e-2776-4d73-9312-ca2168702c96-cert\") pod \"ingress-canary-hlz49\" (UID: \"8683a07e-2776-4d73-9312-ca2168702c96\") " pod="openshift-ingress-canary/ingress-canary-hlz49" Apr 21 10:04:43.889451 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889379 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-tmp-dir\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.889451 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889398 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzdrl\" (UniqueName: \"kubernetes.io/projected/8683a07e-2776-4d73-9312-ca2168702c96-kube-api-access-bzdrl\") pod \"ingress-canary-hlz49\" (UID: \"8683a07e-2776-4d73-9312-ca2168702c96\") " pod="openshift-ingress-canary/ingress-canary-hlz49" Apr 21 10:04:43.889451 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889435 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxb5v\" (UniqueName: \"kubernetes.io/projected/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-kube-api-access-mxb5v\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.889575 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889489 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e51c2a85-13d4-47bb-916a-fb2ecece2c78-data-volume\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.889575 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889507 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e51c2a85-13d4-47bb-916a-fb2ecece2c78-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.889575 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889523 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgtld\" (UniqueName: \"kubernetes.io/projected/e51c2a85-13d4-47bb-916a-fb2ecece2c78-kube-api-access-rgtld\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.889665 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889576 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-metrics-tls\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.889665 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889592 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e51c2a85-13d4-47bb-916a-fb2ecece2c78-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.889665 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.889613 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-config-volume\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.990898 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.990841 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bzdrl\" (UniqueName: \"kubernetes.io/projected/8683a07e-2776-4d73-9312-ca2168702c96-kube-api-access-bzdrl\") pod \"ingress-canary-hlz49\" (UID: \"8683a07e-2776-4d73-9312-ca2168702c96\") " pod="openshift-ingress-canary/ingress-canary-hlz49" Apr 21 10:04:43.990898 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.990870 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mxb5v\" (UniqueName: \"kubernetes.io/projected/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-kube-api-access-mxb5v\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.990898 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.990888 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e51c2a85-13d4-47bb-916a-fb2ecece2c78-data-volume\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.991117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991001 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e51c2a85-13d4-47bb-916a-fb2ecece2c78-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.991117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991029 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rgtld\" (UniqueName: \"kubernetes.io/projected/e51c2a85-13d4-47bb-916a-fb2ecece2c78-kube-api-access-rgtld\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.991117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991070 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-metrics-tls\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.991117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991088 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e51c2a85-13d4-47bb-916a-fb2ecece2c78-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.991117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991117 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-config-volume\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.991377 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991168 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e51c2a85-13d4-47bb-916a-fb2ecece2c78-crio-socket\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.991377 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991201 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8683a07e-2776-4d73-9312-ca2168702c96-cert\") pod \"ingress-canary-hlz49\" (UID: \"8683a07e-2776-4d73-9312-ca2168702c96\") " pod="openshift-ingress-canary/ingress-canary-hlz49" Apr 21 10:04:43.991377 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991226 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-tmp-dir\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.991377 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991347 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e51c2a85-13d4-47bb-916a-fb2ecece2c78-data-volume\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.991574 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991443 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e51c2a85-13d4-47bb-916a-fb2ecece2c78-crio-socket\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.991574 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991564 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-tmp-dir\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.991876 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.991843 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-config-volume\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.992046 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.992023 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e51c2a85-13d4-47bb-916a-fb2ecece2c78-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.995187 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.995162 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-metrics-tls\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.995293 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.995223 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8683a07e-2776-4d73-9312-ca2168702c96-cert\") pod \"ingress-canary-hlz49\" (UID: \"8683a07e-2776-4d73-9312-ca2168702c96\") " pod="openshift-ingress-canary/ingress-canary-hlz49" Apr 21 10:04:43.995293 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.995221 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e51c2a85-13d4-47bb-916a-fb2ecece2c78-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:43.999171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.999143 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxb5v\" (UniqueName: \"kubernetes.io/projected/da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997-kube-api-access-mxb5v\") pod \"dns-default-4fwj6\" (UID: \"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997\") " pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:43.999661 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:43.999640 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgtld\" (UniqueName: \"kubernetes.io/projected/e51c2a85-13d4-47bb-916a-fb2ecece2c78-kube-api-access-rgtld\") pod \"insights-runtime-extractor-ckmzt\" (UID: \"e51c2a85-13d4-47bb-916a-fb2ecece2c78\") " pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:44.000325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.000294 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzdrl\" (UniqueName: \"kubernetes.io/projected/8683a07e-2776-4d73-9312-ca2168702c96-kube-api-access-bzdrl\") pod \"ingress-canary-hlz49\" (UID: \"8683a07e-2776-4d73-9312-ca2168702c96\") " pod="openshift-ingress-canary/ingress-canary-hlz49" Apr 21 10:04:44.119282 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.119242 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:44.131928 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.131900 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-ckmzt" Apr 21 10:04:44.157698 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.152356 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hlz49" Apr 21 10:04:44.313743 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.313716 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-ckmzt"] Apr 21 10:04:44.322623 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.322600 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4fwj6"] Apr 21 10:04:44.325522 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:44.325497 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda8bd8dd_c9e0_4cbd_be5a_cf95d4d01997.slice/crio-d51183a605989e2894947554539fcb405c06108b970465f65d2cacaf9e04d088 WatchSource:0}: Error finding container d51183a605989e2894947554539fcb405c06108b970465f65d2cacaf9e04d088: Status 404 returned error can't find the container with id d51183a605989e2894947554539fcb405c06108b970465f65d2cacaf9e04d088 Apr 21 10:04:44.333945 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.333914 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hlz49"] Apr 21 10:04:44.336550 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:44.336530 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8683a07e_2776_4d73_9312_ca2168702c96.slice/crio-963b8512d4cc77210ac82782985e45ccbf3ad5ae84a49590c8faaf5d0f2fa1a3 WatchSource:0}: Error finding container 963b8512d4cc77210ac82782985e45ccbf3ad5ae84a49590c8faaf5d0f2fa1a3: Status 404 returned error can't find the container with id 963b8512d4cc77210ac82782985e45ccbf3ad5ae84a49590c8faaf5d0f2fa1a3 Apr 21 10:04:44.405695 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.405673 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-dbcfbcdb9-mp2qc"] Apr 21 10:04:44.441010 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.440988 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-dbcfbcdb9-mp2qc"] Apr 21 10:04:44.441103 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.441055 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.443986 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.443962 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 21 10:04:44.444089 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.443995 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 21 10:04:44.444154 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.444108 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 21 10:04:44.444923 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.444896 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 21 10:04:44.444923 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.444896 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-8zt54\"" Apr 21 10:04:44.445048 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.444973 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 21 10:04:44.445261 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.445236 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 21 10:04:44.445261 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.445252 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 21 10:04:44.495436 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.495408 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpn42\" (UniqueName: \"kubernetes.io/projected/27975ac4-8436-4fda-b1bc-5543c0f650f6-kube-api-access-cpn42\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.495709 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.495447 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-oauth-serving-cert\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.495709 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.495561 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-oauth-config\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.495709 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.495606 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-serving-cert\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.495709 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.495670 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-config\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.495709 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.495702 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-service-ca\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.596342 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.596314 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-config\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.596440 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.596343 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-service-ca\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.596440 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.596386 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cpn42\" (UniqueName: \"kubernetes.io/projected/27975ac4-8436-4fda-b1bc-5543c0f650f6-kube-api-access-cpn42\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.596523 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.596436 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-oauth-serving-cert\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.596618 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.596595 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-oauth-config\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.596657 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.596643 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-serving-cert\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.597034 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.597004 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-config\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.597136 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.597079 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-oauth-serving-cert\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.597136 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.597111 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-service-ca\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.599135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.599113 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-oauth-config\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.603570 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.603552 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-serving-cert\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.612121 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.612093 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpn42\" (UniqueName: \"kubernetes.io/projected/27975ac4-8436-4fda-b1bc-5543c0f650f6-kube-api-access-cpn42\") pod \"console-dbcfbcdb9-mp2qc\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.672588 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.672565 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hlz49" event={"ID":"8683a07e-2776-4d73-9312-ca2168702c96","Type":"ContainerStarted","Data":"963b8512d4cc77210ac82782985e45ccbf3ad5ae84a49590c8faaf5d0f2fa1a3"} Apr 21 10:04:44.673562 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.673536 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4fwj6" event={"ID":"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997","Type":"ContainerStarted","Data":"d51183a605989e2894947554539fcb405c06108b970465f65d2cacaf9e04d088"} Apr 21 10:04:44.674724 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.674706 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-ckmzt" event={"ID":"e51c2a85-13d4-47bb-916a-fb2ecece2c78","Type":"ContainerStarted","Data":"f8edb6b9727d9c080f6cdeae285e3a358b3f7eadb117f1af86f7e144a421bc5f"} Apr 21 10:04:44.674797 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.674730 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-ckmzt" event={"ID":"e51c2a85-13d4-47bb-916a-fb2ecece2c78","Type":"ContainerStarted","Data":"8c00e727dba68775db830d9de7e0dc9c55c1721006e14192ba27e5f0739ffd4f"} Apr 21 10:04:44.753382 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.753363 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:44.883880 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:44.883854 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-dbcfbcdb9-mp2qc"] Apr 21 10:04:44.886527 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:44.886497 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27975ac4_8436_4fda_b1bc_5543c0f650f6.slice/crio-048011befffe113514b00602fa02451a6928da4d13dc37bcf472bca8efe0c6c1 WatchSource:0}: Error finding container 048011befffe113514b00602fa02451a6928da4d13dc37bcf472bca8efe0c6c1: Status 404 returned error can't find the container with id 048011befffe113514b00602fa02451a6928da4d13dc37bcf472bca8efe0c6c1 Apr 21 10:04:45.458284 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.458238 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:04:45.458421 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.458237 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:04:45.461473 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.461315 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 21 10:04:45.462430 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.462406 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 21 10:04:45.462550 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.462448 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-fjqp7\"" Apr 21 10:04:45.462794 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.462772 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-qrq7w\"" Apr 21 10:04:45.462996 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.462979 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 21 10:04:45.678853 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.678817 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dbcfbcdb9-mp2qc" event={"ID":"27975ac4-8436-4fda-b1bc-5543c0f650f6","Type":"ContainerStarted","Data":"048011befffe113514b00602fa02451a6928da4d13dc37bcf472bca8efe0c6c1"} Apr 21 10:04:45.682072 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:45.682042 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-ckmzt" event={"ID":"e51c2a85-13d4-47bb-916a-fb2ecece2c78","Type":"ContainerStarted","Data":"8d2c7d3e84b72b8c0a4a85669b40e448c68a5cc8469c367e75a74670d595543e"} Apr 21 10:04:46.916367 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.916291 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-5676c8c784-92v6x"] Apr 21 10:04:46.919220 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.919196 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:46.925294 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.923011 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 21 10:04:46.925294 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.923866 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-dockercfg-s6f4w\"" Apr 21 10:04:46.925294 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.923998 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 21 10:04:46.925294 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.924295 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-tls\"" Apr 21 10:04:46.925294 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.924621 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-kube-rbac-proxy-config\"" Apr 21 10:04:46.925294 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.924878 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 21 10:04:46.930111 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:46.930090 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-5676c8c784-92v6x"] Apr 21 10:04:47.016320 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.016287 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.016320 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.016321 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5vv5\" (UniqueName: \"kubernetes.io/projected/56c34418-0154-4f7b-874a-e64fa17eecab-kube-api-access-p5vv5\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.016543 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.016351 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-tls\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.016543 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.016413 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/56c34418-0154-4f7b-874a-e64fa17eecab-metrics-client-ca\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.116753 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.116726 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.116859 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.116762 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p5vv5\" (UniqueName: \"kubernetes.io/projected/56c34418-0154-4f7b-874a-e64fa17eecab-kube-api-access-p5vv5\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.116859 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.116798 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-tls\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.116859 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.116822 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/56c34418-0154-4f7b-874a-e64fa17eecab-metrics-client-ca\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.117004 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:47.116983 2570 secret.go:189] Couldn't get secret openshift-monitoring/prometheus-operator-tls: secret "prometheus-operator-tls" not found Apr 21 10:04:47.117076 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:47.117064 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-tls podName:56c34418-0154-4f7b-874a-e64fa17eecab nodeName:}" failed. No retries permitted until 2026-04-21 10:04:47.617044333 +0000 UTC m=+53.698260155 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-operator-tls" (UniqueName: "kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-tls") pod "prometheus-operator-5676c8c784-92v6x" (UID: "56c34418-0154-4f7b-874a-e64fa17eecab") : secret "prometheus-operator-tls" not found Apr 21 10:04:47.117580 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.117552 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/56c34418-0154-4f7b-874a-e64fa17eecab-metrics-client-ca\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.119163 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.119137 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.126355 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.126329 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5vv5\" (UniqueName: \"kubernetes.io/projected/56c34418-0154-4f7b-874a-e64fa17eecab-kube-api-access-p5vv5\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.621568 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.621539 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-tls\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.624202 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.624177 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/56c34418-0154-4f7b-874a-e64fa17eecab-prometheus-operator-tls\") pod \"prometheus-operator-5676c8c784-92v6x\" (UID: \"56c34418-0154-4f7b-874a-e64fa17eecab\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:47.832465 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:47.832435 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" Apr 21 10:04:48.304690 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.304638 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-5676c8c784-92v6x"] Apr 21 10:04:48.310174 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:48.310080 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56c34418_0154_4f7b_874a_e64fa17eecab.slice/crio-f3b98635c6ca442378c8b2baa0b5cc5542c756208934b81c579f907f5a82d0f8 WatchSource:0}: Error finding container f3b98635c6ca442378c8b2baa0b5cc5542c756208934b81c579f907f5a82d0f8: Status 404 returned error can't find the container with id f3b98635c6ca442378c8b2baa0b5cc5542c756208934b81c579f907f5a82d0f8 Apr 21 10:04:48.691904 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.691809 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-ckmzt" event={"ID":"e51c2a85-13d4-47bb-916a-fb2ecece2c78","Type":"ContainerStarted","Data":"f8c145431e6b5bcd1140d731ccbbbc43f323d3be68f7897dd29e36e7709bf714"} Apr 21 10:04:48.692795 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.692769 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" event={"ID":"56c34418-0154-4f7b-874a-e64fa17eecab","Type":"ContainerStarted","Data":"f3b98635c6ca442378c8b2baa0b5cc5542c756208934b81c579f907f5a82d0f8"} Apr 21 10:04:48.693990 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.693968 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hlz49" event={"ID":"8683a07e-2776-4d73-9312-ca2168702c96","Type":"ContainerStarted","Data":"79c0c8c8573b8cab4ec31d718ba6552cb0ab56ee3ef42a23f455c228b35f2895"} Apr 21 10:04:48.695535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.695500 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4fwj6" event={"ID":"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997","Type":"ContainerStarted","Data":"cd2418f474598145d191cf1a6231acedbef070c1b4475bcbff99370ae093acd4"} Apr 21 10:04:48.695535 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.695534 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4fwj6" event={"ID":"da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997","Type":"ContainerStarted","Data":"966d862fe2043f21b65e7d9f42f1a38b9fab83401c925a549ed3f2b8b0cae621"} Apr 21 10:04:48.695763 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.695635 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:48.696696 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.696677 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dbcfbcdb9-mp2qc" event={"ID":"27975ac4-8436-4fda-b1bc-5543c0f650f6","Type":"ContainerStarted","Data":"d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0"} Apr 21 10:04:48.710389 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.710349 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-ckmzt" podStartSLOduration=2.055122258 podStartE2EDuration="5.710337277s" podCreationTimestamp="2026-04-21 10:04:43 +0000 UTC" firstStartedPulling="2026-04-21 10:04:44.510935567 +0000 UTC m=+50.592151376" lastFinishedPulling="2026-04-21 10:04:48.16615057 +0000 UTC m=+54.247366395" observedRunningTime="2026-04-21 10:04:48.708779337 +0000 UTC m=+54.789995168" watchObservedRunningTime="2026-04-21 10:04:48.710337277 +0000 UTC m=+54.791553108" Apr 21 10:04:48.723832 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.723793 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hlz49" podStartSLOduration=1.904612008 podStartE2EDuration="5.723780803s" podCreationTimestamp="2026-04-21 10:04:43 +0000 UTC" firstStartedPulling="2026-04-21 10:04:44.33807419 +0000 UTC m=+50.419290001" lastFinishedPulling="2026-04-21 10:04:48.157242986 +0000 UTC m=+54.238458796" observedRunningTime="2026-04-21 10:04:48.723729868 +0000 UTC m=+54.804945701" watchObservedRunningTime="2026-04-21 10:04:48.723780803 +0000 UTC m=+54.804996637" Apr 21 10:04:48.741700 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.741657 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-4fwj6" podStartSLOduration=1.911455656 podStartE2EDuration="5.741642856s" podCreationTimestamp="2026-04-21 10:04:43 +0000 UTC" firstStartedPulling="2026-04-21 10:04:44.327053399 +0000 UTC m=+50.408269210" lastFinishedPulling="2026-04-21 10:04:48.1572406 +0000 UTC m=+54.238456410" observedRunningTime="2026-04-21 10:04:48.74016635 +0000 UTC m=+54.821382183" watchObservedRunningTime="2026-04-21 10:04:48.741642856 +0000 UTC m=+54.822858688" Apr 21 10:04:48.763002 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:48.762959 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-dbcfbcdb9-mp2qc" podStartSLOduration=1.4801110880000001 podStartE2EDuration="4.762946389s" podCreationTimestamp="2026-04-21 10:04:44 +0000 UTC" firstStartedPulling="2026-04-21 10:04:44.888316593 +0000 UTC m=+50.969532404" lastFinishedPulling="2026-04-21 10:04:48.171151881 +0000 UTC m=+54.252367705" observedRunningTime="2026-04-21 10:04:48.762024205 +0000 UTC m=+54.843240037" watchObservedRunningTime="2026-04-21 10:04:48.762946389 +0000 UTC m=+54.844162221" Apr 21 10:04:49.369690 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.369572 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-56cdf5dd7b-bd6vq"] Apr 21 10:04:49.372949 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.372927 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.382488 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.382465 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-56cdf5dd7b-bd6vq"] Apr 21 10:04:49.384123 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.384084 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 21 10:04:49.435714 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.435693 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-oauth-config\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.435818 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.435730 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-config\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.435818 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.435747 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xznfr\" (UniqueName: \"kubernetes.io/projected/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-kube-api-access-xznfr\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.435925 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.435832 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-service-ca\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.435925 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.435863 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-oauth-serving-cert\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.435925 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.435892 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-serving-cert\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.435925 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.435908 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-trusted-ca-bundle\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.536314 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.536226 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-serving-cert\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.536314 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.536286 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-trusted-ca-bundle\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.536452 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.536349 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-oauth-config\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.536452 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.536398 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-config\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.536452 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.536424 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xznfr\" (UniqueName: \"kubernetes.io/projected/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-kube-api-access-xznfr\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.536605 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.536473 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-service-ca\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.536605 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.536495 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-oauth-serving-cert\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.537798 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.537769 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-config\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.537911 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.537895 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-service-ca\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.538009 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.537986 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-oauth-serving-cert\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.538424 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.538405 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-trusted-ca-bundle\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.538960 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.538940 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-serving-cert\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.539440 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.539418 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-oauth-config\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.547430 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.547409 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xznfr\" (UniqueName: \"kubernetes.io/projected/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-kube-api-access-xznfr\") pod \"console-56cdf5dd7b-bd6vq\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.684976 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.684889 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:49.701114 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.701076 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" event={"ID":"56c34418-0154-4f7b-874a-e64fa17eecab","Type":"ContainerStarted","Data":"dc53d60c3b0a51ff687e8a625902522f6cec14b4e4a643b8328936492aadd99c"} Apr 21 10:04:49.701241 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.701121 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" event={"ID":"56c34418-0154-4f7b-874a-e64fa17eecab","Type":"ContainerStarted","Data":"bd48e267d64225687f04bad19d7cb479b3f629fe21670cbcc98f6fad1efc3c7d"} Apr 21 10:04:49.720027 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.719974 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-5676c8c784-92v6x" podStartSLOduration=2.582725969 podStartE2EDuration="3.719958445s" podCreationTimestamp="2026-04-21 10:04:46 +0000 UTC" firstStartedPulling="2026-04-21 10:04:48.312227225 +0000 UTC m=+54.393443039" lastFinishedPulling="2026-04-21 10:04:49.449459702 +0000 UTC m=+55.530675515" observedRunningTime="2026-04-21 10:04:49.719394008 +0000 UTC m=+55.800609840" watchObservedRunningTime="2026-04-21 10:04:49.719958445 +0000 UTC m=+55.801174276" Apr 21 10:04:49.799534 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:49.799504 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-56cdf5dd7b-bd6vq"] Apr 21 10:04:49.803487 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:49.803460 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4eaa8a73_b945_4279_a1ba_e5588cdfdb4e.slice/crio-abe80a9ae7ae0d0df2c2f8f0213311cdd54d55578f559102f57274f39e19da9b WatchSource:0}: Error finding container abe80a9ae7ae0d0df2c2f8f0213311cdd54d55578f559102f57274f39e19da9b: Status 404 returned error can't find the container with id abe80a9ae7ae0d0df2c2f8f0213311cdd54d55578f559102f57274f39e19da9b Apr 21 10:04:50.704950 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:50.704915 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-56cdf5dd7b-bd6vq" event={"ID":"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e","Type":"ContainerStarted","Data":"966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c"} Apr 21 10:04:50.704950 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:50.704956 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-56cdf5dd7b-bd6vq" event={"ID":"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e","Type":"ContainerStarted","Data":"abe80a9ae7ae0d0df2c2f8f0213311cdd54d55578f559102f57274f39e19da9b"} Apr 21 10:04:50.721965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:50.721915 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-56cdf5dd7b-bd6vq" podStartSLOduration=1.721901526 podStartE2EDuration="1.721901526s" podCreationTimestamp="2026-04-21 10:04:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:04:50.7208158 +0000 UTC m=+56.802031644" watchObservedRunningTime="2026-04-21 10:04:50.721901526 +0000 UTC m=+56.803117359" Apr 21 10:04:51.275283 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.275234 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-qn6fw"] Apr 21 10:04:51.312285 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.312245 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-69db897b98-dkjlx"] Apr 21 10:04:51.312425 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.312408 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.315457 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.315435 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 21 10:04:51.315574 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.315525 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 21 10:04:51.315656 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.315642 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-48j4v\"" Apr 21 10:04:51.315728 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.315715 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 21 10:04:51.325354 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.325334 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-69db897b98-dkjlx"] Apr 21 10:04:51.325481 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.325431 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.327824 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.327806 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-state-metrics-tls\"" Apr 21 10:04:51.327824 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.327822 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-state-metrics-dockercfg-mj8tf\"" Apr 21 10:04:51.327958 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.327822 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-state-metrics-kube-rbac-proxy-config\"" Apr 21 10:04:51.327958 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.327935 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-state-metrics-custom-resource-state-configmap\"" Apr 21 10:04:51.350818 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.350793 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.350917 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.350827 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-tls\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.350917 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.350843 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-root\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.350917 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.350860 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-tls\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.351029 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.350950 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-metrics-client-ca\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.351029 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.350974 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-metrics-client-ca\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.351029 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.350991 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-sys\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.351029 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.351021 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8xm\" (UniqueName: \"kubernetes.io/projected/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-api-access-vw8xm\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.351171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.351068 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-wtmp\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.351171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.351099 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-accelerators-collector-config\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.351171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.351135 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.351171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.351153 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqnlv\" (UniqueName: \"kubernetes.io/projected/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-kube-api-access-lqnlv\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.351346 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.351174 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-textfile\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.351346 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.351194 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.351346 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.351216 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-volume-directive-shadow\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.452027 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.451995 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.452027 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452023 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lqnlv\" (UniqueName: \"kubernetes.io/projected/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-kube-api-access-lqnlv\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452216 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452041 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-textfile\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452216 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452062 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452216 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452119 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-volume-directive-shadow\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.452216 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452178 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.452467 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452219 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-tls\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.452467 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452351 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-root\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452467 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452384 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-tls\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452467 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452411 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-textfile\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452467 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452416 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-metrics-client-ca\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.452467 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452466 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-metrics-client-ca\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452788 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452472 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-root\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452788 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:51.452477 2570 secret.go:189] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Apr 21 10:04:51.452788 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452529 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-sys\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452788 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:04:51.452556 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-tls podName:0ee7a123-6bbc-419a-999a-fd25ccd1bba1 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:51.952537694 +0000 UTC m=+58.033753505 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-tls") pod "node-exporter-qn6fw" (UID: "0ee7a123-6bbc-419a-999a-fd25ccd1bba1") : secret "node-exporter-tls" not found Apr 21 10:04:51.452788 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452494 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-sys\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452788 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452601 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8xm\" (UniqueName: \"kubernetes.io/projected/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-api-access-vw8xm\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.452788 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452641 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-wtmp\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.452788 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452695 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-accelerators-collector-config\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.453151 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452813 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.453151 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.452817 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-wtmp\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.453151 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.453005 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-metrics-client-ca\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.453151 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.453137 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-metrics-client-ca\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.453383 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.453165 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-accelerators-collector-config\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.453506 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.453486 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-volume-directive-shadow\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.455086 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.455063 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.455524 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.455505 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-tls\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.455603 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.455501 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.460184 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.460154 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqnlv\" (UniqueName: \"kubernetes.io/projected/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-kube-api-access-lqnlv\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.461204 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.461184 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8xm\" (UniqueName: \"kubernetes.io/projected/6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae-kube-api-access-vw8xm\") pod \"kube-state-metrics-69db897b98-dkjlx\" (UID: \"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae\") " pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.633879 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.633850 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" Apr 21 10:04:51.752062 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.752032 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-69db897b98-dkjlx"] Apr 21 10:04:51.755396 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:51.755368 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6090567f_b8cb_4bc0_9ad5_3cb61bcdaeae.slice/crio-0b689d129cb90b4306f6333613d1f42593841234b13819eeb60a0d9062c343b5 WatchSource:0}: Error finding container 0b689d129cb90b4306f6333613d1f42593841234b13819eeb60a0d9062c343b5: Status 404 returned error can't find the container with id 0b689d129cb90b4306f6333613d1f42593841234b13819eeb60a0d9062c343b5 Apr 21 10:04:51.956761 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.956703 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-tls\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:51.958918 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:51.958901 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/0ee7a123-6bbc-419a-999a-fd25ccd1bba1-node-exporter-tls\") pod \"node-exporter-qn6fw\" (UID: \"0ee7a123-6bbc-419a-999a-fd25ccd1bba1\") " pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:52.221347 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:52.221259 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-qn6fw" Apr 21 10:04:52.231315 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:52.230662 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ee7a123_6bbc_419a_999a_fd25ccd1bba1.slice/crio-943d847ea4ed39473fae86ababefe2ac815912290fd2d34d272760366cd45ec4 WatchSource:0}: Error finding container 943d847ea4ed39473fae86ababefe2ac815912290fd2d34d272760366cd45ec4: Status 404 returned error can't find the container with id 943d847ea4ed39473fae86ababefe2ac815912290fd2d34d272760366cd45ec4 Apr 21 10:04:52.634517 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:52.634487 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbn7w" Apr 21 10:04:52.712970 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:52.712926 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-qn6fw" event={"ID":"0ee7a123-6bbc-419a-999a-fd25ccd1bba1","Type":"ContainerStarted","Data":"943d847ea4ed39473fae86ababefe2ac815912290fd2d34d272760366cd45ec4"} Apr 21 10:04:52.714243 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:52.714201 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" event={"ID":"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae","Type":"ContainerStarted","Data":"0b689d129cb90b4306f6333613d1f42593841234b13819eeb60a0d9062c343b5"} Apr 21 10:04:53.359574 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.359546 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp"] Apr 21 10:04:53.364406 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.364388 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.368650 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.368626 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-web\"" Apr 21 10:04:53.368787 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.368767 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-grpc-tls-163gtq1gub7sp\"" Apr 21 10:04:53.368856 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.368798 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-tls\"" Apr 21 10:04:53.368918 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.368864 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy\"" Apr 21 10:04:53.368918 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.368883 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-dockercfg-wjbfv\"" Apr 21 10:04:53.369016 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.369004 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-metrics\"" Apr 21 10:04:53.369123 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.369106 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"thanos-querier-kube-rbac-proxy-rules\"" Apr 21 10:04:53.374963 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.374945 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp"] Apr 21 10:04:53.468448 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.468418 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.468562 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.468506 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-tls\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.468562 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.468537 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.468657 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.468559 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-metrics-client-ca\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.468657 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.468600 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-grpc-tls\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.468722 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.468666 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.468722 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.468690 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.468722 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.468705 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5mzn\" (UniqueName: \"kubernetes.io/projected/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-kube-api-access-w5mzn\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.569790 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.569744 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.569790 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.569777 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.569961 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.569796 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-w5mzn\" (UniqueName: \"kubernetes.io/projected/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-kube-api-access-w5mzn\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.569961 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.569821 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.570064 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.570045 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-tls\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.570113 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.570084 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.570176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.570114 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-metrics-client-ca\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.570176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.570147 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-grpc-tls\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.570802 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.570776 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-metrics-client-ca\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.572071 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.572049 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.572735 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.572707 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.572829 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.572814 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-tls\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.572887 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.572867 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.572920 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.572867 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.572980 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.572964 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-secret-grpc-tls\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.578421 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.578401 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5mzn\" (UniqueName: \"kubernetes.io/projected/2acd23ca-ac79-49f3-9242-9cd19cdf3c0e-kube-api-access-w5mzn\") pod \"thanos-querier-6bcd6c69d5-4x5xp\" (UID: \"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e\") " pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.681631 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.681598 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:53.718975 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.718939 2570 generic.go:358] "Generic (PLEG): container finished" podID="0ee7a123-6bbc-419a-999a-fd25ccd1bba1" containerID="9feb45a3468aac24737a1bdef546b179917a86d04dd49544026db44d3cb1be27" exitCode=0 Apr 21 10:04:53.719131 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.719018 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-qn6fw" event={"ID":"0ee7a123-6bbc-419a-999a-fd25ccd1bba1","Type":"ContainerDied","Data":"9feb45a3468aac24737a1bdef546b179917a86d04dd49544026db44d3cb1be27"} Apr 21 10:04:53.724808 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.721599 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" event={"ID":"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae","Type":"ContainerStarted","Data":"6da50dba0333f95d5d7121df3f1f31a9f2fc755e4687d41e1233a82292c6748d"} Apr 21 10:04:53.724808 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.721632 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" event={"ID":"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae","Type":"ContainerStarted","Data":"4363d7e85666e2ba22e4683e665c919152cabc7f8e8c884570f6603561dc8de8"} Apr 21 10:04:53.724808 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.721646 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" event={"ID":"6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae","Type":"ContainerStarted","Data":"3b02db81440126b40e82f9f0d8873d0ab6a032036c78a7e8f14506ed8c06b2f5"} Apr 21 10:04:53.754203 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.754160 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-69db897b98-dkjlx" podStartSLOduration=1.6424573009999999 podStartE2EDuration="2.754146183s" podCreationTimestamp="2026-04-21 10:04:51 +0000 UTC" firstStartedPulling="2026-04-21 10:04:51.757309446 +0000 UTC m=+57.838525256" lastFinishedPulling="2026-04-21 10:04:52.868998328 +0000 UTC m=+58.950214138" observedRunningTime="2026-04-21 10:04:53.753041424 +0000 UTC m=+59.834257252" watchObservedRunningTime="2026-04-21 10:04:53.754146183 +0000 UTC m=+59.835362015" Apr 21 10:04:53.809894 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:53.809861 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp"] Apr 21 10:04:53.814185 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:53.814147 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2acd23ca_ac79_49f3_9242_9cd19cdf3c0e.slice/crio-0711755d62ffd01d69af85b0924e6c902457cc48f1ecc920822a7fc3eae66e34 WatchSource:0}: Error finding container 0711755d62ffd01d69af85b0924e6c902457cc48f1ecc920822a7fc3eae66e34: Status 404 returned error can't find the container with id 0711755d62ffd01d69af85b0924e6c902457cc48f1ecc920822a7fc3eae66e34 Apr 21 10:04:54.726368 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:54.726300 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-qn6fw" event={"ID":"0ee7a123-6bbc-419a-999a-fd25ccd1bba1","Type":"ContainerStarted","Data":"0e36fcbaee93c5232bb3f41774115c59b73d49b104813c1617764a2b17ca64e0"} Apr 21 10:04:54.726368 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:54.726345 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-qn6fw" event={"ID":"0ee7a123-6bbc-419a-999a-fd25ccd1bba1","Type":"ContainerStarted","Data":"f696b2f1a45bfcd906971298fb7f3e97fba082c9f86481e63afc0ca7fcc28758"} Apr 21 10:04:54.727621 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:54.727588 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" event={"ID":"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e","Type":"ContainerStarted","Data":"0711755d62ffd01d69af85b0924e6c902457cc48f1ecc920822a7fc3eae66e34"} Apr 21 10:04:54.747680 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:54.747627 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-qn6fw" podStartSLOduration=2.669477711 podStartE2EDuration="3.74760972s" podCreationTimestamp="2026-04-21 10:04:51 +0000 UTC" firstStartedPulling="2026-04-21 10:04:52.232916449 +0000 UTC m=+58.314132259" lastFinishedPulling="2026-04-21 10:04:53.311048443 +0000 UTC m=+59.392264268" observedRunningTime="2026-04-21 10:04:54.746912805 +0000 UTC m=+60.828128638" watchObservedRunningTime="2026-04-21 10:04:54.74760972 +0000 UTC m=+60.828825553" Apr 21 10:04:54.753673 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:54.753648 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:54.753791 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:54.753680 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:54.758993 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:54.758972 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:55.734340 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:55.734309 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:04:56.734969 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:56.734938 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" event={"ID":"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e","Type":"ContainerStarted","Data":"dca76f2d6571a7ec7e216929814fe481beead9096fae2b9eeffa1e4a6f5b18a6"} Apr 21 10:04:56.734969 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:56.734974 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" event={"ID":"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e","Type":"ContainerStarted","Data":"0c16569fdd5e8b34f0e913f32c45367bb6a8ec63f5970325a1722bcbfe567c82"} Apr 21 10:04:56.735383 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:56.734983 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" event={"ID":"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e","Type":"ContainerStarted","Data":"fb072dec30634bcb9a8de7108627c37971a60e73b6f9289a3584c9df2d77c6e5"} Apr 21 10:04:57.740739 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:57.740700 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" event={"ID":"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e","Type":"ContainerStarted","Data":"e55d76779ad05e18d99155362d4aac9c98eda51b9fadcf8bf7bd11df88beb64d"} Apr 21 10:04:57.741188 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:57.740749 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" event={"ID":"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e","Type":"ContainerStarted","Data":"1306561fe322903d9953a37fd28f94a8774ed15786ef7a107d76c010bb3f1893"} Apr 21 10:04:57.741188 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:57.740764 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" event={"ID":"2acd23ca-ac79-49f3-9242-9cd19cdf3c0e","Type":"ContainerStarted","Data":"71149aa32ce28eabc54098b0b162c77e0091ca5494cb120e0fbe10d60bbae374"} Apr 21 10:04:57.741188 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:57.740878 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:04:57.765648 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:57.765592 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" podStartSLOduration=1.8188669910000002 podStartE2EDuration="4.765576582s" podCreationTimestamp="2026-04-21 10:04:53 +0000 UTC" firstStartedPulling="2026-04-21 10:04:53.818214145 +0000 UTC m=+59.899429959" lastFinishedPulling="2026-04-21 10:04:56.764923736 +0000 UTC m=+62.846139550" observedRunningTime="2026-04-21 10:04:57.763100795 +0000 UTC m=+63.844316664" watchObservedRunningTime="2026-04-21 10:04:57.765576582 +0000 UTC m=+63.846792418" Apr 21 10:04:58.703135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:58.703105 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-4fwj6" Apr 21 10:04:59.182000 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.181966 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-56cdf5dd7b-bd6vq"] Apr 21 10:04:59.213823 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.213798 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-574d4c5d68-cxcdm"] Apr 21 10:04:59.235368 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.235347 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-574d4c5d68-cxcdm"] Apr 21 10:04:59.235468 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.235446 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.322239 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.322208 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-console-config\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.322401 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.322249 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-serving-cert\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.322401 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.322285 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqr4j\" (UniqueName: \"kubernetes.io/projected/97661741-da08-4a3a-b481-6e7f7efe69ec-kube-api-access-qqr4j\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.322401 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.322368 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-oauth-config\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.322507 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.322426 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-trusted-ca-bundle\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.322549 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.322507 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-oauth-serving-cert\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.322549 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.322530 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-service-ca\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.423678 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.423641 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-trusted-ca-bundle\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.423834 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.423711 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-oauth-serving-cert\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.423834 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.423733 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-service-ca\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.423834 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.423755 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-console-config\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.423834 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.423788 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-serving-cert\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.423992 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.423834 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qqr4j\" (UniqueName: \"kubernetes.io/projected/97661741-da08-4a3a-b481-6e7f7efe69ec-kube-api-access-qqr4j\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.424288 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.424238 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-oauth-config\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.424546 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.424518 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-service-ca\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.424634 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.424579 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-trusted-ca-bundle\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.424634 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.424597 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-oauth-serving-cert\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.426595 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.426577 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-serving-cert\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.426677 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.426593 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-oauth-config\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.432958 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.432908 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqr4j\" (UniqueName: \"kubernetes.io/projected/97661741-da08-4a3a-b481-6e7f7efe69ec-kube-api-access-qqr4j\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.447859 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.447836 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-console-config\") pod \"console-574d4c5d68-cxcdm\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.544198 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.544174 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:04:59.671040 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.671010 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-574d4c5d68-cxcdm"] Apr 21 10:04:59.675145 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:04:59.675117 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97661741_da08_4a3a_b481_6e7f7efe69ec.slice/crio-d03888fc419f285b1563a374fe4ed666ba5ea56b3029bad24f721e482be778c9 WatchSource:0}: Error finding container d03888fc419f285b1563a374fe4ed666ba5ea56b3029bad24f721e482be778c9: Status 404 returned error can't find the container with id d03888fc419f285b1563a374fe4ed666ba5ea56b3029bad24f721e482be778c9 Apr 21 10:04:59.684998 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.684957 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:04:59.749689 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:04:59.749661 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-574d4c5d68-cxcdm" event={"ID":"97661741-da08-4a3a-b481-6e7f7efe69ec","Type":"ContainerStarted","Data":"d03888fc419f285b1563a374fe4ed666ba5ea56b3029bad24f721e482be778c9"} Apr 21 10:05:00.229004 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.228967 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:05:00.231431 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.231411 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 21 10:05:00.241671 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.241654 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd143ad0-d084-434c-87b4-bd355e2b205a-metrics-certs\") pod \"network-metrics-daemon-ff5dl\" (UID: \"bd143ad0-d084-434c-87b4-bd355e2b205a\") " pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:05:00.330112 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.330087 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:05:00.332632 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.332616 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 21 10:05:00.343171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.343154 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 21 10:05:00.353542 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.353521 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nrcg\" (UniqueName: \"kubernetes.io/projected/d0ada080-47c9-4134-a37c-b3bd33e4eb38-kube-api-access-6nrcg\") pod \"network-check-target-2j8xh\" (UID: \"d0ada080-47c9-4134-a37c-b3bd33e4eb38\") " pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:05:00.474098 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.474073 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-fjqp7\"" Apr 21 10:05:00.480444 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.480397 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-qrq7w\"" Apr 21 10:05:00.482901 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.482880 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ff5dl" Apr 21 10:05:00.488564 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.488547 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:05:00.630167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.630110 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ff5dl"] Apr 21 10:05:00.634777 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:05:00.634747 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd143ad0_d084_434c_87b4_bd355e2b205a.slice/crio-fb106acaa64fd51ab968eef73a026d7e6f302c51159de1df9247c432b837bc08 WatchSource:0}: Error finding container fb106acaa64fd51ab968eef73a026d7e6f302c51159de1df9247c432b837bc08: Status 404 returned error can't find the container with id fb106acaa64fd51ab968eef73a026d7e6f302c51159de1df9247c432b837bc08 Apr 21 10:05:00.644225 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.644203 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-2j8xh"] Apr 21 10:05:00.754372 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.754254 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-2j8xh" event={"ID":"d0ada080-47c9-4134-a37c-b3bd33e4eb38","Type":"ContainerStarted","Data":"d881e027834f79ce019177e92f7b012373d0775432e1e179d5fcf21f60e44b72"} Apr 21 10:05:00.755938 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.755906 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-574d4c5d68-cxcdm" event={"ID":"97661741-da08-4a3a-b481-6e7f7efe69ec","Type":"ContainerStarted","Data":"9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4"} Apr 21 10:05:00.757386 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.757365 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ff5dl" event={"ID":"bd143ad0-d084-434c-87b4-bd355e2b205a","Type":"ContainerStarted","Data":"fb106acaa64fd51ab968eef73a026d7e6f302c51159de1df9247c432b837bc08"} Apr 21 10:05:00.786901 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:00.786840 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-574d4c5d68-cxcdm" podStartSLOduration=1.786823984 podStartE2EDuration="1.786823984s" podCreationTimestamp="2026-04-21 10:04:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:05:00.785222492 +0000 UTC m=+66.866438325" watchObservedRunningTime="2026-04-21 10:05:00.786823984 +0000 UTC m=+66.868039817" Apr 21 10:05:02.765614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:02.765576 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ff5dl" event={"ID":"bd143ad0-d084-434c-87b4-bd355e2b205a","Type":"ContainerStarted","Data":"99b17cbbc72d7fb3108a02c4442ee14b13954beb9a27fb0c29202ee300a3ace6"} Apr 21 10:05:02.765614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:02.765616 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ff5dl" event={"ID":"bd143ad0-d084-434c-87b4-bd355e2b205a","Type":"ContainerStarted","Data":"e50e988d353572f12c09b1f89f7148a04a7722a4b107a62bd64cf772c2f05603"} Apr 21 10:05:02.783490 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:02.783432 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-ff5dl" podStartSLOduration=67.698768808 podStartE2EDuration="1m8.783414929s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:05:00.636754661 +0000 UTC m=+66.717970475" lastFinishedPulling="2026-04-21 10:05:01.721400777 +0000 UTC m=+67.802616596" observedRunningTime="2026-04-21 10:05:02.78141873 +0000 UTC m=+68.862634564" watchObservedRunningTime="2026-04-21 10:05:02.783414929 +0000 UTC m=+68.864630762" Apr 21 10:05:03.749481 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:03.749405 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-6bcd6c69d5-4x5xp" Apr 21 10:05:03.770287 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:03.770246 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-2j8xh" event={"ID":"d0ada080-47c9-4134-a37c-b3bd33e4eb38","Type":"ContainerStarted","Data":"6ce5e110a3de52c97d527088c2c0d9d76743906b38b3fd4a8894274a96d22152"} Apr 21 10:05:03.770635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:03.770396 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:05:03.788729 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:03.788687 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-2j8xh" podStartSLOduration=67.064179482 podStartE2EDuration="1m9.788675464s" podCreationTimestamp="2026-04-21 10:03:54 +0000 UTC" firstStartedPulling="2026-04-21 10:05:00.649323539 +0000 UTC m=+66.730539354" lastFinishedPulling="2026-04-21 10:05:03.37381952 +0000 UTC m=+69.455035336" observedRunningTime="2026-04-21 10:05:03.787286089 +0000 UTC m=+69.868501916" watchObservedRunningTime="2026-04-21 10:05:03.788675464 +0000 UTC m=+69.869891296" Apr 21 10:05:08.082264 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.082232 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-574d4c5d68-cxcdm"] Apr 21 10:05:08.105077 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.105052 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-54567b64f6-lq4ds"] Apr 21 10:05:08.108370 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.108350 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.118473 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.118454 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54567b64f6-lq4ds"] Apr 21 10:05:08.191944 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.191903 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-config\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.191944 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.191945 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-oauth-serving-cert\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.192196 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.192008 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-trusted-ca-bundle\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.192196 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.192034 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-service-ca\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.192196 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.192101 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7k8z\" (UniqueName: \"kubernetes.io/projected/082ead76-09e8-47ef-a9fa-ccf7bf22026f-kube-api-access-g7k8z\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.192196 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.192135 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-oauth-config\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.192196 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.192164 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-serving-cert\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.292820 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.292790 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-trusted-ca-bundle\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.292952 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.292823 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-service-ca\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.292952 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.292847 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g7k8z\" (UniqueName: \"kubernetes.io/projected/082ead76-09e8-47ef-a9fa-ccf7bf22026f-kube-api-access-g7k8z\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.292952 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.292879 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-oauth-config\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.292952 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.292907 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-serving-cert\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.293163 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.292956 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-config\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.293163 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.292976 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-oauth-serving-cert\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.293650 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.293623 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-service-ca\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.293759 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.293655 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-oauth-serving-cert\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.293810 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.293798 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-trusted-ca-bundle\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.293993 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.293960 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-config\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.295673 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.295620 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-serving-cert\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.295771 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.295747 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-oauth-config\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.301618 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.301595 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7k8z\" (UniqueName: \"kubernetes.io/projected/082ead76-09e8-47ef-a9fa-ccf7bf22026f-kube-api-access-g7k8z\") pod \"console-54567b64f6-lq4ds\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.417081 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.417055 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:08.540918 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.540896 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54567b64f6-lq4ds"] Apr 21 10:05:08.543609 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:05:08.543574 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod082ead76_09e8_47ef_a9fa_ccf7bf22026f.slice/crio-f68a770a7c3033255ed7f267e2e88be3c4b8abe5da6ee897f74bcf93e93f771a WatchSource:0}: Error finding container f68a770a7c3033255ed7f267e2e88be3c4b8abe5da6ee897f74bcf93e93f771a: Status 404 returned error can't find the container with id f68a770a7c3033255ed7f267e2e88be3c4b8abe5da6ee897f74bcf93e93f771a Apr 21 10:05:08.786214 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.786137 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54567b64f6-lq4ds" event={"ID":"082ead76-09e8-47ef-a9fa-ccf7bf22026f","Type":"ContainerStarted","Data":"d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f"} Apr 21 10:05:08.786214 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.786173 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54567b64f6-lq4ds" event={"ID":"082ead76-09e8-47ef-a9fa-ccf7bf22026f","Type":"ContainerStarted","Data":"f68a770a7c3033255ed7f267e2e88be3c4b8abe5da6ee897f74bcf93e93f771a"} Apr 21 10:05:08.804895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:08.804854 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-54567b64f6-lq4ds" podStartSLOduration=0.804839482 podStartE2EDuration="804.839482ms" podCreationTimestamp="2026-04-21 10:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:05:08.803720833 +0000 UTC m=+74.884936665" watchObservedRunningTime="2026-04-21 10:05:08.804839482 +0000 UTC m=+74.886055313" Apr 21 10:05:09.544762 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:09.544732 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:05:18.417942 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:18.417822 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:18.417942 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:18.417863 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:18.422420 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:18.422397 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:18.816389 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:18.816326 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:05:18.865450 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:18.865422 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-dbcfbcdb9-mp2qc"] Apr 21 10:05:24.200592 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.200532 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-56cdf5dd7b-bd6vq" podUID="4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" containerName="console" containerID="cri-o://966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c" gracePeriod=15 Apr 21 10:05:24.439093 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.439073 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-56cdf5dd7b-bd6vq_4eaa8a73-b945-4279-a1ba-e5588cdfdb4e/console/0.log" Apr 21 10:05:24.439208 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.439144 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:05:24.609243 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609216 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-trusted-ca-bundle\") pod \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " Apr 21 10:05:24.609403 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609254 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-serving-cert\") pod \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " Apr 21 10:05:24.609403 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609305 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-service-ca\") pod \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " Apr 21 10:05:24.609403 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609346 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xznfr\" (UniqueName: \"kubernetes.io/projected/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-kube-api-access-xznfr\") pod \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " Apr 21 10:05:24.609403 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609385 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-config\") pod \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " Apr 21 10:05:24.609602 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609455 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-oauth-serving-cert\") pod \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " Apr 21 10:05:24.609602 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609481 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-oauth-config\") pod \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\" (UID: \"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e\") " Apr 21 10:05:24.609748 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609632 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" (UID: "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:24.609882 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609753 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-service-ca" (OuterVolumeSpecName: "service-ca") pod "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" (UID: "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:24.609882 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609811 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" (UID: "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:24.609985 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.609809 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-config" (OuterVolumeSpecName: "console-config") pod "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" (UID: "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:24.611729 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.611700 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" (UID: "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:05:24.611810 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.611742 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" (UID: "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:05:24.611810 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.611761 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-kube-api-access-xznfr" (OuterVolumeSpecName: "kube-api-access-xznfr") pod "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" (UID: "4eaa8a73-b945-4279-a1ba-e5588cdfdb4e"). InnerVolumeSpecName "kube-api-access-xznfr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:05:24.710280 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.710234 2570 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-trusted-ca-bundle\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:24.710280 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.710291 2570 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:24.710496 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.710303 2570 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-service-ca\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:24.710496 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.710313 2570 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xznfr\" (UniqueName: \"kubernetes.io/projected/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-kube-api-access-xznfr\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:24.710496 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.710323 2570 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:24.710496 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.710331 2570 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-oauth-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:24.710496 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.710340 2570 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e-console-oauth-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:24.829858 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.829836 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-56cdf5dd7b-bd6vq_4eaa8a73-b945-4279-a1ba-e5588cdfdb4e/console/0.log" Apr 21 10:05:24.829963 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.829876 2570 generic.go:358] "Generic (PLEG): container finished" podID="4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" containerID="966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c" exitCode=2 Apr 21 10:05:24.829963 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.829945 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-56cdf5dd7b-bd6vq" Apr 21 10:05:24.830042 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.829968 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-56cdf5dd7b-bd6vq" event={"ID":"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e","Type":"ContainerDied","Data":"966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c"} Apr 21 10:05:24.830042 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.830004 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-56cdf5dd7b-bd6vq" event={"ID":"4eaa8a73-b945-4279-a1ba-e5588cdfdb4e","Type":"ContainerDied","Data":"abe80a9ae7ae0d0df2c2f8f0213311cdd54d55578f559102f57274f39e19da9b"} Apr 21 10:05:24.830042 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.830029 2570 scope.go:117] "RemoveContainer" containerID="966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c" Apr 21 10:05:24.838555 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.838539 2570 scope.go:117] "RemoveContainer" containerID="966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c" Apr 21 10:05:24.838830 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:05:24.838811 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c\": container with ID starting with 966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c not found: ID does not exist" containerID="966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c" Apr 21 10:05:24.838896 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.838840 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c"} err="failed to get container status \"966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c\": rpc error: code = NotFound desc = could not find container \"966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c\": container with ID starting with 966ba4459e4c59ff688a21fc96f796225f1667b096ed0f5b690fcec1dcc6b61c not found: ID does not exist" Apr 21 10:05:24.853089 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.853063 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-56cdf5dd7b-bd6vq"] Apr 21 10:05:24.856200 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:24.856182 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-56cdf5dd7b-bd6vq"] Apr 21 10:05:26.462305 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:26.462250 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" path="/var/lib/kubelet/pods/4eaa8a73-b945-4279-a1ba-e5588cdfdb4e/volumes" Apr 21 10:05:33.101347 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.101295 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-574d4c5d68-cxcdm" podUID="97661741-da08-4a3a-b481-6e7f7efe69ec" containerName="console" containerID="cri-o://9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4" gracePeriod=15 Apr 21 10:05:33.346305 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.346284 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-574d4c5d68-cxcdm_97661741-da08-4a3a-b481-6e7f7efe69ec/console/0.log" Apr 21 10:05:33.346394 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.346340 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:05:33.471985 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.471959 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-console-config\") pod \"97661741-da08-4a3a-b481-6e7f7efe69ec\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " Apr 21 10:05:33.472110 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.471991 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqr4j\" (UniqueName: \"kubernetes.io/projected/97661741-da08-4a3a-b481-6e7f7efe69ec-kube-api-access-qqr4j\") pod \"97661741-da08-4a3a-b481-6e7f7efe69ec\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " Apr 21 10:05:33.472110 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472009 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-trusted-ca-bundle\") pod \"97661741-da08-4a3a-b481-6e7f7efe69ec\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " Apr 21 10:05:33.472110 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472031 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-oauth-serving-cert\") pod \"97661741-da08-4a3a-b481-6e7f7efe69ec\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " Apr 21 10:05:33.472255 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472191 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-serving-cert\") pod \"97661741-da08-4a3a-b481-6e7f7efe69ec\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " Apr 21 10:05:33.472340 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472263 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-oauth-config\") pod \"97661741-da08-4a3a-b481-6e7f7efe69ec\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " Apr 21 10:05:33.472340 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472334 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-service-ca\") pod \"97661741-da08-4a3a-b481-6e7f7efe69ec\" (UID: \"97661741-da08-4a3a-b481-6e7f7efe69ec\") " Apr 21 10:05:33.472431 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472383 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-console-config" (OuterVolumeSpecName: "console-config") pod "97661741-da08-4a3a-b481-6e7f7efe69ec" (UID: "97661741-da08-4a3a-b481-6e7f7efe69ec"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:33.472483 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472467 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "97661741-da08-4a3a-b481-6e7f7efe69ec" (UID: "97661741-da08-4a3a-b481-6e7f7efe69ec"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:33.472529 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472481 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "97661741-da08-4a3a-b481-6e7f7efe69ec" (UID: "97661741-da08-4a3a-b481-6e7f7efe69ec"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:33.472597 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472577 2570 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-console-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:33.472597 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472596 2570 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-trusted-ca-bundle\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:33.472695 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.472605 2570 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-oauth-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:33.473059 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.473029 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-service-ca" (OuterVolumeSpecName: "service-ca") pod "97661741-da08-4a3a-b481-6e7f7efe69ec" (UID: "97661741-da08-4a3a-b481-6e7f7efe69ec"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:33.474186 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.474158 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97661741-da08-4a3a-b481-6e7f7efe69ec-kube-api-access-qqr4j" (OuterVolumeSpecName: "kube-api-access-qqr4j") pod "97661741-da08-4a3a-b481-6e7f7efe69ec" (UID: "97661741-da08-4a3a-b481-6e7f7efe69ec"). InnerVolumeSpecName "kube-api-access-qqr4j". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:05:33.474422 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.474387 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "97661741-da08-4a3a-b481-6e7f7efe69ec" (UID: "97661741-da08-4a3a-b481-6e7f7efe69ec"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:05:33.474422 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.474403 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "97661741-da08-4a3a-b481-6e7f7efe69ec" (UID: "97661741-da08-4a3a-b481-6e7f7efe69ec"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:05:33.573336 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.573316 2570 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97661741-da08-4a3a-b481-6e7f7efe69ec-service-ca\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:33.573336 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.573334 2570 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qqr4j\" (UniqueName: \"kubernetes.io/projected/97661741-da08-4a3a-b481-6e7f7efe69ec-kube-api-access-qqr4j\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:33.573457 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.573344 2570 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:33.573457 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.573353 2570 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97661741-da08-4a3a-b481-6e7f7efe69ec-console-oauth-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:33.856316 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.856293 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-574d4c5d68-cxcdm_97661741-da08-4a3a-b481-6e7f7efe69ec/console/0.log" Apr 21 10:05:33.856445 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.856331 2570 generic.go:358] "Generic (PLEG): container finished" podID="97661741-da08-4a3a-b481-6e7f7efe69ec" containerID="9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4" exitCode=2 Apr 21 10:05:33.856445 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.856358 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-574d4c5d68-cxcdm" event={"ID":"97661741-da08-4a3a-b481-6e7f7efe69ec","Type":"ContainerDied","Data":"9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4"} Apr 21 10:05:33.856445 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.856380 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-574d4c5d68-cxcdm" event={"ID":"97661741-da08-4a3a-b481-6e7f7efe69ec","Type":"ContainerDied","Data":"d03888fc419f285b1563a374fe4ed666ba5ea56b3029bad24f721e482be778c9"} Apr 21 10:05:33.856445 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.856393 2570 scope.go:117] "RemoveContainer" containerID="9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4" Apr 21 10:05:33.856445 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.856403 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-574d4c5d68-cxcdm" Apr 21 10:05:33.864461 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.864447 2570 scope.go:117] "RemoveContainer" containerID="9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4" Apr 21 10:05:33.864712 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:05:33.864693 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4\": container with ID starting with 9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4 not found: ID does not exist" containerID="9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4" Apr 21 10:05:33.864757 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.864720 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4"} err="failed to get container status \"9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4\": rpc error: code = NotFound desc = could not find container \"9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4\": container with ID starting with 9882bc352c4a790b72319267ead2685566de6894a87d91c8e794447a972a5ee4 not found: ID does not exist" Apr 21 10:05:33.876702 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.876677 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-574d4c5d68-cxcdm"] Apr 21 10:05:33.879845 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:33.879824 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-574d4c5d68-cxcdm"] Apr 21 10:05:34.461753 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:34.461724 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97661741-da08-4a3a-b481-6e7f7efe69ec" path="/var/lib/kubelet/pods/97661741-da08-4a3a-b481-6e7f7efe69ec/volumes" Apr 21 10:05:34.776329 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:34.776239 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-2j8xh" Apr 21 10:05:43.884844 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:43.884803 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-dbcfbcdb9-mp2qc" podUID="27975ac4-8436-4fda-b1bc-5543c0f650f6" containerName="console" containerID="cri-o://d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0" gracePeriod=15 Apr 21 10:05:44.126138 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.126117 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-dbcfbcdb9-mp2qc_27975ac4-8436-4fda-b1bc-5543c0f650f6/console/0.log" Apr 21 10:05:44.126249 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.126177 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:05:44.249135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249064 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-service-ca\") pod \"27975ac4-8436-4fda-b1bc-5543c0f650f6\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " Apr 21 10:05:44.249135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249103 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-oauth-config\") pod \"27975ac4-8436-4fda-b1bc-5543c0f650f6\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " Apr 21 10:05:44.249135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249128 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-oauth-serving-cert\") pod \"27975ac4-8436-4fda-b1bc-5543c0f650f6\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " Apr 21 10:05:44.249426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249288 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpn42\" (UniqueName: \"kubernetes.io/projected/27975ac4-8436-4fda-b1bc-5543c0f650f6-kube-api-access-cpn42\") pod \"27975ac4-8436-4fda-b1bc-5543c0f650f6\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " Apr 21 10:05:44.249426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249336 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-config\") pod \"27975ac4-8436-4fda-b1bc-5543c0f650f6\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " Apr 21 10:05:44.249426 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249395 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-serving-cert\") pod \"27975ac4-8436-4fda-b1bc-5543c0f650f6\" (UID: \"27975ac4-8436-4fda-b1bc-5543c0f650f6\") " Apr 21 10:05:44.249562 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249433 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-service-ca" (OuterVolumeSpecName: "service-ca") pod "27975ac4-8436-4fda-b1bc-5543c0f650f6" (UID: "27975ac4-8436-4fda-b1bc-5543c0f650f6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:44.249562 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249510 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "27975ac4-8436-4fda-b1bc-5543c0f650f6" (UID: "27975ac4-8436-4fda-b1bc-5543c0f650f6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:44.249662 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249609 2570 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-service-ca\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:44.249702 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.249654 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-config" (OuterVolumeSpecName: "console-config") pod "27975ac4-8436-4fda-b1bc-5543c0f650f6" (UID: "27975ac4-8436-4fda-b1bc-5543c0f650f6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:44.251388 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.251358 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "27975ac4-8436-4fda-b1bc-5543c0f650f6" (UID: "27975ac4-8436-4fda-b1bc-5543c0f650f6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:05:44.251502 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.251390 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "27975ac4-8436-4fda-b1bc-5543c0f650f6" (UID: "27975ac4-8436-4fda-b1bc-5543c0f650f6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:05:44.251502 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.251418 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27975ac4-8436-4fda-b1bc-5543c0f650f6-kube-api-access-cpn42" (OuterVolumeSpecName: "kube-api-access-cpn42") pod "27975ac4-8436-4fda-b1bc-5543c0f650f6" (UID: "27975ac4-8436-4fda-b1bc-5543c0f650f6"). InnerVolumeSpecName "kube-api-access-cpn42". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:05:44.350152 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.350113 2570 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:44.350152 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.350142 2570 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-oauth-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:44.350152 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.350153 2570 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-oauth-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:44.350356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.350163 2570 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-cpn42\" (UniqueName: \"kubernetes.io/projected/27975ac4-8436-4fda-b1bc-5543c0f650f6-kube-api-access-cpn42\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:44.350356 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.350172 2570 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/27975ac4-8436-4fda-b1bc-5543c0f650f6-console-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:05:44.888235 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.888211 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-dbcfbcdb9-mp2qc_27975ac4-8436-4fda-b1bc-5543c0f650f6/console/0.log" Apr 21 10:05:44.888614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.888248 2570 generic.go:358] "Generic (PLEG): container finished" podID="27975ac4-8436-4fda-b1bc-5543c0f650f6" containerID="d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0" exitCode=2 Apr 21 10:05:44.888614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.888305 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dbcfbcdb9-mp2qc" event={"ID":"27975ac4-8436-4fda-b1bc-5543c0f650f6","Type":"ContainerDied","Data":"d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0"} Apr 21 10:05:44.888614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.888331 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dbcfbcdb9-mp2qc" event={"ID":"27975ac4-8436-4fda-b1bc-5543c0f650f6","Type":"ContainerDied","Data":"048011befffe113514b00602fa02451a6928da4d13dc37bcf472bca8efe0c6c1"} Apr 21 10:05:44.888614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.888346 2570 scope.go:117] "RemoveContainer" containerID="d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0" Apr 21 10:05:44.888614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.888344 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dbcfbcdb9-mp2qc" Apr 21 10:05:44.896136 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.896119 2570 scope.go:117] "RemoveContainer" containerID="d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0" Apr 21 10:05:44.896367 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:05:44.896350 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0\": container with ID starting with d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0 not found: ID does not exist" containerID="d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0" Apr 21 10:05:44.896413 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.896373 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0"} err="failed to get container status \"d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0\": rpc error: code = NotFound desc = could not find container \"d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0\": container with ID starting with d41b8ab3967cf55235508cc5e5343c97db1aebb9477f799ea7fd2cc81a21cbe0 not found: ID does not exist" Apr 21 10:05:44.903937 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.903916 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-dbcfbcdb9-mp2qc"] Apr 21 10:05:44.911543 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:44.911523 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-dbcfbcdb9-mp2qc"] Apr 21 10:05:46.462398 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:05:46.462360 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27975ac4-8436-4fda-b1bc-5543c0f650f6" path="/var/lib/kubelet/pods/27975ac4-8436-4fda-b1bc-5543c0f650f6/volumes" Apr 21 10:06:17.431830 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.431798 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-6d8f8bb896-9tddf"] Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432048 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" containerName="console" Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432060 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" containerName="console" Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432069 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="97661741-da08-4a3a-b481-6e7f7efe69ec" containerName="console" Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432074 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="97661741-da08-4a3a-b481-6e7f7efe69ec" containerName="console" Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432085 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="27975ac4-8436-4fda-b1bc-5543c0f650f6" containerName="console" Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432091 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="27975ac4-8436-4fda-b1bc-5543c0f650f6" containerName="console" Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432134 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="27975ac4-8436-4fda-b1bc-5543c0f650f6" containerName="console" Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432142 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="4eaa8a73-b945-4279-a1ba-e5588cdfdb4e" containerName="console" Apr 21 10:06:17.432351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.432148 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="97661741-da08-4a3a-b481-6e7f7efe69ec" containerName="console" Apr 21 10:06:17.435145 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.435126 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.444725 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.444695 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d8f8bb896-9tddf"] Apr 21 10:06:17.469200 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.469174 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp6fb\" (UniqueName: \"kubernetes.io/projected/88dc251f-76bf-48c5-8c41-a28c5927407f-kube-api-access-rp6fb\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.469342 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.469208 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-serving-cert\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.469342 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.469234 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-console-config\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.469342 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.469295 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-oauth-config\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.469342 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.469340 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-service-ca\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.469499 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.469361 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-oauth-serving-cert\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.469499 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.469377 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-trusted-ca-bundle\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570130 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570106 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-service-ca\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570220 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570138 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-oauth-serving-cert\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570220 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570159 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-trusted-ca-bundle\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570220 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570206 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rp6fb\" (UniqueName: \"kubernetes.io/projected/88dc251f-76bf-48c5-8c41-a28c5927407f-kube-api-access-rp6fb\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570399 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570233 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-serving-cert\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570399 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570256 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-console-config\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570399 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570310 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-oauth-config\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570895 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570865 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-service-ca\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.570997 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.570922 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-oauth-serving-cert\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.571108 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.571013 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-console-config\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.571327 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.571303 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-trusted-ca-bundle\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.572991 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.572964 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-serving-cert\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.572991 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.572980 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-oauth-config\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.578813 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.578790 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp6fb\" (UniqueName: \"kubernetes.io/projected/88dc251f-76bf-48c5-8c41-a28c5927407f-kube-api-access-rp6fb\") pod \"console-6d8f8bb896-9tddf\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.745047 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.744949 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:17.859572 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.859544 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d8f8bb896-9tddf"] Apr 21 10:06:17.861857 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:06:17.861832 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88dc251f_76bf_48c5_8c41_a28c5927407f.slice/crio-7df8953c3da616233df2e8e26aef2be5a7e9ee177ea9f321795ab4826edff17c WatchSource:0}: Error finding container 7df8953c3da616233df2e8e26aef2be5a7e9ee177ea9f321795ab4826edff17c: Status 404 returned error can't find the container with id 7df8953c3da616233df2e8e26aef2be5a7e9ee177ea9f321795ab4826edff17c Apr 21 10:06:17.976503 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.976464 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d8f8bb896-9tddf" event={"ID":"88dc251f-76bf-48c5-8c41-a28c5927407f","Type":"ContainerStarted","Data":"e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9"} Apr 21 10:06:17.976503 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:17.976499 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d8f8bb896-9tddf" event={"ID":"88dc251f-76bf-48c5-8c41-a28c5927407f","Type":"ContainerStarted","Data":"7df8953c3da616233df2e8e26aef2be5a7e9ee177ea9f321795ab4826edff17c"} Apr 21 10:06:18.014429 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:18.014353 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6d8f8bb896-9tddf" podStartSLOduration=1.014340964 podStartE2EDuration="1.014340964s" podCreationTimestamp="2026-04-21 10:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:06:18.013060554 +0000 UTC m=+144.094276413" watchObservedRunningTime="2026-04-21 10:06:18.014340964 +0000 UTC m=+144.095556797" Apr 21 10:06:27.746171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:27.746143 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:27.746171 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:27.746176 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:27.750758 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:27.750737 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:28.010251 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:28.010180 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:06:28.056448 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:28.056421 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-54567b64f6-lq4ds"] Apr 21 10:06:53.080336 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.080251 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-54567b64f6-lq4ds" podUID="082ead76-09e8-47ef-a9fa-ccf7bf22026f" containerName="console" containerID="cri-o://d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f" gracePeriod=15 Apr 21 10:06:53.314703 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.314681 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-54567b64f6-lq4ds_082ead76-09e8-47ef-a9fa-ccf7bf22026f/console/0.log" Apr 21 10:06:53.314806 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.314739 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:06:53.411254 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411225 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-config\") pod \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " Apr 21 10:06:53.411409 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411260 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-oauth-serving-cert\") pod \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " Apr 21 10:06:53.411409 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411317 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-trusted-ca-bundle\") pod \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " Apr 21 10:06:53.411409 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411335 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-service-ca\") pod \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " Apr 21 10:06:53.411409 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411353 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-oauth-config\") pod \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " Apr 21 10:06:53.411409 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411393 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7k8z\" (UniqueName: \"kubernetes.io/projected/082ead76-09e8-47ef-a9fa-ccf7bf22026f-kube-api-access-g7k8z\") pod \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " Apr 21 10:06:53.411667 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411421 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-serving-cert\") pod \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\" (UID: \"082ead76-09e8-47ef-a9fa-ccf7bf22026f\") " Apr 21 10:06:53.411718 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411678 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-config" (OuterVolumeSpecName: "console-config") pod "082ead76-09e8-47ef-a9fa-ccf7bf22026f" (UID: "082ead76-09e8-47ef-a9fa-ccf7bf22026f"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:06:53.411772 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411750 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-service-ca" (OuterVolumeSpecName: "service-ca") pod "082ead76-09e8-47ef-a9fa-ccf7bf22026f" (UID: "082ead76-09e8-47ef-a9fa-ccf7bf22026f"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:06:53.411807 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411760 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "082ead76-09e8-47ef-a9fa-ccf7bf22026f" (UID: "082ead76-09e8-47ef-a9fa-ccf7bf22026f"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:06:53.411807 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.411765 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "082ead76-09e8-47ef-a9fa-ccf7bf22026f" (UID: "082ead76-09e8-47ef-a9fa-ccf7bf22026f"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:06:53.413763 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.413731 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/082ead76-09e8-47ef-a9fa-ccf7bf22026f-kube-api-access-g7k8z" (OuterVolumeSpecName: "kube-api-access-g7k8z") pod "082ead76-09e8-47ef-a9fa-ccf7bf22026f" (UID: "082ead76-09e8-47ef-a9fa-ccf7bf22026f"). InnerVolumeSpecName "kube-api-access-g7k8z". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:06:53.413763 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.413741 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "082ead76-09e8-47ef-a9fa-ccf7bf22026f" (UID: "082ead76-09e8-47ef-a9fa-ccf7bf22026f"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:06:53.413893 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.413762 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "082ead76-09e8-47ef-a9fa-ccf7bf22026f" (UID: "082ead76-09e8-47ef-a9fa-ccf7bf22026f"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:06:53.512588 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.512563 2570 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-g7k8z\" (UniqueName: \"kubernetes.io/projected/082ead76-09e8-47ef-a9fa-ccf7bf22026f-kube-api-access-g7k8z\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:06:53.512588 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.512586 2570 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:06:53.512699 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.512596 2570 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:06:53.512699 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.512605 2570 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-oauth-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:06:53.512699 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.512614 2570 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-trusted-ca-bundle\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:06:53.512699 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.512623 2570 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/082ead76-09e8-47ef-a9fa-ccf7bf22026f-service-ca\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:06:53.512699 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:53.512631 2570 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/082ead76-09e8-47ef-a9fa-ccf7bf22026f-console-oauth-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:06:54.083821 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.083795 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-54567b64f6-lq4ds_082ead76-09e8-47ef-a9fa-ccf7bf22026f/console/0.log" Apr 21 10:06:54.084236 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.083841 2570 generic.go:358] "Generic (PLEG): container finished" podID="082ead76-09e8-47ef-a9fa-ccf7bf22026f" containerID="d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f" exitCode=2 Apr 21 10:06:54.084236 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.083876 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54567b64f6-lq4ds" event={"ID":"082ead76-09e8-47ef-a9fa-ccf7bf22026f","Type":"ContainerDied","Data":"d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f"} Apr 21 10:06:54.084236 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.083903 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54567b64f6-lq4ds" event={"ID":"082ead76-09e8-47ef-a9fa-ccf7bf22026f","Type":"ContainerDied","Data":"f68a770a7c3033255ed7f267e2e88be3c4b8abe5da6ee897f74bcf93e93f771a"} Apr 21 10:06:54.084236 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.083917 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54567b64f6-lq4ds" Apr 21 10:06:54.084236 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.083921 2570 scope.go:117] "RemoveContainer" containerID="d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f" Apr 21 10:06:54.092308 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.092291 2570 scope.go:117] "RemoveContainer" containerID="d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f" Apr 21 10:06:54.092558 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:06:54.092539 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f\": container with ID starting with d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f not found: ID does not exist" containerID="d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f" Apr 21 10:06:54.092606 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.092568 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f"} err="failed to get container status \"d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f\": rpc error: code = NotFound desc = could not find container \"d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f\": container with ID starting with d91aac4729c12cd0767b05f004369519dc0a2c3be5e3a0987cd32ceab27eff1f not found: ID does not exist" Apr 21 10:06:54.104809 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.104786 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-54567b64f6-lq4ds"] Apr 21 10:06:54.110067 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.110049 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-54567b64f6-lq4ds"] Apr 21 10:06:54.462355 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:06:54.462309 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="082ead76-09e8-47ef-a9fa-ccf7bf22026f" path="/var/lib/kubelet/pods/082ead76-09e8-47ef-a9fa-ccf7bf22026f/volumes" Apr 21 10:07:00.627448 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.627412 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-bfr9j"] Apr 21 10:07:00.634381 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.634347 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="082ead76-09e8-47ef-a9fa-ccf7bf22026f" containerName="console" Apr 21 10:07:00.634381 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.634381 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="082ead76-09e8-47ef-a9fa-ccf7bf22026f" containerName="console" Apr 21 10:07:00.634572 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.634465 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="082ead76-09e8-47ef-a9fa-ccf7bf22026f" containerName="console" Apr 21 10:07:00.638858 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.638838 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.643178 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.643131 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 21 10:07:00.643551 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.643528 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-bfr9j"] Apr 21 10:07:00.761210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.761173 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-dbus\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.761387 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.761232 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-kubelet-config\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.761387 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.761292 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-original-pull-secret\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.862052 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.862026 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-dbus\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.862137 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.862074 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-kubelet-config\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.862137 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.862108 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-original-pull-secret\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.862241 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.862219 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-kubelet-config\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.862295 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.862228 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-dbus\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.864487 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.864472 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/f603c14a-fa13-45be-b47d-b74ac1ec1ff5-original-pull-secret\") pod \"global-pull-secret-syncer-bfr9j\" (UID: \"f603c14a-fa13-45be-b47d-b74ac1ec1ff5\") " pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:00.948294 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:00.948220 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-bfr9j" Apr 21 10:07:01.073686 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:01.073654 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-bfr9j"] Apr 21 10:07:01.077074 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:07:01.077050 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf603c14a_fa13_45be_b47d_b74ac1ec1ff5.slice/crio-37e77d11b60b962dfadbc588ee4e8d4a6f6b60015b0039385f7f0f39451d6cc8 WatchSource:0}: Error finding container 37e77d11b60b962dfadbc588ee4e8d4a6f6b60015b0039385f7f0f39451d6cc8: Status 404 returned error can't find the container with id 37e77d11b60b962dfadbc588ee4e8d4a6f6b60015b0039385f7f0f39451d6cc8 Apr 21 10:07:01.103132 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:01.103106 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-bfr9j" event={"ID":"f603c14a-fa13-45be-b47d-b74ac1ec1ff5","Type":"ContainerStarted","Data":"37e77d11b60b962dfadbc588ee4e8d4a6f6b60015b0039385f7f0f39451d6cc8"} Apr 21 10:07:05.118020 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:05.117987 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-bfr9j" event={"ID":"f603c14a-fa13-45be-b47d-b74ac1ec1ff5","Type":"ContainerStarted","Data":"1f080824a0db35a51a4f4540d9a4cc1c81a283d2f4ae23629f7ba0b1049589ed"} Apr 21 10:07:05.134346 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:05.134292 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-bfr9j" podStartSLOduration=1.4105671 podStartE2EDuration="5.134261156s" podCreationTimestamp="2026-04-21 10:07:00 +0000 UTC" firstStartedPulling="2026-04-21 10:07:01.07868767 +0000 UTC m=+187.159903479" lastFinishedPulling="2026-04-21 10:07:04.802381709 +0000 UTC m=+190.883597535" observedRunningTime="2026-04-21 10:07:05.133577496 +0000 UTC m=+191.214793330" watchObservedRunningTime="2026-04-21 10:07:05.134261156 +0000 UTC m=+191.215476985" Apr 21 10:07:52.883243 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:52.883211 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65"] Apr 21 10:07:52.886319 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:52.886299 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:52.889287 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:52.889256 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 21 10:07:52.889408 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:52.889304 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 21 10:07:52.890440 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:52.890427 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-lttsf\"" Apr 21 10:07:52.896854 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:52.896833 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65"] Apr 21 10:07:53.008553 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.008527 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-bundle\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.008664 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.008561 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfc5c\" (UniqueName: \"kubernetes.io/projected/880a8226-41ef-438f-8f83-51d9e9710a0c-kube-api-access-zfc5c\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.008664 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.008595 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-util\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.109487 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.109459 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zfc5c\" (UniqueName: \"kubernetes.io/projected/880a8226-41ef-438f-8f83-51d9e9710a0c-kube-api-access-zfc5c\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.109592 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.109508 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-util\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.109592 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.109554 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-bundle\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.109988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.109966 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-bundle\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.109988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.109981 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-util\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.119280 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.119245 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfc5c\" (UniqueName: \"kubernetes.io/projected/880a8226-41ef-438f-8f83-51d9e9710a0c-kube-api-access-zfc5c\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.195713 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.195648 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:07:53.314826 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:53.314744 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65"] Apr 21 10:07:53.317392 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:07:53.317358 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod880a8226_41ef_438f_8f83_51d9e9710a0c.slice/crio-719ebcf22e8b8109c86f691634f9194f400e7e2b2d1a092704c1514b1c270c7f WatchSource:0}: Error finding container 719ebcf22e8b8109c86f691634f9194f400e7e2b2d1a092704c1514b1c270c7f: Status 404 returned error can't find the container with id 719ebcf22e8b8109c86f691634f9194f400e7e2b2d1a092704c1514b1c270c7f Apr 21 10:07:54.253972 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:54.253931 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" event={"ID":"880a8226-41ef-438f-8f83-51d9e9710a0c","Type":"ContainerStarted","Data":"719ebcf22e8b8109c86f691634f9194f400e7e2b2d1a092704c1514b1c270c7f"} Apr 21 10:07:59.271764 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:59.271727 2570 generic.go:358] "Generic (PLEG): container finished" podID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerID="1f42090a4fb66d47cc7e62484a8e354fe7ca0400abc03701edf347c85b21f160" exitCode=0 Apr 21 10:07:59.272164 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:07:59.271787 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" event={"ID":"880a8226-41ef-438f-8f83-51d9e9710a0c","Type":"ContainerDied","Data":"1f42090a4fb66d47cc7e62484a8e354fe7ca0400abc03701edf347c85b21f160"} Apr 21 10:08:13.309445 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:13.309410 2570 generic.go:358] "Generic (PLEG): container finished" podID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerID="d16d214a922afc9e6198f97def844065a0d759f1639ed6dbccbda7bb3b031907" exitCode=0 Apr 21 10:08:13.309963 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:13.309457 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" event={"ID":"880a8226-41ef-438f-8f83-51d9e9710a0c","Type":"ContainerDied","Data":"d16d214a922afc9e6198f97def844065a0d759f1639ed6dbccbda7bb3b031907"} Apr 21 10:08:20.331704 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:20.331649 2570 generic.go:358] "Generic (PLEG): container finished" podID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerID="d279e22ba8eff0c6cd318020fbeed029870034fdd55dbf06c68870fc20cef757" exitCode=0 Apr 21 10:08:20.332093 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:20.331730 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" event={"ID":"880a8226-41ef-438f-8f83-51d9e9710a0c","Type":"ContainerDied","Data":"d279e22ba8eff0c6cd318020fbeed029870034fdd55dbf06c68870fc20cef757"} Apr 21 10:08:21.454937 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.454914 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:08:21.501956 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.501934 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-bundle\") pod \"880a8226-41ef-438f-8f83-51d9e9710a0c\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " Apr 21 10:08:21.502076 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.501968 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfc5c\" (UniqueName: \"kubernetes.io/projected/880a8226-41ef-438f-8f83-51d9e9710a0c-kube-api-access-zfc5c\") pod \"880a8226-41ef-438f-8f83-51d9e9710a0c\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " Apr 21 10:08:21.502076 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.501991 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-util\") pod \"880a8226-41ef-438f-8f83-51d9e9710a0c\" (UID: \"880a8226-41ef-438f-8f83-51d9e9710a0c\") " Apr 21 10:08:21.502586 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.502562 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-bundle" (OuterVolumeSpecName: "bundle") pod "880a8226-41ef-438f-8f83-51d9e9710a0c" (UID: "880a8226-41ef-438f-8f83-51d9e9710a0c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 10:08:21.504350 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.504329 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880a8226-41ef-438f-8f83-51d9e9710a0c-kube-api-access-zfc5c" (OuterVolumeSpecName: "kube-api-access-zfc5c") pod "880a8226-41ef-438f-8f83-51d9e9710a0c" (UID: "880a8226-41ef-438f-8f83-51d9e9710a0c"). InnerVolumeSpecName "kube-api-access-zfc5c". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:08:21.506876 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.506856 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-util" (OuterVolumeSpecName: "util") pod "880a8226-41ef-438f-8f83-51d9e9710a0c" (UID: "880a8226-41ef-438f-8f83-51d9e9710a0c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 10:08:21.602988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.602969 2570 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-bundle\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:08:21.603071 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.602990 2570 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zfc5c\" (UniqueName: \"kubernetes.io/projected/880a8226-41ef-438f-8f83-51d9e9710a0c-kube-api-access-zfc5c\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:08:21.603071 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:21.602999 2570 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/880a8226-41ef-438f-8f83-51d9e9710a0c-util\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:08:22.340512 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:22.340473 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" event={"ID":"880a8226-41ef-438f-8f83-51d9e9710a0c","Type":"ContainerDied","Data":"719ebcf22e8b8109c86f691634f9194f400e7e2b2d1a092704c1514b1c270c7f"} Apr 21 10:08:22.340512 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:22.340505 2570 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="719ebcf22e8b8109c86f691634f9194f400e7e2b2d1a092704c1514b1c270c7f" Apr 21 10:08:22.340512 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:22.340512 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29ctpq65" Apr 21 10:08:24.711694 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.711656 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz"] Apr 21 10:08:24.712073 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.711925 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerName="extract" Apr 21 10:08:24.712073 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.711937 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerName="extract" Apr 21 10:08:24.712073 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.711949 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerName="pull" Apr 21 10:08:24.712073 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.711954 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerName="pull" Apr 21 10:08:24.712073 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.711963 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerName="util" Apr 21 10:08:24.712073 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.711968 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerName="util" Apr 21 10:08:24.712073 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.712011 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="880a8226-41ef-438f-8f83-51d9e9710a0c" containerName="extract" Apr 21 10:08:24.715067 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.715048 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:24.718627 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.718564 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-keda\"/\"kube-root-ca.crt\"" Apr 21 10:08:24.718627 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.718597 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-keda\"/\"openshift-service-ca.crt\"" Apr 21 10:08:24.718627 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.718619 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"kedaorg-certs\"" Apr 21 10:08:24.718865 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.718626 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"custom-metrics-autoscaler-operator-dockercfg-xsntm\"" Apr 21 10:08:24.729874 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.729851 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz"] Apr 21 10:08:24.828489 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.828462 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln4hf\" (UniqueName: \"kubernetes.io/projected/51c7cba6-a16b-4c9c-951b-68692250f6e2-kube-api-access-ln4hf\") pod \"custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz\" (UID: \"51c7cba6-a16b-4c9c-951b-68692250f6e2\") " pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:24.828607 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.828496 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/secret/51c7cba6-a16b-4c9c-951b-68692250f6e2-certificates\") pod \"custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz\" (UID: \"51c7cba6-a16b-4c9c-951b-68692250f6e2\") " pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:24.929667 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.929643 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ln4hf\" (UniqueName: \"kubernetes.io/projected/51c7cba6-a16b-4c9c-951b-68692250f6e2-kube-api-access-ln4hf\") pod \"custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz\" (UID: \"51c7cba6-a16b-4c9c-951b-68692250f6e2\") " pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:24.929757 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.929675 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/secret/51c7cba6-a16b-4c9c-951b-68692250f6e2-certificates\") pod \"custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz\" (UID: \"51c7cba6-a16b-4c9c-951b-68692250f6e2\") " pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:24.932134 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.932116 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certificates\" (UniqueName: \"kubernetes.io/secret/51c7cba6-a16b-4c9c-951b-68692250f6e2-certificates\") pod \"custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz\" (UID: \"51c7cba6-a16b-4c9c-951b-68692250f6e2\") " pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:24.938256 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:24.938233 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln4hf\" (UniqueName: \"kubernetes.io/projected/51c7cba6-a16b-4c9c-951b-68692250f6e2-kube-api-access-ln4hf\") pod \"custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz\" (UID: \"51c7cba6-a16b-4c9c-951b-68692250f6e2\") " pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:25.024554 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:25.024497 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:25.142743 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:25.142712 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz"] Apr 21 10:08:25.146806 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:08:25.146773 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51c7cba6_a16b_4c9c_951b_68692250f6e2.slice/crio-278d5f0f2ab8977f0c2ad8378bfe0edf45424ec2cbc8ce3c19c6a2d400a2f0d1 WatchSource:0}: Error finding container 278d5f0f2ab8977f0c2ad8378bfe0edf45424ec2cbc8ce3c19c6a2d400a2f0d1: Status 404 returned error can't find the container with id 278d5f0f2ab8977f0c2ad8378bfe0edf45424ec2cbc8ce3c19c6a2d400a2f0d1 Apr 21 10:08:25.350089 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:25.350002 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" event={"ID":"51c7cba6-a16b-4c9c-951b-68692250f6e2","Type":"ContainerStarted","Data":"278d5f0f2ab8977f0c2ad8378bfe0edf45424ec2cbc8ce3c19c6a2d400a2f0d1"} Apr 21 10:08:28.985145 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:28.985110 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-keda/keda-operator-ffbb595cb-9kgk6"] Apr 21 10:08:28.988361 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:28.988340 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:28.990840 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:28.990818 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-keda\"/\"keda-ocp-cabundle\"" Apr 21 10:08:28.990967 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:28.990824 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"keda-operator-certs\"" Apr 21 10:08:28.990967 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:28.990869 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"keda-operator-dockercfg-wfrx8\"" Apr 21 10:08:28.995849 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:28.995828 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-operator-ffbb595cb-9kgk6"] Apr 21 10:08:29.061819 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.061798 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmd7c\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-kube-api-access-jmd7c\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.061926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.061844 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cabundle0\" (UniqueName: \"kubernetes.io/configmap/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-cabundle0\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.061926 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.061871 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.162911 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.162884 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jmd7c\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-kube-api-access-jmd7c\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.163051 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.162932 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cabundle0\" (UniqueName: \"kubernetes.io/configmap/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-cabundle0\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.163112 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.163067 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.163235 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.163219 2570 secret.go:281] references non-existent secret key: ca.crt Apr 21 10:08:29.163300 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.163241 2570 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: ca.crt Apr 21 10:08:29.163300 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.163255 2570 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-operator-ffbb595cb-9kgk6: references non-existent secret key: ca.crt Apr 21 10:08:29.163372 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.163332 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates podName:36e8d4ce-9db5-41a3-a45c-c918eb1da42d nodeName:}" failed. No retries permitted until 2026-04-21 10:08:29.663315294 +0000 UTC m=+275.744531107 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates") pod "keda-operator-ffbb595cb-9kgk6" (UID: "36e8d4ce-9db5-41a3-a45c-c918eb1da42d") : references non-existent secret key: ca.crt Apr 21 10:08:29.163608 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.163593 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cabundle0\" (UniqueName: \"kubernetes.io/configmap/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-cabundle0\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.171070 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.171043 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmd7c\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-kube-api-access-jmd7c\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.249155 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.249086 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv"] Apr 21 10:08:29.252435 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.252417 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.256405 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.256380 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"keda-metrics-apiserver-certs\"" Apr 21 10:08:29.283411 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.283390 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv"] Apr 21 10:08:29.362118 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.362089 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" event={"ID":"51c7cba6-a16b-4c9c-951b-68692250f6e2","Type":"ContainerStarted","Data":"e52a433662248b374261541896d4f84a560aa0b42291399e899269affc9df516"} Apr 21 10:08:29.362350 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.362331 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:29.364681 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.364662 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjgz2\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-kube-api-access-qjgz2\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.364762 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.364714 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.364802 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.364772 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"temp-vol\" (UniqueName: \"kubernetes.io/empty-dir/f55be428-8a94-4bd2-b578-ace085b45c4d-temp-vol\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.383074 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.383028 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" podStartSLOduration=2.019071761 podStartE2EDuration="5.383017793s" podCreationTimestamp="2026-04-21 10:08:24 +0000 UTC" firstStartedPulling="2026-04-21 10:08:25.149019373 +0000 UTC m=+271.230235183" lastFinishedPulling="2026-04-21 10:08:28.512965405 +0000 UTC m=+274.594181215" observedRunningTime="2026-04-21 10:08:29.381888417 +0000 UTC m=+275.463104250" watchObservedRunningTime="2026-04-21 10:08:29.383017793 +0000 UTC m=+275.464233625" Apr 21 10:08:29.465381 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.465356 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qjgz2\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-kube-api-access-qjgz2\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.465482 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.465422 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.465547 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.465501 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"temp-vol\" (UniqueName: \"kubernetes.io/empty-dir/f55be428-8a94-4bd2-b578-ace085b45c4d-temp-vol\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.465611 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.465568 2570 secret.go:281] references non-existent secret key: tls.crt Apr 21 10:08:29.465611 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.465585 2570 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: tls.crt Apr 21 10:08:29.465611 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.465600 2570 projected.go:264] Couldn't get secret openshift-keda/keda-metrics-apiserver-certs: secret "keda-metrics-apiserver-certs" not found Apr 21 10:08:29.465757 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.465614 2570 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv: [references non-existent secret key: tls.crt, secret "keda-metrics-apiserver-certs" not found] Apr 21 10:08:29.465757 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.465668 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates podName:f55be428-8a94-4bd2-b578-ace085b45c4d nodeName:}" failed. No retries permitted until 2026-04-21 10:08:29.965653334 +0000 UTC m=+276.046869144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates") pod "keda-metrics-apiserver-7c9f485588-xdtsv" (UID: "f55be428-8a94-4bd2-b578-ace085b45c4d") : [references non-existent secret key: tls.crt, secret "keda-metrics-apiserver-certs" not found] Apr 21 10:08:29.465906 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.465880 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"temp-vol\" (UniqueName: \"kubernetes.io/empty-dir/f55be428-8a94-4bd2-b578-ace085b45c4d-temp-vol\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.473756 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.473737 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjgz2\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-kube-api-access-qjgz2\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.541612 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.541562 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-keda/keda-admission-cf49989db-xldfn"] Apr 21 10:08:29.544650 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.544635 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:29.547049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.547033 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"keda-admission-webhooks-certs\"" Apr 21 10:08:29.552891 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.552871 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-admission-cf49989db-xldfn"] Apr 21 10:08:29.666872 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.666847 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/05014fb5-5feb-4853-9590-37531f6dddb7-certificates\") pod \"keda-admission-cf49989db-xldfn\" (UID: \"05014fb5-5feb-4853-9590-37531f6dddb7\") " pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:29.667018 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.666877 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4pl6\" (UniqueName: \"kubernetes.io/projected/05014fb5-5feb-4853-9590-37531f6dddb7-kube-api-access-m4pl6\") pod \"keda-admission-cf49989db-xldfn\" (UID: \"05014fb5-5feb-4853-9590-37531f6dddb7\") " pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:29.667018 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.666909 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:29.667018 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.666995 2570 secret.go:281] references non-existent secret key: ca.crt Apr 21 10:08:29.667018 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.667007 2570 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: ca.crt Apr 21 10:08:29.667018 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.667015 2570 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-operator-ffbb595cb-9kgk6: references non-existent secret key: ca.crt Apr 21 10:08:29.667175 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.667071 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates podName:36e8d4ce-9db5-41a3-a45c-c918eb1da42d nodeName:}" failed. No retries permitted until 2026-04-21 10:08:30.667056066 +0000 UTC m=+276.748271892 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates") pod "keda-operator-ffbb595cb-9kgk6" (UID: "36e8d4ce-9db5-41a3-a45c-c918eb1da42d") : references non-existent secret key: ca.crt Apr 21 10:08:29.768084 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.768047 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/05014fb5-5feb-4853-9590-37531f6dddb7-certificates\") pod \"keda-admission-cf49989db-xldfn\" (UID: \"05014fb5-5feb-4853-9590-37531f6dddb7\") " pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:29.768253 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.768090 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m4pl6\" (UniqueName: \"kubernetes.io/projected/05014fb5-5feb-4853-9590-37531f6dddb7-kube-api-access-m4pl6\") pod \"keda-admission-cf49989db-xldfn\" (UID: \"05014fb5-5feb-4853-9590-37531f6dddb7\") " pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:29.772015 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.771984 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/05014fb5-5feb-4853-9590-37531f6dddb7-certificates\") pod \"keda-admission-cf49989db-xldfn\" (UID: \"05014fb5-5feb-4853-9590-37531f6dddb7\") " pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:29.777286 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.777237 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4pl6\" (UniqueName: \"kubernetes.io/projected/05014fb5-5feb-4853-9590-37531f6dddb7-kube-api-access-m4pl6\") pod \"keda-admission-cf49989db-xldfn\" (UID: \"05014fb5-5feb-4853-9590-37531f6dddb7\") " pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:29.854196 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.854164 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:29.970803 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.970754 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:29.970965 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.970860 2570 secret.go:281] references non-existent secret key: tls.crt Apr 21 10:08:29.970965 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.970872 2570 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: tls.crt Apr 21 10:08:29.970965 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.970888 2570 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv: references non-existent secret key: tls.crt Apr 21 10:08:29.970965 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:29.970938 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates podName:f55be428-8a94-4bd2-b578-ace085b45c4d nodeName:}" failed. No retries permitted until 2026-04-21 10:08:30.970926112 +0000 UTC m=+277.052141922 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates") pod "keda-metrics-apiserver-7c9f485588-xdtsv" (UID: "f55be428-8a94-4bd2-b578-ace085b45c4d") : references non-existent secret key: tls.crt Apr 21 10:08:29.998469 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:29.998415 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-admission-cf49989db-xldfn"] Apr 21 10:08:30.365410 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:30.365375 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-admission-cf49989db-xldfn" event={"ID":"05014fb5-5feb-4853-9590-37531f6dddb7","Type":"ContainerStarted","Data":"026ea05531864273b4ac7f20e1c8a82e4ffd93771d21c2c65d0ef66e00ecec63"} Apr 21 10:08:30.675885 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:30.675809 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:30.676024 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:30.675926 2570 secret.go:281] references non-existent secret key: ca.crt Apr 21 10:08:30.676024 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:30.675943 2570 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: ca.crt Apr 21 10:08:30.676024 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:30.675953 2570 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-operator-ffbb595cb-9kgk6: references non-existent secret key: ca.crt Apr 21 10:08:30.676024 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:30.676000 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates podName:36e8d4ce-9db5-41a3-a45c-c918eb1da42d nodeName:}" failed. No retries permitted until 2026-04-21 10:08:32.675987895 +0000 UTC m=+278.757203704 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates") pod "keda-operator-ffbb595cb-9kgk6" (UID: "36e8d4ce-9db5-41a3-a45c-c918eb1da42d") : references non-existent secret key: ca.crt Apr 21 10:08:30.978950 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:30.978865 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:30.979125 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:30.979013 2570 secret.go:281] references non-existent secret key: tls.crt Apr 21 10:08:30.979125 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:30.979037 2570 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: tls.crt Apr 21 10:08:30.979125 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:30.979059 2570 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv: references non-existent secret key: tls.crt Apr 21 10:08:30.979311 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:30.979132 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates podName:f55be428-8a94-4bd2-b578-ace085b45c4d nodeName:}" failed. No retries permitted until 2026-04-21 10:08:32.979107541 +0000 UTC m=+279.060323350 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates") pod "keda-metrics-apiserver-7c9f485588-xdtsv" (UID: "f55be428-8a94-4bd2-b578-ace085b45c4d") : references non-existent secret key: tls.crt Apr 21 10:08:32.371621 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:32.371589 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-admission-cf49989db-xldfn" event={"ID":"05014fb5-5feb-4853-9590-37531f6dddb7","Type":"ContainerStarted","Data":"55f851bef6fd2b8205e9e62a4933543281cd6404248ed4ca9b20fa2ac04b9267"} Apr 21 10:08:32.371988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:32.371726 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:32.387965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:32.387914 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-keda/keda-admission-cf49989db-xldfn" podStartSLOduration=2.049323795 podStartE2EDuration="3.387901206s" podCreationTimestamp="2026-04-21 10:08:29 +0000 UTC" firstStartedPulling="2026-04-21 10:08:30.004642466 +0000 UTC m=+276.085858277" lastFinishedPulling="2026-04-21 10:08:31.343219878 +0000 UTC m=+277.424435688" observedRunningTime="2026-04-21 10:08:32.386676433 +0000 UTC m=+278.467892268" watchObservedRunningTime="2026-04-21 10:08:32.387901206 +0000 UTC m=+278.469117037" Apr 21 10:08:32.692708 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:32.692619 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:32.692860 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:32.692770 2570 secret.go:281] references non-existent secret key: ca.crt Apr 21 10:08:32.692860 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:32.692789 2570 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: ca.crt Apr 21 10:08:32.692860 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:32.692800 2570 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-operator-ffbb595cb-9kgk6: references non-existent secret key: ca.crt Apr 21 10:08:32.693016 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:32.692862 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates podName:36e8d4ce-9db5-41a3-a45c-c918eb1da42d nodeName:}" failed. No retries permitted until 2026-04-21 10:08:36.692843667 +0000 UTC m=+282.774059481 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates") pod "keda-operator-ffbb595cb-9kgk6" (UID: "36e8d4ce-9db5-41a3-a45c-c918eb1da42d") : references non-existent secret key: ca.crt Apr 21 10:08:32.994771 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:32.994683 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:32.994922 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:32.994819 2570 secret.go:281] references non-existent secret key: tls.crt Apr 21 10:08:32.994922 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:32.994836 2570 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: tls.crt Apr 21 10:08:32.994922 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:32.994856 2570 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv: references non-existent secret key: tls.crt Apr 21 10:08:32.994922 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:08:32.994911 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates podName:f55be428-8a94-4bd2-b578-ace085b45c4d nodeName:}" failed. No retries permitted until 2026-04-21 10:08:36.99489586 +0000 UTC m=+283.076111670 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates") pod "keda-metrics-apiserver-7c9f485588-xdtsv" (UID: "f55be428-8a94-4bd2-b578-ace085b45c4d") : references non-existent secret key: tls.crt Apr 21 10:08:36.720202 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:36.720164 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:36.722857 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:36.722834 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/36e8d4ce-9db5-41a3-a45c-c918eb1da42d-certificates\") pod \"keda-operator-ffbb595cb-9kgk6\" (UID: \"36e8d4ce-9db5-41a3-a45c-c918eb1da42d\") " pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:36.799518 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:36.799494 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:36.915113 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:36.915083 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-operator-ffbb595cb-9kgk6"] Apr 21 10:08:36.917938 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:08:36.917894 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36e8d4ce_9db5_41a3_a45c_c918eb1da42d.slice/crio-af182cefba53449eb91e045dac19e4ddd90a92a041edacae02a807031b2fdfc4 WatchSource:0}: Error finding container af182cefba53449eb91e045dac19e4ddd90a92a041edacae02a807031b2fdfc4: Status 404 returned error can't find the container with id af182cefba53449eb91e045dac19e4ddd90a92a041edacae02a807031b2fdfc4 Apr 21 10:08:37.022785 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:37.022718 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:37.025248 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:37.025215 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/f55be428-8a94-4bd2-b578-ace085b45c4d-certificates\") pod \"keda-metrics-apiserver-7c9f485588-xdtsv\" (UID: \"f55be428-8a94-4bd2-b578-ace085b45c4d\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:37.062037 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:37.062013 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:37.172701 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:37.172672 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv"] Apr 21 10:08:37.175103 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:08:37.175075 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf55be428_8a94_4bd2_b578_ace085b45c4d.slice/crio-2dac8e3379cfaf34946962da6db277bc50d303f860af7ab14071183a03ba5f47 WatchSource:0}: Error finding container 2dac8e3379cfaf34946962da6db277bc50d303f860af7ab14071183a03ba5f47: Status 404 returned error can't find the container with id 2dac8e3379cfaf34946962da6db277bc50d303f860af7ab14071183a03ba5f47 Apr 21 10:08:37.386527 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:37.386500 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" event={"ID":"f55be428-8a94-4bd2-b578-ace085b45c4d","Type":"ContainerStarted","Data":"2dac8e3379cfaf34946962da6db277bc50d303f860af7ab14071183a03ba5f47"} Apr 21 10:08:37.387410 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:37.387388 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" event={"ID":"36e8d4ce-9db5-41a3-a45c-c918eb1da42d","Type":"ContainerStarted","Data":"af182cefba53449eb91e045dac19e4ddd90a92a041edacae02a807031b2fdfc4"} Apr 21 10:08:41.401813 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:41.401780 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" event={"ID":"f55be428-8a94-4bd2-b578-ace085b45c4d","Type":"ContainerStarted","Data":"de930aa52555693b5f23dac0b74c30a1a8b5479100f5d807cb8d3896815d85f0"} Apr 21 10:08:41.402289 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:41.401887 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:41.403106 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:41.403082 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" event={"ID":"36e8d4ce-9db5-41a3-a45c-c918eb1da42d","Type":"ContainerStarted","Data":"101621ba23db823f660c32cfbd825c1310a56ff6b22ab3a63aa26ffdf5edea5e"} Apr 21 10:08:41.403293 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:41.403257 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:08:41.418842 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:41.418802 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" podStartSLOduration=8.934743412 podStartE2EDuration="12.418791071s" podCreationTimestamp="2026-04-21 10:08:29 +0000 UTC" firstStartedPulling="2026-04-21 10:08:37.176523944 +0000 UTC m=+283.257739755" lastFinishedPulling="2026-04-21 10:08:40.660571602 +0000 UTC m=+286.741787414" observedRunningTime="2026-04-21 10:08:41.417348485 +0000 UTC m=+287.498564317" watchObservedRunningTime="2026-04-21 10:08:41.418791071 +0000 UTC m=+287.500006903" Apr 21 10:08:41.434616 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:41.434573 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" podStartSLOduration=9.687880464 podStartE2EDuration="13.43456171s" podCreationTimestamp="2026-04-21 10:08:28 +0000 UTC" firstStartedPulling="2026-04-21 10:08:36.919293853 +0000 UTC m=+283.000509680" lastFinishedPulling="2026-04-21 10:08:40.665975115 +0000 UTC m=+286.747190926" observedRunningTime="2026-04-21 10:08:41.433579912 +0000 UTC m=+287.514795750" watchObservedRunningTime="2026-04-21 10:08:41.43456171 +0000 UTC m=+287.515777541" Apr 21 10:08:50.367851 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:50.367821 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-keda/custom-metrics-autoscaler-operator-bbf89fd5d-d4lgz" Apr 21 10:08:52.410845 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:52.410817 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-xdtsv" Apr 21 10:08:53.377673 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:53.377641 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-keda/keda-admission-cf49989db-xldfn" Apr 21 10:08:54.337996 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:08:54.337971 2570 kubelet.go:1628] "Image garbage collection succeeded" Apr 21 10:09:02.408475 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:02.408441 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-keda/keda-operator-ffbb595cb-9kgk6" Apr 21 10:09:35.774769 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.774737 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve/kserve-controller-manager-84685cd884-bsvvv"] Apr 21 10:09:35.780244 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.780223 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:35.783145 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.783114 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kserve\"/\"openshift-service-ca.crt\"" Apr 21 10:09:35.783257 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.783121 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"kserve-controller-manager-dockercfg-kpbh7\"" Apr 21 10:09:35.783257 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.783126 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"kserve-webhook-server-cert\"" Apr 21 10:09:35.783382 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.783318 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kserve\"/\"kube-root-ca.crt\"" Apr 21 10:09:35.783662 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.783645 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b"] Apr 21 10:09:35.789646 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.788775 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:35.789954 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.789930 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/kserve-controller-manager-84685cd884-bsvvv"] Apr 21 10:09:35.791889 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.791867 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"llmisvc-webhook-server-cert\"" Apr 21 10:09:35.791998 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.791888 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"llmisvc-controller-manager-dockercfg-zqdzk\"" Apr 21 10:09:35.804074 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.804051 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b"] Apr 21 10:09:35.815355 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.815331 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve/seaweedfs-86cc847c5c-sxtd8"] Apr 21 10:09:35.817628 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.817610 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:35.822139 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.822118 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"mlpipeline-s3-artifact\"" Apr 21 10:09:35.822316 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.822142 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"default-dockercfg-kw59d\"" Apr 21 10:09:35.828918 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.828883 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/seaweedfs-86cc847c5c-sxtd8"] Apr 21 10:09:35.920364 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.920341 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8lkq\" (UniqueName: \"kubernetes.io/projected/5a4abe0e-37cf-4685-86d4-c476caea9e49-kube-api-access-n8lkq\") pod \"seaweedfs-86cc847c5c-sxtd8\" (UID: \"5a4abe0e-37cf-4685-86d4-c476caea9e49\") " pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:35.920498 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.920374 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-cert\") pod \"llmisvc-controller-manager-68cc5db7c4-rbt9b\" (UID: \"8f4d63b7-adda-4734-a417-0e8e17d8a6bb\") " pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:35.920498 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.920396 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/5a4abe0e-37cf-4685-86d4-c476caea9e49-data\") pod \"seaweedfs-86cc847c5c-sxtd8\" (UID: \"5a4abe0e-37cf-4685-86d4-c476caea9e49\") " pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:35.920498 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.920414 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lphs5\" (UniqueName: \"kubernetes.io/projected/44eff084-d7a3-48cf-a3d8-270b030d90f7-kube-api-access-lphs5\") pod \"kserve-controller-manager-84685cd884-bsvvv\" (UID: \"44eff084-d7a3-48cf-a3d8-270b030d90f7\") " pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:35.920614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.920506 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44eff084-d7a3-48cf-a3d8-270b030d90f7-cert\") pod \"kserve-controller-manager-84685cd884-bsvvv\" (UID: \"44eff084-d7a3-48cf-a3d8-270b030d90f7\") " pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:35.920614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:35.920557 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsck6\" (UniqueName: \"kubernetes.io/projected/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-kube-api-access-dsck6\") pod \"llmisvc-controller-manager-68cc5db7c4-rbt9b\" (UID: \"8f4d63b7-adda-4734-a417-0e8e17d8a6bb\") " pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:36.021036 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.021010 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dsck6\" (UniqueName: \"kubernetes.io/projected/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-kube-api-access-dsck6\") pod \"llmisvc-controller-manager-68cc5db7c4-rbt9b\" (UID: \"8f4d63b7-adda-4734-a417-0e8e17d8a6bb\") " pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:36.021176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.021053 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-n8lkq\" (UniqueName: \"kubernetes.io/projected/5a4abe0e-37cf-4685-86d4-c476caea9e49-kube-api-access-n8lkq\") pod \"seaweedfs-86cc847c5c-sxtd8\" (UID: \"5a4abe0e-37cf-4685-86d4-c476caea9e49\") " pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:36.021176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.021074 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-cert\") pod \"llmisvc-controller-manager-68cc5db7c4-rbt9b\" (UID: \"8f4d63b7-adda-4734-a417-0e8e17d8a6bb\") " pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:36.021176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.021095 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/5a4abe0e-37cf-4685-86d4-c476caea9e49-data\") pod \"seaweedfs-86cc847c5c-sxtd8\" (UID: \"5a4abe0e-37cf-4685-86d4-c476caea9e49\") " pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:36.021176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.021124 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lphs5\" (UniqueName: \"kubernetes.io/projected/44eff084-d7a3-48cf-a3d8-270b030d90f7-kube-api-access-lphs5\") pod \"kserve-controller-manager-84685cd884-bsvvv\" (UID: \"44eff084-d7a3-48cf-a3d8-270b030d90f7\") " pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:36.021176 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.021163 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44eff084-d7a3-48cf-a3d8-270b030d90f7-cert\") pod \"kserve-controller-manager-84685cd884-bsvvv\" (UID: \"44eff084-d7a3-48cf-a3d8-270b030d90f7\") " pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:36.021454 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:09:36.021198 2570 secret.go:189] Couldn't get secret kserve/llmisvc-webhook-server-cert: secret "llmisvc-webhook-server-cert" not found Apr 21 10:09:36.021454 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:09:36.021304 2570 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-cert podName:8f4d63b7-adda-4734-a417-0e8e17d8a6bb nodeName:}" failed. No retries permitted until 2026-04-21 10:09:36.521258101 +0000 UTC m=+342.602473925 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-cert") pod "llmisvc-controller-manager-68cc5db7c4-rbt9b" (UID: "8f4d63b7-adda-4734-a417-0e8e17d8a6bb") : secret "llmisvc-webhook-server-cert" not found Apr 21 10:09:36.021603 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.021582 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/5a4abe0e-37cf-4685-86d4-c476caea9e49-data\") pod \"seaweedfs-86cc847c5c-sxtd8\" (UID: \"5a4abe0e-37cf-4685-86d4-c476caea9e49\") " pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:36.023655 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.023639 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44eff084-d7a3-48cf-a3d8-270b030d90f7-cert\") pod \"kserve-controller-manager-84685cd884-bsvvv\" (UID: \"44eff084-d7a3-48cf-a3d8-270b030d90f7\") " pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:36.032553 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.032505 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsck6\" (UniqueName: \"kubernetes.io/projected/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-kube-api-access-dsck6\") pod \"llmisvc-controller-manager-68cc5db7c4-rbt9b\" (UID: \"8f4d63b7-adda-4734-a417-0e8e17d8a6bb\") " pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:36.036486 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.036466 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lphs5\" (UniqueName: \"kubernetes.io/projected/44eff084-d7a3-48cf-a3d8-270b030d90f7-kube-api-access-lphs5\") pod \"kserve-controller-manager-84685cd884-bsvvv\" (UID: \"44eff084-d7a3-48cf-a3d8-270b030d90f7\") " pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:36.039797 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.039779 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8lkq\" (UniqueName: \"kubernetes.io/projected/5a4abe0e-37cf-4685-86d4-c476caea9e49-kube-api-access-n8lkq\") pod \"seaweedfs-86cc847c5c-sxtd8\" (UID: \"5a4abe0e-37cf-4685-86d4-c476caea9e49\") " pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:36.096707 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.096662 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:36.126692 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.126671 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:36.225239 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.225221 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/kserve-controller-manager-84685cd884-bsvvv"] Apr 21 10:09:36.227808 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:09:36.227783 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44eff084_d7a3_48cf_a3d8_270b030d90f7.slice/crio-dcbe822383bbc9df241b84bf6008e43642199e5a375215b26502c931e72405ef WatchSource:0}: Error finding container dcbe822383bbc9df241b84bf6008e43642199e5a375215b26502c931e72405ef: Status 404 returned error can't find the container with id dcbe822383bbc9df241b84bf6008e43642199e5a375215b26502c931e72405ef Apr 21 10:09:36.229213 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.229194 2570 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 10:09:36.254648 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.254628 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/seaweedfs-86cc847c5c-sxtd8"] Apr 21 10:09:36.256845 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:09:36.256820 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a4abe0e_37cf_4685_86d4_c476caea9e49.slice/crio-f608117ba705926a9d5dab6d2f6efc88cee933c96d29e2b493899310183b145a WatchSource:0}: Error finding container f608117ba705926a9d5dab6d2f6efc88cee933c96d29e2b493899310183b145a: Status 404 returned error can't find the container with id f608117ba705926a9d5dab6d2f6efc88cee933c96d29e2b493899310183b145a Apr 21 10:09:36.525256 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.525232 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-cert\") pod \"llmisvc-controller-manager-68cc5db7c4-rbt9b\" (UID: \"8f4d63b7-adda-4734-a417-0e8e17d8a6bb\") " pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:36.527590 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.527570 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f4d63b7-adda-4734-a417-0e8e17d8a6bb-cert\") pod \"llmisvc-controller-manager-68cc5db7c4-rbt9b\" (UID: \"8f4d63b7-adda-4734-a417-0e8e17d8a6bb\") " pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:36.546068 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.546036 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/seaweedfs-86cc847c5c-sxtd8" event={"ID":"5a4abe0e-37cf-4685-86d4-c476caea9e49","Type":"ContainerStarted","Data":"f608117ba705926a9d5dab6d2f6efc88cee933c96d29e2b493899310183b145a"} Apr 21 10:09:36.546931 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.546911 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" event={"ID":"44eff084-d7a3-48cf-a3d8-270b030d90f7","Type":"ContainerStarted","Data":"dcbe822383bbc9df241b84bf6008e43642199e5a375215b26502c931e72405ef"} Apr 21 10:09:36.703781 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.703750 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:36.868341 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:36.868280 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b"] Apr 21 10:09:37.560206 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:37.560162 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" event={"ID":"8f4d63b7-adda-4734-a417-0e8e17d8a6bb","Type":"ContainerStarted","Data":"c311499858f7437dc9dda542c46bc2e75fde9353cd68b10d4f3848c08e742018"} Apr 21 10:09:41.575530 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.575495 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" event={"ID":"8f4d63b7-adda-4734-a417-0e8e17d8a6bb","Type":"ContainerStarted","Data":"f5555dc1da0af5741c2c518607def9ce79efb2bd67f6a8e3a882138c7f113beb"} Apr 21 10:09:41.575971 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.575583 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:09:41.576823 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.576799 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" event={"ID":"44eff084-d7a3-48cf-a3d8-270b030d90f7","Type":"ContainerStarted","Data":"1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d"} Apr 21 10:09:41.576940 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.576880 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:09:41.578012 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.577991 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/seaweedfs-86cc847c5c-sxtd8" event={"ID":"5a4abe0e-37cf-4685-86d4-c476caea9e49","Type":"ContainerStarted","Data":"d98cbb32cdf8045c5d996007381ba60ca0531fcf3c95178033c69ef1bcb35cc0"} Apr 21 10:09:41.578125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.578114 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:09:41.592414 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.592374 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" podStartSLOduration=2.334909178 podStartE2EDuration="6.592363998s" podCreationTimestamp="2026-04-21 10:09:35 +0000 UTC" firstStartedPulling="2026-04-21 10:09:36.876220232 +0000 UTC m=+342.957436046" lastFinishedPulling="2026-04-21 10:09:41.13367505 +0000 UTC m=+347.214890866" observedRunningTime="2026-04-21 10:09:41.590329466 +0000 UTC m=+347.671545298" watchObservedRunningTime="2026-04-21 10:09:41.592363998 +0000 UTC m=+347.673579867" Apr 21 10:09:41.604960 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.604919 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve/seaweedfs-86cc847c5c-sxtd8" podStartSLOduration=1.7658377189999999 podStartE2EDuration="6.604908769s" podCreationTimestamp="2026-04-21 10:09:35 +0000 UTC" firstStartedPulling="2026-04-21 10:09:36.258153428 +0000 UTC m=+342.339369242" lastFinishedPulling="2026-04-21 10:09:41.097224469 +0000 UTC m=+347.178440292" observedRunningTime="2026-04-21 10:09:41.604117797 +0000 UTC m=+347.685333629" watchObservedRunningTime="2026-04-21 10:09:41.604908769 +0000 UTC m=+347.686124654" Apr 21 10:09:41.618813 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:41.618778 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" podStartSLOduration=2.683155783 podStartE2EDuration="6.61876794s" podCreationTimestamp="2026-04-21 10:09:35 +0000 UTC" firstStartedPulling="2026-04-21 10:09:36.229342181 +0000 UTC m=+342.310557991" lastFinishedPulling="2026-04-21 10:09:40.164954333 +0000 UTC m=+346.246170148" observedRunningTime="2026-04-21 10:09:41.617716118 +0000 UTC m=+347.698931951" watchObservedRunningTime="2026-04-21 10:09:41.61876794 +0000 UTC m=+347.699983772" Apr 21 10:09:47.583139 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:09:47.583046 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve/seaweedfs-86cc847c5c-sxtd8" Apr 21 10:10:12.583660 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:12.583630 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve/llmisvc-controller-manager-68cc5db7c4-rbt9b" Apr 21 10:10:12.586614 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:12.586595 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:10:13.765665 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.765630 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve/kserve-controller-manager-84685cd884-bsvvv"] Apr 21 10:10:13.766046 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.765847 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" podUID="44eff084-d7a3-48cf-a3d8-270b030d90f7" containerName="manager" containerID="cri-o://1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d" gracePeriod=10 Apr 21 10:10:13.785201 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.785178 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve/kserve-controller-manager-84685cd884-qk94h"] Apr 21 10:10:13.788338 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.788259 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:13.798893 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.798869 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/kserve-controller-manager-84685cd884-qk94h"] Apr 21 10:10:13.870494 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.870471 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmz8b\" (UniqueName: \"kubernetes.io/projected/f6dd56d8-2f6d-4945-9a19-4f3743409c0e-kube-api-access-lmz8b\") pod \"kserve-controller-manager-84685cd884-qk94h\" (UID: \"f6dd56d8-2f6d-4945-9a19-4f3743409c0e\") " pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:13.870618 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.870554 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6dd56d8-2f6d-4945-9a19-4f3743409c0e-cert\") pod \"kserve-controller-manager-84685cd884-qk94h\" (UID: \"f6dd56d8-2f6d-4945-9a19-4f3743409c0e\") " pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:13.971828 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.971793 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lmz8b\" (UniqueName: \"kubernetes.io/projected/f6dd56d8-2f6d-4945-9a19-4f3743409c0e-kube-api-access-lmz8b\") pod \"kserve-controller-manager-84685cd884-qk94h\" (UID: \"f6dd56d8-2f6d-4945-9a19-4f3743409c0e\") " pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:13.971971 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.971877 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6dd56d8-2f6d-4945-9a19-4f3743409c0e-cert\") pod \"kserve-controller-manager-84685cd884-qk94h\" (UID: \"f6dd56d8-2f6d-4945-9a19-4f3743409c0e\") " pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:13.974405 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.974372 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6dd56d8-2f6d-4945-9a19-4f3743409c0e-cert\") pod \"kserve-controller-manager-84685cd884-qk94h\" (UID: \"f6dd56d8-2f6d-4945-9a19-4f3743409c0e\") " pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:13.981020 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:13.980986 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmz8b\" (UniqueName: \"kubernetes.io/projected/f6dd56d8-2f6d-4945-9a19-4f3743409c0e-kube-api-access-lmz8b\") pod \"kserve-controller-manager-84685cd884-qk94h\" (UID: \"f6dd56d8-2f6d-4945-9a19-4f3743409c0e\") " pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:14.001284 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.001250 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:10:14.072751 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.072697 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44eff084-d7a3-48cf-a3d8-270b030d90f7-cert\") pod \"44eff084-d7a3-48cf-a3d8-270b030d90f7\" (UID: \"44eff084-d7a3-48cf-a3d8-270b030d90f7\") " Apr 21 10:10:14.072751 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.072730 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lphs5\" (UniqueName: \"kubernetes.io/projected/44eff084-d7a3-48cf-a3d8-270b030d90f7-kube-api-access-lphs5\") pod \"44eff084-d7a3-48cf-a3d8-270b030d90f7\" (UID: \"44eff084-d7a3-48cf-a3d8-270b030d90f7\") " Apr 21 10:10:14.074941 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.074914 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44eff084-d7a3-48cf-a3d8-270b030d90f7-cert" (OuterVolumeSpecName: "cert") pod "44eff084-d7a3-48cf-a3d8-270b030d90f7" (UID: "44eff084-d7a3-48cf-a3d8-270b030d90f7"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:10:14.074941 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.074922 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44eff084-d7a3-48cf-a3d8-270b030d90f7-kube-api-access-lphs5" (OuterVolumeSpecName: "kube-api-access-lphs5") pod "44eff084-d7a3-48cf-a3d8-270b030d90f7" (UID: "44eff084-d7a3-48cf-a3d8-270b030d90f7"). InnerVolumeSpecName "kube-api-access-lphs5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:10:14.154159 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.154134 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:14.173959 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.173939 2570 reconciler_common.go:299] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44eff084-d7a3-48cf-a3d8-270b030d90f7-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:10:14.173959 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.173960 2570 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lphs5\" (UniqueName: \"kubernetes.io/projected/44eff084-d7a3-48cf-a3d8-270b030d90f7-kube-api-access-lphs5\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:10:14.272553 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.272521 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/kserve-controller-manager-84685cd884-qk94h"] Apr 21 10:10:14.275707 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:10:14.275664 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6dd56d8_2f6d_4945_9a19_4f3743409c0e.slice/crio-eb00320136e72b640070023718fbef7ca98e37002bda9a2ea3cd9491d0ef48c4 WatchSource:0}: Error finding container eb00320136e72b640070023718fbef7ca98e37002bda9a2ea3cd9491d0ef48c4: Status 404 returned error can't find the container with id eb00320136e72b640070023718fbef7ca98e37002bda9a2ea3cd9491d0ef48c4 Apr 21 10:10:14.687036 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.687007 2570 generic.go:358] "Generic (PLEG): container finished" podID="44eff084-d7a3-48cf-a3d8-270b030d90f7" containerID="1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d" exitCode=0 Apr 21 10:10:14.687117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.687067 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" Apr 21 10:10:14.687117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.687081 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" event={"ID":"44eff084-d7a3-48cf-a3d8-270b030d90f7","Type":"ContainerDied","Data":"1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d"} Apr 21 10:10:14.687117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.687113 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/kserve-controller-manager-84685cd884-bsvvv" event={"ID":"44eff084-d7a3-48cf-a3d8-270b030d90f7","Type":"ContainerDied","Data":"dcbe822383bbc9df241b84bf6008e43642199e5a375215b26502c931e72405ef"} Apr 21 10:10:14.687301 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.687134 2570 scope.go:117] "RemoveContainer" containerID="1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d" Apr 21 10:10:14.688497 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.688465 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/kserve-controller-manager-84685cd884-qk94h" event={"ID":"f6dd56d8-2f6d-4945-9a19-4f3743409c0e","Type":"ContainerStarted","Data":"eb00320136e72b640070023718fbef7ca98e37002bda9a2ea3cd9491d0ef48c4"} Apr 21 10:10:14.694210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.694195 2570 scope.go:117] "RemoveContainer" containerID="1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d" Apr 21 10:10:14.694516 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:10:14.694498 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d\": container with ID starting with 1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d not found: ID does not exist" containerID="1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d" Apr 21 10:10:14.694592 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.694529 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d"} err="failed to get container status \"1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d\": rpc error: code = NotFound desc = could not find container \"1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d\": container with ID starting with 1e61fb4ac886cf1644e20f2ae8e84efe7d1b883adf9af96e644774302a57822d not found: ID does not exist" Apr 21 10:10:14.702573 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.702550 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve/kserve-controller-manager-84685cd884-bsvvv"] Apr 21 10:10:14.706083 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:14.706060 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve/kserve-controller-manager-84685cd884-bsvvv"] Apr 21 10:10:15.693851 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:15.693815 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/kserve-controller-manager-84685cd884-qk94h" event={"ID":"f6dd56d8-2f6d-4945-9a19-4f3743409c0e","Type":"ContainerStarted","Data":"7302fb46df709196aa52f44e703fd442f2e0af833bd387612b180c87dba879f9"} Apr 21 10:10:15.694244 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:15.693936 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:15.710224 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:15.710175 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve/kserve-controller-manager-84685cd884-qk94h" podStartSLOduration=2.347487675 podStartE2EDuration="2.71016138s" podCreationTimestamp="2026-04-21 10:10:13 +0000 UTC" firstStartedPulling="2026-04-21 10:10:14.276927919 +0000 UTC m=+380.358143733" lastFinishedPulling="2026-04-21 10:10:14.639601614 +0000 UTC m=+380.720817438" observedRunningTime="2026-04-21 10:10:15.708145401 +0000 UTC m=+381.789361234" watchObservedRunningTime="2026-04-21 10:10:15.71016138 +0000 UTC m=+381.791377212" Apr 21 10:10:16.462284 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:16.462239 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44eff084-d7a3-48cf-a3d8-270b030d90f7" path="/var/lib/kubelet/pods/44eff084-d7a3-48cf-a3d8-270b030d90f7/volumes" Apr 21 10:10:35.730125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.730082 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-5c6b449c9b-fp756"] Apr 21 10:10:35.730500 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.730396 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="44eff084-d7a3-48cf-a3d8-270b030d90f7" containerName="manager" Apr 21 10:10:35.730500 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.730408 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="44eff084-d7a3-48cf-a3d8-270b030d90f7" containerName="manager" Apr 21 10:10:35.730500 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.730474 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="44eff084-d7a3-48cf-a3d8-270b030d90f7" containerName="manager" Apr 21 10:10:35.734608 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.734586 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.744063 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.744040 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c6b449c9b-fp756"] Apr 21 10:10:35.820229 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.820206 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-oauth-config\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.820354 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.820235 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm29p\" (UniqueName: \"kubernetes.io/projected/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-kube-api-access-sm29p\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.820354 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.820255 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-serving-cert\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.820429 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.820374 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-service-ca\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.820462 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.820432 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-oauth-serving-cert\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.820495 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.820464 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-trusted-ca-bundle\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.820530 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.820507 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-config\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.921475 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.921449 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-service-ca\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.921581 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.921484 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-oauth-serving-cert\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.921581 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.921507 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-trusted-ca-bundle\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.921581 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.921533 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-config\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.921581 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.921559 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-oauth-config\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.921581 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.921577 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sm29p\" (UniqueName: \"kubernetes.io/projected/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-kube-api-access-sm29p\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.921815 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.921699 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-serving-cert\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.922385 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.922318 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-service-ca\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.922385 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.922338 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-oauth-serving-cert\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.922385 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.922355 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-config\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.922730 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.922681 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-trusted-ca-bundle\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.924217 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.924189 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-oauth-config\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.924358 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.924340 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-console-serving-cert\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:35.929629 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:35.929608 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm29p\" (UniqueName: \"kubernetes.io/projected/a6f54ef2-5197-4211-bbcc-a5f2ebee0897-kube-api-access-sm29p\") pod \"console-5c6b449c9b-fp756\" (UID: \"a6f54ef2-5197-4211-bbcc-a5f2ebee0897\") " pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:36.044824 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:36.044760 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:36.163762 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:36.163739 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c6b449c9b-fp756"] Apr 21 10:10:36.166344 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:10:36.166318 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6f54ef2_5197_4211_bbcc_a5f2ebee0897.slice/crio-9c2d3a131b45bb747113cfe52554804ce991f6810a41abb38ae68b9518bde6d9 WatchSource:0}: Error finding container 9c2d3a131b45bb747113cfe52554804ce991f6810a41abb38ae68b9518bde6d9: Status 404 returned error can't find the container with id 9c2d3a131b45bb747113cfe52554804ce991f6810a41abb38ae68b9518bde6d9 Apr 21 10:10:36.755909 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:36.755873 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c6b449c9b-fp756" event={"ID":"a6f54ef2-5197-4211-bbcc-a5f2ebee0897","Type":"ContainerStarted","Data":"00f60a6c285614a69a1f92d08518861eebd929160b758c490a9fa26795e509de"} Apr 21 10:10:36.755909 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:36.755912 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c6b449c9b-fp756" event={"ID":"a6f54ef2-5197-4211-bbcc-a5f2ebee0897","Type":"ContainerStarted","Data":"9c2d3a131b45bb747113cfe52554804ce991f6810a41abb38ae68b9518bde6d9"} Apr 21 10:10:36.775329 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:36.775261 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5c6b449c9b-fp756" podStartSLOduration=1.775249526 podStartE2EDuration="1.775249526s" podCreationTimestamp="2026-04-21 10:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:10:36.773485112 +0000 UTC m=+402.854700940" watchObservedRunningTime="2026-04-21 10:10:36.775249526 +0000 UTC m=+402.856465358" Apr 21 10:10:46.044965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:46.044921 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:46.044965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:46.044970 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:46.049918 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:46.049898 2570 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:46.701556 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:46.701520 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve/kserve-controller-manager-84685cd884-qk94h" Apr 21 10:10:46.789916 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:46.789890 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5c6b449c9b-fp756" Apr 21 10:10:46.837793 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:10:46.837765 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-6d8f8bb896-9tddf"] Apr 21 10:11:11.861669 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:11.861627 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-6d8f8bb896-9tddf" podUID="88dc251f-76bf-48c5-8c41-a28c5927407f" containerName="console" containerID="cri-o://e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9" gracePeriod=15 Apr 21 10:11:12.099084 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.099063 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6d8f8bb896-9tddf_88dc251f-76bf-48c5-8c41-a28c5927407f/console/0.log" Apr 21 10:11:12.099195 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.099123 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:11:12.178520 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.178455 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp6fb\" (UniqueName: \"kubernetes.io/projected/88dc251f-76bf-48c5-8c41-a28c5927407f-kube-api-access-rp6fb\") pod \"88dc251f-76bf-48c5-8c41-a28c5927407f\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " Apr 21 10:11:12.178520 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.178486 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-oauth-serving-cert\") pod \"88dc251f-76bf-48c5-8c41-a28c5927407f\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " Apr 21 10:11:12.178520 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.178505 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-oauth-config\") pod \"88dc251f-76bf-48c5-8c41-a28c5927407f\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " Apr 21 10:11:12.178715 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.178530 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-console-config\") pod \"88dc251f-76bf-48c5-8c41-a28c5927407f\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " Apr 21 10:11:12.178715 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.178569 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-trusted-ca-bundle\") pod \"88dc251f-76bf-48c5-8c41-a28c5927407f\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " Apr 21 10:11:12.178715 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.178609 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-service-ca\") pod \"88dc251f-76bf-48c5-8c41-a28c5927407f\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " Apr 21 10:11:12.178715 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.178680 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-serving-cert\") pod \"88dc251f-76bf-48c5-8c41-a28c5927407f\" (UID: \"88dc251f-76bf-48c5-8c41-a28c5927407f\") " Apr 21 10:11:12.178966 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.178941 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "88dc251f-76bf-48c5-8c41-a28c5927407f" (UID: "88dc251f-76bf-48c5-8c41-a28c5927407f"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:11:12.179050 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.179020 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-service-ca" (OuterVolumeSpecName: "service-ca") pod "88dc251f-76bf-48c5-8c41-a28c5927407f" (UID: "88dc251f-76bf-48c5-8c41-a28c5927407f"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:11:12.179102 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.179035 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "88dc251f-76bf-48c5-8c41-a28c5927407f" (UID: "88dc251f-76bf-48c5-8c41-a28c5927407f"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:11:12.179102 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.179040 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-console-config" (OuterVolumeSpecName: "console-config") pod "88dc251f-76bf-48c5-8c41-a28c5927407f" (UID: "88dc251f-76bf-48c5-8c41-a28c5927407f"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:11:12.180998 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.180979 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "88dc251f-76bf-48c5-8c41-a28c5927407f" (UID: "88dc251f-76bf-48c5-8c41-a28c5927407f"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:11:12.181243 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.181215 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "88dc251f-76bf-48c5-8c41-a28c5927407f" (UID: "88dc251f-76bf-48c5-8c41-a28c5927407f"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:11:12.181324 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.181260 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88dc251f-76bf-48c5-8c41-a28c5927407f-kube-api-access-rp6fb" (OuterVolumeSpecName: "kube-api-access-rp6fb") pod "88dc251f-76bf-48c5-8c41-a28c5927407f" (UID: "88dc251f-76bf-48c5-8c41-a28c5927407f"). InnerVolumeSpecName "kube-api-access-rp6fb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:11:12.279324 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.279294 2570 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:11:12.279324 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.279319 2570 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rp6fb\" (UniqueName: \"kubernetes.io/projected/88dc251f-76bf-48c5-8c41-a28c5927407f-kube-api-access-rp6fb\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:11:12.279324 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.279329 2570 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-oauth-serving-cert\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:11:12.279521 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.279338 2570 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88dc251f-76bf-48c5-8c41-a28c5927407f-console-oauth-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:11:12.279521 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.279349 2570 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-console-config\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:11:12.279521 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.279358 2570 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-trusted-ca-bundle\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:11:12.279521 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.279366 2570 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88dc251f-76bf-48c5-8c41-a28c5927407f-service-ca\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:11:12.866720 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.866695 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-6d8f8bb896-9tddf_88dc251f-76bf-48c5-8c41-a28c5927407f/console/0.log" Apr 21 10:11:12.867088 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.866739 2570 generic.go:358] "Generic (PLEG): container finished" podID="88dc251f-76bf-48c5-8c41-a28c5927407f" containerID="e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9" exitCode=2 Apr 21 10:11:12.867088 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.866803 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d8f8bb896-9tddf" Apr 21 10:11:12.867088 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.866826 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d8f8bb896-9tddf" event={"ID":"88dc251f-76bf-48c5-8c41-a28c5927407f","Type":"ContainerDied","Data":"e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9"} Apr 21 10:11:12.867088 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.866868 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d8f8bb896-9tddf" event={"ID":"88dc251f-76bf-48c5-8c41-a28c5927407f","Type":"ContainerDied","Data":"7df8953c3da616233df2e8e26aef2be5a7e9ee177ea9f321795ab4826edff17c"} Apr 21 10:11:12.867088 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.866884 2570 scope.go:117] "RemoveContainer" containerID="e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9" Apr 21 10:11:12.874827 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.874811 2570 scope.go:117] "RemoveContainer" containerID="e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9" Apr 21 10:11:12.875083 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:11:12.875065 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9\": container with ID starting with e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9 not found: ID does not exist" containerID="e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9" Apr 21 10:11:12.875155 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.875090 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9"} err="failed to get container status \"e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9\": rpc error: code = NotFound desc = could not find container \"e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9\": container with ID starting with e1c5bfc3f97aa8752918117792039669ef1aee3710f32c11a0ba21f792bbb7c9 not found: ID does not exist" Apr 21 10:11:12.884730 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.884709 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-6d8f8bb896-9tddf"] Apr 21 10:11:12.888448 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:12.888429 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-6d8f8bb896-9tddf"] Apr 21 10:11:14.462628 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:14.462594 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88dc251f-76bf-48c5-8c41-a28c5927407f" path="/var/lib/kubelet/pods/88dc251f-76bf-48c5-8c41-a28c5927407f/volumes" Apr 21 10:11:24.437358 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.437328 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf"] Apr 21 10:11:24.437805 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.437654 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="88dc251f-76bf-48c5-8c41-a28c5927407f" containerName="console" Apr 21 10:11:24.437805 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.437665 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="88dc251f-76bf-48c5-8c41-a28c5927407f" containerName="console" Apr 21 10:11:24.437805 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.437712 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="88dc251f-76bf-48c5-8c41-a28c5927407f" containerName="console" Apr 21 10:11:24.440843 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.440821 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" Apr 21 10:11:24.443362 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.443344 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve-ci-e2e-test\"/\"default-dockercfg-5m4mw\"" Apr 21 10:11:24.448333 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.448312 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf"] Apr 21 10:11:24.450925 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.450908 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" Apr 21 10:11:24.573778 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.573754 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf"] Apr 21 10:11:24.575124 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:11:24.575099 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6eb46ba0_c96c_4ed7_97e7_2167ef4bad46.slice/crio-5d333d64a26b188aab5d517736384adf40fca5908c024271c5aaf019a00e6a33 WatchSource:0}: Error finding container 5d333d64a26b188aab5d517736384adf40fca5908c024271c5aaf019a00e6a33: Status 404 returned error can't find the container with id 5d333d64a26b188aab5d517736384adf40fca5908c024271c5aaf019a00e6a33 Apr 21 10:11:24.664029 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.663997 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5"] Apr 21 10:11:24.668492 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.668477 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:11:24.678289 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.678249 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5"] Apr 21 10:11:24.750126 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.750051 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4"] Apr 21 10:11:24.753264 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.753246 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" Apr 21 10:11:24.762344 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.762323 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4"] Apr 21 10:11:24.763929 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.763911 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" Apr 21 10:11:24.764027 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.763912 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/e7ba429b-b424-44b9-b5d6-0bdb6b58d59e-kserve-provision-location\") pod \"isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5\" (UID: \"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e\") " pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:11:24.865365 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.865328 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/e7ba429b-b424-44b9-b5d6-0bdb6b58d59e-kserve-provision-location\") pod \"isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5\" (UID: \"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e\") " pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:11:24.865848 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.865754 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/e7ba429b-b424-44b9-b5d6-0bdb6b58d59e-kserve-provision-location\") pod \"isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5\" (UID: \"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e\") " pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:11:24.889991 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.889970 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4"] Apr 21 10:11:24.892539 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:11:24.892510 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda79d6900_c4d7_4a38_ab14_f96d815d0752.slice/crio-7d63e7ab32a5e2f0cb2f238dc64195c90dc6404221402e7eb8ccb19114dd3721 WatchSource:0}: Error finding container 7d63e7ab32a5e2f0cb2f238dc64195c90dc6404221402e7eb8ccb19114dd3721: Status 404 returned error can't find the container with id 7d63e7ab32a5e2f0cb2f238dc64195c90dc6404221402e7eb8ccb19114dd3721 Apr 21 10:11:24.908190 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.908164 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" event={"ID":"6eb46ba0-c96c-4ed7-97e7-2167ef4bad46","Type":"ContainerStarted","Data":"5d333d64a26b188aab5d517736384adf40fca5908c024271c5aaf019a00e6a33"} Apr 21 10:11:24.909300 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.909261 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" event={"ID":"a79d6900-c4d7-4a38-ab14-f96d815d0752","Type":"ContainerStarted","Data":"7d63e7ab32a5e2f0cb2f238dc64195c90dc6404221402e7eb8ccb19114dd3721"} Apr 21 10:11:24.979672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:24.979651 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:11:25.124199 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:25.123997 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5"] Apr 21 10:11:25.127225 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:11:25.127191 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7ba429b_b424_44b9_b5d6_0bdb6b58d59e.slice/crio-f759ad882045baa0bb7988020f284572af8fbb25d15bf839374467e56b349124 WatchSource:0}: Error finding container f759ad882045baa0bb7988020f284572af8fbb25d15bf839374467e56b349124: Status 404 returned error can't find the container with id f759ad882045baa0bb7988020f284572af8fbb25d15bf839374467e56b349124 Apr 21 10:11:25.921084 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:25.921025 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" event={"ID":"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e","Type":"ContainerStarted","Data":"f759ad882045baa0bb7988020f284572af8fbb25d15bf839374467e56b349124"} Apr 21 10:11:42.006249 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.006211 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" event={"ID":"a79d6900-c4d7-4a38-ab14-f96d815d0752","Type":"ContainerStarted","Data":"c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2"} Apr 21 10:11:42.006776 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.006488 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" Apr 21 10:11:42.007765 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.007735 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" event={"ID":"6eb46ba0-c96c-4ed7-97e7-2167ef4bad46","Type":"ContainerStarted","Data":"82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da"} Apr 21 10:11:42.007941 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.007924 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" Apr 21 10:11:42.008031 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.007989 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.29:8080: connect: connection refused" Apr 21 10:11:42.008768 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.008745 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.27:8080: connect: connection refused" Apr 21 10:11:42.009057 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.009035 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" event={"ID":"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e","Type":"ContainerStarted","Data":"b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b"} Apr 21 10:11:42.021921 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.021880 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" podStartSLOduration=1.5460617079999999 podStartE2EDuration="18.02186969s" podCreationTimestamp="2026-04-21 10:11:24 +0000 UTC" firstStartedPulling="2026-04-21 10:11:24.894320949 +0000 UTC m=+450.975536760" lastFinishedPulling="2026-04-21 10:11:41.370128932 +0000 UTC m=+467.451344742" observedRunningTime="2026-04-21 10:11:42.020075679 +0000 UTC m=+468.101291511" watchObservedRunningTime="2026-04-21 10:11:42.02186969 +0000 UTC m=+468.103085520" Apr 21 10:11:42.047304 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:42.047244 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" podStartSLOduration=1.254988147 podStartE2EDuration="18.047228506s" podCreationTimestamp="2026-04-21 10:11:24 +0000 UTC" firstStartedPulling="2026-04-21 10:11:24.576919724 +0000 UTC m=+450.658135549" lastFinishedPulling="2026-04-21 10:11:41.369160085 +0000 UTC m=+467.450375908" observedRunningTime="2026-04-21 10:11:42.046645715 +0000 UTC m=+468.127861563" watchObservedRunningTime="2026-04-21 10:11:42.047228506 +0000 UTC m=+468.128444367" Apr 21 10:11:43.012468 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:43.012427 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.27:8080: connect: connection refused" Apr 21 10:11:43.012842 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:43.012435 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.29:8080: connect: connection refused" Apr 21 10:11:45.020362 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:45.020329 2570 generic.go:358] "Generic (PLEG): container finished" podID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerID="b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b" exitCode=0 Apr 21 10:11:45.020667 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:45.020387 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" event={"ID":"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e","Type":"ContainerDied","Data":"b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b"} Apr 21 10:11:51.043549 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:51.043478 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" event={"ID":"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e","Type":"ContainerStarted","Data":"c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0"} Apr 21 10:11:51.043858 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:51.043804 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:11:51.045039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:51.044999 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.28:8080: connect: connection refused" Apr 21 10:11:51.061444 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:51.061405 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podStartSLOduration=1.441375625 podStartE2EDuration="27.061394125s" podCreationTimestamp="2026-04-21 10:11:24 +0000 UTC" firstStartedPulling="2026-04-21 10:11:25.129712505 +0000 UTC m=+451.210928320" lastFinishedPulling="2026-04-21 10:11:50.749731009 +0000 UTC m=+476.830946820" observedRunningTime="2026-04-21 10:11:51.060162258 +0000 UTC m=+477.141378090" watchObservedRunningTime="2026-04-21 10:11:51.061394125 +0000 UTC m=+477.142609956" Apr 21 10:11:52.047582 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:52.047544 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.28:8080: connect: connection refused" Apr 21 10:11:53.012863 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:53.012819 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.29:8080: connect: connection refused" Apr 21 10:11:53.013023 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:11:53.012828 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.27:8080: connect: connection refused" Apr 21 10:12:02.048146 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:02.048099 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.28:8080: connect: connection refused" Apr 21 10:12:03.012559 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:03.012519 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.29:8080: connect: connection refused" Apr 21 10:12:03.012734 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:03.012523 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.27:8080: connect: connection refused" Apr 21 10:12:12.047908 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:12.047865 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.28:8080: connect: connection refused" Apr 21 10:12:13.012684 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:13.012649 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.29:8080: connect: connection refused" Apr 21 10:12:13.012847 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:13.012646 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.27:8080: connect: connection refused" Apr 21 10:12:22.048282 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:22.048217 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.28:8080: connect: connection refused" Apr 21 10:12:23.012878 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:23.012835 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.29:8080: connect: connection refused" Apr 21 10:12:23.013057 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:23.012835 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.27:8080: connect: connection refused" Apr 21 10:12:32.048251 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:32.048209 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.28:8080: connect: connection refused" Apr 21 10:12:33.013452 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:33.013412 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" Apr 21 10:12:33.013642 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:33.013468 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" Apr 21 10:12:42.048530 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:42.048482 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.28:8080: connect: connection refused" Apr 21 10:12:52.048441 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:52.048409 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:12:58.842612 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.842568 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf"] Apr 21 10:12:58.843135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.842805 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" containerID="cri-o://82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da" gracePeriod=30 Apr 21 10:12:58.890519 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.890489 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4"] Apr 21 10:12:58.893720 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.893701 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" Apr 21 10:12:58.897323 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.897303 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4"] Apr 21 10:12:58.897508 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.897486 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" containerID="cri-o://c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2" gracePeriod=30 Apr 21 10:12:58.904031 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.904011 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4"] Apr 21 10:12:58.905997 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.905806 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" Apr 21 10:12:58.956397 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.955952 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm"] Apr 21 10:12:58.959645 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.959620 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" Apr 21 10:12:58.966394 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.966027 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm"] Apr 21 10:12:58.977569 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:58.976086 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" Apr 21 10:12:59.041358 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.041329 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4"] Apr 21 10:12:59.044819 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:12:59.044457 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9e68348_4996_4f6f_9459_c93d4897bef4.slice/crio-5617bdbd9b142ba2bb2c63bdd736082970bd52637d6668f7ef119b0748baaee6 WatchSource:0}: Error finding container 5617bdbd9b142ba2bb2c63bdd736082970bd52637d6668f7ef119b0748baaee6: Status 404 returned error can't find the container with id 5617bdbd9b142ba2bb2c63bdd736082970bd52637d6668f7ef119b0748baaee6 Apr 21 10:12:59.111032 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.110968 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm"] Apr 21 10:12:59.114569 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:12:59.114540 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod432e0079_4268_4490_8730_9129ff8e25d7.slice/crio-8fd6f45d36ca1db92f42b957985bc0c7533f63dae1d5c06793cc5d34d0bfe9c6 WatchSource:0}: Error finding container 8fd6f45d36ca1db92f42b957985bc0c7533f63dae1d5c06793cc5d34d0bfe9c6: Status 404 returned error can't find the container with id 8fd6f45d36ca1db92f42b957985bc0c7533f63dae1d5c06793cc5d34d0bfe9c6 Apr 21 10:12:59.263248 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.263206 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" event={"ID":"432e0079-4268-4490-8730-9129ff8e25d7","Type":"ContainerStarted","Data":"3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f"} Apr 21 10:12:59.263248 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.263250 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" event={"ID":"432e0079-4268-4490-8730-9129ff8e25d7","Type":"ContainerStarted","Data":"8fd6f45d36ca1db92f42b957985bc0c7533f63dae1d5c06793cc5d34d0bfe9c6"} Apr 21 10:12:59.263515 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.263401 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" Apr 21 10:12:59.264710 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.264680 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.31:8080: connect: connection refused" Apr 21 10:12:59.264835 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.264790 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" event={"ID":"f9e68348-4996-4f6f-9459-c93d4897bef4","Type":"ContainerStarted","Data":"d45e6f9576e72d039e2bf3e6cd4c88d5fce179a07e6fc212974c3e142abd7a75"} Apr 21 10:12:59.264835 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.264821 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" event={"ID":"f9e68348-4996-4f6f-9459-c93d4897bef4","Type":"ContainerStarted","Data":"5617bdbd9b142ba2bb2c63bdd736082970bd52637d6668f7ef119b0748baaee6"} Apr 21 10:12:59.264969 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.264926 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" Apr 21 10:12:59.265917 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.265895 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.30:8080: connect: connection refused" Apr 21 10:12:59.279536 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.279494 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" podStartSLOduration=1.279481436 podStartE2EDuration="1.279481436s" podCreationTimestamp="2026-04-21 10:12:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:12:59.277776792 +0000 UTC m=+545.358992624" watchObservedRunningTime="2026-04-21 10:12:59.279481436 +0000 UTC m=+545.360697267" Apr 21 10:12:59.292294 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:12:59.292210 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" podStartSLOduration=1.292193755 podStartE2EDuration="1.292193755s" podCreationTimestamp="2026-04-21 10:12:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:12:59.291312807 +0000 UTC m=+545.372528635" watchObservedRunningTime="2026-04-21 10:12:59.292193755 +0000 UTC m=+545.373409589" Apr 21 10:13:00.268106 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:00.268061 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.31:8080: connect: connection refused" Apr 21 10:13:00.268569 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:00.268136 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.30:8080: connect: connection refused" Apr 21 10:13:02.196982 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.196959 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" Apr 21 10:13:02.218958 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.218938 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" Apr 21 10:13:02.275761 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.275682 2570 generic.go:358] "Generic (PLEG): container finished" podID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerID="82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da" exitCode=0 Apr 21 10:13:02.275761 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.275741 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" Apr 21 10:13:02.275965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.275766 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" event={"ID":"6eb46ba0-c96c-4ed7-97e7-2167ef4bad46","Type":"ContainerDied","Data":"82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da"} Apr 21 10:13:02.275965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.275800 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf" event={"ID":"6eb46ba0-c96c-4ed7-97e7-2167ef4bad46","Type":"ContainerDied","Data":"5d333d64a26b188aab5d517736384adf40fca5908c024271c5aaf019a00e6a33"} Apr 21 10:13:02.275965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.275817 2570 scope.go:117] "RemoveContainer" containerID="82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da" Apr 21 10:13:02.277023 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.277004 2570 generic.go:358] "Generic (PLEG): container finished" podID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerID="c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2" exitCode=0 Apr 21 10:13:02.277136 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.277037 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" event={"ID":"a79d6900-c4d7-4a38-ab14-f96d815d0752","Type":"ContainerDied","Data":"c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2"} Apr 21 10:13:02.277136 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.277063 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" event={"ID":"a79d6900-c4d7-4a38-ab14-f96d815d0752","Type":"ContainerDied","Data":"7d63e7ab32a5e2f0cb2f238dc64195c90dc6404221402e7eb8ccb19114dd3721"} Apr 21 10:13:02.277136 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.277076 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4" Apr 21 10:13:02.283921 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.283904 2570 scope.go:117] "RemoveContainer" containerID="82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da" Apr 21 10:13:02.284147 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:13:02.284130 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da\": container with ID starting with 82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da not found: ID does not exist" containerID="82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da" Apr 21 10:13:02.284207 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.284162 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da"} err="failed to get container status \"82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da\": rpc error: code = NotFound desc = could not find container \"82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da\": container with ID starting with 82ef8c9ab1817bade5fe5802e59c532260871c0c02f35cbb05eb4554a9e372da not found: ID does not exist" Apr 21 10:13:02.284207 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.284179 2570 scope.go:117] "RemoveContainer" containerID="c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2" Apr 21 10:13:02.292555 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.292493 2570 scope.go:117] "RemoveContainer" containerID="c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2" Apr 21 10:13:02.293919 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:13:02.293888 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2\": container with ID starting with c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2 not found: ID does not exist" containerID="c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2" Apr 21 10:13:02.293991 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.293916 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2"} err="failed to get container status \"c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2\": rpc error: code = NotFound desc = could not find container \"c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2\": container with ID starting with c7e363331cb8db405c4b64c7bd8ee7fc69772ccd6dc0e82f2c320bdd261ed0b2 not found: ID does not exist" Apr 21 10:13:02.301525 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.301503 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf"] Apr 21 10:13:02.303735 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.303711 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-ce2f4-predictor-54fdc77bf9-zfzjf"] Apr 21 10:13:02.311450 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.311431 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4"] Apr 21 10:13:02.316713 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.316695 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-ce2f4-predictor-668bdfbd7d-swrp4"] Apr 21 10:13:02.466892 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.466865 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" path="/var/lib/kubelet/pods/6eb46ba0-c96c-4ed7-97e7-2167ef4bad46/volumes" Apr 21 10:13:02.467098 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:02.467085 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" path="/var/lib/kubelet/pods/a79d6900-c4d7-4a38-ab14-f96d815d0752/volumes" Apr 21 10:13:10.268802 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:10.268761 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.30:8080: connect: connection refused" Apr 21 10:13:10.269177 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:10.268761 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.31:8080: connect: connection refused" Apr 21 10:13:20.269193 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:20.269148 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.31:8080: connect: connection refused" Apr 21 10:13:20.269653 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:20.269152 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.30:8080: connect: connection refused" Apr 21 10:13:24.871325 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.871293 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg"] Apr 21 10:13:24.871775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.871657 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" Apr 21 10:13:24.871775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.871670 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" Apr 21 10:13:24.871775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.871679 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" Apr 21 10:13:24.871775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.871684 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" Apr 21 10:13:24.871775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.871745 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="6eb46ba0-c96c-4ed7-97e7-2167ef4bad46" containerName="kserve-container" Apr 21 10:13:24.871775 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.871756 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="a79d6900-c4d7-4a38-ab14-f96d815d0752" containerName="kserve-container" Apr 21 10:13:24.874562 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.874543 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" Apr 21 10:13:24.886497 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.886474 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" Apr 21 10:13:24.888316 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.888259 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg"] Apr 21 10:13:24.898956 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.898647 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5"] Apr 21 10:13:24.899078 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.898953 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" containerID="cri-o://c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0" gracePeriod=30 Apr 21 10:13:24.964210 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.964119 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9"] Apr 21 10:13:24.972667 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.972643 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" Apr 21 10:13:24.975167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.974968 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9"] Apr 21 10:13:24.985123 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:24.985097 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" Apr 21 10:13:25.037730 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.037699 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg"] Apr 21 10:13:25.039061 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:13:25.039026 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod068d811a_b067_458c_add3_a22af4aa4df7.slice/crio-588d93c985a4d54b54da7f5ce666e00df1c39e23b9b509dcf69adc6d281b1c6f WatchSource:0}: Error finding container 588d93c985a4d54b54da7f5ce666e00df1c39e23b9b509dcf69adc6d281b1c6f: Status 404 returned error can't find the container with id 588d93c985a4d54b54da7f5ce666e00df1c39e23b9b509dcf69adc6d281b1c6f Apr 21 10:13:25.129772 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.129751 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9"] Apr 21 10:13:25.131908 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:13:25.131880 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c79d93d_f477_4d0d_b9c0_223b085c9b11.slice/crio-71a2d7cfa4e3ad7f60272a953a648a134eb65f81beb585ed14a0bfc8af6c15da WatchSource:0}: Error finding container 71a2d7cfa4e3ad7f60272a953a648a134eb65f81beb585ed14a0bfc8af6c15da: Status 404 returned error can't find the container with id 71a2d7cfa4e3ad7f60272a953a648a134eb65f81beb585ed14a0bfc8af6c15da Apr 21 10:13:25.352757 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.352718 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" event={"ID":"068d811a-b067-458c-add3-a22af4aa4df7","Type":"ContainerStarted","Data":"606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84"} Apr 21 10:13:25.352757 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.352763 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" event={"ID":"068d811a-b067-458c-add3-a22af4aa4df7","Type":"ContainerStarted","Data":"588d93c985a4d54b54da7f5ce666e00df1c39e23b9b509dcf69adc6d281b1c6f"} Apr 21 10:13:25.353036 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.352779 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" Apr 21 10:13:25.354018 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.353993 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" event={"ID":"2c79d93d-f477-4d0d-b9c0-223b085c9b11","Type":"ContainerStarted","Data":"518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825"} Apr 21 10:13:25.354124 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.354024 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" event={"ID":"2c79d93d-f477-4d0d-b9c0-223b085c9b11","Type":"ContainerStarted","Data":"71a2d7cfa4e3ad7f60272a953a648a134eb65f81beb585ed14a0bfc8af6c15da"} Apr 21 10:13:25.354231 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.354209 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" Apr 21 10:13:25.354327 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.354294 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.32:8080: connect: connection refused" Apr 21 10:13:25.354992 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.354973 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.33:8080: connect: connection refused" Apr 21 10:13:25.368776 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.368739 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" podStartSLOduration=1.3687235389999999 podStartE2EDuration="1.368723539s" podCreationTimestamp="2026-04-21 10:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:13:25.367420595 +0000 UTC m=+571.448636442" watchObservedRunningTime="2026-04-21 10:13:25.368723539 +0000 UTC m=+571.449939371" Apr 21 10:13:25.379838 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:25.379772 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" podStartSLOduration=1.379761757 podStartE2EDuration="1.379761757s" podCreationTimestamp="2026-04-21 10:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:13:25.379374655 +0000 UTC m=+571.460590479" watchObservedRunningTime="2026-04-21 10:13:25.379761757 +0000 UTC m=+571.460977583" Apr 21 10:13:26.357351 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:26.357310 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.32:8080: connect: connection refused" Apr 21 10:13:26.357712 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:26.357313 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.33:8080: connect: connection refused" Apr 21 10:13:29.250803 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.250531 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:13:29.338541 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.338466 2570 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/e7ba429b-b424-44b9-b5d6-0bdb6b58d59e-kserve-provision-location\") pod \"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e\" (UID: \"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e\") " Apr 21 10:13:29.338743 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.338720 2570 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7ba429b-b424-44b9-b5d6-0bdb6b58d59e-kserve-provision-location" (OuterVolumeSpecName: "kserve-provision-location") pod "e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" (UID: "e7ba429b-b424-44b9-b5d6-0bdb6b58d59e"). InnerVolumeSpecName "kserve-provision-location". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 10:13:29.368377 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.368348 2570 generic.go:358] "Generic (PLEG): container finished" podID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerID="c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0" exitCode=0 Apr 21 10:13:29.368538 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.368432 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" Apr 21 10:13:29.368538 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.368443 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" event={"ID":"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e","Type":"ContainerDied","Data":"c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0"} Apr 21 10:13:29.368538 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.368485 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5" event={"ID":"e7ba429b-b424-44b9-b5d6-0bdb6b58d59e","Type":"ContainerDied","Data":"f759ad882045baa0bb7988020f284572af8fbb25d15bf839374467e56b349124"} Apr 21 10:13:29.368538 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.368501 2570 scope.go:117] "RemoveContainer" containerID="c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0" Apr 21 10:13:29.379882 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.378071 2570 scope.go:117] "RemoveContainer" containerID="b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b" Apr 21 10:13:29.386069 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.386053 2570 scope.go:117] "RemoveContainer" containerID="c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0" Apr 21 10:13:29.386341 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:13:29.386323 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0\": container with ID starting with c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0 not found: ID does not exist" containerID="c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0" Apr 21 10:13:29.386422 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.386352 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0"} err="failed to get container status \"c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0\": rpc error: code = NotFound desc = could not find container \"c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0\": container with ID starting with c2058f0cec1631c4ccdcc42f9e0d9fb9296bf78aec5796e88a875eea0b8dcfe0 not found: ID does not exist" Apr 21 10:13:29.386422 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.386374 2570 scope.go:117] "RemoveContainer" containerID="b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b" Apr 21 10:13:29.386617 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:13:29.386600 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b\": container with ID starting with b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b not found: ID does not exist" containerID="b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b" Apr 21 10:13:29.386656 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.386622 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b"} err="failed to get container status \"b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b\": rpc error: code = NotFound desc = could not find container \"b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b\": container with ID starting with b096b721e8f99133ab11b364b9e9efa00b733e8b826a00f443d3b314304d194b not found: ID does not exist" Apr 21 10:13:29.389524 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.389503 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5"] Apr 21 10:13:29.394827 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.394806 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/isvc-sklearn-graph-1-predictor-5dfd9559ff-wznm5"] Apr 21 10:13:29.439235 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:29.439212 2570 reconciler_common.go:299] "Volume detached for volume \"kserve-provision-location\" (UniqueName: \"kubernetes.io/empty-dir/e7ba429b-b424-44b9-b5d6-0bdb6b58d59e-kserve-provision-location\") on node \"ip-10-0-142-243.ec2.internal\" DevicePath \"\"" Apr 21 10:13:30.268820 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:30.268776 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.31:8080: connect: connection refused" Apr 21 10:13:30.269197 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:30.268776 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.30:8080: connect: connection refused" Apr 21 10:13:30.462814 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:30.462779 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" path="/var/lib/kubelet/pods/e7ba429b-b424-44b9-b5d6-0bdb6b58d59e/volumes" Apr 21 10:13:36.357345 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:36.357299 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.33:8080: connect: connection refused" Apr 21 10:13:36.357709 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:36.357463 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.32:8080: connect: connection refused" Apr 21 10:13:40.269179 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:40.269138 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.31:8080: connect: connection refused" Apr 21 10:13:40.269574 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:40.269138 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.30:8080: connect: connection refused" Apr 21 10:13:46.357649 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:46.357607 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.33:8080: connect: connection refused" Apr 21 10:13:46.358031 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:46.357608 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.32:8080: connect: connection refused" Apr 21 10:13:50.268892 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:50.268857 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" Apr 21 10:13:50.269328 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:50.269311 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" Apr 21 10:13:56.357262 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:56.357218 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.33:8080: connect: connection refused" Apr 21 10:13:56.357669 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:13:56.357395 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.32:8080: connect: connection refused" Apr 21 10:14:06.358100 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:14:06.358060 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.33:8080: connect: connection refused" Apr 21 10:14:06.360386 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:14:06.358060 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.32:8080: connect: connection refused" Apr 21 10:14:16.358504 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:14:16.358424 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" Apr 21 10:14:16.358946 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:14:16.358788 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" Apr 21 10:22:23.763166 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.763131 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4"] Apr 21 10:22:23.765585 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.763455 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" containerID="cri-o://d45e6f9576e72d039e2bf3e6cd4c88d5fce179a07e6fc212974c3e142abd7a75" gracePeriod=30 Apr 21 10:22:23.811987 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.811946 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm"] Apr 21 10:22:23.812326 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.812255 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" containerID="cri-o://3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f" gracePeriod=30 Apr 21 10:22:23.838062 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.838038 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7"] Apr 21 10:22:23.838420 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.838401 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="storage-initializer" Apr 21 10:22:23.838420 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.838417 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="storage-initializer" Apr 21 10:22:23.838555 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.838436 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" Apr 21 10:22:23.838555 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.838442 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" Apr 21 10:22:23.838555 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.838489 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="e7ba429b-b424-44b9-b5d6-0bdb6b58d59e" containerName="kserve-container" Apr 21 10:22:23.841190 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.841176 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" Apr 21 10:22:23.851776 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.851759 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" Apr 21 10:22:23.858022 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.857997 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7"] Apr 21 10:22:23.899829 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.899797 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6"] Apr 21 10:22:23.903445 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.903416 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" Apr 21 10:22:23.913626 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.913599 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6"] Apr 21 10:22:23.919284 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.918873 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" Apr 21 10:22:23.994837 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:23.994808 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7"] Apr 21 10:22:23.998404 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:22:23.998370 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa63b43b_d3fd_4623_b5a0_f7257873bb51.slice/crio-487ddbb0309de178c167373831954bb0757fa8c1c3fcf730bc105a3ab60eb472 WatchSource:0}: Error finding container 487ddbb0309de178c167373831954bb0757fa8c1c3fcf730bc105a3ab60eb472: Status 404 returned error can't find the container with id 487ddbb0309de178c167373831954bb0757fa8c1c3fcf730bc105a3ab60eb472 Apr 21 10:22:24.000369 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:24.000349 2570 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 10:22:24.054980 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:24.054930 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6"] Apr 21 10:22:24.058739 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:22:24.058715 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c9b59d1_7716_4b56_b3cb_28602899877e.slice/crio-ca54aec974e80a4b0428fb3fab7a9682c1a453d14a7279d2d915a72109f1be21 WatchSource:0}: Error finding container ca54aec974e80a4b0428fb3fab7a9682c1a453d14a7279d2d915a72109f1be21: Status 404 returned error can't find the container with id ca54aec974e80a4b0428fb3fab7a9682c1a453d14a7279d2d915a72109f1be21 Apr 21 10:22:24.097011 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:24.096980 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" event={"ID":"9c9b59d1-7716-4b56-b3cb-28602899877e","Type":"ContainerStarted","Data":"ca54aec974e80a4b0428fb3fab7a9682c1a453d14a7279d2d915a72109f1be21"} Apr 21 10:22:24.098083 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:24.098057 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" event={"ID":"aa63b43b-d3fd-4623-b5a0-f7257873bb51","Type":"ContainerStarted","Data":"487ddbb0309de178c167373831954bb0757fa8c1c3fcf730bc105a3ab60eb472"} Apr 21 10:22:25.104372 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:25.104318 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" event={"ID":"aa63b43b-d3fd-4623-b5a0-f7257873bb51","Type":"ContainerStarted","Data":"36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90"} Apr 21 10:22:25.104892 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:25.104594 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" Apr 21 10:22:25.106115 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:25.106079 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.34:8080: connect: connection refused" Apr 21 10:22:25.106497 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:25.106473 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" event={"ID":"9c9b59d1-7716-4b56-b3cb-28602899877e","Type":"ContainerStarted","Data":"9dec8829576b2a73b2ee80b20c23803d25805e97094e3dbf966db77fadc014e9"} Apr 21 10:22:25.106724 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:25.106706 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" Apr 21 10:22:25.107961 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:25.107937 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.35:8080: connect: connection refused" Apr 21 10:22:25.119224 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:25.119182 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podStartSLOduration=2.119169961 podStartE2EDuration="2.119169961s" podCreationTimestamp="2026-04-21 10:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:22:25.118025378 +0000 UTC m=+1111.199241234" watchObservedRunningTime="2026-04-21 10:22:25.119169961 +0000 UTC m=+1111.200385793" Apr 21 10:22:25.131499 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:25.131461 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podStartSLOduration=2.131449981 podStartE2EDuration="2.131449981s" podCreationTimestamp="2026-04-21 10:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:22:25.129953904 +0000 UTC m=+1111.211169737" watchObservedRunningTime="2026-04-21 10:22:25.131449981 +0000 UTC m=+1111.212665813" Apr 21 10:22:26.110887 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:26.110852 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.35:8080: connect: connection refused" Apr 21 10:22:26.111302 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:26.110951 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.34:8080: connect: connection refused" Apr 21 10:22:26.763444 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:26.763416 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" Apr 21 10:22:27.114135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.114098 2570 generic.go:358] "Generic (PLEG): container finished" podID="432e0079-4268-4490-8730-9129ff8e25d7" containerID="3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f" exitCode=0 Apr 21 10:22:27.114604 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.114177 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" event={"ID":"432e0079-4268-4490-8730-9129ff8e25d7","Type":"ContainerDied","Data":"3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f"} Apr 21 10:22:27.114604 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.114185 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" Apr 21 10:22:27.114604 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.114211 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm" event={"ID":"432e0079-4268-4490-8730-9129ff8e25d7","Type":"ContainerDied","Data":"8fd6f45d36ca1db92f42b957985bc0c7533f63dae1d5c06793cc5d34d0bfe9c6"} Apr 21 10:22:27.114604 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.114226 2570 scope.go:117] "RemoveContainer" containerID="3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f" Apr 21 10:22:27.115605 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.115579 2570 generic.go:358] "Generic (PLEG): container finished" podID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerID="d45e6f9576e72d039e2bf3e6cd4c88d5fce179a07e6fc212974c3e142abd7a75" exitCode=0 Apr 21 10:22:27.115679 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.115607 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" event={"ID":"f9e68348-4996-4f6f-9459-c93d4897bef4","Type":"ContainerDied","Data":"d45e6f9576e72d039e2bf3e6cd4c88d5fce179a07e6fc212974c3e142abd7a75"} Apr 21 10:22:27.122514 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.122423 2570 scope.go:117] "RemoveContainer" containerID="3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f" Apr 21 10:22:27.122741 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:22:27.122715 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f\": container with ID starting with 3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f not found: ID does not exist" containerID="3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f" Apr 21 10:22:27.122809 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.122751 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f"} err="failed to get container status \"3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f\": rpc error: code = NotFound desc = could not find container \"3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f\": container with ID starting with 3341b8a5dec6ca3c53b1f3bc12ff8e31b0e4111bf2c94e3658262f29cd80085f not found: ID does not exist" Apr 21 10:22:27.135041 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.135019 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm"] Apr 21 10:22:27.138721 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.138701 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b4346-predictor-5f9c4767fb-4xxhm"] Apr 21 10:22:27.212708 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:27.212682 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" Apr 21 10:22:28.120481 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:28.120444 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" event={"ID":"f9e68348-4996-4f6f-9459-c93d4897bef4","Type":"ContainerDied","Data":"5617bdbd9b142ba2bb2c63bdd736082970bd52637d6668f7ef119b0748baaee6"} Apr 21 10:22:28.120481 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:28.120485 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4" Apr 21 10:22:28.121020 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:28.120489 2570 scope.go:117] "RemoveContainer" containerID="d45e6f9576e72d039e2bf3e6cd4c88d5fce179a07e6fc212974c3e142abd7a75" Apr 21 10:22:28.140158 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:28.140130 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4"] Apr 21 10:22:28.145334 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:28.145314 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b4346-predictor-6c9795b877-t7kx4"] Apr 21 10:22:28.462431 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:28.462357 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="432e0079-4268-4490-8730-9129ff8e25d7" path="/var/lib/kubelet/pods/432e0079-4268-4490-8730-9129ff8e25d7/volumes" Apr 21 10:22:28.462589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:28.462576 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" path="/var/lib/kubelet/pods/f9e68348-4996-4f6f-9459-c93d4897bef4/volumes" Apr 21 10:22:36.111020 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:36.110990 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.34:8080: connect: connection refused" Apr 21 10:22:36.111411 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:36.110988 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.35:8080: connect: connection refused" Apr 21 10:22:46.110982 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:46.110940 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.35:8080: connect: connection refused" Apr 21 10:22:46.111381 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:46.110949 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.34:8080: connect: connection refused" Apr 21 10:22:49.622135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.622102 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg"] Apr 21 10:22:49.622524 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.622344 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" containerID="cri-o://606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84" gracePeriod=30 Apr 21 10:22:49.694587 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.694549 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8"] Apr 21 10:22:49.694937 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.694920 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" Apr 21 10:22:49.695016 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.694940 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" Apr 21 10:22:49.695016 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.694972 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" Apr 21 10:22:49.695016 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.694982 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" Apr 21 10:22:49.695589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.695063 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="432e0079-4268-4490-8730-9129ff8e25d7" containerName="kserve-container" Apr 21 10:22:49.695589 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.695080 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="f9e68348-4996-4f6f-9459-c93d4897bef4" containerName="kserve-container" Apr 21 10:22:49.698051 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.698027 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" Apr 21 10:22:49.708693 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.708657 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8"] Apr 21 10:22:49.710857 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.710696 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" Apr 21 10:22:49.730237 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.730202 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9"] Apr 21 10:22:49.730521 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.730494 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" containerID="cri-o://518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825" gracePeriod=30 Apr 21 10:22:49.788232 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.788195 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4"] Apr 21 10:22:49.799802 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.799121 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4"] Apr 21 10:22:49.799802 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.799255 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" Apr 21 10:22:49.814083 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.814060 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" Apr 21 10:22:49.851007 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.850976 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8"] Apr 21 10:22:49.854510 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:22:49.854474 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod456d8fbb_19dc_4af0_804a_4847333b944a.slice/crio-ca001ab8f05fb47c696a5d3e66e100a0ededd46f7d7833a5a05bb0c8e25b266b WatchSource:0}: Error finding container ca001ab8f05fb47c696a5d3e66e100a0ededd46f7d7833a5a05bb0c8e25b266b: Status 404 returned error can't find the container with id ca001ab8f05fb47c696a5d3e66e100a0ededd46f7d7833a5a05bb0c8e25b266b Apr 21 10:22:49.949331 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:49.949236 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4"] Apr 21 10:22:49.952102 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:22:49.952076 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod575bdf2e_3de3_453a_9809_f22e191e5ce2.slice/crio-343a3e35768b7daf8a3f733e7c787d2f52b204d327707c31a4cbedadbfb8fb8a WatchSource:0}: Error finding container 343a3e35768b7daf8a3f733e7c787d2f52b204d327707c31a4cbedadbfb8fb8a: Status 404 returned error can't find the container with id 343a3e35768b7daf8a3f733e7c787d2f52b204d327707c31a4cbedadbfb8fb8a Apr 21 10:22:50.193041 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.192963 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" event={"ID":"456d8fbb-19dc-4af0-804a-4847333b944a","Type":"ContainerStarted","Data":"e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8"} Apr 21 10:22:50.193041 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.193003 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" event={"ID":"456d8fbb-19dc-4af0-804a-4847333b944a","Type":"ContainerStarted","Data":"ca001ab8f05fb47c696a5d3e66e100a0ededd46f7d7833a5a05bb0c8e25b266b"} Apr 21 10:22:50.193237 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.193141 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" Apr 21 10:22:50.194477 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.194448 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.36:8080: connect: connection refused" Apr 21 10:22:50.194477 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.194468 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" event={"ID":"575bdf2e-3de3-453a-9809-f22e191e5ce2","Type":"ContainerStarted","Data":"5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9"} Apr 21 10:22:50.194631 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.194491 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" event={"ID":"575bdf2e-3de3-453a-9809-f22e191e5ce2","Type":"ContainerStarted","Data":"343a3e35768b7daf8a3f733e7c787d2f52b204d327707c31a4cbedadbfb8fb8a"} Apr 21 10:22:50.194694 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.194679 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" Apr 21 10:22:50.195561 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.195532 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.37:8080: connect: connection refused" Apr 21 10:22:50.207418 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.206873 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podStartSLOduration=1.206857512 podStartE2EDuration="1.206857512s" podCreationTimestamp="2026-04-21 10:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:22:50.205839217 +0000 UTC m=+1136.287055051" watchObservedRunningTime="2026-04-21 10:22:50.206857512 +0000 UTC m=+1136.288073345" Apr 21 10:22:50.219803 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:50.219766 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podStartSLOduration=1.219754538 podStartE2EDuration="1.219754538s" podCreationTimestamp="2026-04-21 10:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:22:50.218353233 +0000 UTC m=+1136.299569064" watchObservedRunningTime="2026-04-21 10:22:50.219754538 +0000 UTC m=+1136.300970369" Apr 21 10:22:51.199230 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:51.198817 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.36:8080: connect: connection refused" Apr 21 10:22:51.199230 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:51.199194 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.37:8080: connect: connection refused" Apr 21 10:22:53.071909 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.070714 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" Apr 21 10:22:53.207740 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.207647 2570 generic.go:358] "Generic (PLEG): container finished" podID="068d811a-b067-458c-add3-a22af4aa4df7" containerID="606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84" exitCode=0 Apr 21 10:22:53.207740 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.207689 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" event={"ID":"068d811a-b067-458c-add3-a22af4aa4df7","Type":"ContainerDied","Data":"606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84"} Apr 21 10:22:53.207740 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.207710 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" event={"ID":"068d811a-b067-458c-add3-a22af4aa4df7","Type":"ContainerDied","Data":"588d93c985a4d54b54da7f5ce666e00df1c39e23b9b509dcf69adc6d281b1c6f"} Apr 21 10:22:53.207740 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.207726 2570 scope.go:117] "RemoveContainer" containerID="606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84" Apr 21 10:22:53.207740 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.207730 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg" Apr 21 10:22:53.216213 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.216196 2570 scope.go:117] "RemoveContainer" containerID="606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84" Apr 21 10:22:53.216518 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:22:53.216500 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84\": container with ID starting with 606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84 not found: ID does not exist" containerID="606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84" Apr 21 10:22:53.216565 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.216525 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84"} err="failed to get container status \"606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84\": rpc error: code = NotFound desc = could not find container \"606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84\": container with ID starting with 606395b72a780c3fc7d9b3b65bbab49e0020b3d52223e01b993cc30f433f8b84 not found: ID does not exist" Apr 21 10:22:53.227258 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.227234 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg"] Apr 21 10:22:53.231065 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.231044 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-cd9c7-predictor-75bcf7484c-pw9cg"] Apr 21 10:22:53.471170 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:53.471150 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" Apr 21 10:22:54.212103 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.212072 2570 generic.go:358] "Generic (PLEG): container finished" podID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerID="518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825" exitCode=0 Apr 21 10:22:54.212553 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.212134 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" Apr 21 10:22:54.212553 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.212159 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" event={"ID":"2c79d93d-f477-4d0d-b9c0-223b085c9b11","Type":"ContainerDied","Data":"518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825"} Apr 21 10:22:54.212553 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.212209 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9" event={"ID":"2c79d93d-f477-4d0d-b9c0-223b085c9b11","Type":"ContainerDied","Data":"71a2d7cfa4e3ad7f60272a953a648a134eb65f81beb585ed14a0bfc8af6c15da"} Apr 21 10:22:54.212553 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.212232 2570 scope.go:117] "RemoveContainer" containerID="518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825" Apr 21 10:22:54.220460 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.220444 2570 scope.go:117] "RemoveContainer" containerID="518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825" Apr 21 10:22:54.220690 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:22:54.220673 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825\": container with ID starting with 518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825 not found: ID does not exist" containerID="518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825" Apr 21 10:22:54.220753 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.220700 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825"} err="failed to get container status \"518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825\": rpc error: code = NotFound desc = could not find container \"518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825\": container with ID starting with 518844ee75178989fe59e5e90b90c57b78d1c037e1ea9355b882fdcda1c27825 not found: ID does not exist" Apr 21 10:22:54.231970 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.231946 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9"] Apr 21 10:22:54.235556 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.235538 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-cd9c7-predictor-fdc5479f4-bh6v9"] Apr 21 10:22:54.463106 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.463053 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="068d811a-b067-458c-add3-a22af4aa4df7" path="/var/lib/kubelet/pods/068d811a-b067-458c-add3-a22af4aa4df7/volumes" Apr 21 10:22:54.463345 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:54.463329 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" path="/var/lib/kubelet/pods/2c79d93d-f477-4d0d-b9c0-223b085c9b11/volumes" Apr 21 10:22:56.110960 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:56.110925 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.35:8080: connect: connection refused" Apr 21 10:22:56.111540 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:22:56.110925 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.34:8080: connect: connection refused" Apr 21 10:23:01.199622 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:01.199579 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.36:8080: connect: connection refused" Apr 21 10:23:01.200102 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:01.199581 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.37:8080: connect: connection refused" Apr 21 10:23:06.111262 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:06.111217 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.34:8080: connect: connection refused" Apr 21 10:23:06.111262 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:06.111217 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.35:8080: connect: connection refused" Apr 21 10:23:11.199691 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:11.199650 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.37:8080: connect: connection refused" Apr 21 10:23:11.199691 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:11.199669 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.36:8080: connect: connection refused" Apr 21 10:23:16.112144 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:16.112058 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" Apr 21 10:23:16.112571 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:16.112434 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" Apr 21 10:23:21.199263 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:21.199219 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.37:8080: connect: connection refused" Apr 21 10:23:21.199628 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:21.199259 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.36:8080: connect: connection refused" Apr 21 10:23:31.199481 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:31.199438 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.37:8080: connect: connection refused" Apr 21 10:23:31.199947 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:31.199435 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.36:8080: connect: connection refused" Apr 21 10:23:41.199448 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:41.199417 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" Apr 21 10:23:41.199841 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:41.199822 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" Apr 21 10:23:44.109714 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.109679 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7"] Apr 21 10:23:44.110080 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.109963 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" containerID="cri-o://36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90" gracePeriod=30 Apr 21 10:23:44.124676 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.124651 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8"] Apr 21 10:23:44.125008 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.124995 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" Apr 21 10:23:44.125050 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.125013 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" Apr 21 10:23:44.125050 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.125037 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" Apr 21 10:23:44.125050 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.125043 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" Apr 21 10:23:44.125140 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.125107 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="068d811a-b067-458c-add3-a22af4aa4df7" containerName="kserve-container" Apr 21 10:23:44.125140 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.125118 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="2c79d93d-f477-4d0d-b9c0-223b085c9b11" containerName="kserve-container" Apr 21 10:23:44.128055 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.128039 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" Apr 21 10:23:44.136743 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.136720 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8"] Apr 21 10:23:44.138525 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.138507 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" Apr 21 10:23:44.192024 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.191986 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6"] Apr 21 10:23:44.192506 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.192358 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" containerID="cri-o://9dec8829576b2a73b2ee80b20c23803d25805e97094e3dbf966db77fadc014e9" gracePeriod=30 Apr 21 10:23:44.210681 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.210649 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827"] Apr 21 10:23:44.215726 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.215706 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" Apr 21 10:23:44.221477 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.221446 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827"] Apr 21 10:23:44.227304 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.227260 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" Apr 21 10:23:44.283470 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.283421 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8"] Apr 21 10:23:44.285574 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:23:44.285544 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bf9dfa1_c1fe_478c_8bb3_f7f0bb782b4e.slice/crio-a58fb6026f2918d7cc7581381b28c27cc33ca8c9f3958de6a1956c608a0bca85 WatchSource:0}: Error finding container a58fb6026f2918d7cc7581381b28c27cc33ca8c9f3958de6a1956c608a0bca85: Status 404 returned error can't find the container with id a58fb6026f2918d7cc7581381b28c27cc33ca8c9f3958de6a1956c608a0bca85 Apr 21 10:23:44.365253 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.365181 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827"] Apr 21 10:23:44.368138 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:23:44.368112 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8236e00_02b1_4a39_8b3e_69a361739c6f.slice/crio-52539c0acb0368f5690a365409aa5717642e689498bfe6d1a22a06af670e54b4 WatchSource:0}: Error finding container 52539c0acb0368f5690a365409aa5717642e689498bfe6d1a22a06af670e54b4: Status 404 returned error can't find the container with id 52539c0acb0368f5690a365409aa5717642e689498bfe6d1a22a06af670e54b4 Apr 21 10:23:44.378618 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.378575 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" event={"ID":"c8236e00-02b1-4a39-8b3e-69a361739c6f","Type":"ContainerStarted","Data":"52539c0acb0368f5690a365409aa5717642e689498bfe6d1a22a06af670e54b4"} Apr 21 10:23:44.379835 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:44.379813 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" event={"ID":"1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e","Type":"ContainerStarted","Data":"a58fb6026f2918d7cc7581381b28c27cc33ca8c9f3958de6a1956c608a0bca85"} Apr 21 10:23:45.384807 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:45.384768 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" event={"ID":"c8236e00-02b1-4a39-8b3e-69a361739c6f","Type":"ContainerStarted","Data":"3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985"} Apr 21 10:23:45.385257 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:45.384949 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" Apr 21 10:23:45.386117 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:45.386090 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" event={"ID":"1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e","Type":"ContainerStarted","Data":"88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1"} Apr 21 10:23:45.386308 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:45.386293 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" Apr 21 10:23:45.386397 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:45.386359 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.39:8080: connect: connection refused" Apr 21 10:23:45.387153 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:45.387132 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.38:8080: connect: connection refused" Apr 21 10:23:45.399974 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:45.399912 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" podStartSLOduration=1.399895725 podStartE2EDuration="1.399895725s" podCreationTimestamp="2026-04-21 10:23:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:23:45.398428067 +0000 UTC m=+1191.479643900" watchObservedRunningTime="2026-04-21 10:23:45.399895725 +0000 UTC m=+1191.481111559" Apr 21 10:23:45.412844 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:45.412806 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" podStartSLOduration=1.41279575 podStartE2EDuration="1.41279575s" podCreationTimestamp="2026-04-21 10:23:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:23:45.412028658 +0000 UTC m=+1191.493244494" watchObservedRunningTime="2026-04-21 10:23:45.41279575 +0000 UTC m=+1191.494011579" Apr 21 10:23:46.111513 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:46.111475 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.34:8080: connect: connection refused" Apr 21 10:23:46.111680 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:46.111475 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.35:8080: connect: connection refused" Apr 21 10:23:46.390100 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:46.390003 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.39:8080: connect: connection refused" Apr 21 10:23:46.390100 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:46.390017 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.38:8080: connect: connection refused" Apr 21 10:23:47.156236 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.156215 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" Apr 21 10:23:47.394608 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.394568 2570 generic.go:358] "Generic (PLEG): container finished" podID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerID="36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90" exitCode=0 Apr 21 10:23:47.395039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.394632 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" Apr 21 10:23:47.395039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.394657 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" event={"ID":"aa63b43b-d3fd-4623-b5a0-f7257873bb51","Type":"ContainerDied","Data":"36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90"} Apr 21 10:23:47.395039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.394693 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7" event={"ID":"aa63b43b-d3fd-4623-b5a0-f7257873bb51","Type":"ContainerDied","Data":"487ddbb0309de178c167373831954bb0757fa8c1c3fcf730bc105a3ab60eb472"} Apr 21 10:23:47.395039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.394709 2570 scope.go:117] "RemoveContainer" containerID="36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90" Apr 21 10:23:47.396404 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.396381 2570 generic.go:358] "Generic (PLEG): container finished" podID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerID="9dec8829576b2a73b2ee80b20c23803d25805e97094e3dbf966db77fadc014e9" exitCode=0 Apr 21 10:23:47.396508 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.396429 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" event={"ID":"9c9b59d1-7716-4b56-b3cb-28602899877e","Type":"ContainerDied","Data":"9dec8829576b2a73b2ee80b20c23803d25805e97094e3dbf966db77fadc014e9"} Apr 21 10:23:47.411752 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.411727 2570 scope.go:117] "RemoveContainer" containerID="36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90" Apr 21 10:23:47.412087 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:23:47.412059 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90\": container with ID starting with 36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90 not found: ID does not exist" containerID="36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90" Apr 21 10:23:47.412177 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.412100 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90"} err="failed to get container status \"36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90\": rpc error: code = NotFound desc = could not find container \"36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90\": container with ID starting with 36345830fdf198143f620e57b58e234be09e30e51540302f82c5ef41faf05d90 not found: ID does not exist" Apr 21 10:23:47.422570 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.422415 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7"] Apr 21 10:23:47.425774 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.425752 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-42a8c-predictor-5bffc845f9-8clt7"] Apr 21 10:23:47.429905 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:47.429879 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" Apr 21 10:23:48.401566 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:48.401538 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" Apr 21 10:23:48.402059 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:48.401530 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6" event={"ID":"9c9b59d1-7716-4b56-b3cb-28602899877e","Type":"ContainerDied","Data":"ca54aec974e80a4b0428fb3fab7a9682c1a453d14a7279d2d915a72109f1be21"} Apr 21 10:23:48.402059 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:48.401666 2570 scope.go:117] "RemoveContainer" containerID="9dec8829576b2a73b2ee80b20c23803d25805e97094e3dbf966db77fadc014e9" Apr 21 10:23:48.422361 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:48.422334 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6"] Apr 21 10:23:48.426166 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:48.426142 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-42a8c-predictor-5d964846d6-lkxg6"] Apr 21 10:23:48.462153 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:48.462128 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" path="/var/lib/kubelet/pods/9c9b59d1-7716-4b56-b3cb-28602899877e/volumes" Apr 21 10:23:48.462407 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:48.462393 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" path="/var/lib/kubelet/pods/aa63b43b-d3fd-4623-b5a0-f7257873bb51/volumes" Apr 21 10:23:56.390083 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:56.390037 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.39:8080: connect: connection refused" Apr 21 10:23:56.390479 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:23:56.390037 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.38:8080: connect: connection refused" Apr 21 10:24:06.390790 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:06.390746 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.39:8080: connect: connection refused" Apr 21 10:24:06.391291 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:06.390745 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.38:8080: connect: connection refused" Apr 21 10:24:09.834437 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.834407 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4"] Apr 21 10:24:09.834877 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.834702 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" containerID="cri-o://5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9" gracePeriod=30 Apr 21 10:24:09.890422 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.890390 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8"] Apr 21 10:24:09.890662 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.890627 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" containerID="cri-o://e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8" gracePeriod=30 Apr 21 10:24:09.908891 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.908865 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq"] Apr 21 10:24:09.909205 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.909193 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" Apr 21 10:24:09.909261 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.909207 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" Apr 21 10:24:09.909261 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.909228 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" Apr 21 10:24:09.909261 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.909234 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" Apr 21 10:24:09.909394 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.909303 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="9c9b59d1-7716-4b56-b3cb-28602899877e" containerName="kserve-container" Apr 21 10:24:09.909394 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.909313 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="aa63b43b-d3fd-4623-b5a0-f7257873bb51" containerName="kserve-container" Apr 21 10:24:09.913592 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.913576 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" Apr 21 10:24:09.921199 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.921178 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq"] Apr 21 10:24:09.923132 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:09.923110 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" Apr 21 10:24:10.041226 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.041195 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6"] Apr 21 10:24:10.045996 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.045974 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" Apr 21 10:24:10.052751 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.052718 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6"] Apr 21 10:24:10.057095 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.057075 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq"] Apr 21 10:24:10.059256 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:24:10.059234 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5905aa4b_bad0_42a9_943f_01f5147fcd54.slice/crio-2a7d2a03dee1e9b372a7640cfd620c33acf10a8f137a894687adfc65b000aa18 WatchSource:0}: Error finding container 2a7d2a03dee1e9b372a7640cfd620c33acf10a8f137a894687adfc65b000aa18: Status 404 returned error can't find the container with id 2a7d2a03dee1e9b372a7640cfd620c33acf10a8f137a894687adfc65b000aa18 Apr 21 10:24:10.059476 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.059460 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" Apr 21 10:24:10.189003 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.188963 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6"] Apr 21 10:24:10.191074 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:24:10.191049 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87b361ed_8851_4e8a_82fb_20213dfbf680.slice/crio-f941df91afd2f9bcaedf8f5406ef676bcf018f1d518da6ae1c7c8e9e9e4b698a WatchSource:0}: Error finding container f941df91afd2f9bcaedf8f5406ef676bcf018f1d518da6ae1c7c8e9e9e4b698a: Status 404 returned error can't find the container with id f941df91afd2f9bcaedf8f5406ef676bcf018f1d518da6ae1c7c8e9e9e4b698a Apr 21 10:24:10.480212 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.480177 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" event={"ID":"87b361ed-8851-4e8a-82fb-20213dfbf680","Type":"ContainerStarted","Data":"b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88"} Apr 21 10:24:10.480414 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.480217 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" event={"ID":"87b361ed-8851-4e8a-82fb-20213dfbf680","Type":"ContainerStarted","Data":"f941df91afd2f9bcaedf8f5406ef676bcf018f1d518da6ae1c7c8e9e9e4b698a"} Apr 21 10:24:10.480414 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.480383 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" Apr 21 10:24:10.481688 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.481661 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" event={"ID":"5905aa4b-bad0-42a9-943f-01f5147fcd54","Type":"ContainerStarted","Data":"810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b"} Apr 21 10:24:10.481785 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.481695 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" event={"ID":"5905aa4b-bad0-42a9-943f-01f5147fcd54","Type":"ContainerStarted","Data":"2a7d2a03dee1e9b372a7640cfd620c33acf10a8f137a894687adfc65b000aa18"} Apr 21 10:24:10.481878 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.481856 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" Apr 21 10:24:10.481983 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.481962 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.41:8080: connect: connection refused" Apr 21 10:24:10.482853 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.482831 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.40:8080: connect: connection refused" Apr 21 10:24:10.496562 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.496525 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" podStartSLOduration=0.496514204 podStartE2EDuration="496.514204ms" podCreationTimestamp="2026-04-21 10:24:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:24:10.495583509 +0000 UTC m=+1216.576799342" watchObservedRunningTime="2026-04-21 10:24:10.496514204 +0000 UTC m=+1216.577730036" Apr 21 10:24:10.509872 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:10.509824 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" podStartSLOduration=1.509814445 podStartE2EDuration="1.509814445s" podCreationTimestamp="2026-04-21 10:24:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:24:10.509247361 +0000 UTC m=+1216.590463194" watchObservedRunningTime="2026-04-21 10:24:10.509814445 +0000 UTC m=+1216.591030278" Apr 21 10:24:11.199385 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:11.198855 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.36:8080: connect: connection refused" Apr 21 10:24:11.199385 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:11.199323 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.37:8080: connect: connection refused" Apr 21 10:24:11.485125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:11.485037 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.41:8080: connect: connection refused" Apr 21 10:24:11.485298 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:11.485046 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.40:8080: connect: connection refused" Apr 21 10:24:13.041101 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.041076 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" Apr 21 10:24:13.493009 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.492971 2570 generic.go:358] "Generic (PLEG): container finished" podID="456d8fbb-19dc-4af0-804a-4847333b944a" containerID="e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8" exitCode=0 Apr 21 10:24:13.493189 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.493031 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" event={"ID":"456d8fbb-19dc-4af0-804a-4847333b944a","Type":"ContainerDied","Data":"e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8"} Apr 21 10:24:13.493189 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.493054 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" event={"ID":"456d8fbb-19dc-4af0-804a-4847333b944a","Type":"ContainerDied","Data":"ca001ab8f05fb47c696a5d3e66e100a0ededd46f7d7833a5a05bb0c8e25b266b"} Apr 21 10:24:13.493189 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.493069 2570 scope.go:117] "RemoveContainer" containerID="e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8" Apr 21 10:24:13.493189 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.493031 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8" Apr 21 10:24:13.502960 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.502940 2570 scope.go:117] "RemoveContainer" containerID="e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8" Apr 21 10:24:13.503369 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:24:13.503246 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8\": container with ID starting with e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8 not found: ID does not exist" containerID="e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8" Apr 21 10:24:13.503369 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.503339 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8"} err="failed to get container status \"e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8\": rpc error: code = NotFound desc = could not find container \"e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8\": container with ID starting with e5c44ff5bb19524350992924f0ee727df96473582f32848127ede609e90e05c8 not found: ID does not exist" Apr 21 10:24:13.517041 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.517017 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8"] Apr 21 10:24:13.520915 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.520896 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-3f0d8-predictor-d56d46fd9-5wdh8"] Apr 21 10:24:13.687681 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:13.687655 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" Apr 21 10:24:14.462742 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.462711 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" path="/var/lib/kubelet/pods/456d8fbb-19dc-4af0-804a-4847333b944a/volumes" Apr 21 10:24:14.496922 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.496890 2570 generic.go:358] "Generic (PLEG): container finished" podID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerID="5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9" exitCode=0 Apr 21 10:24:14.497026 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.496950 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" Apr 21 10:24:14.497026 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.496976 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" event={"ID":"575bdf2e-3de3-453a-9809-f22e191e5ce2","Type":"ContainerDied","Data":"5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9"} Apr 21 10:24:14.497026 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.497006 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4" event={"ID":"575bdf2e-3de3-453a-9809-f22e191e5ce2","Type":"ContainerDied","Data":"343a3e35768b7daf8a3f733e7c787d2f52b204d327707c31a4cbedadbfb8fb8a"} Apr 21 10:24:14.497197 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.497026 2570 scope.go:117] "RemoveContainer" containerID="5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9" Apr 21 10:24:14.507485 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.507463 2570 scope.go:117] "RemoveContainer" containerID="5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9" Apr 21 10:24:14.507741 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:24:14.507720 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9\": container with ID starting with 5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9 not found: ID does not exist" containerID="5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9" Apr 21 10:24:14.507825 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.507747 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9"} err="failed to get container status \"5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9\": rpc error: code = NotFound desc = could not find container \"5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9\": container with ID starting with 5b1488318dc7ed1c5a91a3db4b9e2943f2b98ddff21215685f89a84ca0e1afe9 not found: ID does not exist" Apr 21 10:24:14.513894 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.513870 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4"] Apr 21 10:24:14.517239 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:14.517216 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-3f0d8-predictor-567d69bfc6-x22j4"] Apr 21 10:24:16.390320 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:16.390260 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.39:8080: connect: connection refused" Apr 21 10:24:16.390718 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:16.390261 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.38:8080: connect: connection refused" Apr 21 10:24:16.468102 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:16.468065 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" path="/var/lib/kubelet/pods/575bdf2e-3de3-453a-9809-f22e191e5ce2/volumes" Apr 21 10:24:21.485413 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:21.485374 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.40:8080: connect: connection refused" Apr 21 10:24:21.485803 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:21.485374 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.41:8080: connect: connection refused" Apr 21 10:24:26.391023 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:26.390983 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.39:8080: connect: connection refused" Apr 21 10:24:26.391459 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:26.390982 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.38:8080: connect: connection refused" Apr 21 10:24:31.485861 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:31.485817 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.40:8080: connect: connection refused" Apr 21 10:24:31.486298 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:31.485823 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.41:8080: connect: connection refused" Apr 21 10:24:36.391151 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:36.391122 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" Apr 21 10:24:36.391613 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:36.391171 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" Apr 21 10:24:41.486039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:41.485996 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.40:8080: connect: connection refused" Apr 21 10:24:41.486513 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:41.486004 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.41:8080: connect: connection refused" Apr 21 10:24:51.485633 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:51.485591 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.40:8080: connect: connection refused" Apr 21 10:24:51.485633 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:24:51.485611 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.41:8080: connect: connection refused" Apr 21 10:25:01.486360 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:25:01.486323 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" Apr 21 10:25:01.486807 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:25:01.486420 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" Apr 21 10:33:09.015702 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.015620 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8"] Apr 21 10:33:09.018398 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.016138 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" containerID="cri-o://88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1" gracePeriod=30 Apr 21 10:33:09.074621 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.074584 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827"] Apr 21 10:33:09.074902 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.074874 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" containerID="cri-o://3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985" gracePeriod=30 Apr 21 10:33:09.076826 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.076785 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94"] Apr 21 10:33:09.077359 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.077339 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" Apr 21 10:33:09.077491 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.077361 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" Apr 21 10:33:09.077491 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.077407 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" Apr 21 10:33:09.077491 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.077415 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" Apr 21 10:33:09.077491 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.077485 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="456d8fbb-19dc-4af0-804a-4847333b944a" containerName="kserve-container" Apr 21 10:33:09.077692 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.077497 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="575bdf2e-3de3-453a-9809-f22e191e5ce2" containerName="kserve-container" Apr 21 10:33:09.080451 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.080429 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" Apr 21 10:33:09.091157 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.091135 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" Apr 21 10:33:09.092844 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.092816 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94"] Apr 21 10:33:09.125083 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.125025 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g"] Apr 21 10:33:09.128503 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.128481 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" Apr 21 10:33:09.134667 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.134645 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g"] Apr 21 10:33:09.140244 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.140224 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" Apr 21 10:33:09.232942 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.232909 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94"] Apr 21 10:33:09.236875 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:33:09.236662 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod635fbfdb_7f13_4ee2_92ad_91300e3bd4f1.slice/crio-68d2597d833bba2c0a7ccaa9929c0a33aa012c40acccdf97acf4853bcc6e9489 WatchSource:0}: Error finding container 68d2597d833bba2c0a7ccaa9929c0a33aa012c40acccdf97acf4853bcc6e9489: Status 404 returned error can't find the container with id 68d2597d833bba2c0a7ccaa9929c0a33aa012c40acccdf97acf4853bcc6e9489 Apr 21 10:33:09.238889 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.238868 2570 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 10:33:09.252410 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.252384 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" event={"ID":"635fbfdb-7f13-4ee2-92ad-91300e3bd4f1","Type":"ContainerStarted","Data":"68d2597d833bba2c0a7ccaa9929c0a33aa012c40acccdf97acf4853bcc6e9489"} Apr 21 10:33:09.283312 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:09.283289 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g"] Apr 21 10:33:09.298921 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:33:09.298892 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod809b854e_ae0f_40f4_a846_c002b8003e75.slice/crio-a934ffcfa8b5da66d9af03bd8bc3edec25185fd7c5c19c7f714d186ab5437cdb WatchSource:0}: Error finding container a934ffcfa8b5da66d9af03bd8bc3edec25185fd7c5c19c7f714d186ab5437cdb: Status 404 returned error can't find the container with id a934ffcfa8b5da66d9af03bd8bc3edec25185fd7c5c19c7f714d186ab5437cdb Apr 21 10:33:10.256293 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:10.256233 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" event={"ID":"635fbfdb-7f13-4ee2-92ad-91300e3bd4f1","Type":"ContainerStarted","Data":"a9f0b2328251f9a394827a4de808a55abf953f77c96144af33897745c553ce6f"} Apr 21 10:33:10.257642 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:10.257605 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" event={"ID":"809b854e-ae0f-40f4-a846-c002b8003e75","Type":"ContainerStarted","Data":"6e1a746d4fea3e62644ec53b633f0b9bc5e9d770c757c32dcbaeec532ecb26a2"} Apr 21 10:33:10.257642 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:10.257634 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" event={"ID":"809b854e-ae0f-40f4-a846-c002b8003e75","Type":"ContainerStarted","Data":"a934ffcfa8b5da66d9af03bd8bc3edec25185fd7c5c19c7f714d186ab5437cdb"} Apr 21 10:33:10.257815 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:10.257802 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" Apr 21 10:33:10.258965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:10.258939 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.43:8080: connect: connection refused" Apr 21 10:33:10.272443 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:10.272399 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podStartSLOduration=1.272386964 podStartE2EDuration="1.272386964s" podCreationTimestamp="2026-04-21 10:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:33:10.271154066 +0000 UTC m=+1756.352369903" watchObservedRunningTime="2026-04-21 10:33:10.272386964 +0000 UTC m=+1756.353602796" Apr 21 10:33:10.284970 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:10.284921 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podStartSLOduration=1.2849084419999999 podStartE2EDuration="1.284908442s" podCreationTimestamp="2026-04-21 10:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:33:10.284023993 +0000 UTC m=+1756.365239826" watchObservedRunningTime="2026-04-21 10:33:10.284908442 +0000 UTC m=+1756.366124271" Apr 21 10:33:11.264032 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:11.263983 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.43:8080: connect: connection refused" Apr 21 10:33:11.264652 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:11.264630 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" Apr 21 10:33:11.265734 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:11.265695 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.42:8080: connect: connection refused" Apr 21 10:33:12.266305 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:12.266248 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.42:8080: connect: connection refused" Apr 21 10:33:12.840077 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:12.840056 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" Apr 21 10:33:12.843090 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:12.843073 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" Apr 21 10:33:13.270502 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.270467 2570 generic.go:358] "Generic (PLEG): container finished" podID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerID="88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1" exitCode=0 Apr 21 10:33:13.270983 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.270535 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" Apr 21 10:33:13.270983 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.270563 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" event={"ID":"1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e","Type":"ContainerDied","Data":"88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1"} Apr 21 10:33:13.270983 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.270610 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8" event={"ID":"1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e","Type":"ContainerDied","Data":"a58fb6026f2918d7cc7581381b28c27cc33ca8c9f3958de6a1956c608a0bca85"} Apr 21 10:33:13.270983 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.270629 2570 scope.go:117] "RemoveContainer" containerID="88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1" Apr 21 10:33:13.271790 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.271763 2570 generic.go:358] "Generic (PLEG): container finished" podID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerID="3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985" exitCode=0 Apr 21 10:33:13.271901 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.271828 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" Apr 21 10:33:13.271901 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.271834 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" event={"ID":"c8236e00-02b1-4a39-8b3e-69a361739c6f","Type":"ContainerDied","Data":"3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985"} Apr 21 10:33:13.271901 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.271884 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827" event={"ID":"c8236e00-02b1-4a39-8b3e-69a361739c6f","Type":"ContainerDied","Data":"52539c0acb0368f5690a365409aa5717642e689498bfe6d1a22a06af670e54b4"} Apr 21 10:33:13.279004 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.278985 2570 scope.go:117] "RemoveContainer" containerID="88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1" Apr 21 10:33:13.279263 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:33:13.279240 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1\": container with ID starting with 88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1 not found: ID does not exist" containerID="88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1" Apr 21 10:33:13.279379 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.279288 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1"} err="failed to get container status \"88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1\": rpc error: code = NotFound desc = could not find container \"88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1\": container with ID starting with 88c5634ef5981759b37cfd6f25867ffc068860d5f233f42a6b4c107f4bfbd3c1 not found: ID does not exist" Apr 21 10:33:13.279379 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.279332 2570 scope.go:117] "RemoveContainer" containerID="3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985" Apr 21 10:33:13.286771 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.286758 2570 scope.go:117] "RemoveContainer" containerID="3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985" Apr 21 10:33:13.286995 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:33:13.286979 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985\": container with ID starting with 3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985 not found: ID does not exist" containerID="3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985" Apr 21 10:33:13.287045 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.287001 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985"} err="failed to get container status \"3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985\": rpc error: code = NotFound desc = could not find container \"3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985\": container with ID starting with 3b7b2bc204289c67ad3c93f4342f7b7cc34a0a345e5301c3258e3ff552f7b985 not found: ID does not exist" Apr 21 10:33:13.294319 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.294297 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827"] Apr 21 10:33:13.298997 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.298975 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-c50a0-predictor-ff85bf888-w2827"] Apr 21 10:33:13.308250 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.308211 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8"] Apr 21 10:33:13.309777 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:13.309760 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-c50a0-predictor-b6f8dfcbf-tfgv8"] Apr 21 10:33:14.462612 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:14.462572 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" path="/var/lib/kubelet/pods/1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e/volumes" Apr 21 10:33:14.463083 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:14.462930 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" path="/var/lib/kubelet/pods/c8236e00-02b1-4a39-8b3e-69a361739c6f/volumes" Apr 21 10:33:21.264135 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:21.264094 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.43:8080: connect: connection refused" Apr 21 10:33:22.267330 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:22.267288 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.42:8080: connect: connection refused" Apr 21 10:33:31.264807 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:31.264763 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.43:8080: connect: connection refused" Apr 21 10:33:32.266358 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:32.266318 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.42:8080: connect: connection refused" Apr 21 10:33:34.792047 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.792011 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55"] Apr 21 10:33:34.792586 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.792565 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" Apr 21 10:33:34.792664 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.792588 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" Apr 21 10:33:34.792664 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.792628 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" Apr 21 10:33:34.792664 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.792638 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" Apr 21 10:33:34.792807 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.792725 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="c8236e00-02b1-4a39-8b3e-69a361739c6f" containerName="kserve-container" Apr 21 10:33:34.792807 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.792742 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="1bf9dfa1-c1fe-478c-8bb3-f7f0bb782b4e" containerName="kserve-container" Apr 21 10:33:34.795678 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.795654 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" Apr 21 10:33:34.806953 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.806928 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" Apr 21 10:33:34.808106 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.807378 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq"] Apr 21 10:33:34.808106 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.807652 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" containerID="cri-o://810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b" gracePeriod=30 Apr 21 10:33:34.809487 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.809465 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55"] Apr 21 10:33:34.923019 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.922983 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6"] Apr 21 10:33:34.923992 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.923241 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" containerID="cri-o://b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88" gracePeriod=30 Apr 21 10:33:34.925618 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.925576 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp"] Apr 21 10:33:34.930019 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.930002 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" Apr 21 10:33:34.936126 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.936091 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp"] Apr 21 10:33:34.939994 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.939740 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55"] Apr 21 10:33:34.942855 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:34.942837 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" Apr 21 10:33:34.944944 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:33:34.944605 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod697712a4_0529_45ac_8e52_1df71636f57f.slice/crio-bcd8cfead546c26ca9619693b5a0b4ebe2f5904d33b742c54ee994b75d004880 WatchSource:0}: Error finding container bcd8cfead546c26ca9619693b5a0b4ebe2f5904d33b742c54ee994b75d004880: Status 404 returned error can't find the container with id bcd8cfead546c26ca9619693b5a0b4ebe2f5904d33b742c54ee994b75d004880 Apr 21 10:33:35.080970 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.080942 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp"] Apr 21 10:33:35.082813 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:33:35.082776 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c15a100_2884_48e2_b71a_d9600360d8e9.slice/crio-1b50fa356c7b23ffbae7ab81c8d922cdf088e79218fc23609ff285f71950d364 WatchSource:0}: Error finding container 1b50fa356c7b23ffbae7ab81c8d922cdf088e79218fc23609ff285f71950d364: Status 404 returned error can't find the container with id 1b50fa356c7b23ffbae7ab81c8d922cdf088e79218fc23609ff285f71950d364 Apr 21 10:33:35.348453 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.348366 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" event={"ID":"4c15a100-2884-48e2-b71a-d9600360d8e9","Type":"ContainerStarted","Data":"660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53"} Apr 21 10:33:35.348453 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.348401 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" event={"ID":"4c15a100-2884-48e2-b71a-d9600360d8e9","Type":"ContainerStarted","Data":"1b50fa356c7b23ffbae7ab81c8d922cdf088e79218fc23609ff285f71950d364"} Apr 21 10:33:35.348680 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.348577 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" Apr 21 10:33:35.349950 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.349917 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" event={"ID":"697712a4-0529-45ac-8e52-1df71636f57f","Type":"ContainerStarted","Data":"1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166"} Apr 21 10:33:35.350071 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.349954 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" event={"ID":"697712a4-0529-45ac-8e52-1df71636f57f","Type":"ContainerStarted","Data":"bcd8cfead546c26ca9619693b5a0b4ebe2f5904d33b742c54ee994b75d004880"} Apr 21 10:33:35.350123 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.350112 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" Apr 21 10:33:35.350172 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.350150 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.45:8080: connect: connection refused" Apr 21 10:33:35.351028 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.351004 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.44:8080: connect: connection refused" Apr 21 10:33:35.362874 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.362836 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podStartSLOduration=1.362824774 podStartE2EDuration="1.362824774s" podCreationTimestamp="2026-04-21 10:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:33:35.362657885 +0000 UTC m=+1781.443873720" watchObservedRunningTime="2026-04-21 10:33:35.362824774 +0000 UTC m=+1781.444040605" Apr 21 10:33:35.377919 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:35.377882 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podStartSLOduration=1.377871936 podStartE2EDuration="1.377871936s" podCreationTimestamp="2026-04-21 10:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:33:35.376989447 +0000 UTC m=+1781.458205279" watchObservedRunningTime="2026-04-21 10:33:35.377871936 +0000 UTC m=+1781.459087767" Apr 21 10:33:36.353572 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:36.353532 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.44:8080: connect: connection refused" Apr 21 10:33:36.353948 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:36.353616 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.45:8080: connect: connection refused" Apr 21 10:33:37.964001 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:37.963979 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" Apr 21 10:33:38.361534 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.361499 2570 generic.go:358] "Generic (PLEG): container finished" podID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerID="b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88" exitCode=0 Apr 21 10:33:38.361738 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.361557 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" Apr 21 10:33:38.361738 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.361587 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" event={"ID":"87b361ed-8851-4e8a-82fb-20213dfbf680","Type":"ContainerDied","Data":"b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88"} Apr 21 10:33:38.361738 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.361628 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6" event={"ID":"87b361ed-8851-4e8a-82fb-20213dfbf680","Type":"ContainerDied","Data":"f941df91afd2f9bcaedf8f5406ef676bcf018f1d518da6ae1c7c8e9e9e4b698a"} Apr 21 10:33:38.361738 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.361648 2570 scope.go:117] "RemoveContainer" containerID="b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88" Apr 21 10:33:38.370115 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.370099 2570 scope.go:117] "RemoveContainer" containerID="b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88" Apr 21 10:33:38.370400 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:33:38.370382 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88\": container with ID starting with b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88 not found: ID does not exist" containerID="b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88" Apr 21 10:33:38.370456 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.370408 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88"} err="failed to get container status \"b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88\": rpc error: code = NotFound desc = could not find container \"b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88\": container with ID starting with b3f55a6cc6d28d4b2a8f23bdde09dfdbc9d5ff9851fa9df5d5b45388e9148a88 not found: ID does not exist" Apr 21 10:33:38.382608 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.382586 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6"] Apr 21 10:33:38.385992 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.385973 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-258c6-predictor-79cbb6745d-lwxf6"] Apr 21 10:33:38.463773 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.463742 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" path="/var/lib/kubelet/pods/87b361ed-8851-4e8a-82fb-20213dfbf680/volumes" Apr 21 10:33:38.548681 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:38.548661 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" Apr 21 10:33:39.366320 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.366281 2570 generic.go:358] "Generic (PLEG): container finished" podID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerID="810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b" exitCode=0 Apr 21 10:33:39.366791 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.366360 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" Apr 21 10:33:39.366791 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.366362 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" event={"ID":"5905aa4b-bad0-42a9-943f-01f5147fcd54","Type":"ContainerDied","Data":"810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b"} Apr 21 10:33:39.366791 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.366488 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq" event={"ID":"5905aa4b-bad0-42a9-943f-01f5147fcd54","Type":"ContainerDied","Data":"2a7d2a03dee1e9b372a7640cfd620c33acf10a8f137a894687adfc65b000aa18"} Apr 21 10:33:39.366791 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.366516 2570 scope.go:117] "RemoveContainer" containerID="810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b" Apr 21 10:33:39.374730 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.374715 2570 scope.go:117] "RemoveContainer" containerID="810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b" Apr 21 10:33:39.374957 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:33:39.374939 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b\": container with ID starting with 810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b not found: ID does not exist" containerID="810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b" Apr 21 10:33:39.374998 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.374966 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b"} err="failed to get container status \"810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b\": rpc error: code = NotFound desc = could not find container \"810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b\": container with ID starting with 810d53cb1f44f3ec41c5b7253bb9877c342a25dbdd5efed90e718f2b550d398b not found: ID does not exist" Apr 21 10:33:39.386541 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.386521 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq"] Apr 21 10:33:39.387855 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:39.387837 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-258c6-predictor-868bd945f5-k9mzq"] Apr 21 10:33:40.462040 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:40.462003 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" path="/var/lib/kubelet/pods/5905aa4b-bad0-42a9-943f-01f5147fcd54/volumes" Apr 21 10:33:41.264061 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:41.264016 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.43:8080: connect: connection refused" Apr 21 10:33:42.266601 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:42.266552 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.42:8080: connect: connection refused" Apr 21 10:33:46.354403 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:46.354316 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.44:8080: connect: connection refused" Apr 21 10:33:46.354769 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:46.354316 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.45:8080: connect: connection refused" Apr 21 10:33:51.264492 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:51.264444 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.43:8080: connect: connection refused" Apr 21 10:33:52.267254 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:52.267207 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.42:8080: connect: connection refused" Apr 21 10:33:56.354517 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:56.354470 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.44:8080: connect: connection refused" Apr 21 10:33:56.354960 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:33:56.354472 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.45:8080: connect: connection refused" Apr 21 10:34:01.265456 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:01.265429 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" Apr 21 10:34:02.267398 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:02.267366 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" Apr 21 10:34:06.353993 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:06.353945 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.44:8080: connect: connection refused" Apr 21 10:34:06.354389 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:06.353946 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.45:8080: connect: connection refused" Apr 21 10:34:16.354314 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:16.354257 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.44:8080: connect: connection refused" Apr 21 10:34:16.354770 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:16.354257 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.45:8080: connect: connection refused" Apr 21 10:34:26.354930 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:26.354899 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" Apr 21 10:34:26.355336 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:26.354969 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" Apr 21 10:34:29.400702 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.400669 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8"] Apr 21 10:34:29.401078 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.400997 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" Apr 21 10:34:29.401078 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.401007 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" Apr 21 10:34:29.401078 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.401022 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" Apr 21 10:34:29.401078 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.401028 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" Apr 21 10:34:29.401078 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.401077 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="87b361ed-8851-4e8a-82fb-20213dfbf680" containerName="kserve-container" Apr 21 10:34:29.401283 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.401086 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="5905aa4b-bad0-42a9-943f-01f5147fcd54" containerName="kserve-container" Apr 21 10:34:29.404050 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.404029 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94"] Apr 21 10:34:29.404262 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.404229 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" containerID="cri-o://a9f0b2328251f9a394827a4de808a55abf953f77c96144af33897745c553ce6f" gracePeriod=30 Apr 21 10:34:29.404411 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.404236 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" Apr 21 10:34:29.412667 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.412643 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8"] Apr 21 10:34:29.414965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.414946 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" Apr 21 10:34:29.488570 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.488535 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g"] Apr 21 10:34:29.494659 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.494338 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g"] Apr 21 10:34:29.495480 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.494873 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g"] Apr 21 10:34:29.495480 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.494917 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" containerID="cri-o://6e1a746d4fea3e62644ec53b633f0b9bc5e9d770c757c32dcbaeec532ecb26a2" gracePeriod=30 Apr 21 10:34:29.495480 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.495082 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" Apr 21 10:34:29.509510 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.509454 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" Apr 21 10:34:29.558740 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.558709 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8"] Apr 21 10:34:29.562527 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:34:29.562501 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b96aab8_e9c2_49e3_8530_50ff3a36fa65.slice/crio-9c8fdaac0e3621d5adb96b1cc37f53d87b00307458844badcac8b40abeeae8bf WatchSource:0}: Error finding container 9c8fdaac0e3621d5adb96b1cc37f53d87b00307458844badcac8b40abeeae8bf: Status 404 returned error can't find the container with id 9c8fdaac0e3621d5adb96b1cc37f53d87b00307458844badcac8b40abeeae8bf Apr 21 10:34:29.644842 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:29.644823 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g"] Apr 21 10:34:29.646826 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:34:29.646806 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77237322_1ae3_48fb_9167_1ac10f80be3d.slice/crio-5cbaf5c09aa3e1922b127da1b62618096323a3d65794e012309d3ac87dec9604 WatchSource:0}: Error finding container 5cbaf5c09aa3e1922b127da1b62618096323a3d65794e012309d3ac87dec9604: Status 404 returned error can't find the container with id 5cbaf5c09aa3e1922b127da1b62618096323a3d65794e012309d3ac87dec9604 Apr 21 10:34:30.531085 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.531051 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" event={"ID":"77237322-1ae3-48fb-9167-1ac10f80be3d","Type":"ContainerStarted","Data":"77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7"} Apr 21 10:34:30.531085 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.531093 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" event={"ID":"77237322-1ae3-48fb-9167-1ac10f80be3d","Type":"ContainerStarted","Data":"5cbaf5c09aa3e1922b127da1b62618096323a3d65794e012309d3ac87dec9604"} Apr 21 10:34:30.531626 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.531109 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" Apr 21 10:34:30.532456 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.532423 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" event={"ID":"5b96aab8-e9c2-49e3-8530-50ff3a36fa65","Type":"ContainerStarted","Data":"2ed159dcb1069403d5c59b47a85e41d5d54934c18aa238a0608400dde2ebf87f"} Apr 21 10:34:30.532456 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.532456 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" event={"ID":"5b96aab8-e9c2-49e3-8530-50ff3a36fa65","Type":"ContainerStarted","Data":"9c8fdaac0e3621d5adb96b1cc37f53d87b00307458844badcac8b40abeeae8bf"} Apr 21 10:34:30.532622 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.532600 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" Apr 21 10:34:30.532732 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.532643 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.47:8080: connect: connection refused" Apr 21 10:34:30.533547 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.533523 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.46:8080: connect: connection refused" Apr 21 10:34:30.546027 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.545977 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" podStartSLOduration=1.545955648 podStartE2EDuration="1.545955648s" podCreationTimestamp="2026-04-21 10:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:34:30.544785203 +0000 UTC m=+1836.626001210" watchObservedRunningTime="2026-04-21 10:34:30.545955648 +0000 UTC m=+1836.627171481" Apr 21 10:34:30.559913 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:30.559876 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" podStartSLOduration=1.5598660149999999 podStartE2EDuration="1.559866015s" podCreationTimestamp="2026-04-21 10:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:34:30.55860784 +0000 UTC m=+1836.639823672" watchObservedRunningTime="2026-04-21 10:34:30.559866015 +0000 UTC m=+1836.641081847" Apr 21 10:34:31.264591 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:31.264553 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.43:8080: connect: connection refused" Apr 21 10:34:31.535653 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:31.535552 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.47:8080: connect: connection refused" Apr 21 10:34:31.535653 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:31.535598 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.46:8080: connect: connection refused" Apr 21 10:34:32.267235 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.267195 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.42:8080: connect: connection refused" Apr 21 10:34:32.540430 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.540398 2570 generic.go:358] "Generic (PLEG): container finished" podID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerID="a9f0b2328251f9a394827a4de808a55abf953f77c96144af33897745c553ce6f" exitCode=0 Apr 21 10:34:32.540812 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.540473 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" event={"ID":"635fbfdb-7f13-4ee2-92ad-91300e3bd4f1","Type":"ContainerDied","Data":"a9f0b2328251f9a394827a4de808a55abf953f77c96144af33897745c553ce6f"} Apr 21 10:34:32.541647 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.541623 2570 generic.go:358] "Generic (PLEG): container finished" podID="809b854e-ae0f-40f4-a846-c002b8003e75" containerID="6e1a746d4fea3e62644ec53b633f0b9bc5e9d770c757c32dcbaeec532ecb26a2" exitCode=0 Apr 21 10:34:32.541763 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.541686 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" event={"ID":"809b854e-ae0f-40f4-a846-c002b8003e75","Type":"ContainerDied","Data":"6e1a746d4fea3e62644ec53b633f0b9bc5e9d770c757c32dcbaeec532ecb26a2"} Apr 21 10:34:32.541763 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.541714 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" event={"ID":"809b854e-ae0f-40f4-a846-c002b8003e75","Type":"ContainerDied","Data":"a934ffcfa8b5da66d9af03bd8bc3edec25185fd7c5c19c7f714d186ab5437cdb"} Apr 21 10:34:32.541763 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.541727 2570 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a934ffcfa8b5da66d9af03bd8bc3edec25185fd7c5c19c7f714d186ab5437cdb" Apr 21 10:34:32.542500 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.542484 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" Apr 21 10:34:32.836374 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:32.836350 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" Apr 21 10:34:33.545898 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:33.545866 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" Apr 21 10:34:33.545898 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:33.545888 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g" Apr 21 10:34:33.546405 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:33.545866 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94" event={"ID":"635fbfdb-7f13-4ee2-92ad-91300e3bd4f1","Type":"ContainerDied","Data":"68d2597d833bba2c0a7ccaa9929c0a33aa012c40acccdf97acf4853bcc6e9489"} Apr 21 10:34:33.546405 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:33.545992 2570 scope.go:117] "RemoveContainer" containerID="a9f0b2328251f9a394827a4de808a55abf953f77c96144af33897745c553ce6f" Apr 21 10:34:33.570679 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:33.570658 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g"] Apr 21 10:34:33.574185 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:33.574162 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-e8429-predictor-7f7756df8b-vh46g"] Apr 21 10:34:33.583580 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:33.583556 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94"] Apr 21 10:34:33.585537 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:33.585518 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-e8429-predictor-7dc75fffbc-spl94"] Apr 21 10:34:34.463228 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:34.463194 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" path="/var/lib/kubelet/pods/635fbfdb-7f13-4ee2-92ad-91300e3bd4f1/volumes" Apr 21 10:34:34.463470 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:34.463455 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" path="/var/lib/kubelet/pods/809b854e-ae0f-40f4-a846-c002b8003e75/volumes" Apr 21 10:34:41.536209 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:41.536160 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.46:8080: connect: connection refused" Apr 21 10:34:41.536607 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:41.536160 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.47:8080: connect: connection refused" Apr 21 10:34:51.535635 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:51.535591 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.47:8080: connect: connection refused" Apr 21 10:34:51.536146 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:34:51.535782 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.46:8080: connect: connection refused" Apr 21 10:35:01.536392 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:35:01.536349 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.47:8080: connect: connection refused" Apr 21 10:35:01.536770 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:35:01.536349 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.46:8080: connect: connection refused" Apr 21 10:35:11.536249 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:35:11.536203 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.47:8080: connect: connection refused" Apr 21 10:35:11.536718 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:35:11.536210 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.46:8080: connect: connection refused" Apr 21 10:35:21.536167 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:35:21.536133 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" Apr 21 10:35:21.536570 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:35:21.536409 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" Apr 21 10:39:54.553388 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:39:54.553354 2570 scope.go:117] "RemoveContainer" containerID="6e1a746d4fea3e62644ec53b633f0b9bc5e9d770c757c32dcbaeec532ecb26a2" Apr 21 10:43:54.356302 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:54.356242 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8"] Apr 21 10:43:54.356814 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:54.356505 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" containerID="cri-o://2ed159dcb1069403d5c59b47a85e41d5d54934c18aa238a0608400dde2ebf87f" gracePeriod=30 Apr 21 10:43:54.395611 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:54.395578 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g"] Apr 21 10:43:54.395833 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:54.395812 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" containerID="cri-o://77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7" gracePeriod=30 Apr 21 10:43:57.366764 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:57.366726 2570 generic.go:358] "Generic (PLEG): container finished" podID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerID="2ed159dcb1069403d5c59b47a85e41d5d54934c18aa238a0608400dde2ebf87f" exitCode=0 Apr 21 10:43:57.367110 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:57.366795 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" event={"ID":"5b96aab8-e9c2-49e3-8530-50ff3a36fa65","Type":"ContainerDied","Data":"2ed159dcb1069403d5c59b47a85e41d5d54934c18aa238a0608400dde2ebf87f"} Apr 21 10:43:57.415481 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:57.415457 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" Apr 21 10:43:57.527617 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:57.527596 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" Apr 21 10:43:58.370963 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.370925 2570 generic.go:358] "Generic (PLEG): container finished" podID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerID="77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7" exitCode=0 Apr 21 10:43:58.371423 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.370984 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" Apr 21 10:43:58.371423 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.371006 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" event={"ID":"77237322-1ae3-48fb-9167-1ac10f80be3d","Type":"ContainerDied","Data":"77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7"} Apr 21 10:43:58.371423 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.371041 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g" event={"ID":"77237322-1ae3-48fb-9167-1ac10f80be3d","Type":"ContainerDied","Data":"5cbaf5c09aa3e1922b127da1b62618096323a3d65794e012309d3ac87dec9604"} Apr 21 10:43:58.371423 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.371060 2570 scope.go:117] "RemoveContainer" containerID="77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7" Apr 21 10:43:58.372333 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.372195 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" event={"ID":"5b96aab8-e9c2-49e3-8530-50ff3a36fa65","Type":"ContainerDied","Data":"9c8fdaac0e3621d5adb96b1cc37f53d87b00307458844badcac8b40abeeae8bf"} Apr 21 10:43:58.372333 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.372291 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8" Apr 21 10:43:58.381451 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.381435 2570 scope.go:117] "RemoveContainer" containerID="77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7" Apr 21 10:43:58.381682 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:43:58.381657 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7\": container with ID starting with 77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7 not found: ID does not exist" containerID="77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7" Apr 21 10:43:58.381774 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.381686 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7"} err="failed to get container status \"77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7\": rpc error: code = NotFound desc = could not find container \"77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7\": container with ID starting with 77dd4050b51e2b4dbdcbed78d3e25d32144301b9eb848aae76b8f0bd57565ac7 not found: ID does not exist" Apr 21 10:43:58.381774 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.381702 2570 scope.go:117] "RemoveContainer" containerID="2ed159dcb1069403d5c59b47a85e41d5d54934c18aa238a0608400dde2ebf87f" Apr 21 10:43:58.398923 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.398897 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g"] Apr 21 10:43:58.402024 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.402001 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-932d4-predictor-dcd4cd5db-xc22g"] Apr 21 10:43:58.415235 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.415213 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8"] Apr 21 10:43:58.419785 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.419760 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-932d4-predictor-c79874459-rn8v8"] Apr 21 10:43:58.463744 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.463675 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" path="/var/lib/kubelet/pods/5b96aab8-e9c2-49e3-8530-50ff3a36fa65/volumes" Apr 21 10:43:58.463916 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:43:58.463903 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" path="/var/lib/kubelet/pods/77237322-1ae3-48fb-9167-1ac10f80be3d/volumes" Apr 21 10:51:04.335916 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:04.335881 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55"] Apr 21 10:51:04.338321 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:04.336123 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" containerID="cri-o://1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166" gracePeriod=30 Apr 21 10:51:04.390281 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:04.390230 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp"] Apr 21 10:51:04.390497 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:04.390475 2570 kuberuntime_container.go:864] "Killing container with a grace period" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" containerID="cri-o://660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53" gracePeriod=30 Apr 21 10:51:06.354081 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:06.354043 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.44:8080: connect: connection refused" Apr 21 10:51:06.354477 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:06.354049 2570 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.45:8080: connect: connection refused" Apr 21 10:51:07.409422 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.409399 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" Apr 21 10:51:07.532586 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.532563 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" Apr 21 10:51:07.721841 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.721752 2570 generic.go:358] "Generic (PLEG): container finished" podID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerID="660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53" exitCode=0 Apr 21 10:51:07.721841 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.721818 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" Apr 21 10:51:07.721841 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.721829 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" event={"ID":"4c15a100-2884-48e2-b71a-d9600360d8e9","Type":"ContainerDied","Data":"660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53"} Apr 21 10:51:07.722147 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.721864 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp" event={"ID":"4c15a100-2884-48e2-b71a-d9600360d8e9","Type":"ContainerDied","Data":"1b50fa356c7b23ffbae7ab81c8d922cdf088e79218fc23609ff285f71950d364"} Apr 21 10:51:07.722147 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.721876 2570 scope.go:117] "RemoveContainer" containerID="660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53" Apr 21 10:51:07.723049 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.723026 2570 generic.go:358] "Generic (PLEG): container finished" podID="697712a4-0529-45ac-8e52-1df71636f57f" containerID="1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166" exitCode=0 Apr 21 10:51:07.723144 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.723059 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" event={"ID":"697712a4-0529-45ac-8e52-1df71636f57f","Type":"ContainerDied","Data":"1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166"} Apr 21 10:51:07.723144 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.723080 2570 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" Apr 21 10:51:07.723144 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.723089 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55" event={"ID":"697712a4-0529-45ac-8e52-1df71636f57f","Type":"ContainerDied","Data":"bcd8cfead546c26ca9619693b5a0b4ebe2f5904d33b742c54ee994b75d004880"} Apr 21 10:51:07.730916 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.730871 2570 scope.go:117] "RemoveContainer" containerID="660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53" Apr 21 10:51:07.731150 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:51:07.731122 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53\": container with ID starting with 660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53 not found: ID does not exist" containerID="660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53" Apr 21 10:51:07.731209 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.731156 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53"} err="failed to get container status \"660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53\": rpc error: code = NotFound desc = could not find container \"660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53\": container with ID starting with 660436b11966812529d29bf5faf23204853293458501fb50f903488dfe5a4f53 not found: ID does not exist" Apr 21 10:51:07.731209 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.731178 2570 scope.go:117] "RemoveContainer" containerID="1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166" Apr 21 10:51:07.738781 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.738765 2570 scope.go:117] "RemoveContainer" containerID="1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166" Apr 21 10:51:07.739020 ip-10-0-142-243 kubenswrapper[2570]: E0421 10:51:07.739002 2570 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166\": container with ID starting with 1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166 not found: ID does not exist" containerID="1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166" Apr 21 10:51:07.739065 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.739029 2570 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166"} err="failed to get container status \"1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166\": rpc error: code = NotFound desc = could not find container \"1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166\": container with ID starting with 1f068f4771a0feae24a3a70e92f058b595de251290e99494420e2fa5f851f166 not found: ID does not exist" Apr 21 10:51:07.746438 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.746415 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55"] Apr 21 10:51:07.750192 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.750173 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-b394f-predictor-547b67dbcf-vjh55"] Apr 21 10:51:07.760295 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.760258 2570 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp"] Apr 21 10:51:07.764132 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:07.764102 2570 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kserve-ci-e2e-test/error-404-isvc-b394f-predictor-5978fdbfd6-dhspp"] Apr 21 10:51:08.462363 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:08.462253 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" path="/var/lib/kubelet/pods/4c15a100-2884-48e2-b71a-d9600360d8e9/volumes" Apr 21 10:51:08.462703 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:08.462520 2570 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="697712a4-0529-45ac-8e52-1df71636f57f" path="/var/lib/kubelet/pods/697712a4-0529-45ac-8e52-1df71636f57f/volumes" Apr 21 10:51:30.057641 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.057575 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-95dzh/must-gather-zd7dm"] Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058011 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058030 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058053 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058063 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058071 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058078 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058090 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058098 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058110 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058117 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" Apr 21 10:51:30.058125 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058128 2570 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" Apr 21 10:51:30.058672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058136 2570 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" Apr 21 10:51:30.058672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058209 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="5b96aab8-e9c2-49e3-8530-50ff3a36fa65" containerName="kserve-container" Apr 21 10:51:30.058672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058224 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="809b854e-ae0f-40f4-a846-c002b8003e75" containerName="kserve-container" Apr 21 10:51:30.058672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058235 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="697712a4-0529-45ac-8e52-1df71636f57f" containerName="kserve-container" Apr 21 10:51:30.058672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058244 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="635fbfdb-7f13-4ee2-92ad-91300e3bd4f1" containerName="kserve-container" Apr 21 10:51:30.058672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058253 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="77237322-1ae3-48fb-9167-1ac10f80be3d" containerName="kserve-container" Apr 21 10:51:30.058672 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.058284 2570 memory_manager.go:356] "RemoveStaleState removing state" podUID="4c15a100-2884-48e2-b71a-d9600360d8e9" containerName="kserve-container" Apr 21 10:51:30.062622 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.062601 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95dzh/must-gather-zd7dm" Apr 21 10:51:30.065048 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.065017 2570 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-95dzh\"/\"default-dockercfg-nrjgl\"" Apr 21 10:51:30.065147 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.065047 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-95dzh\"/\"kube-root-ca.crt\"" Apr 21 10:51:30.065147 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.065017 2570 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-95dzh\"/\"openshift-service-ca.crt\"" Apr 21 10:51:30.067870 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.067842 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-95dzh/must-gather-zd7dm"] Apr 21 10:51:30.157046 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.157006 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2kgp\" (UniqueName: \"kubernetes.io/projected/cd09d1af-6666-4ea2-b234-d709bea6308e-kube-api-access-k2kgp\") pod \"must-gather-zd7dm\" (UID: \"cd09d1af-6666-4ea2-b234-d709bea6308e\") " pod="openshift-must-gather-95dzh/must-gather-zd7dm" Apr 21 10:51:30.157206 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.157059 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cd09d1af-6666-4ea2-b234-d709bea6308e-must-gather-output\") pod \"must-gather-zd7dm\" (UID: \"cd09d1af-6666-4ea2-b234-d709bea6308e\") " pod="openshift-must-gather-95dzh/must-gather-zd7dm" Apr 21 10:51:30.258328 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.258258 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-k2kgp\" (UniqueName: \"kubernetes.io/projected/cd09d1af-6666-4ea2-b234-d709bea6308e-kube-api-access-k2kgp\") pod \"must-gather-zd7dm\" (UID: \"cd09d1af-6666-4ea2-b234-d709bea6308e\") " pod="openshift-must-gather-95dzh/must-gather-zd7dm" Apr 21 10:51:30.258483 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.258356 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cd09d1af-6666-4ea2-b234-d709bea6308e-must-gather-output\") pod \"must-gather-zd7dm\" (UID: \"cd09d1af-6666-4ea2-b234-d709bea6308e\") " pod="openshift-must-gather-95dzh/must-gather-zd7dm" Apr 21 10:51:30.258671 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.258654 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cd09d1af-6666-4ea2-b234-d709bea6308e-must-gather-output\") pod \"must-gather-zd7dm\" (UID: \"cd09d1af-6666-4ea2-b234-d709bea6308e\") " pod="openshift-must-gather-95dzh/must-gather-zd7dm" Apr 21 10:51:30.266814 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.266787 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2kgp\" (UniqueName: \"kubernetes.io/projected/cd09d1af-6666-4ea2-b234-d709bea6308e-kube-api-access-k2kgp\") pod \"must-gather-zd7dm\" (UID: \"cd09d1af-6666-4ea2-b234-d709bea6308e\") " pod="openshift-must-gather-95dzh/must-gather-zd7dm" Apr 21 10:51:30.382909 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.382876 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95dzh/must-gather-zd7dm" Apr 21 10:51:30.502301 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.502251 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-95dzh/must-gather-zd7dm"] Apr 21 10:51:30.505478 ip-10-0-142-243 kubenswrapper[2570]: W0421 10:51:30.505443 2570 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd09d1af_6666_4ea2_b234_d709bea6308e.slice/crio-bc8ef8425e5843a4b54ea9e38fa97250321df078c4f3e7c921b1331e458dde38 WatchSource:0}: Error finding container bc8ef8425e5843a4b54ea9e38fa97250321df078c4f3e7c921b1331e458dde38: Status 404 returned error can't find the container with id bc8ef8425e5843a4b54ea9e38fa97250321df078c4f3e7c921b1331e458dde38 Apr 21 10:51:30.507659 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.507640 2570 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 10:51:30.796187 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:30.796101 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95dzh/must-gather-zd7dm" event={"ID":"cd09d1af-6666-4ea2-b234-d709bea6308e","Type":"ContainerStarted","Data":"bc8ef8425e5843a4b54ea9e38fa97250321df078c4f3e7c921b1331e458dde38"} Apr 21 10:51:31.802782 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:31.802735 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95dzh/must-gather-zd7dm" event={"ID":"cd09d1af-6666-4ea2-b234-d709bea6308e","Type":"ContainerStarted","Data":"f6bd6b4863f41c7688c2ad36642ed1f750207d61c343407a8d9f35f739204282"} Apr 21 10:51:31.802782 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:31.802774 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95dzh/must-gather-zd7dm" event={"ID":"cd09d1af-6666-4ea2-b234-d709bea6308e","Type":"ContainerStarted","Data":"58f21240959b90ee1fbfe494930fc61720db3b9083851fe0b3d903347a6559a9"} Apr 21 10:51:31.822812 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:31.821582 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-95dzh/must-gather-zd7dm" podStartSLOduration=0.900628546 podStartE2EDuration="1.821563337s" podCreationTimestamp="2026-04-21 10:51:30 +0000 UTC" firstStartedPulling="2026-04-21 10:51:30.507795277 +0000 UTC m=+2856.589011087" lastFinishedPulling="2026-04-21 10:51:31.428730065 +0000 UTC m=+2857.509945878" observedRunningTime="2026-04-21 10:51:31.818691942 +0000 UTC m=+2857.899907774" watchObservedRunningTime="2026-04-21 10:51:31.821563337 +0000 UTC m=+2857.902779181" Apr 21 10:51:32.869446 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:32.869416 2570 ???:1] "http: TLS handshake error from 10.0.142.243:47506: EOF" Apr 21 10:51:32.875432 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:32.875397 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-bfr9j_f603c14a-fa13-45be-b47d-b74ac1ec1ff5/global-pull-secret-syncer/0.log" Apr 21 10:51:33.067087 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:33.067037 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-pmfn6_924092ef-7a63-4db6-afcd-c4fb3be74aec/konnectivity-agent/0.log" Apr 21 10:51:33.197112 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:33.197027 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-142-243.ec2.internal_b2386108552c5c182289de71222968c1/haproxy/0.log" Apr 21 10:51:36.556820 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:36.556781 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-69db897b98-dkjlx_6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae/kube-state-metrics/0.log" Apr 21 10:51:36.582886 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:36.582860 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-69db897b98-dkjlx_6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae/kube-rbac-proxy-main/0.log" Apr 21 10:51:36.604523 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:36.604489 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-69db897b98-dkjlx_6090567f-b8cb-4bc0-9ad5-3cb61bcdaeae/kube-rbac-proxy-self/0.log" Apr 21 10:51:36.860106 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:36.860074 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-qn6fw_0ee7a123-6bbc-419a-999a-fd25ccd1bba1/node-exporter/0.log" Apr 21 10:51:36.884555 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:36.884526 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-qn6fw_0ee7a123-6bbc-419a-999a-fd25ccd1bba1/kube-rbac-proxy/0.log" Apr 21 10:51:36.908550 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:36.908512 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-qn6fw_0ee7a123-6bbc-419a-999a-fd25ccd1bba1/init-textfile/0.log" Apr 21 10:51:37.190674 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:37.190580 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-5676c8c784-92v6x_56c34418-0154-4f7b-874a-e64fa17eecab/prometheus-operator/0.log" Apr 21 10:51:37.207107 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:37.207082 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-5676c8c784-92v6x_56c34418-0154-4f7b-874a-e64fa17eecab/kube-rbac-proxy/0.log" Apr 21 10:51:37.335128 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:37.335102 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6bcd6c69d5-4x5xp_2acd23ca-ac79-49f3-9242-9cd19cdf3c0e/thanos-query/0.log" Apr 21 10:51:37.358899 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:37.358871 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6bcd6c69d5-4x5xp_2acd23ca-ac79-49f3-9242-9cd19cdf3c0e/kube-rbac-proxy-web/0.log" Apr 21 10:51:37.378754 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:37.378728 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6bcd6c69d5-4x5xp_2acd23ca-ac79-49f3-9242-9cd19cdf3c0e/kube-rbac-proxy/0.log" Apr 21 10:51:37.399043 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:37.399013 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6bcd6c69d5-4x5xp_2acd23ca-ac79-49f3-9242-9cd19cdf3c0e/prom-label-proxy/0.log" Apr 21 10:51:37.418552 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:37.418515 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6bcd6c69d5-4x5xp_2acd23ca-ac79-49f3-9242-9cd19cdf3c0e/kube-rbac-proxy-rules/0.log" Apr 21 10:51:37.441813 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:37.441737 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6bcd6c69d5-4x5xp_2acd23ca-ac79-49f3-9242-9cd19cdf3c0e/kube-rbac-proxy-metrics/0.log" Apr 21 10:51:39.366157 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:39.366118 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5c6b449c9b-fp756_a6f54ef2-5197-4211-bbcc-a5f2ebee0897/console/0.log" Apr 21 10:51:40.161090 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.161055 2570 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc"] Apr 21 10:51:40.166153 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.166117 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.175668 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.175069 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc"] Apr 21 10:51:40.245186 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.245147 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkn74\" (UniqueName: \"kubernetes.io/projected/25de0a4c-d910-4300-a730-28e3026f0f65-kube-api-access-qkn74\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.245396 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.245235 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-podres\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.245396 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.245294 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-lib-modules\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.245396 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.245364 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-proc\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.245541 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.245410 2570 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-sys\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.346849 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.346811 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-proc\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.346988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.346856 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-sys\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.346988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.346890 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qkn74\" (UniqueName: \"kubernetes.io/projected/25de0a4c-d910-4300-a730-28e3026f0f65-kube-api-access-qkn74\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.346988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.346929 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-podres\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.346988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.346933 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-proc\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.346988 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.346974 2570 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-lib-modules\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.347162 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.346994 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-sys\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.347162 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.347092 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-lib-modules\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.347162 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.347144 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/25de0a4c-d910-4300-a730-28e3026f0f65-podres\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.355014 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.354981 2570 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkn74\" (UniqueName: \"kubernetes.io/projected/25de0a4c-d910-4300-a730-28e3026f0f65-kube-api-access-qkn74\") pod \"perf-node-gather-daemonset-cq8lc\" (UID: \"25de0a4c-d910-4300-a730-28e3026f0f65\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.390853 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.390833 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-4fwj6_da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997/dns/0.log" Apr 21 10:51:40.410889 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.410867 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-4fwj6_da8bd8dd-c9e0-4cbd-be5a-cf95d4d01997/kube-rbac-proxy/0.log" Apr 21 10:51:40.486933 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.486883 2570 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.585523 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.585432 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-rrdxp_7c500689-bb9d-4394-869d-32c31072db85/dns-node-resolver/0.log" Apr 21 10:51:40.610643 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.610620 2570 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc"] Apr 21 10:51:40.845144 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.845068 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" event={"ID":"25de0a4c-d910-4300-a730-28e3026f0f65","Type":"ContainerStarted","Data":"89c02ef9154848a1664560f3c4ae7ec430ad3af0dc40ad2b1e9fe20b8528718d"} Apr 21 10:51:40.845144 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.845110 2570 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" event={"ID":"25de0a4c-d910-4300-a730-28e3026f0f65","Type":"ContainerStarted","Data":"ff7249b77fc084a781e26da0c354a8d31afd222d0b5db874fbd26fc3ff872aaf"} Apr 21 10:51:40.845349 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.845159 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:40.862323 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:40.862227 2570 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" podStartSLOduration=0.862208681 podStartE2EDuration="862.208681ms" podCreationTimestamp="2026-04-21 10:51:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:51:40.860362589 +0000 UTC m=+2866.941578421" watchObservedRunningTime="2026-04-21 10:51:40.862208681 +0000 UTC m=+2866.943424514" Apr 21 10:51:41.041111 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:41.041081 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-77lqp_2ae59855-346f-4974-9fe0-7f7658ee240f/node-ca/0.log" Apr 21 10:51:42.137904 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:42.137875 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-hlz49_8683a07e-2776-4d73-9312-ca2168702c96/serve-healthcheck-canary/0.log" Apr 21 10:51:42.514757 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:42.514693 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-ckmzt_e51c2a85-13d4-47bb-916a-fb2ecece2c78/kube-rbac-proxy/0.log" Apr 21 10:51:42.536160 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:42.536130 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-ckmzt_e51c2a85-13d4-47bb-916a-fb2ecece2c78/exporter/0.log" Apr 21 10:51:42.558138 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:42.558110 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-ckmzt_e51c2a85-13d4-47bb-916a-fb2ecece2c78/extractor/0.log" Apr 21 10:51:44.686082 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:44.686047 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve_kserve-controller-manager-84685cd884-qk94h_f6dd56d8-2f6d-4945-9a19-4f3743409c0e/manager/0.log" Apr 21 10:51:44.705394 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:44.705367 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve_llmisvc-controller-manager-68cc5db7c4-rbt9b_8f4d63b7-adda-4734-a417-0e8e17d8a6bb/manager/0.log" Apr 21 10:51:45.221321 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:45.221292 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve_seaweedfs-86cc847c5c-sxtd8_5a4abe0e-37cf-4685-86d4-c476caea9e49/seaweedfs/0.log" Apr 21 10:51:46.858448 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:46.858421 2570 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-cq8lc" Apr 21 10:51:50.475431 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.475407 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rczjx_a458360c-0ec8-4ac8-9524-f44a7102175d/kube-multus-additional-cni-plugins/0.log" Apr 21 10:51:50.494830 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.494809 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rczjx_a458360c-0ec8-4ac8-9524-f44a7102175d/egress-router-binary-copy/0.log" Apr 21 10:51:50.513459 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.513438 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rczjx_a458360c-0ec8-4ac8-9524-f44a7102175d/cni-plugins/0.log" Apr 21 10:51:50.531379 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.531361 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rczjx_a458360c-0ec8-4ac8-9524-f44a7102175d/bond-cni-plugin/0.log" Apr 21 10:51:50.550986 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.550965 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rczjx_a458360c-0ec8-4ac8-9524-f44a7102175d/routeoverride-cni/0.log" Apr 21 10:51:50.568870 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.568851 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rczjx_a458360c-0ec8-4ac8-9524-f44a7102175d/whereabouts-cni-bincopy/0.log" Apr 21 10:51:50.587420 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.587398 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-rczjx_a458360c-0ec8-4ac8-9524-f44a7102175d/whereabouts-cni/0.log" Apr 21 10:51:50.621565 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.621540 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jrkkg_9f83c2fe-affe-48ec-a969-287e374d052c/kube-multus/0.log" Apr 21 10:51:50.701920 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.701890 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-ff5dl_bd143ad0-d084-434c-87b4-bd355e2b205a/network-metrics-daemon/0.log" Apr 21 10:51:50.721827 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:50.721799 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-ff5dl_bd143ad0-d084-434c-87b4-bd355e2b205a/kube-rbac-proxy/0.log" Apr 21 10:51:52.267593 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:52.267545 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qbn7w_e59564ca-d2b4-41c4-a36e-b94f5fc6c40b/ovn-controller/0.log" Apr 21 10:51:52.300092 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:52.300067 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qbn7w_e59564ca-d2b4-41c4-a36e-b94f5fc6c40b/ovn-acl-logging/0.log" Apr 21 10:51:52.326299 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:52.326258 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qbn7w_e59564ca-d2b4-41c4-a36e-b94f5fc6c40b/kube-rbac-proxy-node/0.log" Apr 21 10:51:52.352756 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:52.352683 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qbn7w_e59564ca-d2b4-41c4-a36e-b94f5fc6c40b/kube-rbac-proxy-ovn-metrics/0.log" Apr 21 10:51:52.373744 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:52.373727 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qbn7w_e59564ca-d2b4-41c4-a36e-b94f5fc6c40b/northd/0.log" Apr 21 10:51:52.395039 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:52.395018 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qbn7w_e59564ca-d2b4-41c4-a36e-b94f5fc6c40b/nbdb/0.log" Apr 21 10:51:52.420662 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:52.420614 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qbn7w_e59564ca-d2b4-41c4-a36e-b94f5fc6c40b/sbdb/0.log" Apr 21 10:51:52.546965 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:52.546897 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qbn7w_e59564ca-d2b4-41c4-a36e-b94f5fc6c40b/ovnkube-controller/0.log" Apr 21 10:51:53.516954 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:53.516924 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-2j8xh_d0ada080-47c9-4134-a37c-b3bd33e4eb38/network-check-target-container/0.log" Apr 21 10:51:54.470254 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:54.470230 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-lggw9_3392dd0f-93d6-42ee-a4cb-f9aeb2243a3a/iptables-alerter/0.log" Apr 21 10:51:55.055715 ip-10-0-142-243 kubenswrapper[2570]: I0421 10:51:55.055684 2570 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_tuned-ngjbx_6d247fd6-2379-4000-acc6-743c1e73fbf9/tuned/0.log"