Apr 21 10:03:28.731430 ip-10-0-142-73 systemd[1]: Starting Kubernetes Kubelet... Apr 21 10:03:29.136283 ip-10-0-142-73 kubenswrapper[2575]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 10:03:29.136283 ip-10-0-142-73 kubenswrapper[2575]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 21 10:03:29.136283 ip-10-0-142-73 kubenswrapper[2575]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 10:03:29.136283 ip-10-0-142-73 kubenswrapper[2575]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 21 10:03:29.136283 ip-10-0-142-73 kubenswrapper[2575]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 21 10:03:29.137437 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.137216 2575 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 21 10:03:29.139526 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139508 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:29.139526 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139525 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139530 2575 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139533 2575 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139536 2575 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139539 2575 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139542 2575 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139553 2575 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139557 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139561 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139563 2575 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139566 2575 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139569 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139571 2575 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139574 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139577 2575 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139579 2575 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139582 2575 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139585 2575 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139588 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139595 2575 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:29.139602 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139598 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139601 2575 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139604 2575 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139607 2575 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139609 2575 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139612 2575 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139615 2575 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139618 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139621 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139623 2575 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139626 2575 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139628 2575 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139631 2575 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139633 2575 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139636 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139639 2575 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139641 2575 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139644 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139646 2575 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:29.140075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139649 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139651 2575 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139654 2575 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139657 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139661 2575 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139665 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139668 2575 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139678 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139681 2575 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139683 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139686 2575 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139689 2575 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139692 2575 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139695 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139697 2575 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139701 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139704 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139707 2575 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139710 2575 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:29.140541 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139712 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139715 2575 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139718 2575 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139720 2575 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139723 2575 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139726 2575 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139728 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139731 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139733 2575 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139736 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139738 2575 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139741 2575 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139743 2575 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139746 2575 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139748 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139751 2575 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139754 2575 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139757 2575 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139759 2575 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:29.141009 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139762 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139766 2575 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139769 2575 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139774 2575 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139776 2575 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139779 2575 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139782 2575 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.139785 2575 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140412 2575 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140420 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140423 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140426 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140429 2575 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140431 2575 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140435 2575 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140437 2575 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140440 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140442 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140445 2575 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:29.141486 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140448 2575 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140451 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140453 2575 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140456 2575 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140459 2575 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140461 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140464 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140466 2575 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140469 2575 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140471 2575 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140475 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140478 2575 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140480 2575 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140484 2575 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140488 2575 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140491 2575 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140493 2575 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140496 2575 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140498 2575 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140501 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:29.141946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140503 2575 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140506 2575 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140510 2575 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140513 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140516 2575 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140519 2575 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140522 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140525 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140527 2575 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140530 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140532 2575 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140535 2575 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140537 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140540 2575 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140543 2575 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140546 2575 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140549 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140552 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140555 2575 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:29.142463 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140558 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140561 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140563 2575 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140566 2575 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140569 2575 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140572 2575 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140574 2575 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140577 2575 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140579 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140582 2575 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140584 2575 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140587 2575 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140590 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140592 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140595 2575 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140597 2575 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140599 2575 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140602 2575 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140604 2575 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140607 2575 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:29.142925 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140609 2575 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140612 2575 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140614 2575 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140617 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140619 2575 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140622 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140625 2575 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140627 2575 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140631 2575 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140633 2575 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140636 2575 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140638 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140641 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140644 2575 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140647 2575 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.140650 2575 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140726 2575 flags.go:64] FLAG: --address="0.0.0.0" Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140733 2575 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140740 2575 flags.go:64] FLAG: --anonymous-auth="true" Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140744 2575 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 21 10:03:29.143416 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140749 2575 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140752 2575 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140757 2575 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140761 2575 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140764 2575 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140767 2575 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140771 2575 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140774 2575 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140778 2575 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140781 2575 flags.go:64] FLAG: --cgroup-root="" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140784 2575 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140787 2575 flags.go:64] FLAG: --client-ca-file="" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140790 2575 flags.go:64] FLAG: --cloud-config="" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140792 2575 flags.go:64] FLAG: --cloud-provider="external" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140796 2575 flags.go:64] FLAG: --cluster-dns="[]" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140801 2575 flags.go:64] FLAG: --cluster-domain="" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140804 2575 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140807 2575 flags.go:64] FLAG: --config-dir="" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140810 2575 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140813 2575 flags.go:64] FLAG: --container-log-max-files="5" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140818 2575 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140821 2575 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140825 2575 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140828 2575 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 21 10:03:29.143907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140831 2575 flags.go:64] FLAG: --contention-profiling="false" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140834 2575 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140837 2575 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140840 2575 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140843 2575 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140848 2575 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140851 2575 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140854 2575 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140857 2575 flags.go:64] FLAG: --enable-load-reader="false" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140860 2575 flags.go:64] FLAG: --enable-server="true" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140863 2575 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140868 2575 flags.go:64] FLAG: --event-burst="100" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140871 2575 flags.go:64] FLAG: --event-qps="50" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140876 2575 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140879 2575 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140882 2575 flags.go:64] FLAG: --eviction-hard="" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140886 2575 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140889 2575 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140892 2575 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140895 2575 flags.go:64] FLAG: --eviction-soft="" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140898 2575 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140901 2575 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140905 2575 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140908 2575 flags.go:64] FLAG: --experimental-mounter-path="" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140911 2575 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 21 10:03:29.144538 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140914 2575 flags.go:64] FLAG: --fail-swap-on="true" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140917 2575 flags.go:64] FLAG: --feature-gates="" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140921 2575 flags.go:64] FLAG: --file-check-frequency="20s" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140924 2575 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140928 2575 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140931 2575 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140934 2575 flags.go:64] FLAG: --healthz-port="10248" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140937 2575 flags.go:64] FLAG: --help="false" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140941 2575 flags.go:64] FLAG: --hostname-override="ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140945 2575 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140951 2575 flags.go:64] FLAG: --http-check-frequency="20s" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140956 2575 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140962 2575 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140967 2575 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140973 2575 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140978 2575 flags.go:64] FLAG: --image-service-endpoint="" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140982 2575 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140985 2575 flags.go:64] FLAG: --kube-api-burst="100" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140989 2575 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140992 2575 flags.go:64] FLAG: --kube-api-qps="50" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140996 2575 flags.go:64] FLAG: --kube-reserved="" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.140999 2575 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141002 2575 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 21 10:03:29.145170 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141005 2575 flags.go:64] FLAG: --kubelet-cgroups="" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141008 2575 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141011 2575 flags.go:64] FLAG: --lock-file="" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141014 2575 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141017 2575 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141021 2575 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141026 2575 flags.go:64] FLAG: --log-json-split-stream="false" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141030 2575 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141033 2575 flags.go:64] FLAG: --log-text-split-stream="false" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141035 2575 flags.go:64] FLAG: --logging-format="text" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141038 2575 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141042 2575 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141045 2575 flags.go:64] FLAG: --manifest-url="" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141048 2575 flags.go:64] FLAG: --manifest-url-header="" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141052 2575 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141056 2575 flags.go:64] FLAG: --max-open-files="1000000" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141060 2575 flags.go:64] FLAG: --max-pods="110" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141063 2575 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141066 2575 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141069 2575 flags.go:64] FLAG: --memory-manager-policy="None" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141072 2575 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141075 2575 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141078 2575 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141081 2575 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141091 2575 flags.go:64] FLAG: --node-status-max-images="50" Apr 21 10:03:29.145735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141108 2575 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141113 2575 flags.go:64] FLAG: --oom-score-adj="-999" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141118 2575 flags.go:64] FLAG: --pod-cidr="" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141122 2575 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141129 2575 flags.go:64] FLAG: --pod-manifest-path="" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141132 2575 flags.go:64] FLAG: --pod-max-pids="-1" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141136 2575 flags.go:64] FLAG: --pods-per-core="0" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141139 2575 flags.go:64] FLAG: --port="10250" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141142 2575 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141145 2575 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-06698680f5cd8a571" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141148 2575 flags.go:64] FLAG: --qos-reserved="" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141151 2575 flags.go:64] FLAG: --read-only-port="10255" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141154 2575 flags.go:64] FLAG: --register-node="true" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141157 2575 flags.go:64] FLAG: --register-schedulable="true" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141161 2575 flags.go:64] FLAG: --register-with-taints="" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141165 2575 flags.go:64] FLAG: --registry-burst="10" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141168 2575 flags.go:64] FLAG: --registry-qps="5" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141171 2575 flags.go:64] FLAG: --reserved-cpus="" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141174 2575 flags.go:64] FLAG: --reserved-memory="" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141177 2575 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141185 2575 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141189 2575 flags.go:64] FLAG: --rotate-certificates="false" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141192 2575 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141195 2575 flags.go:64] FLAG: --runonce="false" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141198 2575 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 21 10:03:29.146357 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141201 2575 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141205 2575 flags.go:64] FLAG: --seccomp-default="false" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141208 2575 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141211 2575 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141214 2575 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141217 2575 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141221 2575 flags.go:64] FLAG: --storage-driver-password="root" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141224 2575 flags.go:64] FLAG: --storage-driver-secure="false" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141227 2575 flags.go:64] FLAG: --storage-driver-table="stats" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141230 2575 flags.go:64] FLAG: --storage-driver-user="root" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141233 2575 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141236 2575 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141239 2575 flags.go:64] FLAG: --system-cgroups="" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141242 2575 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141248 2575 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141251 2575 flags.go:64] FLAG: --tls-cert-file="" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141254 2575 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141258 2575 flags.go:64] FLAG: --tls-min-version="" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141262 2575 flags.go:64] FLAG: --tls-private-key-file="" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141265 2575 flags.go:64] FLAG: --topology-manager-policy="none" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141267 2575 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141273 2575 flags.go:64] FLAG: --topology-manager-scope="container" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141277 2575 flags.go:64] FLAG: --v="2" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141281 2575 flags.go:64] FLAG: --version="false" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141289 2575 flags.go:64] FLAG: --vmodule="" Apr 21 10:03:29.146990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141293 2575 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141297 2575 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141390 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141394 2575 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141397 2575 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141400 2575 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141403 2575 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141405 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141408 2575 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141410 2575 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141414 2575 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141418 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141421 2575 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141424 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141426 2575 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141432 2575 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141435 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141437 2575 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141440 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141442 2575 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:29.147618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141445 2575 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141447 2575 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141450 2575 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141452 2575 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141455 2575 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141457 2575 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141459 2575 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141462 2575 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141466 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141468 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141471 2575 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141473 2575 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141476 2575 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141479 2575 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141482 2575 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141485 2575 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141488 2575 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141491 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141494 2575 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:29.148102 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141496 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141499 2575 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141502 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141504 2575 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141507 2575 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141510 2575 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141512 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141514 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141517 2575 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141520 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141522 2575 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141525 2575 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141529 2575 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141532 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141535 2575 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141538 2575 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141541 2575 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141544 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141547 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141549 2575 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:29.148633 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141552 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141556 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141558 2575 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141561 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141563 2575 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141566 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141568 2575 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141572 2575 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141575 2575 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141578 2575 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141581 2575 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141583 2575 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141586 2575 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141589 2575 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141591 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141594 2575 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141596 2575 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141599 2575 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141602 2575 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141604 2575 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:29.149141 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141606 2575 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141609 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141612 2575 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141615 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141618 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141620 2575 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141623 2575 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141625 2575 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.141628 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.141636 2575 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.148134 2575 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.148255 2575 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148306 2575 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148311 2575 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148315 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148318 2575 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:29.149618 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148321 2575 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148324 2575 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148327 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148330 2575 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148332 2575 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148335 2575 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148337 2575 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148340 2575 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148343 2575 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148345 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148348 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148351 2575 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148353 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148356 2575 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148358 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148360 2575 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148363 2575 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148366 2575 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148369 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148372 2575 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:29.150025 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148375 2575 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148377 2575 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148380 2575 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148382 2575 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148385 2575 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148388 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148390 2575 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148395 2575 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148400 2575 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148403 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148405 2575 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148407 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148410 2575 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148413 2575 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148415 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148418 2575 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148420 2575 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148423 2575 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148426 2575 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148428 2575 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:29.150537 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148431 2575 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148434 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148436 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148439 2575 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148441 2575 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148444 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148446 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148449 2575 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148452 2575 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148454 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148457 2575 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148460 2575 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148462 2575 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148465 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148468 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148470 2575 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148473 2575 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148475 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148478 2575 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:29.151026 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148481 2575 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148484 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148486 2575 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148489 2575 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148491 2575 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148494 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148497 2575 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148499 2575 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148502 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148504 2575 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148507 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148509 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148512 2575 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148514 2575 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148517 2575 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148519 2575 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148522 2575 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148525 2575 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148527 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:29.151584 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148529 2575 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148532 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148536 2575 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148540 2575 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.148545 2575 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148643 2575 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148649 2575 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148652 2575 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148655 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148658 2575 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148661 2575 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148664 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148667 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148669 2575 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148672 2575 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 21 10:03:29.152075 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148674 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148677 2575 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148679 2575 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148682 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148684 2575 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148687 2575 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148689 2575 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148692 2575 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148695 2575 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148697 2575 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148700 2575 feature_gate.go:328] unrecognized feature gate: Example2 Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148702 2575 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148706 2575 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148710 2575 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148712 2575 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148715 2575 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148717 2575 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148720 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148722 2575 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148725 2575 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 21 10:03:29.152462 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148727 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148730 2575 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148733 2575 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148735 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148738 2575 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148741 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148744 2575 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148746 2575 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148749 2575 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148751 2575 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148754 2575 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148756 2575 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148759 2575 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148761 2575 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148780 2575 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148784 2575 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148787 2575 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148790 2575 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148793 2575 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 21 10:03:29.152946 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148796 2575 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148800 2575 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148815 2575 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148818 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148821 2575 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148824 2575 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148826 2575 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148829 2575 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148832 2575 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148834 2575 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148837 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148839 2575 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148842 2575 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148845 2575 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148847 2575 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148850 2575 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148853 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148855 2575 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148859 2575 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 21 10:03:29.153403 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148861 2575 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148864 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148866 2575 feature_gate.go:328] unrecognized feature gate: Example Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148869 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148871 2575 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148874 2575 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148876 2575 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148879 2575 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148881 2575 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148884 2575 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148886 2575 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148889 2575 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148892 2575 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148894 2575 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148897 2575 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148899 2575 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148902 2575 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 21 10:03:29.153860 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:29.148905 2575 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 21 10:03:29.154292 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.148909 2575 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 21 10:03:29.154292 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.149028 2575 server.go:962] "Client rotation is on, will bootstrap in background" Apr 21 10:03:29.154292 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.151144 2575 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 21 10:03:29.154292 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.151993 2575 server.go:1019] "Starting client certificate rotation" Apr 21 10:03:29.154292 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.152106 2575 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 21 10:03:29.154292 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.152945 2575 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 21 10:03:29.174994 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.174972 2575 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 21 10:03:29.177475 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.177447 2575 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 21 10:03:29.194164 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.194139 2575 log.go:25] "Validated CRI v1 runtime API" Apr 21 10:03:29.199365 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.199344 2575 log.go:25] "Validated CRI v1 image API" Apr 21 10:03:29.200632 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.200613 2575 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 21 10:03:29.204887 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.204866 2575 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 21 10:03:29.205705 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.205679 2575 fs.go:135] Filesystem UUIDs: map[4d9c8fdd-832f-4435-a7ea-0ba778760620:/dev/nvme0n1p3 7B77-95E7:/dev/nvme0n1p2 d4612c3b-1f3b-48bb-b49e-b6668bbccf86:/dev/nvme0n1p4] Apr 21 10:03:29.205755 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.205706 2575 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 21 10:03:29.212521 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.212406 2575 manager.go:217] Machine: {Timestamp:2026-04-21 10:03:29.210063374 +0000 UTC m=+0.370927721 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3099324 MemoryCapacity:33164496896 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec28670ebc3e85cfdef708cb9a2d6f24 SystemUUID:ec28670e-bc3e-85cf-def7-08cb9a2d6f24 BootID:9e2fd5be-2bd6-4130-b713-4bd87e2a0ac0 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16582246400 Type:vfs Inodes:4048400 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6632902656 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16582250496 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:c3:b5:2d:ce:0d Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:c3:b5:2d:ce:0d Speed:0 Mtu:9001} {Name:ovs-system MacAddress:ea:b7:02:a0:33:9c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33164496896 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 21 10:03:29.212521 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.212517 2575 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 21 10:03:29.212631 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.212607 2575 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 21 10:03:29.213744 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.213716 2575 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 21 10:03:29.213893 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.213746 2575 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-142-73.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 21 10:03:29.213940 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.213903 2575 topology_manager.go:138] "Creating topology manager with none policy" Apr 21 10:03:29.213940 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.213912 2575 container_manager_linux.go:306] "Creating device plugin manager" Apr 21 10:03:29.213940 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.213925 2575 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 21 10:03:29.214733 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.214722 2575 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 21 10:03:29.216136 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.216126 2575 state_mem.go:36] "Initialized new in-memory state store" Apr 21 10:03:29.216249 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.216240 2575 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 21 10:03:29.218530 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.218520 2575 kubelet.go:491] "Attempting to sync node with API server" Apr 21 10:03:29.218567 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.218534 2575 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 21 10:03:29.218567 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.218546 2575 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 21 10:03:29.218567 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.218556 2575 kubelet.go:397] "Adding apiserver pod source" Apr 21 10:03:29.218567 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.218563 2575 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 21 10:03:29.219558 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.219539 2575 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 21 10:03:29.219617 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.219565 2575 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 21 10:03:29.222174 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.222157 2575 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 21 10:03:29.223527 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.223514 2575 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 21 10:03:29.225063 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225051 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225067 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225073 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225079 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225085 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225091 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225112 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225118 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225125 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 21 10:03:29.225138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225131 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 21 10:03:29.225491 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225146 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 21 10:03:29.225491 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225155 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 21 10:03:29.225970 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225954 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 21 10:03:29.225970 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.225969 2575 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 21 10:03:29.229743 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.229725 2575 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 21 10:03:29.229836 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.229765 2575 server.go:1295] "Started kubelet" Apr 21 10:03:29.229887 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.229854 2575 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 21 10:03:29.229949 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.229906 2575 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 21 10:03:29.229982 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.229972 2575 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 21 10:03:29.230708 ip-10-0-142-73 systemd[1]: Started Kubernetes Kubelet. Apr 21 10:03:29.231297 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.231114 2575 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 21 10:03:29.233728 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.233708 2575 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-142-73.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 21 10:03:29.234584 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.234553 2575 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 21 10:03:29.234584 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.234565 2575 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-142-73.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 21 10:03:29.235872 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.235856 2575 server.go:317] "Adding debug handlers to kubelet server" Apr 21 10:03:29.239994 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.239974 2575 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 21 10:03:29.240144 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.239307 2575 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-142-73.ec2.internal.18a85716fc32b221 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-142-73.ec2.internal,UID:ip-10-0-142-73.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-10-0-142-73.ec2.internal,},FirstTimestamp:2026-04-21 10:03:29.229738529 +0000 UTC m=+0.390602870,LastTimestamp:2026-04-21 10:03:29.229738529 +0000 UTC m=+0.390602870,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-142-73.ec2.internal,}" Apr 21 10:03:29.240356 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.240338 2575 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 21 10:03:29.240671 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.240645 2575 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 21 10:03:29.241138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241118 2575 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 21 10:03:29.241138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241142 2575 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 21 10:03:29.241277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241119 2575 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 21 10:03:29.241277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241187 2575 factory.go:55] Registering systemd factory Apr 21 10:03:29.241277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241209 2575 factory.go:223] Registration of the systemd container factory successfully Apr 21 10:03:29.241277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241230 2575 reconstruct.go:97] "Volume reconstruction finished" Apr 21 10:03:29.241277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241238 2575 reconciler.go:26] "Reconciler: start to sync state" Apr 21 10:03:29.241487 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241423 2575 factory.go:153] Registering CRI-O factory Apr 21 10:03:29.241487 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241436 2575 factory.go:223] Registration of the crio container factory successfully Apr 21 10:03:29.241487 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241485 2575 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 21 10:03:29.241648 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241506 2575 factory.go:103] Registering Raw factory Apr 21 10:03:29.241648 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241517 2575 manager.go:1196] Started watching for new ooms in manager Apr 21 10:03:29.241773 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.241129 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:29.241921 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.241908 2575 manager.go:319] Starting recovery of all containers Apr 21 10:03:29.246375 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.246346 2575 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Apr 21 10:03:29.246483 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.246402 2575 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"ip-10-0-142-73.ec2.internal\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Apr 21 10:03:29.248692 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.248663 2575 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-lbfjd" Apr 21 10:03:29.251987 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.251971 2575 manager.go:324] Recovery completed Apr 21 10:03:29.253888 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.253870 2575 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-lbfjd" Apr 21 10:03:29.256361 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.256349 2575 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:29.259440 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.259427 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:29.259508 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.259454 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:29.259508 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.259464 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:29.260013 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.259998 2575 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 21 10:03:29.260073 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.260013 2575 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 21 10:03:29.260073 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.260030 2575 state_mem.go:36] "Initialized new in-memory state store" Apr 21 10:03:29.261780 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.261719 2575 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-142-73.ec2.internal.18a85716fdf7ec5c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-142-73.ec2.internal,UID:ip-10-0-142-73.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node ip-10-0-142-73.ec2.internal status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:ip-10-0-142-73.ec2.internal,},FirstTimestamp:2026-04-21 10:03:29.259441244 +0000 UTC m=+0.420305585,LastTimestamp:2026-04-21 10:03:29.259441244 +0000 UTC m=+0.420305585,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-142-73.ec2.internal,}" Apr 21 10:03:29.262477 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.262463 2575 policy_none.go:49] "None policy: Start" Apr 21 10:03:29.262528 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.262479 2575 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 21 10:03:29.262528 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.262490 2575 state_mem.go:35] "Initializing new in-memory state store" Apr 21 10:03:29.305692 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.305673 2575 manager.go:341] "Starting Device Plugin manager" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.305721 2575 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.305732 2575 server.go:85] "Starting device plugin registration server" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.305971 2575 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.305983 2575 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.306074 2575 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.306177 2575 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.306187 2575 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.306972 2575 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 21 10:03:29.326277 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.307018 2575 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:29.399532 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.399441 2575 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 21 10:03:29.400661 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.400638 2575 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 21 10:03:29.400778 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.400670 2575 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 21 10:03:29.400778 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.400691 2575 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 21 10:03:29.400778 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.400702 2575 kubelet.go:2451] "Starting kubelet main sync loop" Apr 21 10:03:29.400778 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.400744 2575 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 21 10:03:29.403639 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.403619 2575 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:29.406277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.406263 2575 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:29.407322 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.407308 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:29.407409 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.407336 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:29.407409 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.407350 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:29.407409 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.407371 2575 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.415212 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.415196 2575 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.415292 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.415218 2575 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-142-73.ec2.internal\": node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:29.436491 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.436463 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:29.502196 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.502164 2575 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal"] Apr 21 10:03:29.502311 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.502239 2575 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:29.503918 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.503902 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:29.504001 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.503932 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:29.504001 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.503943 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:29.505962 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.505948 2575 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:29.506126 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.506111 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.506174 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.506138 2575 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:29.507193 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.507172 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:29.507193 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.507186 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:29.507346 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.507204 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:29.507346 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.507212 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:29.507346 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.507219 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:29.507346 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.507222 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:29.509366 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.509351 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.509457 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.509377 2575 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 21 10:03:29.510000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.509986 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientMemory" Apr 21 10:03:29.510073 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.510013 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasNoDiskPressure" Apr 21 10:03:29.510073 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.510025 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeHasSufficientPID" Apr 21 10:03:29.534720 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.534694 2575 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-142-73.ec2.internal\" not found" node="ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.536903 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.536888 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:29.539371 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.539354 2575 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-142-73.ec2.internal\" not found" node="ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.543571 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.543555 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/dbfdb56ede086eca491c9eac6de95d77-config\") pod \"kube-apiserver-proxy-ip-10-0-142-73.ec2.internal\" (UID: \"dbfdb56ede086eca491c9eac6de95d77\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.543621 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.543580 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/069e66fead1a70015c7c87c70fe13edb-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal\" (UID: \"069e66fead1a70015c7c87c70fe13edb\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.543621 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.543599 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/069e66fead1a70015c7c87c70fe13edb-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal\" (UID: \"069e66fead1a70015c7c87c70fe13edb\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.637549 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.637508 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:29.643917 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.643891 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/dbfdb56ede086eca491c9eac6de95d77-config\") pod \"kube-apiserver-proxy-ip-10-0-142-73.ec2.internal\" (UID: \"dbfdb56ede086eca491c9eac6de95d77\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.643986 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.643927 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/069e66fead1a70015c7c87c70fe13edb-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal\" (UID: \"069e66fead1a70015c7c87c70fe13edb\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.643986 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.643945 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/069e66fead1a70015c7c87c70fe13edb-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal\" (UID: \"069e66fead1a70015c7c87c70fe13edb\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.644058 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.643998 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/dbfdb56ede086eca491c9eac6de95d77-config\") pod \"kube-apiserver-proxy-ip-10-0-142-73.ec2.internal\" (UID: \"dbfdb56ede086eca491c9eac6de95d77\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.644058 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.644052 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/069e66fead1a70015c7c87c70fe13edb-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal\" (UID: \"069e66fead1a70015c7c87c70fe13edb\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.644147 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.644084 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/069e66fead1a70015c7c87c70fe13edb-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal\" (UID: \"069e66fead1a70015c7c87c70fe13edb\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.738348 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.738277 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:29.836831 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.836800 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.839361 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.839342 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:29.841469 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:29.841454 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:29.939462 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:29.939403 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:30.039977 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:30.039894 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:30.093564 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.093537 2575 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:30.140283 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:30.140234 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:30.152250 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.152221 2575 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 21 10:03:30.152419 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.152403 2575 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 10:03:30.152484 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.152407 2575 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 21 10:03:30.240240 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.240212 2575 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 21 10:03:30.240666 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:30.240642 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:30.252054 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.252033 2575 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 21 10:03:30.256434 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.256410 2575 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-20 09:58:29 +0000 UTC" deadline="2027-12-31 06:58:18.483757572 +0000 UTC" Apr 21 10:03:30.256434 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.256434 2575 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="14852h54m48.22732651s" Apr 21 10:03:30.268668 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.268637 2575 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:30.272086 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.272068 2575 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-czv5s" Apr 21 10:03:30.279432 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.279413 2575 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-czv5s" Apr 21 10:03:30.328713 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:30.328679 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod069e66fead1a70015c7c87c70fe13edb.slice/crio-07ec29455f8464c4eb096d5d14e0f04807ebaddb645c3f41e11e6abc55407032 WatchSource:0}: Error finding container 07ec29455f8464c4eb096d5d14e0f04807ebaddb645c3f41e11e6abc55407032: Status 404 returned error can't find the container with id 07ec29455f8464c4eb096d5d14e0f04807ebaddb645c3f41e11e6abc55407032 Apr 21 10:03:30.333518 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.333503 2575 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 10:03:30.338713 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:30.338684 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbfdb56ede086eca491c9eac6de95d77.slice/crio-c0eec7e63ea78de28645dda5bbabef0b03ba76c79d3c53e31ecfc4d285acbadf WatchSource:0}: Error finding container c0eec7e63ea78de28645dda5bbabef0b03ba76c79d3c53e31ecfc4d285acbadf: Status 404 returned error can't find the container with id c0eec7e63ea78de28645dda5bbabef0b03ba76c79d3c53e31ecfc4d285acbadf Apr 21 10:03:30.341472 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:30.341452 2575 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-142-73.ec2.internal\" not found" Apr 21 10:03:30.403636 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.403582 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" event={"ID":"dbfdb56ede086eca491c9eac6de95d77","Type":"ContainerStarted","Data":"c0eec7e63ea78de28645dda5bbabef0b03ba76c79d3c53e31ecfc4d285acbadf"} Apr 21 10:03:30.404544 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.404523 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" event={"ID":"069e66fead1a70015c7c87c70fe13edb","Type":"ContainerStarted","Data":"07ec29455f8464c4eb096d5d14e0f04807ebaddb645c3f41e11e6abc55407032"} Apr 21 10:03:30.408268 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.408245 2575 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:30.441184 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.441163 2575 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" Apr 21 10:03:30.447543 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.447527 2575 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 21 10:03:30.448494 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.448483 2575 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" Apr 21 10:03:30.469029 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:30.469009 2575 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 21 10:03:31.220173 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.219935 2575 apiserver.go:52] "Watching apiserver" Apr 21 10:03:31.228951 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.228510 2575 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 21 10:03:31.230027 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.229459 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5bsvf","kube-system/konnectivity-agent-l7wqf","kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal","openshift-cluster-node-tuning-operator/tuned-k9p7d","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal","openshift-multus/multus-additional-cni-plugins-pj28j","openshift-multus/multus-zk4px","openshift-multus/network-metrics-daemon-n8tpq","openshift-network-operator/iptables-alerter-b5sx5","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782","openshift-image-registry/node-ca-d5g7x","openshift-network-diagnostics/network-check-target-cnk4k"] Apr 21 10:03:31.232776 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.232747 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.235119 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.235072 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:31.235718 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.235696 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 21 10:03:31.235805 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.235732 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 21 10:03:31.235805 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.235748 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-2hhlj\"" Apr 21 10:03:31.235905 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.235864 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 21 10:03:31.235955 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.235929 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 21 10:03:31.236005 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.235963 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 21 10:03:31.236134 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.236050 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 21 10:03:31.237370 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.237349 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-ffrvb\"" Apr 21 10:03:31.237628 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.237610 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 21 10:03:31.237956 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.237905 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 21 10:03:31.239677 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.239650 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.239813 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.239796 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.242196 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242175 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:03:31.242561 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242539 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-l622j\"" Apr 21 10:03:31.242635 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242612 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 21 10:03:31.242713 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242696 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 21 10:03:31.242713 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242709 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 21 10:03:31.242837 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242777 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 21 10:03:31.242903 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242859 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 21 10:03:31.242903 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242858 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 21 10:03:31.243011 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.242954 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-ttg7n\"" Apr 21 10:03:31.243491 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.243470 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.245611 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.245468 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 21 10:03:31.245611 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.245516 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-wz6ln\"" Apr 21 10:03:31.245963 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.245946 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:31.246049 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.246024 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:31.248606 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.248587 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.251119 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.251084 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 21 10:03:31.251212 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.251126 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.251402 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.251386 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-w5c75\"" Apr 21 10:03:31.251566 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.251541 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 21 10:03:31.251741 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.251719 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:03:31.252983 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.252962 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysctl-conf\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253083 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.252999 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-run-ovn-kubernetes\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.253083 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253024 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58nrd\" (UniqueName: \"kubernetes.io/projected/157148fd-caf0-4611-9379-850dc886dfd6-kube-api-access-58nrd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.253083 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253048 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-kubernetes\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253083 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253071 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-systemd\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253118 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-host\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253141 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-system-cni-dir\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.253284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253162 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-daemon-config\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.253284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253187 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-etc-kubernetes\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.253284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253209 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-run\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253233 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.253284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253256 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2nq8\" (UniqueName: \"kubernetes.io/projected/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-kube-api-access-m2nq8\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.253284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253280 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ee38a6d0-8717-4092-9b02-1541c50c1659-tmp\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253297 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysconfig\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253310 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-etc-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253336 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/157148fd-caf0-4611-9379-850dc886dfd6-ovn-node-metrics-cert\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253359 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-modprobe-d\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253385 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcb5g\" (UniqueName: \"kubernetes.io/projected/ee38a6d0-8717-4092-9b02-1541c50c1659-kube-api-access-dcb5g\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253401 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-os-release\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253420 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-cni-binary-copy\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253442 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-cni-bin\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253463 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-cnibin\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253495 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-os-release\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253517 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-multus-certs\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253540 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-env-overrides\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253553 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-sys\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.253568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253569 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-slash\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253583 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381-konnectivity-ca\") pod \"konnectivity-agent-l7wqf\" (UID: \"0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381\") " pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253602 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253618 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fnvd\" (UniqueName: \"kubernetes.io/projected/c7353367-eaa9-4d89-9444-14f3bfbfc166-kube-api-access-5fnvd\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253632 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-cni-multus\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253646 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-netns\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253660 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-run-netns\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253674 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-cni-netd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253704 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253720 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-tuned\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253739 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-cni-binary-copy\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253766 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-k8s-cni-cncf-io\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253787 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-systemd-units\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253810 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-systemd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253825 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-kubelet\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253840 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-var-lib-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254159 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253862 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253877 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-system-cni-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253894 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-cnibin\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253914 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-ovnkube-script-lib\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253936 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysctl-d\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253961 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-socket-dir-parent\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.253984 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-cni-bin\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254012 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-kubelet\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254036 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-hostroot\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254051 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-log-socket\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254071 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-node-log\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254086 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-ovnkube-config\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254126 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381-agent-certs\") pod \"konnectivity-agent-l7wqf\" (UID: \"0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381\") " pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254155 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-lib-modules\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254174 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-var-lib-kubelet\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254189 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-conf-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254234 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-ovn\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.254866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254262 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.255705 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254291 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-cni-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.255705 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.254772 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 21 10:03:31.255705 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.255037 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 21 10:03:31.255705 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.255133 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-z99xn\"" Apr 21 10:03:31.255705 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.255289 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 21 10:03:31.257323 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.257304 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:31.257419 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.257363 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:31.257581 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.257527 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.260921 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.260901 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 21 10:03:31.261025 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.260951 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-pgq9t\"" Apr 21 10:03:31.261082 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.261034 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 21 10:03:31.261740 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.261720 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 21 10:03:31.281195 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.281162 2575 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-20 09:58:30 +0000 UTC" deadline="2027-10-09 13:28:32.493836077 +0000 UTC" Apr 21 10:03:31.281306 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.281199 2575 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="12867h25m1.212640287s" Apr 21 10:03:31.342533 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.342506 2575 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 21 10:03:31.352795 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.352760 2575 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:31.354516 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354489 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-env-overrides\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.354646 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354525 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/3ef9e3d9-e961-4953-8328-d92b08011fc3-iptables-alerter-script\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.354646 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354549 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b776059e-ddd7-4893-92ef-c96817a1ae4f-serviceca\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.354646 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354575 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-sys\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.354646 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354619 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-kubelet-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.354795 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354653 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-slash\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.354795 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354688 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-slash\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.354795 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354706 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381-konnectivity-ca\") pod \"konnectivity-agent-l7wqf\" (UID: \"0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381\") " pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:31.354795 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354741 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-sys\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.354795 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354751 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.354795 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354785 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5fnvd\" (UniqueName: \"kubernetes.io/projected/c7353367-eaa9-4d89-9444-14f3bfbfc166-kube-api-access-5fnvd\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.355018 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354811 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-cni-multus\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.355018 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354842 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rzr5\" (UniqueName: \"kubernetes.io/projected/3ef9e3d9-e961-4953-8328-d92b08011fc3-kube-api-access-6rzr5\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.355018 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354869 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-netns\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.355018 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354909 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-run-netns\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355018 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354933 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-cni-netd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355018 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354962 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355018 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.354989 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-tuned\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.355018 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355016 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-cni-binary-copy\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355044 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-k8s-cni-cncf-io\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355068 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-systemd-units\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355090 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-systemd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355132 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd9h8\" (UniqueName: \"kubernetes.io/projected/94c5d86e-5a9a-4b0f-9998-b84c78940f55-kube-api-access-wd9h8\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355135 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-env-overrides\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355174 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-kubelet\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355202 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-cni-multus\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355215 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-var-lib-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355222 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355255 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-run-netns\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355268 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-netns\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355281 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-var-lib-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355298 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-cni-netd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355321 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381-konnectivity-ca\") pod \"konnectivity-agent-l7wqf\" (UID: \"0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381\") " pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355340 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-systemd-units\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355353 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-k8s-cni-cncf-io\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.355393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355388 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-systemd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355395 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355400 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-kubelet\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355418 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355448 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-system-cni-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355453 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355474 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-cnibin\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355498 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-ovnkube-script-lib\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355516 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-system-cni-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355530 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9tgz\" (UniqueName: \"kubernetes.io/projected/9276c6c8-dd13-4c6c-a3d3-3318a3731823-kube-api-access-p9tgz\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355566 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysctl-d\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355567 2575 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355596 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-socket-dir-parent\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355601 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-cnibin\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355621 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-cni-bin\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355646 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-kubelet\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355676 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-hostroot\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355701 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-device-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.356211 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355729 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-log-socket\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355742 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysctl-d\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355752 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-node-log\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355781 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-ovnkube-config\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355823 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-cni-binary-copy\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355834 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-kubelet\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355865 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-node-log\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355872 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-hostroot\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355897 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381-agent-certs\") pod \"konnectivity-agent-l7wqf\" (UID: \"0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381\") " pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355919 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-socket-dir-parent\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355940 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-var-lib-cni-bin\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.355994 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-lib-modules\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356028 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-var-lib-kubelet\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356053 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-conf-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356086 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-ovn\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356130 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356150 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-ovnkube-script-lib\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356158 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-cni-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356206 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-cni-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356246 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356270 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/157148fd-caf0-4611-9379-850dc886dfd6-ovnkube-config\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356272 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zswrq\" (UniqueName: \"kubernetes.io/projected/b776059e-ddd7-4893-92ef-c96817a1ae4f-kube-api-access-zswrq\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356311 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-var-lib-kubelet\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356327 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-log-socket\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356363 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-run-ovn\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356347 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-conf-dir\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356435 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysctl-conf\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356451 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-lib-modules\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356474 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-registration-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356571 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysctl-conf\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356601 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-sys-fs\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356687 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b776059e-ddd7-4893-92ef-c96817a1ae4f-host\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356737 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-run-ovn-kubernetes\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356769 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-58nrd\" (UniqueName: \"kubernetes.io/projected/157148fd-caf0-4611-9379-850dc886dfd6-kube-api-access-58nrd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356782 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.357824 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356792 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-kubernetes\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356823 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-systemd\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356836 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-kubernetes\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356846 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-run-ovn-kubernetes\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356848 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-host\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356888 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-host\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356896 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-system-cni-dir\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356918 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-daemon-config\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356926 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-systemd\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356938 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-etc-kubernetes\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356958 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-run\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356966 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-system-cni-dir\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.356980 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357020 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m2nq8\" (UniqueName: \"kubernetes.io/projected/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-kube-api-access-m2nq8\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357054 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357075 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-run\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357079 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-etc-selinux\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.358536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357010 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-etc-kubernetes\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357147 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ee38a6d0-8717-4092-9b02-1541c50c1659-tmp\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357193 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3ef9e3d9-e961-4953-8328-d92b08011fc3-host-slash\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357239 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysconfig\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357265 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-etc-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357289 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/157148fd-caf0-4611-9379-850dc886dfd6-ovn-node-metrics-cert\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357313 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-modprobe-d\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357339 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dcb5g\" (UniqueName: \"kubernetes.io/projected/ee38a6d0-8717-4092-9b02-1541c50c1659-kube-api-access-dcb5g\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357341 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357363 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-os-release\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357390 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-cni-binary-copy\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357372 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-sysconfig\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357417 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-socket-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357437 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-modprobe-d\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357443 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-cni-bin\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357464 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-os-release\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357479 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-cnibin\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.359325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357354 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-etc-openvswitch\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357505 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-os-release\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357525 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c7353367-eaa9-4d89-9444-14f3bfbfc166-cnibin\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357530 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-multus-certs\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357566 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-host-run-multus-certs\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357481 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/157148fd-caf0-4611-9379-850dc886dfd6-host-cni-bin\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.357660 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-os-release\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.358041 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c7353367-eaa9-4d89-9444-14f3bfbfc166-cni-binary-copy\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.358446 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-multus-daemon-config\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.359022 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/ee38a6d0-8717-4092-9b02-1541c50c1659-etc-tuned\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.359382 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381-agent-certs\") pod \"konnectivity-agent-l7wqf\" (UID: \"0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381\") " pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:31.360107 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.359494 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/ee38a6d0-8717-4092-9b02-1541c50c1659-tmp\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.360648 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.360180 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/157148fd-caf0-4611-9379-850dc886dfd6-ovn-node-metrics-cert\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.363276 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.362874 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fnvd\" (UniqueName: \"kubernetes.io/projected/c7353367-eaa9-4d89-9444-14f3bfbfc166-kube-api-access-5fnvd\") pod \"multus-additional-cni-plugins-pj28j\" (UID: \"c7353367-eaa9-4d89-9444-14f3bfbfc166\") " pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.364154 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.364131 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-58nrd\" (UniqueName: \"kubernetes.io/projected/157148fd-caf0-4611-9379-850dc886dfd6-kube-api-access-58nrd\") pod \"ovnkube-node-5bsvf\" (UID: \"157148fd-caf0-4611-9379-850dc886dfd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.364492 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.364470 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2nq8\" (UniqueName: \"kubernetes.io/projected/e945e18f-ffc9-4607-92c7-e5f597ffb6a0-kube-api-access-m2nq8\") pod \"multus-zk4px\" (UID: \"e945e18f-ffc9-4607-92c7-e5f597ffb6a0\") " pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.365155 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.365135 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcb5g\" (UniqueName: \"kubernetes.io/projected/ee38a6d0-8717-4092-9b02-1541c50c1659-kube-api-access-dcb5g\") pod \"tuned-k9p7d\" (UID: \"ee38a6d0-8717-4092-9b02-1541c50c1659\") " pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.457996 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.457956 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wd9h8\" (UniqueName: \"kubernetes.io/projected/94c5d86e-5a9a-4b0f-9998-b84c78940f55-kube-api-access-wd9h8\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.457996 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.457998 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p9tgz\" (UniqueName: \"kubernetes.io/projected/9276c6c8-dd13-4c6c-a3d3-3318a3731823-kube-api-access-p9tgz\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:31.458226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458018 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-device-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458043 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:31.458226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458061 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zswrq\" (UniqueName: \"kubernetes.io/projected/b776059e-ddd7-4893-92ef-c96817a1ae4f-kube-api-access-zswrq\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.458226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458077 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-registration-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458118 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-sys-fs\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458186 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-sys-fs\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458182 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b776059e-ddd7-4893-92ef-c96817a1ae4f-host\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.458226 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.458222 2575 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458238 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b776059e-ddd7-4893-92ef-c96817a1ae4f-host\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458233 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-registration-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458238 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458279 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-etc-selinux\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458304 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-device-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.458312 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs podName:9276c6c8-dd13-4c6c-a3d3-3318a3731823 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:31.958289262 +0000 UTC m=+3.119153610 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs") pod "network-metrics-daemon-n8tpq" (UID: "9276c6c8-dd13-4c6c-a3d3-3318a3731823") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458330 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-etc-selinux\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458350 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3ef9e3d9-e961-4953-8328-d92b08011fc3-host-slash\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458400 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-socket-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458366 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3ef9e3d9-e961-4953-8328-d92b08011fc3-host-slash\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458432 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/3ef9e3d9-e961-4953-8328-d92b08011fc3-iptables-alerter-script\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458459 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b776059e-ddd7-4893-92ef-c96817a1ae4f-serviceca\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458487 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-kubelet-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458515 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-socket-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458520 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6rzr5\" (UniqueName: \"kubernetes.io/projected/3ef9e3d9-e961-4953-8328-d92b08011fc3-kube-api-access-6rzr5\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.458584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458572 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/94c5d86e-5a9a-4b0f-9998-b84c78940f55-kubelet-dir\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.459114 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458877 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b776059e-ddd7-4893-92ef-c96817a1ae4f-serviceca\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.459114 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.458971 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/3ef9e3d9-e961-4953-8328-d92b08011fc3-iptables-alerter-script\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.466930 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.466898 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:31.466930 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.466925 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:31.466930 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.466939 2575 projected.go:194] Error preparing data for projected volume kube-api-access-cqpv5 for pod openshift-network-diagnostics/network-check-target-cnk4k: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:31.467243 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.467008 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5 podName:9db6586f-b886-4d23-84d5-48c25a2a6ac9 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:31.966989466 +0000 UTC m=+3.127853806 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqpv5" (UniqueName: "kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5") pod "network-check-target-cnk4k" (UID: "9db6586f-b886-4d23-84d5-48c25a2a6ac9") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:31.468728 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.468707 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zswrq\" (UniqueName: \"kubernetes.io/projected/b776059e-ddd7-4893-92ef-c96817a1ae4f-kube-api-access-zswrq\") pod \"node-ca-d5g7x\" (UID: \"b776059e-ddd7-4893-92ef-c96817a1ae4f\") " pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.469129 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.469089 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rzr5\" (UniqueName: \"kubernetes.io/projected/3ef9e3d9-e961-4953-8328-d92b08011fc3-kube-api-access-6rzr5\") pod \"iptables-alerter-b5sx5\" (UID: \"3ef9e3d9-e961-4953-8328-d92b08011fc3\") " pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.469226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.469089 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd9h8\" (UniqueName: \"kubernetes.io/projected/94c5d86e-5a9a-4b0f-9998-b84c78940f55-kube-api-access-wd9h8\") pod \"aws-ebs-csi-driver-node-5d782\" (UID: \"94c5d86e-5a9a-4b0f-9998-b84c78940f55\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.469365 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.469345 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9tgz\" (UniqueName: \"kubernetes.io/projected/9276c6c8-dd13-4c6c-a3d3-3318a3731823-kube-api-access-p9tgz\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:31.545743 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.545616 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:31.555465 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.555440 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:31.565285 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.565258 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" Apr 21 10:03:31.570897 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.570879 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pj28j" Apr 21 10:03:31.578514 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.578490 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zk4px" Apr 21 10:03:31.587079 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.587057 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-b5sx5" Apr 21 10:03:31.594651 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.594626 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" Apr 21 10:03:31.600183 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.600163 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-d5g7x" Apr 21 10:03:31.962513 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:31.962444 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:31.962643 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.962583 2575 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:31.962682 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:31.962645 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs podName:9276c6c8-dd13-4c6c-a3d3-3318a3731823 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:32.962626445 +0000 UTC m=+4.123490778 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs") pod "network-metrics-daemon-n8tpq" (UID: "9276c6c8-dd13-4c6c-a3d3-3318a3731823") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:31.984341 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:31.984307 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode945e18f_ffc9_4607_92c7_e5f597ffb6a0.slice/crio-865503bb63ba9d1098326515309dee899176ed0f9278a6761aa38b2524cf7629 WatchSource:0}: Error finding container 865503bb63ba9d1098326515309dee899176ed0f9278a6761aa38b2524cf7629: Status 404 returned error can't find the container with id 865503bb63ba9d1098326515309dee899176ed0f9278a6761aa38b2524cf7629 Apr 21 10:03:31.986354 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:31.986332 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee38a6d0_8717_4092_9b02_1541c50c1659.slice/crio-6247b45a9c265f505766d5b2fb29d2e579d0f534bb270aecbc3aa64721a32b48 WatchSource:0}: Error finding container 6247b45a9c265f505766d5b2fb29d2e579d0f534bb270aecbc3aa64721a32b48: Status 404 returned error can't find the container with id 6247b45a9c265f505766d5b2fb29d2e579d0f534bb270aecbc3aa64721a32b48 Apr 21 10:03:31.987125 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:31.987078 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ef9e3d9_e961_4953_8328_d92b08011fc3.slice/crio-97221be9a982db64bf73ff7eaae17cc0ef583fe443d55a444881a4e5c76b58d7 WatchSource:0}: Error finding container 97221be9a982db64bf73ff7eaae17cc0ef583fe443d55a444881a4e5c76b58d7: Status 404 returned error can't find the container with id 97221be9a982db64bf73ff7eaae17cc0ef583fe443d55a444881a4e5c76b58d7 Apr 21 10:03:31.987908 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:31.987821 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94c5d86e_5a9a_4b0f_9998_b84c78940f55.slice/crio-c3bb8081eab0a1a4316c6f08b21a3da640fe1f654f8389d90aadad24ecc376d7 WatchSource:0}: Error finding container c3bb8081eab0a1a4316c6f08b21a3da640fe1f654f8389d90aadad24ecc376d7: Status 404 returned error can't find the container with id c3bb8081eab0a1a4316c6f08b21a3da640fe1f654f8389d90aadad24ecc376d7 Apr 21 10:03:31.989706 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:31.989690 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7353367_eaa9_4d89_9444_14f3bfbfc166.slice/crio-8b6c15e913955ee307ae33d96c40a94f3bdbad75c6728015668d08123cbe9a8b WatchSource:0}: Error finding container 8b6c15e913955ee307ae33d96c40a94f3bdbad75c6728015668d08123cbe9a8b: Status 404 returned error can't find the container with id 8b6c15e913955ee307ae33d96c40a94f3bdbad75c6728015668d08123cbe9a8b Apr 21 10:03:32.011579 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:32.011526 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod157148fd_caf0_4611_9379_850dc886dfd6.slice/crio-03eeb4042d6b96c88280f3548939af23d659e59325427b97884dfcc3605bddf3 WatchSource:0}: Error finding container 03eeb4042d6b96c88280f3548939af23d659e59325427b97884dfcc3605bddf3: Status 404 returned error can't find the container with id 03eeb4042d6b96c88280f3548939af23d659e59325427b97884dfcc3605bddf3 Apr 21 10:03:32.012484 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:32.012460 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a0fa3aa_3e6a_4ae6_b76a_43b4bb05c381.slice/crio-b1fcb85ffe42f81236bf629d48f9d41f3dc2779c84d9ce8fd6a95883181f0ec6 WatchSource:0}: Error finding container b1fcb85ffe42f81236bf629d48f9d41f3dc2779c84d9ce8fd6a95883181f0ec6: Status 404 returned error can't find the container with id b1fcb85ffe42f81236bf629d48f9d41f3dc2779c84d9ce8fd6a95883181f0ec6 Apr 21 10:03:32.013015 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:32.012994 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb776059e_ddd7_4893_92ef_c96817a1ae4f.slice/crio-9d5a51f41bd0ce14e43c0ce04603551c2f357dea51fa7ddf07d11432e33555de WatchSource:0}: Error finding container 9d5a51f41bd0ce14e43c0ce04603551c2f357dea51fa7ddf07d11432e33555de: Status 404 returned error can't find the container with id 9d5a51f41bd0ce14e43c0ce04603551c2f357dea51fa7ddf07d11432e33555de Apr 21 10:03:32.014942 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.014800 2575 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 21 10:03:32.063024 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.062999 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:32.063184 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:32.063165 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:32.063251 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:32.063190 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:32.063251 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:32.063204 2575 projected.go:194] Error preparing data for projected volume kube-api-access-cqpv5 for pod openshift-network-diagnostics/network-check-target-cnk4k: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:32.063414 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:32.063271 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5 podName:9db6586f-b886-4d23-84d5-48c25a2a6ac9 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:33.063249158 +0000 UTC m=+4.224113500 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqpv5" (UniqueName: "kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5") pod "network-check-target-cnk4k" (UID: "9db6586f-b886-4d23-84d5-48c25a2a6ac9") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:32.282355 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.282266 2575 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-20 09:58:30 +0000 UTC" deadline="2027-12-07 23:42:33.667052433 +0000 UTC" Apr 21 10:03:32.282355 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.282301 2575 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="14293h39m1.384754248s" Apr 21 10:03:32.401598 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.401048 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:32.401598 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:32.401231 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:32.411115 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.411039 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zk4px" event={"ID":"e945e18f-ffc9-4607-92c7-e5f597ffb6a0","Type":"ContainerStarted","Data":"865503bb63ba9d1098326515309dee899176ed0f9278a6761aa38b2524cf7629"} Apr 21 10:03:32.419468 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.419361 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" event={"ID":"dbfdb56ede086eca491c9eac6de95d77","Type":"ContainerStarted","Data":"b48d90a367819061338216d045eea4a1cddbcb03e0749cf8d1768b76bf69f035"} Apr 21 10:03:32.423977 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.423953 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-d5g7x" event={"ID":"b776059e-ddd7-4893-92ef-c96817a1ae4f","Type":"ContainerStarted","Data":"9d5a51f41bd0ce14e43c0ce04603551c2f357dea51fa7ddf07d11432e33555de"} Apr 21 10:03:32.426921 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.426890 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-l7wqf" event={"ID":"0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381","Type":"ContainerStarted","Data":"b1fcb85ffe42f81236bf629d48f9d41f3dc2779c84d9ce8fd6a95883181f0ec6"} Apr 21 10:03:32.433684 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.433656 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" event={"ID":"94c5d86e-5a9a-4b0f-9998-b84c78940f55","Type":"ContainerStarted","Data":"c3bb8081eab0a1a4316c6f08b21a3da640fe1f654f8389d90aadad24ecc376d7"} Apr 21 10:03:32.436882 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.436855 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"03eeb4042d6b96c88280f3548939af23d659e59325427b97884dfcc3605bddf3"} Apr 21 10:03:32.439169 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.439147 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerStarted","Data":"8b6c15e913955ee307ae33d96c40a94f3bdbad75c6728015668d08123cbe9a8b"} Apr 21 10:03:32.444002 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.443977 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-b5sx5" event={"ID":"3ef9e3d9-e961-4953-8328-d92b08011fc3","Type":"ContainerStarted","Data":"97221be9a982db64bf73ff7eaae17cc0ef583fe443d55a444881a4e5c76b58d7"} Apr 21 10:03:32.446240 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.446218 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" event={"ID":"ee38a6d0-8717-4092-9b02-1541c50c1659","Type":"ContainerStarted","Data":"6247b45a9c265f505766d5b2fb29d2e579d0f534bb270aecbc3aa64721a32b48"} Apr 21 10:03:32.971862 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:32.971830 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:32.972029 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:32.972009 2575 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:32.972142 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:32.972083 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs podName:9276c6c8-dd13-4c6c-a3d3-3318a3731823 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:34.972062216 +0000 UTC m=+6.132926567 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs") pod "network-metrics-daemon-n8tpq" (UID: "9276c6c8-dd13-4c6c-a3d3-3318a3731823") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:33.072437 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:33.072255 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:33.072609 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:33.072444 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:33.072609 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:33.072472 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:33.072609 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:33.072485 2575 projected.go:194] Error preparing data for projected volume kube-api-access-cqpv5 for pod openshift-network-diagnostics/network-check-target-cnk4k: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:33.072609 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:33.072543 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5 podName:9db6586f-b886-4d23-84d5-48c25a2a6ac9 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:35.072524308 +0000 UTC m=+6.233388642 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqpv5" (UniqueName: "kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5") pod "network-check-target-cnk4k" (UID: "9db6586f-b886-4d23-84d5-48c25a2a6ac9") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:33.402340 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:33.402261 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:33.402785 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:33.402398 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:33.453461 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:33.453423 2575 generic.go:358] "Generic (PLEG): container finished" podID="069e66fead1a70015c7c87c70fe13edb" containerID="1d70b08262bc162ca5457f2e2d7d718b549b6eb4e848cd94f471a805c47fb1de" exitCode=0 Apr 21 10:03:33.454679 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:33.454590 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" event={"ID":"069e66fead1a70015c7c87c70fe13edb","Type":"ContainerDied","Data":"1d70b08262bc162ca5457f2e2d7d718b549b6eb4e848cd94f471a805c47fb1de"} Apr 21 10:03:33.471314 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:33.471254 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-142-73.ec2.internal" podStartSLOduration=3.471234404 podStartE2EDuration="3.471234404s" podCreationTimestamp="2026-04-21 10:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:03:32.433412998 +0000 UTC m=+3.594277349" watchObservedRunningTime="2026-04-21 10:03:33.471234404 +0000 UTC m=+4.632098757" Apr 21 10:03:34.401675 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:34.401640 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:34.401847 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:34.401790 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:34.474888 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:34.474794 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" event={"ID":"069e66fead1a70015c7c87c70fe13edb","Type":"ContainerStarted","Data":"9bca4ce53cc64bd110d681ecd3f8bdb11585786edc16e6d739cb982fea9c8159"} Apr 21 10:03:34.994807 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:34.994772 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:34.994993 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:34.994944 2575 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:34.995055 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:34.995006 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs podName:9276c6c8-dd13-4c6c-a3d3-3318a3731823 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:38.994987104 +0000 UTC m=+10.155851439 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs") pod "network-metrics-daemon-n8tpq" (UID: "9276c6c8-dd13-4c6c-a3d3-3318a3731823") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:35.095738 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:35.095603 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:35.095907 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:35.095744 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:35.095907 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:35.095763 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:35.095907 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:35.095775 2575 projected.go:194] Error preparing data for projected volume kube-api-access-cqpv5 for pod openshift-network-diagnostics/network-check-target-cnk4k: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:35.095907 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:35.095832 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5 podName:9db6586f-b886-4d23-84d5-48c25a2a6ac9 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:39.095812411 +0000 UTC m=+10.256676743 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqpv5" (UniqueName: "kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5") pod "network-check-target-cnk4k" (UID: "9db6586f-b886-4d23-84d5-48c25a2a6ac9") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:35.404597 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:35.404046 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:35.404597 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:35.404179 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:36.401621 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:36.401573 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:36.402006 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:36.401737 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:37.401314 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:37.401274 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:37.401487 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:37.401387 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:38.401147 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:38.401107 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:38.401586 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:38.401279 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:39.027592 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:39.027534 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:39.027745 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:39.027723 2575 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:39.027856 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:39.027822 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs podName:9276c6c8-dd13-4c6c-a3d3-3318a3731823 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:47.02779955 +0000 UTC m=+18.188663894 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs") pod "network-metrics-daemon-n8tpq" (UID: "9276c6c8-dd13-4c6c-a3d3-3318a3731823") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:39.128233 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:39.128195 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:39.128421 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:39.128373 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:39.128421 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:39.128394 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:39.128421 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:39.128404 2575 projected.go:194] Error preparing data for projected volume kube-api-access-cqpv5 for pod openshift-network-diagnostics/network-check-target-cnk4k: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:39.128551 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:39.128454 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5 podName:9db6586f-b886-4d23-84d5-48c25a2a6ac9 nodeName:}" failed. No retries permitted until 2026-04-21 10:03:47.128438978 +0000 UTC m=+18.289303306 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqpv5" (UniqueName: "kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5") pod "network-check-target-cnk4k" (UID: "9db6586f-b886-4d23-84d5-48c25a2a6ac9") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:39.402577 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:39.402489 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:39.403020 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:39.402607 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:40.401519 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:40.401486 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:40.401692 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:40.401608 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:41.401656 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:41.401466 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:41.402141 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:41.401756 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:42.401706 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:42.401668 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:42.402139 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:42.401816 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:43.401254 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:43.401219 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:43.401447 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:43.401348 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:44.401245 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:44.401161 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:44.401688 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:44.401298 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:45.401635 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:45.401598 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:45.402052 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:45.401726 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:46.401163 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:46.401129 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:46.401413 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:46.401268 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:47.085462 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:47.085422 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:47.085818 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:47.085573 2575 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:47.085818 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:47.085639 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs podName:9276c6c8-dd13-4c6c-a3d3-3318a3731823 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.085622962 +0000 UTC m=+34.246487289 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs") pod "network-metrics-daemon-n8tpq" (UID: "9276c6c8-dd13-4c6c-a3d3-3318a3731823") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:03:47.185775 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:47.185738 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:47.185961 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:47.185876 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 21 10:03:47.185961 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:47.185894 2575 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 21 10:03:47.185961 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:47.185903 2575 projected.go:194] Error preparing data for projected volume kube-api-access-cqpv5 for pod openshift-network-diagnostics/network-check-target-cnk4k: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:47.185961 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:47.185952 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5 podName:9db6586f-b886-4d23-84d5-48c25a2a6ac9 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.185938654 +0000 UTC m=+34.346802996 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqpv5" (UniqueName: "kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5") pod "network-check-target-cnk4k" (UID: "9db6586f-b886-4d23-84d5-48c25a2a6ac9") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 21 10:03:47.401126 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:47.401026 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:47.401275 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:47.401170 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:48.401077 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:48.401046 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:48.401551 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:48.401187 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:49.402329 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.402091 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:49.402913 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:49.402432 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:49.506068 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.506036 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"618700b1e2912c1d767c42562e6453ceb4777d35fc2bfc7e52d1aaaca072dfb3"} Apr 21 10:03:49.506189 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.506073 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"9bc4c969eef1a5c1b18fdcacd9ed4b4bffadb195d605ba624c22b8e661dd7645"} Apr 21 10:03:49.506189 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.506084 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"2c9b744ec46c612d10f4c31c0f9b50f069b1f0c8f57dfaf25d0e7cc6859a9275"} Apr 21 10:03:49.506189 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.506110 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"2ec9f29a800d64aed666fa45c00e3b38713f56c44ebf19400057cc6a75f01097"} Apr 21 10:03:49.507399 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.507372 2575 generic.go:358] "Generic (PLEG): container finished" podID="c7353367-eaa9-4d89-9444-14f3bfbfc166" containerID="041fc35b0a1986ce52e85f2b208b831212a11d9594ce882b64cba72500bf375d" exitCode=0 Apr 21 10:03:49.507486 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.507443 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerDied","Data":"041fc35b0a1986ce52e85f2b208b831212a11d9594ce882b64cba72500bf375d"} Apr 21 10:03:49.508784 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.508708 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" event={"ID":"ee38a6d0-8717-4092-9b02-1541c50c1659","Type":"ContainerStarted","Data":"d8f5738d964baa9630342ca07de6206f0ce6966d588f410d1a1af93c740a52e9"} Apr 21 10:03:49.510233 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.510138 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zk4px" event={"ID":"e945e18f-ffc9-4607-92c7-e5f597ffb6a0","Type":"ContainerStarted","Data":"044701b7c862399b768208afa9ca37de9c79422505ad6f361797f2f4e17587db"} Apr 21 10:03:49.512623 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.512599 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-d5g7x" event={"ID":"b776059e-ddd7-4893-92ef-c96817a1ae4f","Type":"ContainerStarted","Data":"12205924b88d4749604d562b619a199e047cc3d29dc2f0f72061f79ec2b99371"} Apr 21 10:03:49.514188 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.514167 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-l7wqf" event={"ID":"0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381","Type":"ContainerStarted","Data":"62e426be120c959aa53c56e1360e37c92436bea3614ab8852ed853d0a903dc10"} Apr 21 10:03:49.515504 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.515481 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" event={"ID":"94c5d86e-5a9a-4b0f-9998-b84c78940f55","Type":"ContainerStarted","Data":"79617d9f4539ec7ac1a4d5096a88769d97034640394be756a5821fc3866f2be9"} Apr 21 10:03:49.524887 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.524849 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-142-73.ec2.internal" podStartSLOduration=19.524838389 podStartE2EDuration="19.524838389s" podCreationTimestamp="2026-04-21 10:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:03:34.488434453 +0000 UTC m=+5.649298804" watchObservedRunningTime="2026-04-21 10:03:49.524838389 +0000 UTC m=+20.685702740" Apr 21 10:03:49.536572 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.536535 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-l7wqf" podStartSLOduration=3.667156731 podStartE2EDuration="20.536523882s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:03:32.015423682 +0000 UTC m=+3.176288011" lastFinishedPulling="2026-04-21 10:03:48.884790819 +0000 UTC m=+20.045655162" observedRunningTime="2026-04-21 10:03:49.535734807 +0000 UTC m=+20.696599158" watchObservedRunningTime="2026-04-21 10:03:49.536523882 +0000 UTC m=+20.697388223" Apr 21 10:03:49.550444 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.550392 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-zk4px" podStartSLOduration=3.6224768689999998 podStartE2EDuration="20.550378146s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:03:31.986356213 +0000 UTC m=+3.147220543" lastFinishedPulling="2026-04-21 10:03:48.914257492 +0000 UTC m=+20.075121820" observedRunningTime="2026-04-21 10:03:49.550347019 +0000 UTC m=+20.711211369" watchObservedRunningTime="2026-04-21 10:03:49.550378146 +0000 UTC m=+20.711242499" Apr 21 10:03:49.576065 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.576021 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-k9p7d" podStartSLOduration=3.659665351 podStartE2EDuration="20.576004381s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:03:31.988679684 +0000 UTC m=+3.149544026" lastFinishedPulling="2026-04-21 10:03:48.905018709 +0000 UTC m=+20.065883056" observedRunningTime="2026-04-21 10:03:49.563613344 +0000 UTC m=+20.724477694" watchObservedRunningTime="2026-04-21 10:03:49.576004381 +0000 UTC m=+20.736868732" Apr 21 10:03:49.576536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:49.576508 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-d5g7x" podStartSLOduration=8.489944174 podStartE2EDuration="20.576500401s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:03:32.015481727 +0000 UTC m=+3.176346056" lastFinishedPulling="2026-04-21 10:03:44.102037955 +0000 UTC m=+15.262902283" observedRunningTime="2026-04-21 10:03:49.57615022 +0000 UTC m=+20.737014572" watchObservedRunningTime="2026-04-21 10:03:49.576500401 +0000 UTC m=+20.737364751" Apr 21 10:03:50.037726 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.037518 2575 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 21 10:03:50.318511 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.318406 2575 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-21T10:03:50.037722477Z","UUID":"494a6daa-716c-4458-82a1-e589c82c9903","Handler":null,"Name":"","Endpoint":""} Apr 21 10:03:50.320192 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.320154 2575 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 21 10:03:50.320192 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.320187 2575 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 21 10:03:50.400895 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.400866 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:50.401008 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:50.400989 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:50.520340 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.520300 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" event={"ID":"94c5d86e-5a9a-4b0f-9998-b84c78940f55","Type":"ContainerStarted","Data":"4991f99e31fd4540b30c07ce8c153c1628b6d22556e5aa8bc7b274eca81508b4"} Apr 21 10:03:50.523790 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.523750 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"64a28993798d65cb30fd388edb7d154695605617e030875ca6f348886b8d6e2a"} Apr 21 10:03:50.523790 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.523791 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"02d15c278ed29c581acbd653e635928dd01eeeb839187a91ae2e4ddde9ecd952"} Apr 21 10:03:50.525338 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.525295 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-b5sx5" event={"ID":"3ef9e3d9-e961-4953-8328-d92b08011fc3","Type":"ContainerStarted","Data":"c2d637cc2be41ae569ca3b50cc64b927943bdd7051cb6d7e33f2b0fe7ec61f74"} Apr 21 10:03:50.538017 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:50.537950 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-b5sx5" podStartSLOduration=4.62363915 podStartE2EDuration="21.537931729s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:03:31.989039568 +0000 UTC m=+3.149903908" lastFinishedPulling="2026-04-21 10:03:48.903332146 +0000 UTC m=+20.064196487" observedRunningTime="2026-04-21 10:03:50.537573297 +0000 UTC m=+21.698437648" watchObservedRunningTime="2026-04-21 10:03:50.537931729 +0000 UTC m=+21.698796080" Apr 21 10:03:51.401037 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:51.401002 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:51.401338 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:51.401138 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:51.529439 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:51.529400 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" event={"ID":"94c5d86e-5a9a-4b0f-9998-b84c78940f55","Type":"ContainerStarted","Data":"e34ae5322d4c6bd45f7d19da26411724cc9c65bc705e232a0553e6ee8704b5f1"} Apr 21 10:03:51.562144 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:51.562074 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-5d782" podStartSLOduration=3.7588361949999998 podStartE2EDuration="22.562059116s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:03:32.010338199 +0000 UTC m=+3.171202527" lastFinishedPulling="2026-04-21 10:03:50.81356112 +0000 UTC m=+21.974425448" observedRunningTime="2026-04-21 10:03:51.561823853 +0000 UTC m=+22.722688301" watchObservedRunningTime="2026-04-21 10:03:51.562059116 +0000 UTC m=+22.722923467" Apr 21 10:03:51.924449 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:51.924371 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-krcxr"] Apr 21 10:03:51.928440 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:51.928416 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:51.930902 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:51.930878 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-855nf\"" Apr 21 10:03:51.931323 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:51.931298 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 21 10:03:51.931436 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:51.931298 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 21 10:03:52.019187 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.019155 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-tmp-dir\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.019360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.019212 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-hosts-file\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.019360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.019241 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr7rz\" (UniqueName: \"kubernetes.io/projected/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-kube-api-access-hr7rz\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.120393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.120352 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-tmp-dir\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.120570 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.120409 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-hosts-file\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.120570 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.120440 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hr7rz\" (UniqueName: \"kubernetes.io/projected/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-kube-api-access-hr7rz\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.120570 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.120507 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-hosts-file\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.121371 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.121343 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-tmp-dir\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.130504 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.130478 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr7rz\" (UniqueName: \"kubernetes.io/projected/e6f11c13-e164-47be-a16a-c5ea7b32f9e4-kube-api-access-hr7rz\") pod \"node-resolver-krcxr\" (UID: \"e6f11c13-e164-47be-a16a-c5ea7b32f9e4\") " pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.238316 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.238238 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-krcxr" Apr 21 10:03:52.401550 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.401508 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:52.401725 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:52.401634 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:52.474277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.474240 2575 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:52.474994 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.474970 2575 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:52.535109 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.534997 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"343a7baebaa93d1cbe1117ef7352f01e8529824ca1c24e2cad599d0b9214c083"} Apr 21 10:03:52.535794 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.535306 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:52.535794 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.535717 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-l7wqf" Apr 21 10:03:52.553474 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.553447 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-t8wp8"] Apr 21 10:03:52.555960 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.555941 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:52.556068 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:52.556015 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-t8wp8" podUID="37ae773e-f021-4a13-81c5-678bcf84335e" Apr 21 10:03:52.624456 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.624414 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/37ae773e-f021-4a13-81c5-678bcf84335e-dbus\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:52.624642 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.624522 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:52.624642 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.624575 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/37ae773e-f021-4a13-81c5-678bcf84335e-kubelet-config\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:52.725109 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.725057 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:52.725285 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.725135 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/37ae773e-f021-4a13-81c5-678bcf84335e-kubelet-config\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:52.725285 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.725185 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/37ae773e-f021-4a13-81c5-678bcf84335e-dbus\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:52.725285 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:52.725197 2575 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 10:03:52.725285 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.725222 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/37ae773e-f021-4a13-81c5-678bcf84335e-kubelet-config\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:52.725285 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:52.725263 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret podName:37ae773e-f021-4a13-81c5-678bcf84335e nodeName:}" failed. No retries permitted until 2026-04-21 10:03:53.225246145 +0000 UTC m=+24.386110491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret") pod "global-pull-secret-syncer-t8wp8" (UID: "37ae773e-f021-4a13-81c5-678bcf84335e") : object "kube-system"/"original-pull-secret" not registered Apr 21 10:03:52.725555 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:52.725427 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/37ae773e-f021-4a13-81c5-678bcf84335e-dbus\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:53.228542 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:53.228502 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:53.228717 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:53.228650 2575 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 10:03:53.228773 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:53.228717 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret podName:37ae773e-f021-4a13-81c5-678bcf84335e nodeName:}" failed. No retries permitted until 2026-04-21 10:03:54.228702531 +0000 UTC m=+25.389566860 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret") pod "global-pull-secret-syncer-t8wp8" (UID: "37ae773e-f021-4a13-81c5-678bcf84335e") : object "kube-system"/"original-pull-secret" not registered Apr 21 10:03:53.401586 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:53.401554 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:53.401758 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:53.401664 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:53.577928 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:03:53.577896 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6f11c13_e164_47be_a16a_c5ea7b32f9e4.slice/crio-016111d966126264af5baa868806e9f1921d8873b476621f2686ba03b812a8ad WatchSource:0}: Error finding container 016111d966126264af5baa868806e9f1921d8873b476621f2686ba03b812a8ad: Status 404 returned error can't find the container with id 016111d966126264af5baa868806e9f1921d8873b476621f2686ba03b812a8ad Apr 21 10:03:54.237929 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.237687 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:54.237929 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:54.237820 2575 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 10:03:54.237929 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:54.237930 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret podName:37ae773e-f021-4a13-81c5-678bcf84335e nodeName:}" failed. No retries permitted until 2026-04-21 10:03:56.237915768 +0000 UTC m=+27.398780097 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret") pod "global-pull-secret-syncer-t8wp8" (UID: "37ae773e-f021-4a13-81c5-678bcf84335e") : object "kube-system"/"original-pull-secret" not registered Apr 21 10:03:54.401267 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.401234 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:54.401402 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.401234 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:54.401402 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:54.401347 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:54.401485 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:54.401412 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-t8wp8" podUID="37ae773e-f021-4a13-81c5-678bcf84335e" Apr 21 10:03:54.538819 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.538747 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-krcxr" event={"ID":"e6f11c13-e164-47be-a16a-c5ea7b32f9e4","Type":"ContainerStarted","Data":"b8ec78d5fcc2ca707f87aa386f7308caf46b5212de29c1a01e1d0d55d0744e22"} Apr 21 10:03:54.538819 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.538785 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-krcxr" event={"ID":"e6f11c13-e164-47be-a16a-c5ea7b32f9e4","Type":"ContainerStarted","Data":"016111d966126264af5baa868806e9f1921d8873b476621f2686ba03b812a8ad"} Apr 21 10:03:54.541717 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.541682 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" event={"ID":"157148fd-caf0-4611-9379-850dc886dfd6","Type":"ContainerStarted","Data":"63361ad7aaaa7633d83375a37a265401c4f4abf0a86d6010968f61318fb2757a"} Apr 21 10:03:54.541966 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.541933 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:54.541966 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.541959 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:54.542143 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.541973 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:54.543480 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.543453 2575 generic.go:358] "Generic (PLEG): container finished" podID="c7353367-eaa9-4d89-9444-14f3bfbfc166" containerID="d2b5582d95fb9fde5332c0fcf78e2447b05a1493558cf335cd7e63121e3f60e9" exitCode=0 Apr 21 10:03:54.543582 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.543537 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerDied","Data":"d2b5582d95fb9fde5332c0fcf78e2447b05a1493558cf335cd7e63121e3f60e9"} Apr 21 10:03:54.556829 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.556807 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:54.557597 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.557580 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:03:54.559247 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.559212 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-krcxr" podStartSLOduration=3.559200697 podStartE2EDuration="3.559200697s" podCreationTimestamp="2026-04-21 10:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:03:54.558823381 +0000 UTC m=+25.719687733" watchObservedRunningTime="2026-04-21 10:03:54.559200697 +0000 UTC m=+25.720065048" Apr 21 10:03:54.622935 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:54.622884 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" podStartSLOduration=8.406905882 podStartE2EDuration="25.622869333s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:03:32.015516926 +0000 UTC m=+3.176381262" lastFinishedPulling="2026-04-21 10:03:49.231480372 +0000 UTC m=+20.392344713" observedRunningTime="2026-04-21 10:03:54.622540615 +0000 UTC m=+25.783405001" watchObservedRunningTime="2026-04-21 10:03:54.622869333 +0000 UTC m=+25.783733676" Apr 21 10:03:55.401817 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.401786 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:55.401973 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:55.401883 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:55.546691 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.546611 2575 generic.go:358] "Generic (PLEG): container finished" podID="c7353367-eaa9-4d89-9444-14f3bfbfc166" containerID="4cc6f39faa3591b76322ef76af29b22a21aba7de3332e21c6f52fa8ffeff117a" exitCode=0 Apr 21 10:03:55.546828 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.546704 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerDied","Data":"4cc6f39faa3591b76322ef76af29b22a21aba7de3332e21c6f52fa8ffeff117a"} Apr 21 10:03:55.707533 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.707504 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-t8wp8"] Apr 21 10:03:55.708160 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.707623 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:55.708160 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:55.707704 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-t8wp8" podUID="37ae773e-f021-4a13-81c5-678bcf84335e" Apr 21 10:03:55.712072 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.712046 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-cnk4k"] Apr 21 10:03:55.712236 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.712153 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:55.712292 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:55.712249 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:55.712712 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.712688 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n8tpq"] Apr 21 10:03:55.712823 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:55.712809 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:55.712953 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:55.712921 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:56.256562 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:56.256536 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:56.256739 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:56.256676 2575 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 10:03:56.256781 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:56.256741 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret podName:37ae773e-f021-4a13-81c5-678bcf84335e nodeName:}" failed. No retries permitted until 2026-04-21 10:04:00.256725769 +0000 UTC m=+31.417590097 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret") pod "global-pull-secret-syncer-t8wp8" (UID: "37ae773e-f021-4a13-81c5-678bcf84335e") : object "kube-system"/"original-pull-secret" not registered Apr 21 10:03:57.401136 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:57.401084 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:57.401567 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:57.401084 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:57.401567 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:57.401212 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-t8wp8" podUID="37ae773e-f021-4a13-81c5-678bcf84335e" Apr 21 10:03:57.401567 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:57.401084 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:57.401567 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:57.401302 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:03:57.401567 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:57.401387 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:57.552611 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:57.552570 2575 generic.go:358] "Generic (PLEG): container finished" podID="c7353367-eaa9-4d89-9444-14f3bfbfc166" containerID="ce4a12ac5891c9481f34f418acb9f65cf289999d91e475ee754be9f9d4ce3a4b" exitCode=0 Apr 21 10:03:57.552765 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:57.552631 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerDied","Data":"ce4a12ac5891c9481f34f418acb9f65cf289999d91e475ee754be9f9d4ce3a4b"} Apr 21 10:03:59.401812 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:59.401785 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:03:59.402774 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:59.401872 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:03:59.402774 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:59.401908 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-t8wp8" podUID="37ae773e-f021-4a13-81c5-678bcf84335e" Apr 21 10:03:59.402774 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:59.401941 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:03:59.402774 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:03:59.401972 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:03:59.402774 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:03:59.402059 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:04:00.283674 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:00.283632 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:04:00.283852 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:00.283817 2575 secret.go:189] Couldn't get secret kube-system/original-pull-secret: object "kube-system"/"original-pull-secret" not registered Apr 21 10:04:00.283920 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:00.283903 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret podName:37ae773e-f021-4a13-81c5-678bcf84335e nodeName:}" failed. No retries permitted until 2026-04-21 10:04:08.28388027 +0000 UTC m=+39.444744617 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "original-pull-secret" (UniqueName: "kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret") pod "global-pull-secret-syncer-t8wp8" (UID: "37ae773e-f021-4a13-81c5-678bcf84335e") : object "kube-system"/"original-pull-secret" not registered Apr 21 10:04:01.401536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:01.401322 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:04:01.402005 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:01.401322 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:04:01.402005 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:01.401641 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cnk4k" podUID="9db6586f-b886-4d23-84d5-48c25a2a6ac9" Apr 21 10:04:01.402005 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:01.401349 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:04:01.402005 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:01.401717 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="kube-system/global-pull-secret-syncer-t8wp8" podUID="37ae773e-f021-4a13-81c5-678bcf84335e" Apr 21 10:04:01.402005 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:01.401805 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n8tpq" podUID="9276c6c8-dd13-4c6c-a3d3-3318a3731823" Apr 21 10:04:02.132921 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.132873 2575 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-142-73.ec2.internal" event="NodeReady" Apr 21 10:04:02.133113 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.133035 2575 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 21 10:04:02.164935 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.164901 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-6659b7df7b-wn6k4"] Apr 21 10:04:02.198535 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.198505 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2"] Apr 21 10:04:02.198743 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.198657 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.202474 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.202447 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-qvkrb\"" Apr 21 10:04:02.202773 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.202753 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 21 10:04:02.202854 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.202810 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 21 10:04:02.203284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.203265 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 21 10:04:02.213040 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.212284 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-operator-585dfdc468-4hz8p"] Apr 21 10:04:02.213040 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.212426 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2" Apr 21 10:04:02.213489 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.213462 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 21 10:04:02.216218 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.216148 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-storage-operator\"/\"volume-data-source-validator-dockercfg-g749d\"" Apr 21 10:04:02.216812 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.216790 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-storage-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.216907 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.216831 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-storage-operator\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.231650 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.231180 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8"] Apr 21 10:04:02.232074 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.231597 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.236644 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.236526 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"operator-dockercfg-g7zxg\"" Apr 21 10:04:02.236769 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.236738 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"openshift-insights-serving-cert\"" Apr 21 10:04:02.236827 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.236799 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.236880 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.236868 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"service-ca-bundle\"" Apr 21 10:04:02.237029 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.237008 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.241675 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.241622 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"trusted-ca-bundle\"" Apr 21 10:04:02.247490 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.247461 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl"] Apr 21 10:04:02.261673 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.261646 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2"] Apr 21 10:04:02.261673 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.261679 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-6659b7df7b-wn6k4"] Apr 21 10:04:02.261871 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.261692 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl"] Apr 21 10:04:02.261871 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.261707 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-r495v"] Apr 21 10:04:02.261871 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.261775 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.261871 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.261858 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.264580 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.264560 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.264833 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.264816 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.265587 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.265401 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Apr 21 10:04:02.265587 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.265419 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-vxvfm\"" Apr 21 10:04:02.265587 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.265431 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-dockercfg-gdf9d\"" Apr 21 10:04:02.265587 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.265461 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-tls\"" Apr 21 10:04:02.265587 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.265516 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemetry-config\"" Apr 21 10:04:02.265587 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.265422 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.265943 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.265686 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.265943 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.265691 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Apr 21 10:04:02.280188 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.280166 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-mrf6q"] Apr 21 10:04:02.280314 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.280297 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.282896 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.282877 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 21 10:04:02.283006 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.282941 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-nglr4\"" Apr 21 10:04:02.283006 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.282883 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 21 10:04:02.295111 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.295077 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q"] Apr 21 10:04:02.295247 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.295229 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.297742 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.297604 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Apr 21 10:04:02.297742 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.297644 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-qtbfw\"" Apr 21 10:04:02.297944 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.297928 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Apr 21 10:04:02.298016 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.297984 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.298112 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298079 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.298284 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298268 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-installation-pull-secrets\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.298345 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298291 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7a64cde-eca1-4b8d-843b-90f038244fe0-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.298345 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298312 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/f7a64cde-eca1-4b8d-843b-90f038244fe0-snapshots\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.298345 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298340 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.298490 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298362 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f88tn\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-kube-api-access-f88tn\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.298490 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298426 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54tb9\" (UniqueName: \"kubernetes.io/projected/7e56e8b5-c500-424b-9f59-b08a49f14cce-kube-api-access-54tb9\") pod \"volume-data-source-validator-7c6cbb6c87-6nlj2\" (UID: \"7e56e8b5-c500-424b-9f59-b08a49f14cce\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2" Apr 21 10:04:02.298593 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298501 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/540077c6-a9f4-4a48-9f62-58c5f7d5c095-ca-trust-extracted\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.298593 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298543 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7a64cde-eca1-4b8d-843b-90f038244fe0-service-ca-bundle\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.298683 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298594 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-trusted-ca\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.298683 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298619 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f7a64cde-eca1-4b8d-843b-90f038244fe0-tmp\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.298683 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298643 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a64cde-eca1-4b8d-843b-90f038244fe0-serving-cert\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.298825 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298691 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vspsp\" (UniqueName: \"kubernetes.io/projected/f7a64cde-eca1-4b8d-843b-90f038244fe0-kube-api-access-vspsp\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.298825 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298753 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-certificates\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.298920 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298831 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-image-registry-private-configuration\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.298920 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.298860 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-bound-sa-token\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.303043 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.303022 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Apr 21 10:04:02.312858 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.312835 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-56b8857454-d4h69"] Apr 21 10:04:02.313022 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.312993 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:02.315274 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.315242 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Apr 21 10:04:02.315460 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.315423 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"default-dockercfg-xdcvf\"" Apr 21 10:04:02.315719 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.315697 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Apr 21 10:04:02.330889 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.330862 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct"] Apr 21 10:04:02.331042 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.331026 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.333998 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.333982 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Apr 21 10:04:02.334471 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.334394 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"default-ingress-cert\"" Apr 21 10:04:02.334864 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.334646 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.334864 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.334659 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Apr 21 10:04:02.335000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.334916 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.335000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.334940 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Apr 21 10:04:02.335000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.334984 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-5tz9l\"" Apr 21 10:04:02.352589 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.352567 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b"] Apr 21 10:04:02.352747 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.352729 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.355248 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.355131 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Apr 21 10:04:02.355248 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.355139 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-fb27b\"" Apr 21 10:04:02.355248 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.355210 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.355475 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.355457 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Apr 21 10:04:02.355565 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.355533 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.366972 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.366943 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b5lnk"] Apr 21 10:04:02.367604 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.367178 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:02.371800 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.371566 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.371800 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.371602 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Apr 21 10:04:02.371800 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.371611 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-hvnc6\"" Apr 21 10:04:02.372432 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.372256 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.386682 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.386620 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn"] Apr 21 10:04:02.386808 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.386777 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:02.389477 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.389456 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 21 10:04:02.389593 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.389526 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-sjbvv\"" Apr 21 10:04:02.389593 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.389570 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.389757 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.389464 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.399904 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.399879 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:02.400015 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.399920 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-bound-sa-token\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.400015 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.399946 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/2b707e63-9bbb-4244-adcd-2119a1c7187b-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.400015 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.399975 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.400204 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400031 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/f7a64cde-eca1-4b8d-843b-90f038244fe0-snapshots\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.400204 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400060 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f88tn\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-kube-api-access-f88tn\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.400204 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400086 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwtbc\" (UniqueName: \"kubernetes.io/projected/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-kube-api-access-rwtbc\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.400204 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400129 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-262hv\" (UniqueName: \"kubernetes.io/projected/cc037d14-2ee7-421f-875a-4ec33fac21bb-kube-api-access-262hv\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.400204 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400159 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-54tb9\" (UniqueName: \"kubernetes.io/projected/7e56e8b5-c500-424b-9f59-b08a49f14cce-kube-api-access-54tb9\") pod \"volume-data-source-validator-7c6cbb6c87-6nlj2\" (UID: \"7e56e8b5-c500-424b-9f59-b08a49f14cce\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2" Apr 21 10:04:02.400204 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400185 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-serving-cert\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.400479 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400209 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccb482af-112d-452e-a16f-49fe93eedf37-config-volume\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.400479 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400294 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.400479 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400345 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-certificates\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.400479 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400374 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-default-certificate\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.400479 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400407 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc567\" (UniqueName: \"kubernetes.io/projected/91ca9a98-134d-41d4-95a3-460d18d5986b-kube-api-access-hc567\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.400479 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400433 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.400479 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400462 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1d46c6c5-053d-4431-833a-439ecc5708f6-nginx-conf\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:02.400803 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400507 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-image-registry-private-configuration\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.400803 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400661 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-installation-pull-secrets\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.400803 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400692 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91ca9a98-134d-41d4-95a3-460d18d5986b-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.400803 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400704 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/f7a64cde-eca1-4b8d-843b-90f038244fe0-snapshots\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.400803 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400722 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7a64cde-eca1-4b8d-843b-90f038244fe0-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.400803 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400757 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-config\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400816 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd8r5\" (UniqueName: \"kubernetes.io/projected/ccb482af-112d-452e-a16f-49fe93eedf37-kube-api-access-hd8r5\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400850 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400888 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400930 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7a64cde-eca1-4b8d-843b-90f038244fe0-service-ca-bundle\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400938 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-certificates\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400959 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccb482af-112d-452e-a16f-49fe93eedf37-tmp-dir\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.400987 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-stats-auth\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401014 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/540077c6-a9f4-4a48-9f62-58c5f7d5c095-ca-trust-extracted\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.400960 2575 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.401041 2575 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-6659b7df7b-wn6k4: secret "image-registry-tls" not found Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401045 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fz8q\" (UniqueName: \"kubernetes.io/projected/2b707e63-9bbb-4244-adcd-2119a1c7187b-kube-api-access-5fz8q\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401089 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-trusted-ca\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.401137 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.401139 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls podName:540077c6-a9f4-4a48-9f62-58c5f7d5c095 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:02.901118429 +0000 UTC m=+34.061982757 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls") pod "image-registry-6659b7df7b-wn6k4" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095") : secret "image-registry-tls" not found Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401174 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a64cde-eca1-4b8d-843b-90f038244fe0-serving-cert\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401384 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/540077c6-a9f4-4a48-9f62-58c5f7d5c095-ca-trust-extracted\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401392 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7a64cde-eca1-4b8d-843b-90f038244fe0-service-ca-bundle\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401425 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vspsp\" (UniqueName: \"kubernetes.io/projected/f7a64cde-eca1-4b8d-843b-90f038244fe0-kube-api-access-vspsp\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401454 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f7a64cde-eca1-4b8d-843b-90f038244fe0-tmp\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401481 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ca9a98-134d-41d4-95a3-460d18d5986b-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401509 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-trusted-ca\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401665 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7a64cde-eca1-4b8d-843b-90f038244fe0-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.402059 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.401712 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f7a64cde-eca1-4b8d-843b-90f038244fe0-tmp\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.402521 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.402147 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-trusted-ca\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.405579 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.405542 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a64cde-eca1-4b8d-843b-90f038244fe0-serving-cert\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.405579 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.405553 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-installation-pull-secrets\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.405579 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.405556 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-image-registry-private-configuration\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.407625 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.407600 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58"] Apr 21 10:04:02.407822 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.407753 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.410607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.410251 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-ca\"" Apr 21 10:04:02.410607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.410268 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"open-cluster-management-agent-addon\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.410607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.410405 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-service-proxy-server-certificates\"" Apr 21 10:04:02.410607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.410503 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-hub-kubeconfig\"" Apr 21 10:04:02.410607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.410561 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"open-cluster-management-image-pull-credentials\"" Apr 21 10:04:02.410893 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.410673 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"open-cluster-management-agent-addon\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.410893 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.410786 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-open-cluster-management.io-proxy-agent-signer-client-cert\"" Apr 21 10:04:02.416509 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.416488 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-54tb9\" (UniqueName: \"kubernetes.io/projected/7e56e8b5-c500-424b-9f59-b08a49f14cce-kube-api-access-54tb9\") pod \"volume-data-source-validator-7c6cbb6c87-6nlj2\" (UID: \"7e56e8b5-c500-424b-9f59-b08a49f14cce\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2" Apr 21 10:04:02.416592 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.416520 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-bound-sa-token\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.416679 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.416657 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vspsp\" (UniqueName: \"kubernetes.io/projected/f7a64cde-eca1-4b8d-843b-90f038244fe0-kube-api-access-vspsp\") pod \"insights-operator-585dfdc468-4hz8p\" (UID: \"f7a64cde-eca1-4b8d-843b-90f038244fe0\") " pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.416785 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.416758 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f88tn\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-kube-api-access-f88tn\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.432728 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.432582 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-w964t"] Apr 21 10:04:02.433006 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.432984 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" Apr 21 10:04:02.435393 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.435362 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"managed-serviceaccount-dockercfg-2kt5s\"" Apr 21 10:04:02.435498 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.435452 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"managed-serviceaccount-hub-kubeconfig\"" Apr 21 10:04:02.448401 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.448378 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6"] Apr 21 10:04:02.448560 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.448471 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-w964t" Apr 21 10:04:02.451054 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.451005 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 21 10:04:02.451054 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.451019 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 21 10:04:02.451054 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.451018 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"network-diagnostics-dockercfg-dd4hz\"" Apr 21 10:04:02.471537 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471516 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-r495v"] Apr 21 10:04:02.471663 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471542 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-585dfdc468-4hz8p"] Apr 21 10:04:02.471663 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471561 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58"] Apr 21 10:04:02.471663 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471572 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6"] Apr 21 10:04:02.471663 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471584 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-w964t"] Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471707 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471750 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b5lnk"] Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471770 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b"] Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471783 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q"] Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471795 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8"] Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471808 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/router-default-56b8857454-d4h69"] Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471820 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct"] Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471832 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn"] Apr 21 10:04:02.471844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.471850 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-mrf6q"] Apr 21 10:04:02.474027 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.474009 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"work-manager-hub-kubeconfig\"" Apr 21 10:04:02.501971 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.501930 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-default-certificate\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.501971 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.501970 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hc567\" (UniqueName: \"kubernetes.io/projected/91ca9a98-134d-41d4-95a3-460d18d5986b-kube-api-access-hc567\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.502259 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.501999 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.502259 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502025 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtqbm\" (UniqueName: \"kubernetes.io/projected/0057318f-db18-4772-bd50-0cf37022dc7c-kube-api-access-xtqbm\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.502259 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502056 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1d46c6c5-053d-4431-833a-439ecc5708f6-nginx-conf\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:02.502259 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502076 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hub\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-hub\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.502259 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.502203 2575 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 10:04:02.502520 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502278 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91ca9a98-134d-41d4-95a3-460d18d5986b-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.502520 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.502297 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.002275641 +0000 UTC m=+34.163139976 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : secret "router-metrics-certs-default" not found Apr 21 10:04:02.502520 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502335 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-config\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.502520 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502369 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hd8r5\" (UniqueName: \"kubernetes.io/projected/ccb482af-112d-452e-a16f-49fe93eedf37-kube-api-access-hd8r5\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.502520 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502395 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.502520 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502426 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.502520 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502478 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccb482af-112d-452e-a16f-49fe93eedf37-tmp-dir\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.502520 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502503 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-stats-auth\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502535 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b468p\" (UniqueName: \"kubernetes.io/projected/eb96ed3e-5f61-48e5-829d-2b68b67c919a-kube-api-access-b468p\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.502570 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.002552747 +0000 UTC m=+34.163417077 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : configmap references non-existent config key: service-ca.crt Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502600 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/6a157b32-8d6c-48fb-b43a-de74eace04a2-hub-kubeconfig\") pod \"managed-serviceaccount-addon-agent-5984d6bcfc-g5x58\" (UID: \"6a157b32-8d6c-48fb-b43a-de74eace04a2\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502636 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csps9\" (UniqueName: \"kubernetes.io/projected/8372d417-f5b6-4e27-86b6-8c2b69e1103f-kube-api-access-csps9\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502691 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5fz8q\" (UniqueName: \"kubernetes.io/projected/2b707e63-9bbb-4244-adcd-2119a1c7187b-kube-api-access-5fz8q\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502735 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0057318f-db18-4772-bd50-0cf37022dc7c-config\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502768 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jjft\" (UniqueName: \"kubernetes.io/projected/6a157b32-8d6c-48fb-b43a-de74eace04a2-kube-api-access-4jjft\") pod \"managed-serviceaccount-addon-agent-5984d6bcfc-g5x58\" (UID: \"6a157b32-8d6c-48fb-b43a-de74eace04a2\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502798 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ca9a98-134d-41d4-95a3-460d18d5986b-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502823 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-trusted-ca\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502865 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-ca\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502888 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/ccb482af-112d-452e-a16f-49fe93eedf37-tmp-dir\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502901 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502923 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1d46c6c5-053d-4431-833a-439ecc5708f6-nginx-conf\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:02.502930 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502930 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.502969 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/2b707e63-9bbb-4244-adcd-2119a1c7187b-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503074 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.503081 2575 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503121 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503152 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/8372d417-f5b6-4e27-86b6-8c2b69e1103f-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.503158 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert podName:1d46c6c5-053d-4431-833a-439ecc5708f6 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.003143741 +0000 UTC m=+34.164008073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-5mv9q" (UID: "1d46c6c5-053d-4431-833a-439ecc5708f6") : secret "networking-console-plugin-cert" not found Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.503338 2575 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.503393 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls podName:2b707e63-9bbb-4244-adcd-2119a1c7187b nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.003379569 +0000 UTC m=+34.164243908 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-mscz8" (UID: "2b707e63-9bbb-4244-adcd-2119a1c7187b") : secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503431 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0057318f-db18-4772-bd50-0cf37022dc7c-serving-cert\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503479 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rwtbc\" (UniqueName: \"kubernetes.io/projected/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-kube-api-access-rwtbc\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503505 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-262hv\" (UniqueName: \"kubernetes.io/projected/cc037d14-2ee7-421f-875a-4ec33fac21bb-kube-api-access-262hv\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503518 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91ca9a98-134d-41d4-95a3-460d18d5986b-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503536 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-serving-cert\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.503607 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.503569 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccb482af-112d-452e-a16f-49fe93eedf37-config-volume\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.504380 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.504019 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccb482af-112d-452e-a16f-49fe93eedf37-config-volume\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.504380 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.504037 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-trusted-ca\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.504380 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.504139 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.504380 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.504182 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5r4s\" (UniqueName: \"kubernetes.io/projected/eff9f25f-8452-4746-806c-70ba95f122a8-kube-api-access-t5r4s\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:02.504380 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.504211 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.504380 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.504224 2575 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 10:04:02.504380 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.504259 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls podName:ccb482af-112d-452e-a16f-49fe93eedf37 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.00424741 +0000 UTC m=+34.165111747 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls") pod "dns-default-r495v" (UID: "ccb482af-112d-452e-a16f-49fe93eedf37") : secret "dns-default-metrics-tls" not found Apr 21 10:04:02.504380 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.504322 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/2b707e63-9bbb-4244-adcd-2119a1c7187b-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.504806 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.504750 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-config\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.505436 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.505412 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-stats-auth\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.505436 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.505430 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91ca9a98-134d-41d4-95a3-460d18d5986b-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.506118 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.506082 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-serving-cert\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.506260 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.506240 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-default-certificate\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.511961 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.511923 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc567\" (UniqueName: \"kubernetes.io/projected/91ca9a98-134d-41d4-95a3-460d18d5986b-kube-api-access-hc567\") pod \"kube-storage-version-migrator-operator-6769c5d45-7gxkl\" (UID: \"91ca9a98-134d-41d4-95a3-460d18d5986b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.512139 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.511995 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd8r5\" (UniqueName: \"kubernetes.io/projected/ccb482af-112d-452e-a16f-49fe93eedf37-kube-api-access-hd8r5\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:02.512139 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.512119 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-262hv\" (UniqueName: \"kubernetes.io/projected/cc037d14-2ee7-421f-875a-4ec33fac21bb-kube-api-access-262hv\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:02.513116 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.513069 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fz8q\" (UniqueName: \"kubernetes.io/projected/2b707e63-9bbb-4244-adcd-2119a1c7187b-kube-api-access-5fz8q\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:02.513811 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.513788 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwtbc\" (UniqueName: \"kubernetes.io/projected/c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3-kube-api-access-rwtbc\") pod \"console-operator-9d4b6777b-mrf6q\" (UID: \"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3\") " pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.526424 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.526396 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2" Apr 21 10:04:02.543000 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.542972 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-585dfdc468-4hz8p" Apr 21 10:04:02.579373 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.579333 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.605771 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.605991 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"klusterlet-config\" (UniqueName: \"kubernetes.io/secret/32458ba5-062e-4e68-9793-601d24c3f0fa-klusterlet-config\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606067 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606146 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b468p\" (UniqueName: \"kubernetes.io/projected/eb96ed3e-5f61-48e5-829d-2b68b67c919a-kube-api-access-b468p\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606184 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/6a157b32-8d6c-48fb-b43a-de74eace04a2-hub-kubeconfig\") pod \"managed-serviceaccount-addon-agent-5984d6bcfc-g5x58\" (UID: \"6a157b32-8d6c-48fb-b43a-de74eace04a2\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606224 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-csps9\" (UniqueName: \"kubernetes.io/projected/8372d417-f5b6-4e27-86b6-8c2b69e1103f-kube-api-access-csps9\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606267 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0057318f-db18-4772-bd50-0cf37022dc7c-config\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606303 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4jjft\" (UniqueName: \"kubernetes.io/projected/6a157b32-8d6c-48fb-b43a-de74eace04a2-kube-api-access-4jjft\") pod \"managed-serviceaccount-addon-agent-5984d6bcfc-g5x58\" (UID: \"6a157b32-8d6c-48fb-b43a-de74eace04a2\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606356 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-ca\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606404 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606456 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606487 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/8372d417-f5b6-4e27-86b6-8c2b69e1103f-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606534 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0057318f-db18-4772-bd50-0cf37022dc7c-serving-cert\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606586 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzvcl\" (UniqueName: \"kubernetes.io/projected/32458ba5-062e-4e68-9793-601d24c3f0fa-kube-api-access-lzvcl\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.606760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606662 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-t5r4s\" (UniqueName: \"kubernetes.io/projected/eff9f25f-8452-4746-806c-70ba95f122a8-kube-api-access-t5r4s\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:02.607522 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606804 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.607522 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606844 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/32458ba5-062e-4e68-9793-601d24c3f0fa-tmp\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.607522 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606884 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7t5f\" (UniqueName: \"kubernetes.io/projected/6b8f4897-7439-4c3d-9a79-d48ee11b3137-kube-api-access-r7t5f\") pod \"network-check-source-8894fc9bd-w964t\" (UID: \"6b8f4897-7439-4c3d-9a79-d48ee11b3137\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-w964t" Apr 21 10:04:02.607522 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606947 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xtqbm\" (UniqueName: \"kubernetes.io/projected/0057318f-db18-4772-bd50-0cf37022dc7c-kube-api-access-xtqbm\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.607522 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.606977 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hub\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-hub\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.607522 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.607209 2575 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 10:04:02.607522 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.607210 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0057318f-db18-4772-bd50-0cf37022dc7c-config\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.607522 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.607274 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert podName:eff9f25f-8452-4746-806c-70ba95f122a8 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.107254999 +0000 UTC m=+34.268119327 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert") pod "ingress-canary-b5lnk" (UID: "eff9f25f-8452-4746-806c-70ba95f122a8") : secret "canary-serving-cert" not found Apr 21 10:04:02.607956 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.607922 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/8372d417-f5b6-4e27-86b6-8c2b69e1103f-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.608090 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.607983 2575 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 10:04:02.608090 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.608046 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls podName:eb96ed3e-5f61-48e5-829d-2b68b67c919a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.108029809 +0000 UTC m=+34.268894140 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-gfm8b" (UID: "eb96ed3e-5f61-48e5-829d-2b68b67c919a") : secret "samples-operator-tls" not found Apr 21 10:04:02.610378 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.610351 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.610478 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.610400 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hub\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-hub\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.611233 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.611212 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.611337 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.611253 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca\" (UniqueName: \"kubernetes.io/secret/8372d417-f5b6-4e27-86b6-8c2b69e1103f-ca\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.613740 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.613715 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/6a157b32-8d6c-48fb-b43a-de74eace04a2-hub-kubeconfig\") pod \"managed-serviceaccount-addon-agent-5984d6bcfc-g5x58\" (UID: \"6a157b32-8d6c-48fb-b43a-de74eace04a2\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" Apr 21 10:04:02.613911 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.613889 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0057318f-db18-4772-bd50-0cf37022dc7c-serving-cert\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.615141 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.615121 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b468p\" (UniqueName: \"kubernetes.io/projected/eb96ed3e-5f61-48e5-829d-2b68b67c919a-kube-api-access-b468p\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:02.615261 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.615205 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jjft\" (UniqueName: \"kubernetes.io/projected/6a157b32-8d6c-48fb-b43a-de74eace04a2-kube-api-access-4jjft\") pod \"managed-serviceaccount-addon-agent-5984d6bcfc-g5x58\" (UID: \"6a157b32-8d6c-48fb-b43a-de74eace04a2\") " pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" Apr 21 10:04:02.616500 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.616481 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtqbm\" (UniqueName: \"kubernetes.io/projected/0057318f-db18-4772-bd50-0cf37022dc7c-kube-api-access-xtqbm\") pod \"service-ca-operator-d6fc45fc5-lpqct\" (UID: \"0057318f-db18-4772-bd50-0cf37022dc7c\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.617767 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.617746 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5r4s\" (UniqueName: \"kubernetes.io/projected/eff9f25f-8452-4746-806c-70ba95f122a8-kube-api-access-t5r4s\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:02.617767 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.617762 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-csps9\" (UniqueName: \"kubernetes.io/projected/8372d417-f5b6-4e27-86b6-8c2b69e1103f-kube-api-access-csps9\") pod \"cluster-proxy-proxy-agent-799b464b89-gxrxn\" (UID: \"8372d417-f5b6-4e27-86b6-8c2b69e1103f\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.663005 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.662918 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" Apr 21 10:04:02.707516 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.707479 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lzvcl\" (UniqueName: \"kubernetes.io/projected/32458ba5-062e-4e68-9793-601d24c3f0fa-kube-api-access-lzvcl\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.707725 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.707543 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/32458ba5-062e-4e68-9793-601d24c3f0fa-tmp\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.707725 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.707606 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-r7t5f\" (UniqueName: \"kubernetes.io/projected/6b8f4897-7439-4c3d-9a79-d48ee11b3137-kube-api-access-r7t5f\") pod \"network-check-source-8894fc9bd-w964t\" (UID: \"6b8f4897-7439-4c3d-9a79-d48ee11b3137\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-w964t" Apr 21 10:04:02.707725 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.707688 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"klusterlet-config\" (UniqueName: \"kubernetes.io/secret/32458ba5-062e-4e68-9793-601d24c3f0fa-klusterlet-config\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.707936 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.707915 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/32458ba5-062e-4e68-9793-601d24c3f0fa-tmp\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.712703 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.712671 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"klusterlet-config\" (UniqueName: \"kubernetes.io/secret/32458ba5-062e-4e68-9793-601d24c3f0fa-klusterlet-config\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.716306 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.716279 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzvcl\" (UniqueName: \"kubernetes.io/projected/32458ba5-062e-4e68-9793-601d24c3f0fa-kube-api-access-lzvcl\") pod \"klusterlet-addon-workmgr-69f755bfdd-sv5v6\" (UID: \"32458ba5-062e-4e68-9793-601d24c3f0fa\") " pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.716586 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.716570 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7t5f\" (UniqueName: \"kubernetes.io/projected/6b8f4897-7439-4c3d-9a79-d48ee11b3137-kube-api-access-r7t5f\") pod \"network-check-source-8894fc9bd-w964t\" (UID: \"6b8f4897-7439-4c3d-9a79-d48ee11b3137\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-w964t" Apr 21 10:04:02.747765 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.747694 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:02.757477 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.757455 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" Apr 21 10:04:02.773186 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.773155 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-w964t" Apr 21 10:04:02.780954 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.780931 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:02.909605 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:02.909571 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:02.909815 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.909730 2575 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 10:04:02.909815 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.909748 2575 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-6659b7df7b-wn6k4: secret "image-registry-tls" not found Apr 21 10:04:02.909815 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:02.909813 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls podName:540077c6-a9f4-4a48-9f62-58c5f7d5c095 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:03.909785494 +0000 UTC m=+35.070649823 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls") pod "image-registry-6659b7df7b-wn6k4" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095") : secret "image-registry-tls" not found Apr 21 10:04:03.010176 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.010069 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:03.010176 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.010156 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:03.010398 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010217 2575 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 21 10:04:03.010398 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010297 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert podName:1d46c6c5-053d-4431-833a-439ecc5708f6 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.010274819 +0000 UTC m=+35.171139174 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-5mv9q" (UID: "1d46c6c5-053d-4431-833a-439ecc5708f6") : secret "networking-console-plugin-cert" not found Apr 21 10:04:03.010398 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010301 2575 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 10:04:03.010398 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.010221 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:03.010398 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010303 2575 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:03.010398 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010350 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls podName:ccb482af-112d-452e-a16f-49fe93eedf37 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.010335569 +0000 UTC m=+35.171199914 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls") pod "dns-default-r495v" (UID: "ccb482af-112d-452e-a16f-49fe93eedf37") : secret "dns-default-metrics-tls" not found Apr 21 10:04:03.010398 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010400 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls podName:2b707e63-9bbb-4244-adcd-2119a1c7187b nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.010382907 +0000 UTC m=+35.171247236 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-mscz8" (UID: "2b707e63-9bbb-4244-adcd-2119a1c7187b") : secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:03.010710 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.010436 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:03.010710 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.010497 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:03.010710 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010570 2575 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 10:04:03.010710 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010586 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.010574165 +0000 UTC m=+35.171438512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : configmap references non-existent config key: service-ca.crt Apr 21 10:04:03.010710 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.010614 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.0106013 +0000 UTC m=+35.171465633 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : secret "router-metrics-certs-default" not found Apr 21 10:04:03.111401 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.111368 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:04:03.111580 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.111508 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:03.111580 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.111519 2575 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:03.111580 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.111560 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:03.111775 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.111598 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs podName:9276c6c8-dd13-4c6c-a3d3-3318a3731823 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:35.111578622 +0000 UTC m=+66.272442966 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs") pod "network-metrics-daemon-n8tpq" (UID: "9276c6c8-dd13-4c6c-a3d3-3318a3731823") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 21 10:04:03.111775 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.111668 2575 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 10:04:03.111775 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.111720 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls podName:eb96ed3e-5f61-48e5-829d-2b68b67c919a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.111704633 +0000 UTC m=+35.272569022 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-gfm8b" (UID: "eb96ed3e-5f61-48e5-829d-2b68b67c919a") : secret "samples-operator-tls" not found Apr 21 10:04:03.111775 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.111723 2575 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 10:04:03.111775 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.111748 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert podName:eff9f25f-8452-4746-806c-70ba95f122a8 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:04.111740553 +0000 UTC m=+35.272604881 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert") pod "ingress-canary-b5lnk" (UID: "eff9f25f-8452-4746-806c-70ba95f122a8") : secret "canary-serving-cert" not found Apr 21 10:04:03.213244 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.213088 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:04:03.216195 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.216157 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqpv5\" (UniqueName: \"kubernetes.io/projected/9db6586f-b886-4d23-84d5-48c25a2a6ac9-kube-api-access-cqpv5\") pod \"network-check-target-cnk4k\" (UID: \"9db6586f-b886-4d23-84d5-48c25a2a6ac9\") " pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:04:03.401967 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.401662 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:04:03.401967 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.401914 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:04:03.402453 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.402321 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:04:03.406566 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.405754 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-qrq7w\"" Apr 21 10:04:03.406566 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.406116 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 21 10:04:03.406566 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.406361 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-fjqp7\"" Apr 21 10:04:03.406566 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.406384 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 21 10:04:03.506468 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.504561 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:04:03.527478 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.527344 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-mrf6q"] Apr 21 10:04:03.531812 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.531765 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl"] Apr 21 10:04:03.534854 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.534832 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-w964t"] Apr 21 10:04:03.536053 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.536033 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58"] Apr 21 10:04:03.539639 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.539609 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2"] Apr 21 10:04:03.560793 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.560773 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6"] Apr 21 10:04:03.568648 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.568621 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn"] Apr 21 10:04:03.576243 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.576221 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct"] Apr 21 10:04:03.579117 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.579080 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-585dfdc468-4hz8p"] Apr 21 10:04:03.583277 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.583248 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b8f4897_7439_4c3d_9a79_d48ee11b3137.slice/crio-03cff24e2010a06c2d779ba2a08b2a536ede5d1e052f69e13dd99f9f8d5f7de7 WatchSource:0}: Error finding container 03cff24e2010a06c2d779ba2a08b2a536ede5d1e052f69e13dd99f9f8d5f7de7: Status 404 returned error can't find the container with id 03cff24e2010a06c2d779ba2a08b2a536ede5d1e052f69e13dd99f9f8d5f7de7 Apr 21 10:04:03.583623 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.583597 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91ca9a98_134d_41d4_95a3_460d18d5986b.slice/crio-198a118f9d3cbcd24da6f77e91d5afe137958bb2d989082e084abd2ce313ee31 WatchSource:0}: Error finding container 198a118f9d3cbcd24da6f77e91d5afe137958bb2d989082e084abd2ce313ee31: Status 404 returned error can't find the container with id 198a118f9d3cbcd24da6f77e91d5afe137958bb2d989082e084abd2ce313ee31 Apr 21 10:04:03.584063 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.584043 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a157b32_8d6c_48fb_b43a_de74eace04a2.slice/crio-9365828110ec5fd615d534d09966ad97e6ba5d1596fbdc1bd397de5ff3660a9b WatchSource:0}: Error finding container 9365828110ec5fd615d534d09966ad97e6ba5d1596fbdc1bd397de5ff3660a9b: Status 404 returned error can't find the container with id 9365828110ec5fd615d534d09966ad97e6ba5d1596fbdc1bd397de5ff3660a9b Apr 21 10:04:03.585447 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.585272 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e56e8b5_c500_424b_9f59_b08a49f14cce.slice/crio-2b365715de0c555d6b87565fa2d5ad570e27830b968114558e72cfed240e7244 WatchSource:0}: Error finding container 2b365715de0c555d6b87565fa2d5ad570e27830b968114558e72cfed240e7244: Status 404 returned error can't find the container with id 2b365715de0c555d6b87565fa2d5ad570e27830b968114558e72cfed240e7244 Apr 21 10:04:03.586769 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.586305 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8891fb4_b21c_41fd_9b9a_1add8bc8f1c3.slice/crio-355b6689256013d78ccbffefacc80d78126cd1d0965b70d9cc38b9020d5a4eec WatchSource:0}: Error finding container 355b6689256013d78ccbffefacc80d78126cd1d0965b70d9cc38b9020d5a4eec: Status 404 returned error can't find the container with id 355b6689256013d78ccbffefacc80d78126cd1d0965b70d9cc38b9020d5a4eec Apr 21 10:04:03.587614 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.587176 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32458ba5_062e_4e68_9793_601d24c3f0fa.slice/crio-80fb61f49f106f050001ba38c24cae99aa101b0058cf2d03616f749c11e0afbe WatchSource:0}: Error finding container 80fb61f49f106f050001ba38c24cae99aa101b0058cf2d03616f749c11e0afbe: Status 404 returned error can't find the container with id 80fb61f49f106f050001ba38c24cae99aa101b0058cf2d03616f749c11e0afbe Apr 21 10:04:03.589409 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.588936 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8372d417_f5b6_4e27_86b6_8c2b69e1103f.slice/crio-b9109879008a1186ee6d5c47fe2f73fdf03a4f2151af319e4378dc3ea80d75c2 WatchSource:0}: Error finding container b9109879008a1186ee6d5c47fe2f73fdf03a4f2151af319e4378dc3ea80d75c2: Status 404 returned error can't find the container with id b9109879008a1186ee6d5c47fe2f73fdf03a4f2151af319e4378dc3ea80d75c2 Apr 21 10:04:03.590074 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.590030 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0057318f_db18_4772_bd50_0cf37022dc7c.slice/crio-8990eec15d663ab3c8c4ac2db1333c8bdc1d4dc05810938f62c2ddd929cf46de WatchSource:0}: Error finding container 8990eec15d663ab3c8c4ac2db1333c8bdc1d4dc05810938f62c2ddd929cf46de: Status 404 returned error can't find the container with id 8990eec15d663ab3c8c4ac2db1333c8bdc1d4dc05810938f62c2ddd929cf46de Apr 21 10:04:03.591456 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.591428 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7a64cde_eca1_4b8d_843b_90f038244fe0.slice/crio-5cda69075c19d7883554d78971a0dc78f22b9bb36f1b56f3bb0f58051c6d52ce WatchSource:0}: Error finding container 5cda69075c19d7883554d78971a0dc78f22b9bb36f1b56f3bb0f58051c6d52ce: Status 404 returned error can't find the container with id 5cda69075c19d7883554d78971a0dc78f22b9bb36f1b56f3bb0f58051c6d52ce Apr 21 10:04:03.736247 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.736061 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-cnk4k"] Apr 21 10:04:03.749322 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:03.749291 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db6586f_b886_4d23_84d5_48c25a2a6ac9.slice/crio-b22c7bae832a126628ef5e660dec1b2205ac583efae407a9c463503d3901379d WatchSource:0}: Error finding container b22c7bae832a126628ef5e660dec1b2205ac583efae407a9c463503d3901379d: Status 404 returned error can't find the container with id b22c7bae832a126628ef5e660dec1b2205ac583efae407a9c463503d3901379d Apr 21 10:04:03.920487 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:03.920458 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:03.920625 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.920605 2575 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 10:04:03.920625 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.920623 2575 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-6659b7df7b-wn6k4: secret "image-registry-tls" not found Apr 21 10:04:03.920695 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:03.920678 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls podName:540077c6-a9f4-4a48-9f62-58c5f7d5c095 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:05.920663235 +0000 UTC m=+37.081527578 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls") pod "image-registry-6659b7df7b-wn6k4" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095") : secret "image-registry-tls" not found Apr 21 10:04:04.020913 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.020877 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:04.021076 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.020934 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:04.021076 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.020987 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:04.021076 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.021021 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:04.021076 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021058 2575 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 21 10:04:04.021076 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.021073 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:04.021377 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021157 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert podName:1d46c6c5-053d-4431-833a-439ecc5708f6 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:06.021134803 +0000 UTC m=+37.181999139 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-5mv9q" (UID: "1d46c6c5-053d-4431-833a-439ecc5708f6") : secret "networking-console-plugin-cert" not found Apr 21 10:04:04.021377 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021226 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:06.021207979 +0000 UTC m=+37.182072326 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : configmap references non-existent config key: service-ca.crt Apr 21 10:04:04.021377 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021260 2575 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 10:04:04.021377 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021291 2575 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 10:04:04.021377 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021298 2575 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:04.021377 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021309 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls podName:ccb482af-112d-452e-a16f-49fe93eedf37 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:06.021298402 +0000 UTC m=+37.182162730 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls") pod "dns-default-r495v" (UID: "ccb482af-112d-452e-a16f-49fe93eedf37") : secret "dns-default-metrics-tls" not found Apr 21 10:04:04.021377 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021328 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:06.02131735 +0000 UTC m=+37.182181680 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : secret "router-metrics-certs-default" not found Apr 21 10:04:04.021377 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.021344 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls podName:2b707e63-9bbb-4244-adcd-2119a1c7187b nodeName:}" failed. No retries permitted until 2026-04-21 10:04:06.021335698 +0000 UTC m=+37.182200027 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-mscz8" (UID: "2b707e63-9bbb-4244-adcd-2119a1c7187b") : secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:04.122051 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.122017 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:04.122232 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.122070 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:04.122232 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.122188 2575 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 10:04:04.122359 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.122238 2575 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 10:04:04.122359 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.122255 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert podName:eff9f25f-8452-4746-806c-70ba95f122a8 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:06.122236322 +0000 UTC m=+37.283100651 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert") pod "ingress-canary-b5lnk" (UID: "eff9f25f-8452-4746-806c-70ba95f122a8") : secret "canary-serving-cert" not found Apr 21 10:04:04.122359 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:04.122283 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls podName:eb96ed3e-5f61-48e5-829d-2b68b67c919a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:06.122270385 +0000 UTC m=+37.283134712 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-gfm8b" (UID: "eb96ed3e-5f61-48e5-829d-2b68b67c919a") : secret "samples-operator-tls" not found Apr 21 10:04:04.579332 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.579291 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" event={"ID":"32458ba5-062e-4e68-9793-601d24c3f0fa","Type":"ContainerStarted","Data":"80fb61f49f106f050001ba38c24cae99aa101b0058cf2d03616f749c11e0afbe"} Apr 21 10:04:04.581446 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.581365 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-cnk4k" event={"ID":"9db6586f-b886-4d23-84d5-48c25a2a6ac9","Type":"ContainerStarted","Data":"b22c7bae832a126628ef5e660dec1b2205ac583efae407a9c463503d3901379d"} Apr 21 10:04:04.583990 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.583949 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-4hz8p" event={"ID":"f7a64cde-eca1-4b8d-843b-90f038244fe0","Type":"ContainerStarted","Data":"5cda69075c19d7883554d78971a0dc78f22b9bb36f1b56f3bb0f58051c6d52ce"} Apr 21 10:04:04.586457 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.586394 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-w964t" event={"ID":"6b8f4897-7439-4c3d-9a79-d48ee11b3137","Type":"ContainerStarted","Data":"03cff24e2010a06c2d779ba2a08b2a536ede5d1e052f69e13dd99f9f8d5f7de7"} Apr 21 10:04:04.588531 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.588468 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" event={"ID":"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3","Type":"ContainerStarted","Data":"355b6689256013d78ccbffefacc80d78126cd1d0965b70d9cc38b9020d5a4eec"} Apr 21 10:04:04.597596 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.596556 2575 generic.go:358] "Generic (PLEG): container finished" podID="c7353367-eaa9-4d89-9444-14f3bfbfc166" containerID="7a5b0405663db0bfdea254b003c73ca509b1b3971fdc69e5c4d441cfcea9a005" exitCode=0 Apr 21 10:04:04.597596 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.596654 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerDied","Data":"7a5b0405663db0bfdea254b003c73ca509b1b3971fdc69e5c4d441cfcea9a005"} Apr 21 10:04:04.600610 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.600521 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" event={"ID":"91ca9a98-134d-41d4-95a3-460d18d5986b","Type":"ContainerStarted","Data":"198a118f9d3cbcd24da6f77e91d5afe137958bb2d989082e084abd2ce313ee31"} Apr 21 10:04:04.614898 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.614808 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" event={"ID":"0057318f-db18-4772-bd50-0cf37022dc7c","Type":"ContainerStarted","Data":"8990eec15d663ab3c8c4ac2db1333c8bdc1d4dc05810938f62c2ddd929cf46de"} Apr 21 10:04:04.616913 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.616888 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2" event={"ID":"7e56e8b5-c500-424b-9f59-b08a49f14cce","Type":"ContainerStarted","Data":"2b365715de0c555d6b87565fa2d5ad570e27830b968114558e72cfed240e7244"} Apr 21 10:04:04.618364 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.618319 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" event={"ID":"6a157b32-8d6c-48fb-b43a-de74eace04a2","Type":"ContainerStarted","Data":"9365828110ec5fd615d534d09966ad97e6ba5d1596fbdc1bd397de5ff3660a9b"} Apr 21 10:04:04.620199 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:04.620169 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" event={"ID":"8372d417-f5b6-4e27-86b6-8c2b69e1103f","Type":"ContainerStarted","Data":"b9109879008a1186ee6d5c47fe2f73fdf03a4f2151af319e4378dc3ea80d75c2"} Apr 21 10:04:05.686056 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:05.686009 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerStarted","Data":"403644d1e8de98b8fd88c8838eb319b2ff5125389409b64b258009119dad1da6"} Apr 21 10:04:05.949978 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:05.949025 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:05.949978 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:05.949371 2575 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 10:04:05.949978 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:05.949388 2575 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-6659b7df7b-wn6k4: secret "image-registry-tls" not found Apr 21 10:04:05.949978 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:05.949531 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls podName:540077c6-a9f4-4a48-9f62-58c5f7d5c095 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:09.949511026 +0000 UTC m=+41.110375373 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls") pod "image-registry-6659b7df7b-wn6k4" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095") : secret "image-registry-tls" not found Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:06.050693 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:06.050797 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:06.050894 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:06.051019 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:06.051065 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.051244 2575 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.051336 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls podName:2b707e63-9bbb-4244-adcd-2119a1c7187b nodeName:}" failed. No retries permitted until 2026-04-21 10:04:10.051315744 +0000 UTC m=+41.212180076 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-mscz8" (UID: "2b707e63-9bbb-4244-adcd-2119a1c7187b") : secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.051807 2575 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.051857 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls podName:ccb482af-112d-452e-a16f-49fe93eedf37 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:10.051841266 +0000 UTC m=+41.212705597 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls") pod "dns-default-r495v" (UID: "ccb482af-112d-452e-a16f-49fe93eedf37") : secret "dns-default-metrics-tls" not found Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.051967 2575 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.052023 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:10.052010087 +0000 UTC m=+41.212874419 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : secret "router-metrics-certs-default" not found Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.052135 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:10.052123646 +0000 UTC m=+41.212987978 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : configmap references non-existent config key: service-ca.crt Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.052235 2575 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 21 10:04:06.052325 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.052276 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert podName:1d46c6c5-053d-4431-833a-439ecc5708f6 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:10.052264219 +0000 UTC m=+41.213128550 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-5mv9q" (UID: "1d46c6c5-053d-4431-833a-439ecc5708f6") : secret "networking-console-plugin-cert" not found Apr 21 10:04:06.152588 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:06.152108 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:06.152588 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:06.152172 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:06.152588 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.152222 2575 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 10:04:06.152588 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.152302 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert podName:eff9f25f-8452-4746-806c-70ba95f122a8 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:10.152280371 +0000 UTC m=+41.313144705 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert") pod "ingress-canary-b5lnk" (UID: "eff9f25f-8452-4746-806c-70ba95f122a8") : secret "canary-serving-cert" not found Apr 21 10:04:06.152588 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.152372 2575 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 10:04:06.152588 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:06.152423 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls podName:eb96ed3e-5f61-48e5-829d-2b68b67c919a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:10.152407199 +0000 UTC m=+41.313271530 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-gfm8b" (UID: "eb96ed3e-5f61-48e5-829d-2b68b67c919a") : secret "samples-operator-tls" not found Apr 21 10:04:07.707092 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:07.707052 2575 generic.go:358] "Generic (PLEG): container finished" podID="c7353367-eaa9-4d89-9444-14f3bfbfc166" containerID="403644d1e8de98b8fd88c8838eb319b2ff5125389409b64b258009119dad1da6" exitCode=0 Apr 21 10:04:07.707604 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:07.707126 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerDied","Data":"403644d1e8de98b8fd88c8838eb319b2ff5125389409b64b258009119dad1da6"} Apr 21 10:04:08.379308 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:08.378278 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:04:08.383560 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:08.383503 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/37ae773e-f021-4a13-81c5-678bcf84335e-original-pull-secret\") pod \"global-pull-secret-syncer-t8wp8\" (UID: \"37ae773e-f021-4a13-81c5-678bcf84335e\") " pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:04:08.628900 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:08.628867 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-t8wp8" Apr 21 10:04:09.994752 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:09.994701 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:09.995263 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:09.994875 2575 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 10:04:09.995263 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:09.994900 2575 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-6659b7df7b-wn6k4: secret "image-registry-tls" not found Apr 21 10:04:09.995263 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:09.994983 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls podName:540077c6-a9f4-4a48-9f62-58c5f7d5c095 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:17.994960271 +0000 UTC m=+49.155824614 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls") pod "image-registry-6659b7df7b-wn6k4" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095") : secret "image-registry-tls" not found Apr 21 10:04:10.096261 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:10.096215 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:10.096457 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:10.096289 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:10.096457 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:10.096352 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:10.096457 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096368 2575 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 21 10:04:10.096457 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:10.096395 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:10.096457 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096431 2575 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:10.096457 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096445 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert podName:1d46c6c5-053d-4431-833a-439ecc5708f6 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:18.096424288 +0000 UTC m=+49.257288618 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-5mv9q" (UID: "1d46c6c5-053d-4431-833a-439ecc5708f6") : secret "networking-console-plugin-cert" not found Apr 21 10:04:10.096761 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096488 2575 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 10:04:10.096761 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096505 2575 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 10:04:10.096761 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096506 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls podName:2b707e63-9bbb-4244-adcd-2119a1c7187b nodeName:}" failed. No retries permitted until 2026-04-21 10:04:18.096494898 +0000 UTC m=+49.257359226 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-mscz8" (UID: "2b707e63-9bbb-4244-adcd-2119a1c7187b") : secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:10.096761 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:10.096588 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:10.096761 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096617 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:18.09660335 +0000 UTC m=+49.257467678 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : secret "router-metrics-certs-default" not found Apr 21 10:04:10.096761 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096642 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls podName:ccb482af-112d-452e-a16f-49fe93eedf37 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:18.096630705 +0000 UTC m=+49.257495044 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls") pod "dns-default-r495v" (UID: "ccb482af-112d-452e-a16f-49fe93eedf37") : secret "dns-default-metrics-tls" not found Apr 21 10:04:10.096761 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.096664 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:18.096651646 +0000 UTC m=+49.257515975 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : configmap references non-existent config key: service-ca.crt Apr 21 10:04:10.197943 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:10.197908 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:10.198143 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:10.197965 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:10.198143 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.198068 2575 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 10:04:10.198247 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.198152 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert podName:eff9f25f-8452-4746-806c-70ba95f122a8 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:18.198136376 +0000 UTC m=+49.359000704 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert") pod "ingress-canary-b5lnk" (UID: "eff9f25f-8452-4746-806c-70ba95f122a8") : secret "canary-serving-cert" not found Apr 21 10:04:10.198247 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.198078 2575 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 10:04:10.198247 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:10.198228 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls podName:eb96ed3e-5f61-48e5-829d-2b68b67c919a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:18.198214928 +0000 UTC m=+49.359079269 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-gfm8b" (UID: "eb96ed3e-5f61-48e5-829d-2b68b67c919a") : secret "samples-operator-tls" not found Apr 21 10:04:16.025527 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.025482 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-t8wp8"] Apr 21 10:04:16.282821 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:16.282789 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37ae773e_f021_4a13_81c5_678bcf84335e.slice/crio-98701f3ea3245861ed55316f8c9d6e17478bed8b38a22d099739746a8f3d9991 WatchSource:0}: Error finding container 98701f3ea3245861ed55316f8c9d6e17478bed8b38a22d099739746a8f3d9991: Status 404 returned error can't find the container with id 98701f3ea3245861ed55316f8c9d6e17478bed8b38a22d099739746a8f3d9991 Apr 21 10:04:16.730313 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.730273 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pj28j" event={"ID":"c7353367-eaa9-4d89-9444-14f3bfbfc166","Type":"ContainerStarted","Data":"a6055b019b5bc54045c614ebae2a75582ac2f2c5e1d8f940ad13049e8526f143"} Apr 21 10:04:16.732069 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.732019 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" event={"ID":"91ca9a98-134d-41d4-95a3-460d18d5986b","Type":"ContainerStarted","Data":"28854bbff905a8f63604b9f5b2b053d8313b222005e26ade10f746cac4550c47"} Apr 21 10:04:16.734223 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.734180 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" event={"ID":"0057318f-db18-4772-bd50-0cf37022dc7c","Type":"ContainerStarted","Data":"debb7ab0f1dc13ab8d8cabd6f5fd989bc5812489d61ebc616c644a83575f2ba4"} Apr 21 10:04:16.735837 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.735809 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2" event={"ID":"7e56e8b5-c500-424b-9f59-b08a49f14cce","Type":"ContainerStarted","Data":"70c587aaa61f840b9279f38f9571600644f051c4bc569eb734dd7794c4ff9e23"} Apr 21 10:04:16.737294 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.737072 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" event={"ID":"6a157b32-8d6c-48fb-b43a-de74eace04a2","Type":"ContainerStarted","Data":"41588600c24d80b47e665efa1c3da73f3b1885137adf91da95b7e5ec94576b60"} Apr 21 10:04:16.738626 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.738446 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" event={"ID":"8372d417-f5b6-4e27-86b6-8c2b69e1103f","Type":"ContainerStarted","Data":"5768fd00c4032e2b85ab437c986d08c4a8f73d93729a824f637004b9a06827db"} Apr 21 10:04:16.739821 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.739800 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" event={"ID":"32458ba5-062e-4e68-9793-601d24c3f0fa","Type":"ContainerStarted","Data":"b5a21ee1af530eff63f10816ca1753fd724f4beae07c6755ddaae105c6f71b77"} Apr 21 10:04:16.740458 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.740441 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:16.742344 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.742313 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" Apr 21 10:04:16.742667 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.742484 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-cnk4k" event={"ID":"9db6586f-b886-4d23-84d5-48c25a2a6ac9","Type":"ContainerStarted","Data":"8e0b451d9d2dfc4656f0a4addc149cd13fbc54e1263e02e3e914b7a1a50cd656"} Apr 21 10:04:16.742855 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.742836 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:04:16.743978 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.743959 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-4hz8p" event={"ID":"f7a64cde-eca1-4b8d-843b-90f038244fe0","Type":"ContainerStarted","Data":"2364df08b9966ff5c04dee6b98a471190febf5f7b7bb666a390a37c9e988d3ce"} Apr 21 10:04:16.745212 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.745192 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-t8wp8" event={"ID":"37ae773e-f021-4a13-81c5-678bcf84335e","Type":"ContainerStarted","Data":"98701f3ea3245861ed55316f8c9d6e17478bed8b38a22d099739746a8f3d9991"} Apr 21 10:04:16.746571 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.746549 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-w964t" event={"ID":"6b8f4897-7439-4c3d-9a79-d48ee11b3137","Type":"ContainerStarted","Data":"cb830ae5d2db3ab3db90d258439c5d6d10534baf351247363109d2ef84b94863"} Apr 21 10:04:16.748169 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.748147 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/0.log" Apr 21 10:04:16.748254 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.748183 2575 generic.go:358] "Generic (PLEG): container finished" podID="c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3" containerID="d1ff43406d5a5c832e8213b0a9d1ff38880299d1a3c4b1a6153f0950bda6887c" exitCode=255 Apr 21 10:04:16.748254 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.748213 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" event={"ID":"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3","Type":"ContainerDied","Data":"d1ff43406d5a5c832e8213b0a9d1ff38880299d1a3c4b1a6153f0950bda6887c"} Apr 21 10:04:16.748407 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.748394 2575 scope.go:117] "RemoveContainer" containerID="d1ff43406d5a5c832e8213b0a9d1ff38880299d1a3c4b1a6153f0950bda6887c" Apr 21 10:04:16.758949 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.758902 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-pj28j" podStartSLOduration=16.132365495 podStartE2EDuration="47.758888159s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:03:32.01039646 +0000 UTC m=+3.171260795" lastFinishedPulling="2026-04-21 10:04:03.636919127 +0000 UTC m=+34.797783459" observedRunningTime="2026-04-21 10:04:16.75673279 +0000 UTC m=+47.917597142" watchObservedRunningTime="2026-04-21 10:04:16.758888159 +0000 UTC m=+47.919752511" Apr 21 10:04:16.802070 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.802008 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-cnk4k" podStartSLOduration=35.671090662 podStartE2EDuration="47.801988995s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.751072421 +0000 UTC m=+34.911936748" lastFinishedPulling="2026-04-21 10:04:15.881970745 +0000 UTC m=+47.042835081" observedRunningTime="2026-04-21 10:04:16.776815914 +0000 UTC m=+47.937680265" watchObservedRunningTime="2026-04-21 10:04:16.801988995 +0000 UTC m=+47.962853350" Apr 21 10:04:16.823248 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.823205 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="open-cluster-management-agent-addon/managed-serviceaccount-addon-agent-5984d6bcfc-g5x58" podStartSLOduration=4.542574783 podStartE2EDuration="16.823172061s" podCreationTimestamp="2026-04-21 10:04:00 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.586613818 +0000 UTC m=+34.747478161" lastFinishedPulling="2026-04-21 10:04:15.867211094 +0000 UTC m=+47.028075439" observedRunningTime="2026-04-21 10:04:16.803183588 +0000 UTC m=+47.964047938" watchObservedRunningTime="2026-04-21 10:04:16.823172061 +0000 UTC m=+47.984036412" Apr 21 10:04:16.838788 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.838672 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-w964t" podStartSLOduration=32.628261691 podStartE2EDuration="44.838657391s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.585862447 +0000 UTC m=+34.746726785" lastFinishedPulling="2026-04-21 10:04:15.796258158 +0000 UTC m=+46.957122485" observedRunningTime="2026-04-21 10:04:16.837822111 +0000 UTC m=+47.998686486" watchObservedRunningTime="2026-04-21 10:04:16.838657391 +0000 UTC m=+47.999521741" Apr 21 10:04:16.856165 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.856122 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" podStartSLOduration=32.576361692 podStartE2EDuration="44.856085352s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.586610782 +0000 UTC m=+34.747475113" lastFinishedPulling="2026-04-21 10:04:15.866334429 +0000 UTC m=+47.027198773" observedRunningTime="2026-04-21 10:04:16.854641013 +0000 UTC m=+48.015505363" watchObservedRunningTime="2026-04-21 10:04:16.856085352 +0000 UTC m=+48.016949701" Apr 21 10:04:16.879173 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.879119 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" podStartSLOduration=32.619283594 podStartE2EDuration="44.87907912s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.607187133 +0000 UTC m=+34.768051462" lastFinishedPulling="2026-04-21 10:04:15.866982659 +0000 UTC m=+47.027846988" observedRunningTime="2026-04-21 10:04:16.877996732 +0000 UTC m=+48.038861084" watchObservedRunningTime="2026-04-21 10:04:16.87907912 +0000 UTC m=+48.039943472" Apr 21 10:04:16.897695 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.897631 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-6nlj2" podStartSLOduration=35.54088264 podStartE2EDuration="44.89761488s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.60658755 +0000 UTC m=+34.767451895" lastFinishedPulling="2026-04-21 10:04:12.963319791 +0000 UTC m=+44.124184135" observedRunningTime="2026-04-21 10:04:16.897383857 +0000 UTC m=+48.058248208" watchObservedRunningTime="2026-04-21 10:04:16.89761488 +0000 UTC m=+48.058479236" Apr 21 10:04:16.917202 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.917155 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-operator-585dfdc468-4hz8p" podStartSLOduration=32.730594329 podStartE2EDuration="44.917139549s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.607146786 +0000 UTC m=+34.768011119" lastFinishedPulling="2026-04-21 10:04:15.793691994 +0000 UTC m=+46.954556339" observedRunningTime="2026-04-21 10:04:16.916143881 +0000 UTC m=+48.077008232" watchObservedRunningTime="2026-04-21 10:04:16.917139549 +0000 UTC m=+48.078003898" Apr 21 10:04:16.945189 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:16.944904 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="open-cluster-management-agent-addon/klusterlet-addon-workmgr-69f755bfdd-sv5v6" podStartSLOduration=4.679689841 podStartE2EDuration="16.9448825s" podCreationTimestamp="2026-04-21 10:04:00 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.607084016 +0000 UTC m=+34.767948351" lastFinishedPulling="2026-04-21 10:04:15.872276682 +0000 UTC m=+47.033141010" observedRunningTime="2026-04-21 10:04:16.944592293 +0000 UTC m=+48.105456666" watchObservedRunningTime="2026-04-21 10:04:16.9448825 +0000 UTC m=+48.105746852" Apr 21 10:04:17.754286 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:17.754200 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:04:17.755137 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:17.755090 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/0.log" Apr 21 10:04:17.755325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:17.755160 2575 generic.go:358] "Generic (PLEG): container finished" podID="c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3" containerID="637ca51b2c409a3e55c32eefa0ba9e304581011c338f18a6696a64a6212a2932" exitCode=255 Apr 21 10:04:17.755325 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:17.755299 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" event={"ID":"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3","Type":"ContainerDied","Data":"637ca51b2c409a3e55c32eefa0ba9e304581011c338f18a6696a64a6212a2932"} Apr 21 10:04:17.755441 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:17.755346 2575 scope.go:117] "RemoveContainer" containerID="d1ff43406d5a5c832e8213b0a9d1ff38880299d1a3c4b1a6153f0950bda6887c" Apr 21 10:04:17.756042 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:17.755568 2575 scope.go:117] "RemoveContainer" containerID="637ca51b2c409a3e55c32eefa0ba9e304581011c338f18a6696a64a6212a2932" Apr 21 10:04:17.756042 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:17.755790 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-mrf6q_openshift-console-operator(c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3)\"" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" podUID="c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3" Apr 21 10:04:18.077839 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.077801 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:18.078148 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.078090 2575 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Apr 21 10:04:18.078148 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.078137 2575 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-6659b7df7b-wn6k4: secret "image-registry-tls" not found Apr 21 10:04:18.078342 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.078195 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls podName:540077c6-a9f4-4a48-9f62-58c5f7d5c095 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:34.078177788 +0000 UTC m=+65.239042125 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls") pod "image-registry-6659b7df7b-wn6k4" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095") : secret "image-registry-tls" not found Apr 21 10:04:18.079415 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.079391 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr"] Apr 21 10:04:18.103247 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.102180 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr"] Apr 21 10:04:18.103247 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.102325 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" Apr 21 10:04:18.110661 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.110545 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Apr 21 10:04:18.111769 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.110811 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-g4mrk\"" Apr 21 10:04:18.111769 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.111030 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Apr 21 10:04:18.179468 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.179427 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:18.179667 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.179493 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:18.179667 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.179553 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:18.179667 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.179580 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbv89\" (UniqueName: \"kubernetes.io/projected/da1e2988-6ac0-4b0c-821b-5d0d7989985f-kube-api-access-cbv89\") pod \"migrator-74bb7799d9-246pr\" (UID: \"da1e2988-6ac0-4b0c-821b-5d0d7989985f\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" Apr 21 10:04:18.179667 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179611 2575 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 21 10:04:18.179667 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.179655 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:18.179838 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179659 2575 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Apr 21 10:04:18.179838 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179695 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls podName:ccb482af-112d-452e-a16f-49fe93eedf37 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:34.179674517 +0000 UTC m=+65.340538852 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls") pod "dns-default-r495v" (UID: "ccb482af-112d-452e-a16f-49fe93eedf37") : secret "dns-default-metrics-tls" not found Apr 21 10:04:18.179838 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179715 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:34.179706774 +0000 UTC m=+65.340571103 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : secret "router-metrics-certs-default" not found Apr 21 10:04:18.179838 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179735 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle podName:cc037d14-2ee7-421f-875a-4ec33fac21bb nodeName:}" failed. No retries permitted until 2026-04-21 10:04:34.179728574 +0000 UTC m=+65.340592901 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle") pod "router-default-56b8857454-d4h69" (UID: "cc037d14-2ee7-421f-875a-4ec33fac21bb") : configmap references non-existent config key: service-ca.crt Apr 21 10:04:18.179838 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179763 2575 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: secret "networking-console-plugin-cert" not found Apr 21 10:04:18.179838 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.179769 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:18.179838 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179792 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert podName:1d46c6c5-053d-4431-833a-439ecc5708f6 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:34.179780341 +0000 UTC m=+65.340644670 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert") pod "networking-console-plugin-cb95c66f6-5mv9q" (UID: "1d46c6c5-053d-4431-833a-439ecc5708f6") : secret "networking-console-plugin-cert" not found Apr 21 10:04:18.179838 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179825 2575 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:18.180136 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.179860 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls podName:2b707e63-9bbb-4244-adcd-2119a1c7187b nodeName:}" failed. No retries permitted until 2026-04-21 10:04:34.179849565 +0000 UTC m=+65.340713901 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-mscz8" (UID: "2b707e63-9bbb-4244-adcd-2119a1c7187b") : secret "cluster-monitoring-operator-tls" not found Apr 21 10:04:18.281263 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.281224 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:18.281434 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.281293 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:18.281434 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.281394 2575 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 21 10:04:18.281434 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.281427 2575 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 21 10:04:18.281557 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.281448 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cbv89\" (UniqueName: \"kubernetes.io/projected/da1e2988-6ac0-4b0c-821b-5d0d7989985f-kube-api-access-cbv89\") pod \"migrator-74bb7799d9-246pr\" (UID: \"da1e2988-6ac0-4b0c-821b-5d0d7989985f\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" Apr 21 10:04:18.281557 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.281466 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert podName:eff9f25f-8452-4746-806c-70ba95f122a8 nodeName:}" failed. No retries permitted until 2026-04-21 10:04:34.281446205 +0000 UTC m=+65.442310534 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert") pod "ingress-canary-b5lnk" (UID: "eff9f25f-8452-4746-806c-70ba95f122a8") : secret "canary-serving-cert" not found Apr 21 10:04:18.281557 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.281551 2575 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls podName:eb96ed3e-5f61-48e5-829d-2b68b67c919a nodeName:}" failed. No retries permitted until 2026-04-21 10:04:34.281532751 +0000 UTC m=+65.442397085 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-gfm8b" (UID: "eb96ed3e-5f61-48e5-829d-2b68b67c919a") : secret "samples-operator-tls" not found Apr 21 10:04:18.289894 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.289868 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbv89\" (UniqueName: \"kubernetes.io/projected/da1e2988-6ac0-4b0c-821b-5d0d7989985f-kube-api-access-cbv89\") pod \"migrator-74bb7799d9-246pr\" (UID: \"da1e2988-6ac0-4b0c-821b-5d0d7989985f\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" Apr 21 10:04:18.418592 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.418520 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" Apr 21 10:04:18.562027 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.561989 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr"] Apr 21 10:04:18.564813 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:18.564788 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda1e2988_6ac0_4b0c_821b_5d0d7989985f.slice/crio-42f5e52eb225d4003ca54f23b114894d82e9d6464f4ae035c426d6d29d8d9d44 WatchSource:0}: Error finding container 42f5e52eb225d4003ca54f23b114894d82e9d6464f4ae035c426d6d29d8d9d44: Status 404 returned error can't find the container with id 42f5e52eb225d4003ca54f23b114894d82e9d6464f4ae035c426d6d29d8d9d44 Apr 21 10:04:18.759864 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.759824 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" event={"ID":"da1e2988-6ac0-4b0c-821b-5d0d7989985f","Type":"ContainerStarted","Data":"42f5e52eb225d4003ca54f23b114894d82e9d6464f4ae035c426d6d29d8d9d44"} Apr 21 10:04:18.761208 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.761186 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:04:18.761756 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:18.761737 2575 scope.go:117] "RemoveContainer" containerID="637ca51b2c409a3e55c32eefa0ba9e304581011c338f18a6696a64a6212a2932" Apr 21 10:04:18.761908 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:18.761892 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-mrf6q_openshift-console-operator(c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3)\"" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" podUID="c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3" Apr 21 10:04:20.845730 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.845703 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-krcxr_e6f11c13-e164-47be-a16a-c5ea7b32f9e4/dns-node-resolver/0.log" Apr 21 10:04:20.883707 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.883673 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-865cb79987-k8jpg"] Apr 21 10:04:20.886794 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.886771 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:20.889367 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.889345 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Apr 21 10:04:20.890282 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.890258 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Apr 21 10:04:20.890367 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.890326 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Apr 21 10:04:20.890451 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.890396 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-97lrs\"" Apr 21 10:04:20.890514 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.890450 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Apr 21 10:04:20.895453 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:20.895431 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-k8jpg"] Apr 21 10:04:21.008233 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.008193 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/33c0eac6-4107-440f-9e3b-d1935f422428-signing-cabundle\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.008425 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.008334 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/33c0eac6-4107-440f-9e3b-d1935f422428-signing-key\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.008425 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.008386 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6c48\" (UniqueName: \"kubernetes.io/projected/33c0eac6-4107-440f-9e3b-d1935f422428-kube-api-access-n6c48\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.109541 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.109446 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/33c0eac6-4107-440f-9e3b-d1935f422428-signing-cabundle\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.109713 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.109574 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/33c0eac6-4107-440f-9e3b-d1935f422428-signing-key\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.109713 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.109619 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-n6c48\" (UniqueName: \"kubernetes.io/projected/33c0eac6-4107-440f-9e3b-d1935f422428-kube-api-access-n6c48\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.110260 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.110231 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/33c0eac6-4107-440f-9e3b-d1935f422428-signing-cabundle\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.112287 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.112257 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/33c0eac6-4107-440f-9e3b-d1935f422428-signing-key\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.119706 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.119677 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6c48\" (UniqueName: \"kubernetes.io/projected/33c0eac6-4107-440f-9e3b-d1935f422428-kube-api-access-n6c48\") pod \"service-ca-865cb79987-k8jpg\" (UID: \"33c0eac6-4107-440f-9e3b-d1935f422428\") " pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.197696 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.197658 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-k8jpg" Apr 21 10:04:21.392419 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.392393 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-k8jpg"] Apr 21 10:04:21.771535 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.771401 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" event={"ID":"da1e2988-6ac0-4b0c-821b-5d0d7989985f","Type":"ContainerStarted","Data":"d1984922abf185a5eb52daf77a3a42df5f7810e4d8aea868d3facb0bd04ce546"} Apr 21 10:04:21.771535 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.771445 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" event={"ID":"da1e2988-6ac0-4b0c-821b-5d0d7989985f","Type":"ContainerStarted","Data":"a3c379fec5e0e83128bf9d476bd5b8ef32f74e2e3b226bf0e8709035ff7a1661"} Apr 21 10:04:21.772905 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.772877 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-t8wp8" event={"ID":"37ae773e-f021-4a13-81c5-678bcf84335e","Type":"ContainerStarted","Data":"39bfbc98cad297ce42f88acefd0636d95ba553f78d969bdaab7af14d6c61c1ff"} Apr 21 10:04:21.774378 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.774347 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-k8jpg" event={"ID":"33c0eac6-4107-440f-9e3b-d1935f422428","Type":"ContainerStarted","Data":"b31ddcbb28c98283a511b07b9cc20fefdde961bcb99d6c5aa0afbb7ab91c44a8"} Apr 21 10:04:21.774490 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.774381 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-k8jpg" event={"ID":"33c0eac6-4107-440f-9e3b-d1935f422428","Type":"ContainerStarted","Data":"b4912987b8fe8feb1af4dffeb192a4e644a7d83a45c53434aed3ee5a2e711d2a"} Apr 21 10:04:21.792494 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.792249 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-246pr" podStartSLOduration=1.124263339 podStartE2EDuration="3.792232525s" podCreationTimestamp="2026-04-21 10:04:18 +0000 UTC" firstStartedPulling="2026-04-21 10:04:18.56684134 +0000 UTC m=+49.727705671" lastFinishedPulling="2026-04-21 10:04:21.234810528 +0000 UTC m=+52.395674857" observedRunningTime="2026-04-21 10:04:21.7916718 +0000 UTC m=+52.952536152" watchObservedRunningTime="2026-04-21 10:04:21.792232525 +0000 UTC m=+52.953096875" Apr 21 10:04:21.812876 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.812819 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-t8wp8" podStartSLOduration=24.872975447 podStartE2EDuration="29.812802573s" podCreationTimestamp="2026-04-21 10:03:52 +0000 UTC" firstStartedPulling="2026-04-21 10:04:16.305023394 +0000 UTC m=+47.465887737" lastFinishedPulling="2026-04-21 10:04:21.244850527 +0000 UTC m=+52.405714863" observedRunningTime="2026-04-21 10:04:21.812564698 +0000 UTC m=+52.973429048" watchObservedRunningTime="2026-04-21 10:04:21.812802573 +0000 UTC m=+52.973666926" Apr 21 10:04:21.830995 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.830936 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-865cb79987-k8jpg" podStartSLOduration=1.830920431 podStartE2EDuration="1.830920431s" podCreationTimestamp="2026-04-21 10:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:04:21.830331827 +0000 UTC m=+52.991196203" watchObservedRunningTime="2026-04-21 10:04:21.830920431 +0000 UTC m=+52.991784782" Apr 21 10:04:21.846888 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:21.846860 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-d5g7x_b776059e-ddd7-4893-92ef-c96817a1ae4f/node-ca/0.log" Apr 21 10:04:22.607316 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:22.607057 2575 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:22.607316 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:22.607117 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:22.607647 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:22.607629 2575 scope.go:117] "RemoveContainer" containerID="637ca51b2c409a3e55c32eefa0ba9e304581011c338f18a6696a64a6212a2932" Apr 21 10:04:22.607869 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:04:22.607845 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-mrf6q_openshift-console-operator(c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3)\"" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" podUID="c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3" Apr 21 10:04:22.650942 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:22.650908 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-246pr_da1e2988-6ac0-4b0c-821b-5d0d7989985f/migrator/0.log" Apr 21 10:04:22.848073 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:22.847971 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-246pr_da1e2988-6ac0-4b0c-821b-5d0d7989985f/graceful-termination/0.log" Apr 21 10:04:23.048064 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:23.047941 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-7gxkl_91ca9a98-134d-41d4-95a3-460d18d5986b/kube-storage-version-migrator-operator/0.log" Apr 21 10:04:23.784302 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:23.784264 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" event={"ID":"8372d417-f5b6-4e27-86b6-8c2b69e1103f","Type":"ContainerStarted","Data":"389686ffcb5e09301c5854ef7868b7b9c676e9728971930801e0a06a532c4676"} Apr 21 10:04:23.784302 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:23.784303 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" event={"ID":"8372d417-f5b6-4e27-86b6-8c2b69e1103f","Type":"ContainerStarted","Data":"3a140adbfdd966eeff4a5fd2bed296f7a8839fdf71e662f55277c13f0e7e4cb2"} Apr 21 10:04:23.803452 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:23.803397 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" podStartSLOduration=4.526081953 podStartE2EDuration="23.803379301s" podCreationTimestamp="2026-04-21 10:04:00 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.606933307 +0000 UTC m=+34.767797638" lastFinishedPulling="2026-04-21 10:04:22.884230653 +0000 UTC m=+54.045094986" observedRunningTime="2026-04-21 10:04:23.802271333 +0000 UTC m=+54.963135688" watchObservedRunningTime="2026-04-21 10:04:23.803379301 +0000 UTC m=+54.964243656" Apr 21 10:04:26.566732 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:26.566701 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5bsvf" Apr 21 10:04:32.749890 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:32.749830 2575 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" podUID="8372d417-f5b6-4e27-86b6-8c2b69e1103f" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 21 10:04:34.141808 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.141764 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:34.144150 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.144128 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"image-registry-6659b7df7b-wn6k4\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:34.242493 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.242457 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:34.242493 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.242509 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:34.242756 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.242554 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:34.242756 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.242581 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:34.242756 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.242624 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:34.243381 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.243355 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc037d14-2ee7-421f-875a-4ec33fac21bb-service-ca-bundle\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:34.244922 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.244901 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/1d46c6c5-053d-4431-833a-439ecc5708f6-networking-console-plugin-cert\") pod \"networking-console-plugin-cb95c66f6-5mv9q\" (UID: \"1d46c6c5-053d-4431-833a-439ecc5708f6\") " pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:34.245056 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.245032 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b707e63-9bbb-4244-adcd-2119a1c7187b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-mscz8\" (UID: \"2b707e63-9bbb-4244-adcd-2119a1c7187b\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:34.245267 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.245246 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccb482af-112d-452e-a16f-49fe93eedf37-metrics-tls\") pod \"dns-default-r495v\" (UID: \"ccb482af-112d-452e-a16f-49fe93eedf37\") " pod="openshift-dns/dns-default-r495v" Apr 21 10:04:34.245399 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.245380 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc037d14-2ee7-421f-875a-4ec33fac21bb-metrics-certs\") pod \"router-default-56b8857454-d4h69\" (UID: \"cc037d14-2ee7-421f-875a-4ec33fac21bb\") " pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:34.313013 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.312982 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-qvkrb\"" Apr 21 10:04:34.320765 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.320739 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:34.343797 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.343765 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:34.343953 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.343806 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:34.346323 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.346294 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eff9f25f-8452-4746-806c-70ba95f122a8-cert\") pod \"ingress-canary-b5lnk\" (UID: \"eff9f25f-8452-4746-806c-70ba95f122a8\") " pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:34.346426 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.346301 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb96ed3e-5f61-48e5-829d-2b68b67c919a-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-gfm8b\" (UID: \"eb96ed3e-5f61-48e5-829d-2b68b67c919a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:34.377897 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.377867 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-dockercfg-gdf9d\"" Apr 21 10:04:34.385850 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.385822 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" Apr 21 10:04:34.392514 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.392468 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-nglr4\"" Apr 21 10:04:34.400892 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.400370 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-r495v" Apr 21 10:04:34.401455 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.401435 2575 scope.go:117] "RemoveContainer" containerID="637ca51b2c409a3e55c32eefa0ba9e304581011c338f18a6696a64a6212a2932" Apr 21 10:04:34.428058 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.427387 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"default-dockercfg-xdcvf\"" Apr 21 10:04:34.434285 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.434245 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" Apr 21 10:04:34.443364 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.443334 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-5tz9l\"" Apr 21 10:04:34.457929 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.453413 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:34.464566 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.464466 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-6659b7df7b-wn6k4"] Apr 21 10:04:34.469428 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:34.469393 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod540077c6_a9f4_4a48_9f62_58c5f7d5c095.slice/crio-6d5b16a9b3f9101c3457ea90aa759de99a379a9eb7f553f6262a983ba4a8b8bb WatchSource:0}: Error finding container 6d5b16a9b3f9101c3457ea90aa759de99a379a9eb7f553f6262a983ba4a8b8bb: Status 404 returned error can't find the container with id 6d5b16a9b3f9101c3457ea90aa759de99a379a9eb7f553f6262a983ba4a8b8bb Apr 21 10:04:34.484379 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.483922 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-hvnc6\"" Apr 21 10:04:34.491748 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.491307 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" Apr 21 10:04:34.506109 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.505341 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-sjbvv\"" Apr 21 10:04:34.513893 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.510898 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b5lnk" Apr 21 10:04:34.621477 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.621273 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8"] Apr 21 10:04:34.646756 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.643934 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-r495v"] Apr 21 10:04:34.663004 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:34.662970 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccb482af_112d_452e_a16f_49fe93eedf37.slice/crio-7fa5d4b62a4ab09028c6d0ead4a2ef3c78d076b6e9719660c9c7062d4138904b WatchSource:0}: Error finding container 7fa5d4b62a4ab09028c6d0ead4a2ef3c78d076b6e9719660c9c7062d4138904b: Status 404 returned error can't find the container with id 7fa5d4b62a4ab09028c6d0ead4a2ef3c78d076b6e9719660c9c7062d4138904b Apr 21 10:04:34.679363 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.674749 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q"] Apr 21 10:04:34.680382 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:34.680343 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d46c6c5_053d_4431_833a_439ecc5708f6.slice/crio-5c22a1a2280680f3946ff8b911b19514c964fd5a695b3b412db868336ec375ac WatchSource:0}: Error finding container 5c22a1a2280680f3946ff8b911b19514c964fd5a695b3b412db868336ec375ac: Status 404 returned error can't find the container with id 5c22a1a2280680f3946ff8b911b19514c964fd5a695b3b412db868336ec375ac Apr 21 10:04:34.703158 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.702706 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/router-default-56b8857454-d4h69"] Apr 21 10:04:34.707892 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:34.707840 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc037d14_2ee7_421f_875a_4ec33fac21bb.slice/crio-8d5f9726d44ca41c7329f87cffda73034143c53f159c86e76df3da3eef9d011a WatchSource:0}: Error finding container 8d5f9726d44ca41c7329f87cffda73034143c53f159c86e76df3da3eef9d011a: Status 404 returned error can't find the container with id 8d5f9726d44ca41c7329f87cffda73034143c53f159c86e76df3da3eef9d011a Apr 21 10:04:34.736452 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.736337 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b"] Apr 21 10:04:34.749863 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.749790 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b5lnk"] Apr 21 10:04:34.753457 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:34.753422 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeff9f25f_8452_4746_806c_70ba95f122a8.slice/crio-66730b4523e36f3ef768f78b56a8fc9ecbd24c53bd7963d3cc1c2fc6ab11e549 WatchSource:0}: Error finding container 66730b4523e36f3ef768f78b56a8fc9ecbd24c53bd7963d3cc1c2fc6ab11e549: Status 404 returned error can't find the container with id 66730b4523e36f3ef768f78b56a8fc9ecbd24c53bd7963d3cc1c2fc6ab11e549 Apr 21 10:04:34.817390 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.816943 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-56b8857454-d4h69" event={"ID":"cc037d14-2ee7-421f-875a-4ec33fac21bb","Type":"ContainerStarted","Data":"0e1287002c2cb1a3f7f4f93cc7807c66dabc9c79cb4b84b1c4541e2be47b1dac"} Apr 21 10:04:34.817390 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.816988 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-56b8857454-d4h69" event={"ID":"cc037d14-2ee7-421f-875a-4ec33fac21bb","Type":"ContainerStarted","Data":"8d5f9726d44ca41c7329f87cffda73034143c53f159c86e76df3da3eef9d011a"} Apr 21 10:04:34.818646 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.818620 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" event={"ID":"eb96ed3e-5f61-48e5-829d-2b68b67c919a","Type":"ContainerStarted","Data":"8036f1a16e851081f03d4f2adf165c91ea739779f0c88baf75f2ef880945afa1"} Apr 21 10:04:34.819756 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.819720 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" event={"ID":"1d46c6c5-053d-4431-833a-439ecc5708f6","Type":"ContainerStarted","Data":"5c22a1a2280680f3946ff8b911b19514c964fd5a695b3b412db868336ec375ac"} Apr 21 10:04:34.821425 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.821402 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:04:34.821536 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.821515 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" event={"ID":"c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3","Type":"ContainerStarted","Data":"9eaa846353f35aaa3c9cad89e8c3ea4da032ab015de837b330eb0a772dad4f42"} Apr 21 10:04:34.821869 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.821837 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:34.822614 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.822590 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" event={"ID":"2b707e63-9bbb-4244-adcd-2119a1c7187b","Type":"ContainerStarted","Data":"686a21e9b9c62043d464d7c067dbb2a033d08cddc45dbe2ed2810ef3bd2256a3"} Apr 21 10:04:34.824580 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.824555 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b5lnk" event={"ID":"eff9f25f-8452-4746-806c-70ba95f122a8","Type":"ContainerStarted","Data":"66730b4523e36f3ef768f78b56a8fc9ecbd24c53bd7963d3cc1c2fc6ab11e549"} Apr 21 10:04:34.825830 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.825794 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-r495v" event={"ID":"ccb482af-112d-452e-a16f-49fe93eedf37","Type":"ContainerStarted","Data":"7fa5d4b62a4ab09028c6d0ead4a2ef3c78d076b6e9719660c9c7062d4138904b"} Apr 21 10:04:34.827140 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.827118 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" event={"ID":"540077c6-a9f4-4a48-9f62-58c5f7d5c095","Type":"ContainerStarted","Data":"937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72"} Apr 21 10:04:34.827237 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.827145 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" event={"ID":"540077c6-a9f4-4a48-9f62-58c5f7d5c095","Type":"ContainerStarted","Data":"6d5b16a9b3f9101c3457ea90aa759de99a379a9eb7f553f6262a983ba4a8b8bb"} Apr 21 10:04:34.827237 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.827225 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:04:34.840208 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.840166 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-56b8857454-d4h69" podStartSLOduration=62.840151078 podStartE2EDuration="1m2.840151078s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:04:34.839802526 +0000 UTC m=+66.000666878" watchObservedRunningTime="2026-04-21 10:04:34.840151078 +0000 UTC m=+66.001015425" Apr 21 10:04:34.860319 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.860254 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" podStartSLOduration=65.860231422 podStartE2EDuration="1m5.860231422s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:04:34.859835889 +0000 UTC m=+66.020700244" watchObservedRunningTime="2026-04-21 10:04:34.860231422 +0000 UTC m=+66.021095773" Apr 21 10:04:34.878498 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.878448 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" podStartSLOduration=50.617721461 podStartE2EDuration="1m2.878431117s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:03.606508645 +0000 UTC m=+34.767372973" lastFinishedPulling="2026-04-21 10:04:15.867218286 +0000 UTC m=+47.028082629" observedRunningTime="2026-04-21 10:04:34.877402622 +0000 UTC m=+66.038266972" watchObservedRunningTime="2026-04-21 10:04:34.878431117 +0000 UTC m=+66.039295467" Apr 21 10:04:34.878919 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:34.878898 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-9d4b6777b-mrf6q" Apr 21 10:04:35.157421 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.157319 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:04:35.160218 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.160188 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 21 10:04:35.170360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.170293 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9276c6c8-dd13-4c6c-a3d3-3318a3731823-metrics-certs\") pod \"network-metrics-daemon-n8tpq\" (UID: \"9276c6c8-dd13-4c6c-a3d3-3318a3731823\") " pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:04:35.322517 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.322475 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-fjqp7\"" Apr 21 10:04:35.330657 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.330631 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n8tpq" Apr 21 10:04:35.459258 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.459171 2575 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:35.463430 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.463215 2575 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:35.511480 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.511406 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n8tpq"] Apr 21 10:04:35.832435 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.832391 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:35.833875 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:35.833851 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-56b8857454-d4h69" Apr 21 10:04:36.018502 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:36.018452 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9276c6c8_dd13_4c6c_a3d3_3318a3731823.slice/crio-e1c922cafb99f212c170496366bb29ba4da5e3039a0583fa0bff961eccea1b83 WatchSource:0}: Error finding container e1c922cafb99f212c170496366bb29ba4da5e3039a0583fa0bff961eccea1b83: Status 404 returned error can't find the container with id e1c922cafb99f212c170496366bb29ba4da5e3039a0583fa0bff961eccea1b83 Apr 21 10:04:36.835980 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:36.835942 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n8tpq" event={"ID":"9276c6c8-dd13-4c6c-a3d3-3318a3731823","Type":"ContainerStarted","Data":"e1c922cafb99f212c170496366bb29ba4da5e3039a0583fa0bff961eccea1b83"} Apr 21 10:04:38.844720 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.844678 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n8tpq" event={"ID":"9276c6c8-dd13-4c6c-a3d3-3318a3731823","Type":"ContainerStarted","Data":"ba2cdd616fd3b1b857650d5d3e77365ceb9bfd4022a148ebb3db8c86d25d2c8a"} Apr 21 10:04:38.844720 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.844724 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n8tpq" event={"ID":"9276c6c8-dd13-4c6c-a3d3-3318a3731823","Type":"ContainerStarted","Data":"fc6b9ca4f108f323ad95ff528e887abc383bbf5852ffa70ac74ac83596498a3b"} Apr 21 10:04:38.846487 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.846446 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" event={"ID":"2b707e63-9bbb-4244-adcd-2119a1c7187b","Type":"ContainerStarted","Data":"e459b7e5252da2364be8c296ab8866cd42518dffbfb37a9c64c06ff1473ffdcc"} Apr 21 10:04:38.847909 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.847881 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b5lnk" event={"ID":"eff9f25f-8452-4746-806c-70ba95f122a8","Type":"ContainerStarted","Data":"08d71f73e876119446a4effdd56cf186f7d4cea996aa67891587be6b29faabbb"} Apr 21 10:04:38.849517 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.849492 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-r495v" event={"ID":"ccb482af-112d-452e-a16f-49fe93eedf37","Type":"ContainerStarted","Data":"37a722f3d1c2aee826a3dcc3c5946fb7943867799b39840b8a2af74664d64c25"} Apr 21 10:04:38.849628 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.849525 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-r495v" event={"ID":"ccb482af-112d-452e-a16f-49fe93eedf37","Type":"ContainerStarted","Data":"866a7669c2e72cf4aaba683dd116cb8b84a6749b708338798c74eaefa5af4ff7"} Apr 21 10:04:38.849628 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.849581 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-r495v" Apr 21 10:04:38.851173 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.851142 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" event={"ID":"eb96ed3e-5f61-48e5-829d-2b68b67c919a","Type":"ContainerStarted","Data":"24d177be1bb59638bf183c1d0df3fd1526d874786827cd2948bfae194113aa75"} Apr 21 10:04:38.851268 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.851173 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" event={"ID":"eb96ed3e-5f61-48e5-829d-2b68b67c919a","Type":"ContainerStarted","Data":"69a83d116dff9feccb9bcc3d5e7e5093f21081ef40c3207ee20b277445ac9f1f"} Apr 21 10:04:38.852537 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.852512 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" event={"ID":"1d46c6c5-053d-4431-833a-439ecc5708f6","Type":"ContainerStarted","Data":"633000ea0b8a66bebb670f18fc1d342f7a299975898917aa3d5ce762d50999d9"} Apr 21 10:04:38.860848 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.860804 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-n8tpq" podStartSLOduration=67.558451612 podStartE2EDuration="1m9.860791867s" podCreationTimestamp="2026-04-21 10:03:29 +0000 UTC" firstStartedPulling="2026-04-21 10:04:36.023875375 +0000 UTC m=+67.184739718" lastFinishedPulling="2026-04-21 10:04:38.326215642 +0000 UTC m=+69.487079973" observedRunningTime="2026-04-21 10:04:38.860440732 +0000 UTC m=+70.021305084" watchObservedRunningTime="2026-04-21 10:04:38.860791867 +0000 UTC m=+70.021656223" Apr 21 10:04:38.876606 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.876558 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b5lnk" podStartSLOduration=33.381980225 podStartE2EDuration="36.876543874s" podCreationTimestamp="2026-04-21 10:04:02 +0000 UTC" firstStartedPulling="2026-04-21 10:04:34.755063183 +0000 UTC m=+65.915927514" lastFinishedPulling="2026-04-21 10:04:38.249626821 +0000 UTC m=+69.410491163" observedRunningTime="2026-04-21 10:04:38.874962825 +0000 UTC m=+70.035827177" watchObservedRunningTime="2026-04-21 10:04:38.876543874 +0000 UTC m=+70.037408224" Apr 21 10:04:38.892499 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.892445 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-gfm8b" podStartSLOduration=63.442410734 podStartE2EDuration="1m6.892427364s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:34.799563392 +0000 UTC m=+65.960427734" lastFinishedPulling="2026-04-21 10:04:38.249580031 +0000 UTC m=+69.410444364" observedRunningTime="2026-04-21 10:04:38.891384039 +0000 UTC m=+70.052248390" watchObservedRunningTime="2026-04-21 10:04:38.892427364 +0000 UTC m=+70.053291716" Apr 21 10:04:38.907331 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.906443 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-mscz8" podStartSLOduration=63.291056345 podStartE2EDuration="1m6.906426798s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:34.634349145 +0000 UTC m=+65.795213474" lastFinishedPulling="2026-04-21 10:04:38.249719587 +0000 UTC m=+69.410583927" observedRunningTime="2026-04-21 10:04:38.90581372 +0000 UTC m=+70.066678074" watchObservedRunningTime="2026-04-21 10:04:38.906426798 +0000 UTC m=+70.067291153" Apr 21 10:04:38.923588 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.923544 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-r495v" podStartSLOduration=33.339928486 podStartE2EDuration="36.92352939s" podCreationTimestamp="2026-04-21 10:04:02 +0000 UTC" firstStartedPulling="2026-04-21 10:04:34.666026587 +0000 UTC m=+65.826890919" lastFinishedPulling="2026-04-21 10:04:38.249627485 +0000 UTC m=+69.410491823" observedRunningTime="2026-04-21 10:04:38.921638418 +0000 UTC m=+70.082502768" watchObservedRunningTime="2026-04-21 10:04:38.92352939 +0000 UTC m=+70.084393740" Apr 21 10:04:38.937793 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:38.937746 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-console/networking-console-plugin-cb95c66f6-5mv9q" podStartSLOduration=63.370982834 podStartE2EDuration="1m6.93773146s" podCreationTimestamp="2026-04-21 10:03:32 +0000 UTC" firstStartedPulling="2026-04-21 10:04:34.683017439 +0000 UTC m=+65.843881781" lastFinishedPulling="2026-04-21 10:04:38.249766074 +0000 UTC m=+69.410630407" observedRunningTime="2026-04-21 10:04:38.93622669 +0000 UTC m=+70.097091040" watchObservedRunningTime="2026-04-21 10:04:38.93773146 +0000 UTC m=+70.098595809" Apr 21 10:04:42.749286 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:42.749245 2575 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" podUID="8372d417-f5b6-4e27-86b6-8c2b69e1103f" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 21 10:04:43.008577 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.008494 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-m6sp6"] Apr 21 10:04:43.066240 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.066188 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-m6sp6"] Apr 21 10:04:43.066428 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.066388 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.069305 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.069277 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 21 10:04:43.070363 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.070338 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 21 10:04:43.070485 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.070456 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-l8m2h\"" Apr 21 10:04:43.121764 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.121734 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e686f946-4a0a-455c-b0b1-444547a37100-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.121945 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.121794 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e686f946-4a0a-455c-b0b1-444547a37100-crio-socket\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.121945 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.121815 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e686f946-4a0a-455c-b0b1-444547a37100-data-volume\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.121945 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.121835 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e686f946-4a0a-455c-b0b1-444547a37100-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.121945 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.121859 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2wnd\" (UniqueName: \"kubernetes.io/projected/e686f946-4a0a-455c-b0b1-444547a37100-kube-api-access-k2wnd\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.124868 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.124843 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-6bcc868b7-2dq9x"] Apr 21 10:04:43.143820 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.143790 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-6bcc868b7-2dq9x" Apr 21 10:04:43.147368 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.147331 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 21 10:04:43.147500 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.147484 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-ntfjv\"" Apr 21 10:04:43.147592 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.147575 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 21 10:04:43.148747 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.148729 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j"] Apr 21 10:04:43.170819 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.170778 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-6bcc868b7-2dq9x"] Apr 21 10:04:43.170819 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.170815 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j"] Apr 21 10:04:43.171001 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.170923 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" Apr 21 10:04:43.173403 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.173377 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-admission-webhook-dockercfg-tll6m\"" Apr 21 10:04:43.173605 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.173588 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-admission-webhook-tls\"" Apr 21 10:04:43.223119 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223056 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e686f946-4a0a-455c-b0b1-444547a37100-crio-socket\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.223119 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223121 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e686f946-4a0a-455c-b0b1-444547a37100-data-volume\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.223360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223155 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e686f946-4a0a-455c-b0b1-444547a37100-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.223360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223191 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84f9p\" (UniqueName: \"kubernetes.io/projected/e80439e8-0f75-4677-94bd-88728581a3b0-kube-api-access-84f9p\") pod \"downloads-6bcc868b7-2dq9x\" (UID: \"e80439e8-0f75-4677-94bd-88728581a3b0\") " pod="openshift-console/downloads-6bcc868b7-2dq9x" Apr 21 10:04:43.223360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223191 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/e686f946-4a0a-455c-b0b1-444547a37100-crio-socket\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.223360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223226 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-k2wnd\" (UniqueName: \"kubernetes.io/projected/e686f946-4a0a-455c-b0b1-444547a37100-kube-api-access-k2wnd\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.223360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223304 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e686f946-4a0a-455c-b0b1-444547a37100-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.223677 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223357 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/e472dacd-629f-4f47-a0de-d0c4b1ad5701-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-j962j\" (UID: \"e472dacd-629f-4f47-a0de-d0c4b1ad5701\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" Apr 21 10:04:43.223677 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223534 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/e686f946-4a0a-455c-b0b1-444547a37100-data-volume\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.223951 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.223900 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/e686f946-4a0a-455c-b0b1-444547a37100-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.225623 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.225602 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/e686f946-4a0a-455c-b0b1-444547a37100-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.234527 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.234471 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2wnd\" (UniqueName: \"kubernetes.io/projected/e686f946-4a0a-455c-b0b1-444547a37100-kube-api-access-k2wnd\") pod \"insights-runtime-extractor-m6sp6\" (UID: \"e686f946-4a0a-455c-b0b1-444547a37100\") " pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.324049 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.324012 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-84f9p\" (UniqueName: \"kubernetes.io/projected/e80439e8-0f75-4677-94bd-88728581a3b0-kube-api-access-84f9p\") pod \"downloads-6bcc868b7-2dq9x\" (UID: \"e80439e8-0f75-4677-94bd-88728581a3b0\") " pod="openshift-console/downloads-6bcc868b7-2dq9x" Apr 21 10:04:43.324250 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.324072 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/e472dacd-629f-4f47-a0de-d0c4b1ad5701-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-j962j\" (UID: \"e472dacd-629f-4f47-a0de-d0c4b1ad5701\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" Apr 21 10:04:43.326444 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.326411 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/e472dacd-629f-4f47-a0de-d0c4b1ad5701-tls-certificates\") pod \"prometheus-operator-admission-webhook-57cf98b594-j962j\" (UID: \"e472dacd-629f-4f47-a0de-d0c4b1ad5701\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" Apr 21 10:04:43.336497 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.336469 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-84f9p\" (UniqueName: \"kubernetes.io/projected/e80439e8-0f75-4677-94bd-88728581a3b0-kube-api-access-84f9p\") pod \"downloads-6bcc868b7-2dq9x\" (UID: \"e80439e8-0f75-4677-94bd-88728581a3b0\") " pod="openshift-console/downloads-6bcc868b7-2dq9x" Apr 21 10:04:43.376584 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.376554 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-m6sp6" Apr 21 10:04:43.454588 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.454501 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-6bcc868b7-2dq9x" Apr 21 10:04:43.480253 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.480207 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" Apr 21 10:04:43.516851 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.515980 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-m6sp6"] Apr 21 10:04:43.518036 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:43.518002 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode686f946_4a0a_455c_b0b1_444547a37100.slice/crio-e9c84e3246f0aac39748efaa5600f4819f7e3bd44bd2d31ab50c56569a54c7eb WatchSource:0}: Error finding container e9c84e3246f0aac39748efaa5600f4819f7e3bd44bd2d31ab50c56569a54c7eb: Status 404 returned error can't find the container with id e9c84e3246f0aac39748efaa5600f4819f7e3bd44bd2d31ab50c56569a54c7eb Apr 21 10:04:43.611353 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.611317 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-6bcc868b7-2dq9x"] Apr 21 10:04:43.613919 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:43.613881 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode80439e8_0f75_4677_94bd_88728581a3b0.slice/crio-b30fcc49a1d018ef263070a0ac5fb0754938e94952e830596716768a8631ea8e WatchSource:0}: Error finding container b30fcc49a1d018ef263070a0ac5fb0754938e94952e830596716768a8631ea8e: Status 404 returned error can't find the container with id b30fcc49a1d018ef263070a0ac5fb0754938e94952e830596716768a8631ea8e Apr 21 10:04:43.640859 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.640831 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j"] Apr 21 10:04:43.644917 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:43.644886 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode472dacd_629f_4f47_a0de_d0c4b1ad5701.slice/crio-bdeb3d1b7c6d712ed4e25b26b34dce6c91cbebb1c7cf0e754b3e12411cbacfa7 WatchSource:0}: Error finding container bdeb3d1b7c6d712ed4e25b26b34dce6c91cbebb1c7cf0e754b3e12411cbacfa7: Status 404 returned error can't find the container with id bdeb3d1b7c6d712ed4e25b26b34dce6c91cbebb1c7cf0e754b3e12411cbacfa7 Apr 21 10:04:43.874836 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.874723 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-m6sp6" event={"ID":"e686f946-4a0a-455c-b0b1-444547a37100","Type":"ContainerStarted","Data":"029118e549ab82cbd4cb9bc9ed5151da6842508f09f57ced0da1004392447899"} Apr 21 10:04:43.874836 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.874775 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-m6sp6" event={"ID":"e686f946-4a0a-455c-b0b1-444547a37100","Type":"ContainerStarted","Data":"e9c84e3246f0aac39748efaa5600f4819f7e3bd44bd2d31ab50c56569a54c7eb"} Apr 21 10:04:43.875866 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.875836 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" event={"ID":"e472dacd-629f-4f47-a0de-d0c4b1ad5701","Type":"ContainerStarted","Data":"bdeb3d1b7c6d712ed4e25b26b34dce6c91cbebb1c7cf0e754b3e12411cbacfa7"} Apr 21 10:04:43.876844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:43.876821 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-6bcc868b7-2dq9x" event={"ID":"e80439e8-0f75-4677-94bd-88728581a3b0","Type":"ContainerStarted","Data":"b30fcc49a1d018ef263070a0ac5fb0754938e94952e830596716768a8631ea8e"} Apr 21 10:04:45.886995 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:45.886956 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-m6sp6" event={"ID":"e686f946-4a0a-455c-b0b1-444547a37100","Type":"ContainerStarted","Data":"4bcda2933ba06bed08bf835afc1457f8c01ee11aaa1db56bdcbbd950660b1256"} Apr 21 10:04:45.888568 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:45.888533 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" event={"ID":"e472dacd-629f-4f47-a0de-d0c4b1ad5701","Type":"ContainerStarted","Data":"99d07a647fe09d6d2bbf3e46824a4c185a8497fb31254402cab31e126400f3ad"} Apr 21 10:04:45.888825 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:45.888793 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" Apr 21 10:04:45.894246 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:45.894222 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" Apr 21 10:04:45.904908 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:45.904840 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-57cf98b594-j962j" podStartSLOduration=1.147074583 podStartE2EDuration="2.904821465s" podCreationTimestamp="2026-04-21 10:04:43 +0000 UTC" firstStartedPulling="2026-04-21 10:04:43.649641678 +0000 UTC m=+74.810506006" lastFinishedPulling="2026-04-21 10:04:45.407388547 +0000 UTC m=+76.568252888" observedRunningTime="2026-04-21 10:04:45.902878233 +0000 UTC m=+77.063742585" watchObservedRunningTime="2026-04-21 10:04:45.904821465 +0000 UTC m=+77.065685816" Apr 21 10:04:47.759066 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:47.758979 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-cnk4k" Apr 21 10:04:47.896765 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:47.896729 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-m6sp6" event={"ID":"e686f946-4a0a-455c-b0b1-444547a37100","Type":"ContainerStarted","Data":"724c23b3cfdd60d4946c99cbf64ca7a608851376220eb6bca436bae72b511726"} Apr 21 10:04:47.912947 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:47.912894 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-m6sp6" podStartSLOduration=2.19976417 podStartE2EDuration="5.912879535s" podCreationTimestamp="2026-04-21 10:04:42 +0000 UTC" firstStartedPulling="2026-04-21 10:04:43.65504052 +0000 UTC m=+74.815904849" lastFinishedPulling="2026-04-21 10:04:47.368155877 +0000 UTC m=+78.529020214" observedRunningTime="2026-04-21 10:04:47.912612859 +0000 UTC m=+79.073477223" watchObservedRunningTime="2026-04-21 10:04:47.912879535 +0000 UTC m=+79.073743889" Apr 21 10:04:48.859771 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:48.859539 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-r495v" Apr 21 10:04:51.260041 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.258937 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-qh8t6"] Apr 21 10:04:51.299965 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.299906 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.303803 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.303776 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 21 10:04:51.303963 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.303776 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 21 10:04:51.304352 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.304319 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 21 10:04:51.304475 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.304379 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 21 10:04:51.304533 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.304334 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-48j4v\"" Apr 21 10:04:51.400252 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400219 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-textfile\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.400445 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400275 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7ddf49bd-b530-4121-814e-61f13607452d-metrics-client-ca\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.400445 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400304 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-accelerators-collector-config\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.400445 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400335 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-wtmp\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.400445 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400361 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-sys\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.400445 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400404 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-tls\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.400445 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400430 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5lrj\" (UniqueName: \"kubernetes.io/projected/7ddf49bd-b530-4121-814e-61f13607452d-kube-api-access-t5lrj\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.400748 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400463 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.400748 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.400491 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-root\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501488 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501450 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-tls\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501662 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501505 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-t5lrj\" (UniqueName: \"kubernetes.io/projected/7ddf49bd-b530-4121-814e-61f13607452d-kube-api-access-t5lrj\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501662 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501544 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501662 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501573 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-root\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501662 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501632 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-textfile\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501877 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501672 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7ddf49bd-b530-4121-814e-61f13607452d-metrics-client-ca\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501877 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501698 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-accelerators-collector-config\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501877 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501729 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-wtmp\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501877 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501757 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-sys\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.501877 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.501843 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-sys\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.502472 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.502272 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-textfile\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.502844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.502658 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-root\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.502844 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.502799 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-wtmp\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.503652 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.503626 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-accelerators-collector-config\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.503760 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.503677 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7ddf49bd-b530-4121-814e-61f13607452d-metrics-client-ca\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.505555 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.505522 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-tls\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.506267 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.506242 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/7ddf49bd-b530-4121-814e-61f13607452d-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.510616 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.510562 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5lrj\" (UniqueName: \"kubernetes.io/projected/7ddf49bd-b530-4121-814e-61f13607452d-kube-api-access-t5lrj\") pod \"node-exporter-qh8t6\" (UID: \"7ddf49bd-b530-4121-814e-61f13607452d\") " pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.613547 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.613508 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-qh8t6" Apr 21 10:04:51.625970 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:04:51.625934 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ddf49bd_b530_4121_814e_61f13607452d.slice/crio-2c07c4d4819c76efe0125018f0b215261340e65f93bb10e0269ff7807dae9c4e WatchSource:0}: Error finding container 2c07c4d4819c76efe0125018f0b215261340e65f93bb10e0269ff7807dae9c4e: Status 404 returned error can't find the container with id 2c07c4d4819c76efe0125018f0b215261340e65f93bb10e0269ff7807dae9c4e Apr 21 10:04:51.911879 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:51.911836 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-qh8t6" event={"ID":"7ddf49bd-b530-4121-814e-61f13607452d","Type":"ContainerStarted","Data":"2c07c4d4819c76efe0125018f0b215261340e65f93bb10e0269ff7807dae9c4e"} Apr 21 10:04:52.749240 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:52.749194 2575 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" podUID="8372d417-f5b6-4e27-86b6-8c2b69e1103f" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 21 10:04:52.749645 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:52.749274 2575 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" Apr 21 10:04:52.749968 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:52.749917 2575 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="service-proxy" containerStatusID={"Type":"cri-o","ID":"389686ffcb5e09301c5854ef7868b7b9c676e9728971930801e0a06a532c4676"} pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" containerMessage="Container service-proxy failed liveness probe, will be restarted" Apr 21 10:04:52.750090 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:52.750003 2575 kuberuntime_container.go:864] "Killing container with a grace period" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" podUID="8372d417-f5b6-4e27-86b6-8c2b69e1103f" containerName="service-proxy" containerID="cri-o://389686ffcb5e09301c5854ef7868b7b9c676e9728971930801e0a06a532c4676" gracePeriod=30 Apr 21 10:04:52.919052 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:52.919016 2575 generic.go:358] "Generic (PLEG): container finished" podID="8372d417-f5b6-4e27-86b6-8c2b69e1103f" containerID="389686ffcb5e09301c5854ef7868b7b9c676e9728971930801e0a06a532c4676" exitCode=2 Apr 21 10:04:52.919235 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:52.919126 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" event={"ID":"8372d417-f5b6-4e27-86b6-8c2b69e1103f","Type":"ContainerDied","Data":"389686ffcb5e09301c5854ef7868b7b9c676e9728971930801e0a06a532c4676"} Apr 21 10:04:53.925442 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:53.925388 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-799b464b89-gxrxn" event={"ID":"8372d417-f5b6-4e27-86b6-8c2b69e1103f","Type":"ContainerStarted","Data":"358f90301586016ea41bb62d1d12c05a412dde7a01cc2c3e1b53ab234b077b9a"} Apr 21 10:04:53.927169 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:53.927133 2575 generic.go:358] "Generic (PLEG): container finished" podID="7ddf49bd-b530-4121-814e-61f13607452d" containerID="c8a4ff1f0e68b82284290bb3bdce3a503828e92e116287b18afa2ffdd8d89e02" exitCode=0 Apr 21 10:04:53.927319 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:53.927222 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-qh8t6" event={"ID":"7ddf49bd-b530-4121-814e-61f13607452d","Type":"ContainerDied","Data":"c8a4ff1f0e68b82284290bb3bdce3a503828e92e116287b18afa2ffdd8d89e02"} Apr 21 10:04:54.325466 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:54.325423 2575 patch_prober.go:28] interesting pod/image-registry-6659b7df7b-wn6k4 container/registry namespace/openshift-image-registry: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 21 10:04:54.325633 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:54.325493 2575 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" podUID="540077c6-a9f4-4a48-9f62-58c5f7d5c095" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 21 10:04:55.836627 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:04:55.836598 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:05:00.953138 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:00.953071 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-6bcc868b7-2dq9x" event={"ID":"e80439e8-0f75-4677-94bd-88728581a3b0","Type":"ContainerStarted","Data":"2b682d1ff41b324dc9a72c84d84bc6819cf3231b51d1716e8bf53c26ba5ac626"} Apr 21 10:05:00.953628 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:00.953268 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/downloads-6bcc868b7-2dq9x" Apr 21 10:05:00.955828 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:00.955801 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-qh8t6" event={"ID":"7ddf49bd-b530-4121-814e-61f13607452d","Type":"ContainerStarted","Data":"b98b96bd6cdc3622732144c60f8d9758426a53c2bb7cd2bbafbd243d4c845315"} Apr 21 10:05:00.955949 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:00.955835 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-qh8t6" event={"ID":"7ddf49bd-b530-4121-814e-61f13607452d","Type":"ContainerStarted","Data":"18421daeb30a1cdb1fcdea71d326a96f3998758dcc864a2b2e22cf64583b6afa"} Apr 21 10:05:00.968802 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:00.968777 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-6bcc868b7-2dq9x" Apr 21 10:05:00.972636 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:00.972585 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-6bcc868b7-2dq9x" podStartSLOduration=1.345274925 podStartE2EDuration="17.972570443s" podCreationTimestamp="2026-04-21 10:04:43 +0000 UTC" firstStartedPulling="2026-04-21 10:04:43.615777461 +0000 UTC m=+74.776641805" lastFinishedPulling="2026-04-21 10:05:00.243072992 +0000 UTC m=+91.403937323" observedRunningTime="2026-04-21 10:05:00.971268249 +0000 UTC m=+92.132132600" watchObservedRunningTime="2026-04-21 10:05:00.972570443 +0000 UTC m=+92.133434794" Apr 21 10:05:00.990183 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:00.990127 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-qh8t6" podStartSLOduration=8.749647067 podStartE2EDuration="9.990111657s" podCreationTimestamp="2026-04-21 10:04:51 +0000 UTC" firstStartedPulling="2026-04-21 10:04:51.628378163 +0000 UTC m=+82.789242505" lastFinishedPulling="2026-04-21 10:04:52.86884276 +0000 UTC m=+84.029707095" observedRunningTime="2026-04-21 10:05:00.989462749 +0000 UTC m=+92.150327101" watchObservedRunningTime="2026-04-21 10:05:00.990111657 +0000 UTC m=+92.150976005" Apr 21 10:05:05.580818 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:05.580779 2575 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-6659b7df7b-wn6k4"] Apr 21 10:05:28.041214 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:28.041176 2575 generic.go:358] "Generic (PLEG): container finished" podID="0057318f-db18-4772-bd50-0cf37022dc7c" containerID="debb7ab0f1dc13ab8d8cabd6f5fd989bc5812489d61ebc616c644a83575f2ba4" exitCode=0 Apr 21 10:05:28.041620 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:28.041227 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" event={"ID":"0057318f-db18-4772-bd50-0cf37022dc7c","Type":"ContainerDied","Data":"debb7ab0f1dc13ab8d8cabd6f5fd989bc5812489d61ebc616c644a83575f2ba4"} Apr 21 10:05:28.041620 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:28.041586 2575 scope.go:117] "RemoveContainer" containerID="debb7ab0f1dc13ab8d8cabd6f5fd989bc5812489d61ebc616c644a83575f2ba4" Apr 21 10:05:29.045277 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:29.045242 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-lpqct" event={"ID":"0057318f-db18-4772-bd50-0cf37022dc7c","Type":"ContainerStarted","Data":"95496f0c6e48325d22d6931a613680e4cedff9152aa19b9cd15843595383700e"} Apr 21 10:05:30.604599 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.604533 2575 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" podUID="540077c6-a9f4-4a48-9f62-58c5f7d5c095" containerName="registry" containerID="cri-o://937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72" gracePeriod=30 Apr 21 10:05:30.874011 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.873986 2575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:05:30.950611 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.950575 2575 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") pod \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " Apr 21 10:05:30.950811 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.950636 2575 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-trusted-ca\") pod \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " Apr 21 10:05:30.950811 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.950676 2575 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-installation-pull-secrets\") pod \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " Apr 21 10:05:30.950811 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.950703 2575 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/540077c6-a9f4-4a48-9f62-58c5f7d5c095-ca-trust-extracted\") pod \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " Apr 21 10:05:30.950811 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.950735 2575 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f88tn\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-kube-api-access-f88tn\") pod \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " Apr 21 10:05:30.950811 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.950758 2575 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-certificates\") pod \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " Apr 21 10:05:30.951242 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.951135 2575 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "540077c6-a9f4-4a48-9f62-58c5f7d5c095" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:30.951371 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.951345 2575 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "540077c6-a9f4-4a48-9f62-58c5f7d5c095" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 21 10:05:30.953289 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.953240 2575 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-kube-api-access-f88tn" (OuterVolumeSpecName: "kube-api-access-f88tn") pod "540077c6-a9f4-4a48-9f62-58c5f7d5c095" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095"). InnerVolumeSpecName "kube-api-access-f88tn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:05:30.953407 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.953314 2575 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "540077c6-a9f4-4a48-9f62-58c5f7d5c095" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:05:30.953407 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.953317 2575 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "540077c6-a9f4-4a48-9f62-58c5f7d5c095" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:05:30.959887 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:30.959858 2575 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/540077c6-a9f4-4a48-9f62-58c5f7d5c095-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "540077c6-a9f4-4a48-9f62-58c5f7d5c095" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 21 10:05:31.051621 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.051584 2575 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-bound-sa-token\") pod \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " Apr 21 10:05:31.051811 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.051644 2575 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-image-registry-private-configuration\") pod \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\" (UID: \"540077c6-a9f4-4a48-9f62-58c5f7d5c095\") " Apr 21 10:05:31.051886 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.051872 2575 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-installation-pull-secrets\") on node \"ip-10-0-142-73.ec2.internal\" DevicePath \"\"" Apr 21 10:05:31.051933 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.051892 2575 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/540077c6-a9f4-4a48-9f62-58c5f7d5c095-ca-trust-extracted\") on node \"ip-10-0-142-73.ec2.internal\" DevicePath \"\"" Apr 21 10:05:31.051933 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.051905 2575 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-f88tn\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-kube-api-access-f88tn\") on node \"ip-10-0-142-73.ec2.internal\" DevicePath \"\"" Apr 21 10:05:31.051933 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.051919 2575 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-certificates\") on node \"ip-10-0-142-73.ec2.internal\" DevicePath \"\"" Apr 21 10:05:31.052069 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.051932 2575 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-registry-tls\") on node \"ip-10-0-142-73.ec2.internal\" DevicePath \"\"" Apr 21 10:05:31.052069 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.051947 2575 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/540077c6-a9f4-4a48-9f62-58c5f7d5c095-trusted-ca\") on node \"ip-10-0-142-73.ec2.internal\" DevicePath \"\"" Apr 21 10:05:31.052774 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.052746 2575 generic.go:358] "Generic (PLEG): container finished" podID="540077c6-a9f4-4a48-9f62-58c5f7d5c095" containerID="937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72" exitCode=0 Apr 21 10:05:31.052888 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.052808 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" event={"ID":"540077c6-a9f4-4a48-9f62-58c5f7d5c095","Type":"ContainerDied","Data":"937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72"} Apr 21 10:05:31.052888 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.052812 2575 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" Apr 21 10:05:31.052888 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.052842 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-6659b7df7b-wn6k4" event={"ID":"540077c6-a9f4-4a48-9f62-58c5f7d5c095","Type":"ContainerDied","Data":"6d5b16a9b3f9101c3457ea90aa759de99a379a9eb7f553f6262a983ba4a8b8bb"} Apr 21 10:05:31.052888 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.052864 2575 scope.go:117] "RemoveContainer" containerID="937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72" Apr 21 10:05:31.054060 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.054034 2575 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "540077c6-a9f4-4a48-9f62-58c5f7d5c095" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 21 10:05:31.054152 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.054034 2575 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "540077c6-a9f4-4a48-9f62-58c5f7d5c095" (UID: "540077c6-a9f4-4a48-9f62-58c5f7d5c095"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 21 10:05:31.060896 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.060878 2575 scope.go:117] "RemoveContainer" containerID="937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72" Apr 21 10:05:31.061189 ip-10-0-142-73 kubenswrapper[2575]: E0421 10:05:31.061159 2575 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72\": container with ID starting with 937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72 not found: ID does not exist" containerID="937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72" Apr 21 10:05:31.061245 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.061199 2575 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72"} err="failed to get container status \"937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72\": rpc error: code = NotFound desc = could not find container \"937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72\": container with ID starting with 937dc0a895c54dd68bf1dcc96d8c48474db8c3348bc1054e8c41693282ab9e72 not found: ID does not exist" Apr 21 10:05:31.152369 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.152285 2575 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/540077c6-a9f4-4a48-9f62-58c5f7d5c095-bound-sa-token\") on node \"ip-10-0-142-73.ec2.internal\" DevicePath \"\"" Apr 21 10:05:31.152369 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.152313 2575 reconciler_common.go:299] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/540077c6-a9f4-4a48-9f62-58c5f7d5c095-image-registry-private-configuration\") on node \"ip-10-0-142-73.ec2.internal\" DevicePath \"\"" Apr 21 10:05:31.373517 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.373478 2575 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-6659b7df7b-wn6k4"] Apr 21 10:05:31.375247 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.375219 2575 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-6659b7df7b-wn6k4"] Apr 21 10:05:31.405424 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:31.405351 2575 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="540077c6-a9f4-4a48-9f62-58c5f7d5c095" path="/var/lib/kubelet/pods/540077c6-a9f4-4a48-9f62-58c5f7d5c095/volumes" Apr 21 10:05:47.100014 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:47.099980 2575 generic.go:358] "Generic (PLEG): container finished" podID="91ca9a98-134d-41d4-95a3-460d18d5986b" containerID="28854bbff905a8f63604b9f5b2b053d8313b222005e26ade10f746cac4550c47" exitCode=0 Apr 21 10:05:47.100473 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:47.100034 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" event={"ID":"91ca9a98-134d-41d4-95a3-460d18d5986b","Type":"ContainerDied","Data":"28854bbff905a8f63604b9f5b2b053d8313b222005e26ade10f746cac4550c47"} Apr 21 10:05:47.100473 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:47.100380 2575 scope.go:117] "RemoveContainer" containerID="28854bbff905a8f63604b9f5b2b053d8313b222005e26ade10f746cac4550c47" Apr 21 10:05:48.103735 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:48.103701 2575 generic.go:358] "Generic (PLEG): container finished" podID="f7a64cde-eca1-4b8d-843b-90f038244fe0" containerID="2364df08b9966ff5c04dee6b98a471190febf5f7b7bb666a390a37c9e988d3ce" exitCode=0 Apr 21 10:05:48.104209 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:48.103776 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-4hz8p" event={"ID":"f7a64cde-eca1-4b8d-843b-90f038244fe0","Type":"ContainerDied","Data":"2364df08b9966ff5c04dee6b98a471190febf5f7b7bb666a390a37c9e988d3ce"} Apr 21 10:05:48.104209 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:48.104181 2575 scope.go:117] "RemoveContainer" containerID="2364df08b9966ff5c04dee6b98a471190febf5f7b7bb666a390a37c9e988d3ce" Apr 21 10:05:48.105669 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:48.105604 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-7gxkl" event={"ID":"91ca9a98-134d-41d4-95a3-460d18d5986b","Type":"ContainerStarted","Data":"922bc23d6abcb55af818ee9b784bc33b3a8372679661423a865f48785dca51e4"} Apr 21 10:05:49.110253 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:05:49.110212 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-4hz8p" event={"ID":"f7a64cde-eca1-4b8d-843b-90f038244fe0","Type":"ContainerStarted","Data":"fb9ae26d2004ee7addb76368cf3d776d5cc0f346b0487e8fc4b2cbe0ac13eef8"} Apr 21 10:08:29.256144 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:08:29.256084 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:08:29.256674 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:08:29.256454 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:08:29.266253 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:08:29.266228 2575 kubelet.go:1628] "Image garbage collection succeeded" Apr 21 10:13:29.282717 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:13:29.282684 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:13:29.283157 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:13:29.283026 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:18:29.305433 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:18:29.305351 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:18:29.307422 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:18:29.307391 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:23:29.325755 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:23:29.325717 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:23:29.334127 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:23:29.334086 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:28:29.346690 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:28:29.346652 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:28:29.359849 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:28:29.359820 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:33:29.372194 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:33:29.372059 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:33:29.381301 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:33:29.381279 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:38:29.393260 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:38:29.393152 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:38:29.402022 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:38:29.401998 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:43:29.414988 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:43:29.414890 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:43:29.422965 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:43:29.422946 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:48:29.436874 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:48:29.436769 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:48:29.443781 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:48:29.443760 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:51:32.967881 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:32.967790 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-t8wp8_37ae773e-f021-4a13-81c5-678bcf84335e/global-pull-secret-syncer/0.log" Apr 21 10:51:33.030658 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:33.030622 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-l7wqf_0a0fa3aa-3e6a-4ae6-b76a-43b4bb05c381/konnectivity-agent/0.log" Apr 21 10:51:33.239757 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:33.239674 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-142-73.ec2.internal_dbfdb56ede086eca491c9eac6de95d77/haproxy/0.log" Apr 21 10:51:36.533485 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:36.533450 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-mscz8_2b707e63-9bbb-4244-adcd-2119a1c7187b/cluster-monitoring-operator/0.log" Apr 21 10:51:36.781619 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:36.781541 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-qh8t6_7ddf49bd-b530-4121-814e-61f13607452d/node-exporter/0.log" Apr 21 10:51:36.801757 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:36.801721 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-qh8t6_7ddf49bd-b530-4121-814e-61f13607452d/kube-rbac-proxy/0.log" Apr 21 10:51:36.825878 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:36.825851 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-qh8t6_7ddf49bd-b530-4121-814e-61f13607452d/init-textfile/0.log" Apr 21 10:51:37.233238 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:37.233207 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-admission-webhook-57cf98b594-j962j_e472dacd-629f-4f47-a0de-d0c4b1ad5701/prometheus-operator-admission-webhook/0.log" Apr 21 10:51:38.591165 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:38.591134 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-console_networking-console-plugin-cb95c66f6-5mv9q_1d46c6c5-053d-4431-833a-439ecc5708f6/networking-console-plugin/0.log" Apr 21 10:51:39.001633 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:39.001546 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/1.log" Apr 21 10:51:39.011029 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:39.010996 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-mrf6q_c8891fb4-b21c-41fd-9b9a-1add8bc8f1c3/console-operator/2.log" Apr 21 10:51:39.405898 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:39.405870 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_downloads-6bcc868b7-2dq9x_e80439e8-0f75-4677-94bd-88728581a3b0/download-server/0.log" Apr 21 10:51:39.790604 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:39.790516 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_volume-data-source-validator-7c6cbb6c87-6nlj2_7e56e8b5-c500-424b-9f59-b08a49f14cce/volume-data-source-validator/0.log" Apr 21 10:51:40.163801 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.163768 2575 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4"] Apr 21 10:51:40.164073 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.164062 2575 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="540077c6-a9f4-4a48-9f62-58c5f7d5c095" containerName="registry" Apr 21 10:51:40.164146 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.164075 2575 state_mem.go:107] "Deleted CPUSet assignment" podUID="540077c6-a9f4-4a48-9f62-58c5f7d5c095" containerName="registry" Apr 21 10:51:40.164187 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.164153 2575 memory_manager.go:356] "RemoveStaleState removing state" podUID="540077c6-a9f4-4a48-9f62-58c5f7d5c095" containerName="registry" Apr 21 10:51:40.167128 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.167089 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.169378 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.169362 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-95dzh\"/\"kube-root-ca.crt\"" Apr 21 10:51:40.169450 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.169363 2575 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-95dzh\"/\"default-dockercfg-nrjgl\"" Apr 21 10:51:40.170369 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.170353 2575 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-95dzh\"/\"openshift-service-ca.crt\"" Apr 21 10:51:40.173550 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.173527 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4"] Apr 21 10:51:40.225230 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.225199 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-sys\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.225230 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.225236 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-podres\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.225439 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.225261 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-proc\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.225439 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.225339 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b49sh\" (UniqueName: \"kubernetes.io/projected/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-kube-api-access-b49sh\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.225439 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.225385 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-lib-modules\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.325845 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.325809 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-sys\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.325845 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.325849 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-podres\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.326043 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.325875 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-proc\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.326043 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.325920 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b49sh\" (UniqueName: \"kubernetes.io/projected/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-kube-api-access-b49sh\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.326043 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.325929 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-sys\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.326043 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.325939 2575 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-lib-modules\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.326043 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.325984 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-podres\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.326043 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.326014 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-proc\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.326043 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.326040 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-lib-modules\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.334385 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.334359 2575 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b49sh\" (UniqueName: \"kubernetes.io/projected/0d566ec0-a36d-4284-9bd3-fb5cb5803f2f-kube-api-access-b49sh\") pod \"perf-node-gather-daemonset-2lsw4\" (UID: \"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f\") " pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.477303 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.477203 2575 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:40.484598 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.484570 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-r495v_ccb482af-112d-452e-a16f-49fe93eedf37/dns/0.log" Apr 21 10:51:40.505283 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.505258 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-r495v_ccb482af-112d-452e-a16f-49fe93eedf37/kube-rbac-proxy/0.log" Apr 21 10:51:40.557294 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.557085 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-krcxr_e6f11c13-e164-47be-a16a-c5ea7b32f9e4/dns-node-resolver/0.log" Apr 21 10:51:40.604077 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.604047 2575 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4"] Apr 21 10:51:40.607349 ip-10-0-142-73 kubenswrapper[2575]: W0421 10:51:40.607319 2575 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0d566ec0_a36d_4284_9bd3_fb5cb5803f2f.slice/crio-6e743548b473e5648a86ca0dd5e2ae90963a23e118e0116597719bc70ee8959a WatchSource:0}: Error finding container 6e743548b473e5648a86ca0dd5e2ae90963a23e118e0116597719bc70ee8959a: Status 404 returned error can't find the container with id 6e743548b473e5648a86ca0dd5e2ae90963a23e118e0116597719bc70ee8959a Apr 21 10:51:40.609384 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:40.609365 2575 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 21 10:51:41.057024 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:41.056982 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" event={"ID":"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f","Type":"ContainerStarted","Data":"828c8f777311eb87fe26b0ae236921c4a09e5b909f4f110fc3afdcbbb9f3a5b8"} Apr 21 10:51:41.057024 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:41.057025 2575 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" event={"ID":"0d566ec0-a36d-4284-9bd3-fb5cb5803f2f","Type":"ContainerStarted","Data":"6e743548b473e5648a86ca0dd5e2ae90963a23e118e0116597719bc70ee8959a"} Apr 21 10:51:41.057450 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:41.057130 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:41.069761 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:41.069730 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-d5g7x_b776059e-ddd7-4893-92ef-c96817a1ae4f/node-ca/0.log" Apr 21 10:51:41.073543 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:41.073468 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" podStartSLOduration=1.073453437 podStartE2EDuration="1.073453437s" podCreationTimestamp="2026-04-21 10:51:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-21 10:51:41.073038361 +0000 UTC m=+2892.233902735" watchObservedRunningTime="2026-04-21 10:51:41.073453437 +0000 UTC m=+2892.234317789" Apr 21 10:51:41.780520 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:41.780495 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-56b8857454-d4h69_cc037d14-2ee7-421f-875a-4ec33fac21bb/router/0.log" Apr 21 10:51:42.095945 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:42.095914 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-b5lnk_eff9f25f-8452-4746-806c-70ba95f122a8/serve-healthcheck-canary/0.log" Apr 21 10:51:42.477011 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:42.476921 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-585dfdc468-4hz8p_f7a64cde-eca1-4b8d-843b-90f038244fe0/insights-operator/0.log" Apr 21 10:51:42.477743 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:42.477713 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-585dfdc468-4hz8p_f7a64cde-eca1-4b8d-843b-90f038244fe0/insights-operator/1.log" Apr 21 10:51:42.584665 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:42.584631 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-m6sp6_e686f946-4a0a-455c-b0b1-444547a37100/kube-rbac-proxy/0.log" Apr 21 10:51:42.619293 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:42.619270 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-m6sp6_e686f946-4a0a-455c-b0b1-444547a37100/exporter/0.log" Apr 21 10:51:42.652186 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:42.652156 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-m6sp6_e686f946-4a0a-455c-b0b1-444547a37100/extractor/0.log" Apr 21 10:51:47.069799 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:47.069770 2575 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-95dzh/perf-node-gather-daemonset-2lsw4" Apr 21 10:51:48.893788 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:48.893759 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-246pr_da1e2988-6ac0-4b0c-821b-5d0d7989985f/migrator/0.log" Apr 21 10:51:48.912301 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:48.912269 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-246pr_da1e2988-6ac0-4b0c-821b-5d0d7989985f/graceful-termination/0.log" Apr 21 10:51:49.259609 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:49.259505 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-7gxkl_91ca9a98-134d-41d4-95a3-460d18d5986b/kube-storage-version-migrator-operator/1.log" Apr 21 10:51:49.261180 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:49.261152 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-7gxkl_91ca9a98-134d-41d4-95a3-460d18d5986b/kube-storage-version-migrator-operator/0.log" Apr 21 10:51:50.338595 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.338563 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pj28j_c7353367-eaa9-4d89-9444-14f3bfbfc166/kube-multus-additional-cni-plugins/0.log" Apr 21 10:51:50.357842 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.357816 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pj28j_c7353367-eaa9-4d89-9444-14f3bfbfc166/egress-router-binary-copy/0.log" Apr 21 10:51:50.376555 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.376526 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pj28j_c7353367-eaa9-4d89-9444-14f3bfbfc166/cni-plugins/0.log" Apr 21 10:51:50.395365 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.395329 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pj28j_c7353367-eaa9-4d89-9444-14f3bfbfc166/bond-cni-plugin/0.log" Apr 21 10:51:50.415252 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.415226 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pj28j_c7353367-eaa9-4d89-9444-14f3bfbfc166/routeoverride-cni/0.log" Apr 21 10:51:50.433226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.433201 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pj28j_c7353367-eaa9-4d89-9444-14f3bfbfc166/whereabouts-cni-bincopy/0.log" Apr 21 10:51:50.451226 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.451203 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pj28j_c7353367-eaa9-4d89-9444-14f3bfbfc166/whereabouts-cni/0.log" Apr 21 10:51:50.679391 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.679317 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zk4px_e945e18f-ffc9-4607-92c7-e5f597ffb6a0/kube-multus/0.log" Apr 21 10:51:50.747009 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.746982 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-n8tpq_9276c6c8-dd13-4c6c-a3d3-3318a3731823/network-metrics-daemon/0.log" Apr 21 10:51:50.765715 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:50.765683 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-n8tpq_9276c6c8-dd13-4c6c-a3d3-3318a3731823/kube-rbac-proxy/0.log" Apr 21 10:51:51.582315 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:51.582278 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bsvf_157148fd-caf0-4611-9379-850dc886dfd6/ovn-controller/0.log" Apr 21 10:51:51.623804 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:51.623774 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bsvf_157148fd-caf0-4611-9379-850dc886dfd6/ovn-acl-logging/0.log" Apr 21 10:51:51.644350 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:51.644321 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bsvf_157148fd-caf0-4611-9379-850dc886dfd6/kube-rbac-proxy-node/0.log" Apr 21 10:51:51.664495 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:51.664469 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bsvf_157148fd-caf0-4611-9379-850dc886dfd6/kube-rbac-proxy-ovn-metrics/0.log" Apr 21 10:51:51.681671 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:51.681644 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bsvf_157148fd-caf0-4611-9379-850dc886dfd6/northd/0.log" Apr 21 10:51:51.700700 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:51.700674 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bsvf_157148fd-caf0-4611-9379-850dc886dfd6/nbdb/0.log" Apr 21 10:51:51.721218 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:51.721150 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bsvf_157148fd-caf0-4611-9379-850dc886dfd6/sbdb/0.log" Apr 21 10:51:51.829609 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:51.829577 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5bsvf_157148fd-caf0-4611-9379-850dc886dfd6/ovnkube-controller/0.log" Apr 21 10:51:53.489718 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:53.489673 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-8894fc9bd-w964t_6b8f4897-7439-4c3d-9a79-d48ee11b3137/check-endpoints/0.log" Apr 21 10:51:53.540996 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:53.540965 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-cnk4k_9db6586f-b886-4d23-84d5-48c25a2a6ac9/network-check-target-container/0.log" Apr 21 10:51:54.449590 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:54.449556 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-b5sx5_3ef9e3d9-e961-4953-8328-d92b08011fc3/iptables-alerter/0.log" Apr 21 10:51:55.031360 ip-10-0-142-73 kubenswrapper[2575]: I0421 10:51:55.031333 2575 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_tuned-k9p7d_ee38a6d0-8717-4092-9b02-1541c50c1659/tuned/0.log"