W0423 14:40:09.624892 1 cmd.go:257] Using insecure, self-signed certificates I0423 14:40:10.458964 1 start.go:138] Unable to read service ca bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 14:40:10.459265 1 observer_polling.go:159] Starting file observer I0423 14:40:11.214785 1 operator.go:60] Starting insights-operator v0.0.0-master+$Format:%H$ I0423 14:40:11.214974 1 legacy_config.go:327] Current config: {"report":false,"storagePath":"/var/lib/insights-operator","interval":"2h","endpoint":"https://console.redhat.com/api/ingress/v1/upload","conditionalGathererEndpoint":"https://console.redhat.com/api/gathering/v2/%s/gathering_rules","pull_report":{"endpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports","delay":"60s","timeout":"3000s","min_retry":"30s"},"impersonate":"system:serviceaccount:openshift-insights:gather","enableGlobalObfuscation":false,"ocm":{"scaEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates","scaInterval":"8h","scaDisabled":false,"clusterTransferEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/","clusterTransferInterval":"12h"},"disableInsightsAlerts":false,"processingStatusEndpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status","reportEndpointTechPreview":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/report"} I0423 14:40:11.215521 1 secure_serving.go:57] Forcing use of http/1.1 only W0423 14:40:11.215541 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. W0423 14:40:11.215544 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. W0423 14:40:11.215548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. W0423 14:40:11.215551 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. W0423 14:40:11.215553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. I0423 14:40:11.215552 1 simple_featuregate_reader.go:171] Starting feature-gate-detector W0423 14:40:11.215556 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. I0423 14:40:11.219531 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-insights", Name:"insights-operator", UID:"c0bc65cd-4aa5-4387-bb3d-eb44ff02a2c7", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BuildCSIVolumes", "CPMSMachineNamePrefix", "ConsolePluginContentSecurityPolicy", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "ImageVolume", "IngressControllerLBSubnetsAWS", "KMSv1", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "MetricsCollectionProfiles", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "PinnedImages", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SetEIPForNLBIngressController", "SigstoreImageVerification", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiDisk", "VSphereMultiNetworks"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSServiceLBNetworkSecurityGroup", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureMultiDisk", "BootImageSkewEnforcement", "BootcNodeManagement", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "Example2", "ExternalSnapshotMetadata", "GCPClusterHostedDNS", "GCPClusterHostedDNSInstall", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "ImageModeStatusReporting", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryptionProvider", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "ManagedBootImagesAzure", "ManagedBootImagesvSphere", "MaxUnavailableStatefulSet", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "MultiDiskSetup", "MutatingAdmissionPolicy", "NewOLM", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterOperations", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PreconfiguredUDNAddresses", "SELinuxMount", "ShortCertRotation", "SignatureStores", "SigstoreImageVerificationPKI", "TranslateStreamCloseWebsocketRequests", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VolumeAttributesClass", "VolumeGroupSnapshot"}} I0423 14:40:11.219492 1 operator.go:125] FeatureGates initialized: knownFeatureGates=[AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BuildCSIVolumes CPMSMachineNamePrefix ConsolePluginContentSecurityPolicy ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GatewayAPI GatewayAPIController HighlyAvailableArbiter ImageVolume IngressControllerLBSubnetsAWS KMSv1 MachineConfigNodes ManagedBootImages ManagedBootImagesAWS MetricsCollectionProfiles NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation PinnedImages ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SetEIPForNLBIngressController SigstoreImageVerification StoragePerformantSecurityPolicy UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiDisk VSphereMultiNetworks AWSClusterHostedDNS AWSClusterHostedDNSInstall AWSDedicatedHosts AWSServiceLBNetworkSecurityGroup AutomatedEtcdBackup AzureClusterHostedDNSInstall AzureDedicatedHosts AzureMultiDisk BootImageSkewEnforcement BootcNodeManagement ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterMonitoringConfig ClusterVersionOperatorConfiguration DNSNameResolver DualReplica DyanmicServiceEndpointIBMCloud DynamicResourceAllocation EtcdBackendQuota EventedPLEG Example Example2 ExternalSnapshotMetadata GCPClusterHostedDNS GCPClusterHostedDNSInstall GCPCustomAPIEndpoints GCPCustomAPIEndpointsInstall ImageModeStatusReporting ImageStreamImportMode IngressControllerDynamicConfigurationManager InsightsConfig InsightsConfigAPI InsightsOnDemandDataGather IrreconcilableMachineConfig KMSEncryptionProvider MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController ManagedBootImagesAzure ManagedBootImagesvSphere MaxUnavailableStatefulSet MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure MultiDiskSetup MutatingAdmissionPolicy NewOLM NewOLMCatalogdAPIV1Metas NewOLMOwnSingleNamespace NewOLMPreflightPermissionChecks NewOLMWebhookProviderOpenshiftServiceCA NoRegistryClusterOperations NodeSwap NutanixMultiSubnets OVNObservability OpenShiftPodSecurityAdmission PreconfiguredUDNAddresses SELinuxMount ShortCertRotation SignatureStores SigstoreImageVerificationPKI TranslateStreamCloseWebsocketRequests VSphereConfigurableMaxAllowedBlockVolumesPerNode VSphereHostVMGroupZonal VSphereMixedNodeEnv VolumeAttributesClass VolumeGroupSnapshot] I0423 14:40:11.220617 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController I0423 14:40:11.220640 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController I0423 14:40:11.220627 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0423 14:40:11.220632 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0423 14:40:11.220658 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0423 14:40:11.220660 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0423 14:40:11.221145 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/tmp/serving-cert-1433122005/tls.crt::/tmp/serving-cert-1433122005/tls.key" I0423 14:40:11.221554 1 secure_serving.go:213] Serving securely on [::]:8443 I0423 14:40:11.221599 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" W0423 14:40:11.227284 1 configmapobserver.go:64] Cannot get the configuration config map: configmaps "insights-config" not found. Default configuration is used. I0423 14:40:11.227309 1 secretconfigobserver.go:216] Legacy configuration set: enabled=false endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=false reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0423 14:40:11.227416 1 base_controller.go:76] Waiting for caches to sync for ConfigController I0423 14:40:11.231485 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0423 14:40:11.231502 1 secretconfigobserver.go:204] Legacy configuration updated: enabled=true endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=true reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0423 14:40:11.235499 1 secretconfigobserver.go:119] support secret does not exist I0423 14:40:11.242050 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0423 14:40:11.247505 1 secretconfigobserver.go:119] support secret does not exist I0423 14:40:11.247998 1 recorder.go:161] Pruning old reports every 5h22m30s, max age is 288h0m0s I0423 14:40:11.252132 1 controllerstatus.go:80] name=insightsuploader healthy=true reason= message= I0423 14:40:11.252144 1 insightsuploader.go:86] Reporting status periodically to https://console.redhat.com/api/ingress/v1/upload every 2h0m0s, starting in 1m30s I0423 14:40:11.252146 1 controllerstatus.go:80] name=insightsreport healthy=true reason= message= I0423 14:40:11.252155 1 insightsreport.go:296] Starting report retriever I0423 14:40:11.252162 1 insightsreport.go:298] Insights analysis reports will be downloaded from the https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports endpoint with a delay of 1m0s I0423 14:40:11.252181 1 periodic.go:209] Running clusterconfig gatherer I0423 14:40:11.252227 1 tasks_processing.go:45] number of workers: 64 I0423 14:40:11.252257 1 tasks_processing.go:69] worker 5 listening for tasks. I0423 14:40:11.252264 1 tasks_processing.go:69] worker 2 listening for tasks. I0423 14:40:11.252269 1 tasks_processing.go:69] worker 3 listening for tasks. I0423 14:40:11.252270 1 tasks_processing.go:69] worker 0 listening for tasks. I0423 14:40:11.252275 1 tasks_processing.go:69] worker 1 listening for tasks. I0423 14:40:11.252273 1 tasks_processing.go:69] worker 22 listening for tasks. I0423 14:40:11.252283 1 tasks_processing.go:71] worker 22 working on oauths task. I0423 14:40:11.252283 1 tasks_processing.go:69] worker 4 listening for tasks. I0423 14:40:11.252282 1 tasks_processing.go:69] worker 6 listening for tasks. I0423 14:40:11.252284 1 tasks_processing.go:71] worker 1 working on config_maps task. I0423 14:40:11.252290 1 tasks_processing.go:69] worker 7 listening for tasks. I0423 14:40:11.252292 1 tasks_processing.go:69] worker 39 listening for tasks. I0423 14:40:11.252300 1 tasks_processing.go:69] worker 54 listening for tasks. I0423 14:40:11.252295 1 tasks_processing.go:69] worker 38 listening for tasks. I0423 14:40:11.252302 1 tasks_processing.go:69] worker 8 listening for tasks. I0423 14:40:11.252307 1 tasks_processing.go:69] worker 9 listening for tasks. I0423 14:40:11.252311 1 tasks_processing.go:69] worker 23 listening for tasks. I0423 14:40:11.252309 1 tasks_processing.go:69] worker 53 listening for tasks. I0423 14:40:11.252314 1 tasks_processing.go:69] worker 16 listening for tasks. I0423 14:40:11.252325 1 tasks_processing.go:69] worker 41 listening for tasks. I0423 14:40:11.252327 1 tasks_processing.go:69] worker 21 listening for tasks. I0423 14:40:11.252318 1 tasks_processing.go:69] worker 17 listening for tasks. I0423 14:40:11.252318 1 tasks_processing.go:69] worker 55 listening for tasks. I0423 14:40:11.252335 1 tasks_processing.go:69] worker 57 listening for tasks. I0423 14:40:11.252318 1 tasks_processing.go:69] worker 40 listening for tasks. I0423 14:40:11.252323 1 tasks_processing.go:69] worker 20 listening for tasks. I0423 14:40:11.252327 1 tasks_processing.go:69] worker 56 listening for tasks. I0423 14:40:11.252334 1 tasks_processing.go:69] worker 48 listening for tasks. I0423 14:40:11.252334 1 tasks_processing.go:69] worker 24 listening for tasks. I0423 14:40:11.252342 1 tasks_processing.go:69] worker 49 listening for tasks. I0423 14:40:11.252337 1 tasks_processing.go:69] worker 19 listening for tasks. I0423 14:40:11.252354 1 tasks_processing.go:69] worker 59 listening for tasks. I0423 14:40:11.252354 1 tasks_processing.go:69] worker 18 listening for tasks. I0423 14:40:11.252344 1 tasks_processing.go:69] worker 58 listening for tasks. I0423 14:40:11.252364 1 tasks_processing.go:69] worker 12 listening for tasks. I0423 14:40:11.252369 1 tasks_processing.go:69] worker 61 listening for tasks. I0423 14:40:11.252353 1 tasks_processing.go:69] worker 51 listening for tasks. I0423 14:40:11.252374 1 tasks_processing.go:71] worker 0 working on crds task. I0423 14:40:11.252374 1 tasks_processing.go:71] worker 3 working on certificate_signing_requests task. I0423 14:40:11.252374 1 tasks_processing.go:69] worker 13 listening for tasks. I0423 14:40:11.252383 1 tasks_processing.go:69] worker 10 listening for tasks. I0423 14:40:11.252384 1 tasks_processing.go:69] worker 45 listening for tasks. I0423 14:40:11.252366 1 tasks_processing.go:69] worker 31 listening for tasks. I0423 14:40:11.252392 1 tasks_processing.go:69] worker 46 listening for tasks. I0423 14:40:11.252357 1 tasks_processing.go:69] worker 52 listening for tasks. I0423 14:40:11.252396 1 tasks_processing.go:69] worker 30 listening for tasks. I0423 14:40:11.252396 1 tasks_processing.go:69] worker 29 listening for tasks. I0423 14:40:11.252401 1 tasks_processing.go:69] worker 33 listening for tasks. I0423 14:40:11.252395 1 tasks_processing.go:69] worker 35 listening for tasks. I0423 14:40:11.252403 1 tasks_processing.go:69] worker 34 listening for tasks. I0423 14:40:11.252409 1 tasks_processing.go:71] worker 54 working on lokistack task. I0423 14:40:11.252406 1 tasks_processing.go:69] worker 11 listening for tasks. I0423 14:40:11.252414 1 tasks_processing.go:71] worker 38 working on jaegers task. I0423 14:40:11.252411 1 tasks_processing.go:71] worker 6 working on ingress_certificates task. I0423 14:40:11.252421 1 tasks_processing.go:71] worker 8 working on clusterroles task. I0423 14:40:11.252421 1 tasks_processing.go:71] worker 40 working on operators_pods_and_events task. I0423 14:40:11.252424 1 tasks_processing.go:71] worker 17 working on aggregated_monitoring_cr_names task. I0423 14:40:11.252426 1 tasks_processing.go:71] worker 11 working on machine_sets task. I0423 14:40:11.252433 1 tasks_processing.go:71] worker 41 working on nodenetworkstates task. I0423 14:40:11.252473 1 tasks_processing.go:71] worker 51 working on support_secret task. I0423 14:40:11.252481 1 tasks_processing.go:71] worker 52 working on container_images task. I0423 14:40:11.252413 1 tasks_processing.go:71] worker 20 working on feature_gates task. I0423 14:40:11.252429 1 tasks_processing.go:71] worker 16 working on pod_network_connectivity_checks task. I0423 14:40:11.252421 1 tasks_processing.go:71] worker 23 working on pdbs task. I0423 14:40:11.252512 1 tasks_processing.go:71] worker 45 working on openstack_controlplanes task. I0423 14:40:11.252425 1 tasks_processing.go:71] worker 53 working on active_alerts task. I0423 14:40:11.252363 1 tasks_processing.go:71] worker 5 working on olm_operators task. I0423 14:40:11.252771 1 tasks_processing.go:71] worker 46 working on machine_healthchecks task. I0423 14:40:11.252862 1 tasks_processing.go:71] worker 18 working on cluster_apiserver task. I0423 14:40:11.252978 1 tasks_processing.go:71] worker 30 working on networks task. I0423 14:40:11.252998 1 tasks_processing.go:71] worker 35 working on operators task. I0423 14:40:11.252366 1 tasks_processing.go:69] worker 14 listening for tasks. I0423 14:40:11.253117 1 tasks_processing.go:71] worker 33 working on openstack_dataplanenodesets task. I0423 14:40:11.253128 1 tasks_processing.go:71] worker 29 working on image_registries task. I0423 14:40:11.253192 1 tasks_processing.go:71] worker 7 working on container_runtime_configs task. I0423 14:40:11.252880 1 tasks_processing.go:71] worker 39 working on node_logs task. I0423 14:40:11.253372 1 tasks_processing.go:71] worker 24 working on cost_management_metrics_configs task. I0423 14:40:11.253399 1 tasks_processing.go:71] worker 48 working on machine_config_pools task. I0423 14:40:11.253407 1 tasks_processing.go:71] worker 56 working on infrastructures task. I0423 14:40:11.253110 1 tasks_processing.go:71] worker 14 working on openstack_dataplanedeployments task. I0423 14:40:11.253696 1 tasks_processing.go:71] worker 34 working on image_pruners task. I0423 14:40:11.253756 1 tasks_processing.go:71] worker 12 working on authentication task. I0423 14:40:11.252356 1 tasks_processing.go:69] worker 15 listening for tasks. I0423 14:40:11.254100 1 tasks_processing.go:71] worker 15 working on sap_pods task. I0423 14:40:11.252374 1 tasks_processing.go:69] worker 42 listening for tasks. I0423 14:40:11.252787 1 tasks_processing.go:71] worker 59 working on dvo_metrics task. I0423 14:40:11.252359 1 tasks_processing.go:69] worker 47 listening for tasks. I0423 14:40:11.252383 1 tasks_processing.go:69] worker 36 listening for tasks. I0423 14:40:11.252380 1 tasks_processing.go:69] worker 44 listening for tasks. I0423 14:40:11.252388 1 tasks_processing.go:69] worker 26 listening for tasks. I0423 14:40:11.252389 1 tasks_processing.go:69] worker 63 listening for tasks. I0423 14:40:11.252393 1 tasks_processing.go:69] worker 28 listening for tasks. I0423 14:40:11.252391 1 tasks_processing.go:69] worker 37 listening for tasks. I0423 14:40:11.252392 1 tasks_processing.go:69] worker 43 listening for tasks. I0423 14:40:11.252401 1 tasks_processing.go:69] worker 27 listening for tasks. I0423 14:40:11.252408 1 tasks_processing.go:71] worker 4 working on machine_autoscalers task. I0423 14:40:11.254653 1 tasks_processing.go:71] worker 27 working on nodenetworkconfigurationpolicies task. I0423 14:40:11.252409 1 tasks_processing.go:71] worker 9 working on validating_webhook_configurations task. I0423 14:40:11.252377 1 tasks_processing.go:69] worker 62 listening for tasks. I0423 14:40:11.254786 1 tasks_processing.go:71] worker 62 working on mutating_webhook_configurations task. I0423 14:40:11.252418 1 tasks_processing.go:71] worker 57 working on version task. I0423 14:40:11.252419 1 tasks_processing.go:71] worker 55 working on metrics task. I0423 14:40:11.252528 1 tasks_processing.go:71] worker 13 working on sap_config task. I0423 14:40:11.252531 1 tasks_processing.go:71] worker 58 working on openshift_machine_api_events task. I0423 14:40:11.254948 1 tasks_processing.go:71] worker 36 working on tsdb_status task. I0423 14:40:11.252349 1 tasks_processing.go:69] worker 50 listening for tasks. I0423 14:40:11.252416 1 tasks_processing.go:71] worker 21 working on proxies task. W0423 14:40:11.252790 1 gather_active_alerts.go:54] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 14:40:11.252804 1 tasks_processing.go:71] worker 31 working on install_plans task. I0423 14:40:11.255179 1 tasks_processing.go:71] worker 53 working on machines task. I0423 14:40:11.252369 1 tasks_processing.go:71] worker 2 working on number_of_pods_and_netnamespaces_with_sdn_annotations task. I0423 14:40:11.252375 1 tasks_processing.go:69] worker 32 listening for tasks. I0423 14:40:11.254273 1 tasks_processing.go:71] worker 49 working on nodes task. I0423 14:40:11.255207 1 gather.go:177] gatherer "clusterconfig" function "active_alerts" took 2.509048ms to process 0 records I0423 14:40:11.254309 1 tasks_processing.go:71] worker 61 working on schedulers task. I0423 14:40:11.252361 1 tasks_processing.go:69] worker 25 listening for tasks. I0423 14:40:11.254837 1 tasks_processing.go:71] worker 28 working on image task. I0423 14:40:11.254856 1 tasks_processing.go:71] worker 26 working on machine_configs task. I0423 14:40:11.255704 1 tasks_processing.go:71] worker 32 working on ingress task. I0423 14:40:11.254877 1 tasks_processing.go:71] worker 47 working on storage_cluster task. I0423 14:40:11.254893 1 tasks_processing.go:71] worker 63 working on openstack_version task. I0423 14:40:11.254947 1 tasks_processing.go:71] worker 37 working on service_accounts task. I0423 14:40:11.254955 1 tasks_processing.go:71] worker 42 working on monitoring_persistent_volumes task. I0423 14:40:11.252531 1 tasks_processing.go:71] worker 10 working on sap_datahubs task. I0423 14:40:11.254967 1 tasks_processing.go:71] worker 43 working on silenced_alerts task. I0423 14:40:11.256160 1 tasks_processing.go:71] worker 25 working on storage_classes task. I0423 14:40:11.254979 1 tasks_processing.go:71] worker 44 working on openshift_logging task. W0423 14:40:11.256183 1 gather_silenced_alerts.go:38] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 14:40:11.256192 1 tasks_processing.go:71] worker 43 working on qemu_kubevirt_launcher_logs task. I0423 14:40:11.256234 1 gather.go:177] gatherer "clusterconfig" function "silenced_alerts" took 24.857µs to process 0 records W0423 14:40:11.255018 1 gather_prometheus_tsdb_status.go:38] Unable to load metrics client, tsdb status cannot be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory W0423 14:40:11.255021 1 gather_most_recent_metrics.go:64] Unable to load metrics client, no metrics will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 14:40:11.255075 1 tasks_processing.go:71] worker 50 working on overlapping_namespace_uids task. I0423 14:40:11.252365 1 tasks_processing.go:69] worker 60 listening for tasks. I0423 14:40:11.255992 1 tasks_processing.go:71] worker 19 working on ceph_cluster task. I0423 14:40:11.256279 1 gather.go:177] gatherer "clusterconfig" function "metrics" took 1.306893ms to process 0 records I0423 14:40:11.256299 1 gather.go:177] gatherer "clusterconfig" function "tsdb_status" took 1.299927ms to process 0 records I0423 14:40:11.256316 1 tasks_processing.go:74] worker 36 stopped. I0423 14:40:11.256318 1 tasks_processing.go:74] worker 55 stopped. I0423 14:40:11.256325 1 tasks_processing.go:74] worker 60 stopped. I0423 14:40:11.256571 1 controller.go:128] Initializing last reported time to 0001-01-01T00:00:00Z I0423 14:40:11.256586 1 controller.go:212] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0423 14:40:11.256592 1 controller.go:212] Source periodic-conditional *controllerstatus.Simple is not ready I0423 14:40:11.256596 1 controller.go:212] Source periodic-workloads *controllerstatus.Simple is not ready I0423 14:40:11.256613 1 controller.go:489] The operator is still being initialized I0423 14:40:11.256621 1 controller.go:512] The operator is healthy I0423 14:40:11.258502 1 tasks_processing.go:74] worker 45 stopped. I0423 14:40:11.258518 1 gather.go:177] gatherer "clusterconfig" function "openstack_controlplanes" took 5.971602ms to process 0 records I0423 14:40:11.258666 1 tasks_processing.go:74] worker 22 stopped. I0423 14:40:11.258882 1 recorder.go:75] Recording config/oauth with fingerprint=0ecc89cd2f4c771d3b8337c61e24521adb3ccf2d1bfa865bf887267e30e7278e I0423 14:40:11.258894 1 gather.go:177] gatherer "clusterconfig" function "oauths" took 6.37139ms to process 1 records I0423 14:40:11.262237 1 tasks_processing.go:74] worker 24 stopped. I0423 14:40:11.262248 1 gather.go:177] gatherer "clusterconfig" function "cost_management_metrics_configs" took 8.847318ms to process 0 records I0423 14:40:11.266257 1 tasks_processing.go:74] worker 54 stopped. I0423 14:40:11.266267 1 gather.go:177] gatherer "clusterconfig" function "lokistack" took 13.841015ms to process 0 records I0423 14:40:11.266391 1 tasks_processing.go:74] worker 11 stopped. I0423 14:40:11.266404 1 gather.go:177] gatherer "clusterconfig" function "machine_sets" took 13.954329ms to process 0 records I0423 14:40:11.266421 1 tasks_processing.go:74] worker 4 stopped. I0423 14:40:11.266431 1 gather.go:177] gatherer "clusterconfig" function "machine_autoscalers" took 12.010786ms to process 0 records I0423 14:40:11.266439 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanenodesets" took 13.296536ms to process 0 records I0423 14:40:11.266443 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanedeployments" took 12.824916ms to process 0 records I0423 14:40:11.266446 1 tasks_processing.go:74] worker 33 stopped. I0423 14:40:11.266448 1 tasks_processing.go:74] worker 14 stopped. I0423 14:40:11.270409 1 tasks_processing.go:74] worker 41 stopped. I0423 14:40:11.270421 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkstates" took 17.967874ms to process 0 records I0423 14:40:11.270428 1 gather.go:177] gatherer "clusterconfig" function "container_runtime_configs" took 17.206847ms to process 0 records I0423 14:40:11.270432 1 tasks_processing.go:74] worker 7 stopped. I0423 14:40:11.270495 1 tasks_processing.go:74] worker 46 stopped. E0423 14:40:11.270505 1 gather.go:140] gatherer "clusterconfig" function "machine_healthchecks" failed with the error: machinehealthchecks.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machinehealthchecks" in API group "machine.openshift.io" at the cluster scope I0423 14:40:11.270512 1 gather.go:177] gatherer "clusterconfig" function "machine_healthchecks" took 17.67139ms to process 0 records I0423 14:40:11.270694 1 tasks_processing.go:74] worker 18 stopped. I0423 14:40:11.270807 1 recorder.go:75] Recording config/apiserver with fingerprint=d86f4fe5254f597c076f879b0d642c50c8456ec1b01e1c839ad7b1c85537016e I0423 14:40:11.270818 1 gather.go:177] gatherer "clusterconfig" function "cluster_apiserver" took 17.730871ms to process 1 records I0423 14:40:11.275640 1 tasks_processing.go:74] worker 15 stopped. I0423 14:40:11.275654 1 gather.go:177] gatherer "clusterconfig" function "sap_pods" took 21.528648ms to process 0 records I0423 14:40:11.275682 1 tasks_processing.go:74] worker 48 stopped. I0423 14:40:11.275692 1 gather.go:177] gatherer "clusterconfig" function "machine_config_pools" took 22.263997ms to process 0 records I0423 14:40:11.275776 1 tasks_processing.go:74] worker 51 stopped. E0423 14:40:11.275784 1 gather.go:140] gatherer "clusterconfig" function "support_secret" failed with the error: secrets "support" not found I0423 14:40:11.275790 1 gather.go:177] gatherer "clusterconfig" function "support_secret" took 23.29356ms to process 0 records I0423 14:40:11.275838 1 tasks_processing.go:74] worker 27 stopped. I0423 14:40:11.275847 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkconfigurationpolicies" took 21.172911ms to process 0 records I0423 14:40:11.275862 1 tasks_processing.go:74] worker 56 stopped. I0423 14:40:11.276559 1 recorder.go:75] Recording config/infrastructure with fingerprint=a7fa03692a1eef6399141645eb66f1e81b3992a4eff1468ad82eed76df462d90 I0423 14:40:11.276573 1 gather.go:177] gatherer "clusterconfig" function "infrastructures" took 22.438844ms to process 1 records I0423 14:40:11.276585 1 gather.go:177] gatherer "clusterconfig" function "jaegers" took 23.515488ms to process 0 records I0423 14:40:11.276596 1 tasks_processing.go:74] worker 38 stopped. I0423 14:40:11.276664 1 recorder.go:75] Recording config/pdbs/openshift-image-registry/image-registry with fingerprint=4314340806d6ca22407c62c114c358ca05390c20a0f6d7398fd51deab5793281 I0423 14:40:11.276669 1 tasks_processing.go:74] worker 23 stopped. I0423 14:40:11.276681 1 recorder.go:75] Recording config/pdbs/openshift-ingress/router-default with fingerprint=143019f52698d281905f764f2b7043570c6f25bf61588620f15e3e424a1a10a5 I0423 14:40:11.276695 1 recorder.go:75] Recording config/pdbs/openshift-operator-lifecycle-manager/packageserver-pdb with fingerprint=1aaa146438f91368c94c24a07f6c6e2b36d4d53f40e6d1b27440b3eadf95d9bb I0423 14:40:11.276702 1 gather.go:177] gatherer "clusterconfig" function "pdbs" took 23.477561ms to process 3 records I0423 14:40:11.276710 1 gather.go:177] gatherer "clusterconfig" function "node_logs" took 22.814848ms to process 0 records E0423 14:40:11.276716 1 gather.go:140] gatherer "clusterconfig" function "pod_network_connectivity_checks" failed with the error: the server could not find the requested resource (get podnetworkconnectivitychecks.controlplane.operator.openshift.io) I0423 14:40:11.276722 1 gather.go:177] gatherer "clusterconfig" function "pod_network_connectivity_checks" took 23.619627ms to process 0 records I0423 14:40:11.276722 1 tasks_processing.go:74] worker 39 stopped. E0423 14:40:11.276726 1 gather.go:140] gatherer "clusterconfig" function "machines" failed with the error: machines.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machines" in API group "machine.openshift.io" at the cluster scope I0423 14:40:11.276731 1 gather.go:177] gatherer "clusterconfig" function "machines" took 20.933027ms to process 0 records I0423 14:40:11.276756 1 tasks_processing.go:74] worker 53 stopped. I0423 14:40:11.276758 1 tasks_processing.go:74] worker 16 stopped. I0423 14:40:11.276799 1 tasks_processing.go:74] worker 12 stopped. I0423 14:40:11.276918 1 recorder.go:75] Recording config/authentication with fingerprint=486e0b829fd3125825a8f002c5ffd26f3a0c08b0b120980d16190b2e8c1e9d1c I0423 14:40:11.276931 1 gather.go:177] gatherer "clusterconfig" function "authentication" took 22.594644ms to process 1 records I0423 14:40:11.277016 1 tasks_processing.go:74] worker 30 stopped. I0423 14:40:11.277079 1 recorder.go:75] Recording config/network with fingerprint=0a164c53ecb96e08f34f599abb4b706618f318c95c071a1585d5cfc5a1f3852e I0423 14:40:11.277091 1 gather.go:177] gatherer "clusterconfig" function "networks" took 23.428239ms to process 1 records I0423 14:40:11.277169 1 tasks_processing.go:74] worker 34 stopped. I0423 14:40:11.277245 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/imagepruner/cluster with fingerprint=a10ddc5609b7e21387be99381b7b3e1b39ec79d6e532ec488cec8a8405b57d88 I0423 14:40:11.277257 1 gather.go:177] gatherer "clusterconfig" function "image_pruners" took 22.785981ms to process 1 records I0423 14:40:11.278313 1 tasks_processing.go:74] worker 58 stopped. I0423 14:40:11.278353 1 gather.go:177] gatherer "clusterconfig" function "openshift_machine_api_events" took 23.356973ms to process 0 records I0423 14:40:11.278415 1 tasks_processing.go:74] worker 21 stopped. I0423 14:40:11.278465 1 recorder.go:75] Recording config/proxy with fingerprint=832f7398874007fe38fd6892cdf811255d2b3c808feb5674e79451c9fae30194 I0423 14:40:11.278475 1 gather.go:177] gatherer "clusterconfig" function "proxies" took 23.325815ms to process 1 records I0423 14:40:11.279060 1 tasks_processing.go:74] worker 19 stopped. I0423 14:40:11.279101 1 gather.go:177] gatherer "clusterconfig" function "ceph_cluster" took 22.785618ms to process 0 records I0423 14:40:11.279234 1 tasks_processing.go:74] worker 13 stopped. I0423 14:40:11.279254 1 gather.go:177] gatherer "clusterconfig" function "sap_config" took 24.257869ms to process 0 records I0423 14:40:11.279459 1 tasks_processing.go:74] worker 44 stopped. I0423 14:40:11.279477 1 gather.go:177] gatherer "clusterconfig" function "openshift_logging" took 23.270834ms to process 0 records I0423 14:40:11.279491 1 tasks_processing.go:74] worker 10 stopped. I0423 14:40:11.279504 1 gather.go:177] gatherer "clusterconfig" function "sap_datahubs" took 23.441813ms to process 0 records I0423 14:40:11.280544 1 tasks_processing.go:74] worker 20 stopped. I0423 14:40:11.280692 1 recorder.go:75] Recording config/featuregate with fingerprint=b660db11519b377c37a0ac93d81932790fe3261e763a965aa6d3eafe14487d1b I0423 14:40:11.280710 1 gather.go:177] gatherer "clusterconfig" function "feature_gates" took 28.043883ms to process 1 records I0423 14:40:11.280720 1 gather.go:177] gatherer "clusterconfig" function "storage_cluster" took 24.774002ms to process 0 records I0423 14:40:11.280727 1 tasks_processing.go:74] worker 47 stopped. I0423 14:40:11.280927 1 tasks_processing.go:74] worker 63 stopped. I0423 14:40:11.280938 1 gather.go:177] gatherer "clusterconfig" function "openstack_version" took 25.00194ms to process 0 records I0423 14:40:11.281151 1 tasks_processing.go:74] worker 61 stopped. I0423 14:40:11.281231 1 recorder.go:75] Recording config/schedulers/cluster with fingerprint=e635f00029add8286be17f11ab8c6979c6a4eaf7d43fa26ccac3e389c5442ddc I0423 14:40:11.281244 1 gather.go:177] gatherer "clusterconfig" function "schedulers" took 25.83701ms to process 1 records I0423 14:40:11.281295 1 tasks_processing.go:74] worker 28 stopped. I0423 14:40:11.281367 1 recorder.go:75] Recording config/image with fingerprint=72f79be99e9d81bc527722c7678a9373f6f35d9341fb3539d8ab0958392ad043 I0423 14:40:11.281378 1 gather.go:177] gatherer "clusterconfig" function "image" took 25.966387ms to process 1 records I0423 14:40:11.281522 1 tasks_processing.go:74] worker 29 stopped. W0423 14:40:11.281567 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 14:40:11.281843 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/config/cluster with fingerprint=a452af2b13c467007e87c52574cb70fd57592af863521b9947a9229a367b902a I0423 14:40:11.281856 1 gather.go:177] gatherer "clusterconfig" function "image_registries" took 28.376836ms to process 1 records I0423 14:40:11.283344 1 tasks_processing.go:74] worker 32 stopped. I0423 14:40:11.283453 1 recorder.go:75] Recording config/ingress with fingerprint=aee41deba192d661b105dcc2635ebb8f5c49e700ed24f392efe9f1bda8d10568 I0423 14:40:11.283471 1 gather.go:177] gatherer "clusterconfig" function "ingress" took 27.624401ms to process 1 records I0423 14:40:11.283966 1 tasks_processing.go:74] worker 9 stopped. I0423 14:40:11.284134 1 recorder.go:75] Recording config/validatingwebhookconfigurations/multus.openshift.io with fingerprint=91d3cc843911bd12dd3ec2ec01d924a4ed56a582bbdf2a7784d92fc6804156df I0423 14:40:11.284220 1 recorder.go:75] Recording config/validatingwebhookconfigurations/network-node-identity.openshift.io with fingerprint=a71c9829befa769ce8288dfba5bc5578580413fe46515660c7c16abe7b18ba73 I0423 14:40:11.284248 1 recorder.go:75] Recording config/validatingwebhookconfigurations/performance-addon-operator with fingerprint=7f80d2f7674153b3e29a5ce888193014710c2179f8b39a748517eec7267e0855 I0423 14:40:11.284282 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterrolebindings-validation with fingerprint=6be6bdff24422a72820f462551d3d479083b0aec5bc6958ef22e5cb175b345f7 I0423 14:40:11.284315 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterroles-validation with fingerprint=d592abab383b56bf094ffcb91ce804e3be8301f360d402b638414a92f4ad0de0 I0423 14:40:11.284347 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-ingress-config-validation with fingerprint=febf29d5ef1a8ff7083e1d2b1ca76b694ced4b8d87ab9666dfe08a807249ab19 I0423 14:40:11.284393 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-network-operator-validation with fingerprint=7b087b15db8420a495efbe54c506023dda80255735af8a0f29c41e1b85a935f4 I0423 14:40:11.284437 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-regular-user-validation with fingerprint=0b45fa9f30886b310f875ad26d6537da36f3295c8f569921cef768a42fa29578 I0423 14:40:11.284477 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-scc-validation with fingerprint=27c3a19b4adbbbec0bb889afdfa48ad5769af9195dabfebe6961d2070b171911 I0423 14:40:11.284514 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-serviceaccount-validation with fingerprint=6e654dcb57f121a0eb8a5de16e9ff80e20e2e6f058fabd117a04df2ab59e59aa I0423 14:40:11.284549 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-techpreviewnoupgrade-validation with fingerprint=19cfadb0e8afcdd3ec1e937bdb24b04da1ffb071fe94b0cbc39cc5f9e98ad5ff I0423 14:40:11.284561 1 gather.go:177] gatherer "clusterconfig" function "validating_webhook_configurations" took 29.276719ms to process 11 records I0423 14:40:11.284653 1 tasks_processing.go:74] worker 62 stopped. I0423 14:40:11.284658 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/aws-pod-identity with fingerprint=b638f9909f2d83e1e50b834c6a778444e24e9f7702b19b41cc41df845d207dd2 I0423 14:40:11.284725 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-podimagespec-mutation with fingerprint=79c008b3aabd1e2664fdccc1ea792529b2ce9a3367599f0a1cd9277be0d605ee I0423 14:40:11.284767 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-service-mutation with fingerprint=0fdf58fcc8b0d0717d63f48fcddad0b3ce080c4322c39d815c796233148925be I0423 14:40:11.284784 1 gather.go:177] gatherer "clusterconfig" function "mutating_webhook_configurations" took 29.06249ms to process 3 records I0423 14:40:11.284916 1 tasks_processing.go:74] worker 49 stopped. I0423 14:40:11.285180 1 recorder.go:75] Recording config/node/ip-10-0-0-127.ec2.internal with fingerprint=f83a1a75d4136abd6174f775fbe284e4574222419f2edcce5e8a8939d0ad3e8d I0423 14:40:11.285269 1 recorder.go:75] Recording config/node/ip-10-0-1-163.ec2.internal with fingerprint=8a438f7a9dafb04c5a68e08d2cd63256daf9c62b05272a5d1d80b558949bab02 I0423 14:40:11.285379 1 recorder.go:75] Recording config/node/ip-10-0-2-76.ec2.internal with fingerprint=632ce266daa7ed620d1f9cec70b42f47ad0f1ade5432b8073c89d9b328de89f9 I0423 14:40:11.285391 1 gather.go:177] gatherer "clusterconfig" function "nodes" took 28.96666ms to process 3 records I0423 14:40:11.285741 1 tasks_processing.go:74] worker 42 stopped. I0423 14:40:11.285759 1 gather.go:177] gatherer "clusterconfig" function "monitoring_persistent_volumes" took 29.712755ms to process 0 records I0423 14:40:11.285811 1 gather_logs.go:145] no pods in namespace were found I0423 14:40:11.285827 1 tasks_processing.go:74] worker 43 stopped. I0423 14:40:11.285848 1 gather.go:177] gatherer "clusterconfig" function "qemu_kubevirt_launcher_logs" took 29.629255ms to process 0 records I0423 14:40:11.286020 1 tasks_processing.go:74] worker 8 stopped. I0423 14:40:11.286322 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/admin with fingerprint=109eee05ade67fe446eca4c0a93c5ffc16ad4dc53c328e20b77ff0c7daf27e07 I0423 14:40:11.286538 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/edit with fingerprint=7d3f565e801e5cb6a064b2689f1aff8e009d917f837f4b8dcf690c162c96b760 I0423 14:40:11.286576 1 gather.go:177] gatherer "clusterconfig" function "clusterroles" took 33.589793ms to process 2 records I0423 14:40:11.286612 1 gather.go:177] gatherer "clusterconfig" function "certificate_signing_requests" took 33.635679ms to process 0 records I0423 14:40:11.286809 1 recorder.go:75] Recording config/storage/storageclasses/gp2-csi with fingerprint=5bf4c8b9eb201334f61a325f05de58d1bac57b6b0cb3df49c7286a4d386f25a6 I0423 14:40:11.286856 1 recorder.go:75] Recording config/storage/storageclasses/gp3-csi with fingerprint=67c34eecc1b77de73b9e3a5cb2c5beb91c597eb114ec042e2891b387f5667cc1 I0423 14:40:11.286917 1 gather.go:177] gatherer "clusterconfig" function "storage_classes" took 30.012949ms to process 2 records I0423 14:40:11.286972 1 tasks_processing.go:74] worker 25 stopped. I0423 14:40:11.287012 1 tasks_processing.go:74] worker 0 stopped. I0423 14:40:11.287130 1 tasks_processing.go:74] worker 3 stopped. I0423 14:40:11.288173 1 recorder.go:75] Recording config/crd/volumesnapshots.snapshot.storage.k8s.io with fingerprint=e4045216dd9e9092a567fdc9cc95af59b1037ac3eeb6b000beed03c8d81734e4 I0423 14:40:11.289282 1 recorder.go:75] Recording config/crd/volumesnapshotcontents.snapshot.storage.k8s.io with fingerprint=cad8828a510e964668b356507be4fbe9c8f28e370cd1c561427e6c8f33e6e5ce I0423 14:40:11.289299 1 gather.go:177] gatherer "clusterconfig" function "crds" took 33.945814ms to process 2 records I0423 14:40:11.289312 1 gather.go:177] gatherer "clusterconfig" function "aggregated_monitoring_cr_names" took 35.642944ms to process 0 records I0423 14:40:11.289325 1 recorder.go:75] Recording config/namespaces_with_overlapping_uids with fingerprint=4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945 I0423 14:40:11.289330 1 gather.go:177] gatherer "clusterconfig" function "overlapping_namespace_uids" took 32.870401ms to process 1 records I0423 14:40:11.289326 1 tasks_processing.go:74] worker 17 stopped. I0423 14:40:11.289336 1 tasks_processing.go:74] worker 50 stopped. I0423 14:40:11.290917 1 tasks_processing.go:74] worker 5 stopped. I0423 14:40:11.291029 1 recorder.go:75] Recording config/olm_operators with fingerprint=ea9251847ad24a42e39dd6c697c2d8beddc78ce3eab45c41571c8afcedf1a714 I0423 14:40:11.291048 1 gather.go:177] gatherer "clusterconfig" function "olm_operators" took 38.213453ms to process 1 records I0423 14:40:11.295282 1 sca.go:136] Pulling SCA certificates from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates. Next check is in 8h0m0s I0423 14:40:11.295287 1 cluster_transfer.go:83] checking the availability of cluster transfer. Next check is in 12h0m0s W0423 14:40:11.295412 1 operator.go:288] started I0423 14:40:11.295439 1 base_controller.go:76] Waiting for caches to sync for LoggingSyncer I0423 14:40:11.295794 1 tasks_processing.go:74] worker 52 stopped. I0423 14:40:11.296921 1 recorder.go:75] Recording config/pod/openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-0-127.ec2.internal with fingerprint=279b641949cfae3850de609c4da00df33d03fcdee7ed1d8140e17f04fcf279c3 I0423 14:40:11.296968 1 recorder.go:75] Recording config/running_containers with fingerprint=c0c97838bbfe3c3a06ac29b6f973c3d76bd156705009fa49471b139e798faf37 I0423 14:40:11.296977 1 gather.go:177] gatherer "clusterconfig" function "container_images" took 43.298721ms to process 2 records I0423 14:40:11.305669 1 controller.go:212] Source periodic-workloads *controllerstatus.Simple is not ready I0423 14:40:11.305684 1 controller.go:212] Source scaController *sca.Controller is not ready I0423 14:40:11.305688 1 controller.go:212] Source clusterTransferController *clustertransfer.Controller is not ready I0423 14:40:11.305691 1 controller.go:212] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0423 14:40:11.305693 1 controller.go:212] Source periodic-conditional *controllerstatus.Simple is not ready I0423 14:40:11.305709 1 controller.go:489] The operator is still being initialized I0423 14:40:11.305714 1 controller.go:512] The operator is healthy I0423 14:40:11.308154 1 tasks_processing.go:74] worker 2 stopped. I0423 14:40:11.308169 1 gather.go:177] gatherer "clusterconfig" function "number_of_pods_and_netnamespaces_with_sdn_annotations" took 52.943662ms to process 0 records I0423 14:40:11.309729 1 prometheus_rules.go:88] Prometheus rules successfully created E0423 14:40:11.311473 1 cluster_transfer.go:95] failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%279f9a9ebc-d93b-420a-8221-7e5ebb7b2f94%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.128.0.10:59915->172.30.0.10:53: read: connection refused I0423 14:40:11.311486 1 controllerstatus.go:80] name=clusterTransferController healthy=true reason=Disconnected message=failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%279f9a9ebc-d93b-420a-8221-7e5ebb7b2f94%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.128.0.10:59915->172.30.0.10:53: read: connection refused I0423 14:40:11.314001 1 tasks_processing.go:74] worker 1 stopped. E0423 14:40:11.314013 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "cluster-monitoring-config" not found E0423 14:40:11.314018 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "gateway-mode-config" not found E0423 14:40:11.314021 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "insights-config" not found I0423 14:40:11.314029 1 recorder.go:75] Recording config/configmaps/openshift-config/installer-images/images.json with fingerprint=26b6661162b099a0f5a279859b4f46c867929a79d9a4a41fde4be4e6fe138018 I0423 14:40:11.314068 1 recorder.go:75] Recording config/configmaps/openshift-config/kube-root-ca.crt/ca.crt with fingerprint=d476c7d3f5b104863f08f481b1264dcc68cc272ecefb0ecb709b18a6afab034d I0423 14:40:11.314076 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/invoker with fingerprint=76b482f683cd3ef9da02debac5b26080a5aeb06ff768ee5c21117514dff29d8a I0423 14:40:11.314080 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/version with fingerprint=c93090eb0d2a4736885abeb79c91680cfd01fda46464f83456b085d4dc8239f0 I0423 14:40:11.314083 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 I0423 14:40:11.314118 1 recorder.go:75] Recording config/configmaps/openshift-config/rosa-brand-logo/rosa-brand-logo.svg with fingerprint=6ed8ca4dd7a8eee7249182bc006e9649ce84d76c551ddfaaa33e55d8c4cc1ed0 I0423 14:40:11.314125 1 recorder.go:75] Recording config/configmaps/kube-system/cluster-config-v1/install-config with fingerprint=ab3811c6b83fd7b8e920094cfa3080d1b4ee3c35ec4c8379437b21d27bd6608d I0423 14:40:11.314130 1 gather.go:177] gatherer "clusterconfig" function "config_maps" took 61.699319ms to process 7 records I0423 14:40:11.320983 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0423 14:40:11.320992 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0423 14:40:11.321124 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController I0423 14:40:11.328189 1 base_controller.go:82] Caches are synced for ConfigController I0423 14:40:11.328199 1 base_controller.go:119] Starting #1 worker of ConfigController controller ... I0423 14:40:11.332333 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 14:40:11.369010 1 tasks_processing.go:74] worker 26 stopped. I0423 14:40:11.369033 1 recorder.go:75] Recording aggregated/unused_machine_configs_count with fingerprint=4bfc9fa984e5dfcd45848faaf05269de7619bf42edf9f781751af5ee05c1a499 I0423 14:40:11.369042 1 gather.go:177] gatherer "clusterconfig" function "machine_configs" took 113.343248ms to process 1 records I0423 14:40:11.375373 1 tasks_processing.go:74] worker 57 stopped. I0423 14:40:11.375598 1 recorder.go:75] Recording config/version with fingerprint=8eb6c4bcca5a4ab400a3b2dd89c2f8344a4409e779b6a2356946cfd78f7e699e I0423 14:40:11.375617 1 recorder.go:75] Recording config/id with fingerprint=b8fa6877b52a69c6e62eed406d401beb94b89a5f0a87a0b368a8326a1e249474 I0423 14:40:11.375623 1 gather.go:177] gatherer "clusterconfig" function "version" took 120.392468ms to process 2 records I0423 14:40:11.383995 1 requests.go:205] Asking for SCA certificate with "{"arch": ["x86_64"]}" payload W0423 14:40:11.387964 1 sca.go:161] Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.128.0.10:47283->172.30.0.10:53: read: connection refused I0423 14:40:11.387978 1 controllerstatus.go:80] name=scaController healthy=true reason=NonHTTPError message=Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.128.0.10:47283->172.30.0.10:53: read: connection refused I0423 14:40:11.394219 1 tasks_processing.go:74] worker 6 stopped. E0423 14:40:11.394232 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret 'router-certs-default' in namespace 'openshift-ingress': secrets "router-certs-default" not found E0423 14:40:11.394238 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret '2prtjk59bm7qaqh0k5bvl0hqkh3iqck5-primary-cert-bundle-secret' in namespace 'openshift-ingress-operator': secrets "2prtjk59bm7qaqh0k5bvl0hqkh3iqck5-primary-cert-bundle-secret" not found I0423 14:40:11.394280 1 recorder.go:75] Recording aggregated/ingress_controllers_certs with fingerprint=71c7488ad33475885e3f1ead6081dc0bd52251a6871cfffcd53663fa5b895cc1 I0423 14:40:11.394291 1 gather.go:177] gatherer "clusterconfig" function "ingress_certificates" took 141.789051ms to process 1 records I0423 14:40:11.396322 1 base_controller.go:82] Caches are synced for LoggingSyncer I0423 14:40:11.396336 1 base_controller.go:119] Starting #1 worker of LoggingSyncer controller ... I0423 14:40:11.692838 1 gather_cluster_operator_pods_and_events.go:121] Found 18 pods with 21 containers I0423 14:40:11.692851 1 gather_cluster_operator_pods_and_events.go:235] Maximum buffer size: 1198372 bytes I0423 14:40:11.693933 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-929w7 pod in namespace openshift-dns (previous: false). I0423 14:40:11.920028 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-929w7 pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-929w7\" is waiting to start: ContainerCreating" I0423 14:40:11.920044 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-929w7\" is waiting to start: ContainerCreating" I0423 14:40:11.920053 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-929w7 pod in namespace openshift-dns (previous: false). I0423 14:40:12.097664 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-929w7 pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-929w7\" is waiting to start: ContainerCreating" I0423 14:40:12.097679 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-929w7\" is waiting to start: ContainerCreating" I0423 14:40:12.097708 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-cps8z pod in namespace openshift-dns (previous: false). W0423 14:40:12.280365 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 14:40:12.320658 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-cps8z pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-cps8z\" is waiting to start: ContainerCreating" I0423 14:40:12.320671 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-cps8z\" is waiting to start: ContainerCreating" I0423 14:40:12.320679 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-cps8z pod in namespace openshift-dns (previous: false). I0423 14:40:12.498931 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-cps8z pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-cps8z\" is waiting to start: ContainerCreating" I0423 14:40:12.498945 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-cps8z\" is waiting to start: ContainerCreating" I0423 14:40:12.498975 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-wxthm pod in namespace openshift-dns (previous: false). I0423 14:40:12.707296 1 gather_cluster_operators.go:184] Unable to get configs.samples.operator.openshift.io resource due to: configs.samples.operator.openshift.io "cluster" not found I0423 14:40:12.718880 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-wxthm pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-wxthm\" is waiting to start: ContainerCreating" I0423 14:40:12.718893 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-wxthm\" is waiting to start: ContainerCreating" I0423 14:40:12.718915 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-wxthm pod in namespace openshift-dns (previous: false). I0423 14:40:12.898203 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-wxthm pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-wxthm\" is waiting to start: ContainerCreating" I0423 14:40:12.898217 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-wxthm\" is waiting to start: ContainerCreating" I0423 14:40:12.898227 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-bm4kh pod in namespace openshift-dns (previous: false). I0423 14:40:13.097612 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 14:40:13.097627 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-g45dr pod in namespace openshift-dns (previous: false). I0423 14:40:13.109024 1 tasks_processing.go:74] worker 35 stopped. I0423 14:40:13.109071 1 recorder.go:75] Recording config/clusteroperator/console with fingerprint=f2300918a2ae81361ac14984e3afa3774f3aa54130c167f355215c86bd0c6fc4 I0423 14:40:13.109107 1 recorder.go:75] Recording config/clusteroperator/csi-snapshot-controller with fingerprint=f9111a4fd190a5b0809ba977a1892c5f548842850bc961518cbdf25fa3ace774 I0423 14:40:13.109137 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/csisnapshotcontroller/cluster with fingerprint=5adc514f4b63e2f1ecc68bf6f9c0af70c5eea04522a49524e102721b1c41f80e I0423 14:40:13.109166 1 recorder.go:75] Recording config/clusteroperator/dns with fingerprint=c44b8c8b51c86a470b2f383cf690ade8414e47015d6cd571c6fbbbbc23073018 I0423 14:40:13.109203 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/dns/default with fingerprint=9e7b4ce029030d3d8c3b49af92c556acdcc415000b40d3f969dbdc42c432b47f I0423 14:40:13.109227 1 recorder.go:75] Recording config/clusteroperator/image-registry with fingerprint=79ffd50f0b6c88b18d964cfa85af2c44bf37f3a3c25c3559259f875ea1751639 I0423 14:40:13.109259 1 recorder.go:75] Recording config/clusteroperator/ingress with fingerprint=64833435b9ebf9d75ad437714a3338801124571d184902cd6c0a4ac4c52f1ad0 I0423 14:40:13.109283 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/ingresscontroller/openshift-ingress-operator/default with fingerprint=a6410955466fa6e89a1d0daf42aeb95f881aed4676e65958cafc10e906b7a3a4 I0423 14:40:13.109297 1 recorder.go:75] Recording config/clusteroperator/insights with fingerprint=8aa37c5363880c4bd062836bb1a0f232fb692ea5d8d76945607076d5b19df4e9 I0423 14:40:13.109314 1 recorder.go:75] Recording config/clusteroperator/kube-apiserver with fingerprint=4aae7b227411aa14e708665099024ab6c50813d35bcf41e7960b497459df2564 I0423 14:40:13.109323 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubeapiserver/cluster with fingerprint=51503bf0b784fcf65ea46bcaf1f72ac1a5c4d5dc211934f18f27871efed05762 I0423 14:40:13.109339 1 recorder.go:75] Recording config/clusteroperator/kube-controller-manager with fingerprint=ad1742c935f8f5746cd6e62988209ad27f292019bbdf7436de54a96868995c43 I0423 14:40:13.109349 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubecontrollermanager/cluster with fingerprint=ce90c0d4f367d7da085074268031798382ae7c54fdcb0a21f15a4818fe308c11 I0423 14:40:13.109365 1 recorder.go:75] Recording config/clusteroperator/kube-scheduler with fingerprint=7c6a13169cedacc15176258eed149e8ad71619f198cd5850b5e9d6289547162a I0423 14:40:13.109375 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubescheduler/cluster with fingerprint=f2940fb9fd20c19951dfc295eb363b7fba0c505f5ae61f01967a063099e6b60a I0423 14:40:13.109388 1 recorder.go:75] Recording config/clusteroperator/kube-storage-version-migrator with fingerprint=317a5399b249d36575d223d1c2f7d8aee30e9660db2c9be85265d0ab66f5dce6 I0423 14:40:13.109410 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubestorageversionmigrator/cluster with fingerprint=9351181aa7e6ada41ef581ab31e13516c6b934cc95710154bafb2eb222cb58db I0423 14:40:13.109428 1 recorder.go:75] Recording config/clusteroperator/monitoring with fingerprint=6543611a59be75c6f45fa09267f0e85e51ee767d36b529f39e1e70197d058fa0 I0423 14:40:13.109550 1 recorder.go:75] Recording config/clusteroperator/network with fingerprint=fbe438575946f5ad778f5718c68cae83fb6dbc9477c85720ae0ca1655c1c429d I0423 14:40:13.109565 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/ovn with fingerprint=626a89d20e0deaed5b6dfb533acfe65f4bb1618bd200a703b62e60c5d16d94ab I0423 14:40:13.109573 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/signer with fingerprint=90410b16914712b85b3c4578716ad8c0ae072e688f4cd1e022bf76f20da3506d I0423 14:40:13.109600 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/network/cluster with fingerprint=a93d15eaecb455a0e40ecb2826eeecc1533899204ddd3c3921d15ab70af7ae75 I0423 14:40:13.109622 1 recorder.go:75] Recording config/clusteroperator/node-tuning with fingerprint=6553ca5ba8927869f7ec607c2d793d2fa58ca6b2b3cbc54f429d8d961c2c5934 I0423 14:40:13.109646 1 recorder.go:75] Recording config/clusteroperator/openshift-apiserver with fingerprint=310eb9d48a1121d6e0c833abd13fdfb2b00f18fbc3c0833bb5b4067686a04f1c I0423 14:40:13.109655 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftapiserver/cluster with fingerprint=e712e6cf27339b441e4ed1f4cde91dbde7e952698ba93407e4457db63a4a4c76 I0423 14:40:13.109671 1 recorder.go:75] Recording config/clusteroperator/openshift-controller-manager with fingerprint=4bfd03815a5fcebcf126686696bf4fc67f6267e61b4ea8cedf6c1c9a7174884a I0423 14:40:13.109682 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftcontrollermanager/cluster with fingerprint=d71a0f4672f9b45d9fc8293bf1687afc650fd28d32e2e30de27523fe7b4eadf7 I0423 14:40:13.109695 1 recorder.go:75] Recording config/clusteroperator/openshift-samples with fingerprint=670bba02e6564f5e47dd91ef43f87ed5fb64f892fa0c15ad63f099f960b03b15 I0423 14:40:13.109709 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager with fingerprint=58f9f61d9b0f56ff0a30ccad3c65bcb58da39d8e090905933081d80ae9d3b2d9 I0423 14:40:13.109723 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-catalog with fingerprint=2d213aa59ce4dd9cbc41c1775eba27b52cc32aebb90dad3996230d544fd14ada I0423 14:40:13.109738 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-packageserver with fingerprint=b987ddcf25c7f793e7405dcf304244a2e651cc60ee15fb2ed33447188737521c I0423 14:40:13.109749 1 recorder.go:75] Recording config/clusteroperator/service-ca with fingerprint=f5ceb9ebc8e8949722d0535f5ca3af293bd7d47af716f28cbdf4c9b7a1abd0eb I0423 14:40:13.109782 1 recorder.go:75] Recording config/clusteroperator/storage with fingerprint=d85a228b396eea8cb8f7a00578fd387c197eeaba78e996e4819f33d150b74754 I0423 14:40:13.109798 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/clustercsidriver/ebs.csi.aws.com with fingerprint=510064d6f6bcced87ab5bd2ddaff3d0edd7f93f4a4f7af2641f29fc53ffab21e I0423 14:40:13.109807 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/storage/cluster with fingerprint=8e480f8c1ce1b39baac42d8ec780c57c2592929ae0c801b61ffad49ba13f33ad I0423 14:40:13.109815 1 gather.go:177] gatherer "clusterconfig" function "operators" took 1.856000678s to process 35 records W0423 14:40:13.280803 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 14:40:13.298847 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 14:40:13.298864 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-r7td7 pod in namespace openshift-dns (previous: false). I0423 14:40:13.498236 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 14:40:13.498291 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-7cb4f6c89-j52dm pod in namespace openshift-image-registry (previous: false). I0423 14:40:13.705121 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-7cb4f6c89-j52dm pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-7cb4f6c89-j52dm\" is waiting to start: ContainerCreating" I0423 14:40:13.705135 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-7cb4f6c89-j52dm\" is waiting to start: ContainerCreating" I0423 14:40:13.705178 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-7cb4f6c89-prfnv pod in namespace openshift-image-registry (previous: false). I0423 14:40:13.897866 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-7cb4f6c89-prfnv pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-7cb4f6c89-prfnv\" is waiting to start: ContainerCreating" I0423 14:40:13.897881 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-7cb4f6c89-prfnv\" is waiting to start: ContainerCreating" I0423 14:40:13.897923 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-844ff948f4-qg22z pod in namespace openshift-image-registry (previous: false). I0423 14:40:14.097954 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-844ff948f4-qg22z pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-844ff948f4-qg22z\" is waiting to start: ContainerCreating" I0423 14:40:14.097969 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-844ff948f4-qg22z\" is waiting to start: ContainerCreating" I0423 14:40:14.097979 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-bkv4m pod in namespace openshift-image-registry (previous: false). W0423 14:40:14.280399 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 14:40:14.297927 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 14:40:14.297944 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-nhstg pod in namespace openshift-image-registry (previous: false). I0423 14:40:14.500144 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 14:40:14.500157 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-vlk9n pod in namespace openshift-image-registry (previous: false). I0423 14:40:14.698433 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 14:40:14.698447 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-5c464d6b7f-76x9c pod in namespace openshift-ingress (previous: false). I0423 14:40:14.898808 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-5c464d6b7f-76x9c pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-5c464d6b7f-76x9c\" is waiting to start: ContainerCreating" I0423 14:40:14.898825 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-5c464d6b7f-76x9c\" is waiting to start: ContainerCreating" I0423 14:40:14.898838 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-5c464d6b7f-vqh9q pod in namespace openshift-ingress (previous: false). I0423 14:40:15.097590 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-5c464d6b7f-vqh9q pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-5c464d6b7f-vqh9q\" is waiting to start: ContainerCreating" I0423 14:40:15.097604 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-5c464d6b7f-vqh9q\" is waiting to start: ContainerCreating" I0423 14:40:15.097615 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-6458c485df-44t2t pod in namespace openshift-ingress (previous: false). W0423 14:40:15.280175 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 14:40:15.298942 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-6458c485df-44t2t pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-6458c485df-44t2t\" is waiting to start: ContainerCreating" I0423 14:40:15.298957 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-6458c485df-44t2t\" is waiting to start: ContainerCreating" I0423 14:40:15.298982 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-4wf5j pod in namespace openshift-ingress-canary (previous: false). I0423 14:40:15.496872 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-4wf5j pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-4wf5j\" is waiting to start: ContainerCreating" I0423 14:40:15.496885 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-4wf5j\" is waiting to start: ContainerCreating" I0423 14:40:15.496924 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-q28j4 pod in namespace openshift-ingress-canary (previous: false). I0423 14:40:15.697472 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-q28j4 pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-q28j4\" is waiting to start: ContainerCreating" I0423 14:40:15.697487 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-q28j4\" is waiting to start: ContainerCreating" I0423 14:40:15.697521 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-xf9xk pod in namespace openshift-ingress-canary (previous: false). I0423 14:40:15.897932 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-xf9xk pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-xf9xk\" is waiting to start: ContainerCreating" I0423 14:40:15.897947 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-xf9xk\" is waiting to start: ContainerCreating" I0423 14:40:15.897964 1 tasks_processing.go:74] worker 40 stopped. I0423 14:40:15.898055 1 recorder.go:75] Recording events/openshift-dns-operator with fingerprint=e73266fb36ae20b810dc10a4f18b4db52139e1685e0f2a5a67554c5cd70a4089 I0423 14:40:15.898109 1 recorder.go:75] Recording events/openshift-dns with fingerprint=ec0a4aa93d0043d6819f7959fb7478183b89c21ea8d2555c97e8362254882944 I0423 14:40:15.898199 1 recorder.go:75] Recording events/openshift-image-registry with fingerprint=2b349a6654aa6421af05e2ec243f97b06ac5885dd5f14116411f3012370e75e7 I0423 14:40:15.898235 1 recorder.go:75] Recording events/openshift-ingress-operator with fingerprint=4f9d47f79e75b7e917aff42763a604783d65b3c4df17a185e9edd5fef29c97ff I0423 14:40:15.898304 1 recorder.go:75] Recording events/openshift-ingress with fingerprint=88c098673d64f39861fae47e68bbebbb5e71a92b8f609e4487144be002800958 I0423 14:40:15.898323 1 recorder.go:75] Recording events/openshift-ingress-canary with fingerprint=7e2de850adddad0d5d4a9854f057cf72a0e8a89a6e857b7c2137021efbbf1809 I0423 14:40:15.898475 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-929w7 with fingerprint=f53b91f54a91c716cb6e76f7e10b6e3bd0627ee23d4b2fcce72a1ceba242b8e5 I0423 14:40:15.898563 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-cps8z with fingerprint=de4180644d1316b44879c50dc64b00b6834e9c483e463618bbf418609eacee7f I0423 14:40:15.898638 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-wxthm with fingerprint=031f69983aecbde131ef0f1cc5abab09ed18bb1029992b4d8390299795662636 I0423 14:40:15.898748 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-7cb4f6c89-j52dm with fingerprint=c83963f04a1a39362bdc9b59b9d143cb88b77a2dd7bef780c69c4e751b534bf2 I0423 14:40:15.898838 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-7cb4f6c89-prfnv with fingerprint=a549aa7fb0189596eda73762967b59f776c2347beaeb2506fed2cf8f8e4cda9d I0423 14:40:15.898938 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-844ff948f4-qg22z with fingerprint=d56cffee2a03ac66c23cf9805e5e86b1994761462218e4dedbdfaf359554b543 I0423 14:40:15.898997 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-4wf5j with fingerprint=ea0fe75ea8fafe6a054431fd055d15c2cc9a438d755d073b40c3fc305a6d0076 I0423 14:40:15.899052 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-q28j4 with fingerprint=daf88308a2a1e6a4a53b7bdf404ef4262a4aa62f16dc28f912d4d965dd5511fb I0423 14:40:15.899108 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-xf9xk with fingerprint=fb37d765a2563f7264388b9bde11557e7c8ff6b847ea4c0769e5bde6cc28e565 I0423 14:40:15.899119 1 gather.go:177] gatherer "clusterconfig" function "operators_pods_and_events" took 4.645528468s to process 15 records W0423 14:40:16.280478 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. W0423 14:40:16.280501 1 gather_dvo_metrics.go:117] Unable to read metrics from endpoint "http://deployment-validation-operator-metrics.openshift-deployment-validation-operator.svc:8383": DVO metrics service was not available within the 5s timeout: context deadline exceeded I0423 14:40:16.280518 1 tasks_processing.go:74] worker 59 stopped. E0423 14:40:16.280527 1 gather.go:140] gatherer "clusterconfig" function "dvo_metrics" failed with the error: DVO metrics service was not available within the 5s timeout: context deadline exceeded I0423 14:40:16.280538 1 recorder.go:75] Recording config/dvo_metrics with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 W0423 14:40:16.280556 1 gather.go:155] issue recording gatherer "clusterconfig" function "dvo_metrics" result "config/dvo_metrics" because of the warning: warning: the record with the same fingerprint "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" was already recorded at path "config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt", recording another one with a different path "config/dvo_metrics" I0423 14:40:16.280568 1 gather.go:177] gatherer "clusterconfig" function "dvo_metrics" took 5.026216728s to process 1 records I0423 14:40:22.936449 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 14:40:23.690988 1 tasks_processing.go:74] worker 31 stopped. I0423 14:40:23.691026 1 recorder.go:75] Recording config/installplans with fingerprint=f17dbfacc3bfddf27ca3b213b39495434cd4c4e9e3dbd69566ffb3845bbcf539 I0423 14:40:23.691038 1 gather.go:177] gatherer "clusterconfig" function "install_plans" took 12.435795027s to process 1 records I0423 14:40:24.461931 1 tasks_processing.go:74] worker 37 stopped. I0423 14:40:24.462178 1 recorder.go:75] Recording config/serviceaccounts with fingerprint=751aa8b7d1208af5fd734152a53c8f2354708612b0cd50cfd9c595651a5cc8c8 I0423 14:40:24.462194 1 gather.go:177] gatherer "clusterconfig" function "service_accounts" took 13.205919059s to process 1 records E0423 14:40:24.462259 1 periodic.go:247] "Unhandled Error" err="clusterconfig failed after 13.21s with: function \"machine_healthchecks\" failed with an error, function \"support_secret\" failed with an error, function \"pod_network_connectivity_checks\" failed with an error, function \"machines\" failed with an error, function \"config_maps\" failed with an error, function \"ingress_certificates\" failed with an error, function \"dvo_metrics\" failed with an error" I0423 14:40:24.463373 1 controllerstatus.go:89] name=periodic-clusterconfig healthy=false reason=PeriodicGatherFailed message=Source clusterconfig could not be retrieved: function "machine_healthchecks" failed with an error, function "support_secret" failed with an error, function "pod_network_connectivity_checks" failed with an error, function "machines" failed with an error, function "config_maps" failed with an error, function "ingress_certificates" failed with an error, function "dvo_metrics" failed with an error I0423 14:40:24.463389 1 periodic.go:209] Running workloads gatherer I0423 14:40:24.463407 1 tasks_processing.go:45] number of workers: 2 I0423 14:40:24.463420 1 tasks_processing.go:69] worker 1 listening for tasks. I0423 14:40:24.463425 1 tasks_processing.go:71] worker 1 working on workload_info task. I0423 14:40:24.463433 1 tasks_processing.go:69] worker 0 listening for tasks. I0423 14:40:24.463449 1 tasks_processing.go:71] worker 0 working on helmchart_info task. I0423 14:40:24.489349 1 tasks_processing.go:74] worker 0 stopped. I0423 14:40:24.489373 1 gather.go:177] gatherer "workloads" function "helmchart_info" took 25.88554ms to process 0 records I0423 14:40:24.490093 1 gather_workloads_info.go:278] Loaded pods in 0s, will wait 22s for image data I0423 14:40:24.498211 1 gather_workloads_info.go:387] No image sha256:04c87c054a3f366a7dfbe0a93ebb0c80a098ee16842c5794b67c1202eec61996 (9ms) I0423 14:40:24.506113 1 gather_workloads_info.go:387] No image sha256:765f0d23b637f685f98a31bd47c131b03cf72a40761a3f9a9d6320faa3c33733 (8ms) I0423 14:40:24.514424 1 gather_workloads_info.go:387] No image sha256:2904a78e2eb73fd6a9bb94c105c2a056831fb4113fbb7b0607c50adc9d879c9b (8ms) I0423 14:40:24.523508 1 gather_workloads_info.go:387] No image sha256:1a2532940843248c57d52141185dd71fbc393ab28b65d48f682038632c1dbbad (9ms) I0423 14:40:24.532067 1 gather_workloads_info.go:387] No image sha256:c940ea87e7d133d75ba0002ef00c0806825eed3db8094cdb260d1bac18127733 (9ms) I0423 14:40:24.540148 1 gather_workloads_info.go:387] No image sha256:289816958633a763a72dbc44e1dad40466223164e7e253039514f0d974ea5d21 (8ms) I0423 14:40:24.547528 1 gather_workloads_info.go:387] No image sha256:a258c226562adb14e3a163a1940938526ee6a0928982a7667d85d9a7334ce639 (7ms) I0423 14:40:24.555494 1 gather_workloads_info.go:387] No image sha256:a498046d64605bcccee2440aa4f04a4602baaae263cf01d977ec5208e876b1fd (8ms) I0423 14:40:24.562960 1 gather_workloads_info.go:387] No image sha256:03cf4cd7ef1518610c6c7b3ad27d1622d82e98e3dc6e3f8e5d0fceb5c8d3786e (7ms) I0423 14:40:24.570208 1 gather_workloads_info.go:387] No image sha256:ce98d5d844bfc2ba8de1893866ad38166c95157d54abd8192b181e819bc50bb5 (7ms) I0423 14:40:24.604743 1 gather_workloads_info.go:387] No image sha256:56a85660a445eced5c79a595a0eccf590087c5672d50f49d4c25ad52f9a44f04 (35ms) I0423 14:40:24.698866 1 gather_workloads_info.go:387] No image sha256:4556896f77307821531ef91b7b7faccb82b824ea695693b2989f597f0deca038 (94ms) I0423 14:40:24.798009 1 gather_workloads_info.go:387] No image sha256:943018739e3db1763c3184b460dbc409e058abbac76d57b9927faad317be85e4 (99ms) I0423 14:40:24.898547 1 gather_workloads_info.go:387] No image sha256:47154813651033d59751fb655a384dbffb64dd26f10bd7f3be0c3128d0486356 (101ms) I0423 14:40:25.001792 1 gather_workloads_info.go:387] No image sha256:ca1344cb64140188b7cae7bbc51fb751566c0b0c97d5e39b5850e628032c4a5e (103ms) I0423 14:40:25.099035 1 gather_workloads_info.go:387] No image sha256:91828234f107c068c8a4966d08370ae7b73e637651dbc6d92c18c4553402c22c (97ms) I0423 14:40:25.205572 1 gather_workloads_info.go:387] No image sha256:2e564f336c77116053f34d4201d364d8da04e789cfffa0ea422574c95f2d6404 (107ms) I0423 14:40:25.298657 1 gather_workloads_info.go:387] No image sha256:695cf2f0cc07683c2a3ce1eaf3e56fe18abc6e2bac716f7d9843f5d173b9df52 (93ms) I0423 14:40:25.401752 1 gather_workloads_info.go:387] No image sha256:2e57e192c3c1240fd935dcd55c8fde5e70e78bf81d6176c96edf21fafe59f8ba (103ms) I0423 14:40:25.405052 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 14:40:25.498591 1 gather_workloads_info.go:387] No image sha256:36b9e89c3cfcf1ab9ae500486e38afb6862cba48cb0b4d84a09508ab8f3d299f (97ms) I0423 14:40:25.603518 1 gather_workloads_info.go:387] No image sha256:5a95c19d82767e0235b4edb4a0536482c816904897aae1dc3eb255cb52b87a9f (105ms) I0423 14:40:25.607186 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 14:40:25.698737 1 gather_workloads_info.go:387] No image sha256:653c666f842c13e0baae2e89a9b1efe0e2ef56f621ffb5b32005115d2a26ab8c (95ms) I0423 14:40:25.802251 1 gather_workloads_info.go:387] No image sha256:a56211d075aa43cbb491f669a5b2e46ee023dc95b7d51dbac28f463948c5ad61 (103ms) I0423 14:40:25.898064 1 gather_workloads_info.go:387] No image sha256:7adc1eab05d6724c76ba751f6df816b08d6e70b78dee9eb94fa6fd9690542c98 (96ms) I0423 14:40:26.002066 1 gather_workloads_info.go:387] No image sha256:e84cb128d930bd1ab867cc89b7b7bf2b2c0e41105ab93b5381069945b3ee9c57 (104ms) I0423 14:40:26.098919 1 gather_workloads_info.go:387] No image sha256:c15ca0c0ad60fe8757c2d5d1723fcdd7a1ed6c0251a90d22a7e6cae6811d01aa (97ms) I0423 14:40:26.201997 1 gather_workloads_info.go:387] No image sha256:0a99240166165eb5718e7516a43282fe32df9c7c5e809b31b58abe44e42ff94d (103ms) I0423 14:40:26.298728 1 gather_workloads_info.go:387] No image sha256:2598489729a4b258e4ecda4a06f6875133f2a10ced5c5241f8a57a8a05418e36 (97ms) I0423 14:40:26.398704 1 gather_workloads_info.go:387] No image sha256:a0105d1eb62cf6ac9e5e2ef28d3e89bf6dc514bc594fc7090fe5a5ee18a09c87 (100ms) I0423 14:40:26.498322 1 gather_workloads_info.go:387] No image sha256:5f0b67cfbbc381243fb91ccc17345b56d05f4d717c667e8c644e5bf05633ba71 (100ms) I0423 14:40:26.601005 1 gather_workloads_info.go:387] No image sha256:b3909bf664c77097f75b3768830863d642eed3815dab2bfb4415c771ca2d5007 (103ms) I0423 14:40:26.698762 1 gather_workloads_info.go:387] No image sha256:521712486e2c6e3c020dad6a1cb340db8e55665b69f7c208fab9cd9e965fd588 (98ms) I0423 14:40:26.804544 1 gather_workloads_info.go:387] No image sha256:7b31223098f08328f5ddea8e5b871dbbd5f5a61ec550e8956f66793c0c6031a9 (106ms) I0423 14:40:26.804569 1 tasks_processing.go:74] worker 1 stopped. E0423 14:40:26.804578 1 gather.go:140] gatherer "workloads" function "workload_info" failed with the error: no running pods found for the insights-runtime-extractor statefulset I0423 14:40:26.804868 1 recorder.go:75] Recording config/workload_info with fingerprint=39c2269f37684075c524464e0e32646adb03582bf27cfed7dd9cd6cfb04ef2c4 I0423 14:40:26.804882 1 gather.go:177] gatherer "workloads" function "workload_info" took 2.341137035s to process 1 records E0423 14:40:26.804923 1 periodic.go:247] "Unhandled Error" err="workloads failed after 2.341s with: function \"workload_info\" failed with an error" I0423 14:40:26.806019 1 controllerstatus.go:89] name=periodic-workloads healthy=false reason=PeriodicGatherFailed message=Source workloads could not be retrieved: function "workload_info" failed with an error I0423 14:40:26.806030 1 periodic.go:209] Running conditional gatherer I0423 14:40:26.811523 1 requests.go:294] Making HTTP GET request at: https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules I0423 14:40:26.817922 1 conditional_gatherer.go:107] Get "https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.128.0.10:44838->172.30.0.10:53: read: connection refused E0423 14:40:26.818144 1 conditional_gatherer.go:322] unable to update alerts cache: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 14:40:26.818199 1 conditional_gatherer.go:384] updating version cache for conditional gatherer I0423 14:40:26.823445 1 conditional_gatherer.go:392] cluster version is '4.20.8' E0423 14:40:26.823457 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 14:40:26.823463 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 14:40:26.823466 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 14:40:26.823469 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 14:40:26.823472 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 14:40:26.823475 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 14:40:26.823478 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 14:40:26.823481 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 14:40:26.823483 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing I0423 14:40:26.823497 1 tasks_processing.go:45] number of workers: 3 I0423 14:40:26.823503 1 tasks_processing.go:69] worker 2 listening for tasks. I0423 14:40:26.823507 1 tasks_processing.go:71] worker 2 working on conditional_gatherer_rules task. I0423 14:40:26.823571 1 tasks_processing.go:69] worker 1 listening for tasks. I0423 14:40:26.823581 1 tasks_processing.go:71] worker 1 working on remote_configuration task. I0423 14:40:26.823584 1 recorder.go:75] Recording insights-operator/conditional-gatherer-rules with fingerprint=7034af97e7e41c22e4b775abdd4b9066c8ebb19da33eb7f69f39bfd2eb5f6406 I0423 14:40:26.823585 1 tasks_processing.go:71] worker 2 working on rapid_container_logs task. I0423 14:40:26.823593 1 gather.go:177] gatherer "conditional" function "conditional_gatherer_rules" took 798ns to process 1 records I0423 14:40:26.823596 1 tasks_processing.go:69] worker 0 listening for tasks. I0423 14:40:26.823605 1 tasks_processing.go:74] worker 0 stopped. I0423 14:40:26.823618 1 recorder.go:75] Recording insights-operator/remote-configuration with fingerprint=0394430c431eec4d48bb1811a90918e95161d2282c59af26f2473613cc0959db I0423 14:40:26.823625 1 gather.go:177] gatherer "conditional" function "remote_configuration" took 1.127µs to process 1 records I0423 14:40:26.823630 1 tasks_processing.go:74] worker 1 stopped. I0423 14:40:26.823860 1 tasks_processing.go:74] worker 2 stopped. I0423 14:40:26.823874 1 gather.go:177] gatherer "conditional" function "rapid_container_logs" took 260.306µs to process 0 records I0423 14:40:26.823894 1 controllerstatus.go:89] name=periodic-conditional healthy=false reason=NotAvailable message=Get "https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.128.0.10:44838->172.30.0.10:53: read: connection refused I0423 14:40:26.823928 1 recorder.go:75] Recording insights-operator/remote-configuration.json with fingerprint=359de9c990c741675cec72fda96b5c3682221efdb4799f5eaa6e9805bcd3b5c1 I0423 14:40:26.844695 1 recorder.go:75] Recording insights-operator/gathers with fingerprint=7c659364247dd9835f62465d7dcf808a266c8d79c9f2499e62cfc8bde3b0619d I0423 14:40:26.844850 1 diskrecorder.go:70] Writing 110 records to /var/lib/insights-operator/insights-2026-04-23-144026.tar.gz I0423 14:40:26.851433 1 diskrecorder.go:51] Wrote 110 records to disk in 6ms I0423 14:40:26.851462 1 periodic.go:278] Gathering cluster info every 2h0m0s I0423 14:40:26.851475 1 periodic.go:279] Configuration is dataReporting: interval: 2h0m0s, uploadEndpoint: https://console.redhat.com/api/ingress/v1/upload, storagePath: /var/lib/insights-operator, downloadEndpoint: https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports, conditionalGathererEndpoint: https://console.redhat.com/api/gathering/v2/%s/gathering_rules, obfuscation: [] sca: disabled: false, endpoint: https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates, interval: 8h0m0s alerting: disabled: false clusterTransfer: endpoint: https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/, interval: 12h0m0s proxy: httpProxy: , httpsProxy: , noProxy: I0423 14:40:39.741548 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 14:41:41.253235 1 diskrecorder.go:223] Found files to send: insights-2026-04-23-144026.tar.gz I0423 14:41:41.253261 1 insightsuploader.go:150] Checking archives to upload periodically every 16m1.99328338s I0423 14:41:41.253270 1 insightsuploader.go:165] Uploading latest report since 0001-01-01T00:00:00Z I0423 14:41:41.264678 1 requests.go:46] Uploading application/vnd.redhat.openshift.periodic to https://console.redhat.com/api/ingress/v1/upload I0423 14:41:41.268317 1 requests.go:49] Unable to build a request, possible invalid token: Post "https://console.redhat.com/api/ingress/v1/upload": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.128.0.10:34981->172.30.0.10:53: read: connection refused I0423 14:41:41.268345 1 insightsuploader.go:169] Unable to upload report after 10ms: unable to build request to connect to Insights server: Post "https://console.redhat.com/api/ingress/v1/upload": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.128.0.10:34981->172.30.0.10:53: read: connection refused I0423 14:41:41.268358 1 controllerstatus.go:80] name=insightsuploader healthy=false reason=UploadFailed message=Unable to report: unable to build request to connect to Insights server: Post "https://console.redhat.com/api/ingress/v1/upload": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.128.0.10:34981->172.30.0.10:53: read: connection refused I0423 14:41:45.459510 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.crt" has been created (hash="ac8b83edf8615db1bad0ce594634fc3c59fa105938eb9cbc049a1befbe30fd88") W0423 14:41:45.459544 1 builder.go:160] Restart triggered because of file /var/run/secrets/serving-cert/tls.crt was created I0423 14:41:45.459585 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.key" has been created (hash="965bef36b5fa0e0cb864d1eb1440a29de029c0b96be383906520c8806d26c29f") I0423 14:41:45.459592 1 simple_featuregate_reader.go:177] Shutting down feature-gate-detector