W0420 11:46:30.926885 1 cmd.go:257] Using insecure, self-signed certificates I0420 11:46:32.110004 1 start.go:138] Unable to read service ca bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0420 11:46:32.110327 1 observer_polling.go:159] Starting file observer I0420 11:46:32.558322 1 operator.go:60] Starting insights-operator v0.0.0-master+$Format:%H$ I0420 11:46:32.558506 1 legacy_config.go:327] Current config: {"report":false,"storagePath":"/var/lib/insights-operator","interval":"2h","endpoint":"https://console.redhat.com/api/ingress/v1/upload","conditionalGathererEndpoint":"https://console.redhat.com/api/gathering/v2/%s/gathering_rules","pull_report":{"endpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports","delay":"60s","timeout":"3000s","min_retry":"30s"},"impersonate":"system:serviceaccount:openshift-insights:gather","enableGlobalObfuscation":false,"ocm":{"scaEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates","scaInterval":"8h","scaDisabled":false,"clusterTransferEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/","clusterTransferInterval":"12h"},"disableInsightsAlerts":false,"processingStatusEndpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status","reportEndpointTechPreview":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/report"} I0420 11:46:32.559031 1 simple_featuregate_reader.go:171] Starting feature-gate-detector I0420 11:46:32.559071 1 secure_serving.go:57] Forcing use of http/1.1 only W0420 11:46:32.559102 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. W0420 11:46:32.559107 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. W0420 11:46:32.559112 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. W0420 11:46:32.559115 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. W0420 11:46:32.559118 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. W0420 11:46:32.559120 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. I0420 11:46:32.562909 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-insights", Name:"insights-operator", UID:"f436fd2f-088e-47bb-80d6-b60740c86fa6", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BuildCSIVolumes", "CPMSMachineNamePrefix", "ConsolePluginContentSecurityPolicy", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "ImageVolume", "IngressControllerLBSubnetsAWS", "KMSv1", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "MetricsCollectionProfiles", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "PinnedImages", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SetEIPForNLBIngressController", "SigstoreImageVerification", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiDisk", "VSphereMultiNetworks"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSServiceLBNetworkSecurityGroup", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureMultiDisk", "BootImageSkewEnforcement", "BootcNodeManagement", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "Example2", "ExternalSnapshotMetadata", "GCPClusterHostedDNS", "GCPClusterHostedDNSInstall", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "ImageModeStatusReporting", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryptionProvider", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "ManagedBootImagesAzure", "ManagedBootImagesvSphere", "MaxUnavailableStatefulSet", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "MultiDiskSetup", "MutatingAdmissionPolicy", "NewOLM", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterOperations", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PreconfiguredUDNAddresses", "SELinuxMount", "ShortCertRotation", "SignatureStores", "SigstoreImageVerificationPKI", "TranslateStreamCloseWebsocketRequests", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VolumeAttributesClass", "VolumeGroupSnapshot"}} I0420 11:46:32.562890 1 operator.go:125] FeatureGates initialized: knownFeatureGates=[AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BuildCSIVolumes CPMSMachineNamePrefix ConsolePluginContentSecurityPolicy ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GatewayAPI GatewayAPIController HighlyAvailableArbiter ImageVolume IngressControllerLBSubnetsAWS KMSv1 MachineConfigNodes ManagedBootImages ManagedBootImagesAWS MetricsCollectionProfiles NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation PinnedImages ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SetEIPForNLBIngressController SigstoreImageVerification StoragePerformantSecurityPolicy UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiDisk VSphereMultiNetworks AWSClusterHostedDNS AWSClusterHostedDNSInstall AWSDedicatedHosts AWSServiceLBNetworkSecurityGroup AutomatedEtcdBackup AzureClusterHostedDNSInstall AzureDedicatedHosts AzureMultiDisk BootImageSkewEnforcement BootcNodeManagement ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterMonitoringConfig ClusterVersionOperatorConfiguration DNSNameResolver DualReplica DyanmicServiceEndpointIBMCloud DynamicResourceAllocation EtcdBackendQuota EventedPLEG Example Example2 ExternalSnapshotMetadata GCPClusterHostedDNS GCPClusterHostedDNSInstall GCPCustomAPIEndpoints GCPCustomAPIEndpointsInstall ImageModeStatusReporting ImageStreamImportMode IngressControllerDynamicConfigurationManager InsightsConfig InsightsConfigAPI InsightsOnDemandDataGather IrreconcilableMachineConfig KMSEncryptionProvider MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController ManagedBootImagesAzure ManagedBootImagesvSphere MaxUnavailableStatefulSet MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure MultiDiskSetup MutatingAdmissionPolicy NewOLM NewOLMCatalogdAPIV1Metas NewOLMOwnSingleNamespace NewOLMPreflightPermissionChecks NewOLMWebhookProviderOpenshiftServiceCA NoRegistryClusterOperations NodeSwap NutanixMultiSubnets OVNObservability OpenShiftPodSecurityAdmission PreconfiguredUDNAddresses SELinuxMount ShortCertRotation SignatureStores SigstoreImageVerificationPKI TranslateStreamCloseWebsocketRequests VSphereConfigurableMaxAllowedBlockVolumesPerNode VSphereHostVMGroupZonal VSphereMixedNodeEnv VolumeAttributesClass VolumeGroupSnapshot] I0420 11:46:32.564644 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController I0420 11:46:32.564667 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController I0420 11:46:32.564666 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0420 11:46:32.564674 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0420 11:46:32.564685 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0420 11:46:32.564687 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0420 11:46:32.564908 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/tmp/serving-cert-3982279924/tls.crt::/tmp/serving-cert-3982279924/tls.key" I0420 11:46:32.565165 1 secure_serving.go:213] Serving securely on [::]:8443 I0420 11:46:32.565191 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" W0420 11:46:32.568538 1 configmapobserver.go:64] Cannot get the configuration config map: configmaps "insights-config" not found. Default configuration is used. I0420 11:46:32.568564 1 secretconfigobserver.go:216] Legacy configuration set: enabled=false endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=false reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0420 11:46:32.568597 1 base_controller.go:76] Waiting for caches to sync for ConfigController I0420 11:46:32.574158 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0420 11:46:32.574173 1 secretconfigobserver.go:204] Legacy configuration updated: enabled=true endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=true reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0420 11:46:32.578744 1 secretconfigobserver.go:119] support secret does not exist I0420 11:46:32.583170 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0420 11:46:32.587635 1 secretconfigobserver.go:119] support secret does not exist I0420 11:46:32.591002 1 recorder.go:161] Pruning old reports every 6h8m1s, max age is 288h0m0s I0420 11:46:32.596004 1 periodic.go:209] Running clusterconfig gatherer I0420 11:46:32.596004 1 controllerstatus.go:80] name=insightsuploader healthy=true reason= message= I0420 11:46:32.596031 1 insightsuploader.go:86] Reporting status periodically to https://console.redhat.com/api/ingress/v1/upload every 2h0m0s, starting in 1m30s I0420 11:46:32.596054 1 controllerstatus.go:80] name=insightsreport healthy=true reason= message= I0420 11:46:32.596061 1 tasks_processing.go:45] number of workers: 64 I0420 11:46:32.596064 1 insightsreport.go:296] Starting report retriever I0420 11:46:32.596070 1 insightsreport.go:298] Insights analysis reports will be downloaded from the https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports endpoint with a delay of 1m0s I0420 11:46:32.596085 1 tasks_processing.go:69] worker 4 listening for tasks. I0420 11:46:32.596105 1 tasks_processing.go:69] worker 18 listening for tasks. I0420 11:46:32.596107 1 tasks_processing.go:69] worker 5 listening for tasks. I0420 11:46:32.596113 1 tasks_processing.go:71] worker 18 working on openstack_version task. I0420 11:46:32.596114 1 tasks_processing.go:69] worker 0 listening for tasks. I0420 11:46:32.596119 1 tasks_processing.go:69] worker 6 listening for tasks. I0420 11:46:32.596118 1 tasks_processing.go:69] worker 35 listening for tasks. I0420 11:46:32.596121 1 tasks_processing.go:69] worker 34 listening for tasks. I0420 11:46:32.596113 1 tasks_processing.go:71] worker 4 working on aggregated_monitoring_cr_names task. I0420 11:46:32.596129 1 tasks_processing.go:69] worker 37 listening for tasks. I0420 11:46:32.596130 1 tasks_processing.go:69] worker 2 listening for tasks. I0420 11:46:32.596130 1 tasks_processing.go:69] worker 9 listening for tasks. I0420 11:46:32.596134 1 tasks_processing.go:69] worker 10 listening for tasks. I0420 11:46:32.596124 1 tasks_processing.go:69] worker 36 listening for tasks. I0420 11:46:32.596139 1 tasks_processing.go:69] worker 8 listening for tasks. I0420 11:46:32.596143 1 tasks_processing.go:69] worker 3 listening for tasks. I0420 11:46:32.596146 1 tasks_processing.go:69] worker 26 listening for tasks. I0420 11:46:32.596142 1 tasks_processing.go:69] worker 52 listening for tasks. I0420 11:46:32.596137 1 tasks_processing.go:69] worker 38 listening for tasks. I0420 11:46:32.596151 1 tasks_processing.go:69] worker 7 listening for tasks. I0420 11:46:32.596154 1 tasks_processing.go:69] worker 30 listening for tasks. I0420 11:46:32.596149 1 tasks_processing.go:69] worker 53 listening for tasks. I0420 11:46:32.596155 1 tasks_processing.go:69] worker 54 listening for tasks. I0420 11:46:32.596161 1 tasks_processing.go:69] worker 39 listening for tasks. I0420 11:46:32.596165 1 tasks_processing.go:69] worker 23 listening for tasks. I0420 11:46:32.596167 1 tasks_processing.go:69] worker 59 listening for tasks. I0420 11:46:32.596169 1 tasks_processing.go:69] worker 60 listening for tasks. I0420 11:46:32.596173 1 tasks_processing.go:69] worker 40 listening for tasks. I0420 11:46:32.596175 1 tasks_processing.go:69] worker 24 listening for tasks. I0420 11:46:32.596168 1 tasks_processing.go:69] worker 27 listening for tasks. I0420 11:46:32.596177 1 tasks_processing.go:69] worker 22 listening for tasks. I0420 11:46:32.596183 1 tasks_processing.go:69] worker 41 listening for tasks. I0420 11:46:32.596185 1 tasks_processing.go:69] worker 16 listening for tasks. I0420 11:46:32.596187 1 tasks_processing.go:69] worker 17 listening for tasks. I0420 11:46:32.596189 1 tasks_processing.go:69] worker 42 listening for tasks. I0420 11:46:32.596189 1 tasks_processing.go:69] worker 29 listening for tasks. I0420 11:46:32.596186 1 tasks_processing.go:69] worker 62 listening for tasks. I0420 11:46:32.596194 1 tasks_processing.go:69] worker 14 listening for tasks. I0420 11:46:32.596199 1 tasks_processing.go:69] worker 21 listening for tasks. I0420 11:46:32.596199 1 tasks_processing.go:69] worker 63 listening for tasks. I0420 11:46:32.596203 1 tasks_processing.go:71] worker 38 working on operators_pods_and_events task. I0420 11:46:32.596205 1 tasks_processing.go:69] worker 57 listening for tasks. I0420 11:46:32.596209 1 tasks_processing.go:71] worker 36 working on nodes task. I0420 11:46:32.596210 1 tasks_processing.go:71] worker 63 working on certificate_signing_requests task. I0420 11:46:32.596213 1 tasks_processing.go:71] worker 57 working on openshift_logging task. I0420 11:46:32.596216 1 tasks_processing.go:71] worker 34 working on mutating_webhook_configurations task. I0420 11:46:32.596218 1 tasks_processing.go:69] worker 46 listening for tasks. I0420 11:46:32.596194 1 tasks_processing.go:69] worker 25 listening for tasks. I0420 11:46:32.596180 1 tasks_processing.go:69] worker 61 listening for tasks. I0420 11:46:32.596215 1 tasks_processing.go:69] worker 15 listening for tasks. I0420 11:46:32.596233 1 tasks_processing.go:71] worker 9 working on proxies task. I0420 11:46:32.596230 1 tasks_processing.go:71] worker 6 working on support_secret task. I0420 11:46:32.596236 1 tasks_processing.go:71] worker 5 working on ceph_cluster task. I0420 11:46:32.596244 1 tasks_processing.go:69] worker 49 listening for tasks. I0420 11:46:32.596245 1 tasks_processing.go:71] worker 10 working on networks task. I0420 11:46:32.596253 1 tasks_processing.go:69] worker 50 listening for tasks. I0420 11:46:32.596258 1 tasks_processing.go:69] worker 58 listening for tasks. I0420 11:46:32.596276 1 tasks_processing.go:71] worker 59 working on machine_configs task. I0420 11:46:32.596288 1 tasks_processing.go:71] worker 58 working on machines task. I0420 11:46:32.596288 1 tasks_processing.go:71] worker 0 working on lokistack task. I0420 11:46:32.596163 1 tasks_processing.go:69] worker 55 listening for tasks. I0420 11:46:32.596492 1 tasks_processing.go:71] worker 7 working on active_alerts task. I0420 11:46:32.597044 1 tasks_processing.go:69] worker 32 listening for tasks. I0420 11:46:32.597135 1 tasks_processing.go:69] worker 31 listening for tasks. I0420 11:46:32.597158 1 tasks_processing.go:71] worker 37 working on pod_network_connectivity_checks task. I0420 11:46:32.597272 1 tasks_processing.go:71] worker 30 working on infrastructures task. I0420 11:46:32.596085 1 tasks_processing.go:69] worker 11 listening for tasks. I0420 11:46:32.596211 1 tasks_processing.go:71] worker 8 working on install_plans task. I0420 11:46:32.596499 1 tasks_processing.go:71] worker 55 working on image task. I0420 11:46:32.597386 1 tasks_processing.go:71] worker 53 working on image_pruners task. I0420 11:46:32.596151 1 tasks_processing.go:69] worker 51 listening for tasks. I0420 11:46:32.596181 1 tasks_processing.go:69] worker 28 listening for tasks. I0420 11:46:32.596201 1 tasks_processing.go:71] worker 52 working on olm_operators task. I0420 11:46:32.596191 1 tasks_processing.go:69] worker 20 listening for tasks. I0420 11:46:32.596188 1 tasks_processing.go:69] worker 56 listening for tasks. I0420 11:46:32.596197 1 tasks_processing.go:69] worker 43 listening for tasks. I0420 11:46:32.596202 1 tasks_processing.go:71] worker 14 working on pdbs task. I0420 11:46:32.596203 1 tasks_processing.go:69] worker 44 listening for tasks. I0420 11:46:32.596208 1 tasks_processing.go:71] worker 2 working on ingress_certificates task. I0420 11:46:32.596211 1 tasks_processing.go:69] worker 45 listening for tasks. I0420 11:46:32.596212 1 tasks_processing.go:71] worker 35 working on openshift_machine_api_events task. I0420 11:46:32.596222 1 tasks_processing.go:71] worker 3 working on oauths task. I0420 11:46:32.596227 1 tasks_processing.go:71] worker 26 working on machine_config_pools task. I0420 11:46:32.596230 1 tasks_processing.go:69] worker 47 listening for tasks. I0420 11:46:32.596237 1 tasks_processing.go:69] worker 48 listening for tasks. I0420 11:46:32.596156 1 tasks_processing.go:69] worker 19 listening for tasks. I0420 11:46:32.596486 1 tasks_processing.go:71] worker 24 working on nodenetworkstates task. I0420 11:46:32.598011 1 tasks_processing.go:71] worker 21 working on metrics task. W0420 11:46:32.598168 1 gather_most_recent_metrics.go:64] Unable to load metrics client, no metrics will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0420 11:46:32.598221 1 gather.go:177] gatherer "clusterconfig" function "metrics" took 145.551µs to process 0 records I0420 11:46:32.598272 1 tasks_processing.go:71] worker 32 working on monitoring_persistent_volumes task. I0420 11:46:32.596201 1 tasks_processing.go:71] worker 62 working on dvo_metrics task. I0420 11:46:32.598404 1 tasks_processing.go:71] worker 16 working on silenced_alerts task. I0420 11:46:32.598429 1 tasks_processing.go:71] worker 44 working on number_of_pods_and_netnamespaces_with_sdn_annotations task. W0420 11:46:32.598435 1 gather_silenced_alerts.go:38] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0420 11:46:32.597332 1 tasks_processing.go:71] worker 39 working on storage_cluster task. I0420 11:46:32.598583 1 gather.go:177] gatherer "clusterconfig" function "silenced_alerts" took 134.403µs to process 0 records I0420 11:46:32.596123 1 tasks_processing.go:69] worker 1 listening for tasks. I0420 11:46:32.597498 1 tasks_processing.go:71] worker 23 working on config_maps task. I0420 11:46:32.597499 1 tasks_processing.go:71] worker 60 working on openstack_controlplanes task. I0420 11:46:32.599130 1 tasks_processing.go:71] worker 47 working on overlapping_namespace_uids task. I0420 11:46:32.599145 1 tasks_processing.go:71] worker 28 working on cluster_apiserver task. I0420 11:46:32.599159 1 tasks_processing.go:71] worker 56 working on container_runtime_configs task. I0420 11:46:32.599166 1 tasks_processing.go:71] worker 20 working on image_registries task. I0420 11:46:32.599174 1 tasks_processing.go:71] worker 11 working on jaegers task. I0420 11:46:32.599198 1 tasks_processing.go:71] worker 43 working on cost_management_metrics_configs task. I0420 11:46:32.599226 1 tasks_processing.go:71] worker 21 working on operators task. I0420 11:46:32.599241 1 tasks_processing.go:71] worker 45 working on service_accounts task. I0420 11:46:32.599265 1 tasks_processing.go:71] worker 48 working on node_logs task. I0420 11:46:32.597511 1 tasks_processing.go:71] worker 54 working on storage_classes task. I0420 11:46:32.599364 1 tasks_processing.go:71] worker 51 working on validating_webhook_configurations task. I0420 11:46:32.597524 1 tasks_processing.go:69] worker 33 listening for tasks. I0420 11:46:32.597573 1 tasks_processing.go:71] worker 27 working on machine_sets task. I0420 11:46:32.597596 1 tasks_processing.go:71] worker 61 working on sap_datahubs task. I0420 11:46:32.597606 1 tasks_processing.go:71] worker 46 working on machine_autoscalers task. I0420 11:46:32.597624 1 tasks_processing.go:71] worker 22 working on ingress task. I0420 11:46:32.599769 1 tasks_processing.go:71] worker 0 working on openstack_dataplanenodesets task. I0420 11:46:32.599777 1 tasks_processing.go:71] worker 1 working on clusterroles task. I0420 11:46:32.599790 1 gather.go:177] gatherer "clusterconfig" function "lokistack" took 3.466532ms to process 0 records I0420 11:46:32.597626 1 tasks_processing.go:71] worker 40 working on schedulers task. I0420 11:46:32.597648 1 tasks_processing.go:71] worker 25 working on crds task. I0420 11:46:32.597671 1 tasks_processing.go:71] worker 41 working on tsdb_status task. I0420 11:46:32.597693 1 tasks_processing.go:71] worker 42 working on container_images task. W0420 11:46:32.599939 1 gather_prometheus_tsdb_status.go:38] Unable to load metrics client, tsdb status cannot be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0420 11:46:32.599949 1 tasks_processing.go:74] worker 41 stopped. I0420 11:46:32.599960 1 gather.go:177] gatherer "clusterconfig" function "tsdb_status" took 22.613µs to process 0 records I0420 11:46:32.597696 1 tasks_processing.go:71] worker 49 working on machine_healthchecks task. I0420 11:46:32.597703 1 tasks_processing.go:71] worker 50 working on sap_config task. I0420 11:46:32.597761 1 tasks_processing.go:71] worker 15 working on openstack_dataplanedeployments task. I0420 11:46:32.597762 1 tasks_processing.go:71] worker 17 working on nodenetworkconfigurationpolicies task. I0420 11:46:32.597777 1 tasks_processing.go:71] worker 29 working on authentication task. I0420 11:46:32.596180 1 tasks_processing.go:69] worker 13 listening for tasks. I0420 11:46:32.600340 1 tasks_processing.go:74] worker 13 stopped. I0420 11:46:32.596175 1 tasks_processing.go:69] worker 12 listening for tasks. I0420 11:46:32.600347 1 tasks_processing.go:74] worker 12 stopped. W0420 11:46:32.597196 1 gather_active_alerts.go:54] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0420 11:46:32.600356 1 tasks_processing.go:74] worker 7 stopped. I0420 11:46:32.598481 1 tasks_processing.go:71] worker 31 working on feature_gates task. I0420 11:46:32.600362 1 gather.go:177] gatherer "clusterconfig" function "active_alerts" took 3.309384ms to process 0 records I0420 11:46:32.598554 1 tasks_processing.go:71] worker 16 working on sap_pods task. I0420 11:46:32.599486 1 tasks_processing.go:71] worker 19 working on qemu_kubevirt_launcher_logs task. I0420 11:46:32.599526 1 tasks_processing.go:71] worker 33 working on version task. I0420 11:46:32.601505 1 tasks_processing.go:74] worker 57 stopped. I0420 11:46:32.601518 1 gather.go:177] gatherer "clusterconfig" function "openshift_logging" took 5.280787ms to process 0 records I0420 11:46:32.601689 1 controller.go:128] Initializing last reported time to 0001-01-01T00:00:00Z I0420 11:46:32.601705 1 controller.go:212] Source periodic-workloads *controllerstatus.Simple is not ready I0420 11:46:32.601710 1 controller.go:212] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0420 11:46:32.601713 1 controller.go:212] Source periodic-conditional *controllerstatus.Simple is not ready I0420 11:46:32.601725 1 controller.go:489] The operator is still being initialized I0420 11:46:32.601733 1 controller.go:512] The operator is healthy I0420 11:46:32.603987 1 tasks_processing.go:74] worker 18 stopped. I0420 11:46:32.604005 1 gather.go:177] gatherer "clusterconfig" function "openstack_version" took 7.866357ms to process 0 records I0420 11:46:32.604015 1 gather.go:177] gatherer "clusterconfig" function "ceph_cluster" took 7.752887ms to process 0 records E0420 11:46:32.604023 1 gather.go:140] gatherer "clusterconfig" function "support_secret" failed with the error: secrets "support" not found I0420 11:46:32.604031 1 gather.go:177] gatherer "clusterconfig" function "support_secret" took 7.768703ms to process 0 records I0420 11:46:32.604040 1 tasks_processing.go:74] worker 6 stopped. I0420 11:46:32.604046 1 tasks_processing.go:74] worker 5 stopped. I0420 11:46:32.604127 1 tasks_processing.go:74] worker 9 stopped. I0420 11:46:32.604208 1 recorder.go:75] Recording config/proxy with fingerprint=ff9860328082d31a8382180ead3c16f19e384de12d6c519c950994e18dca6872 I0420 11:46:32.604220 1 gather.go:177] gatherer "clusterconfig" function "proxies" took 7.881627ms to process 1 records I0420 11:46:32.614428 1 tasks_processing.go:74] worker 24 stopped. I0420 11:46:32.614438 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkstates" took 16.467041ms to process 0 records I0420 11:46:32.614446 1 gather.go:177] gatherer "clusterconfig" function "container_runtime_configs" took 15.259099ms to process 0 records I0420 11:46:32.614450 1 gather.go:177] gatherer "clusterconfig" function "cost_management_metrics_configs" took 15.230817ms to process 0 records I0420 11:46:32.614454 1 gather.go:177] gatherer "clusterconfig" function "jaegers" took 15.261692ms to process 0 records I0420 11:46:32.614459 1 tasks_processing.go:74] worker 11 stopped. I0420 11:46:32.614462 1 tasks_processing.go:74] worker 56 stopped. I0420 11:46:32.614465 1 tasks_processing.go:74] worker 43 stopped. I0420 11:46:32.614486 1 tasks_processing.go:74] worker 58 stopped. E0420 11:46:32.614499 1 gather.go:140] gatherer "clusterconfig" function "machines" failed with the error: machines.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machines" in API group "machine.openshift.io" at the cluster scope I0420 11:46:32.614510 1 gather.go:177] gatherer "clusterconfig" function "machines" took 18.189459ms to process 0 records I0420 11:46:32.621304 1 tasks_processing.go:74] worker 30 stopped. I0420 11:46:32.621798 1 recorder.go:75] Recording config/infrastructure with fingerprint=7ff622abaa52410c310dbd29e933561db704d80910f97b284bb953d77a5020d7 I0420 11:46:32.621811 1 gather.go:177] gatherer "clusterconfig" function "infrastructures" took 24.002499ms to process 1 records E0420 11:46:32.621822 1 gather.go:140] gatherer "clusterconfig" function "pod_network_connectivity_checks" failed with the error: the server could not find the requested resource (get podnetworkconnectivitychecks.controlplane.operator.openshift.io) I0420 11:46:32.621827 1 gather.go:177] gatherer "clusterconfig" function "pod_network_connectivity_checks" took 24.237956ms to process 0 records I0420 11:46:32.621832 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanenodesets" took 21.686213ms to process 0 records I0420 11:46:32.621836 1 gather.go:177] gatherer "clusterconfig" function "machine_sets" took 21.937832ms to process 0 records I0420 11:46:32.621842 1 gather.go:177] gatherer "clusterconfig" function "sap_datahubs" took 21.866886ms to process 0 records I0420 11:46:32.621843 1 tasks_processing.go:74] worker 37 stopped. I0420 11:46:32.621849 1 tasks_processing.go:74] worker 0 stopped. I0420 11:46:32.621847 1 gather.go:177] gatherer "clusterconfig" function "machine_autoscalers" took 21.836304ms to process 0 records I0420 11:46:32.621857 1 tasks_processing.go:74] worker 61 stopped. I0420 11:46:32.621857 1 tasks_processing.go:74] worker 27 stopped. I0420 11:46:32.621853 1 tasks_processing.go:74] worker 46 stopped. I0420 11:46:32.621922 1 tasks_processing.go:74] worker 3 stopped. I0420 11:46:32.622064 1 recorder.go:75] Recording config/oauth with fingerprint=25139b86ad84d926eface8357ad96383b60ae161fa399116968490a02f1aa2b6 I0420 11:46:32.622080 1 gather.go:177] gatherer "clusterconfig" function "oauths" took 23.022889ms to process 1 records I0420 11:46:32.622168 1 tasks_processing.go:74] worker 34 stopped. I0420 11:46:32.622271 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/aws-pod-identity with fingerprint=c39e29a041311461953817ec55225a0219695b4d556c8137e855c489b608c6dd I0420 11:46:32.622312 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-podimagespec-mutation with fingerprint=6d816e08e7a95759deda933c601558734221e301a66ac6b24fc31d2f81a671cf I0420 11:46:32.622350 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-service-mutation with fingerprint=1cb9d2ac45a13404ee9eb5e09b73d7576ad3e551679cef75da563defa38592c8 I0420 11:46:32.622361 1 gather.go:177] gatherer "clusterconfig" function "mutating_webhook_configurations" took 25.376098ms to process 3 records I0420 11:46:32.624624 1 tasks_processing.go:74] worker 49 stopped. E0420 11:46:32.624642 1 gather.go:140] gatherer "clusterconfig" function "machine_healthchecks" failed with the error: machinehealthchecks.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machinehealthchecks" in API group "machine.openshift.io" at the cluster scope I0420 11:46:32.624659 1 gather.go:177] gatherer "clusterconfig" function "machine_healthchecks" took 24.57723ms to process 0 records I0420 11:46:32.624827 1 tasks_processing.go:74] worker 53 stopped. I0420 11:46:32.625186 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/imagepruner/cluster with fingerprint=7a1bcc0eff2904f514a13474fa3fc2225f64e0a2e8f531449f65007dc7ae8f83 I0420 11:46:32.625206 1 gather.go:177] gatherer "clusterconfig" function "image_pruners" took 27.141006ms to process 1 records I0420 11:46:32.625289 1 tasks_processing.go:74] worker 55 stopped. I0420 11:46:32.625307 1 recorder.go:75] Recording config/image with fingerprint=ec9802565aacac842c553b1140bf92243f28df680bf5bdfa01095d2fa3c2e680 I0420 11:46:32.625315 1 gather.go:177] gatherer "clusterconfig" function "image" took 27.627259ms to process 1 records I0420 11:46:32.625390 1 tasks_processing.go:74] worker 14 stopped. I0420 11:46:32.625436 1 recorder.go:75] Recording config/pdbs/openshift-image-registry/image-registry with fingerprint=2ceef48460fdbe12cce2ee8617018eaf38f815dfa9e9f721110f1354c795fdb3 I0420 11:46:32.625466 1 recorder.go:75] Recording config/pdbs/openshift-ingress/router-default with fingerprint=e6813cd95c227742999796b210debc03ba6a5266fbfc80b18c17fba8e727bc9c I0420 11:46:32.625493 1 recorder.go:75] Recording config/pdbs/openshift-operator-lifecycle-manager/packageserver-pdb with fingerprint=60b64ff05779b5dcaed4333a57827d39d0ff6cbfeb13bebb9f0737eb94c99ea2 I0420 11:46:32.625503 1 gather.go:177] gatherer "clusterconfig" function "pdbs" took 26.977541ms to process 3 records I0420 11:46:32.625510 1 gather.go:177] gatherer "clusterconfig" function "node_logs" took 26.087495ms to process 0 records I0420 11:46:32.625549 1 tasks_processing.go:74] worker 48 stopped. I0420 11:46:32.625623 1 tasks_processing.go:74] worker 28 stopped. I0420 11:46:32.625662 1 recorder.go:75] Recording config/apiserver with fingerprint=1e5fe8a75b260ee03984ffcf72ddca604b015fa51cc10f80e8efa721aa54f553 I0420 11:46:32.625673 1 gather.go:177] gatherer "clusterconfig" function "cluster_apiserver" took 26.225099ms to process 1 records I0420 11:46:32.625741 1 tasks_processing.go:74] worker 54 stopped. I0420 11:46:32.625753 1 recorder.go:75] Recording config/storage/storageclasses/gp2-csi with fingerprint=c72c25e7c59f02a3518adda5c530c55b777dc6462afa1ad4c9668fb460fd2553 I0420 11:46:32.625770 1 recorder.go:75] Recording config/storage/storageclasses/gp3-csi with fingerprint=11d86687a5005e677e9300d5c0505a757aa98809f0113ee0ca0fc50f300c80ce I0420 11:46:32.625781 1 gather.go:177] gatherer "clusterconfig" function "storage_classes" took 26.084499ms to process 2 records I0420 11:46:32.626286 1 tasks_processing.go:74] worker 50 stopped. I0420 11:46:32.626298 1 gather.go:177] gatherer "clusterconfig" function "sap_config" took 26.206309ms to process 0 records W0420 11:46:32.629060 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0420 11:46:32.629160 1 tasks_processing.go:74] worker 15 stopped. I0420 11:46:32.629176 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanedeployments" took 29.039667ms to process 0 records I0420 11:46:32.629183 1 gather.go:177] gatherer "clusterconfig" function "storage_cluster" took 30.57342ms to process 0 records I0420 11:46:32.629189 1 tasks_processing.go:74] worker 39 stopped. I0420 11:46:32.629241 1 tasks_processing.go:74] worker 51 stopped. I0420 11:46:32.629371 1 recorder.go:75] Recording config/validatingwebhookconfigurations/multus.openshift.io with fingerprint=06dc3bebabe07cc1fbf4424a99ab4f7391ce3e492cd45c71ecb9e48b016a317b I0420 11:46:32.629462 1 recorder.go:75] Recording config/validatingwebhookconfigurations/network-node-identity.openshift.io with fingerprint=ca042dd0050c3ae53b00f3c72e43a353ab83a7d55ef095c8ac4842488e330215 I0420 11:46:32.629483 1 recorder.go:75] Recording config/validatingwebhookconfigurations/performance-addon-operator with fingerprint=d062d4608f14752f186c33b9341f81ec0454a719ecf66bf7de7aeb5d17b661e6 I0420 11:46:32.629504 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterrolebindings-validation with fingerprint=188eb6ca5434036828be3e04af175df1264afed80c002b3b083bba7104222e0f I0420 11:46:32.629527 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterroles-validation with fingerprint=171c63fc9a7fa7daf8f4785a3f0fcf5759ba4a5e0c83ac6d0f63b5f2ec40e5ed I0420 11:46:32.629550 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-ingress-config-validation with fingerprint=47803a05a9fecf12a42dc325793d59d402ea833051cf0d63bfab780533268300 I0420 11:46:32.629572 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-network-operator-validation with fingerprint=a86788af8e03ef6b77acdc0b2e84f9a59da475935e307ded1f14ad5e7f32d5a3 I0420 11:46:32.629602 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-regular-user-validation with fingerprint=09d479e7fe373dfbcf874c83c4eb471e0ab6b8f7b9600416ec1be9215649afec I0420 11:46:32.629625 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-scc-validation with fingerprint=baa104897f1d9c4562928a5e91dd6af590141cdcaf6925c583109f1ef58e4677 I0420 11:46:32.629647 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-serviceaccount-validation with fingerprint=5653fa889f0e28891a30ed4d17a14d918e96413820b617d932311dfa8d83a8cc I0420 11:46:32.629672 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-techpreviewnoupgrade-validation with fingerprint=4388f145549fe55bd538bc1426e2c3926b866624392447f601b893b04a55de7e I0420 11:46:32.629679 1 gather.go:177] gatherer "clusterconfig" function "validating_webhook_configurations" took 29.862679ms to process 11 records I0420 11:46:32.629690 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkconfigurationpolicies" took 29.093639ms to process 0 records I0420 11:46:32.629706 1 tasks_processing.go:74] worker 17 stopped. I0420 11:46:32.629791 1 tasks_processing.go:74] worker 10 stopped. I0420 11:46:32.629792 1 recorder.go:75] Recording config/network with fingerprint=911660a67fc83fb2d43fe6f15a34d39fa55746852ae16e8467dcc90d3a162027 I0420 11:46:32.629813 1 gather.go:177] gatherer "clusterconfig" function "networks" took 33.248702ms to process 1 records I0420 11:46:32.629902 1 recorder.go:75] Recording config/schedulers/cluster with fingerprint=bab002347790381c824020bea7fce66550b6829f2e58e484109567581e54ee05 I0420 11:46:32.629909 1 tasks_processing.go:74] worker 40 stopped. I0420 11:46:32.629914 1 gather.go:177] gatherer "clusterconfig" function "schedulers" took 29.781002ms to process 1 records I0420 11:46:32.629921 1 gather.go:177] gatherer "clusterconfig" function "openstack_controlplanes" took 30.669726ms to process 0 records I0420 11:46:32.629925 1 gather.go:177] gatherer "clusterconfig" function "sap_pods" took 29.502125ms to process 0 records I0420 11:46:32.629946 1 tasks_processing.go:74] worker 60 stopped. I0420 11:46:32.629957 1 tasks_processing.go:74] worker 16 stopped. I0420 11:46:32.630024 1 tasks_processing.go:74] worker 20 stopped. I0420 11:46:32.630267 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/config/cluster with fingerprint=0adab6b8108b4ffd57c361f89a11329c843ce745fd2b4ed527e33706fafbe3a5 I0420 11:46:32.630278 1 gather.go:177] gatherer "clusterconfig" function "image_registries" took 30.718417ms to process 1 records I0420 11:46:32.632032 1 tasks_processing.go:74] worker 63 stopped. I0420 11:46:32.632120 1 gather.go:177] gatherer "clusterconfig" function "certificate_signing_requests" took 35.804593ms to process 0 records I0420 11:46:32.634500 1 tasks_processing.go:74] worker 32 stopped. I0420 11:46:32.634516 1 gather.go:177] gatherer "clusterconfig" function "monitoring_persistent_volumes" took 36.208737ms to process 0 records I0420 11:46:32.634845 1 tasks_processing.go:74] worker 22 stopped. I0420 11:46:32.634993 1 recorder.go:75] Recording config/ingress with fingerprint=81c1d481e7780b9472cd63f29fe54cc305da278e359a8fb2b97aa036e3f068be I0420 11:46:32.635012 1 gather.go:177] gatherer "clusterconfig" function "ingress" took 35.066136ms to process 1 records I0420 11:46:32.635138 1 tasks_processing.go:74] worker 47 stopped. I0420 11:46:32.635178 1 recorder.go:75] Recording config/namespaces_with_overlapping_uids with fingerprint=4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945 I0420 11:46:32.635187 1 gather.go:177] gatherer "clusterconfig" function "overlapping_namespace_uids" took 35.987215ms to process 1 records I0420 11:46:32.635695 1 sca.go:136] Pulling SCA certificates from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates. Next check is in 8h0m0s I0420 11:46:32.635703 1 cluster_transfer.go:83] checking the availability of cluster transfer. Next check is in 12h0m0s W0420 11:46:32.635814 1 operator.go:288] started I0420 11:46:32.635854 1 base_controller.go:76] Waiting for caches to sync for LoggingSyncer I0420 11:46:32.637855 1 tasks_processing.go:74] worker 35 stopped. I0420 11:46:32.637869 1 gather.go:177] gatherer "clusterconfig" function "openshift_machine_api_events" took 39.921467ms to process 0 records I0420 11:46:32.638128 1 tasks_processing.go:74] worker 29 stopped. I0420 11:46:32.638321 1 recorder.go:75] Recording config/authentication with fingerprint=d23e5452b9e793cfa79f6fd6105aa3f9167074d085cff3761a2019a2908ad030 I0420 11:46:32.638332 1 gather.go:177] gatherer "clusterconfig" function "authentication" took 37.807945ms to process 1 records I0420 11:46:32.638413 1 recorder.go:75] Recording config/featuregate with fingerprint=769bb2ab202e665043bab7f8a8f477d8765b69eaac0a2d9fd78af112c8618fe7 I0420 11:46:32.638421 1 gather.go:177] gatherer "clusterconfig" function "feature_gates" took 37.870954ms to process 1 records I0420 11:46:32.638421 1 tasks_processing.go:74] worker 31 stopped. I0420 11:46:32.640702 1 tasks_processing.go:74] worker 4 stopped. I0420 11:46:32.640720 1 gather.go:177] gatherer "clusterconfig" function "aggregated_monitoring_cr_names" took 44.559525ms to process 0 records I0420 11:46:32.649380 1 tasks_processing.go:74] worker 52 stopped. I0420 11:46:32.649457 1 recorder.go:75] Recording config/olm_operators with fingerprint=cf07d5e91c84a66f4bbdaa5197a3523cbad0fbefa7ab195bbfc7beedc7f328d2 I0420 11:46:32.649473 1 gather.go:177] gatherer "clusterconfig" function "olm_operators" took 51.587144ms to process 1 records I0420 11:46:32.650299 1 tasks_processing.go:74] worker 42 stopped. I0420 11:46:32.650366 1 recorder.go:75] Recording config/running_containers with fingerprint=b9ccd910d9e06a93d1837c081c67f743232e5b5847412afd38af182e3bb4a9f4 I0420 11:46:32.650377 1 gather.go:177] gatherer "clusterconfig" function "container_images" took 50.350807ms to process 1 records I0420 11:46:32.653140 1 tasks_processing.go:74] worker 1 stopped. I0420 11:46:32.653179 1 prometheus_rules.go:88] Prometheus rules successfully created I0420 11:46:32.653318 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/admin with fingerprint=8fa37cad31017e243953495df42d5b706af2c3ee803aa67e878f37083de0411b I0420 11:46:32.653409 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/edit with fingerprint=d6c713f39da8c50bf40f858504a9d2ee6076b683e550aa485c268fbba4de1c6a I0420 11:46:32.653421 1 gather.go:177] gatherer "clusterconfig" function "clusterroles" took 53.337825ms to process 2 records I0420 11:46:32.654804 1 tasks_processing.go:74] worker 26 stopped. I0420 11:46:32.654817 1 gather.go:177] gatherer "clusterconfig" function "machine_config_pools" took 56.304264ms to process 0 records I0420 11:46:32.655154 1 tasks_processing.go:74] worker 25 stopped. I0420 11:46:32.655297 1 controller.go:212] Source clusterTransferController *clustertransfer.Controller is not ready I0420 11:46:32.655312 1 controller.go:212] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0420 11:46:32.655316 1 controller.go:212] Source periodic-conditional *controllerstatus.Simple is not ready I0420 11:46:32.655319 1 controller.go:212] Source periodic-workloads *controllerstatus.Simple is not ready I0420 11:46:32.655323 1 controller.go:212] Source scaController *sca.Controller is not ready I0420 11:46:32.655340 1 controller.go:489] The operator is still being initialized I0420 11:46:32.655348 1 controller.go:512] The operator is healthy I0420 11:46:32.655678 1 recorder.go:75] Recording config/crd/volumesnapshots.snapshot.storage.k8s.io with fingerprint=b0b07f8f4f95ff877f8458d88290021c2f7c6edf7cb6b131af564c6ca908a780 I0420 11:46:32.655916 1 recorder.go:75] Recording config/crd/volumesnapshotcontents.snapshot.storage.k8s.io with fingerprint=d5eacf35292949c7812f5e8d8d31a79824178a0865911175561f0e90bbdca327 I0420 11:46:32.655928 1 gather.go:177] gatherer "clusterconfig" function "crds" took 55.246606ms to process 2 records I0420 11:46:32.655943 1 gather.go:177] gatherer "clusterconfig" function "number_of_pods_and_netnamespaces_with_sdn_annotations" took 57.081489ms to process 0 records I0420 11:46:32.655956 1 tasks_processing.go:74] worker 44 stopped. E0420 11:46:32.659738 1 cluster_transfer.go:95] failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%279c48e7b2-21cb-491e-b050-64e49cf67a77%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:49398->172.30.0.10:53: read: connection refused I0420 11:46:32.659751 1 controllerstatus.go:80] name=clusterTransferController healthy=true reason=Disconnected message=failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%279c48e7b2-21cb-491e-b050-64e49cf67a77%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:49398->172.30.0.10:53: read: connection refused I0420 11:46:32.664906 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0420 11:46:32.664917 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0420 11:46:32.664929 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController I0420 11:46:32.665877 1 tasks_processing.go:74] worker 23 stopped. E0420 11:46:32.665890 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "cluster-monitoring-config" not found E0420 11:46:32.665895 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "gateway-mode-config" not found E0420 11:46:32.665899 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "insights-config" not found I0420 11:46:32.665908 1 recorder.go:75] Recording config/configmaps/openshift-config/installer-images/images.json with fingerprint=26b6661162b099a0f5a279859b4f46c867929a79d9a4a41fde4be4e6fe138018 I0420 11:46:32.665932 1 recorder.go:75] Recording config/configmaps/openshift-config/kube-root-ca.crt/ca.crt with fingerprint=d476c7d3f5b104863f08f481b1264dcc68cc272ecefb0ecb709b18a6afab034d I0420 11:46:32.665940 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/invoker with fingerprint=76b482f683cd3ef9da02debac5b26080a5aeb06ff768ee5c21117514dff29d8a I0420 11:46:32.665945 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/version with fingerprint=c93090eb0d2a4736885abeb79c91680cfd01fda46464f83456b085d4dc8239f0 I0420 11:46:32.665949 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 I0420 11:46:32.666005 1 recorder.go:75] Recording config/configmaps/openshift-config/rosa-brand-logo/rosa-brand-logo.svg with fingerprint=6ed8ca4dd7a8eee7249182bc006e9649ce84d76c551ddfaaa33e55d8c4cc1ed0 I0420 11:46:32.666013 1 recorder.go:75] Recording config/configmaps/kube-system/cluster-config-v1/install-config with fingerprint=ab3811c6b83fd7b8e920094cfa3080d1b4ee3c35ec4c8379437b21d27bd6608d I0420 11:46:32.666019 1 gather.go:177] gatherer "clusterconfig" function "config_maps" took 67.15127ms to process 7 records I0420 11:46:32.666991 1 tasks_processing.go:74] worker 2 stopped. E0420 11:46:32.667004 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret 'router-certs-default' in namespace 'openshift-ingress': secrets "router-certs-default" not found E0420 11:46:32.667009 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret '2pprtdr8jhjhlh62vg8sqlta0p1f7rfm-primary-cert-bundle-secret' in namespace 'openshift-ingress-operator': secrets "2pprtdr8jhjhlh62vg8sqlta0p1f7rfm-primary-cert-bundle-secret" not found I0420 11:46:32.667049 1 recorder.go:75] Recording aggregated/ingress_controllers_certs with fingerprint=62d852cbfb8601adf72e0c408f4e71386091c2c3e1680c69e420458731e347a8 I0420 11:46:32.667061 1 gather.go:177] gatherer "clusterconfig" function "ingress_certificates" took 68.568186ms to process 1 records I0420 11:46:32.668979 1 base_controller.go:82] Caches are synced for ConfigController I0420 11:46:32.668989 1 base_controller.go:119] Starting #1 worker of ConfigController controller ... I0420 11:46:32.675335 1 configmapobserver.go:84] configmaps "insights-config" not found I0420 11:46:32.721037 1 tasks_processing.go:74] worker 36 stopped. I0420 11:46:32.721315 1 recorder.go:75] Recording config/node/ip-10-0-0-216.ec2.internal with fingerprint=28d41f704ccb2de6c8322cc7ca9473772bf8ed67e2588162b5613b4ebddf3130 I0420 11:46:32.721369 1 recorder.go:75] Recording config/node/ip-10-0-1-188.ec2.internal with fingerprint=182103ea31ace3688b049d770af64e603b03c5c68517c5c36ce596800701b7d4 I0420 11:46:32.721417 1 recorder.go:75] Recording config/node/ip-10-0-2-213.ec2.internal with fingerprint=95ba9950538bf89f927af4152f07286ce8fc96e68b05715f8ec7564db597dfb3 I0420 11:46:32.721425 1 gather.go:177] gatherer "clusterconfig" function "nodes" took 124.815176ms to process 3 records I0420 11:46:32.728895 1 tasks_processing.go:74] worker 33 stopped. I0420 11:46:32.728940 1 gather_logs.go:145] no pods in namespace were found I0420 11:46:32.729952 1 recorder.go:75] Recording config/version with fingerprint=29a6c5ca9590a1a2fa2a484287ca42958e612ced5aa8fef91e3bbc5b8daea69d I0420 11:46:32.730024 1 recorder.go:75] Recording config/id with fingerprint=09d40997cae3ba755485c68f1077924002ccd275cbd3bfef526168edf2968e4c I0420 11:46:32.730070 1 gather.go:177] gatherer "clusterconfig" function "version" took 128.393736ms to process 2 records I0420 11:46:32.730132 1 gather.go:177] gatherer "clusterconfig" function "qemu_kubevirt_launcher_logs" took 128.553898ms to process 0 records I0420 11:46:32.730199 1 tasks_processing.go:74] worker 19 stopped. I0420 11:46:32.731635 1 requests.go:205] Asking for SCA certificate with "{"arch": ["x86_64"]}" payload W0420 11:46:32.735515 1 sca.go:161] Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:40426->172.30.0.10:53: read: connection refused I0420 11:46:32.735529 1 controllerstatus.go:80] name=scaController healthy=true reason=NonHTTPError message=Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:40426->172.30.0.10:53: read: connection refused I0420 11:46:32.736576 1 base_controller.go:82] Caches are synced for LoggingSyncer I0420 11:46:32.736587 1 base_controller.go:119] Starting #1 worker of LoggingSyncer controller ... I0420 11:46:33.038078 1 gather_cluster_operator_pods_and_events.go:121] Found 18 pods with 21 containers I0420 11:46:33.038108 1 gather_cluster_operator_pods_and_events.go:235] Maximum buffer size: 1198372 bytes I0420 11:46:33.038162 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-6269d pod in namespace openshift-dns (previous: false). I0420 11:46:33.316197 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-6269d pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-6269d\" is waiting to start: ContainerCreating" I0420 11:46:33.316214 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-6269d\" is waiting to start: ContainerCreating" I0420 11:46:33.316222 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-6269d pod in namespace openshift-dns (previous: false). I0420 11:46:33.445258 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-6269d pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-6269d\" is waiting to start: ContainerCreating" I0420 11:46:33.445275 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-6269d\" is waiting to start: ContainerCreating" I0420 11:46:33.445304 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-fqvbb pod in namespace openshift-dns (previous: false). W0420 11:46:33.628885 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0420 11:46:33.667645 1 tasks_processing.go:74] worker 59 stopped. I0420 11:46:33.667676 1 recorder.go:75] Recording aggregated/unused_machine_configs_count with fingerprint=4bfc9fa984e5dfcd45848faaf05269de7619bf42edf9f781751af5ee05c1a499 I0420 11:46:33.667687 1 gather.go:177] gatherer "clusterconfig" function "machine_configs" took 1.071344468s to process 1 records I0420 11:46:33.672610 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-fqvbb pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-fqvbb\" is waiting to start: ContainerCreating" I0420 11:46:33.672623 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-fqvbb\" is waiting to start: ContainerCreating" I0420 11:46:33.672630 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-fqvbb pod in namespace openshift-dns (previous: false). I0420 11:46:33.845249 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-fqvbb pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-fqvbb\" is waiting to start: ContainerCreating" I0420 11:46:33.845267 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-fqvbb\" is waiting to start: ContainerCreating" I0420 11:46:33.845295 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-vdg5g pod in namespace openshift-dns (previous: false). I0420 11:46:34.061835 1 gather_cluster_operators.go:184] Unable to get configs.samples.operator.openshift.io resource due to: configs.samples.operator.openshift.io "cluster" not found I0420 11:46:34.070031 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-vdg5g pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-vdg5g\" is waiting to start: ContainerCreating" I0420 11:46:34.070045 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-vdg5g\" is waiting to start: ContainerCreating" I0420 11:46:34.070053 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-vdg5g pod in namespace openshift-dns (previous: false). I0420 11:46:34.244430 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-vdg5g pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-vdg5g\" is waiting to start: ContainerCreating" I0420 11:46:34.244447 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-vdg5g\" is waiting to start: ContainerCreating" I0420 11:46:34.244457 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-8vvmk pod in namespace openshift-dns (previous: false). I0420 11:46:34.445246 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0420 11:46:34.445263 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-k65kq pod in namespace openshift-dns (previous: false). I0420 11:46:34.465735 1 tasks_processing.go:74] worker 21 stopped. I0420 11:46:34.465792 1 recorder.go:75] Recording config/clusteroperator/console with fingerprint=6a687717cd06998686b52cf496e5a33ad1694f80ce050231d586511093916099 I0420 11:46:34.465836 1 recorder.go:75] Recording config/clusteroperator/csi-snapshot-controller with fingerprint=617f1f6fc9fe82dfd7e845b86b1c9fc87552f795132ae54ea0227052b61d4fd7 I0420 11:46:34.465865 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/csisnapshotcontroller/cluster with fingerprint=5adc514f4b63e2f1ecc68bf6f9c0af70c5eea04522a49524e102721b1c41f80e I0420 11:46:34.465899 1 recorder.go:75] Recording config/clusteroperator/dns with fingerprint=50afa30a2a4abfa768296ebea64dc3bd71373532b6e117604e0f4b81224db89a I0420 11:46:34.465918 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/dns/default with fingerprint=9e7b4ce029030d3d8c3b49af92c556acdcc415000b40d3f969dbdc42c432b47f I0420 11:46:34.465990 1 recorder.go:75] Recording config/clusteroperator/image-registry with fingerprint=21a2c736fb8ebe7a22e36c67c934d27989bd182f2c694e63701b01866674e4d0 I0420 11:46:34.466022 1 recorder.go:75] Recording config/clusteroperator/ingress with fingerprint=9549febe795cce46c3e947d92e3df9ef70020b066c218d7e51334c3a54e07308 I0420 11:46:34.466052 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/ingresscontroller/openshift-ingress-operator/default with fingerprint=93a16b3aeb6171d8c8f7660ff449d6827aff13836e7db88ff4db43b8c7bd0a46 I0420 11:46:34.466069 1 recorder.go:75] Recording config/clusteroperator/insights with fingerprint=7f1ce7c13cb08c0f75ff76c7a3fca8230b478f420263955e8e8e7c67ead9bc10 I0420 11:46:34.466105 1 recorder.go:75] Recording config/clusteroperator/kube-apiserver with fingerprint=8d40a34f2303307b52295625f03ffbbbda8d76160c4b83c40423d1b52f6bc67e I0420 11:46:34.466118 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubeapiserver/cluster with fingerprint=51503bf0b784fcf65ea46bcaf1f72ac1a5c4d5dc211934f18f27871efed05762 I0420 11:46:34.466135 1 recorder.go:75] Recording config/clusteroperator/kube-controller-manager with fingerprint=44cde43430b51ba646a9653afa46edf5cf7362cb2880b86b5dee5f912731726d I0420 11:46:34.466146 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubecontrollermanager/cluster with fingerprint=ce90c0d4f367d7da085074268031798382ae7c54fdcb0a21f15a4818fe308c11 I0420 11:46:34.466162 1 recorder.go:75] Recording config/clusteroperator/kube-scheduler with fingerprint=545ddc3a2877a9e8892fc61a0d754f656e95de345691b4fd1852ba9003352913 I0420 11:46:34.466173 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubescheduler/cluster with fingerprint=f2940fb9fd20c19951dfc295eb363b7fba0c505f5ae61f01967a063099e6b60a I0420 11:46:34.466191 1 recorder.go:75] Recording config/clusteroperator/kube-storage-version-migrator with fingerprint=5406e795cb174556c609a40730f788aef280a5df537e11a7a5a39aff06355a46 I0420 11:46:34.466203 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubestorageversionmigrator/cluster with fingerprint=9351181aa7e6ada41ef581ab31e13516c6b934cc95710154bafb2eb222cb58db I0420 11:46:34.466219 1 recorder.go:75] Recording config/clusteroperator/monitoring with fingerprint=f92696e6643052357c040f6bbc9bfc734196ec980096bbf910d864f15da26364 I0420 11:46:34.466352 1 recorder.go:75] Recording config/clusteroperator/network with fingerprint=7354436c85e262d79c4ee017ad2174e8a2db046a7757218771db395281c35795 I0420 11:46:34.466363 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/ovn with fingerprint=626a89d20e0deaed5b6dfb533acfe65f4bb1618bd200a703b62e60c5d16d94ab I0420 11:46:34.466370 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/signer with fingerprint=90410b16914712b85b3c4578716ad8c0ae072e688f4cd1e022bf76f20da3506d I0420 11:46:34.466389 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/network/cluster with fingerprint=a93d15eaecb455a0e40ecb2826eeecc1533899204ddd3c3921d15ab70af7ae75 I0420 11:46:34.466417 1 recorder.go:75] Recording config/clusteroperator/node-tuning with fingerprint=04b847339b34b62d88860f084031857123dcfa5cc705b0df86fd20f083386350 I0420 11:46:34.466443 1 recorder.go:75] Recording config/clusteroperator/openshift-apiserver with fingerprint=63f6546a5a48860833a01167fc48461ff6b6ba842212d76a7d2cb6ce58c8932b I0420 11:46:34.466457 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftapiserver/cluster with fingerprint=e712e6cf27339b441e4ed1f4cde91dbde7e952698ba93407e4457db63a4a4c76 I0420 11:46:34.466474 1 recorder.go:75] Recording config/clusteroperator/openshift-controller-manager with fingerprint=e0cc34d81a6f343bbbb31c07a88443fd90411430fab3fe757e5ccd5505cc361a I0420 11:46:34.466484 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftcontrollermanager/cluster with fingerprint=d71a0f4672f9b45d9fc8293bf1687afc650fd28d32e2e30de27523fe7b4eadf7 I0420 11:46:34.466503 1 recorder.go:75] Recording config/clusteroperator/openshift-samples with fingerprint=97b0ccf6a087072d199d16f0096d1ad85915d88193dcfb142ae05067bb214880 I0420 11:46:34.466518 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager with fingerprint=366a86bb6235b6cd7b6909c3ec78b30ec0463f00d59ef5ba76f34f89b93fe045 I0420 11:46:34.466535 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-catalog with fingerprint=f5fbe0d43dc92d1d22773995b91daa62442575401f88ede447659ac30cf97829 I0420 11:46:34.466550 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-packageserver with fingerprint=a757b7a85bb90edba4e2ef75d5bc8cef68535b6d93a3ec3d95f91ba57859d4bd I0420 11:46:34.466573 1 recorder.go:75] Recording config/clusteroperator/service-ca with fingerprint=dbd9492af446ac5d2555412407d258d39047cc33f05305d8d3ff15e36a7e2604 I0420 11:46:34.466604 1 recorder.go:75] Recording config/clusteroperator/storage with fingerprint=ece39855d01f59415911e1d375c1b25ee792a0e0cb24e391c10ba1fd997e6192 I0420 11:46:34.466628 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/clustercsidriver/ebs.csi.aws.com with fingerprint=510064d6f6bcced87ab5bd2ddaff3d0edd7f93f4a4f7af2641f29fc53ffab21e I0420 11:46:34.466637 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/storage/cluster with fingerprint=8e480f8c1ce1b39baac42d8ec780c57c2592929ae0c801b61ffad49ba13f33ad I0420 11:46:34.466645 1 gather.go:177] gatherer "clusterconfig" function "operators" took 1.866486592s to process 35 records W0420 11:46:34.628932 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0420 11:46:34.646212 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0420 11:46:34.646225 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-wtrxk pod in namespace openshift-dns (previous: false). I0420 11:46:34.844796 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0420 11:46:34.844845 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-66dc5dfb66-d2lz7 pod in namespace openshift-image-registry (previous: false). I0420 11:46:35.044310 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-66dc5dfb66-d2lz7 pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-66dc5dfb66-d2lz7\" is waiting to start: ContainerCreating" I0420 11:46:35.044327 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-66dc5dfb66-d2lz7\" is waiting to start: ContainerCreating" I0420 11:46:35.044368 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-6f769675c7-76kgg pod in namespace openshift-image-registry (previous: false). I0420 11:46:35.243790 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-6f769675c7-76kgg pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-6f769675c7-76kgg\" is waiting to start: ContainerCreating" I0420 11:46:35.243806 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-6f769675c7-76kgg\" is waiting to start: ContainerCreating" I0420 11:46:35.243839 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-6f769675c7-dgt89 pod in namespace openshift-image-registry (previous: false). I0420 11:46:35.444279 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-6f769675c7-dgt89 pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-6f769675c7-dgt89\" is waiting to start: ContainerCreating" I0420 11:46:35.444294 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-6f769675c7-dgt89\" is waiting to start: ContainerCreating" I0420 11:46:35.444305 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-hjtlq pod in namespace openshift-image-registry (previous: false). W0420 11:46:35.629209 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0420 11:46:35.644929 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0420 11:46:35.644945 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-m9qln pod in namespace openshift-image-registry (previous: false). I0420 11:46:35.844604 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0420 11:46:35.844618 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-zf6xs pod in namespace openshift-image-registry (previous: false). I0420 11:46:36.044703 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0420 11:46:36.044718 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-64c65f7577-hzz7d pod in namespace openshift-ingress (previous: false). I0420 11:46:36.243699 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-64c65f7577-hzz7d pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-64c65f7577-hzz7d\" is waiting to start: ContainerCreating" I0420 11:46:36.243716 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-64c65f7577-hzz7d\" is waiting to start: ContainerCreating" I0420 11:46:36.243726 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-7b9b45cb9b-8l5br pod in namespace openshift-ingress (previous: false). I0420 11:46:36.445301 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-7b9b45cb9b-8l5br pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-7b9b45cb9b-8l5br\" is waiting to start: ContainerCreating" I0420 11:46:36.445318 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-7b9b45cb9b-8l5br\" is waiting to start: ContainerCreating" I0420 11:46:36.445327 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-7b9b45cb9b-ghbfb pod in namespace openshift-ingress (previous: false). W0420 11:46:36.628630 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0420 11:46:36.645420 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-7b9b45cb9b-ghbfb pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-7b9b45cb9b-ghbfb\" is waiting to start: ContainerCreating" I0420 11:46:36.645438 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-7b9b45cb9b-ghbfb\" is waiting to start: ContainerCreating" I0420 11:46:36.645467 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-c8qj7 pod in namespace openshift-ingress-canary (previous: false). I0420 11:46:36.845449 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-c8qj7 pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-c8qj7\" is waiting to start: ContainerCreating" I0420 11:46:36.845462 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-c8qj7\" is waiting to start: ContainerCreating" I0420 11:46:36.845488 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-jd8f9 pod in namespace openshift-ingress-canary (previous: false). I0420 11:46:37.045026 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-jd8f9 pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-jd8f9\" is waiting to start: ContainerCreating" I0420 11:46:37.045044 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-jd8f9\" is waiting to start: ContainerCreating" I0420 11:46:37.045071 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-n6gq7 pod in namespace openshift-ingress-canary (previous: false). I0420 11:46:37.243317 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-n6gq7 pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-n6gq7\" is waiting to start: ContainerCreating" I0420 11:46:37.243332 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-n6gq7\" is waiting to start: ContainerCreating" I0420 11:46:37.243348 1 tasks_processing.go:74] worker 38 stopped. I0420 11:46:37.243433 1 recorder.go:75] Recording events/openshift-dns-operator with fingerprint=9279c5372b87ad65d39746f5ec53ac5ef5fc27d08fe2bb7027157ce8ee6ce997 I0420 11:46:37.243483 1 recorder.go:75] Recording events/openshift-dns with fingerprint=db487117b171d263b06b7c3942598c9c580282580d7b820d0af19bb06c860f5c I0420 11:46:37.243614 1 recorder.go:75] Recording events/openshift-image-registry with fingerprint=cfc8377bffc5d075e129b04f1d5bf94e61317ce2b20bdc65a1bd4e4c83a4e648 I0420 11:46:37.243642 1 recorder.go:75] Recording events/openshift-ingress-operator with fingerprint=a57a1f1679fd6accd29c7d41b393d6ead737bb3f302cbe2ae958c5f0c39a2ee0 I0420 11:46:37.243687 1 recorder.go:75] Recording events/openshift-ingress with fingerprint=c6441115b3cbe5491730ab8d741626b548651f3aa9e6b80b8749b1be7601b445 I0420 11:46:37.243704 1 recorder.go:75] Recording events/openshift-ingress-canary with fingerprint=9113a188c452fd839792746e6921b76cca25d9ee0ed5d9d53649ab008c528576 I0420 11:46:37.245012 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-6269d with fingerprint=1d386cb4b92c389b91add561e9b612b9d137ae7b019e4cefd0146540951cb92f I0420 11:46:37.245116 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-fqvbb with fingerprint=0cd8a0c9ac16893abdbf17f51d36754b87f17b4489a7ea4cbce316f17cd46892 I0420 11:46:37.245192 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-vdg5g with fingerprint=bc3b42a060c510491a7674415a6033bb876abded7c24073631e86e58e7ead7d4 I0420 11:46:37.245292 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-66dc5dfb66-d2lz7 with fingerprint=ce4c7cec61b7bcfd635fd0210235b29039974a9c74d9818a08393eafc49107b8 I0420 11:46:37.245382 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-6f769675c7-76kgg with fingerprint=57d92e88fe3cde462776ed0505c7092082cf5cda60d596b38dd617a1598730b7 I0420 11:46:37.245469 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-6f769675c7-dgt89 with fingerprint=b29da3e891d6706ad59db4e3fae90b738a60490644300b7e05d484c759cad3b5 I0420 11:46:37.245524 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-c8qj7 with fingerprint=3a8636d6446d186ed28d714fb4f90ccea5ea0f19a8e7105028010d70191997b2 I0420 11:46:37.245575 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-jd8f9 with fingerprint=2cf86602e331df67d2bf881528b81174f28899b49839f464bd899b6a8d4e76b3 I0420 11:46:37.245631 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-n6gq7 with fingerprint=89f39a3353b51c1a1b12d43ab57d52e70c06d41d97d6bf17411c1367d1e0b193 I0420 11:46:37.245640 1 gather.go:177] gatherer "clusterconfig" function "operators_pods_and_events" took 4.647132878s to process 15 records W0420 11:46:37.629483 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. W0420 11:46:37.629505 1 gather_dvo_metrics.go:117] Unable to read metrics from endpoint "http://deployment-validation-operator-metrics.openshift-deployment-validation-operator.svc:8383": DVO metrics service was not available within the 5s timeout: context deadline exceeded I0420 11:46:37.629516 1 tasks_processing.go:74] worker 62 stopped. E0420 11:46:37.629525 1 gather.go:140] gatherer "clusterconfig" function "dvo_metrics" failed with the error: DVO metrics service was not available within the 5s timeout: context deadline exceeded I0420 11:46:37.629535 1 recorder.go:75] Recording config/dvo_metrics with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 W0420 11:46:37.629548 1 gather.go:155] issue recording gatherer "clusterconfig" function "dvo_metrics" result "config/dvo_metrics" because of the warning: warning: the record with the same fingerprint "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" was already recorded at path "config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt", recording another one with a different path "config/dvo_metrics" I0420 11:46:37.629557 1 gather.go:177] gatherer "clusterconfig" function "dvo_metrics" took 5.031114785s to process 1 records I0420 11:46:45.042047 1 tasks_processing.go:74] worker 8 stopped. I0420 11:46:45.042085 1 recorder.go:75] Recording config/installplans with fingerprint=f17dbfacc3bfddf27ca3b213b39495434cd4c4e9e3dbd69566ffb3845bbcf539 I0420 11:46:45.042112 1 gather.go:177] gatherer "clusterconfig" function "install_plans" took 12.444720282s to process 1 records I0420 11:46:45.796231 1 configmapobserver.go:84] configmaps "insights-config" not found I0420 11:46:45.806630 1 tasks_processing.go:74] worker 45 stopped. I0420 11:46:45.806949 1 recorder.go:75] Recording config/serviceaccounts with fingerprint=751aa8b7d1208af5fd734152a53c8f2354708612b0cd50cfd9c595651a5cc8c8 I0420 11:46:45.806972 1 gather.go:177] gatherer "clusterconfig" function "service_accounts" took 13.20737034s to process 1 records E0420 11:46:45.807042 1 periodic.go:247] "Unhandled Error" err="clusterconfig failed after 13.211s with: function \"support_secret\" failed with an error, function \"machines\" failed with an error, function \"pod_network_connectivity_checks\" failed with an error, function \"machine_healthchecks\" failed with an error, function \"config_maps\" failed with an error, function \"ingress_certificates\" failed with an error, function \"dvo_metrics\" failed with an error" I0420 11:46:45.808117 1 controllerstatus.go:89] name=periodic-clusterconfig healthy=false reason=PeriodicGatherFailed message=Source clusterconfig could not be retrieved: function "support_secret" failed with an error, function "machines" failed with an error, function "pod_network_connectivity_checks" failed with an error, function "machine_healthchecks" failed with an error, function "config_maps" failed with an error, function "ingress_certificates" failed with an error, function "dvo_metrics" failed with an error I0420 11:46:45.808130 1 periodic.go:209] Running workloads gatherer I0420 11:46:45.808144 1 tasks_processing.go:45] number of workers: 2 I0420 11:46:45.808150 1 tasks_processing.go:69] worker 1 listening for tasks. I0420 11:46:45.808154 1 tasks_processing.go:71] worker 1 working on helmchart_info task. I0420 11:46:45.808161 1 tasks_processing.go:69] worker 0 listening for tasks. I0420 11:46:45.808237 1 tasks_processing.go:71] worker 0 working on workload_info task. I0420 11:46:45.834699 1 gather_workloads_info.go:278] Loaded pods in 0s, will wait 22s for image data I0420 11:46:45.844272 1 gather_workloads_info.go:387] No image sha256:2904a78e2eb73fd6a9bb94c105c2a056831fb4113fbb7b0607c50adc9d879c9b (10ms) I0420 11:46:45.846214 1 tasks_processing.go:74] worker 1 stopped. I0420 11:46:45.846226 1 gather.go:177] gatherer "workloads" function "helmchart_info" took 38.04671ms to process 0 records I0420 11:46:45.853543 1 gather_workloads_info.go:387] No image sha256:04c87c054a3f366a7dfbe0a93ebb0c80a098ee16842c5794b67c1202eec61996 (9ms) I0420 11:46:45.863276 1 gather_workloads_info.go:387] No image sha256:765f0d23b637f685f98a31bd47c131b03cf72a40761a3f9a9d6320faa3c33733 (10ms) I0420 11:46:45.872353 1 gather_workloads_info.go:387] No image sha256:36b9e89c3cfcf1ab9ae500486e38afb6862cba48cb0b4d84a09508ab8f3d299f (9ms) I0420 11:46:45.881427 1 gather_workloads_info.go:387] No image sha256:c15ca0c0ad60fe8757c2d5d1723fcdd7a1ed6c0251a90d22a7e6cae6811d01aa (9ms) I0420 11:46:45.890398 1 gather_workloads_info.go:387] No image sha256:7adc1eab05d6724c76ba751f6df816b08d6e70b78dee9eb94fa6fd9690542c98 (9ms) I0420 11:46:45.899319 1 gather_workloads_info.go:387] No image sha256:0a99240166165eb5718e7516a43282fe32df9c7c5e809b31b58abe44e42ff94d (9ms) I0420 11:46:45.908119 1 gather_workloads_info.go:387] No image sha256:4556896f77307821531ef91b7b7faccb82b824ea695693b2989f597f0deca038 (9ms) I0420 11:46:45.918795 1 gather_workloads_info.go:387] No image sha256:03cf4cd7ef1518610c6c7b3ad27d1622d82e98e3dc6e3f8e5d0fceb5c8d3786e (11ms) I0420 11:46:45.927368 1 gather_workloads_info.go:387] No image sha256:943018739e3db1763c3184b460dbc409e058abbac76d57b9927faad317be85e4 (9ms) I0420 11:46:45.943283 1 gather_workloads_info.go:387] No image sha256:695cf2f0cc07683c2a3ce1eaf3e56fe18abc6e2bac716f7d9843f5d173b9df52 (16ms) I0420 11:46:46.044147 1 gather_workloads_info.go:387] No image sha256:653c666f842c13e0baae2e89a9b1efe0e2ef56f621ffb5b32005115d2a26ab8c (101ms) I0420 11:46:46.145062 1 gather_workloads_info.go:387] No image sha256:a498046d64605bcccee2440aa4f04a4602baaae263cf01d977ec5208e876b1fd (101ms) I0420 11:46:46.244366 1 gather_workloads_info.go:387] No image sha256:2598489729a4b258e4ecda4a06f6875133f2a10ced5c5241f8a57a8a05418e36 (99ms) I0420 11:46:46.344449 1 gather_workloads_info.go:387] No image sha256:a258c226562adb14e3a163a1940938526ee6a0928982a7667d85d9a7334ce639 (100ms) I0420 11:46:46.446459 1 gather_workloads_info.go:387] No image sha256:2e57e192c3c1240fd935dcd55c8fde5e70e78bf81d6176c96edf21fafe59f8ba (102ms) I0420 11:46:46.544505 1 gather_workloads_info.go:387] No image sha256:2e564f336c77116053f34d4201d364d8da04e789cfffa0ea422574c95f2d6404 (98ms) I0420 11:46:46.645406 1 gather_workloads_info.go:387] No image sha256:e84cb128d930bd1ab867cc89b7b7bf2b2c0e41105ab93b5381069945b3ee9c57 (101ms) I0420 11:46:46.744672 1 gather_workloads_info.go:387] No image sha256:5f0b67cfbbc381243fb91ccc17345b56d05f4d717c667e8c644e5bf05633ba71 (99ms) I0420 11:46:46.844580 1 gather_workloads_info.go:387] No image sha256:1a2532940843248c57d52141185dd71fbc393ab28b65d48f682038632c1dbbad (100ms) I0420 11:46:46.944442 1 gather_workloads_info.go:387] No image sha256:289816958633a763a72dbc44e1dad40466223164e7e253039514f0d974ea5d21 (100ms) I0420 11:46:47.044625 1 gather_workloads_info.go:387] No image sha256:a0105d1eb62cf6ac9e5e2ef28d3e89bf6dc514bc594fc7090fe5a5ee18a09c87 (100ms) I0420 11:46:47.144080 1 gather_workloads_info.go:387] No image sha256:5a95c19d82767e0235b4edb4a0536482c816904897aae1dc3eb255cb52b87a9f (99ms) I0420 11:46:47.244675 1 gather_workloads_info.go:387] No image sha256:91828234f107c068c8a4966d08370ae7b73e637651dbc6d92c18c4553402c22c (101ms) I0420 11:46:47.344720 1 gather_workloads_info.go:387] No image sha256:a56211d075aa43cbb491f669a5b2e46ee023dc95b7d51dbac28f463948c5ad61 (100ms) I0420 11:46:47.444781 1 gather_workloads_info.go:387] No image sha256:56a85660a445eced5c79a595a0eccf590087c5672d50f49d4c25ad52f9a44f04 (100ms) I0420 11:46:47.545109 1 gather_workloads_info.go:387] No image sha256:b3909bf664c77097f75b3768830863d642eed3815dab2bfb4415c771ca2d5007 (100ms) I0420 11:46:47.647128 1 gather_workloads_info.go:387] No image sha256:47154813651033d59751fb655a384dbffb64dd26f10bd7f3be0c3128d0486356 (102ms) I0420 11:46:47.744856 1 gather_workloads_info.go:387] No image sha256:c940ea87e7d133d75ba0002ef00c0806825eed3db8094cdb260d1bac18127733 (98ms) I0420 11:46:47.843952 1 gather_workloads_info.go:387] No image sha256:ca1344cb64140188b7cae7bbc51fb751566c0b0c97d5e39b5850e628032c4a5e (99ms) I0420 11:46:47.943751 1 gather_workloads_info.go:387] No image sha256:521712486e2c6e3c020dad6a1cb340db8e55665b69f7c208fab9cd9e965fd588 (100ms) I0420 11:46:47.985235 1 configmapobserver.go:84] configmaps "insights-config" not found I0420 11:46:48.044063 1 gather_workloads_info.go:387] No image sha256:7b31223098f08328f5ddea8e5b871dbbd5f5a61ec550e8956f66793c0c6031a9 (100ms) I0420 11:46:48.044102 1 tasks_processing.go:74] worker 0 stopped. E0420 11:46:48.044112 1 gather.go:140] gatherer "workloads" function "workload_info" failed with the error: no running pods found for the insights-runtime-extractor statefulset I0420 11:46:48.044403 1 recorder.go:75] Recording config/workload_info with fingerprint=80e4ae6114088963732fc1122c7afb0d839dcd2de5fe134489a3f1e9d5321004 I0420 11:46:48.044418 1 gather.go:177] gatherer "workloads" function "workload_info" took 2.235839481s to process 1 records E0420 11:46:48.044441 1 periodic.go:247] "Unhandled Error" err="workloads failed after 2.236s with: function \"workload_info\" failed with an error" I0420 11:46:48.045534 1 controllerstatus.go:89] name=periodic-workloads healthy=false reason=PeriodicGatherFailed message=Source workloads could not be retrieved: function "workload_info" failed with an error I0420 11:46:48.045544 1 periodic.go:209] Running conditional gatherer I0420 11:46:48.051841 1 requests.go:294] Making HTTP GET request at: https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules I0420 11:46:48.058047 1 conditional_gatherer.go:107] Get "https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.129.0.10:50206->172.30.0.10:53: read: connection refused E0420 11:46:48.058305 1 conditional_gatherer.go:322] unable to update alerts cache: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0420 11:46:48.058365 1 conditional_gatherer.go:384] updating version cache for conditional gatherer I0420 11:46:48.064642 1 conditional_gatherer.go:392] cluster version is '4.20.8' E0420 11:46:48.064655 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0420 11:46:48.064662 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0420 11:46:48.064668 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0420 11:46:48.064672 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0420 11:46:48.064675 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0420 11:46:48.064678 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0420 11:46:48.064681 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0420 11:46:48.064684 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0420 11:46:48.064688 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing I0420 11:46:48.064701 1 tasks_processing.go:45] number of workers: 3 I0420 11:46:48.064710 1 tasks_processing.go:69] worker 2 listening for tasks. I0420 11:46:48.064714 1 tasks_processing.go:71] worker 2 working on conditional_gatherer_rules task. I0420 11:46:48.064722 1 tasks_processing.go:69] worker 0 listening for tasks. I0420 11:46:48.064731 1 tasks_processing.go:69] worker 1 listening for tasks. I0420 11:46:48.064734 1 tasks_processing.go:71] worker 0 working on remote_configuration task. I0420 11:46:48.064738 1 tasks_processing.go:74] worker 1 stopped. I0420 11:46:48.064741 1 tasks_processing.go:71] worker 2 working on rapid_container_logs task. I0420 11:46:48.064791 1 recorder.go:75] Recording insights-operator/conditional-gatherer-rules with fingerprint=7034af97e7e41c22e4b775abdd4b9066c8ebb19da33eb7f69f39bfd2eb5f6406 I0420 11:46:48.064803 1 gather.go:177] gatherer "conditional" function "conditional_gatherer_rules" took 753ns to process 1 records I0420 11:46:48.064841 1 recorder.go:75] Recording insights-operator/remote-configuration with fingerprint=0394430c431eec4d48bb1811a90918e95161d2282c59af26f2473613cc0959db I0420 11:46:48.064851 1 gather.go:177] gatherer "conditional" function "remote_configuration" took 1.176µs to process 1 records I0420 11:46:48.064856 1 tasks_processing.go:74] worker 0 stopped. I0420 11:46:48.064983 1 tasks_processing.go:74] worker 2 stopped. I0420 11:46:48.064995 1 gather.go:177] gatherer "conditional" function "rapid_container_logs" took 234.712µs to process 0 records I0420 11:46:48.065013 1 controllerstatus.go:89] name=periodic-conditional healthy=false reason=NotAvailable message=Get "https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.129.0.10:50206->172.30.0.10:53: read: connection refused I0420 11:46:48.065031 1 recorder.go:75] Recording insights-operator/remote-configuration.json with fingerprint=359de9c990c741675cec72fda96b5c3682221efdb4799f5eaa6e9805bcd3b5c1 I0420 11:46:48.086947 1 recorder.go:75] Recording insights-operator/gathers with fingerprint=b678f7d73382031834b0be775ca2e2cdc1cf32157ef324c0e4d0a33b66f27293 I0420 11:46:48.087062 1 diskrecorder.go:70] Writing 109 records to /var/lib/insights-operator/insights-2026-04-20-114648.tar.gz I0420 11:46:48.093604 1 diskrecorder.go:51] Wrote 109 records to disk in 6ms I0420 11:46:48.093631 1 periodic.go:278] Gathering cluster info every 2h0m0s I0420 11:46:48.093645 1 periodic.go:279] Configuration is dataReporting: interval: 2h0m0s, uploadEndpoint: https://console.redhat.com/api/ingress/v1/upload, storagePath: /var/lib/insights-operator, downloadEndpoint: https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports, conditionalGathererEndpoint: https://console.redhat.com/api/gathering/v2/%s/gathering_rules, obfuscation: [] sca: disabled: false, endpoint: https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates, interval: 8h0m0s alerting: disabled: false clusterTransfer: endpoint: https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/, interval: 12h0m0s proxy: httpProxy: , httpsProxy: , noProxy: I0420 11:46:48.186620 1 configmapobserver.go:84] configmaps "insights-config" not found I0420 11:47:01.600306 1 configmapobserver.go:84] configmaps "insights-config" not found I0420 11:48:02.110588 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.crt" has been created (hash="e1e4bb63bf841ed1937fe2442a93f19db73a811192354e23beb11d61891b8fae") W0420 11:48:02.110623 1 builder.go:160] Restart triggered because of file /var/run/secrets/serving-cert/tls.crt was created I0420 11:48:02.110666 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.key" has been created (hash="4ecd6b965f17bb4de46ca198230e4edf79ea9b4e191da728756fa17bacf70950") I0420 11:48:02.110668 1 simple_featuregate_reader.go:177] Shutting down feature-gate-detector I0420 11:48:02.110704 1 genericapiserver.go:693] "[graceful-termination] pre-shutdown hooks completed" name="PreShutdownHooksStopped" I0420 11:48:02.110725 1 genericapiserver.go:548] "[graceful-termination] shutdown event" name="ShutdownInitiated" I0420 11:48:02.110752 1 object_count_tracker.go:151] "StorageObjectCountTracker pruner is exiting" I0420 11:48:02.110755 1 observer_polling.go:111] Observed file "/var/run/configmaps/service-ca-bundle/service-ca.crt" has been created (hash="5f1678d6cd5a38b5c3597a4b977739761b09d83915b9fecf4fd11476625a8c90") I0420 11:48:02.110769 1 configmap_cafile_content.go:226] "Shutting down controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0420 11:48:02.110774 1 genericapiserver.go:651] "[graceful-termination] not going to wait for active watch request(s) to drain" I0420 11:48:02.110783 1 base_controller.go:181] Shutting down ConfigController ... I0420 11:48:02.110783 1 dynamic_serving_content.go:149] "Shutting down controller" name="serving-cert::/tmp/serving-cert-3982279924/tls.crt::/tmp/serving-cert-3982279924/tls.key" I0420 11:48:02.110801 1 tlsconfig.go:258] "Shutting down DynamicServingCertificateController" I0420 11:48:02.110807 1 requestheader_controller.go:194] Shutting down RequestHeaderAuthRequestController I0420 11:48:02.110781 1 configmap_cafile_content.go:226] "Shutting down controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0420 11:48:02.110824 1 base_controller.go:123] Shutting down worker of ConfigController controller ... I0420 11:48:02.110812 1 base_controller.go:181] Shutting down LoggingSyncer ... I0420 11:48:02.110833 1 base_controller.go:113] All ConfigController workers have been terminated I0420 11:48:02.110803 1 periodic.go:170] Shutting down I0420 11:48:02.110842 1 base_controller.go:123] Shutting down worker of LoggingSyncer controller ... I0420 11:48:02.110850 1 base_controller.go:113] All LoggingSyncer workers have been terminated