W0424 09:06:05.522372 1 cmd.go:257] Using insecure, self-signed certificates I0424 09:06:05.965168 1 start.go:138] Unable to read service ca bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0424 09:06:05.965490 1 observer_polling.go:159] Starting file observer I0424 09:06:06.589778 1 operator.go:60] Starting insights-operator v0.0.0-master+$Format:%H$ I0424 09:06:06.589961 1 legacy_config.go:327] Current config: {"report":false,"storagePath":"/var/lib/insights-operator","interval":"2h","endpoint":"https://console.redhat.com/api/ingress/v1/upload","conditionalGathererEndpoint":"https://console.redhat.com/api/gathering/v2/%s/gathering_rules","pull_report":{"endpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports","delay":"60s","timeout":"3000s","min_retry":"30s"},"impersonate":"system:serviceaccount:openshift-insights:gather","enableGlobalObfuscation":false,"ocm":{"scaEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates","scaInterval":"8h","scaDisabled":false,"clusterTransferEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/","clusterTransferInterval":"12h"},"disableInsightsAlerts":false,"processingStatusEndpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status","reportEndpointTechPreview":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/report"} I0424 09:06:06.590575 1 secure_serving.go:57] Forcing use of http/1.1 only W0424 09:06:06.590598 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. W0424 09:06:06.590605 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. W0424 09:06:06.590611 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. W0424 09:06:06.590616 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. W0424 09:06:06.590619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. W0424 09:06:06.590621 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. I0424 09:06:06.590739 1 simple_featuregate_reader.go:171] Starting feature-gate-detector I0424 09:06:06.595294 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-insights", Name:"insights-operator", UID:"1c37d2b2-ff5a-404d-a1a3-a6e4c1136fd9", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BuildCSIVolumes", "CPMSMachineNamePrefix", "ConsolePluginContentSecurityPolicy", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "ImageVolume", "IngressControllerLBSubnetsAWS", "KMSv1", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "MetricsCollectionProfiles", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "PinnedImages", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SetEIPForNLBIngressController", "SigstoreImageVerification", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiDisk", "VSphereMultiNetworks"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSServiceLBNetworkSecurityGroup", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureMultiDisk", "BootImageSkewEnforcement", "BootcNodeManagement", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "Example2", "ExternalSnapshotMetadata", "GCPClusterHostedDNS", "GCPClusterHostedDNSInstall", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "ImageModeStatusReporting", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryptionProvider", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "ManagedBootImagesAzure", "ManagedBootImagesvSphere", "MaxUnavailableStatefulSet", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "MultiDiskSetup", "MutatingAdmissionPolicy", "NewOLM", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterOperations", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PreconfiguredUDNAddresses", "SELinuxMount", "ShortCertRotation", "SignatureStores", "SigstoreImageVerificationPKI", "TranslateStreamCloseWebsocketRequests", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VolumeAttributesClass", "VolumeGroupSnapshot"}} I0424 09:06:06.595287 1 operator.go:125] FeatureGates initialized: knownFeatureGates=[AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BuildCSIVolumes CPMSMachineNamePrefix ConsolePluginContentSecurityPolicy ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GatewayAPI GatewayAPIController HighlyAvailableArbiter ImageVolume IngressControllerLBSubnetsAWS KMSv1 MachineConfigNodes ManagedBootImages ManagedBootImagesAWS MetricsCollectionProfiles NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation PinnedImages ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SetEIPForNLBIngressController SigstoreImageVerification StoragePerformantSecurityPolicy UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiDisk VSphereMultiNetworks AWSClusterHostedDNS AWSClusterHostedDNSInstall AWSDedicatedHosts AWSServiceLBNetworkSecurityGroup AutomatedEtcdBackup AzureClusterHostedDNSInstall AzureDedicatedHosts AzureMultiDisk BootImageSkewEnforcement BootcNodeManagement ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterMonitoringConfig ClusterVersionOperatorConfiguration DNSNameResolver DualReplica DyanmicServiceEndpointIBMCloud DynamicResourceAllocation EtcdBackendQuota EventedPLEG Example Example2 ExternalSnapshotMetadata GCPClusterHostedDNS GCPClusterHostedDNSInstall GCPCustomAPIEndpoints GCPCustomAPIEndpointsInstall ImageModeStatusReporting ImageStreamImportMode IngressControllerDynamicConfigurationManager InsightsConfig InsightsConfigAPI InsightsOnDemandDataGather IrreconcilableMachineConfig KMSEncryptionProvider MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController ManagedBootImagesAzure ManagedBootImagesvSphere MaxUnavailableStatefulSet MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure MultiDiskSetup MutatingAdmissionPolicy NewOLM NewOLMCatalogdAPIV1Metas NewOLMOwnSingleNamespace NewOLMPreflightPermissionChecks NewOLMWebhookProviderOpenshiftServiceCA NoRegistryClusterOperations NodeSwap NutanixMultiSubnets OVNObservability OpenShiftPodSecurityAdmission PreconfiguredUDNAddresses SELinuxMount ShortCertRotation SignatureStores SigstoreImageVerificationPKI TranslateStreamCloseWebsocketRequests VSphereConfigurableMaxAllowedBlockVolumesPerNode VSphereHostVMGroupZonal VSphereMixedNodeEnv VolumeAttributesClass VolumeGroupSnapshot] I0424 09:06:06.597880 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController I0424 09:06:06.597887 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0424 09:06:06.597903 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0424 09:06:06.597903 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0424 09:06:06.597894 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController I0424 09:06:06.597917 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0424 09:06:06.598191 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/tmp/serving-cert-2284781888/tls.crt::/tmp/serving-cert-2284781888/tls.key" I0424 09:06:06.598541 1 secure_serving.go:213] Serving securely on [::]:8443 I0424 09:06:06.598569 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" W0424 09:06:06.604081 1 configmapobserver.go:64] Cannot get the configuration config map: configmaps "insights-config" not found. Default configuration is used. I0424 09:06:06.604108 1 secretconfigobserver.go:216] Legacy configuration set: enabled=false endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=false reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0424 09:06:06.604206 1 base_controller.go:76] Waiting for caches to sync for ConfigController I0424 09:06:06.608937 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0424 09:06:06.608954 1 secretconfigobserver.go:204] Legacy configuration updated: enabled=true endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=true reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0424 09:06:06.613736 1 secretconfigobserver.go:119] support secret does not exist I0424 09:06:06.618032 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0424 09:06:06.622345 1 secretconfigobserver.go:119] support secret does not exist I0424 09:06:06.624624 1 recorder.go:161] Pruning old reports every 5h5m13s, max age is 288h0m0s I0424 09:06:06.629665 1 periodic.go:209] Running clusterconfig gatherer I0424 09:06:06.629681 1 controllerstatus.go:80] name=insightsreport healthy=true reason= message= I0424 09:06:06.629697 1 insightsreport.go:296] Starting report retriever I0424 09:06:06.629704 1 insightsreport.go:298] Insights analysis reports will be downloaded from the https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports endpoint with a delay of 1m0s I0424 09:06:06.629666 1 controllerstatus.go:80] name=insightsuploader healthy=true reason= message= I0424 09:06:06.629717 1 tasks_processing.go:45] number of workers: 64 I0424 09:06:06.629728 1 insightsuploader.go:86] Reporting status periodically to https://console.redhat.com/api/ingress/v1/upload every 2h0m0s, starting in 1m30s I0424 09:06:06.629755 1 tasks_processing.go:69] worker 19 listening for tasks. I0424 09:06:06.629763 1 tasks_processing.go:69] worker 20 listening for tasks. I0424 09:06:06.629765 1 tasks_processing.go:71] worker 19 working on support_secret task. I0424 09:06:06.629772 1 tasks_processing.go:69] worker 21 listening for tasks. I0424 09:06:06.629781 1 tasks_processing.go:69] worker 43 listening for tasks. I0424 09:06:06.629786 1 tasks_processing.go:69] worker 42 listening for tasks. I0424 09:06:06.629790 1 tasks_processing.go:69] worker 44 listening for tasks. I0424 09:06:06.629796 1 tasks_processing.go:69] worker 32 listening for tasks. I0424 09:06:06.629798 1 tasks_processing.go:69] worker 45 listening for tasks. I0424 09:06:06.629801 1 tasks_processing.go:69] worker 31 listening for tasks. I0424 09:06:06.629805 1 tasks_processing.go:69] worker 33 listening for tasks. I0424 09:06:06.629806 1 tasks_processing.go:69] worker 46 listening for tasks. I0424 09:06:06.629812 1 tasks_processing.go:69] worker 27 listening for tasks. I0424 09:06:06.629812 1 tasks_processing.go:69] worker 47 listening for tasks. I0424 09:06:06.629816 1 tasks_processing.go:69] worker 34 listening for tasks. I0424 09:06:06.629819 1 tasks_processing.go:69] worker 48 listening for tasks. I0424 09:06:06.629823 1 tasks_processing.go:69] worker 35 listening for tasks. I0424 09:06:06.629826 1 tasks_processing.go:69] worker 49 listening for tasks. I0424 09:06:06.629832 1 tasks_processing.go:69] worker 36 listening for tasks. I0424 09:06:06.629829 1 tasks_processing.go:69] worker 26 listening for tasks. I0424 09:06:06.629835 1 tasks_processing.go:69] worker 57 listening for tasks. I0424 09:06:06.629840 1 tasks_processing.go:69] worker 37 listening for tasks. I0424 09:06:06.629848 1 tasks_processing.go:69] worker 38 listening for tasks. I0424 09:06:06.629863 1 tasks_processing.go:69] worker 22 listening for tasks. I0424 09:06:06.629875 1 tasks_processing.go:69] worker 23 listening for tasks. I0424 09:06:06.629881 1 tasks_processing.go:69] worker 24 listening for tasks. I0424 09:06:06.629897 1 tasks_processing.go:69] worker 28 listening for tasks. I0424 09:06:06.629906 1 tasks_processing.go:69] worker 9 listening for tasks. I0424 09:06:06.629910 1 tasks_processing.go:69] worker 0 listening for tasks. I0424 09:06:06.629910 1 tasks_processing.go:69] worker 58 listening for tasks. I0424 09:06:06.629909 1 tasks_processing.go:69] worker 54 listening for tasks. I0424 09:06:06.629916 1 tasks_processing.go:69] worker 1 listening for tasks. I0424 09:06:06.629920 1 tasks_processing.go:69] worker 2 listening for tasks. I0424 09:06:06.629921 1 tasks_processing.go:69] worker 53 listening for tasks. I0424 09:06:06.629923 1 tasks_processing.go:69] worker 59 listening for tasks. I0424 09:06:06.629926 1 tasks_processing.go:69] worker 60 listening for tasks. I0424 09:06:06.629931 1 tasks_processing.go:69] worker 61 listening for tasks. I0424 09:06:06.629933 1 tasks_processing.go:69] worker 30 listening for tasks. I0424 09:06:06.629957 1 tasks_processing.go:69] worker 7 listening for tasks. I0424 09:06:06.629964 1 tasks_processing.go:69] worker 5 listening for tasks. I0424 09:06:06.629971 1 tasks_processing.go:69] worker 14 listening for tasks. I0424 09:06:06.629975 1 tasks_processing.go:69] worker 8 listening for tasks. I0424 09:06:06.629982 1 tasks_processing.go:69] worker 10 listening for tasks. I0424 09:06:06.629994 1 tasks_processing.go:69] worker 12 listening for tasks. I0424 09:06:06.629997 1 tasks_processing.go:69] worker 41 listening for tasks. I0424 09:06:06.630004 1 tasks_processing.go:69] worker 11 listening for tasks. I0424 09:06:06.630015 1 tasks_processing.go:69] worker 51 listening for tasks. I0424 09:06:06.630016 1 tasks_processing.go:69] worker 17 listening for tasks. I0424 09:06:06.630024 1 tasks_processing.go:69] worker 15 listening for tasks. I0424 09:06:06.630039 1 tasks_processing.go:69] worker 18 listening for tasks. I0424 09:06:06.629758 1 tasks_processing.go:69] worker 63 listening for tasks. I0424 09:06:06.630064 1 tasks_processing.go:69] worker 13 listening for tasks. I0424 09:06:06.630072 1 tasks_processing.go:69] worker 16 listening for tasks. I0424 09:06:06.630097 1 tasks_processing.go:69] worker 4 listening for tasks. I0424 09:06:06.630104 1 tasks_processing.go:69] worker 50 listening for tasks. I0424 09:06:06.630102 1 tasks_processing.go:69] worker 6 listening for tasks. I0424 09:06:06.630115 1 tasks_processing.go:71] worker 21 working on cluster_apiserver task. I0424 09:06:06.630123 1 tasks_processing.go:71] worker 20 working on mutating_webhook_configurations task. I0424 09:06:06.630128 1 tasks_processing.go:71] worker 6 working on image task. I0424 09:06:06.630142 1 tasks_processing.go:69] worker 3 listening for tasks. I0424 09:06:06.630143 1 tasks_processing.go:71] worker 58 working on jaegers task. I0424 09:06:06.630144 1 tasks_processing.go:69] worker 62 listening for tasks. I0424 09:06:06.630150 1 tasks_processing.go:71] worker 10 working on openshift_logging task. I0424 09:06:06.630154 1 tasks_processing.go:71] worker 62 working on nodenetworkstates task. I0424 09:06:06.630155 1 tasks_processing.go:71] worker 54 working on metrics task. I0424 09:06:06.630147 1 tasks_processing.go:71] worker 3 working on machines task. I0424 09:06:06.630166 1 tasks_processing.go:71] worker 30 working on version task. I0424 09:06:06.630176 1 tasks_processing.go:71] worker 1 working on cost_management_metrics_configs task. I0424 09:06:06.630180 1 tasks_processing.go:69] worker 29 listening for tasks. W0424 09:06:06.630191 1 gather_most_recent_metrics.go:64] Unable to load metrics client, no metrics will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0424 09:06:06.630197 1 tasks_processing.go:71] worker 53 working on dvo_metrics task. I0424 09:06:06.630196 1 tasks_processing.go:69] worker 40 listening for tasks. I0424 09:06:06.630203 1 gather.go:177] gatherer "clusterconfig" function "metrics" took 29.688µs to process 0 records I0424 09:06:06.630214 1 tasks_processing.go:71] worker 5 working on schedulers task. I0424 09:06:06.630214 1 tasks_processing.go:71] worker 11 working on container_runtime_configs task. I0424 09:06:06.630151 1 tasks_processing.go:71] worker 60 working on overlapping_namespace_uids task. I0424 09:06:06.630216 1 tasks_processing.go:69] worker 56 listening for tasks. I0424 09:06:06.630225 1 tasks_processing.go:71] worker 14 working on active_alerts task. I0424 09:06:06.630235 1 tasks_processing.go:71] worker 35 working on openstack_dataplanenodesets task. I0424 09:06:06.630247 1 tasks_processing.go:71] worker 38 working on silenced_alerts task. W0424 09:06:06.630273 1 gather_silenced_alerts.go:38] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0424 09:06:06.630283 1 tasks_processing.go:71] worker 7 working on aggregated_monitoring_cr_names task. I0424 09:06:06.630149 1 tasks_processing.go:69] worker 55 listening for tasks. I0424 09:06:06.630292 1 tasks_processing.go:71] worker 29 working on sap_datahubs task. I0424 09:06:06.630297 1 tasks_processing.go:71] worker 55 working on olm_operators task. I0424 09:06:06.630302 1 tasks_processing.go:71] worker 26 working on operators task. I0424 09:06:06.630208 1 tasks_processing.go:71] worker 59 working on certificate_signing_requests task. I0424 09:06:06.630436 1 tasks_processing.go:71] worker 15 working on validating_webhook_configurations task. I0424 09:06:06.630444 1 tasks_processing.go:71] worker 43 working on machine_autoscalers task. I0424 09:06:06.630450 1 tasks_processing.go:71] worker 4 working on clusterroles task. I0424 09:06:06.630538 1 tasks_processing.go:71] worker 18 working on openstack_version task. W0424 09:06:06.630267 1 gather_active_alerts.go:54] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0424 09:06:06.630235 1 tasks_processing.go:69] worker 39 listening for tasks. I0424 09:06:06.630578 1 tasks_processing.go:71] worker 39 working on oauths task. I0424 09:06:06.630163 1 tasks_processing.go:71] worker 61 working on image_pruners task. I0424 09:06:06.630777 1 tasks_processing.go:71] worker 8 working on qemu_kubevirt_launcher_logs task. I0424 09:06:06.630811 1 tasks_processing.go:71] worker 28 working on storage_cluster task. I0424 09:06:06.630896 1 tasks_processing.go:71] worker 22 working on lokistack task. I0424 09:06:06.630169 1 tasks_processing.go:71] worker 2 working on crds task. I0424 09:06:06.630209 1 tasks_processing.go:71] worker 41 working on feature_gates task. I0424 09:06:06.630293 1 tasks_processing.go:71] worker 49 working on ceph_cluster task. I0424 09:06:06.630886 1 tasks_processing.go:69] worker 52 listening for tasks. I0424 09:06:06.631486 1 gather.go:177] gatherer "clusterconfig" function "silenced_alerts" took 31.317µs to process 0 records I0424 09:06:06.631506 1 gather.go:177] gatherer "clusterconfig" function "active_alerts" took 329.236µs to process 0 records I0424 09:06:06.631518 1 tasks_processing.go:71] worker 9 working on operators_pods_and_events task. I0424 09:06:06.631533 1 tasks_processing.go:71] worker 23 working on pod_network_connectivity_checks task. I0424 09:06:06.631571 1 tasks_processing.go:71] worker 52 working on monitoring_persistent_volumes task. I0424 09:06:06.631776 1 tasks_processing.go:71] worker 40 working on nodenetworkconfigurationpolicies task. I0424 09:06:06.630159 1 tasks_processing.go:69] worker 25 listening for tasks. I0424 09:06:06.630427 1 tasks_processing.go:71] worker 63 working on install_plans task. I0424 09:06:06.631920 1 tasks_processing.go:71] worker 50 working on number_of_pods_and_netnamespaces_with_sdn_annotations task. I0424 09:06:06.631962 1 tasks_processing.go:71] worker 33 working on service_accounts task. I0424 09:06:06.630298 1 tasks_processing.go:71] worker 36 working on proxies task. I0424 09:06:06.631577 1 tasks_processing.go:71] worker 0 working on authentication task. I0424 09:06:06.632184 1 tasks_processing.go:71] worker 32 working on machine_healthchecks task. I0424 09:06:06.630390 1 tasks_processing.go:71] worker 37 working on tsdb_status task. I0424 09:06:06.630430 1 tasks_processing.go:71] worker 51 working on config_maps task. I0424 09:06:06.630412 1 tasks_processing.go:71] worker 57 working on openstack_dataplanedeployments task. I0424 09:06:06.630288 1 tasks_processing.go:71] worker 38 working on pdbs task. I0424 09:06:06.630439 1 tasks_processing.go:71] worker 13 working on machine_config_pools task. I0424 09:06:06.632133 1 tasks_processing.go:71] worker 24 working on machine_sets task. W0424 09:06:06.632378 1 gather_prometheus_tsdb_status.go:38] Unable to load metrics client, tsdb status cannot be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0424 09:06:06.632448 1 tasks_processing.go:74] worker 14 stopped. I0424 09:06:06.630284 1 tasks_processing.go:71] worker 56 working on ingress task. I0424 09:06:06.632435 1 tasks_processing.go:71] worker 48 working on networks task. I0424 09:06:06.630444 1 tasks_processing.go:71] worker 16 working on node_logs task. I0424 09:06:06.630433 1 tasks_processing.go:71] worker 17 working on infrastructures task. I0424 09:06:06.632200 1 tasks_processing.go:71] worker 27 working on ingress_certificates task. I0424 09:06:06.632239 1 tasks_processing.go:71] worker 34 working on nodes task. I0424 09:06:06.632266 1 tasks_processing.go:71] worker 46 working on openshift_machine_api_events task. I0424 09:06:06.630205 1 tasks_processing.go:71] worker 12 working on image_registries task. I0424 09:06:06.632203 1 tasks_processing.go:71] worker 45 working on sap_pods task. I0424 09:06:06.632351 1 tasks_processing.go:71] worker 47 working on container_images task. I0424 09:06:06.630419 1 tasks_processing.go:71] worker 54 working on sap_config task. I0424 09:06:06.632464 1 gather.go:177] gatherer "clusterconfig" function "tsdb_status" took 91.032µs to process 0 records I0424 09:06:06.632506 1 tasks_processing.go:71] worker 44 working on machine_configs task. I0424 09:06:06.632512 1 tasks_processing.go:71] worker 42 working on storage_classes task. I0424 09:06:06.632609 1 tasks_processing.go:74] worker 25 stopped. I0424 09:06:06.632311 1 tasks_processing.go:71] worker 31 working on openstack_controlplanes task. I0424 09:06:06.632454 1 tasks_processing.go:74] worker 37 stopped. I0424 09:06:06.634277 1 tasks_processing.go:74] worker 58 stopped. I0424 09:06:06.634291 1 gather.go:177] gatherer "clusterconfig" function "jaegers" took 4.124483ms to process 0 records I0424 09:06:06.634592 1 tasks_processing.go:74] worker 11 stopped. I0424 09:06:06.634608 1 gather.go:177] gatherer "clusterconfig" function "container_runtime_configs" took 4.354064ms to process 0 records I0424 09:06:06.634897 1 controller.go:128] Initializing last reported time to 0001-01-01T00:00:00Z E0424 09:06:06.634910 1 gather.go:140] gatherer "clusterconfig" function "support_secret" failed with the error: secrets "support" not found I0424 09:06:06.634919 1 controller.go:212] Source periodic-workloads *controllerstatus.Simple is not ready I0424 09:06:06.634921 1 gather.go:177] gatherer "clusterconfig" function "support_secret" took 5.12371ms to process 0 records I0424 09:06:06.634927 1 controller.go:212] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0424 09:06:06.634931 1 controller.go:212] Source periodic-conditional *controllerstatus.Simple is not ready I0424 09:06:06.634948 1 controller.go:489] The operator is still being initialized I0424 09:06:06.634956 1 controller.go:512] The operator is healthy I0424 09:06:06.634902 1 tasks_processing.go:74] worker 19 stopped. I0424 09:06:06.636177 1 tasks_processing.go:74] worker 35 stopped. I0424 09:06:06.636190 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanenodesets" took 5.92244ms to process 0 records I0424 09:06:06.636445 1 tasks_processing.go:74] worker 21 stopped. I0424 09:06:06.636612 1 recorder.go:75] Recording config/apiserver with fingerprint=93c24f0de17a58fe5940b0c6c1cb436e468f17dd4941b5865040718dd8d6e71a I0424 09:06:06.636626 1 gather.go:177] gatherer "clusterconfig" function "cluster_apiserver" took 6.316442ms to process 1 records I0424 09:06:06.636951 1 tasks_processing.go:74] worker 43 stopped. I0424 09:06:06.636961 1 gather.go:177] gatherer "clusterconfig" function "machine_autoscalers" took 6.492821ms to process 0 records I0424 09:06:06.638659 1 tasks_processing.go:74] worker 62 stopped. I0424 09:06:06.638666 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkstates" took 8.49609ms to process 0 records I0424 09:06:06.638673 1 gather.go:177] gatherer "clusterconfig" function "machine_sets" took 6.293562ms to process 0 records I0424 09:06:06.638680 1 tasks_processing.go:74] worker 24 stopped. I0424 09:06:06.638774 1 tasks_processing.go:74] worker 3 stopped. E0424 09:06:06.638786 1 gather.go:140] gatherer "clusterconfig" function "machines" failed with the error: machines.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machines" in API group "machine.openshift.io" at the cluster scope I0424 09:06:06.638795 1 gather.go:177] gatherer "clusterconfig" function "machines" took 8.592339ms to process 0 records I0424 09:06:06.656917 1 tasks_processing.go:74] worker 49 stopped. I0424 09:06:06.656933 1 gather.go:177] gatherer "clusterconfig" function "ceph_cluster" took 25.483504ms to process 0 records I0424 09:06:06.656943 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanedeployments" took 24.558023ms to process 0 records I0424 09:06:06.656950 1 tasks_processing.go:74] worker 57 stopped. I0424 09:06:06.656978 1 tasks_processing.go:74] worker 40 stopped. I0424 09:06:06.656988 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkconfigurationpolicies" took 25.190024ms to process 0 records I0424 09:06:06.656996 1 gather.go:177] gatherer "clusterconfig" function "storage_cluster" took 26.156026ms to process 0 records I0424 09:06:06.657002 1 tasks_processing.go:74] worker 28 stopped. I0424 09:06:06.657058 1 tasks_processing.go:74] worker 32 stopped. E0424 09:06:06.657097 1 gather.go:140] gatherer "clusterconfig" function "machine_healthchecks" failed with the error: machinehealthchecks.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machinehealthchecks" in API group "machine.openshift.io" at the cluster scope I0424 09:06:06.657109 1 gather.go:177] gatherer "clusterconfig" function "machine_healthchecks" took 24.586524ms to process 0 records I0424 09:06:06.657209 1 tasks_processing.go:74] worker 5 stopped. I0424 09:06:06.657253 1 recorder.go:75] Recording config/schedulers/cluster with fingerprint=19a80918e832b0bb1df92e32d570297e8a9a0486929da077b0dd30bd13360ca8 I0424 09:06:06.657270 1 gather.go:177] gatherer "clusterconfig" function "schedulers" took 26.885049ms to process 1 records I0424 09:06:06.657279 1 gather.go:177] gatherer "clusterconfig" function "openshift_logging" took 26.987145ms to process 0 records I0424 09:06:06.657284 1 gather.go:177] gatherer "clusterconfig" function "cost_management_metrics_configs" took 26.957322ms to process 0 records I0424 09:06:06.657289 1 tasks_processing.go:74] worker 1 stopped. I0424 09:06:06.657289 1 tasks_processing.go:74] worker 10 stopped. I0424 09:06:06.657384 1 tasks_processing.go:74] worker 55 stopped. I0424 09:06:06.657414 1 gather.go:177] gatherer "clusterconfig" function "olm_operators" took 27.079501ms to process 0 records I0424 09:06:06.657686 1 tasks_processing.go:74] worker 23 stopped. E0424 09:06:06.657698 1 gather.go:140] gatherer "clusterconfig" function "pod_network_connectivity_checks" failed with the error: the server could not find the requested resource (get podnetworkconnectivitychecks.controlplane.operator.openshift.io) I0424 09:06:06.657708 1 gather.go:177] gatherer "clusterconfig" function "pod_network_connectivity_checks" took 26.144407ms to process 0 records I0424 09:06:06.657753 1 tasks_processing.go:74] worker 36 stopped. I0424 09:06:06.657819 1 recorder.go:75] Recording config/proxy with fingerprint=5993599663bc7739eae79d5ba918063e1c0b57a189ef80abd4fe9ec6f1acb977 I0424 09:06:06.657834 1 gather.go:177] gatherer "clusterconfig" function "proxies" took 25.675689ms to process 1 records I0424 09:06:06.657922 1 tasks_processing.go:74] worker 0 stopped. I0424 09:06:06.658017 1 recorder.go:75] Recording config/authentication with fingerprint=bfedab60c69106f7fc696df31a2d31e2a9bc962cf402a335f478cc7c0b84eb04 I0424 09:06:06.658028 1 gather.go:177] gatherer "clusterconfig" function "authentication" took 25.632891ms to process 1 records I0424 09:06:06.658192 1 recorder.go:75] Recording config/oauth with fingerprint=e32e011cefb808adae3e9afe6054d4f1922c37f5b73b832b4f79b8eae3265c8a I0424 09:06:06.658206 1 gather.go:177] gatherer "clusterconfig" function "oauths" took 27.426398ms to process 1 records I0424 09:06:06.658194 1 tasks_processing.go:74] worker 39 stopped. I0424 09:06:06.658213 1 gather.go:177] gatherer "clusterconfig" function "lokistack" took 27.226237ms to process 0 records I0424 09:06:06.658228 1 gather.go:177] gatherer "clusterconfig" function "sap_datahubs" took 27.846909ms to process 0 records I0424 09:06:06.658234 1 tasks_processing.go:74] worker 22 stopped. I0424 09:06:06.658238 1 tasks_processing.go:74] worker 29 stopped. I0424 09:06:06.658273 1 tasks_processing.go:74] worker 17 stopped. I0424 09:06:06.658752 1 recorder.go:75] Recording config/infrastructure with fingerprint=21dcefed596d7790b7fab690cc183ef098142df5830088478e78152e31d36c4c I0424 09:06:06.658765 1 gather.go:177] gatherer "clusterconfig" function "infrastructures" took 25.554907ms to process 1 records I0424 09:06:06.658860 1 tasks_processing.go:74] worker 48 stopped. I0424 09:06:06.658873 1 recorder.go:75] Recording config/network with fingerprint=c4f0602480cfa4153f50866e4624ecb6bf575c3291a17bacdb883b769d38d787 I0424 09:06:06.658879 1 gather.go:177] gatherer "clusterconfig" function "networks" took 25.913429ms to process 1 records I0424 09:06:06.660868 1 tasks_processing.go:74] worker 16 stopped. I0424 09:06:06.660946 1 gather.go:177] gatherer "clusterconfig" function "node_logs" took 28.213218ms to process 0 records I0424 09:06:06.661166 1 recorder.go:75] Recording config/pdbs/openshift-image-registry/image-registry with fingerprint=9e58b9300d6141b2857fb2d8239b97f7027da33d405cdd182ba588f33f0ed73e I0424 09:06:06.661238 1 recorder.go:75] Recording config/pdbs/openshift-ingress/router-default with fingerprint=54077c8605dc0a359642dcf9aeb33f45f3da70334d79b61241aabf2741a67b52 I0424 09:06:06.661303 1 recorder.go:75] Recording config/pdbs/openshift-operator-lifecycle-manager/packageserver-pdb with fingerprint=3890ef1efd33ce240705c2c224c69480180090dcc68341cce97c7dd2d76591a0 I0424 09:06:06.661348 1 gather.go:177] gatherer "clusterconfig" function "pdbs" took 28.079431ms to process 3 records I0424 09:06:06.661376 1 gather.go:177] gatherer "clusterconfig" function "openstack_controlplanes" took 27.38422ms to process 0 records I0424 09:06:06.661427 1 gather.go:177] gatherer "clusterconfig" function "openstack_version" took 30.564316ms to process 0 records I0424 09:06:06.661460 1 gather.go:177] gatherer "clusterconfig" function "sap_pods" took 27.988665ms to process 0 records I0424 09:06:06.661485 1 tasks_processing.go:74] worker 31 stopped. I0424 09:06:06.661555 1 tasks_processing.go:74] worker 38 stopped. I0424 09:06:06.661573 1 tasks_processing.go:74] worker 18 stopped. I0424 09:06:06.661589 1 tasks_processing.go:74] worker 45 stopped. I0424 09:06:06.661645 1 recorder.go:75] Recording config/featuregate with fingerprint=fc564a9c00b9c52b07365375184b0316d38f535f4a7ba62674abb0f02668f8cc I0424 09:06:06.661651 1 tasks_processing.go:74] worker 41 stopped. I0424 09:06:06.661660 1 gather.go:177] gatherer "clusterconfig" function "feature_gates" took 29.948166ms to process 1 records I0424 09:06:06.661699 1 tasks_processing.go:74] worker 6 stopped. I0424 09:06:06.661816 1 recorder.go:75] Recording config/image with fingerprint=5c5f220aacefb1ce5ef4e39c87e7c48f39071e017a820e371891a5a01ed5fa7d I0424 09:06:06.661880 1 gather.go:177] gatherer "clusterconfig" function "image" took 31.301841ms to process 1 records I0424 09:06:06.661894 1 gather.go:177] gatherer "clusterconfig" function "sap_config" took 28.068886ms to process 0 records I0424 09:06:06.661921 1 tasks_processing.go:74] worker 54 stopped. I0424 09:06:06.661994 1 tasks_processing.go:74] worker 42 stopped. I0424 09:06:06.662710 1 recorder.go:75] Recording config/storage/storageclasses/gp2-csi with fingerprint=0509a2192b6590c0a612e5a3b06ce3fd0f11821061f2f37fc4d5a22cbac8db62 I0424 09:06:06.662741 1 recorder.go:75] Recording config/storage/storageclasses/gp3-csi with fingerprint=a51d94d1ecf39ec2064fda3a6f0113cdebd7ac884e1178cd11a9f0c147305d21 I0424 09:06:06.662749 1 gather.go:177] gatherer "clusterconfig" function "storage_classes" took 27.94502ms to process 2 records I0424 09:06:06.662833 1 tasks_processing.go:74] worker 56 stopped. I0424 09:06:06.662883 1 recorder.go:75] Recording config/ingress with fingerprint=06b5ca66ad0497941306511c0672830ea8e773eee96add73715dce923bdffa4e I0424 09:06:06.662897 1 gather.go:177] gatherer "clusterconfig" function "ingress" took 28.980381ms to process 1 records I0424 09:06:06.662994 1 gather_logs.go:145] no pods in namespace were found I0424 09:06:06.663008 1 tasks_processing.go:74] worker 8 stopped. I0424 09:06:06.663018 1 gather.go:177] gatherer "clusterconfig" function "qemu_kubevirt_launcher_logs" took 32.213524ms to process 0 records I0424 09:06:06.663072 1 tasks_processing.go:74] worker 52 stopped. I0424 09:06:06.663087 1 gather.go:177] gatherer "clusterconfig" function "monitoring_persistent_volumes" took 31.485443ms to process 0 records I0424 09:06:06.663150 1 tasks_processing.go:74] worker 15 stopped. I0424 09:06:06.663304 1 recorder.go:75] Recording config/validatingwebhookconfigurations/multus.openshift.io with fingerprint=10e635606bd34f44e375f6190e6d4449ea02de59d807098b354f9a615990f64c I0424 09:06:06.663424 1 recorder.go:75] Recording config/validatingwebhookconfigurations/network-node-identity.openshift.io with fingerprint=208980643f30eee89dbfdba5498a3bb28e0e3f88ea163ff443499f51459de425 I0424 09:06:06.663455 1 recorder.go:75] Recording config/validatingwebhookconfigurations/performance-addon-operator with fingerprint=e480586c196c654ad726baa177acfa33b04766c731a44c446915b97b77721e1c I0424 09:06:06.663490 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterrolebindings-validation with fingerprint=1c6a8fe82d47d5c36ee0f2902aceaf6707cd5cd21b2cc15e1fdb8d683eaac542 I0424 09:06:06.663524 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterroles-validation with fingerprint=16ee62dc8936a2a479ad7e953269d086d430b3019e69f4c69e58581abf6d01c6 I0424 09:06:06.663561 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-ingress-config-validation with fingerprint=311a8de94ca3efedd959eb73d5bd55fa8d1d107497e9e2373c3734cfc2190ac7 I0424 09:06:06.663620 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-network-operator-validation with fingerprint=611f6336a2964734051c3c6e1ff552fd64320feebf0d38ac6dfed0290ac8fb28 I0424 09:06:06.663669 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-regular-user-validation with fingerprint=e27e9386960b767128de91e0b673af57effbe9b5bd35b5cec79deedbf16ad51a I0424 09:06:06.663716 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-scc-validation with fingerprint=9b4b51953ab9c1ed45970d6113b95fa2adfbe5d0e69edb6c69512046310f8f1d I0424 09:06:06.663749 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-serviceaccount-validation with fingerprint=bf0747ac8535119a954af9a180dbbaa146d6e851a2cf399c3ba29321b25b665e I0424 09:06:06.663782 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-techpreviewnoupgrade-validation with fingerprint=2e188b2be4110929bf7a332e5ad9ed94fa586d5bb344be15dc5c8c2a12b36464 I0424 09:06:06.663794 1 gather.go:177] gatherer "clusterconfig" function "validating_webhook_configurations" took 32.625058ms to process 11 records I0424 09:06:06.663816 1 recorder.go:75] Recording config/namespaces_with_overlapping_uids with fingerprint=4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945 I0424 09:06:06.663827 1 gather.go:177] gatherer "clusterconfig" function "overlapping_namespace_uids" took 33.430449ms to process 1 records I0424 09:06:06.663836 1 tasks_processing.go:74] worker 60 stopped. W0424 09:06:06.664083 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0424 09:06:06.664222 1 tasks_processing.go:74] worker 34 stopped. I0424 09:06:06.664624 1 recorder.go:75] Recording config/node/ip-10-0-0-87.ec2.internal with fingerprint=9e3253e53d806f978b9b39a96426df9f5893f2e8af3f5d47f983e9e2672f9a8e I0424 09:06:06.664682 1 recorder.go:75] Recording config/node/ip-10-0-1-210.ec2.internal with fingerprint=fbc39a14e7fbaf7db2f37e0b517075b412f91ab5ad5cd9a9f2bda5dbe8022ffb I0424 09:06:06.664739 1 recorder.go:75] Recording config/node/ip-10-0-2-43.ec2.internal with fingerprint=9a67526332d61bab37f4b91fa95354f6c38eed036bee4ce40615630504dd78e9 I0424 09:06:06.664747 1 gather.go:177] gatherer "clusterconfig" function "nodes" took 31.230801ms to process 3 records I0424 09:06:06.664833 1 tasks_processing.go:74] worker 12 stopped. I0424 09:06:06.665141 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/config/cluster with fingerprint=8909d62a2bea8d70fd7d78730f0f7c38add98ccb6733a410bf7ffea5817c45f7 I0424 09:06:06.665362 1 gather.go:177] gatherer "clusterconfig" function "image_registries" took 31.254823ms to process 1 records I0424 09:06:06.665478 1 tasks_processing.go:74] worker 20 stopped. I0424 09:06:06.665725 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/aws-pod-identity with fingerprint=8d18d24944d7ed54eb5a2710071d1b23e0c07e370397db40c540f4b0986964fe I0424 09:06:06.665866 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-podimagespec-mutation with fingerprint=a1c2133af6a93f7ff2f4cc0854cf4e312b9bb5c0c5d372ba74a50a8d1154459b I0424 09:06:06.665913 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-service-mutation with fingerprint=a3414c42135ecfcb996600d1beb17cd8a6b35869cd22e1c6551af775a7e79112 I0424 09:06:06.665963 1 gather.go:177] gatherer "clusterconfig" function "mutating_webhook_configurations" took 34.319204ms to process 3 records I0424 09:06:06.665976 1 gather.go:177] gatherer "clusterconfig" function "certificate_signing_requests" took 34.694023ms to process 0 records I0424 09:06:06.665985 1 tasks_processing.go:74] worker 59 stopped. I0424 09:06:06.673973 1 tasks_processing.go:74] worker 61 stopped. I0424 09:06:06.674094 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/imagepruner/cluster with fingerprint=6ba87edb025debe97431f6c2f23ff1552ba8be0983702a3800befd1a87535cdc I0424 09:06:06.674113 1 gather.go:177] gatherer "clusterconfig" function "image_pruners" took 43.210539ms to process 1 records I0424 09:06:06.674124 1 gather.go:177] gatherer "clusterconfig" function "openshift_machine_api_events" took 41.095975ms to process 0 records I0424 09:06:06.674135 1 tasks_processing.go:74] worker 46 stopped. I0424 09:06:06.674540 1 tasks_processing.go:74] worker 4 stopped. I0424 09:06:06.674704 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/admin with fingerprint=ea916f026a0510280233baa8a64ba66ccea45353f259d31acc16e1702f1fd7c0 I0424 09:06:06.674803 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/edit with fingerprint=96b1e303bfdfd4cb3a680c97b0f392f6ebe2db09085035ae87e5a0825baebf05 I0424 09:06:06.674813 1 gather.go:177] gatherer "clusterconfig" function "clusterroles" took 44.081297ms to process 2 records I0424 09:06:06.675691 1 tasks_processing.go:74] worker 7 stopped. I0424 09:06:06.675703 1 gather.go:177] gatherer "clusterconfig" function "aggregated_monitoring_cr_names" took 45.394334ms to process 0 records I0424 09:06:06.676261 1 sca.go:136] Pulling SCA certificates from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates. Next check is in 8h0m0s I0424 09:06:06.676266 1 cluster_transfer.go:83] checking the availability of cluster transfer. Next check is in 12h0m0s W0424 09:06:06.676387 1 operator.go:288] started I0424 09:06:06.676423 1 base_controller.go:76] Waiting for caches to sync for LoggingSyncer I0424 09:06:06.677705 1 tasks_processing.go:74] worker 2 stopped. I0424 09:06:06.679330 1 recorder.go:75] Recording config/crd/volumesnapshots.snapshot.storage.k8s.io with fingerprint=ac5048d4b5a5ed2d42ddbbafb884518609adc8ea446595a02d4e188e0a3bd89c I0424 09:06:06.679554 1 recorder.go:75] Recording config/crd/volumesnapshotcontents.snapshot.storage.k8s.io with fingerprint=f3d49a81563f3a85af2e988c695831839a4165f874c1b2afef920ccbc1595822 I0424 09:06:06.679566 1 gather.go:177] gatherer "clusterconfig" function "crds" took 46.289569ms to process 2 records I0424 09:06:06.679687 1 tasks_processing.go:74] worker 30 stopped. I0424 09:06:06.679899 1 recorder.go:75] Recording config/version with fingerprint=b89b09ca3ce1010ea94e1f8da5bb9f3ad39922abea475f38085704e99877b0b5 I0424 09:06:06.679908 1 recorder.go:75] Recording config/id with fingerprint=d33ab0550bce0bf6bc65010f5841687cd8b1a380635a3e42cdf16899ebd2aaaa I0424 09:06:06.679914 1 gather.go:177] gatherer "clusterconfig" function "version" took 49.506532ms to process 2 records I0424 09:06:06.681174 1 tasks_processing.go:74] worker 47 stopped. I0424 09:06:06.682467 1 recorder.go:75] Recording config/pod/openshift-console-operator/console-operator-575cd97545-n5h8g with fingerprint=a98fa8b42e852cade70ebc24c49dc1be760d9eed97f13751a4dc511f314126aa I0424 09:06:06.682518 1 recorder.go:75] Recording config/running_containers with fingerprint=b9d23371a03ce23c45c19f682856acf0aa057be3c0e9826500d74a3e7c86311d I0424 09:06:06.682527 1 gather.go:177] gatherer "clusterconfig" function "container_images" took 47.822967ms to process 2 records I0424 09:06:06.686954 1 tasks_processing.go:74] worker 50 stopped. I0424 09:06:06.686965 1 gather.go:177] gatherer "clusterconfig" function "number_of_pods_and_netnamespaces_with_sdn_annotations" took 55.015045ms to process 0 records I0424 09:06:06.692574 1 prometheus_rules.go:88] Prometheus rules successfully created I0424 09:06:06.697888 1 controller.go:212] Source scaController *sca.Controller is not ready I0424 09:06:06.697899 1 controller.go:212] Source clusterTransferController *clustertransfer.Controller is not ready I0424 09:06:06.697903 1 controller.go:212] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0424 09:06:06.697906 1 controller.go:212] Source periodic-conditional *controllerstatus.Simple is not ready I0424 09:06:06.697908 1 controller.go:212] Source periodic-workloads *controllerstatus.Simple is not ready I0424 09:06:06.697924 1 controller.go:489] The operator is still being initialized I0424 09:06:06.697929 1 controller.go:512] The operator is healthy I0424 09:06:06.697962 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0424 09:06:06.697998 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController I0424 09:06:06.698001 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0424 09:06:06.704330 1 base_controller.go:82] Caches are synced for ConfigController I0424 09:06:06.704343 1 base_controller.go:119] Starting #1 worker of ConfigController controller ... E0424 09:06:06.704960 1 cluster_transfer.go:95] failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%274d8dd466-8450-4012-a91b-24f949e8f1c5%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:55171->172.30.0.10:53: read: connection refused I0424 09:06:06.704974 1 controllerstatus.go:80] name=clusterTransferController healthy=true reason=Disconnected message=failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%274d8dd466-8450-4012-a91b-24f949e8f1c5%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:55171->172.30.0.10:53: read: connection refused I0424 09:06:06.714993 1 tasks_processing.go:74] worker 51 stopped. E0424 09:06:06.715007 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "cluster-monitoring-config" not found E0424 09:06:06.715012 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "gateway-mode-config" not found E0424 09:06:06.715016 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "insights-config" not found I0424 09:06:06.715025 1 recorder.go:75] Recording config/configmaps/openshift-config/installer-images/images.json with fingerprint=26b6661162b099a0f5a279859b4f46c867929a79d9a4a41fde4be4e6fe138018 I0424 09:06:06.715048 1 recorder.go:75] Recording config/configmaps/openshift-config/kube-root-ca.crt/ca.crt with fingerprint=d476c7d3f5b104863f08f481b1264dcc68cc272ecefb0ecb709b18a6afab034d I0424 09:06:06.715055 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/invoker with fingerprint=76b482f683cd3ef9da02debac5b26080a5aeb06ff768ee5c21117514dff29d8a I0424 09:06:06.715059 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/version with fingerprint=c93090eb0d2a4736885abeb79c91680cfd01fda46464f83456b085d4dc8239f0 I0424 09:06:06.715063 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 I0424 09:06:06.715099 1 recorder.go:75] Recording config/configmaps/openshift-config/rosa-brand-logo/rosa-brand-logo.svg with fingerprint=6ed8ca4dd7a8eee7249182bc006e9649ce84d76c551ddfaaa33e55d8c4cc1ed0 I0424 09:06:06.715106 1 recorder.go:75] Recording config/configmaps/kube-system/cluster-config-v1/install-config with fingerprint=ab3811c6b83fd7b8e920094cfa3080d1b4ee3c35ec4c8379437b21d27bd6608d I0424 09:06:06.715111 1 gather.go:177] gatherer "clusterconfig" function "config_maps" took 82.596587ms to process 7 records I0424 09:06:06.715459 1 configmapobserver.go:84] configmaps "insights-config" not found I0424 09:06:06.777478 1 base_controller.go:82] Caches are synced for LoggingSyncer I0424 09:06:06.777489 1 base_controller.go:119] Starting #1 worker of LoggingSyncer controller ... I0424 09:06:06.800823 1 tasks_processing.go:74] worker 27 stopped. E0424 09:06:06.800842 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret 'router-certs-default' in namespace 'openshift-ingress': secrets "router-certs-default" not found E0424 09:06:06.800854 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret '2pse0mo5k01bfflh5njr161oj1tspn3j-primary-cert-bundle-secret' in namespace 'openshift-ingress-operator': secrets "2pse0mo5k01bfflh5njr161oj1tspn3j-primary-cert-bundle-secret" not found I0424 09:06:06.800905 1 recorder.go:75] Recording aggregated/ingress_controllers_certs with fingerprint=29ab9c784d437b70d6cc451d4eeb1a9365bf5d0a653d63b5156435d22eab6b27 I0424 09:06:06.800918 1 gather.go:177] gatherer "clusterconfig" function "ingress_certificates" took 168.043167ms to process 1 records I0424 09:06:06.830021 1 requests.go:205] Asking for SCA certificate with "{"arch": ["x86_64"]}" payload W0424 09:06:06.833191 1 sca.go:161] Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:33093->172.30.0.10:53: read: connection refused I0424 09:06:06.833204 1 controllerstatus.go:80] name=scaController healthy=true reason=NonHTTPError message=Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:33093->172.30.0.10:53: read: connection refused I0424 09:06:06.842636 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0424 09:06:06.847962 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0424 09:06:07.072616 1 gather_cluster_operator_pods_and_events.go:121] Found 18 pods with 21 containers I0424 09:06:07.072628 1 gather_cluster_operator_pods_and_events.go:235] Maximum buffer size: 1198372 bytes I0424 09:06:07.073475 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-4mtts pod in namespace openshift-dns (previous: false). I0424 09:06:07.355926 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-4mtts pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-4mtts\" is waiting to start: ContainerCreating" I0424 09:06:07.355944 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-4mtts\" is waiting to start: ContainerCreating" I0424 09:06:07.355952 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-4mtts pod in namespace openshift-dns (previous: false). I0424 09:06:07.479275 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-4mtts pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-4mtts\" is waiting to start: ContainerCreating" I0424 09:06:07.479291 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-4mtts\" is waiting to start: ContainerCreating" I0424 09:06:07.479302 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-6znpb pod in namespace openshift-dns (previous: false). W0424 09:06:07.661244 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0424 09:06:07.710160 1 tasks_processing.go:74] worker 13 stopped. I0424 09:06:07.710180 1 gather.go:177] gatherer "clusterconfig" function "machine_config_pools" took 1.077231527s to process 0 records I0424 09:06:07.721702 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-6znpb pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-6znpb\" is waiting to start: ContainerCreating" I0424 09:06:07.721716 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-6znpb\" is waiting to start: ContainerCreating" I0424 09:06:07.721724 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-6znpb pod in namespace openshift-dns (previous: false). I0424 09:06:07.722194 1 tasks_processing.go:74] worker 44 stopped. I0424 09:06:07.722259 1 recorder.go:75] Recording aggregated/unused_machine_configs_count with fingerprint=4bfc9fa984e5dfcd45848faaf05269de7619bf42edf9f781751af5ee05c1a499 I0424 09:06:07.722272 1 gather.go:177] gatherer "clusterconfig" function "machine_configs" took 1.08874682s to process 1 records I0424 09:06:07.878251 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-6znpb pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-6znpb\" is waiting to start: ContainerCreating" I0424 09:06:07.878266 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-6znpb\" is waiting to start: ContainerCreating" I0424 09:06:07.878277 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-p9n9z pod in namespace openshift-dns (previous: false). I0424 09:06:08.109768 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-p9n9z pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-p9n9z\" is waiting to start: ContainerCreating" I0424 09:06:08.109787 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-p9n9z\" is waiting to start: ContainerCreating" I0424 09:06:08.109794 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-p9n9z pod in namespace openshift-dns (previous: false). I0424 09:06:08.112737 1 gather_cluster_operators.go:184] Unable to get configs.samples.operator.openshift.io resource due to: configs.samples.operator.openshift.io "cluster" not found I0424 09:06:08.277142 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-p9n9z pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-p9n9z\" is waiting to start: ContainerCreating" I0424 09:06:08.277158 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-p9n9z\" is waiting to start: ContainerCreating" I0424 09:06:08.277168 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-gqtwz pod in namespace openshift-dns (previous: false). I0424 09:06:08.477494 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0424 09:06:08.477511 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-vlf4r pod in namespace openshift-dns (previous: false). I0424 09:06:08.494628 1 tasks_processing.go:74] worker 26 stopped. I0424 09:06:08.494679 1 recorder.go:75] Recording config/clusteroperator/console with fingerprint=9ebe67613d4b5f6cdc3f280283cec6c67ef78d525c0890736ebf3b37a6a11fca I0424 09:06:08.494710 1 recorder.go:75] Recording config/clusteroperator/csi-snapshot-controller with fingerprint=d4510cd01002f1aceb6c0a288021476b3207703e38792ef578af6cdd9776d4e4 I0424 09:06:08.494739 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/csisnapshotcontroller/cluster with fingerprint=5adc514f4b63e2f1ecc68bf6f9c0af70c5eea04522a49524e102721b1c41f80e I0424 09:06:08.494768 1 recorder.go:75] Recording config/clusteroperator/dns with fingerprint=48c8a47ac7bf77e1973bd266b6472a2fa1b5c859da5a67f9ffee8c7e19aabe27 I0424 09:06:08.494784 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/dns/default with fingerprint=9e7b4ce029030d3d8c3b49af92c556acdcc415000b40d3f969dbdc42c432b47f I0424 09:06:08.494822 1 recorder.go:75] Recording config/clusteroperator/image-registry with fingerprint=b77b24206f41dc07f59b8214fa52a852649065abe2d3c83b7a8fab0858d5946c I0424 09:06:08.494852 1 recorder.go:75] Recording config/clusteroperator/ingress with fingerprint=3739d27a17b33c5c17f00d6252efa348d9d7228fd3688ad0762ac79f8f6bcd19 I0424 09:06:08.494876 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/ingresscontroller/openshift-ingress-operator/default with fingerprint=3f6958cbd6ced07caa9c48ee9e63461ec0a35403750aff5ba4611ceec6614a82 I0424 09:06:08.494889 1 recorder.go:75] Recording config/clusteroperator/insights with fingerprint=0354711e046c0e7edfdada61b20195439d8c234ea505484da05815e7b18e74f1 I0424 09:06:08.494906 1 recorder.go:75] Recording config/clusteroperator/kube-apiserver with fingerprint=ca4cc49651524eaf4b0cee3e9a7b480622af56b0d84f5d8e469ec47023cfe368 I0424 09:06:08.494915 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubeapiserver/cluster with fingerprint=51503bf0b784fcf65ea46bcaf1f72ac1a5c4d5dc211934f18f27871efed05762 I0424 09:06:08.494931 1 recorder.go:75] Recording config/clusteroperator/kube-controller-manager with fingerprint=6261f0532fbd9ceca3140acda754e5d848658e2dac87939c4ee584ec608e5fdd I0424 09:06:08.494940 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubecontrollermanager/cluster with fingerprint=ce90c0d4f367d7da085074268031798382ae7c54fdcb0a21f15a4818fe308c11 I0424 09:06:08.494955 1 recorder.go:75] Recording config/clusteroperator/kube-scheduler with fingerprint=172210f05f3c4b740b3cf8c3b52cbe59a3ead483a9c4a1da9848f2e60952d332 I0424 09:06:08.494965 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubescheduler/cluster with fingerprint=f2940fb9fd20c19951dfc295eb363b7fba0c505f5ae61f01967a063099e6b60a I0424 09:06:08.494977 1 recorder.go:75] Recording config/clusteroperator/kube-storage-version-migrator with fingerprint=7c58f7a1940d41d66d5ae553f9f8661e9f562a3edad3579bc3926a6bb41cd545 I0424 09:06:08.494986 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubestorageversionmigrator/cluster with fingerprint=9351181aa7e6ada41ef581ab31e13516c6b934cc95710154bafb2eb222cb58db I0424 09:06:08.495002 1 recorder.go:75] Recording config/clusteroperator/monitoring with fingerprint=27d0e6dc11f15f136842081c71809c69c5f0e93d2fb78803b21d696d716410ae I0424 09:06:08.495122 1 recorder.go:75] Recording config/clusteroperator/network with fingerprint=23024b109fc5f2d2ab88e17da33715041917afab28c6d263a2b43b38cd65b268 I0424 09:06:08.495131 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/ovn with fingerprint=626a89d20e0deaed5b6dfb533acfe65f4bb1618bd200a703b62e60c5d16d94ab I0424 09:06:08.495139 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/signer with fingerprint=90410b16914712b85b3c4578716ad8c0ae072e688f4cd1e022bf76f20da3506d I0424 09:06:08.495158 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/network/cluster with fingerprint=a93d15eaecb455a0e40ecb2826eeecc1533899204ddd3c3921d15ab70af7ae75 I0424 09:06:08.495179 1 recorder.go:75] Recording config/clusteroperator/node-tuning with fingerprint=dee966adf51b73da6d85577fc14a1ef9a0489758f0e96744c981ef6a963051ff I0424 09:06:08.495202 1 recorder.go:75] Recording config/clusteroperator/openshift-apiserver with fingerprint=7796280a8e745e190e07ab70865bc1d021dfd6b7ee6bff404db79d8696423b58 I0424 09:06:08.495211 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftapiserver/cluster with fingerprint=e712e6cf27339b441e4ed1f4cde91dbde7e952698ba93407e4457db63a4a4c76 I0424 09:06:08.495226 1 recorder.go:75] Recording config/clusteroperator/openshift-controller-manager with fingerprint=05bd5376cfceabc30dc9c91b5deb4b7ba206a6e444214247e5a51daa4bba53f1 I0424 09:06:08.495235 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftcontrollermanager/cluster with fingerprint=d71a0f4672f9b45d9fc8293bf1687afc650fd28d32e2e30de27523fe7b4eadf7 I0424 09:06:08.495249 1 recorder.go:75] Recording config/clusteroperator/openshift-samples with fingerprint=91776f6ae73ba8084178d3c229a70bed3e46dbe521948f100505139568ef2175 I0424 09:06:08.495263 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager with fingerprint=7c320a91b9a546824e0deb6f51b70e8c1a46645039b6af5f2512596a193524db I0424 09:06:08.495279 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-catalog with fingerprint=6c3ee61578cd09fbef23f54ac13730b4e1079f7a662d5f4b456285db0e8db9a9 I0424 09:06:08.495294 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-packageserver with fingerprint=e888f5daa5348ed212f825be31fd22ef77f0732f80898b901ba8c09013650686 I0424 09:06:08.495307 1 recorder.go:75] Recording config/clusteroperator/service-ca with fingerprint=79a3b2eaf501fddf741c34782deeb84da69ecd68b42a68da071b2315c775df2e I0424 09:06:08.495330 1 recorder.go:75] Recording config/clusteroperator/storage with fingerprint=8aa35147b2d0e5fef7824143c93c071116014c41baf37e69123166855238e195 I0424 09:06:08.495346 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/clustercsidriver/ebs.csi.aws.com with fingerprint=510064d6f6bcced87ab5bd2ddaff3d0edd7f93f4a4f7af2641f29fc53ffab21e I0424 09:06:08.495354 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/storage/cluster with fingerprint=8e480f8c1ce1b39baac42d8ec780c57c2592929ae0c801b61ffad49ba13f33ad I0424 09:06:08.495360 1 gather.go:177] gatherer "clusterconfig" function "operators" took 1.86430902s to process 35 records W0424 09:06:08.660970 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0424 09:06:08.677672 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0424 09:06:08.677685 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-vmsjh pod in namespace openshift-dns (previous: false). I0424 09:06:08.877153 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0424 09:06:08.877168 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-6474dd756f-2ngq2 pod in namespace openshift-image-registry (previous: false). I0424 09:06:09.077018 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-6474dd756f-2ngq2 pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-6474dd756f-2ngq2\" is waiting to start: ContainerCreating" I0424 09:06:09.077033 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-6474dd756f-2ngq2\" is waiting to start: ContainerCreating" I0424 09:06:09.077042 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-7947f68d9d-kflcb pod in namespace openshift-image-registry (previous: false). I0424 09:06:09.280377 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-7947f68d9d-kflcb pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-7947f68d9d-kflcb\" is waiting to start: ContainerCreating" I0424 09:06:09.280407 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-7947f68d9d-kflcb\" is waiting to start: ContainerCreating" I0424 09:06:09.280418 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-7947f68d9d-r7g8p pod in namespace openshift-image-registry (previous: false). I0424 09:06:09.477473 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-7947f68d9d-r7g8p pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-7947f68d9d-r7g8p\" is waiting to start: ContainerCreating" I0424 09:06:09.477490 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-7947f68d9d-r7g8p\" is waiting to start: ContainerCreating" I0424 09:06:09.477500 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-blflt pod in namespace openshift-image-registry (previous: false). W0424 09:06:09.673878 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0424 09:06:09.681485 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0424 09:06:09.681505 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-bw4s9 pod in namespace openshift-image-registry (previous: false). I0424 09:06:09.877126 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0424 09:06:09.877142 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-g2lbw pod in namespace openshift-image-registry (previous: false). I0424 09:06:10.076659 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0424 09:06:10.076675 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-5bc69d7bc7-gbb6q pod in namespace openshift-ingress (previous: false). I0424 09:06:10.277244 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-5bc69d7bc7-gbb6q pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-5bc69d7bc7-gbb6q\" is waiting to start: ContainerCreating" I0424 09:06:10.277259 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-5bc69d7bc7-gbb6q\" is waiting to start: ContainerCreating" I0424 09:06:10.277268 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-c58dd8557-8ffb2 pod in namespace openshift-ingress (previous: false). I0424 09:06:10.477015 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-c58dd8557-8ffb2 pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-c58dd8557-8ffb2\" is waiting to start: ContainerCreating" I0424 09:06:10.477031 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-c58dd8557-8ffb2\" is waiting to start: ContainerCreating" I0424 09:06:10.477043 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-c58dd8557-wdg7v pod in namespace openshift-ingress (previous: false). W0424 09:06:10.660607 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0424 09:06:10.677018 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-c58dd8557-wdg7v pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-c58dd8557-wdg7v\" is waiting to start: ContainerCreating" I0424 09:06:10.677037 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-c58dd8557-wdg7v\" is waiting to start: ContainerCreating" I0424 09:06:10.677050 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-72zsn pod in namespace openshift-ingress-canary (previous: false). I0424 09:06:10.876431 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-72zsn pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-72zsn\" is waiting to start: ContainerCreating" I0424 09:06:10.876447 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-72zsn\" is waiting to start: ContainerCreating" I0424 09:06:10.876458 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-9ddc7 pod in namespace openshift-ingress-canary (previous: false). I0424 09:06:11.078521 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-9ddc7 pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-9ddc7\" is waiting to start: ContainerCreating" I0424 09:06:11.078537 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-9ddc7\" is waiting to start: ContainerCreating" I0424 09:06:11.078547 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-qhgb7 pod in namespace openshift-ingress-canary (previous: false). I0424 09:06:11.278022 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-qhgb7 pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-qhgb7\" is waiting to start: ContainerCreating" I0424 09:06:11.278040 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-qhgb7\" is waiting to start: ContainerCreating" I0424 09:06:11.278055 1 tasks_processing.go:74] worker 9 stopped. I0424 09:06:11.278154 1 recorder.go:75] Recording events/openshift-dns with fingerprint=1d0fabc96f4d38159b20d413948c54cf79ac1c1fe82a176ad000ae5cf35b6f82 I0424 09:06:11.278269 1 recorder.go:75] Recording events/openshift-image-registry with fingerprint=b82c7be83a013fcc1ed96119d3643bc7ef7c62c4cdf43a1b208ab779a07f5ef4 I0424 09:06:11.278300 1 recorder.go:75] Recording events/openshift-ingress-operator with fingerprint=62f01320497ea2568acb61ee4ce14ae29158f570882924def4e6afda7debbb4b I0424 09:06:11.278355 1 recorder.go:75] Recording events/openshift-ingress with fingerprint=1a28814ad5dd2efe79c94b2e8b2e50ac3f2d8b8e62c595ccc2bcd49b0f281b78 I0424 09:06:11.278375 1 recorder.go:75] Recording events/openshift-ingress-canary with fingerprint=c09c17b7d6f9e703e1e0ce12558337be89a345179dc0c842bbd493d753b90c7d I0424 09:06:11.278383 1 gather.go:177] gatherer "clusterconfig" function "operators_pods_and_events" took 4.646518199s to process 5 records W0424 09:06:11.661297 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. W0424 09:06:11.661327 1 gather_dvo_metrics.go:117] Unable to read metrics from endpoint "http://deployment-validation-operator-metrics.openshift-deployment-validation-operator.svc:8383": DVO metrics service was not available within the 5s timeout: context deadline exceeded I0424 09:06:11.661346 1 tasks_processing.go:74] worker 53 stopped. E0424 09:06:11.661357 1 gather.go:140] gatherer "clusterconfig" function "dvo_metrics" failed with the error: DVO metrics service was not available within the 5s timeout: context deadline exceeded I0424 09:06:11.661371 1 recorder.go:75] Recording config/dvo_metrics with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 W0424 09:06:11.661406 1 gather.go:155] issue recording gatherer "clusterconfig" function "dvo_metrics" result "config/dvo_metrics" because of the warning: warning: the record with the same fingerprint "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" was already recorded at path "config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt", recording another one with a different path "config/dvo_metrics" I0424 09:06:11.661421 1 gather.go:177] gatherer "clusterconfig" function "dvo_metrics" took 5.031134762s to process 1 records I0424 09:06:19.070781 1 tasks_processing.go:74] worker 63 stopped. I0424 09:06:19.070816 1 recorder.go:75] Recording config/installplans with fingerprint=7b887df561a3a9e6ef0dc672845aa5d56e348505006b7496d3a2f83892b0c95b I0424 09:06:19.070827 1 gather.go:177] gatherer "clusterconfig" function "install_plans" took 12.438923792s to process 1 records I0424 09:06:19.838290 1 tasks_processing.go:74] worker 33 stopped. I0424 09:06:19.838561 1 recorder.go:75] Recording config/serviceaccounts with fingerprint=6f02bbed9776fbc53f719d04062abe458034f9ac95b2963e439b1381b6c8e3c4 I0424 09:06:19.838576 1 gather.go:177] gatherer "clusterconfig" function "service_accounts" took 13.206303827s to process 1 records E0424 09:06:19.838632 1 periodic.go:247] "Unhandled Error" err="clusterconfig failed after 13.208s with: function \"support_secret\" failed with an error, function \"machines\" failed with an error, function \"machine_healthchecks\" failed with an error, function \"pod_network_connectivity_checks\" failed with an error, function \"config_maps\" failed with an error, function \"ingress_certificates\" failed with an error, function \"dvo_metrics\" failed with an error" I0424 09:06:19.838837 1 configmapobserver.go:84] configmaps "insights-config" not found I0424 09:06:19.839882 1 controllerstatus.go:89] name=periodic-clusterconfig healthy=false reason=PeriodicGatherFailed message=Source clusterconfig could not be retrieved: function "support_secret" failed with an error, function "machines" failed with an error, function "machine_healthchecks" failed with an error, function "pod_network_connectivity_checks" failed with an error, function "config_maps" failed with an error, function "ingress_certificates" failed with an error, function "dvo_metrics" failed with an error I0424 09:06:19.839894 1 periodic.go:209] Running workloads gatherer I0424 09:06:19.839910 1 tasks_processing.go:45] number of workers: 2 I0424 09:06:19.839915 1 tasks_processing.go:69] worker 1 listening for tasks. I0424 09:06:19.839919 1 tasks_processing.go:71] worker 1 working on helmchart_info task. I0424 09:06:19.839931 1 tasks_processing.go:69] worker 0 listening for tasks. I0424 09:06:19.840000 1 tasks_processing.go:71] worker 0 working on workload_info task. I0424 09:06:19.862788 1 gather_workloads_info.go:278] Loaded pods in 0s, will wait 22s for image data I0424 09:06:19.866068 1 tasks_processing.go:74] worker 1 stopped. I0424 09:06:19.866083 1 gather.go:177] gatherer "workloads" function "helmchart_info" took 26.137475ms to process 0 records I0424 09:06:19.873971 1 gather_workloads_info.go:387] No image sha256:2904a78e2eb73fd6a9bb94c105c2a056831fb4113fbb7b0607c50adc9d879c9b (12ms) I0424 09:06:19.881905 1 gather_workloads_info.go:387] No image sha256:04c87c054a3f366a7dfbe0a93ebb0c80a098ee16842c5794b67c1202eec61996 (8ms) I0424 09:06:19.889581 1 gather_workloads_info.go:387] No image sha256:765f0d23b637f685f98a31bd47c131b03cf72a40761a3f9a9d6320faa3c33733 (8ms) I0424 09:06:19.897295 1 gather_workloads_info.go:387] No image sha256:7adc1eab05d6724c76ba751f6df816b08d6e70b78dee9eb94fa6fd9690542c98 (8ms) I0424 09:06:19.905290 1 gather_workloads_info.go:387] No image sha256:36b9e89c3cfcf1ab9ae500486e38afb6862cba48cb0b4d84a09508ab8f3d299f (8ms) I0424 09:06:19.912739 1 gather_workloads_info.go:387] No image sha256:a498046d64605bcccee2440aa4f04a4602baaae263cf01d977ec5208e876b1fd (7ms) I0424 09:06:19.920114 1 gather_workloads_info.go:387] No image sha256:ca1344cb64140188b7cae7bbc51fb751566c0b0c97d5e39b5850e628032c4a5e (7ms) I0424 09:06:19.927600 1 gather_workloads_info.go:387] No image sha256:91828234f107c068c8a4966d08370ae7b73e637651dbc6d92c18c4553402c22c (7ms) I0424 09:06:19.936091 1 gather_workloads_info.go:387] No image sha256:56a85660a445eced5c79a595a0eccf590087c5672d50f49d4c25ad52f9a44f04 (8ms) I0424 09:06:19.943519 1 gather_workloads_info.go:387] No image sha256:5f0b67cfbbc381243fb91ccc17345b56d05f4d717c667e8c644e5bf05633ba71 (7ms) I0424 09:06:19.970557 1 gather_workloads_info.go:387] No image sha256:289816958633a763a72dbc44e1dad40466223164e7e253039514f0d974ea5d21 (27ms) I0424 09:06:20.071465 1 gather_workloads_info.go:387] No image sha256:03cf4cd7ef1518610c6c7b3ad27d1622d82e98e3dc6e3f8e5d0fceb5c8d3786e (101ms) I0424 09:06:20.170864 1 gather_workloads_info.go:387] No image sha256:c15ca0c0ad60fe8757c2d5d1723fcdd7a1ed6c0251a90d22a7e6cae6811d01aa (99ms) I0424 09:06:20.271418 1 gather_workloads_info.go:387] No image sha256:47154813651033d59751fb655a384dbffb64dd26f10bd7f3be0c3128d0486356 (101ms) I0424 09:06:20.371136 1 gather_workloads_info.go:387] No image sha256:a258c226562adb14e3a163a1940938526ee6a0928982a7667d85d9a7334ce639 (100ms) I0424 09:06:20.471128 1 gather_workloads_info.go:387] No image sha256:4556896f77307821531ef91b7b7faccb82b824ea695693b2989f597f0deca038 (100ms) I0424 09:06:20.571212 1 gather_workloads_info.go:387] No image sha256:2598489729a4b258e4ecda4a06f6875133f2a10ced5c5241f8a57a8a05418e36 (100ms) I0424 09:06:20.671373 1 gather_workloads_info.go:387] No image sha256:5a95c19d82767e0235b4edb4a0536482c816904897aae1dc3eb255cb52b87a9f (100ms) I0424 09:06:20.771163 1 gather_workloads_info.go:387] No image sha256:695cf2f0cc07683c2a3ce1eaf3e56fe18abc6e2bac716f7d9843f5d173b9df52 (100ms) I0424 09:06:20.871081 1 gather_workloads_info.go:387] No image sha256:2e564f336c77116053f34d4201d364d8da04e789cfffa0ea422574c95f2d6404 (100ms) I0424 09:06:20.971101 1 gather_workloads_info.go:387] No image sha256:7b31223098f08328f5ddea8e5b871dbbd5f5a61ec550e8956f66793c0c6031a9 (100ms) I0424 09:06:21.071547 1 gather_workloads_info.go:387] No image sha256:521712486e2c6e3c020dad6a1cb340db8e55665b69f7c208fab9cd9e965fd588 (100ms) I0424 09:06:21.171561 1 gather_workloads_info.go:387] No image sha256:1a2532940843248c57d52141185dd71fbc393ab28b65d48f682038632c1dbbad (100ms) I0424 09:06:21.271746 1 gather_workloads_info.go:387] No image sha256:2e57e192c3c1240fd935dcd55c8fde5e70e78bf81d6176c96edf21fafe59f8ba (100ms) I0424 09:06:21.372076 1 gather_workloads_info.go:387] No image sha256:a0105d1eb62cf6ac9e5e2ef28d3e89bf6dc514bc594fc7090fe5a5ee18a09c87 (100ms) I0424 09:06:21.471732 1 gather_workloads_info.go:387] No image sha256:653c666f842c13e0baae2e89a9b1efe0e2ef56f621ffb5b32005115d2a26ab8c (100ms) I0424 09:06:21.574408 1 gather_workloads_info.go:387] No image sha256:c940ea87e7d133d75ba0002ef00c0806825eed3db8094cdb260d1bac18127733 (103ms) I0424 09:06:21.671115 1 gather_workloads_info.go:387] No image sha256:0a99240166165eb5718e7516a43282fe32df9c7c5e809b31b58abe44e42ff94d (97ms) I0424 09:06:21.761270 1 configmapobserver.go:84] configmaps "insights-config" not found I0424 09:06:21.770612 1 gather_workloads_info.go:387] No image sha256:b3909bf664c77097f75b3768830863d642eed3815dab2bfb4415c771ca2d5007 (99ms) I0424 09:06:21.871466 1 gather_workloads_info.go:387] No image sha256:e84cb128d930bd1ab867cc89b7b7bf2b2c0e41105ab93b5381069945b3ee9c57 (101ms) I0424 09:06:21.962254 1 configmapobserver.go:84] configmaps "insights-config" not found I0424 09:06:21.970243 1 gather_workloads_info.go:387] No image sha256:a56211d075aa43cbb491f669a5b2e46ee023dc95b7d51dbac28f463948c5ad61 (99ms) I0424 09:06:22.071295 1 gather_workloads_info.go:387] No image sha256:943018739e3db1763c3184b460dbc409e058abbac76d57b9927faad317be85e4 (101ms) I0424 09:06:22.071321 1 tasks_processing.go:74] worker 0 stopped. E0424 09:06:22.071330 1 gather.go:140] gatherer "workloads" function "workload_info" failed with the error: no running pods found for the insights-runtime-extractor statefulset I0424 09:06:22.071636 1 recorder.go:75] Recording config/workload_info with fingerprint=0340a73ca39b428df2cf896b5572b82173f823f73dafa62d0deb0e5008b091f5 I0424 09:06:22.071650 1 gather.go:177] gatherer "workloads" function "workload_info" took 2.231306522s to process 1 records E0424 09:06:22.071674 1 periodic.go:247] "Unhandled Error" err="workloads failed after 2.231s with: function \"workload_info\" failed with an error" I0424 09:06:22.072774 1 controllerstatus.go:89] name=periodic-workloads healthy=false reason=PeriodicGatherFailed message=Source workloads could not be retrieved: function "workload_info" failed with an error I0424 09:06:22.072787 1 periodic.go:209] Running conditional gatherer I0424 09:06:22.079004 1 requests.go:294] Making HTTP GET request at: https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules I0424 09:06:22.086729 1 conditional_gatherer.go:107] Get "https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.129.0.10:55096->172.30.0.10:53: read: connection refused E0424 09:06:22.086973 1 conditional_gatherer.go:322] unable to update alerts cache: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0424 09:06:22.087032 1 conditional_gatherer.go:384] updating version cache for conditional gatherer I0424 09:06:22.093222 1 conditional_gatherer.go:392] cluster version is '4.20.8' E0424 09:06:22.093237 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0424 09:06:22.093244 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0424 09:06:22.093248 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0424 09:06:22.093250 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0424 09:06:22.093253 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0424 09:06:22.093256 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0424 09:06:22.093259 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0424 09:06:22.093262 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0424 09:06:22.093265 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing I0424 09:06:22.093278 1 tasks_processing.go:45] number of workers: 3 I0424 09:06:22.093289 1 tasks_processing.go:69] worker 2 listening for tasks. I0424 09:06:22.093293 1 tasks_processing.go:71] worker 2 working on remote_configuration task. I0424 09:06:22.093299 1 tasks_processing.go:69] worker 0 listening for tasks. I0424 09:06:22.093311 1 tasks_processing.go:69] worker 1 listening for tasks. I0424 09:06:22.093322 1 tasks_processing.go:74] worker 1 stopped. I0424 09:06:22.093315 1 tasks_processing.go:71] worker 0 working on rapid_container_logs task. I0424 09:06:22.093327 1 tasks_processing.go:71] worker 2 working on conditional_gatherer_rules task. I0424 09:06:22.093344 1 recorder.go:75] Recording insights-operator/remote-configuration with fingerprint=0394430c431eec4d48bb1811a90918e95161d2282c59af26f2473613cc0959db I0424 09:06:22.093355 1 gather.go:177] gatherer "conditional" function "remote_configuration" took 1.03µs to process 1 records I0424 09:06:22.093414 1 tasks_processing.go:74] worker 2 stopped. I0424 09:06:22.093427 1 recorder.go:75] Recording insights-operator/conditional-gatherer-rules with fingerprint=7034af97e7e41c22e4b775abdd4b9066c8ebb19da33eb7f69f39bfd2eb5f6406 I0424 09:06:22.093433 1 gather.go:177] gatherer "conditional" function "conditional_gatherer_rules" took 1.25µs to process 1 records I0424 09:06:22.093574 1 tasks_processing.go:74] worker 0 stopped. I0424 09:06:22.093586 1 gather.go:177] gatherer "conditional" function "rapid_container_logs" took 240.036µs to process 0 records I0424 09:06:22.093607 1 controllerstatus.go:89] name=periodic-conditional healthy=false reason=NotAvailable message=Get "https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.129.0.10:55096->172.30.0.10:53: read: connection refused I0424 09:06:22.093624 1 recorder.go:75] Recording insights-operator/remote-configuration.json with fingerprint=359de9c990c741675cec72fda96b5c3682221efdb4799f5eaa6e9805bcd3b5c1 I0424 09:06:22.116687 1 recorder.go:75] Recording insights-operator/gathers with fingerprint=2cd1d132e997684bef7233153f45cd6e6c5b3a64d20a5cfa07f443645f221da4 I0424 09:06:22.116804 1 diskrecorder.go:70] Writing 99 records to /var/lib/insights-operator/insights-2026-04-24-090622.tar.gz I0424 09:06:22.122830 1 diskrecorder.go:51] Wrote 99 records to disk in 6ms I0424 09:06:22.122856 1 periodic.go:278] Gathering cluster info every 2h0m0s I0424 09:06:22.122872 1 periodic.go:279] Configuration is dataReporting: interval: 2h0m0s, uploadEndpoint: https://console.redhat.com/api/ingress/v1/upload, storagePath: /var/lib/insights-operator, downloadEndpoint: https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports, conditionalGathererEndpoint: https://console.redhat.com/api/gathering/v2/%s/gathering_rules, obfuscation: [] sca: disabled: false, endpoint: https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates, interval: 8h0m0s alerting: disabled: false clusterTransfer: endpoint: https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/, interval: 12h0m0s proxy: httpProxy: , httpsProxy: , noProxy: I0424 09:06:32.040385 1 configmapobserver.go:84] configmaps "insights-config" not found I0424 09:07:15.965851 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.crt" has been created (hash="c332f7e818188c15e07785e67aedee0cf12602442cdc6a55043eeeafe57f8ea1") W0424 09:07:15.965888 1 builder.go:160] Restart triggered because of file /var/run/secrets/serving-cert/tls.crt was created I0424 09:07:15.965946 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.key" has been created (hash="974ea0bda3e4debd89f4d7b6b90b7114dc655689f9ae32f224b3abad379a25d8") I0424 09:07:15.965955 1 base_controller.go:181] Shutting down LoggingSyncer ... I0424 09:07:15.965954 1 genericapiserver.go:548] "[graceful-termination] shutdown event" name="ShutdownInitiated" I0424 09:07:15.965989 1 genericapiserver.go:693] "[graceful-termination] pre-shutdown hooks completed" name="PreShutdownHooksStopped" I0424 09:07:15.965993 1 base_controller.go:123] Shutting down worker of LoggingSyncer controller ... I0424 09:07:15.965995 1 periodic.go:170] Shutting down I0424 09:07:15.966007 1 base_controller.go:113] All LoggingSyncer workers have been terminated