W0428 11:25:37.553395 1 cmd.go:257] Using insecure, self-signed certificates I0428 11:25:38.348572 1 start.go:138] Unable to read service ca bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0428 11:25:38.348886 1 observer_polling.go:159] Starting file observer I0428 11:25:38.791853 1 operator.go:60] Starting insights-operator v0.0.0-master+$Format:%H$ I0428 11:25:38.792038 1 legacy_config.go:327] Current config: {"report":false,"storagePath":"/var/lib/insights-operator","interval":"2h","endpoint":"https://console.redhat.com/api/ingress/v1/upload","conditionalGathererEndpoint":"https://console.redhat.com/api/gathering/v2/%s/gathering_rules","pull_report":{"endpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports","delay":"60s","timeout":"3000s","min_retry":"30s"},"impersonate":"system:serviceaccount:openshift-insights:gather","enableGlobalObfuscation":false,"ocm":{"scaEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates","scaInterval":"8h","scaDisabled":false,"clusterTransferEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/","clusterTransferInterval":"12h"},"disableInsightsAlerts":false,"processingStatusEndpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status","reportEndpointTechPreview":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/report"} I0428 11:25:38.792585 1 secure_serving.go:57] Forcing use of http/1.1 only W0428 11:25:38.792606 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. W0428 11:25:38.792611 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. W0428 11:25:38.792616 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. W0428 11:25:38.792619 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. W0428 11:25:38.792621 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. W0428 11:25:38.792624 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. I0428 11:25:38.792702 1 simple_featuregate_reader.go:171] Starting feature-gate-detector I0428 11:25:38.796391 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-insights", Name:"insights-operator", UID:"00cf46e7-7e1f-44a0-9de7-78f99e24ad01", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BuildCSIVolumes", "CPMSMachineNamePrefix", "ConsolePluginContentSecurityPolicy", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "ImageVolume", "IngressControllerLBSubnetsAWS", "KMSv1", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "MetricsCollectionProfiles", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "PinnedImages", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SetEIPForNLBIngressController", "SigstoreImageVerification", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiDisk", "VSphereMultiNetworks"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSServiceLBNetworkSecurityGroup", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureMultiDisk", "BootImageSkewEnforcement", "BootcNodeManagement", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "Example2", "ExternalSnapshotMetadata", "GCPClusterHostedDNS", "GCPClusterHostedDNSInstall", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "ImageModeStatusReporting", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryptionProvider", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "ManagedBootImagesAzure", "ManagedBootImagesvSphere", "MaxUnavailableStatefulSet", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "MultiDiskSetup", "MutatingAdmissionPolicy", "NewOLM", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterOperations", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PreconfiguredUDNAddresses", "SELinuxMount", "ShortCertRotation", "SignatureStores", "SigstoreImageVerificationPKI", "TranslateStreamCloseWebsocketRequests", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VolumeAttributesClass", "VolumeGroupSnapshot"}} I0428 11:25:38.796378 1 operator.go:125] FeatureGates initialized: knownFeatureGates=[AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BuildCSIVolumes CPMSMachineNamePrefix ConsolePluginContentSecurityPolicy ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GatewayAPI GatewayAPIController HighlyAvailableArbiter ImageVolume IngressControllerLBSubnetsAWS KMSv1 MachineConfigNodes ManagedBootImages ManagedBootImagesAWS MetricsCollectionProfiles NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation PinnedImages ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SetEIPForNLBIngressController SigstoreImageVerification StoragePerformantSecurityPolicy UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiDisk VSphereMultiNetworks AWSClusterHostedDNS AWSClusterHostedDNSInstall AWSDedicatedHosts AWSServiceLBNetworkSecurityGroup AutomatedEtcdBackup AzureClusterHostedDNSInstall AzureDedicatedHosts AzureMultiDisk BootImageSkewEnforcement BootcNodeManagement ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterMonitoringConfig ClusterVersionOperatorConfiguration DNSNameResolver DualReplica DyanmicServiceEndpointIBMCloud DynamicResourceAllocation EtcdBackendQuota EventedPLEG Example Example2 ExternalSnapshotMetadata GCPClusterHostedDNS GCPClusterHostedDNSInstall GCPCustomAPIEndpoints GCPCustomAPIEndpointsInstall ImageModeStatusReporting ImageStreamImportMode IngressControllerDynamicConfigurationManager InsightsConfig InsightsConfigAPI InsightsOnDemandDataGather IrreconcilableMachineConfig KMSEncryptionProvider MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController ManagedBootImagesAzure ManagedBootImagesvSphere MaxUnavailableStatefulSet MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure MultiDiskSetup MutatingAdmissionPolicy NewOLM NewOLMCatalogdAPIV1Metas NewOLMOwnSingleNamespace NewOLMPreflightPermissionChecks NewOLMWebhookProviderOpenshiftServiceCA NoRegistryClusterOperations NodeSwap NutanixMultiSubnets OVNObservability OpenShiftPodSecurityAdmission PreconfiguredUDNAddresses SELinuxMount ShortCertRotation SignatureStores SigstoreImageVerificationPKI TranslateStreamCloseWebsocketRequests VSphereConfigurableMaxAllowedBlockVolumesPerNode VSphereHostVMGroupZonal VSphereMixedNodeEnv VolumeAttributesClass VolumeGroupSnapshot] I0428 11:25:38.796803 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController I0428 11:25:38.796822 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController I0428 11:25:38.796819 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0428 11:25:38.796819 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0428 11:25:38.796836 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0428 11:25:38.796842 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0428 11:25:38.797026 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/tmp/serving-cert-3840697712/tls.crt::/tmp/serving-cert-3840697712/tls.key" I0428 11:25:38.797290 1 secure_serving.go:213] Serving securely on [::]:8443 I0428 11:25:38.797319 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" W0428 11:25:38.804210 1 configmapobserver.go:64] Cannot get the configuration config map: configmaps "insights-config" not found. Default configuration is used. I0428 11:25:38.804242 1 secretconfigobserver.go:216] Legacy configuration set: enabled=false endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=false reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0428 11:25:38.804347 1 base_controller.go:76] Waiting for caches to sync for ConfigController I0428 11:25:38.809541 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0428 11:25:38.809556 1 secretconfigobserver.go:204] Legacy configuration updated: enabled=true endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=true reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0428 11:25:38.813461 1 secretconfigobserver.go:119] support secret does not exist I0428 11:25:38.817513 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0428 11:25:38.821294 1 secretconfigobserver.go:119] support secret does not exist I0428 11:25:38.824006 1 recorder.go:161] Pruning old reports every 6h50m21s, max age is 288h0m0s I0428 11:25:38.828895 1 controllerstatus.go:80] name=insightsuploader healthy=true reason= message= I0428 11:25:38.828909 1 insightsuploader.go:86] Reporting status periodically to https://console.redhat.com/api/ingress/v1/upload every 2h0m0s, starting in 1m30s I0428 11:25:38.828915 1 controllerstatus.go:80] name=insightsreport healthy=true reason= message= I0428 11:25:38.828925 1 insightsreport.go:296] Starting report retriever I0428 11:25:38.828931 1 insightsreport.go:298] Insights analysis reports will be downloaded from the https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports endpoint with a delay of 1m0s I0428 11:25:38.828940 1 periodic.go:209] Running clusterconfig gatherer I0428 11:25:38.828982 1 tasks_processing.go:45] number of workers: 64 I0428 11:25:38.829009 1 tasks_processing.go:69] worker 3 listening for tasks. I0428 11:25:38.829018 1 tasks_processing.go:69] worker 12 listening for tasks. I0428 11:25:38.829021 1 tasks_processing.go:69] worker 0 listening for tasks. I0428 11:25:38.829026 1 tasks_processing.go:69] worker 8 listening for tasks. I0428 11:25:38.829029 1 tasks_processing.go:69] worker 1 listening for tasks. I0428 11:25:38.829032 1 tasks_processing.go:69] worker 9 listening for tasks. I0428 11:25:38.829038 1 tasks_processing.go:69] worker 10 listening for tasks. I0428 11:25:38.829039 1 tasks_processing.go:69] worker 2 listening for tasks. I0428 11:25:38.829037 1 tasks_processing.go:69] worker 7 listening for tasks. I0428 11:25:38.829043 1 tasks_processing.go:69] worker 11 listening for tasks. I0428 11:25:38.829051 1 tasks_processing.go:69] worker 6 listening for tasks. I0428 11:25:38.829050 1 tasks_processing.go:69] worker 5 listening for tasks. I0428 11:25:38.829056 1 tasks_processing.go:69] worker 4 listening for tasks. I0428 11:25:38.829052 1 tasks_processing.go:69] worker 29 listening for tasks. I0428 11:25:38.829065 1 tasks_processing.go:69] worker 25 listening for tasks. I0428 11:25:38.829070 1 tasks_processing.go:69] worker 21 listening for tasks. I0428 11:25:38.829074 1 tasks_processing.go:69] worker 18 listening for tasks. I0428 11:25:38.829077 1 tasks_processing.go:69] worker 26 listening for tasks. I0428 11:25:38.829082 1 tasks_processing.go:69] worker 23 listening for tasks. I0428 11:25:38.829082 1 tasks_processing.go:69] worker 15 listening for tasks. I0428 11:25:38.829083 1 tasks_processing.go:69] worker 17 listening for tasks. I0428 11:25:38.829077 1 tasks_processing.go:69] worker 14 listening for tasks. I0428 11:25:38.829084 1 tasks_processing.go:69] worker 20 listening for tasks. I0428 11:25:38.829095 1 tasks_processing.go:71] worker 14 working on active_alerts task. I0428 11:25:38.829097 1 tasks_processing.go:71] worker 1 working on monitoring_persistent_volumes task. I0428 11:25:38.829100 1 tasks_processing.go:69] worker 31 listening for tasks. I0428 11:25:38.829101 1 tasks_processing.go:71] worker 20 working on ceph_cluster task. I0428 11:25:38.829090 1 tasks_processing.go:69] worker 22 listening for tasks. I0428 11:25:38.829108 1 tasks_processing.go:69] worker 32 listening for tasks. I0428 11:25:38.829108 1 tasks_processing.go:69] worker 62 listening for tasks. I0428 11:25:38.829097 1 tasks_processing.go:69] worker 27 listening for tasks. I0428 11:25:38.829106 1 tasks_processing.go:69] worker 28 listening for tasks. I0428 11:25:38.829118 1 tasks_processing.go:71] worker 29 working on openshift_logging task. I0428 11:25:38.829118 1 tasks_processing.go:71] worker 6 working on feature_gates task. I0428 11:25:38.829125 1 tasks_processing.go:69] worker 63 listening for tasks. I0428 11:25:38.829128 1 tasks_processing.go:69] worker 35 listening for tasks. W0428 11:25:38.829126 1 gather_active_alerts.go:54] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0428 11:25:38.829133 1 tasks_processing.go:69] worker 19 listening for tasks. I0428 11:25:38.829134 1 tasks_processing.go:71] worker 3 working on nodenetworkconfigurationpolicies task. I0428 11:25:38.829138 1 tasks_processing.go:69] worker 36 listening for tasks. I0428 11:25:38.829132 1 tasks_processing.go:71] worker 0 working on overlapping_namespace_uids task. I0428 11:25:38.829142 1 tasks_processing.go:69] worker 45 listening for tasks. I0428 11:25:38.829143 1 gather.go:177] gatherer "clusterconfig" function "active_alerts" took 35.471µs to process 0 records I0428 11:25:38.829149 1 tasks_processing.go:69] worker 61 listening for tasks. I0428 11:25:38.829152 1 tasks_processing.go:69] worker 37 listening for tasks. I0428 11:25:38.829142 1 tasks_processing.go:69] worker 60 listening for tasks. I0428 11:25:38.829149 1 tasks_processing.go:71] worker 8 working on aggregated_monitoring_cr_names task. I0428 11:25:38.829158 1 tasks_processing.go:71] worker 17 working on container_runtime_configs task. I0428 11:25:38.829161 1 tasks_processing.go:69] worker 38 listening for tasks. I0428 11:25:38.829063 1 tasks_processing.go:69] worker 24 listening for tasks. I0428 11:25:38.829168 1 tasks_processing.go:69] worker 49 listening for tasks. I0428 11:25:38.829172 1 tasks_processing.go:71] worker 10 working on infrastructures task. I0428 11:25:38.829176 1 tasks_processing.go:69] worker 39 listening for tasks. I0428 11:25:38.829178 1 tasks_processing.go:69] worker 50 listening for tasks. I0428 11:25:38.829167 1 tasks_processing.go:71] worker 25 working on container_images task. I0428 11:25:38.829185 1 tasks_processing.go:69] worker 40 listening for tasks. I0428 11:25:38.829186 1 tasks_processing.go:69] worker 53 listening for tasks. I0428 11:25:38.829071 1 tasks_processing.go:69] worker 13 listening for tasks. I0428 11:25:38.829082 1 tasks_processing.go:71] worker 18 working on storage_cluster task. I0428 11:25:38.829190 1 tasks_processing.go:69] worker 52 listening for tasks. I0428 11:25:38.829089 1 tasks_processing.go:71] worker 23 working on networks task. I0428 11:25:38.829211 1 tasks_processing.go:69] worker 42 listening for tasks. I0428 11:25:38.829205 1 tasks_processing.go:71] worker 21 working on qemu_kubevirt_launcher_logs task. I0428 11:25:38.829214 1 tasks_processing.go:69] worker 51 listening for tasks. I0428 11:25:38.829238 1 tasks_processing.go:69] worker 41 listening for tasks. I0428 11:25:38.829090 1 tasks_processing.go:69] worker 55 listening for tasks. I0428 11:25:38.829104 1 tasks_processing.go:69] worker 56 listening for tasks. I0428 11:25:38.829115 1 tasks_processing.go:69] worker 33 listening for tasks. I0428 11:25:38.829251 1 tasks_processing.go:69] worker 43 listening for tasks. I0428 11:25:38.829114 1 tasks_processing.go:71] worker 4 working on operators task. I0428 11:25:38.829266 1 tasks_processing.go:69] worker 44 listening for tasks. I0428 11:25:38.829272 1 tasks_processing.go:71] worker 11 working on certificate_signing_requests task. I0428 11:25:38.829082 1 tasks_processing.go:71] worker 26 working on olm_operators task. I0428 11:25:38.829109 1 tasks_processing.go:71] worker 5 working on number_of_pods_and_netnamespaces_with_sdn_annotations task. I0428 11:25:38.829361 1 tasks_processing.go:71] worker 31 working on nodenetworkstates task. I0428 11:25:38.829365 1 tasks_processing.go:71] worker 45 working on validating_webhook_configurations task. I0428 11:25:38.829127 1 tasks_processing.go:69] worker 58 listening for tasks. I0428 11:25:38.829400 1 tasks_processing.go:71] worker 58 working on version task. I0428 11:25:38.829090 1 tasks_processing.go:69] worker 30 listening for tasks. I0428 11:25:38.829152 1 tasks_processing.go:69] worker 47 listening for tasks. I0428 11:25:38.829154 1 tasks_processing.go:71] worker 15 working on service_accounts task. I0428 11:25:38.829116 1 tasks_processing.go:71] worker 12 working on cluster_apiserver task. I0428 11:25:38.829119 1 tasks_processing.go:69] worker 34 listening for tasks. I0428 11:25:38.829167 1 tasks_processing.go:71] worker 9 working on machine_sets task. I0428 11:25:38.829135 1 tasks_processing.go:69] worker 59 listening for tasks. I0428 11:25:38.829257 1 tasks_processing.go:71] worker 7 working on pdbs task. I0428 11:25:38.829120 1 tasks_processing.go:69] worker 57 listening for tasks. I0428 11:25:38.829357 1 tasks_processing.go:71] worker 44 working on machine_healthchecks task. I0428 11:25:38.829163 1 tasks_processing.go:71] worker 2 working on clusterroles task. I0428 11:25:38.829194 1 tasks_processing.go:69] worker 54 listening for tasks. I0428 11:25:38.829636 1 tasks_processing.go:71] worker 28 working on authentication task. I0428 11:25:38.829357 1 tasks_processing.go:71] worker 36 working on openstack_controlplanes task. I0428 11:25:38.829146 1 tasks_processing.go:69] worker 46 listening for tasks. I0428 11:25:38.829067 1 tasks_processing.go:69] worker 16 listening for tasks. I0428 11:25:38.829674 1 tasks_processing.go:71] worker 22 working on ingress_certificates task. I0428 11:25:38.829678 1 tasks_processing.go:71] worker 62 working on lokistack task. I0428 11:25:38.829708 1 tasks_processing.go:71] worker 35 working on storage_classes task. I0428 11:25:38.829741 1 tasks_processing.go:71] worker 14 working on nodes task. I0428 11:25:38.829774 1 tasks_processing.go:71] worker 61 working on config_maps task. I0428 11:25:38.829160 1 tasks_processing.go:69] worker 48 listening for tasks. I0428 11:25:38.829869 1 tasks_processing.go:71] worker 19 working on sap_datahubs task. I0428 11:25:38.829897 1 tasks_processing.go:71] worker 27 working on openstack_dataplanenodesets task. I0428 11:25:38.829940 1 tasks_processing.go:71] worker 63 working on support_secret task. I0428 11:25:38.829972 1 tasks_processing.go:71] worker 48 working on silenced_alerts task. I0428 11:25:38.829979 1 tasks_processing.go:71] worker 37 working on image_pruners task. W0428 11:25:38.830003 1 gather_silenced_alerts.go:38] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0428 11:25:38.830016 1 tasks_processing.go:74] worker 48 stopped. I0428 11:25:38.830027 1 tasks_processing.go:71] worker 49 working on image_registries task. I0428 11:25:38.830043 1 tasks_processing.go:71] worker 47 working on metrics task. I0428 11:25:38.830055 1 tasks_processing.go:71] worker 38 working on pod_network_connectivity_checks task. W0428 11:25:38.830066 1 gather_most_recent_metrics.go:64] Unable to load metrics client, no metrics will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0428 11:25:38.830073 1 tasks_processing.go:71] worker 34 working on mutating_webhook_configurations task. I0428 11:25:38.830127 1 tasks_processing.go:71] worker 39 working on machine_configs task. I0428 11:25:38.830151 1 tasks_processing.go:71] worker 60 working on jaegers task. I0428 11:25:38.830172 1 tasks_processing.go:71] worker 32 working on openstack_version task. I0428 11:25:38.830185 1 tasks_processing.go:71] worker 50 working on sap_pods task. I0428 11:25:38.830201 1 tasks_processing.go:71] worker 40 working on proxies task. I0428 11:25:38.830350 1 tasks_processing.go:71] worker 41 working on oauths task. I0428 11:25:38.830370 1 tasks_processing.go:71] worker 52 working on sap_config task. I0428 11:25:38.830379 1 tasks_processing.go:71] worker 42 working on openstack_dataplanedeployments task. I0428 11:25:38.830393 1 tasks_processing.go:71] worker 24 working on tsdb_status task. W0428 11:25:38.830410 1 gather_prometheus_tsdb_status.go:38] Unable to load metrics client, tsdb status cannot be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0428 11:25:38.830026 1 gather.go:177] gatherer "clusterconfig" function "silenced_alerts" took 30.129µs to process 0 records I0428 11:25:38.830432 1 gather.go:177] gatherer "clusterconfig" function "metrics" took 20.428µs to process 0 records I0428 11:25:38.830442 1 tasks_processing.go:71] worker 13 working on machine_config_pools task. I0428 11:25:38.830447 1 tasks_processing.go:74] worker 24 stopped. I0428 11:25:38.830032 1 tasks_processing.go:71] worker 30 working on install_plans task. I0428 11:25:38.830502 1 tasks_processing.go:74] worker 47 stopped. I0428 11:25:38.830513 1 tasks_processing.go:71] worker 56 working on operators_pods_and_events task. I0428 11:25:38.830442 1 gather.go:177] gatherer "clusterconfig" function "tsdb_status" took 23.056µs to process 0 records I0428 11:25:38.830524 1 tasks_processing.go:71] worker 51 working on schedulers task. I0428 11:25:38.830524 1 tasks_processing.go:71] worker 55 working on machine_autoscalers task. I0428 11:25:38.830623 1 tasks_processing.go:71] worker 54 working on dvo_metrics task. I0428 11:25:38.830671 1 tasks_processing.go:71] worker 33 working on ingress task. I0428 11:25:38.830676 1 tasks_processing.go:71] worker 16 working on image task. I0428 11:25:38.829975 1 tasks_processing.go:71] worker 53 working on cost_management_metrics_configs task. I0428 11:25:38.830681 1 tasks_processing.go:71] worker 46 working on machines task. I0428 11:25:38.830689 1 tasks_processing.go:71] worker 59 working on crds task. I0428 11:25:38.830688 1 tasks_processing.go:71] worker 43 working on node_logs task. I0428 11:25:38.830694 1 tasks_processing.go:71] worker 57 working on openshift_machine_api_events task. I0428 11:25:38.832408 1 tasks_processing.go:74] worker 29 stopped. I0428 11:25:38.832420 1 gather.go:177] gatherer "clusterconfig" function "openshift_logging" took 3.277194ms to process 0 records I0428 11:25:38.835306 1 tasks_processing.go:74] worker 20 stopped. I0428 11:25:38.835319 1 gather.go:177] gatherer "clusterconfig" function "ceph_cluster" took 6.195455ms to process 0 records I0428 11:25:38.835326 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkconfigurationpolicies" took 6.180981ms to process 0 records I0428 11:25:38.835334 1 tasks_processing.go:74] worker 3 stopped. I0428 11:25:38.835527 1 controller.go:128] Initializing last reported time to 0001-01-01T00:00:00Z I0428 11:25:38.835543 1 controller.go:212] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0428 11:25:38.835548 1 controller.go:212] Source periodic-conditional *controllerstatus.Simple is not ready I0428 11:25:38.835551 1 controller.go:212] Source periodic-workloads *controllerstatus.Simple is not ready I0428 11:25:38.835566 1 controller.go:489] The operator is still being initialized I0428 11:25:38.835574 1 controller.go:512] The operator is healthy I0428 11:25:38.835654 1 tasks_processing.go:74] worker 10 stopped. I0428 11:25:38.836208 1 recorder.go:75] Recording config/infrastructure with fingerprint=3664a187388fbf32474c0c2c7ea61f6fb252784d046567cc689c9240a4219b2e I0428 11:25:38.836240 1 gather.go:177] gatherer "clusterconfig" function "infrastructures" took 6.470555ms to process 1 records I0428 11:25:38.837826 1 tasks_processing.go:74] worker 31 stopped. I0428 11:25:38.837837 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkstates" took 8.456569ms to process 0 records I0428 11:25:38.837843 1 gather.go:177] gatherer "clusterconfig" function "container_runtime_configs" took 8.668529ms to process 0 records I0428 11:25:38.837848 1 gather.go:177] gatherer "clusterconfig" function "machine_sets" took 8.377976ms to process 0 records I0428 11:25:38.837852 1 tasks_processing.go:74] worker 9 stopped. I0428 11:25:38.837856 1 tasks_processing.go:74] worker 17 stopped. I0428 11:25:38.847709 1 tasks_processing.go:74] worker 18 stopped. I0428 11:25:38.847721 1 gather.go:177] gatherer "clusterconfig" function "storage_cluster" took 18.506917ms to process 0 records I0428 11:25:38.847742 1 tasks_processing.go:74] worker 19 stopped. I0428 11:25:38.847749 1 gather.go:177] gatherer "clusterconfig" function "sap_datahubs" took 17.856364ms to process 0 records E0428 11:25:38.847754 1 gather.go:140] gatherer "clusterconfig" function "pod_network_connectivity_checks" failed with the error: the server could not find the requested resource (get podnetworkconnectivitychecks.controlplane.operator.openshift.io) I0428 11:25:38.847760 1 gather.go:177] gatherer "clusterconfig" function "pod_network_connectivity_checks" took 17.690898ms to process 0 records I0428 11:25:38.847765 1 tasks_processing.go:74] worker 38 stopped. I0428 11:25:38.864582 1 tasks_processing.go:74] worker 27 stopped. I0428 11:25:38.864596 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanenodesets" took 34.671176ms to process 0 records I0428 11:25:38.864730 1 tasks_processing.go:74] worker 53 stopped. I0428 11:25:38.864743 1 gather.go:177] gatherer "clusterconfig" function "cost_management_metrics_configs" took 33.97866ms to process 0 records I0428 11:25:38.865436 1 tasks_processing.go:74] worker 36 stopped. I0428 11:25:38.865456 1 gather.go:177] gatherer "clusterconfig" function "openstack_controlplanes" took 35.783216ms to process 0 records I0428 11:25:38.865784 1 tasks_processing.go:74] worker 12 stopped. I0428 11:25:38.866339 1 recorder.go:75] Recording config/apiserver with fingerprint=203441e654cc8693724688745b576ea4eb50834f54ca3c6fdd4770947b9fd63e I0428 11:25:38.866359 1 gather.go:177] gatherer "clusterconfig" function "cluster_apiserver" took 36.343415ms to process 1 records I0428 11:25:38.866376 1 tasks_processing.go:74] worker 33 stopped. I0428 11:25:38.867067 1 recorder.go:75] Recording config/ingress with fingerprint=c2b493301f13c6708df7b69d80eef8d34f3635a5d59a95143a799a9f99e9dadd I0428 11:25:38.867088 1 gather.go:177] gatherer "clusterconfig" function "ingress" took 35.143436ms to process 1 records I0428 11:25:38.867277 1 tasks_processing.go:74] worker 50 stopped. I0428 11:25:38.867293 1 gather.go:177] gatherer "clusterconfig" function "sap_pods" took 37.080339ms to process 0 records I0428 11:25:38.867349 1 tasks_processing.go:74] worker 42 stopped. I0428 11:25:38.867367 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanedeployments" took 36.956847ms to process 0 records I0428 11:25:38.867389 1 tasks_processing.go:74] worker 62 stopped. I0428 11:25:38.867400 1 gather.go:177] gatherer "clusterconfig" function "lokistack" took 37.702965ms to process 0 records I0428 11:25:38.867490 1 tasks_processing.go:74] worker 16 stopped. I0428 11:25:38.867519 1 recorder.go:75] Recording config/image with fingerprint=9216e7911ea779b08417cf9d0365cc80f7b8857c9ae1e37ccd30a0e516c4687c I0428 11:25:38.867528 1 gather.go:177] gatherer "clusterconfig" function "image" took 36.650702ms to process 1 records I0428 11:25:38.867604 1 tasks_processing.go:74] worker 35 stopped. I0428 11:25:38.867608 1 recorder.go:75] Recording config/storage/storageclasses/gp2-csi with fingerprint=ac52f76e3b571c5bd55e9c36a5a405c413b70fab3b948ba98663a0cdba72d37a I0428 11:25:38.867651 1 recorder.go:75] Recording config/storage/storageclasses/gp3-csi with fingerprint=3ffc3afcbb8e1362d040f5bb3e53bb6298382b574ffba573cbe5b81c5268e2e2 I0428 11:25:38.867663 1 gather.go:177] gatherer "clusterconfig" function "storage_classes" took 37.698036ms to process 2 records I0428 11:25:38.867737 1 tasks_processing.go:74] worker 6 stopped. I0428 11:25:38.867793 1 recorder.go:75] Recording config/featuregate with fingerprint=7c33d55d7e84445fc1d1a05315d8a2e84a0ea3dfb1fa5fdde840ca081221cb08 I0428 11:25:38.867806 1 gather.go:177] gatherer "clusterconfig" function "feature_gates" took 38.409763ms to process 1 records I0428 11:25:38.867876 1 recorder.go:75] Recording config/schedulers/cluster with fingerprint=145e7f06c3f57bd24a726fd0c5a82113f27b454ae4b4b6ca2ad321a593dd2df3 I0428 11:25:38.867889 1 gather.go:177] gatherer "clusterconfig" function "schedulers" took 37.042536ms to process 1 records I0428 11:25:38.867904 1 tasks_processing.go:74] worker 51 stopped. I0428 11:25:38.867969 1 tasks_processing.go:74] worker 23 stopped. I0428 11:25:38.868013 1 recorder.go:75] Recording config/network with fingerprint=bcd4e04dfc64182088054b2e83f491edf118f75b887cb24ab443b965062ffaf3 I0428 11:25:38.868025 1 gather.go:177] gatherer "clusterconfig" function "networks" took 38.36476ms to process 1 records I0428 11:25:38.868106 1 tasks_processing.go:74] worker 28 stopped. I0428 11:25:38.868161 1 recorder.go:75] Recording config/authentication with fingerprint=f8a4a3d285a8fa6ffac4bf430ef1fa44d0286f48dfa59bd7a7225161cd30747f I0428 11:25:38.868169 1 gather.go:177] gatherer "clusterconfig" function "authentication" took 38.079692ms to process 1 records I0428 11:25:38.868318 1 tasks_processing.go:74] worker 7 stopped. I0428 11:25:38.868434 1 recorder.go:75] Recording config/pdbs/openshift-image-registry/image-registry with fingerprint=75e847107add39889636429ce81691468b66f12d5d8bb250554fb7c8078b16c9 I0428 11:25:38.868467 1 recorder.go:75] Recording config/pdbs/openshift-ingress/router-default with fingerprint=3d7ca8eb0bc4ce94a58a50f63118029690ed41d40aadd6379de8fd58aa399adf I0428 11:25:38.868490 1 recorder.go:75] Recording config/pdbs/openshift-operator-lifecycle-manager/packageserver-pdb with fingerprint=79933cf78f61ca32ee2c6eb9c78e77b1b480100b5d03f560c8e73ee39346a9e6 I0428 11:25:38.868505 1 gather.go:177] gatherer "clusterconfig" function "pdbs" took 38.791044ms to process 3 records I0428 11:25:38.868513 1 gather.go:177] gatherer "clusterconfig" function "openstack_version" took 38.269924ms to process 0 records I0428 11:25:38.868520 1 gather.go:177] gatherer "clusterconfig" function "openshift_machine_api_events" took 37.544652ms to process 0 records I0428 11:25:38.868525 1 gather.go:177] gatherer "clusterconfig" function "machine_config_pools" took 38.046611ms to process 0 records I0428 11:25:38.868525 1 tasks_processing.go:74] worker 32 stopped. I0428 11:25:38.868530 1 gather.go:177] gatherer "clusterconfig" function "machine_autoscalers" took 37.974066ms to process 0 records I0428 11:25:38.868530 1 tasks_processing.go:74] worker 57 stopped. I0428 11:25:38.868534 1 tasks_processing.go:74] worker 13 stopped. I0428 11:25:38.868540 1 tasks_processing.go:74] worker 55 stopped. I0428 11:25:38.868541 1 tasks_processing.go:74] worker 52 stopped. I0428 11:25:38.868536 1 gather.go:177] gatherer "clusterconfig" function "sap_config" took 38.129606ms to process 0 records E0428 11:25:38.868554 1 gather.go:140] gatherer "clusterconfig" function "machines" failed with the error: machines.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machines" in API group "machine.openshift.io" at the cluster scope I0428 11:25:38.868563 1 gather.go:177] gatherer "clusterconfig" function "machines" took 37.668629ms to process 0 records E0428 11:25:38.868570 1 gather.go:140] gatherer "clusterconfig" function "machine_healthchecks" failed with the error: machinehealthchecks.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machinehealthchecks" in API group "machine.openshift.io" at the cluster scope I0428 11:25:38.868578 1 gather.go:177] gatherer "clusterconfig" function "machine_healthchecks" took 38.959873ms to process 0 records I0428 11:25:38.868581 1 gather_logs.go:145] no pods in namespace were found I0428 11:25:38.868585 1 gather.go:177] gatherer "clusterconfig" function "jaegers" took 38.377263ms to process 0 records I0428 11:25:38.868594 1 tasks_processing.go:74] worker 46 stopped. I0428 11:25:38.868600 1 tasks_processing.go:74] worker 44 stopped. I0428 11:25:38.868590 1 tasks_processing.go:74] worker 60 stopped. I0428 11:25:38.868602 1 gather.go:177] gatherer "clusterconfig" function "qemu_kubevirt_launcher_logs" took 39.350037ms to process 0 records I0428 11:25:38.868611 1 tasks_processing.go:74] worker 21 stopped. I0428 11:25:38.868628 1 tasks_processing.go:74] worker 40 stopped. I0428 11:25:38.868713 1 recorder.go:75] Recording config/proxy with fingerprint=3a7a899b9ec4a0c042f375dd2769068449ffa274ce148de5605f1765e27b07ef I0428 11:25:38.868728 1 gather.go:177] gatherer "clusterconfig" function "proxies" took 38.412475ms to process 1 records I0428 11:25:38.868739 1 gather.go:177] gatherer "clusterconfig" function "monitoring_persistent_volumes" took 39.613288ms to process 0 records I0428 11:25:38.868746 1 tasks_processing.go:74] worker 1 stopped. I0428 11:25:38.868776 1 tasks_processing.go:74] worker 37 stopped. I0428 11:25:38.869051 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/imagepruner/cluster with fingerprint=09b321ff8408ae02892620ad5012407cdfeac028f24e7b63cb59f8a39da5563d I0428 11:25:38.869074 1 gather.go:177] gatherer "clusterconfig" function "image_pruners" took 38.780116ms to process 1 records I0428 11:25:38.869234 1 tasks_processing.go:74] worker 34 stopped. I0428 11:25:38.869812 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/aws-pod-identity with fingerprint=e4dd87733208720a535f60a71f6517d6e051de6d1b07dd8fd82f8fa1eb61bc76 I0428 11:25:38.869981 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-podimagespec-mutation with fingerprint=1f6d5a397c02c5b143f5b2a50139099ecc7ff3c4bfa375c6f24764ce7934402c I0428 11:25:38.870139 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-service-mutation with fingerprint=ec670194a3f17cca57838a6115cd098cb81178a8dce7b5caf76ecec473ed89ef I0428 11:25:38.870166 1 gather.go:177] gatherer "clusterconfig" function "mutating_webhook_configurations" took 38.695671ms to process 3 records I0428 11:25:38.870717 1 recorder.go:75] Recording config/oauth with fingerprint=ad8d26c43bbc67e09c967492fd0e1b486c0df03c89df91cba625e862ae72ac9b I0428 11:25:38.870735 1 gather.go:177] gatherer "clusterconfig" function "oauths" took 38.467516ms to process 1 records E0428 11:25:38.870797 1 gather.go:140] gatherer "clusterconfig" function "support_secret" failed with the error: secrets "support" not found I0428 11:25:38.870809 1 gather.go:177] gatherer "clusterconfig" function "support_secret" took 38.902987ms to process 0 records I0428 11:25:38.870998 1 tasks_processing.go:74] worker 41 stopped. I0428 11:25:38.871002 1 recorder.go:75] Recording config/validatingwebhookconfigurations/multus.openshift.io with fingerprint=c1b5bc1ed9940fe7b3e232d894ae2d4c6f51cd85ae7228ee1ba714d8dbe57411 I0428 11:25:38.871016 1 tasks_processing.go:74] worker 45 stopped. I0428 11:25:38.871020 1 tasks_processing.go:74] worker 63 stopped. I0428 11:25:38.871570 1 recorder.go:75] Recording config/validatingwebhookconfigurations/network-node-identity.openshift.io with fingerprint=9b6b75a5f1490293288e5f6c3a53655941e4886919ffafc455a0831766f83ca6 I0428 11:25:38.871612 1 recorder.go:75] Recording config/validatingwebhookconfigurations/performance-addon-operator with fingerprint=490e9139648878dff083326eafb0b4b0230bb899852bf854798139dfdaeb55af I0428 11:25:38.871699 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterrolebindings-validation with fingerprint=c4aa8dc6633ed0f0408170fa33c847a58c2a49dd0dfe4d597586f474d073919e I0428 11:25:38.872258 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterroles-validation with fingerprint=8622987dd84087c987112eca7539a46a80a7f53a70e04f3eb10ab5968fd30735 I0428 11:25:38.872304 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-ingress-config-validation with fingerprint=b66065e90b5e80135bf20291ada2f545b7dc01c95e57cc0910b8f37ed917bf7f I0428 11:25:38.872331 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-network-operator-validation with fingerprint=b15e52d8ff9b119837d0f8b1c23114d7353c52df0b8d1e236abb68e85e6e9096 I0428 11:25:38.872373 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-regular-user-validation with fingerprint=c5597a4071b9ed8866464410965a83aaef323bce61bfcc6bbf0600c483816402 I0428 11:25:38.872397 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-scc-validation with fingerprint=33e35d9d8fdb3cb588b929ea097812f4d1db6b8508cce931fed4edc6aea3bbeb I0428 11:25:38.872420 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-serviceaccount-validation with fingerprint=ab0586c6e807b6adfb03321979e9df1862eaedc041501ece0d93bfc0c471628b I0428 11:25:38.872450 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-techpreviewnoupgrade-validation with fingerprint=e10bdf9e09cfee9eeb0ef5d45aafbe66691fcde616628254c79308513429e4dd I0428 11:25:38.872462 1 gather.go:177] gatherer "clusterconfig" function "validating_webhook_configurations" took 39.656743ms to process 11 records W0428 11:25:38.873794 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0428 11:25:38.877414 1 sca.go:136] Pulling SCA certificates from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates. Next check is in 8h0m0s I0428 11:25:38.877495 1 cluster_transfer.go:83] checking the availability of cluster transfer. Next check is in 12h0m0s I0428 11:25:38.877544 1 tasks_processing.go:74] worker 14 stopped. W0428 11:25:38.877571 1 operator.go:288] started I0428 11:25:38.877597 1 base_controller.go:76] Waiting for caches to sync for LoggingSyncer I0428 11:25:38.878054 1 recorder.go:75] Recording config/node/ip-10-0-0-93.ec2.internal with fingerprint=05174dfdd1c4d5accd451991208d77d73636c285cb045c34794911782a1bc896 I0428 11:25:38.878159 1 recorder.go:75] Recording config/node/ip-10-0-1-48.ec2.internal with fingerprint=55744a8e0c940a920fe69c8461325dbdd5f57026cfb85d0df199f5148d10bee1 I0428 11:25:38.878280 1 recorder.go:75] Recording config/node/ip-10-0-2-63.ec2.internal with fingerprint=57b50373ed48ba7980fc61f9589ec87845d620b0ee81c5b53a605d32e50e8db1 I0428 11:25:38.878293 1 gather.go:177] gatherer "clusterconfig" function "nodes" took 47.788541ms to process 3 records I0428 11:25:38.878308 1 tasks_processing.go:74] worker 0 stopped. E0428 11:25:38.878319 1 gather.go:140] gatherer "clusterconfig" function "overlapping_namespace_uids" failed with the error: can't read uid range of the openshift-service-ca namespace I0428 11:25:38.878349 1 recorder.go:75] Recording config/namespaces_with_overlapping_uids with fingerprint=4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945 I0428 11:25:38.878367 1 gather.go:177] gatherer "clusterconfig" function "overlapping_namespace_uids" took 49.158318ms to process 1 records I0428 11:25:38.878468 1 tasks_processing.go:74] worker 49 stopped. I0428 11:25:38.879417 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/config/cluster with fingerprint=a5e7d79843e3dbca5be75cf31662cad6dc2ddd5e656a337040b8ad9a4789f630 I0428 11:25:38.879443 1 gather.go:177] gatherer "clusterconfig" function "image_registries" took 48.424587ms to process 1 records I0428 11:25:38.880472 1 tasks_processing.go:74] worker 11 stopped. I0428 11:25:38.880767 1 gather.go:177] gatherer "clusterconfig" function "certificate_signing_requests" took 51.191825ms to process 0 records I0428 11:25:38.896262 1 tasks_processing.go:74] worker 25 stopped. I0428 11:25:38.896871 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0428 11:25:38.896903 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0428 11:25:38.896929 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController I0428 11:25:38.898488 1 recorder.go:75] Recording config/pod/openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-0-93.ec2.internal with fingerprint=1404570aa84f44d5203f060c7a502ef01196dfe80dc410398104f2f7fedb7eb7 I0428 11:25:38.898558 1 recorder.go:75] Recording config/running_containers with fingerprint=ecf0348bbb6800c5d6ff9493613bee6f7220cd4e32f6f7d9e72e14743120bf8b I0428 11:25:38.898569 1 gather.go:177] gatherer "clusterconfig" function "container_images" took 67.068316ms to process 2 records I0428 11:25:38.904443 1 base_controller.go:82] Caches are synced for ConfigController I0428 11:25:38.904455 1 base_controller.go:119] Starting #1 worker of ConfigController controller ... I0428 11:25:38.908404 1 prometheus_rules.go:88] Prometheus rules successfully created I0428 11:25:38.908417 1 tasks_processing.go:74] worker 2 stopped. I0428 11:25:38.908611 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/admin with fingerprint=072f6f3db9b8f4ca0c7fe752856551fcdab38a7e186273d670edf198bff2d153 I0428 11:25:38.908697 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/edit with fingerprint=5e9065357faabc30afef260d90f99864b8e4b29838089b123e0f7fbd08350bca I0428 11:25:38.908706 1 gather.go:177] gatherer "clusterconfig" function "clusterroles" took 78.812095ms to process 2 records I0428 11:25:38.922415 1 tasks_processing.go:74] worker 5 stopped. I0428 11:25:38.922428 1 gather.go:177] gatherer "clusterconfig" function "number_of_pods_and_netnamespaces_with_sdn_annotations" took 93.042187ms to process 0 records I0428 11:25:38.923460 1 configmapobserver.go:84] configmaps "insights-config" not found I0428 11:25:38.923478 1 tasks_processing.go:74] worker 58 stopped. I0428 11:25:38.923732 1 recorder.go:75] Recording config/version with fingerprint=a37c7cb512e1d15e543cec54093e6b383e42036f01cfb8fc5d6f6f357a7d770b I0428 11:25:38.923746 1 recorder.go:75] Recording config/id with fingerprint=da74a2110dc57429e6a77537e8d9f912547d05d0d5a1f8ec790736325b7ee2d5 I0428 11:25:38.923752 1 gather.go:177] gatherer "clusterconfig" function "version" took 94.067712ms to process 2 records I0428 11:25:38.923827 1 tasks_processing.go:74] worker 59 stopped. I0428 11:25:38.924205 1 recorder.go:75] Recording config/crd/volumesnapshots.snapshot.storage.k8s.io with fingerprint=86673e30ebc762ba064299de78e4dd93b8d975afed603d91609392cf47582dd5 I0428 11:25:38.924452 1 recorder.go:75] Recording config/crd/volumesnapshotcontents.snapshot.storage.k8s.io with fingerprint=3c142863d2d579154cf18b6f078bcdfeeded659527fd3ccc93284a5970194267 I0428 11:25:38.924462 1 gather.go:177] gatherer "clusterconfig" function "crds" took 92.798468ms to process 2 records I0428 11:25:38.924518 1 recorder.go:75] Recording config/olm_operators with fingerprint=984b98d660bc399b586a6b7a310d79b12425654aab5f338b9673bb7a398f7d82 I0428 11:25:38.924525 1 gather.go:177] gatherer "clusterconfig" function "olm_operators" took 94.950658ms to process 1 records I0428 11:25:38.924533 1 tasks_processing.go:74] worker 26 stopped. I0428 11:25:38.925266 1 tasks_processing.go:74] worker 39 stopped. I0428 11:25:38.925304 1 recorder.go:75] Recording aggregated/unused_machine_configs_count with fingerprint=4bfc9fa984e5dfcd45848faaf05269de7619bf42edf9f781751af5ee05c1a499 I0428 11:25:38.925322 1 gather.go:177] gatherer "clusterconfig" function "machine_configs" took 95.119683ms to process 1 records I0428 11:25:38.925486 1 controller.go:212] Source scaController *sca.Controller is not ready I0428 11:25:38.925498 1 controller.go:212] Source clusterTransferController *clustertransfer.Controller is not ready I0428 11:25:38.925502 1 controller.go:212] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0428 11:25:38.925505 1 controller.go:212] Source periodic-conditional *controllerstatus.Simple is not ready I0428 11:25:38.925508 1 controller.go:212] Source periodic-workloads *controllerstatus.Simple is not ready I0428 11:25:38.925523 1 controller.go:489] The operator is still being initialized I0428 11:25:38.925531 1 controller.go:512] The operator is healthy I0428 11:25:38.925606 1 tasks_processing.go:74] worker 8 stopped. I0428 11:25:38.925619 1 gather.go:177] gatherer "clusterconfig" function "aggregated_monitoring_cr_names" took 96.43704ms to process 0 records E0428 11:25:38.930441 1 cluster_transfer.go:95] failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%277612f4c3-bede-40f1-8b1b-6a9549cbdf9c%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.130.0.14:60926->172.30.0.10:53: read: connection refused I0428 11:25:38.930454 1 controllerstatus.go:80] name=clusterTransferController healthy=true reason=Disconnected message=failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%277612f4c3-bede-40f1-8b1b-6a9549cbdf9c%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.130.0.14:60926->172.30.0.10:53: read: connection refused I0428 11:25:38.936496 1 requests.go:205] Asking for SCA certificate with "{"arch": ["x86_64"]}" payload I0428 11:25:38.939302 1 tasks_processing.go:74] worker 61 stopped. E0428 11:25:38.939317 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "cluster-monitoring-config" not found E0428 11:25:38.939323 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "gateway-mode-config" not found E0428 11:25:38.939326 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "insights-config" not found I0428 11:25:38.939335 1 recorder.go:75] Recording config/configmaps/openshift-config/installer-images/images.json with fingerprint=26b6661162b099a0f5a279859b4f46c867929a79d9a4a41fde4be4e6fe138018 I0428 11:25:38.939372 1 recorder.go:75] Recording config/configmaps/openshift-config/kube-root-ca.crt/ca.crt with fingerprint=d476c7d3f5b104863f08f481b1264dcc68cc272ecefb0ecb709b18a6afab034d I0428 11:25:38.939383 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/invoker with fingerprint=76b482f683cd3ef9da02debac5b26080a5aeb06ff768ee5c21117514dff29d8a I0428 11:25:38.939387 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/version with fingerprint=c93090eb0d2a4736885abeb79c91680cfd01fda46464f83456b085d4dc8239f0 I0428 11:25:38.939391 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 I0428 11:25:38.939429 1 recorder.go:75] Recording config/configmaps/openshift-config/rosa-brand-logo/rosa-brand-logo.svg with fingerprint=6ed8ca4dd7a8eee7249182bc006e9649ce84d76c551ddfaaa33e55d8c4cc1ed0 I0428 11:25:38.939437 1 recorder.go:75] Recording config/configmaps/kube-system/cluster-config-v1/install-config with fingerprint=ab3811c6b83fd7b8e920094cfa3080d1b4ee3c35ec4c8379437b21d27bd6608d I0428 11:25:38.939442 1 gather.go:177] gatherer "clusterconfig" function "config_maps" took 109.505124ms to process 7 records W0428 11:25:38.939832 1 sca.go:161] Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.130.0.14:44723->172.30.0.10:53: read: connection refused I0428 11:25:38.939844 1 controllerstatus.go:80] name=scaController healthy=true reason=NonHTTPError message=Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.130.0.14:44723->172.30.0.10:53: read: connection refused I0428 11:25:38.950344 1 tasks_processing.go:74] worker 43 stopped. I0428 11:25:38.950360 1 gather.go:177] gatherer "clusterconfig" function "node_logs" took 119.425844ms to process 0 records I0428 11:25:38.978138 1 base_controller.go:82] Caches are synced for LoggingSyncer I0428 11:25:38.978152 1 base_controller.go:119] Starting #1 worker of LoggingSyncer controller ... I0428 11:25:38.982460 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0428 11:25:38.986803 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0428 11:25:39.022055 1 tasks_processing.go:74] worker 22 stopped. E0428 11:25:39.022072 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret 'router-certs-default' in namespace 'openshift-ingress': secrets "router-certs-default" not found E0428 11:25:39.022078 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret '2pv477a5lhnbva0e6ci6jb0asv05tnph-primary-cert-bundle-secret' in namespace 'openshift-ingress-operator': secrets "2pv477a5lhnbva0e6ci6jb0asv05tnph-primary-cert-bundle-secret" not found I0428 11:25:39.022157 1 recorder.go:75] Recording aggregated/ingress_controllers_certs with fingerprint=8e8ceb9e36c60b3aa62f9ee37a94475f6cc2a2f5233f9cf736b5eb630e1365a8 I0428 11:25:39.022171 1 gather.go:177] gatherer "clusterconfig" function "ingress_certificates" took 192.362487ms to process 1 records I0428 11:25:39.278176 1 gather_cluster_operator_pods_and_events.go:121] Found 18 pods with 21 containers I0428 11:25:39.278190 1 gather_cluster_operator_pods_and_events.go:235] Maximum buffer size: 1198372 bytes I0428 11:25:39.278907 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-d6644 pod in namespace openshift-dns (previous: false). I0428 11:25:39.505078 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-d6644 pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-d6644\" is waiting to start: ContainerCreating" I0428 11:25:39.505107 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-d6644\" is waiting to start: ContainerCreating" I0428 11:25:39.505119 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-d6644 pod in namespace openshift-dns (previous: false). I0428 11:25:39.682420 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-d6644 pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-d6644\" is waiting to start: ContainerCreating" I0428 11:25:39.682440 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-d6644\" is waiting to start: ContainerCreating" I0428 11:25:39.682469 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-gg77d pod in namespace openshift-dns (previous: false). W0428 11:25:39.871374 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0428 11:25:39.901896 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-gg77d pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-gg77d\" is waiting to start: ContainerCreating" I0428 11:25:39.902091 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-gg77d\" is waiting to start: ContainerCreating" I0428 11:25:39.902106 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-gg77d pod in namespace openshift-dns (previous: false). I0428 11:25:40.085448 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-gg77d pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-gg77d\" is waiting to start: ContainerCreating" I0428 11:25:40.085466 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-gg77d\" is waiting to start: ContainerCreating" I0428 11:25:40.085496 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-kxz2s pod in namespace openshift-dns (previous: false). I0428 11:25:40.304612 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-kxz2s pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-kxz2s\" is waiting to start: ContainerCreating" I0428 11:25:40.304634 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-kxz2s\" is waiting to start: ContainerCreating" I0428 11:25:40.304642 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-kxz2s pod in namespace openshift-dns (previous: false). I0428 11:25:40.329761 1 gather_cluster_operators.go:184] Unable to get configs.samples.operator.openshift.io resource due to: configs.samples.operator.openshift.io "cluster" not found I0428 11:25:40.487654 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-kxz2s pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-kxz2s\" is waiting to start: ContainerCreating" I0428 11:25:40.487673 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-kxz2s\" is waiting to start: ContainerCreating" I0428 11:25:40.487683 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-f7mmc pod in namespace openshift-dns (previous: false). I0428 11:25:40.683080 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0428 11:25:40.683098 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-nl7c4 pod in namespace openshift-dns (previous: false). W0428 11:25:40.871245 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0428 11:25:40.882382 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0428 11:25:40.882397 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-vtkv2 pod in namespace openshift-dns (previous: false). I0428 11:25:40.933402 1 tasks_processing.go:74] worker 4 stopped. I0428 11:25:40.933455 1 recorder.go:75] Recording config/clusteroperator/console with fingerprint=2a03988145b34a17f8ce94a2eda8d577c314d42756613ae053ff03bf636a18cc I0428 11:25:40.933490 1 recorder.go:75] Recording config/clusteroperator/csi-snapshot-controller with fingerprint=e4011cd35c2d40649530fa6f9f6492de5cea9667f689583dfe17e8e6a04975d5 I0428 11:25:40.933519 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/csisnapshotcontroller/cluster with fingerprint=5adc514f4b63e2f1ecc68bf6f9c0af70c5eea04522a49524e102721b1c41f80e I0428 11:25:40.933547 1 recorder.go:75] Recording config/clusteroperator/dns with fingerprint=6c0a18442dce3cc2f8b1994ca26d31b60a08c59ef7f10690d5d1035dba047543 I0428 11:25:40.933579 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/dns/default with fingerprint=9e7b4ce029030d3d8c3b49af92c556acdcc415000b40d3f969dbdc42c432b47f I0428 11:25:40.933602 1 recorder.go:75] Recording config/clusteroperator/image-registry with fingerprint=7a974c7a0d14444c0c8bb9119db54863275bc08e3ec437d1d983172e13a402ed I0428 11:25:40.933632 1 recorder.go:75] Recording config/clusteroperator/ingress with fingerprint=f44c7a760bc48e63c80e1b87fac01dc389c8e6d9ef959748c522736c09f54ffb I0428 11:25:40.933657 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/ingresscontroller/openshift-ingress-operator/default with fingerprint=a9e33af0455d20b3871de81435ddf9df49a2a3a00bcdc70890666b190a9c54a8 I0428 11:25:40.933671 1 recorder.go:75] Recording config/clusteroperator/insights with fingerprint=ecf7c6e2b73561bdb141df751f42909c094bf6f55ffeaea0a1eaf9628ede3b88 I0428 11:25:40.933712 1 recorder.go:75] Recording config/clusteroperator/kube-apiserver with fingerprint=d80261eb8f525e5cc4c2fe80c42cb1dd9dc9a63c5f0a118390b9d6fab8ba61e5 I0428 11:25:40.933724 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubeapiserver/cluster with fingerprint=51503bf0b784fcf65ea46bcaf1f72ac1a5c4d5dc211934f18f27871efed05762 I0428 11:25:40.933740 1 recorder.go:75] Recording config/clusteroperator/kube-controller-manager with fingerprint=b82ce499eb15153d5ca26be9fd50a5af50c674cbae6cdd4d46d727857b8e6a35 I0428 11:25:40.933751 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubecontrollermanager/cluster with fingerprint=ce90c0d4f367d7da085074268031798382ae7c54fdcb0a21f15a4818fe308c11 I0428 11:25:40.933767 1 recorder.go:75] Recording config/clusteroperator/kube-scheduler with fingerprint=2a5af4faec58018a0f86850a5c3a434fbc4fe40db1eb6935d925bedd517b454c I0428 11:25:40.933774 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubescheduler/cluster with fingerprint=f2940fb9fd20c19951dfc295eb363b7fba0c505f5ae61f01967a063099e6b60a I0428 11:25:40.933786 1 recorder.go:75] Recording config/clusteroperator/kube-storage-version-migrator with fingerprint=970a4348dbc8e5e6a22adad4fa81af559c310cf7f6044aeb5076c49703434a25 I0428 11:25:40.933796 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubestorageversionmigrator/cluster with fingerprint=9351181aa7e6ada41ef581ab31e13516c6b934cc95710154bafb2eb222cb58db I0428 11:25:40.933811 1 recorder.go:75] Recording config/clusteroperator/monitoring with fingerprint=fe6e0098af312c6b1ed9f795b872567e09976736b4c9a74dbcb18003a5e0f1f5 I0428 11:25:40.933931 1 recorder.go:75] Recording config/clusteroperator/network with fingerprint=64dec12f39b7bfc215e1ca3f5b2ec62a497d04a229b0b347096ec09ebf56597b I0428 11:25:40.933940 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/ovn with fingerprint=626a89d20e0deaed5b6dfb533acfe65f4bb1618bd200a703b62e60c5d16d94ab I0428 11:25:40.933947 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/signer with fingerprint=90410b16914712b85b3c4578716ad8c0ae072e688f4cd1e022bf76f20da3506d I0428 11:25:40.933976 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/network/cluster with fingerprint=a93d15eaecb455a0e40ecb2826eeecc1533899204ddd3c3921d15ab70af7ae75 I0428 11:25:40.933998 1 recorder.go:75] Recording config/clusteroperator/node-tuning with fingerprint=683defe2a5f0438faf0bf83393d8fff83b6925c048c843f3f0b93fe8c5ae6fc6 I0428 11:25:40.934021 1 recorder.go:75] Recording config/clusteroperator/openshift-apiserver with fingerprint=253bcb2cbe85980f4ba21126dd77fce31ebf81efaa3799951e6422a9c6d322c8 I0428 11:25:40.934030 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftapiserver/cluster with fingerprint=e712e6cf27339b441e4ed1f4cde91dbde7e952698ba93407e4457db63a4a4c76 I0428 11:25:40.934047 1 recorder.go:75] Recording config/clusteroperator/openshift-controller-manager with fingerprint=b504ff3d1f72a556372b4e700d034fc6209be04b2001342d6d02517f2eaa78e5 I0428 11:25:40.934056 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftcontrollermanager/cluster with fingerprint=d71a0f4672f9b45d9fc8293bf1687afc650fd28d32e2e30de27523fe7b4eadf7 I0428 11:25:40.934069 1 recorder.go:75] Recording config/clusteroperator/openshift-samples with fingerprint=5b53eac6143cbc633b80aeebc2d2f1d3d9724e95124537154c79f80788d2cb7e I0428 11:25:40.934084 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager with fingerprint=e3f9be21d9130d9435e472b0822297ca654be61e9bb1296c6a04d2996991c79c I0428 11:25:40.934100 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-catalog with fingerprint=10ffec34ae2ad8f6599daf8aab9fc9a28f59f44594fce0e0ff541ef83c15fe13 I0428 11:25:40.934115 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-packageserver with fingerprint=b0c69c41197e82b0f507c30ae66ea3a5b161e55119a14b6fcb5ad9196f2ee387 I0428 11:25:40.934135 1 recorder.go:75] Recording config/clusteroperator/service-ca with fingerprint=548657d0878e306b705b7dba55405f6cde52d9910611dccafd657e7ecec6a314 I0428 11:25:40.934144 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/serviceca/cluster with fingerprint=812f7edc2cdb30e61e7f2b29454357a40b1a507a4b0c2b7729193b67f0e3b4aa I0428 11:25:40.934171 1 recorder.go:75] Recording config/clusteroperator/storage with fingerprint=f4399fab136f7b204ee66df84c0319a179673a082954db6ddbaa1d97fd5ff0e9 I0428 11:25:40.934187 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/clustercsidriver/ebs.csi.aws.com with fingerprint=510064d6f6bcced87ab5bd2ddaff3d0edd7f93f4a4f7af2641f29fc53ffab21e I0428 11:25:40.934195 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/storage/cluster with fingerprint=8e480f8c1ce1b39baac42d8ec780c57c2592929ae0c801b61ffad49ba13f33ad I0428 11:25:40.934202 1 gather.go:177] gatherer "clusterconfig" function "operators" took 2.104126533s to process 36 records I0428 11:25:41.082983 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0428 11:25:41.083035 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-5688fdf595-8dx4w pod in namespace openshift-image-registry (previous: false). I0428 11:25:41.283263 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-5688fdf595-8dx4w pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-5688fdf595-8dx4w\" is waiting to start: ContainerCreating" I0428 11:25:41.283281 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-5688fdf595-8dx4w\" is waiting to start: ContainerCreating" I0428 11:25:41.283327 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-76f6f745fb-j2kkz pod in namespace openshift-image-registry (previous: false). I0428 11:25:41.482994 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-76f6f745fb-j2kkz pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-76f6f745fb-j2kkz\" is waiting to start: ContainerCreating" I0428 11:25:41.483010 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-76f6f745fb-j2kkz\" is waiting to start: ContainerCreating" I0428 11:25:41.483042 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-76f6f745fb-td46t pod in namespace openshift-image-registry (previous: false). I0428 11:25:41.681924 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-76f6f745fb-td46t pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-76f6f745fb-td46t\" is waiting to start: ContainerCreating" I0428 11:25:41.681941 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-76f6f745fb-td46t\" is waiting to start: ContainerCreating" I0428 11:25:41.681951 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-g282m pod in namespace openshift-image-registry (previous: false). W0428 11:25:41.871826 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0428 11:25:41.886080 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0428 11:25:41.886094 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-hdk27 pod in namespace openshift-image-registry (previous: false). I0428 11:25:42.081552 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0428 11:25:42.081569 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-wnmcr pod in namespace openshift-image-registry (previous: false). I0428 11:25:42.282049 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0428 11:25:42.282068 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-7576d6b94c-8h6pr pod in namespace openshift-ingress (previous: false). I0428 11:25:42.483158 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-7576d6b94c-8h6pr pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-7576d6b94c-8h6pr\" is waiting to start: ContainerCreating" I0428 11:25:42.483176 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-7576d6b94c-8h6pr\" is waiting to start: ContainerCreating" I0428 11:25:42.483187 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-768746d744-vx4qc pod in namespace openshift-ingress (previous: false). I0428 11:25:42.682384 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-768746d744-vx4qc pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-768746d744-vx4qc\" is waiting to start: ContainerCreating" I0428 11:25:42.682402 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-768746d744-vx4qc\" is waiting to start: ContainerCreating" I0428 11:25:42.682413 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-768746d744-wx9vr pod in namespace openshift-ingress (previous: false). W0428 11:25:42.871517 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0428 11:25:42.881732 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-768746d744-wx9vr pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-768746d744-wx9vr\" is waiting to start: ContainerCreating" I0428 11:25:42.881746 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-768746d744-wx9vr\" is waiting to start: ContainerCreating" I0428 11:25:42.881772 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-46xcr pod in namespace openshift-ingress-canary (previous: false). I0428 11:25:43.081145 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-46xcr pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-46xcr\" is waiting to start: ContainerCreating" I0428 11:25:43.081161 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-46xcr\" is waiting to start: ContainerCreating" I0428 11:25:43.081193 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-5pc9r pod in namespace openshift-ingress-canary (previous: false). I0428 11:25:43.281475 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-5pc9r pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-5pc9r\" is waiting to start: ContainerCreating" I0428 11:25:43.281492 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-5pc9r\" is waiting to start: ContainerCreating" I0428 11:25:43.281523 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-rlg77 pod in namespace openshift-ingress-canary (previous: false). I0428 11:25:43.482427 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for ingress-canary-rlg77 pod in namespace openshift-ingress-canary for failing operator serve-healthcheck-canary (previous: false): "container \"serve-healthcheck-canary\" in pod \"ingress-canary-rlg77\" is waiting to start: ContainerCreating" I0428 11:25:43.482445 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"serve-healthcheck-canary\" in pod \"ingress-canary-rlg77\" is waiting to start: ContainerCreating" I0428 11:25:43.482461 1 tasks_processing.go:74] worker 56 stopped. I0428 11:25:43.482545 1 recorder.go:75] Recording events/openshift-dns-operator with fingerprint=295e9aff805eaace1e4334c22c3b1c8371021febcc4ae082a362d34792fba9cf I0428 11:25:43.482597 1 recorder.go:75] Recording events/openshift-dns with fingerprint=4597ee775bdb118af8988cb0e3fa05ba412428d33fc479abd783b33eaad139d6 I0428 11:25:43.482686 1 recorder.go:75] Recording events/openshift-image-registry with fingerprint=c2342260d3f77d8449e4ecbbc450b311a2277838c2e824e617b1b16177713a01 I0428 11:25:43.482712 1 recorder.go:75] Recording events/openshift-ingress-operator with fingerprint=fe00b3174d330872327efd111a8e50883c0eb11e5cd15f5116edbc81e38caf41 I0428 11:25:43.482759 1 recorder.go:75] Recording events/openshift-ingress with fingerprint=fb5023b37c3781fb8e2629eff531170e60f8857cfdc68a30352746b96fa5494e I0428 11:25:43.482776 1 recorder.go:75] Recording events/openshift-ingress-canary with fingerprint=b5af2d4dae4c8dd007422770ae2ef8fe255e4e78abe58a1318880314ff3c3185 I0428 11:25:43.482906 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-d6644 with fingerprint=46bbf405d66101b85b04e055c7e9d904dcbd7c6ff3853f0b6cf2d21443ed7543 I0428 11:25:43.482990 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-gg77d with fingerprint=6d74549486e67aa1fdb88f2aa9f6eb5f021e576b8f371a8b2c5089980c06cf62 I0428 11:25:43.483062 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-kxz2s with fingerprint=142ce8c5463cf760272cd50fbb5babcc0677f226e8238bc475ecba371884c687 I0428 11:25:43.483166 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-5688fdf595-8dx4w with fingerprint=c751627768c6020d657c8ee5c86ffdcf29a1124c49580942c0975d1b99784404 I0428 11:25:43.483276 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-76f6f745fb-j2kkz with fingerprint=9f33801246113e62c9834c1a11f1f60aa24642d3079c9c5131e85c8ad18e3fcf I0428 11:25:43.483365 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-76f6f745fb-td46t with fingerprint=eeaabeee9925ec4edcd7e2e907fb82dd8246db194aaf018a920fa56fbab1a708 I0428 11:25:43.483420 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-46xcr with fingerprint=6dc35d245273a6616a8a7edff4aae515305ad145cd81bf55e50da23034d4e0ca I0428 11:25:43.483474 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-5pc9r with fingerprint=39f6c387874e037da342aa722b17d5e8bbcdf7ab5614eb85b21a9620790417b2 I0428 11:25:43.483528 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/ingress-canary-rlg77 with fingerprint=4fafafc15b56dd99d971ba2f1c65b8f08634c74a41bc3c1cc2914fffe253575d I0428 11:25:43.483537 1 gather.go:177] gatherer "clusterconfig" function "operators_pods_and_events" took 4.65193766s to process 15 records W0428 11:25:43.871346 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. W0428 11:25:43.871372 1 gather_dvo_metrics.go:117] Unable to read metrics from endpoint "http://deployment-validation-operator-metrics.openshift-deployment-validation-operator.svc:8383": DVO metrics service was not available within the 5s timeout: context deadline exceeded I0428 11:25:43.871383 1 tasks_processing.go:74] worker 54 stopped. E0428 11:25:43.871393 1 gather.go:140] gatherer "clusterconfig" function "dvo_metrics" failed with the error: DVO metrics service was not available within the 5s timeout: context deadline exceeded I0428 11:25:43.871404 1 recorder.go:75] Recording config/dvo_metrics with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 W0428 11:25:43.871417 1 gather.go:155] issue recording gatherer "clusterconfig" function "dvo_metrics" result "config/dvo_metrics" because of the warning: warning: the record with the same fingerprint "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" was already recorded at path "config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt", recording another one with a different path "config/dvo_metrics" I0428 11:25:43.871435 1 gather.go:177] gatherer "clusterconfig" function "dvo_metrics" took 5.040748256s to process 1 records I0428 11:25:49.271705 1 configmapobserver.go:84] configmaps "insights-config" not found I0428 11:25:51.283979 1 tasks_processing.go:74] worker 30 stopped. I0428 11:25:51.284020 1 recorder.go:75] Recording config/installplans with fingerprint=95dfd5f33a9a46199239158a0ec0183b629818eca24ae41f4d706d7a3d604aec I0428 11:25:51.284033 1 gather.go:177] gatherer "clusterconfig" function "install_plans" took 12.45351181s to process 1 records I0428 11:25:52.234800 1 tasks_processing.go:74] worker 15 stopped. I0428 11:25:52.235123 1 recorder.go:75] Recording config/serviceaccounts with fingerprint=5bfe903b1da30878a989600cf95950087bbfed0fc0418c904f1a71b56765f5ff I0428 11:25:52.235140 1 gather.go:177] gatherer "clusterconfig" function "service_accounts" took 13.405368736s to process 1 records E0428 11:25:52.235206 1 periodic.go:247] "Unhandled Error" err="clusterconfig failed after 13.406s with: function \"pod_network_connectivity_checks\" failed with an error, function \"machines\" failed with an error, function \"machine_healthchecks\" failed with an error, function \"support_secret\" failed with an error, function \"overlapping_namespace_uids\" failed with an error, function \"config_maps\" failed with an error, function \"ingress_certificates\" failed with an error, function \"dvo_metrics\" failed with an error" I0428 11:25:52.236325 1 controllerstatus.go:89] name=periodic-clusterconfig healthy=false reason=PeriodicGatherFailed message=Source clusterconfig could not be retrieved: function "pod_network_connectivity_checks" failed with an error, function "machines" failed with an error, function "machine_healthchecks" failed with an error, function "support_secret" failed with an error, function "overlapping_namespace_uids" failed with an error, function "config_maps" failed with an error, function "ingress_certificates" failed with an error, function "dvo_metrics" failed with an error I0428 11:25:52.236341 1 periodic.go:209] Running workloads gatherer I0428 11:25:52.236355 1 tasks_processing.go:45] number of workers: 2 I0428 11:25:52.236363 1 tasks_processing.go:69] worker 1 listening for tasks. I0428 11:25:52.236367 1 tasks_processing.go:71] worker 1 working on workload_info task. I0428 11:25:52.236430 1 tasks_processing.go:69] worker 0 listening for tasks. I0428 11:25:52.236501 1 tasks_processing.go:71] worker 0 working on helmchart_info task. I0428 11:25:52.262865 1 tasks_processing.go:74] worker 0 stopped. I0428 11:25:52.262884 1 gather.go:177] gatherer "workloads" function "helmchart_info" took 26.347865ms to process 0 records I0428 11:25:52.263664 1 gather_workloads_info.go:278] Loaded pods in 0s, will wait 22s for image data I0428 11:25:52.271527 1 gather_workloads_info.go:387] No image sha256:765f0d23b637f685f98a31bd47c131b03cf72a40761a3f9a9d6320faa3c33733 (9ms) I0428 11:25:52.279332 1 gather_workloads_info.go:387] No image sha256:2904a78e2eb73fd6a9bb94c105c2a056831fb4113fbb7b0607c50adc9d879c9b (8ms) I0428 11:25:52.288072 1 gather_workloads_info.go:387] No image sha256:04c87c054a3f366a7dfbe0a93ebb0c80a098ee16842c5794b67c1202eec61996 (9ms) I0428 11:25:52.296262 1 gather_workloads_info.go:387] No image sha256:2e564f336c77116053f34d4201d364d8da04e789cfffa0ea422574c95f2d6404 (8ms) I0428 11:25:52.304045 1 gather_workloads_info.go:387] No image sha256:ae7d3453fd734ecc865e5f9bb16f29244ebffe6291b77e1d4e496f71eb053174 (8ms) I0428 11:25:52.315249 1 gather_workloads_info.go:387] No image sha256:4556896f77307821531ef91b7b7faccb82b824ea695693b2989f597f0deca038 (11ms) I0428 11:25:52.322498 1 gather_workloads_info.go:387] No image sha256:a56211d075aa43cbb491f669a5b2e46ee023dc95b7d51dbac28f463948c5ad61 (7ms) I0428 11:25:52.330250 1 gather_workloads_info.go:387] No image sha256:36b9e89c3cfcf1ab9ae500486e38afb6862cba48cb0b4d84a09508ab8f3d299f (8ms) I0428 11:25:52.338195 1 gather_workloads_info.go:387] No image sha256:7adc1eab05d6724c76ba751f6df816b08d6e70b78dee9eb94fa6fd9690542c98 (8ms) I0428 11:25:52.345930 1 gather_workloads_info.go:387] No image sha256:2598489729a4b258e4ecda4a06f6875133f2a10ced5c5241f8a57a8a05418e36 (8ms) I0428 11:25:52.371840 1 gather_workloads_info.go:387] No image sha256:a0105d1eb62cf6ac9e5e2ef28d3e89bf6dc514bc594fc7090fe5a5ee18a09c87 (26ms) I0428 11:25:52.472450 1 gather_workloads_info.go:387] No image sha256:5a95c19d82767e0235b4edb4a0536482c816904897aae1dc3eb255cb52b87a9f (101ms) I0428 11:25:52.572293 1 gather_workloads_info.go:387] No image sha256:a258c226562adb14e3a163a1940938526ee6a0928982a7667d85d9a7334ce639 (100ms) I0428 11:25:52.672333 1 gather_workloads_info.go:387] No image sha256:943018739e3db1763c3184b460dbc409e058abbac76d57b9927faad317be85e4 (100ms) I0428 11:25:52.772596 1 gather_workloads_info.go:387] No image sha256:e84cb128d930bd1ab867cc89b7b7bf2b2c0e41105ab93b5381069945b3ee9c57 (100ms) I0428 11:25:52.872480 1 gather_workloads_info.go:387] No image sha256:653c666f842c13e0baae2e89a9b1efe0e2ef56f621ffb5b32005115d2a26ab8c (100ms) I0428 11:25:52.972432 1 gather_workloads_info.go:387] No image sha256:7b31223098f08328f5ddea8e5b871dbbd5f5a61ec550e8956f66793c0c6031a9 (100ms) I0428 11:25:53.072980 1 gather_workloads_info.go:387] No image sha256:91828234f107c068c8a4966d08370ae7b73e637651dbc6d92c18c4553402c22c (101ms) I0428 11:25:53.163700 1 configmapobserver.go:84] configmaps "insights-config" not found I0428 11:25:53.172195 1 gather_workloads_info.go:387] No image sha256:521712486e2c6e3c020dad6a1cb340db8e55665b69f7c208fab9cd9e965fd588 (99ms) I0428 11:25:53.272135 1 gather_workloads_info.go:387] No image sha256:47154813651033d59751fb655a384dbffb64dd26f10bd7f3be0c3128d0486356 (100ms) I0428 11:25:53.358822 1 configmapobserver.go:84] configmaps "insights-config" not found I0428 11:25:53.374850 1 gather_workloads_info.go:387] No image sha256:f68c0da8268f23078f5f760a91a3c8212f2195acf18c231f6e034207a67adeec (103ms) I0428 11:25:53.472365 1 gather_workloads_info.go:387] No image sha256:289816958633a763a72dbc44e1dad40466223164e7e253039514f0d974ea5d21 (97ms) I0428 11:25:53.573513 1 gather_workloads_info.go:387] No image sha256:2e57e192c3c1240fd935dcd55c8fde5e70e78bf81d6176c96edf21fafe59f8ba (101ms) I0428 11:25:53.672728 1 gather_workloads_info.go:387] No image sha256:03cf4cd7ef1518610c6c7b3ad27d1622d82e98e3dc6e3f8e5d0fceb5c8d3786e (99ms) I0428 11:25:53.774283 1 gather_workloads_info.go:387] No image sha256:0a99240166165eb5718e7516a43282fe32df9c7c5e809b31b58abe44e42ff94d (102ms) I0428 11:25:53.873468 1 gather_workloads_info.go:387] No image sha256:56a85660a445eced5c79a595a0eccf590087c5672d50f49d4c25ad52f9a44f04 (99ms) I0428 11:25:53.972749 1 gather_workloads_info.go:387] No image sha256:695cf2f0cc07683c2a3ce1eaf3e56fe18abc6e2bac716f7d9843f5d173b9df52 (99ms) I0428 11:25:54.073035 1 gather_workloads_info.go:387] No image sha256:5f0b67cfbbc381243fb91ccc17345b56d05f4d717c667e8c644e5bf05633ba71 (100ms) I0428 11:25:54.176402 1 gather_workloads_info.go:387] No image sha256:a498046d64605bcccee2440aa4f04a4602baaae263cf01d977ec5208e876b1fd (103ms) I0428 11:25:54.272644 1 gather_workloads_info.go:387] No image sha256:ca1344cb64140188b7cae7bbc51fb751566c0b0c97d5e39b5850e628032c4a5e (96ms) I0428 11:25:54.372641 1 gather_workloads_info.go:387] No image sha256:1a2532940843248c57d52141185dd71fbc393ab28b65d48f682038632c1dbbad (100ms) I0428 11:25:54.472924 1 gather_workloads_info.go:387] No image sha256:b3909bf664c77097f75b3768830863d642eed3815dab2bfb4415c771ca2d5007 (100ms) I0428 11:25:54.572630 1 gather_workloads_info.go:387] No image sha256:c940ea87e7d133d75ba0002ef00c0806825eed3db8094cdb260d1bac18127733 (100ms) I0428 11:25:54.572661 1 tasks_processing.go:74] worker 1 stopped. E0428 11:25:54.572670 1 gather.go:140] gatherer "workloads" function "workload_info" failed with the error: no running pods found for the insights-runtime-extractor statefulset I0428 11:25:54.572895 1 recorder.go:75] Recording config/workload_info with fingerprint=e7a1722a5081293a725077e329fa75462ad0a077cea0af131a3779fbcf0877cc I0428 11:25:54.572911 1 gather.go:177] gatherer "workloads" function "workload_info" took 2.336286973s to process 1 records E0428 11:25:54.572934 1 periodic.go:247] "Unhandled Error" err="workloads failed after 2.336s with: function \"workload_info\" failed with an error" I0428 11:25:54.574033 1 controllerstatus.go:89] name=periodic-workloads healthy=false reason=PeriodicGatherFailed message=Source workloads could not be retrieved: function "workload_info" failed with an error I0428 11:25:54.574046 1 periodic.go:209] Running conditional gatherer I0428 11:25:54.579787 1 requests.go:294] Making HTTP GET request at: https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules I0428 11:25:54.588306 1 conditional_gatherer.go:107] Get "https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.130.0.14:51582->172.30.0.10:53: read: connection refused E0428 11:25:54.588542 1 conditional_gatherer.go:322] unable to update alerts cache: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0428 11:25:54.588600 1 conditional_gatherer.go:384] updating version cache for conditional gatherer I0428 11:25:54.594292 1 conditional_gatherer.go:392] cluster version is '4.20.8' E0428 11:25:54.594308 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0428 11:25:54.594315 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0428 11:25:54.594319 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0428 11:25:54.594324 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0428 11:25:54.594327 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0428 11:25:54.594330 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0428 11:25:54.594333 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0428 11:25:54.594336 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0428 11:25:54.594340 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing I0428 11:25:54.594354 1 tasks_processing.go:45] number of workers: 3 I0428 11:25:54.594361 1 tasks_processing.go:69] worker 2 listening for tasks. I0428 11:25:54.594365 1 tasks_processing.go:71] worker 2 working on conditional_gatherer_rules task. I0428 11:25:54.594372 1 tasks_processing.go:69] worker 0 listening for tasks. I0428 11:25:54.594384 1 tasks_processing.go:71] worker 0 working on remote_configuration task. I0428 11:25:54.594390 1 tasks_processing.go:71] worker 2 working on rapid_container_logs task. I0428 11:25:54.594388 1 tasks_processing.go:69] worker 1 listening for tasks. I0428 11:25:54.594398 1 tasks_processing.go:74] worker 1 stopped. I0428 11:25:54.594444 1 recorder.go:75] Recording insights-operator/conditional-gatherer-rules with fingerprint=7034af97e7e41c22e4b775abdd4b9066c8ebb19da33eb7f69f39bfd2eb5f6406 I0428 11:25:54.594458 1 gather.go:177] gatherer "conditional" function "conditional_gatherer_rules" took 824ns to process 1 records I0428 11:25:54.594491 1 recorder.go:75] Recording insights-operator/remote-configuration with fingerprint=0394430c431eec4d48bb1811a90918e95161d2282c59af26f2473613cc0959db I0428 11:25:54.594500 1 gather.go:177] gatherer "conditional" function "remote_configuration" took 1.127µs to process 1 records I0428 11:25:54.594506 1 tasks_processing.go:74] worker 0 stopped. I0428 11:25:54.594640 1 tasks_processing.go:74] worker 2 stopped. I0428 11:25:54.594652 1 gather.go:177] gatherer "conditional" function "rapid_container_logs" took 245.318µs to process 0 records I0428 11:25:54.594671 1 controllerstatus.go:89] name=periodic-conditional healthy=false reason=NotAvailable message=Get "https://console.redhat.com/api/gathering/v2/4.20.8/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.130.0.14:51582->172.30.0.10:53: read: connection refused I0428 11:25:54.594689 1 recorder.go:75] Recording insights-operator/remote-configuration.json with fingerprint=359de9c990c741675cec72fda96b5c3682221efdb4799f5eaa6e9805bcd3b5c1 I0428 11:25:54.616576 1 recorder.go:75] Recording insights-operator/gathers with fingerprint=db1aafebc7d5b2847915e7840820ac46286a7323b899c07afe3e1dc210b06773 I0428 11:25:54.616698 1 diskrecorder.go:70] Writing 111 records to /var/lib/insights-operator/insights-2026-04-28-112554.tar.gz I0428 11:25:54.624585 1 diskrecorder.go:51] Wrote 111 records to disk in 7ms I0428 11:25:54.624616 1 periodic.go:278] Gathering cluster info every 2h0m0s I0428 11:25:54.624638 1 periodic.go:279] Configuration is dataReporting: interval: 2h0m0s, uploadEndpoint: https://console.redhat.com/api/ingress/v1/upload, storagePath: /var/lib/insights-operator, downloadEndpoint: https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports, conditionalGathererEndpoint: https://console.redhat.com/api/gathering/v2/%s/gathering_rules, obfuscation: [] sca: disabled: false, endpoint: https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates, interval: 8h0m0s alerting: disabled: false clusterTransfer: endpoint: https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/, interval: 12h0m0s proxy: httpProxy: , httpsProxy: , noProxy: I0428 11:26:05.788909 1 configmapobserver.go:84] configmaps "insights-config" not found I0428 11:26:58.349517 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.crt" has been created (hash="122d40262d58e69ef5d5d9b9ae26b8319106200f84579ec500b1218f3c56a417") W0428 11:26:58.349559 1 builder.go:160] Restart triggered because of file /var/run/secrets/serving-cert/tls.crt was created I0428 11:26:58.349606 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.key" has been created (hash="bb812f54654c84f4cc0b7cd544001c0e12d93d9a2b04b1f8cf92caf96dc605ac") I0428 11:26:58.349627 1 genericapiserver.go:548] "[graceful-termination] shutdown event" name="ShutdownInitiated" I0428 11:26:58.349664 1 simple_featuregate_reader.go:177] Shutting down feature-gate-detector I0428 11:26:58.349665 1 genericapiserver.go:693] "[graceful-termination] pre-shutdown hooks completed" name="PreShutdownHooksStopped" I0428 11:26:58.349687 1 requestheader_controller.go:194] Shutting down RequestHeaderAuthRequestController I0428 11:26:58.349689 1 observer_polling.go:111] Observed file "/var/run/configmaps/service-ca-bundle/service-ca.crt" has been created (hash="92ae37838fcbe22dfb898e236d153b763cd826f7b112006a7511930bb47b7736") I0428 11:26:58.349704 1 configmap_cafile_content.go:226] "Shutting down controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0428 11:26:58.349714 1 tlsconfig.go:258] "Shutting down DynamicServingCertificateController" I0428 11:26:58.349720 1 dynamic_serving_content.go:149] "Shutting down controller" name="serving-cert::/tmp/serving-cert-3840697712/tls.crt::/tmp/serving-cert-3840697712/tls.key" I0428 11:26:58.349736 1 genericapiserver.go:651] "[graceful-termination] not going to wait for active watch request(s) to drain" I0428 11:26:58.349751 1 base_controller.go:181] Shutting down ConfigController ... I0428 11:26:58.349763 1 base_controller.go:181] Shutting down LoggingSyncer ... I0428 11:26:58.349773 1 object_count_tracker.go:151] "StorageObjectCountTracker pruner is exiting" I0428 11:26:58.349739 1 secure_serving.go:258] Stopped listening on [::]:8443 I0428 11:26:58.349793 1 base_controller.go:123] Shutting down worker of ConfigController controller ...