W0423 19:50:37.585388 1 cmd.go:257] Using insecure, self-signed certificates I0423 19:50:38.496303 1 start.go:138] Unable to read service ca bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 19:50:38.496610 1 observer_polling.go:159] Starting file observer I0423 19:50:38.987489 1 operator.go:76] Starting insights-operator v0.0.0-master+$Format:%H$ I0423 19:50:38.987697 1 legacy_config.go:327] Current config: {"report":false,"storagePath":"/var/lib/insights-operator","interval":"2h","endpoint":"https://console.redhat.com/api/ingress/v1/upload","conditionalGathererEndpoint":"https://console.redhat.com/api/gathering/v2/%s/gathering_rules","pull_report":{"endpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports","delay":"60s","timeout":"3000s","min_retry":"30s"},"impersonate":"system:serviceaccount:openshift-insights:gather","enableGlobalObfuscation":false,"ocm":{"scaEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates","scaInterval":"8h","scaDisabled":false,"clusterTransferEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/","clusterTransferInterval":"12h"},"disableInsightsAlerts":false,"processingStatusEndpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status","reportEndpointTechPreview":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/report"} I0423 19:50:38.988250 1 simple_featuregate_reader.go:171] Starting feature-gate-detector I0423 19:50:38.988257 1 secure_serving.go:57] Forcing use of http/1.1 only W0423 19:50:38.988292 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. W0423 19:50:38.988305 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. W0423 19:50:38.988312 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. W0423 19:50:38.988316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. W0423 19:50:38.988319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. W0423 19:50:38.988323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. I0423 19:50:38.991279 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-insights", Name:"insights-operator", UID:"f703ec93-f128-4eb5-9f23-6631384be241", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BuildCSIVolumes", "CPMSMachineNamePrefix", "ConsolePluginContentSecurityPolicy", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNSInstall", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "HyperShiftOnlyDynamicResourceAllocation", "ImageStreamImportMode", "ImageVolume", "KMSv1", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "ManagedBootImagesAzure", "ManagedBootImagesvSphere", "MetricsCollectionProfiles", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "PinnedImages", "PreconfiguredUDNAddresses", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SigstoreImageVerification", "SigstoreImageVerificationPKI", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiDisk", "VSphereMultiNetworks", "VolumeAttributesClass"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSDualStackInstall", "AWSServiceLBNetworkSecurityGroup", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureDualStackInstall", "AzureMultiDisk", "BootImageSkewEnforcement", "BootcNodeManagement", "CBORServingAndStorage", "CRDCompatibilityRequirementOperator", "ClientsAllowCBOR", "ClientsPreferCBOR", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterAPIMachineManagement", "ClusterAPIMachineManagementVSphere", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "EtcdBackendQuota", "EventTTL", "EventedPLEG", "Example", "Example2", "ExternalSnapshotMetadata", "GCPClusterHostedDNS", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "GCPDualStackInstall", "ImageModeStatusReporting", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryption", "KMSEncryptionProvider", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "ManagedBootImagesCPMS", "MaxUnavailableStatefulSet", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "MultiDiskSetup", "MutableCSINodeAllocatableCount", "MutatingAdmissionPolicy", "NewOLM", "NewOLMBoxCutterRuntime", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterInstall", "NutanixMultiSubnets", "OSStreams", "OVNObservability", "OnPremDNSRecords", "OpenShiftPodSecurityAdmission", "ProvisioningRequestAvailable", "SELinuxMount", "ShortCertRotation", "SignatureStores", "TranslateStreamCloseWebsocketRequests", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VolumeGroupSnapshot"}} I0423 19:50:38.991271 1 operator.go:141] FeatureGates initialized: knownFeatureGates=[AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BuildCSIVolumes CPMSMachineNamePrefix ConsolePluginContentSecurityPolicy ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GCPClusterHostedDNSInstall GatewayAPI GatewayAPIController HighlyAvailableArbiter HyperShiftOnlyDynamicResourceAllocation ImageStreamImportMode ImageVolume KMSv1 MachineConfigNodes ManagedBootImages ManagedBootImagesAWS ManagedBootImagesAzure ManagedBootImagesvSphere MetricsCollectionProfiles NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation PinnedImages PreconfiguredUDNAddresses ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SigstoreImageVerification SigstoreImageVerificationPKI StoragePerformantSecurityPolicy UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiDisk VSphereMultiNetworks VolumeAttributesClass AWSClusterHostedDNS AWSClusterHostedDNSInstall AWSDedicatedHosts AWSDualStackInstall AWSServiceLBNetworkSecurityGroup AutomatedEtcdBackup AzureClusterHostedDNSInstall AzureDedicatedHosts AzureDualStackInstall AzureMultiDisk BootImageSkewEnforcement BootcNodeManagement CBORServingAndStorage CRDCompatibilityRequirementOperator ClientsAllowCBOR ClientsPreferCBOR ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterAPIMachineManagement ClusterAPIMachineManagementVSphere ClusterMonitoringConfig ClusterVersionOperatorConfiguration DNSNameResolver DualReplica DyanmicServiceEndpointIBMCloud EtcdBackendQuota EventTTL EventedPLEG Example Example2 ExternalSnapshotMetadata GCPClusterHostedDNS GCPCustomAPIEndpoints GCPCustomAPIEndpointsInstall GCPDualStackInstall ImageModeStatusReporting IngressControllerDynamicConfigurationManager InsightsConfig InsightsOnDemandDataGather IrreconcilableMachineConfig KMSEncryption KMSEncryptionProvider MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController ManagedBootImagesCPMS MaxUnavailableStatefulSet MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure MultiDiskSetup MutableCSINodeAllocatableCount MutatingAdmissionPolicy NewOLM NewOLMBoxCutterRuntime NewOLMCatalogdAPIV1Metas NewOLMOwnSingleNamespace NewOLMPreflightPermissionChecks NewOLMWebhookProviderOpenshiftServiceCA NoRegistryClusterInstall NutanixMultiSubnets OSStreams OVNObservability OnPremDNSRecords OpenShiftPodSecurityAdmission ProvisioningRequestAvailable SELinuxMount ShortCertRotation SignatureStores TranslateStreamCloseWebsocketRequests VSphereConfigurableMaxAllowedBlockVolumesPerNode VSphereHostVMGroupZonal VSphereMixedNodeEnv VolumeGroupSnapshot] I0423 19:50:38.994315 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController I0423 19:50:38.994353 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController I0423 19:50:38.994352 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0423 19:50:38.994354 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0423 19:50:38.994369 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0423 19:50:38.994375 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0423 19:50:38.994661 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/tmp/serving-cert-1516100057/tls.crt::/tmp/serving-cert-1516100057/tls.key" I0423 19:50:38.994923 1 secure_serving.go:213] Serving securely on [::]:8443 I0423 19:50:38.994958 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" W0423 19:50:38.997487 1 configmapobserver.go:64] Cannot get the configuration config map: configmaps "insights-config" not found. Default configuration is used. I0423 19:50:38.997518 1 secretconfigobserver.go:216] Legacy configuration set: enabled=false endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=false reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0423 19:50:38.997614 1 base_controller.go:76] Waiting for caches to sync for ConfigController I0423 19:50:39.002604 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0423 19:50:39.002621 1 secretconfigobserver.go:204] Legacy configuration updated: enabled=true endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=true reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0423 19:50:39.006832 1 secretconfigobserver.go:119] support secret does not exist I0423 19:50:39.011270 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0423 19:50:39.015559 1 secretconfigobserver.go:119] support secret does not exist I0423 19:50:39.019137 1 recorder.go:176] Pruning old reports every 8h37m4s, max age is 288h0m0s I0423 19:50:39.023922 1 periodic.go:216] Running clusterconfig gatherer I0423 19:50:39.023932 1 controllerstatus.go:80] name=insightsuploader healthy=true reason= message= I0423 19:50:39.023948 1 insightsuploader.go:86] Reporting status periodically to https://console.redhat.com/api/ingress/v1/upload every 2h0m0s, starting in 1m30s I0423 19:50:39.023968 1 tasks_processing.go:45] number of workers: 64 I0423 19:50:39.023980 1 controllerstatus.go:80] name=insightsreport healthy=true reason= message= I0423 19:50:39.023987 1 insightsreport.go:296] Starting report retriever I0423 19:50:39.023987 1 tasks_processing.go:69] worker 2 listening for tasks. I0423 19:50:39.023993 1 insightsreport.go:298] Insights analysis reports will be downloaded from the https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports endpoint with a delay of 1m0s I0423 19:50:39.023991 1 tasks_processing.go:69] worker 13 listening for tasks. I0423 19:50:39.024003 1 tasks_processing.go:69] worker 0 listening for tasks. I0423 19:50:39.024003 1 tasks_processing.go:71] worker 2 working on service_accounts task. I0423 19:50:39.024008 1 tasks_processing.go:69] worker 5 listening for tasks. I0423 19:50:39.024011 1 tasks_processing.go:71] worker 0 working on pod_network_connectivity_checks task. I0423 19:50:39.024014 1 tasks_processing.go:71] worker 5 working on image task. I0423 19:50:39.024021 1 tasks_processing.go:69] worker 48 listening for tasks. I0423 19:50:39.024025 1 tasks_processing.go:69] worker 46 listening for tasks. I0423 19:50:39.024028 1 tasks_processing.go:69] worker 49 listening for tasks. I0423 19:50:39.024015 1 tasks_processing.go:69] worker 47 listening for tasks. I0423 19:50:39.024037 1 tasks_processing.go:69] worker 51 listening for tasks. I0423 19:50:39.024024 1 tasks_processing.go:69] worker 14 listening for tasks. I0423 19:50:39.024032 1 tasks_processing.go:69] worker 50 listening for tasks. I0423 19:50:39.024037 1 tasks_processing.go:69] worker 15 listening for tasks. I0423 19:50:39.024045 1 tasks_processing.go:69] worker 38 listening for tasks. I0423 19:50:39.024038 1 tasks_processing.go:69] worker 37 listening for tasks. I0423 19:50:39.024048 1 tasks_processing.go:69] worker 16 listening for tasks. I0423 19:50:39.024050 1 tasks_processing.go:69] worker 58 listening for tasks. I0423 19:50:39.024056 1 tasks_processing.go:69] worker 17 listening for tasks. I0423 19:50:39.024050 1 tasks_processing.go:69] worker 57 listening for tasks. I0423 19:50:39.024063 1 tasks_processing.go:69] worker 53 listening for tasks. I0423 19:50:39.024055 1 tasks_processing.go:69] worker 59 listening for tasks. I0423 19:50:39.024058 1 tasks_processing.go:69] worker 52 listening for tasks. I0423 19:50:39.024059 1 tasks_processing.go:69] worker 60 listening for tasks. I0423 19:50:39.024075 1 tasks_processing.go:69] worker 63 listening for tasks. I0423 19:50:39.024005 1 tasks_processing.go:71] worker 13 working on networks task. I0423 19:50:39.024080 1 tasks_processing.go:69] worker 54 listening for tasks. I0423 19:50:39.024079 1 tasks_processing.go:69] worker 41 listening for tasks. I0423 19:50:39.024078 1 tasks_processing.go:69] worker 22 listening for tasks. I0423 19:50:39.024086 1 tasks_processing.go:69] worker 8 listening for tasks. I0423 19:50:39.024054 1 tasks_processing.go:69] worker 30 listening for tasks. I0423 19:50:39.024094 1 tasks_processing.go:69] worker 35 listening for tasks. I0423 19:50:39.024099 1 tasks_processing.go:69] worker 45 listening for tasks. I0423 19:50:39.023998 1 tasks_processing.go:69] worker 3 listening for tasks. I0423 19:50:39.024009 1 tasks_processing.go:69] worker 29 listening for tasks. I0423 19:50:39.024108 1 tasks_processing.go:69] worker 12 listening for tasks. I0423 19:50:39.024110 1 tasks_processing.go:69] worker 21 listening for tasks. I0423 19:50:39.024108 1 tasks_processing.go:69] worker 27 listening for tasks. I0423 19:50:39.024111 1 tasks_processing.go:69] worker 42 listening for tasks. I0423 19:50:39.024117 1 tasks_processing.go:69] worker 11 listening for tasks. I0423 19:50:39.024117 1 tasks_processing.go:69] worker 1 listening for tasks. I0423 19:50:39.024121 1 tasks_processing.go:71] worker 27 working on openshift_machine_api_events task. I0423 19:50:39.024125 1 tasks_processing.go:71] worker 45 working on config_maps task. I0423 19:50:39.024125 1 tasks_processing.go:71] worker 12 working on oauths task. I0423 19:50:39.024128 1 tasks_processing.go:71] worker 3 working on olm_operators task. I0423 19:50:39.024130 1 tasks_processing.go:71] worker 60 working on overlapping_namespace_uids task. I0423 19:50:39.024135 1 tasks_processing.go:71] worker 53 working on machines task. I0423 19:50:39.024136 1 tasks_processing.go:71] worker 59 working on cluster_apiserver task. I0423 19:50:39.024143 1 tasks_processing.go:71] worker 42 working on openstack_version task. I0423 19:50:39.024148 1 tasks_processing.go:71] worker 50 working on storage_cluster task. I0423 19:50:39.024154 1 tasks_processing.go:71] worker 48 working on validating_webhook_configurations task. I0423 19:50:39.024125 1 tasks_processing.go:71] worker 8 working on node_logs task. I0423 19:50:39.024195 1 tasks_processing.go:71] worker 35 working on active_alerts task. I0423 19:50:39.024143 1 tasks_processing.go:71] worker 51 working on storage_classes task. I0423 19:50:39.024219 1 tasks_processing.go:71] worker 46 working on clusterroles task. I0423 19:50:39.024376 1 tasks_processing.go:71] worker 11 working on sap_pods task. I0423 19:50:39.024012 1 tasks_processing.go:69] worker 28 listening for tasks. I0423 19:50:39.024860 1 tasks_processing.go:71] worker 28 working on container_runtime_configs task. I0423 19:50:39.024132 1 tasks_processing.go:71] worker 17 working on ingress task. I0423 19:50:39.024464 1 tasks_processing.go:71] worker 15 working on certificate_signing_requests task. I0423 19:50:39.024137 1 tasks_processing.go:71] worker 1 working on install_plans task. I0423 19:50:39.024484 1 tasks_processing.go:71] worker 49 working on machine_autoscalers task. I0423 19:50:39.024144 1 tasks_processing.go:71] worker 14 working on jaegers task. W0423 19:50:39.024220 1 gather_active_alerts.go:54] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 19:50:39.024072 1 tasks_processing.go:69] worker 32 listening for tasks. I0423 19:50:39.024070 1 tasks_processing.go:69] worker 40 listening for tasks. I0423 19:50:39.024074 1 tasks_processing.go:69] worker 23 listening for tasks. I0423 19:50:39.024078 1 tasks_processing.go:69] worker 7 listening for tasks. I0423 19:50:39.024079 1 tasks_processing.go:69] worker 33 listening for tasks. I0423 19:50:39.024561 1 tasks_processing.go:71] worker 63 working on openstack_dataplanenodesets task. I0423 19:50:39.025626 1 tasks_processing.go:71] worker 33 working on machine_sets task. I0423 19:50:39.024115 1 tasks_processing.go:69] worker 10 listening for tasks. I0423 19:50:39.024064 1 tasks_processing.go:69] worker 61 listening for tasks. I0423 19:50:39.024573 1 tasks_processing.go:71] worker 57 working on sap_config task. I0423 19:50:39.024087 1 tasks_processing.go:69] worker 55 listening for tasks. I0423 19:50:39.024086 1 tasks_processing.go:69] worker 34 listening for tasks. I0423 19:50:39.024091 1 tasks_processing.go:69] worker 44 listening for tasks. I0423 19:50:39.024070 1 tasks_processing.go:69] worker 62 listening for tasks. I0423 19:50:39.024091 1 tasks_processing.go:69] worker 25 listening for tasks. I0423 19:50:39.024092 1 tasks_processing.go:69] worker 56 listening for tasks. I0423 19:50:39.024070 1 tasks_processing.go:69] worker 6 listening for tasks. I0423 19:50:39.024100 1 tasks_processing.go:69] worker 26 listening for tasks. I0423 19:50:39.024100 1 tasks_processing.go:69] worker 9 listening for tasks. I0423 19:50:39.024102 1 tasks_processing.go:69] worker 20 listening for tasks. I0423 19:50:39.024097 1 tasks_processing.go:69] worker 43 listening for tasks. I0423 19:50:39.024103 1 tasks_processing.go:69] worker 18 listening for tasks. I0423 19:50:39.024107 1 tasks_processing.go:69] worker 19 listening for tasks. I0423 19:50:39.024118 1 tasks_processing.go:71] worker 52 working on number_of_pods_and_netnamespaces_with_sdn_annotations task. I0423 19:50:39.024119 1 tasks_processing.go:71] worker 21 working on pdbs task. I0423 19:50:39.024124 1 tasks_processing.go:71] worker 29 working on authentication task. I0423 19:50:39.024125 1 tasks_processing.go:71] worker 58 working on container_images task. I0423 19:50:39.024117 1 tasks_processing.go:69] worker 36 listening for tasks. I0423 19:50:39.024128 1 tasks_processing.go:71] worker 30 working on qemu_kubevirt_launcher_logs task. I0423 19:50:39.024131 1 tasks_processing.go:71] worker 16 working on version task. I0423 19:50:39.024003 1 tasks_processing.go:69] worker 4 listening for tasks. I0423 19:50:39.024061 1 tasks_processing.go:69] worker 31 listening for tasks. I0423 19:50:39.024710 1 tasks_processing.go:71] worker 37 working on node_features task. I0423 19:50:39.024525 1 tasks_processing.go:71] worker 47 working on metrics task. W0423 19:50:39.025869 1 gather_most_recent_metrics.go:64] Unable to load metrics client, no metrics will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 19:50:39.024082 1 tasks_processing.go:69] worker 24 listening for tasks. I0423 19:50:39.024535 1 tasks_processing.go:71] worker 54 working on tsdb_status task. I0423 19:50:39.024531 1 tasks_processing.go:71] worker 41 working on ingress_certificates task. W0423 19:50:39.026089 1 gather_prometheus_tsdb_status.go:38] Unable to load metrics client, tsdb status cannot be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 19:50:39.025527 1 tasks_processing.go:71] worker 35 working on machine_configs task. I0423 19:50:39.024064 1 tasks_processing.go:69] worker 39 listening for tasks. I0423 19:50:39.024672 1 tasks_processing.go:71] worker 38 working on sap_datahubs task. I0423 19:50:39.024683 1 tasks_processing.go:71] worker 22 working on aggregated_monitoring_cr_names task. I0423 19:50:39.026058 1 tasks_processing.go:71] worker 36 working on openshift_logging task. I0423 19:50:39.025754 1 gather.go:177] gatherer "clusterconfig" function "active_alerts" took 1.322575ms to process 0 records I0423 19:50:39.026318 1 tasks_processing.go:71] worker 47 working on image_registries task. I0423 19:50:39.026321 1 gather.go:177] gatherer "clusterconfig" function "metrics" took 238.585µs to process 0 records I0423 19:50:39.026339 1 gather.go:177] gatherer "clusterconfig" function "tsdb_status" took 25.289µs to process 0 records I0423 19:50:39.026353 1 tasks_processing.go:74] worker 54 stopped. I0423 19:50:39.026730 1 tasks_processing.go:71] worker 25 working on nodenetworkstates task. I0423 19:50:39.026762 1 tasks_processing.go:71] worker 24 working on infrastructures task. I0423 19:50:39.026772 1 tasks_processing.go:71] worker 6 working on crds task. I0423 19:50:39.026772 1 tasks_processing.go:71] worker 19 working on mutating_webhook_configurations task. I0423 19:50:39.026785 1 tasks_processing.go:71] worker 23 working on operators task. I0423 19:50:39.026797 1 tasks_processing.go:71] worker 43 working on dvo_metrics task. I0423 19:50:39.026807 1 tasks_processing.go:71] worker 10 working on ceph_cluster task. I0423 19:50:39.026809 1 tasks_processing.go:71] worker 34 working on proxies task. I0423 19:50:39.026827 1 tasks_processing.go:71] worker 20 working on machine_config_pools task. I0423 19:50:39.026850 1 tasks_processing.go:71] worker 56 working on operators_pods_and_events task. I0423 19:50:39.026889 1 tasks_processing.go:71] worker 18 working on image_pruners task. I0423 19:50:39.026762 1 tasks_processing.go:71] worker 4 working on schedulers task. I0423 19:50:39.026933 1 tasks_processing.go:71] worker 32 working on openstack_controlplanes task. I0423 19:50:39.026970 1 tasks_processing.go:71] worker 7 working on openstack_dataplanedeployments task. I0423 19:50:39.026768 1 tasks_processing.go:71] worker 31 working on silenced_alerts task. I0423 19:50:39.027009 1 tasks_processing.go:71] worker 40 working on cost_management_metrics_configs task. W0423 19:50:39.027024 1 gather_silenced_alerts.go:38] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 19:50:39.027033 1 tasks_processing.go:71] worker 61 working on feature_gates task. I0423 19:50:39.027037 1 tasks_processing.go:74] worker 31 stopped. I0423 19:50:39.027047 1 gather.go:177] gatherer "clusterconfig" function "silenced_alerts" took 40.427µs to process 0 records I0423 19:50:39.027102 1 tasks_processing.go:71] worker 55 working on machine_healthchecks task. I0423 19:50:39.027178 1 tasks_processing.go:71] worker 62 working on support_secret task. I0423 19:50:39.027215 1 tasks_processing.go:71] worker 44 working on monitoring_persistent_volumes task. I0423 19:50:39.027249 1 tasks_processing.go:71] worker 26 working on nodenetworkconfigurationpolicies task. I0423 19:50:39.027329 1 tasks_processing.go:71] worker 9 working on nodes task. I0423 19:50:39.027403 1 tasks_processing.go:71] worker 39 working on lokistack task. I0423 19:50:39.027669 1 tasks_processing.go:74] worker 42 stopped. I0423 19:50:39.027692 1 gather.go:177] gatherer "clusterconfig" function "openstack_version" took 3.514558ms to process 0 records E0423 19:50:39.027712 1 gather.go:140] gatherer "clusterconfig" function "pod_network_connectivity_checks" failed with the error: the server could not find the requested resource (get podnetworkconnectivitychecks.controlplane.operator.openshift.io) I0423 19:50:39.027721 1 gather.go:177] gatherer "clusterconfig" function "pod_network_connectivity_checks" took 3.659287ms to process 0 records I0423 19:50:39.027729 1 tasks_processing.go:74] worker 0 stopped. I0423 19:50:39.028395 1 tasks_processing.go:74] worker 50 stopped. I0423 19:50:39.028407 1 gather.go:177] gatherer "clusterconfig" function "storage_cluster" took 4.236978ms to process 0 records I0423 19:50:39.029342 1 controller.go:129] Initializing last reported time to 0001-01-01T00:00:00Z I0423 19:50:39.029357 1 controller.go:254] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0423 19:50:39.029361 1 controller.go:254] Source periodic-conditional *controllerstatus.Simple is not ready I0423 19:50:39.029364 1 controller.go:254] Source periodic-workloads *controllerstatus.Simple is not ready I0423 19:50:39.029379 1 controller.go:531] The operator is still being initialized I0423 19:50:39.029386 1 controller.go:554] The operator is healthy I0423 19:50:39.032491 1 tasks_processing.go:74] worker 28 stopped. I0423 19:50:39.032503 1 gather.go:177] gatherer "clusterconfig" function "container_runtime_configs" took 7.59555ms to process 0 records I0423 19:50:39.032524 1 tasks_processing.go:74] worker 11 stopped. I0423 19:50:39.032537 1 gather.go:177] gatherer "clusterconfig" function "sap_pods" took 7.784864ms to process 0 records I0423 19:50:39.032542 1 gather.go:177] gatherer "clusterconfig" function "jaegers" took 6.984633ms to process 0 records I0423 19:50:39.032548 1 tasks_processing.go:74] worker 14 stopped. I0423 19:50:39.032717 1 tasks_processing.go:74] worker 59 stopped. I0423 19:50:39.032858 1 recorder.go:75] Recording config/apiserver with fingerprint=5fe140368858fc8fac681a7aab11382152bf866406f8ca1235355bdd62bfae70 I0423 19:50:39.032871 1 gather.go:177] gatherer "clusterconfig" function "cluster_apiserver" took 8.566409ms to process 1 records I0423 19:50:39.034378 1 tasks_processing.go:74] worker 51 stopped. I0423 19:50:39.034453 1 recorder.go:75] Recording config/storage/storageclasses/gp2-csi with fingerprint=85a7c57e913ccc356f18c30088e7f11b43a20abcb3dfe3717272bae5e85052ac I0423 19:50:39.034470 1 recorder.go:75] Recording config/storage/storageclasses/gp3-csi with fingerprint=ac85d6685d0c305919cb9f51d2e0d404ba6e231897654900e7e8e5c9d1550798 I0423 19:50:39.034478 1 gather.go:177] gatherer "clusterconfig" function "storage_classes" took 10.15129ms to process 2 records I0423 19:50:39.034489 1 gather.go:177] gatherer "clusterconfig" function "openshift_machine_api_events" took 10.246177ms to process 0 records E0423 19:50:39.034494 1 gather.go:140] gatherer "clusterconfig" function "machines" failed with the error: machines.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machines" in API group "machine.openshift.io" at the cluster scope I0423 19:50:39.034501 1 gather.go:177] gatherer "clusterconfig" function "machines" took 10.281103ms to process 0 records I0423 19:50:39.034507 1 tasks_processing.go:74] worker 53 stopped. I0423 19:50:39.034511 1 tasks_processing.go:74] worker 27 stopped. I0423 19:50:39.036121 1 tasks_processing.go:74] worker 63 stopped. I0423 19:50:39.036131 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanenodesets" took 10.505064ms to process 0 records I0423 19:50:39.036139 1 gather.go:177] gatherer "clusterconfig" function "machine_autoscalers" took 10.670718ms to process 0 records I0423 19:50:39.036146 1 tasks_processing.go:74] worker 49 stopped. I0423 19:50:39.036152 1 tasks_processing.go:74] worker 36 stopped. I0423 19:50:39.036163 1 gather.go:177] gatherer "clusterconfig" function "openshift_logging" took 9.876459ms to process 0 records I0423 19:50:39.036332 1 tasks_processing.go:74] worker 17 stopped. I0423 19:50:39.036539 1 recorder.go:75] Recording config/ingress with fingerprint=26fcc93e3b4a4b79158d92195feb6031328ea54bebbbbcacbaa7bb8b4a3e41ba I0423 19:50:39.036557 1 gather.go:177] gatherer "clusterconfig" function "ingress" took 11.254298ms to process 1 records I0423 19:50:39.036659 1 tasks_processing.go:74] worker 12 stopped. I0423 19:50:39.036849 1 recorder.go:75] Recording config/oauth with fingerprint=c95d8cb9eedb289c98e13fbdb12bb13e3a5c8c7c8300c7d0f1e3bab3ab72ab19 I0423 19:50:39.036863 1 gather.go:177] gatherer "clusterconfig" function "oauths" took 12.212401ms to process 1 records I0423 19:50:39.036947 1 tasks_processing.go:74] worker 48 stopped. I0423 19:50:39.037071 1 recorder.go:75] Recording config/validatingwebhookconfigurations/multus.openshift.io with fingerprint=d2f1f7cdc7f3f03fa386ed25f0dbf259df59e13ba44a212a590124880780675c I0423 19:50:39.037153 1 recorder.go:75] Recording config/validatingwebhookconfigurations/network-node-identity.openshift.io with fingerprint=6ec413692e68fb0bc2fb1a469f545779f414c0cf135e312695f590670231ca65 I0423 19:50:39.037172 1 recorder.go:75] Recording config/validatingwebhookconfigurations/performance-addon-operator with fingerprint=20b4e563c470c0de3754a4d6be27e2cc4aee0baed4450d50bed2443584601f4d I0423 19:50:39.037201 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterrolebindings-validation with fingerprint=d7e967bb404fa84d4104f83334b4b0932c2ad5d9c359bb16642e8b9088d979d1 I0423 19:50:39.037232 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterroles-validation with fingerprint=bcc99011bd949a43cc58f229f70fe4d40f988399c830ce4003f747edf03fc0c4 I0423 19:50:39.037261 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-ingress-config-validation with fingerprint=54574d18e1785853917161b489658642d7124027b2b5cde462906710b764be79 I0423 19:50:39.037289 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-network-operator-validation with fingerprint=1f48e8ac54d9682c5dabb11f4d6e51b934a9bcaf41f8f454e783bcddff9e802a I0423 19:50:39.037334 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-regular-user-validation with fingerprint=02332ffe8632fc63e56f846e10536599d0d73330266ab0e173e2ffe9574f34dd I0423 19:50:39.037363 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-scc-validation with fingerprint=b8e4b869fa8b85a57dcef46764ecde1e7d4af46d5e1c7d8394068956599fa6b8 I0423 19:50:39.037393 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-serviceaccount-validation with fingerprint=f2f48ee7f9c25a73f22f0aa851d27e85eb7b3c3023803dadf0ae934c9050370e I0423 19:50:39.037424 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-techpreviewnoupgrade-validation with fingerprint=51adaad3c18400c1ecb1221499cbd5c7bbd4e6664a444d2edcf6c03f7c1e45ad I0423 19:50:39.037431 1 gather.go:177] gatherer "clusterconfig" function "validating_webhook_configurations" took 12.392319ms to process 11 records I0423 19:50:39.048007 1 tasks_processing.go:74] worker 38 stopped. I0423 19:50:39.048021 1 gather.go:177] gatherer "clusterconfig" function "sap_datahubs" took 21.784867ms to process 0 records I0423 19:50:39.048087 1 tasks_processing.go:74] worker 33 stopped. I0423 19:50:39.048098 1 gather.go:177] gatherer "clusterconfig" function "machine_sets" took 22.417682ms to process 0 records I0423 19:50:39.048332 1 tasks_processing.go:74] worker 13 stopped. I0423 19:50:39.048448 1 recorder.go:75] Recording config/network with fingerprint=213ec1a7f631aa6bd42bdbee1b8017af002bd405139c8a59d3ab665f18ba37fe I0423 19:50:39.048459 1 gather.go:177] gatherer "clusterconfig" function "networks" took 24.246963ms to process 1 records I0423 19:50:39.049440 1 tasks_processing.go:74] worker 8 stopped. I0423 19:50:39.049471 1 gather.go:177] gatherer "clusterconfig" function "node_logs" took 25.244723ms to process 0 records I0423 19:50:39.049959 1 tasks_processing.go:74] worker 21 stopped. I0423 19:50:39.050603 1 recorder.go:75] Recording config/pdbs/openshift-image-registry/image-registry with fingerprint=944ddc85923947d9735ea89cf46492b708b5021246b9235a03cbfd9c32a6810e I0423 19:50:39.050980 1 recorder.go:75] Recording config/pdbs/openshift-ingress/router-default with fingerprint=c8a53fcbf9c1c9cb2a95fda70a544aa9045bbe808186d472defeafa2c60aca10 I0423 19:50:39.051012 1 recorder.go:75] Recording config/pdbs/openshift-operator-lifecycle-manager/packageserver-pdb with fingerprint=90055904db729cc051cf686693f3c32bd658cc9596394644b70ccb74dd45b1cb I0423 19:50:39.051022 1 gather.go:177] gatherer "clusterconfig" function "pdbs" took 24.047076ms to process 3 records I0423 19:50:39.051099 1 recorder.go:75] Recording config/image with fingerprint=66819b554da8c9a80408bc1315e60f49daa6ed2c542c579b9afe828d23e1846f I0423 19:50:39.051108 1 gather.go:177] gatherer "clusterconfig" function "image" took 26.23686ms to process 1 records I0423 19:50:39.051119 1 tasks_processing.go:74] worker 5 stopped. I0423 19:50:39.051978 1 tasks_processing.go:74] worker 25 stopped. I0423 19:50:39.051997 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkstates" took 25.236005ms to process 0 records I0423 19:50:39.052013 1 gather.go:177] gatherer "clusterconfig" function "sap_config" took 26.201741ms to process 0 records I0423 19:50:39.052020 1 gather.go:177] gatherer "clusterconfig" function "machine_config_pools" took 25.15943ms to process 0 records I0423 19:50:39.052025 1 gather.go:177] gatherer "clusterconfig" function "ceph_cluster" took 25.208417ms to process 0 records I0423 19:50:39.052027 1 tasks_processing.go:74] worker 57 stopped. I0423 19:50:39.052028 1 tasks_processing.go:74] worker 20 stopped. I0423 19:50:39.052032 1 tasks_processing.go:74] worker 26 stopped. I0423 19:50:39.052029 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkconfigurationpolicies" took 24.770569ms to process 0 records I0423 19:50:39.052036 1 tasks_processing.go:74] worker 10 stopped. E0423 19:50:39.052043 1 gather.go:140] gatherer "clusterconfig" function "machine_healthchecks" failed with the error: machinehealthchecks.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machinehealthchecks" in API group "machine.openshift.io" at the cluster scope I0423 19:50:39.052050 1 gather.go:177] gatherer "clusterconfig" function "machine_healthchecks" took 24.928488ms to process 0 records I0423 19:50:39.052057 1 tasks_processing.go:74] worker 55 stopped. E0423 19:50:39.052286 1 gather_node_features.go:86] GatherNodeFeatures: NodeFeatures resource not found in openshift-nfd namespace (may not be installed) I0423 19:50:39.052299 1 tasks_processing.go:74] worker 37 stopped. I0423 19:50:39.052306 1 gather.go:177] gatherer "clusterconfig" function "node_features" took 26.387442ms to process 0 records I0423 19:50:39.052329 1 tasks_processing.go:74] worker 39 stopped. I0423 19:50:39.052338 1 gather.go:177] gatherer "clusterconfig" function "lokistack" took 24.916328ms to process 0 records I0423 19:50:39.052386 1 tasks_processing.go:74] worker 40 stopped. I0423 19:50:39.052398 1 gather.go:177] gatherer "clusterconfig" function "cost_management_metrics_configs" took 25.363557ms to process 0 records I0423 19:50:39.052425 1 tasks_processing.go:74] worker 7 stopped. I0423 19:50:39.052440 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanedeployments" took 25.444197ms to process 0 records I0423 19:50:39.052576 1 tasks_processing.go:74] worker 62 stopped. E0423 19:50:39.052588 1 gather.go:140] gatherer "clusterconfig" function "support_secret" failed with the error: secrets "support" not found I0423 19:50:39.052598 1 gather.go:177] gatherer "clusterconfig" function "support_secret" took 25.371571ms to process 0 records I0423 19:50:39.052605 1 gather.go:177] gatherer "clusterconfig" function "openstack_controlplanes" took 25.637744ms to process 0 records I0423 19:50:39.052610 1 tasks_processing.go:74] worker 32 stopped. I0423 19:50:39.052622 1 tasks_processing.go:74] worker 60 stopped. I0423 19:50:39.052667 1 recorder.go:75] Recording config/namespaces_with_overlapping_uids with fingerprint=4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945 I0423 19:50:39.052679 1 gather.go:177] gatherer "clusterconfig" function "overlapping_namespace_uids" took 28.48045ms to process 1 records I0423 19:50:39.052845 1 recorder.go:75] Recording config/authentication with fingerprint=c3f641982dc86eda1a50f0b5fca514253c6c5ee0af2015d11021be1d2f94c6c5 I0423 19:50:39.052856 1 gather.go:177] gatherer "clusterconfig" function "authentication" took 26.727224ms to process 1 records I0423 19:50:39.052862 1 gather.go:177] gatherer "clusterconfig" function "monitoring_persistent_volumes" took 25.586274ms to process 0 records I0423 19:50:39.052868 1 tasks_processing.go:74] worker 44 stopped. I0423 19:50:39.052878 1 tasks_processing.go:74] worker 29 stopped. I0423 19:50:39.053070 1 tasks_processing.go:74] worker 4 stopped. I0423 19:50:39.053124 1 recorder.go:75] Recording config/schedulers/cluster with fingerprint=ec76e36d70a3448dc900538260a351ee92ff72d372ac30b6a18c51629f2bc855 I0423 19:50:39.053134 1 gather.go:177] gatherer "clusterconfig" function "schedulers" took 26.143382ms to process 1 records I0423 19:50:39.054709 1 tasks_processing.go:74] worker 15 stopped. I0423 19:50:39.054722 1 gather.go:177] gatherer "clusterconfig" function "certificate_signing_requests" took 29.448541ms to process 0 records W0423 19:50:39.057214 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 19:50:39.062713 1 gather_logs.go:145] no pods in namespace were found I0423 19:50:39.062731 1 tasks_processing.go:74] worker 30 stopped. I0423 19:50:39.062744 1 gather.go:177] gatherer "clusterconfig" function "qemu_kubevirt_launcher_logs" took 36.775139ms to process 0 records I0423 19:50:39.063008 1 tasks_processing.go:74] worker 24 stopped. W0423 19:50:39.063286 1 operator.go:328] started I0423 19:50:39.063361 1 base_controller.go:76] Waiting for caches to sync for LoggingSyncer I0423 19:50:39.063496 1 sca.go:136] Pulling SCA certificates from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates. Next check is in 8h0m0s I0423 19:50:39.063638 1 cluster_transfer.go:83] checking the availability of cluster transfer. Next check is in 12h0m0s I0423 19:50:39.064511 1 recorder.go:75] Recording config/infrastructure with fingerprint=f476dcf555724fc9338683014002002884012bbff208e850554cc4a17b5e497c I0423 19:50:39.064536 1 gather.go:177] gatherer "clusterconfig" function "infrastructures" took 36.230916ms to process 1 records I0423 19:50:39.064673 1 tasks_processing.go:74] worker 19 stopped. I0423 19:50:39.065147 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/aws-pod-identity with fingerprint=7ffe0714e1fb274c98ac7080faed0db699b40414581a3bbadac128c776773c25 I0423 19:50:39.065326 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-podimagespec-mutation with fingerprint=08007b9bfc28ad3c05910320c9376d9a50d4581702c098200f47a6b98ab4d6dd I0423 19:50:39.065449 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-service-mutation with fingerprint=75e8ba33c3bd31ef769ab869c679ea97208effecbf79f05d4988711283464ff9 I0423 19:50:39.065461 1 gather.go:177] gatherer "clusterconfig" function "mutating_webhook_configurations" took 36.236248ms to process 3 records E0423 19:50:39.065474 1 gather.go:140] gatherer "clusterconfig" function "machine_configs" failed with the error: getting MachineConfigPools failed: the server could not find the requested resource (get machineconfigpools.machineconfiguration.openshift.io) I0423 19:50:39.065546 1 tasks_processing.go:74] worker 35 stopped. I0423 19:50:39.065623 1 recorder.go:75] Recording aggregated/unused_machine_configs_count with fingerprint=4bfc9fa984e5dfcd45848faaf05269de7619bf42edf9f781751af5ee05c1a499 I0423 19:50:39.065661 1 gather.go:177] gatherer "clusterconfig" function "machine_configs" took 37.776102ms to process 1 records I0423 19:50:39.065741 1 tasks_processing.go:74] worker 46 stopped. I0423 19:50:39.067290 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/admin with fingerprint=b95936968fd25dfb0b8df5e53f1ac794601bdead5e31dc3a87388c6e6a5e80f0 I0423 19:50:39.067408 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/edit with fingerprint=398977fdb9c2a1697decfe520321ca179eb490dfbfad7fdc027beeeb264ca052 I0423 19:50:39.067420 1 gather.go:177] gatherer "clusterconfig" function "clusterroles" took 39.301843ms to process 2 records I0423 19:50:39.067489 1 recorder.go:75] Recording config/proxy with fingerprint=46e27cff81e15d23d932f7b49a470ce760458e65e1cb6d7169ad29b63f9e3945 I0423 19:50:39.067498 1 gather.go:177] gatherer "clusterconfig" function "proxies" took 37.359715ms to process 1 records I0423 19:50:39.067509 1 tasks_processing.go:74] worker 34 stopped. I0423 19:50:39.067578 1 recorder.go:75] Recording config/featuregate with fingerprint=bd74b4e6ab1d21ce3a39d2c5ee67bae37dba836efb77d7eff7db9de7b02b08d1 I0423 19:50:39.067588 1 gather.go:177] gatherer "clusterconfig" function "feature_gates" took 37.239736ms to process 1 records I0423 19:50:39.067593 1 tasks_processing.go:74] worker 61 stopped. I0423 19:50:39.067636 1 recorder.go:75] Recording config/olm_operators with fingerprint=ef7694193e237f6f3b4f07f3aaed0489caebe4a38ea4c9ce409842a60cce8edf I0423 19:50:39.067659 1 gather.go:177] gatherer "clusterconfig" function "olm_operators" took 41.143148ms to process 1 records I0423 19:50:39.067663 1 tasks_processing.go:74] worker 3 stopped. I0423 19:50:39.067736 1 tasks_processing.go:74] worker 18 stopped. I0423 19:50:39.067844 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/imagepruner/cluster with fingerprint=d6afab1fb63df86f0400c3c6554ee46343281718226bc6c5f6c35338a2ec411e I0423 19:50:39.067855 1 gather.go:177] gatherer "clusterconfig" function "image_pruners" took 38.913931ms to process 1 records I0423 19:50:39.067935 1 tasks_processing.go:74] worker 47 stopped. I0423 19:50:39.068220 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/config/cluster with fingerprint=0b68b7b31af8c3e42b6e52d781f504fc990f7d72666f3f0f067b54ce2f74182b I0423 19:50:39.068232 1 gather.go:177] gatherer "clusterconfig" function "image_registries" took 39.858958ms to process 1 records I0423 19:50:39.070817 1 tasks_processing.go:74] worker 58 stopped. I0423 19:50:39.071966 1 recorder.go:75] Recording config/pod/openshift-console-operator/console-operator-66dbbb6468-vvdg4 with fingerprint=fc38295172c923006b3c42fd980f7ec81a35e2744dd79230e6cbef126217d01c I0423 19:50:39.072017 1 recorder.go:75] Recording config/running_containers with fingerprint=657eeda3ab8afaad742d51077bb3b0f216964ca0094d68bf7a746f49f393e283 I0423 19:50:39.072032 1 gather.go:177] gatherer "clusterconfig" function "container_images" took 44.973864ms to process 2 records I0423 19:50:39.078981 1 tasks_processing.go:74] worker 52 stopped. I0423 19:50:39.079007 1 gather.go:177] gatherer "clusterconfig" function "number_of_pods_and_netnamespaces_with_sdn_annotations" took 53.144626ms to process 0 records I0423 19:50:39.086286 1 prometheus_rules.go:88] Prometheus rules successfully created I0423 19:50:39.091378 1 tasks_processing.go:74] worker 6 stopped. I0423 19:50:39.091581 1 controller.go:254] Source scaController *sca.Controller is not ready I0423 19:50:39.091595 1 controller.go:254] Source clusterTransferController *clustertransfer.Controller is not ready I0423 19:50:39.091600 1 controller.go:254] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0423 19:50:39.091603 1 controller.go:254] Source periodic-conditional *controllerstatus.Simple is not ready I0423 19:50:39.091605 1 controller.go:254] Source periodic-workloads *controllerstatus.Simple is not ready I0423 19:50:39.091620 1 controller.go:531] The operator is still being initialized I0423 19:50:39.091628 1 controller.go:554] The operator is healthy I0423 19:50:39.092113 1 recorder.go:75] Recording config/crd/volumesnapshots.snapshot.storage.k8s.io with fingerprint=a5377b263a81c97ea4d565adc3ad783374bb863c44dbca08c4c0b6acc5821f78 I0423 19:50:39.092426 1 recorder.go:75] Recording config/crd/volumesnapshotcontents.snapshot.storage.k8s.io with fingerprint=a8822d2fbf76e36157bfda1ed325f546b48f7a430f091be5cffb7bdc74fa0e61 I0423 19:50:39.092441 1 gather.go:177] gatherer "clusterconfig" function "crds" took 64.593034ms to process 2 records E0423 19:50:39.093309 1 cluster_transfer.go:95] failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%2778e19fa4-9938-47b8-b984-77864a4a09ea%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:60157->172.30.0.10:53: read: connection refused I0423 19:50:39.093320 1 controllerstatus.go:80] name=clusterTransferController healthy=true reason=Disconnected message=failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%2778e19fa4-9938-47b8-b984-77864a4a09ea%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:60157->172.30.0.10:53: read: connection refused I0423 19:50:39.095408 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0423 19:50:39.095483 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0423 19:50:39.095565 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController I0423 19:50:39.096412 1 tasks_processing.go:74] worker 22 stopped. I0423 19:50:39.096425 1 gather.go:177] gatherer "clusterconfig" function "aggregated_monitoring_cr_names" took 70.149895ms to process 0 records I0423 19:50:39.098278 1 base_controller.go:82] Caches are synced for ConfigController I0423 19:50:39.098290 1 base_controller.go:119] Starting #1 worker of ConfigController controller ... I0423 19:50:39.114778 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 19:50:39.115316 1 tasks_processing.go:74] worker 45 stopped. E0423 19:50:39.115328 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "cluster-monitoring-config" not found E0423 19:50:39.115336 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "gateway-mode-config" not found E0423 19:50:39.115340 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "insights-config" not found I0423 19:50:39.115348 1 recorder.go:75] Recording config/configmaps/openshift-config/installer-images/images.json with fingerprint=ad286723f58bdcfc37aeba3ec5b4110c08e1af59cd34d14b4bfaab02d18e4856 I0423 19:50:39.115382 1 recorder.go:75] Recording config/configmaps/openshift-config/kube-root-ca.crt/ca.crt with fingerprint=d476c7d3f5b104863f08f481b1264dcc68cc272ecefb0ecb709b18a6afab034d I0423 19:50:39.115389 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/invoker with fingerprint=76b482f683cd3ef9da02debac5b26080a5aeb06ff768ee5c21117514dff29d8a I0423 19:50:39.115393 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/version with fingerprint=2ab5ab7b1b10d7fcf1197bb24dea7c90f400e4effc18c7356873209d54fdf84b I0423 19:50:39.115397 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 I0423 19:50:39.115435 1 recorder.go:75] Recording config/configmaps/openshift-config/rosa-brand-logo/rosa-brand-logo.svg with fingerprint=6ed8ca4dd7a8eee7249182bc006e9649ce84d76c551ddfaaa33e55d8c4cc1ed0 I0423 19:50:39.115442 1 recorder.go:75] Recording config/configmaps/kube-system/cluster-config-v1/install-config with fingerprint=ab3811c6b83fd7b8e920094cfa3080d1b4ee3c35ec4c8379437b21d27bd6608d I0423 19:50:39.115448 1 gather.go:177] gatherer "clusterconfig" function "config_maps" took 91.177245ms to process 7 records I0423 19:50:39.142254 1 tasks_processing.go:74] worker 9 stopped. I0423 19:50:39.142521 1 recorder.go:75] Recording config/node/ip-10-0-0-80.ec2.internal with fingerprint=a391cf60681d1a19e649f1b29384db1a2dd5d464eb4e42c3d74aef13c4d13871 I0423 19:50:39.142574 1 recorder.go:75] Recording config/node/ip-10-0-1-141.ec2.internal with fingerprint=139a2046c82aec4db11dec5327bbebb5ca6f46de2b54fed2d2da1a6a369d4e5a I0423 19:50:39.142623 1 recorder.go:75] Recording config/node/ip-10-0-2-131.ec2.internal with fingerprint=ee775db389d548e864b6f39dca6be2340aaa1303d98573fcbd83b837370cc743 I0423 19:50:39.142631 1 gather.go:177] gatherer "clusterconfig" function "nodes" took 114.905941ms to process 3 records I0423 19:50:39.142724 1 tasks_processing.go:74] worker 16 stopped. I0423 19:50:39.142839 1 recorder.go:75] Recording config/version with fingerprint=7b3b2dbeb4b2a654d584b64cd0f29e4b993afa35be0035bd2fdc3c8f814a72df I0423 19:50:39.142849 1 recorder.go:75] Recording config/id with fingerprint=16437d5a0a24f4e8047b80ea80080def56351953a42ee5067a99147967a607db I0423 19:50:39.142855 1 gather.go:177] gatherer "clusterconfig" function "version" took 116.642778ms to process 2 records I0423 19:50:39.161978 1 requests.go:205] Asking for SCA certificate with "{"arch": ["x86_64"]}" payload I0423 19:50:39.163411 1 base_controller.go:82] Caches are synced for LoggingSyncer I0423 19:50:39.163420 1 base_controller.go:119] Starting #1 worker of LoggingSyncer controller ... W0423 19:50:39.165029 1 sca.go:161] Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:52232->172.30.0.10:53: read: connection refused I0423 19:50:39.165041 1 controllerstatus.go:80] name=scaController healthy=true reason=NonHTTPError message=Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.129.0.10:52232->172.30.0.10:53: read: connection refused I0423 19:50:39.201268 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0423 19:50:39.201276 1 tasks_processing.go:74] worker 41 stopped. E0423 19:50:39.201292 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret 'router-certs-default' in namespace 'openshift-ingress': secrets "router-certs-default" not found E0423 19:50:39.201303 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret '2ps2bve4hjh3ovu4l3oim6c2t962kfba-primary-cert-bundle-secret' in namespace 'openshift-ingress-operator': secrets "2ps2bve4hjh3ovu4l3oim6c2t962kfba-primary-cert-bundle-secret" not found I0423 19:50:39.201362 1 recorder.go:75] Recording aggregated/ingress_controllers_certs with fingerprint=125b7556369f1986e59380082449929884981a1c9caf6a79906ea31160276c0c I0423 19:50:39.201376 1 gather.go:177] gatherer "clusterconfig" function "ingress_certificates" took 175.170226ms to process 1 records I0423 19:50:39.219589 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0423 19:50:39.477038 1 gather_cluster_operator_pods_and_events.go:121] Found 18 pods with 21 containers I0423 19:50:39.477052 1 gather_cluster_operator_pods_and_events.go:235] Maximum buffer size: 1198372 bytes I0423 19:50:39.477806 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-7jqn9 pod in namespace openshift-dns (previous: false). I0423 19:50:39.682591 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-7jqn9 pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-7jqn9\" is waiting to start: ContainerCreating" I0423 19:50:39.682609 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-7jqn9\" is waiting to start: ContainerCreating" I0423 19:50:39.682617 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-7jqn9 pod in namespace openshift-dns (previous: false). I0423 19:50:39.885482 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-7jqn9 pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-7jqn9\" is waiting to start: ContainerCreating" I0423 19:50:39.885501 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-7jqn9\" is waiting to start: ContainerCreating" I0423 19:50:39.885528 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-7qsv8 pod in namespace openshift-dns (previous: false). W0423 19:50:40.056258 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 19:50:40.158160 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-7qsv8 pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-7qsv8\" is waiting to start: ContainerCreating" I0423 19:50:40.158177 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-7qsv8\" is waiting to start: ContainerCreating" I0423 19:50:40.158185 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-7qsv8 pod in namespace openshift-dns (previous: false). I0423 19:50:40.283888 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-7qsv8 pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-7qsv8\" is waiting to start: ContainerCreating" I0423 19:50:40.283904 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-7qsv8\" is waiting to start: ContainerCreating" I0423 19:50:40.283931 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-cfmp6 pod in namespace openshift-dns (previous: false). I0423 19:50:40.501961 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-cfmp6 pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-cfmp6\" is waiting to start: ContainerCreating" I0423 19:50:40.501976 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-cfmp6\" is waiting to start: ContainerCreating" I0423 19:50:40.501985 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-cfmp6 pod in namespace openshift-dns (previous: false). I0423 19:50:40.684114 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-cfmp6 pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-cfmp6\" is waiting to start: ContainerCreating" I0423 19:50:40.684127 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-cfmp6\" is waiting to start: ContainerCreating" I0423 19:50:40.684138 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-kzbhx pod in namespace openshift-dns (previous: false). I0423 19:50:40.694579 1 gather_cluster_operators.go:184] Unable to get configs.samples.operator.openshift.io resource due to: configs.samples.operator.openshift.io "cluster" not found I0423 19:50:40.881993 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 19:50:40.882007 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-qrxwr pod in namespace openshift-dns (previous: false). W0423 19:50:41.056872 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 19:50:41.086975 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 19:50:41.086990 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-vlq7n pod in namespace openshift-dns (previous: false). I0423 19:50:41.095589 1 tasks_processing.go:74] worker 23 stopped. I0423 19:50:41.095635 1 recorder.go:75] Recording config/clusteroperator/console with fingerprint=69f4e539df17d3c090b86ef4dc0972eb8f403a193271f629210820dbb90cb6c6 I0423 19:50:41.095699 1 recorder.go:75] Recording config/clusteroperator/csi-snapshot-controller with fingerprint=81415ededb06c947236e9026f363db8a6ce285ccc8c05adeac52ef5f57f40cf3 I0423 19:50:41.095729 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/csisnapshotcontroller/cluster with fingerprint=5adc514f4b63e2f1ecc68bf6f9c0af70c5eea04522a49524e102721b1c41f80e I0423 19:50:41.095756 1 recorder.go:75] Recording config/clusteroperator/dns with fingerprint=43ca073a91d510ba37a27d7e9672a23563e84ea595025930f1e83af46af1c563 I0423 19:50:41.095788 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/dns/default with fingerprint=9e7b4ce029030d3d8c3b49af92c556acdcc415000b40d3f969dbdc42c432b47f I0423 19:50:41.095811 1 recorder.go:75] Recording config/clusteroperator/image-registry with fingerprint=8b2aa9711bd5a57ae202d99fb9414bcd591bae01a6c525625f77bc28a40ed018 I0423 19:50:41.095842 1 recorder.go:75] Recording config/clusteroperator/ingress with fingerprint=013295356e9e461163bfd70897b557976efcbc8e658c71988821dd4d0376dcbc I0423 19:50:41.095865 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/ingresscontroller/openshift-ingress-operator/default with fingerprint=86318544acadb12fbe6d93324617d21c4324004d1de52dd6a94bd9c9de0df2a7 I0423 19:50:41.095896 1 recorder.go:75] Recording config/clusteroperator/insights with fingerprint=67954f8c7c956d395d3ec233c5df50fd09fe4c6dccfb774c2c618fd3a22565a5 I0423 19:50:41.095906 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/insightsoperator/cluster with fingerprint=e5ff11d57817f84a678f6fa9565af55bd1120227c16a21933637ab62675a6d70 I0423 19:50:41.095935 1 recorder.go:75] Recording config/clusteroperator/kube-apiserver with fingerprint=1c62756760d96b39edb28b4d277afeae6f236512d8b8ee4b2808b2e93f42ab3b I0423 19:50:41.095944 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubeapiserver/cluster with fingerprint=51503bf0b784fcf65ea46bcaf1f72ac1a5c4d5dc211934f18f27871efed05762 I0423 19:50:41.095968 1 recorder.go:75] Recording config/clusteroperator/kube-controller-manager with fingerprint=9b01bfea1c0aa003facd1452197dea1ec8237b85cfce25a96b6ae9fa1cf74560 I0423 19:50:41.095979 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubecontrollermanager/cluster with fingerprint=ce90c0d4f367d7da085074268031798382ae7c54fdcb0a21f15a4818fe308c11 I0423 19:50:41.095994 1 recorder.go:75] Recording config/clusteroperator/kube-scheduler with fingerprint=9a2610548d78b21209755bfc026706e088817ee9e4713ddbc68e2662e4a0c3fc I0423 19:50:41.096003 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubescheduler/cluster with fingerprint=f2940fb9fd20c19951dfc295eb363b7fba0c505f5ae61f01967a063099e6b60a I0423 19:50:41.096016 1 recorder.go:75] Recording config/clusteroperator/kube-storage-version-migrator with fingerprint=b922936f2e685d4cf05edef9918ce0098a299e93175eec5bf0b2ed3fd788be97 I0423 19:50:41.096023 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubestorageversionmigrator/cluster with fingerprint=9351181aa7e6ada41ef581ab31e13516c6b934cc95710154bafb2eb222cb58db I0423 19:50:41.096041 1 recorder.go:75] Recording config/clusteroperator/monitoring with fingerprint=d71518555ddd89a6e78bc92e630c6a539bdf6cc35cc88d66e661612b96dd6568 I0423 19:50:41.096339 1 recorder.go:75] Recording config/clusteroperator/network with fingerprint=d763179433d15a9d6c6e85793866bd08ba45ddb9fc237645c96e2ffea6cd64c4 I0423 19:50:41.096357 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/ovn with fingerprint=626a89d20e0deaed5b6dfb533acfe65f4bb1618bd200a703b62e60c5d16d94ab I0423 19:50:41.096366 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/signer with fingerprint=90410b16914712b85b3c4578716ad8c0ae072e688f4cd1e022bf76f20da3506d I0423 19:50:41.096409 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/network/cluster with fingerprint=a93d15eaecb455a0e40ecb2826eeecc1533899204ddd3c3921d15ab70af7ae75 I0423 19:50:41.096587 1 recorder.go:75] Recording config/clusteroperator/node-tuning with fingerprint=dca963ec17532ce8dc576ebb5e6d517197c475606d5b2d18ceef41d39a7f3840 I0423 19:50:41.096749 1 recorder.go:75] Recording config/clusteroperator/openshift-apiserver with fingerprint=cd5128a6d73c45486568489404a71c46519edf0358b48c7ae2bb958945c3fb0d I0423 19:50:41.096809 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftapiserver/cluster with fingerprint=e712e6cf27339b441e4ed1f4cde91dbde7e952698ba93407e4457db63a4a4c76 I0423 19:50:41.096871 1 recorder.go:75] Recording config/clusteroperator/openshift-controller-manager with fingerprint=9f61a6a96b9fc638328288a28e54f5784fa451a253d05be215164ac8eca0a1d9 I0423 19:50:41.096904 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftcontrollermanager/cluster with fingerprint=d71a0f4672f9b45d9fc8293bf1687afc650fd28d32e2e30de27523fe7b4eadf7 I0423 19:50:41.096944 1 recorder.go:75] Recording config/clusteroperator/openshift-samples with fingerprint=2653da4afcd3dc572bb5407894501abc352e2106440c27bce53406fd5e281c03 I0423 19:50:41.096991 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager with fingerprint=d2415dbce18e68a66b0a1e4831f54d4ecf872615d9e75224b5d5518cd5b47d63 I0423 19:50:41.097031 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-catalog with fingerprint=a7ba28e86507c1d6670582bcc5283663c72cffb9f7c9ef42480ccd2a6a9804f1 I0423 19:50:41.097072 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-packageserver with fingerprint=4535eece05c8b7dd7e86ed2297c128bd8fca8374f5746c6828bb69ae0c861d64 I0423 19:50:41.097115 1 recorder.go:75] Recording config/clusteroperator/service-ca with fingerprint=9b442169f8436a05232557bebd9fe71eebd7bf9bb871c37f17d68ab9751f5994 I0423 19:50:41.097176 1 recorder.go:75] Recording config/clusteroperator/storage with fingerprint=064073c1fedd52fe8cabe4e30de00bcefd1ae49bc1a872ad08a92bfd714ca8d7 I0423 19:50:41.097225 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/clustercsidriver/ebs.csi.aws.com with fingerprint=510064d6f6bcced87ab5bd2ddaff3d0edd7f93f4a4f7af2641f29fc53ffab21e I0423 19:50:41.097256 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/storage/cluster with fingerprint=8e480f8c1ce1b39baac42d8ec780c57c2592929ae0c801b61ffad49ba13f33ad I0423 19:50:41.097284 1 gather.go:177] gatherer "clusterconfig" function "operators" took 2.068785745s to process 36 records I0423 19:50:41.286434 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 19:50:41.286482 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-6bdc985f4c-w29zk pod in namespace openshift-image-registry (previous: false). I0423 19:50:41.483712 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-6bdc985f4c-w29zk pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-6bdc985f4c-w29zk\" is waiting to start: ContainerCreating" I0423 19:50:41.483725 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-6bdc985f4c-w29zk\" is waiting to start: ContainerCreating" I0423 19:50:41.483754 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-6bdc985f4c-zpklh pod in namespace openshift-image-registry (previous: false). I0423 19:50:41.690256 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-6bdc985f4c-zpklh pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-6bdc985f4c-zpklh\" is waiting to start: ContainerCreating" I0423 19:50:41.690271 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-6bdc985f4c-zpklh\" is waiting to start: ContainerCreating" I0423 19:50:41.690311 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-b9b695556-wr9bs pod in namespace openshift-image-registry (previous: false). I0423 19:50:41.901910 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-b9b695556-wr9bs pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-b9b695556-wr9bs\" is waiting to start: ContainerCreating" I0423 19:50:41.901923 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-b9b695556-wr9bs\" is waiting to start: ContainerCreating" I0423 19:50:41.901932 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-4qkzz pod in namespace openshift-image-registry (previous: false). W0423 19:50:42.056985 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 19:50:42.083246 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 19:50:42.083264 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-c674v pod in namespace openshift-image-registry (previous: false). I0423 19:50:42.288816 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 19:50:42.288842 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-xv2ch pod in namespace openshift-image-registry (previous: false). I0423 19:50:42.483667 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0423 19:50:42.483684 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-5b9c6dbf75-hdcs7 pod in namespace openshift-ingress (previous: false). I0423 19:50:42.685415 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-5b9c6dbf75-hdcs7 pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-5b9c6dbf75-hdcs7\" is waiting to start: ContainerCreating" I0423 19:50:42.685439 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-5b9c6dbf75-hdcs7\" is waiting to start: ContainerCreating" I0423 19:50:42.685451 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-7878869646-476h2 pod in namespace openshift-ingress (previous: false). I0423 19:50:42.881244 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-7878869646-476h2 pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-7878869646-476h2\" is waiting to start: ContainerCreating" I0423 19:50:42.881265 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-7878869646-476h2\" is waiting to start: ContainerCreating" I0423 19:50:42.881275 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-7878869646-qdsz9 pod in namespace openshift-ingress (previous: false). W0423 19:50:43.056847 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0423 19:50:43.083703 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-7878869646-qdsz9 pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-7878869646-qdsz9\" is waiting to start: ContainerCreating" I0423 19:50:43.083718 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-7878869646-qdsz9\" is waiting to start: ContainerCreating" I0423 19:50:43.083729 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-bjjk2 pod in namespace openshift-ingress-canary (previous: false). I0423 19:50:43.292999 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-qwnb9 pod in namespace openshift-ingress-canary (previous: false). I0423 19:50:43.482934 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-zw5rg pod in namespace openshift-ingress-canary (previous: false). I0423 19:50:43.685164 1 tasks_processing.go:74] worker 56 stopped. I0423 19:50:43.685262 1 recorder.go:75] Recording events/openshift-dns-operator with fingerprint=9e013bcad9e841d849a4fb637ddc09a97a8450adaebbc5ad2cea320a4e232c83 I0423 19:50:43.685330 1 recorder.go:75] Recording events/openshift-dns with fingerprint=aef594784544ce482d5585cb60c7cc6dd230b34a109f3a54d835f60598c95ba4 I0423 19:50:43.685437 1 recorder.go:75] Recording events/openshift-image-registry with fingerprint=f2c47ad642b235355af7be7f378b9f4bc15b34e53f54efdd12d9d076d0209333 I0423 19:50:43.685480 1 recorder.go:75] Recording events/openshift-ingress-operator with fingerprint=680d5c6977147d47b19492015ee23f9a7bfb2b947b11d04b396bb30a835214d3 I0423 19:50:43.685540 1 recorder.go:75] Recording events/openshift-ingress with fingerprint=f0b7a5903a6280ffa96d859a65c8e086a84cc31e1eaf67e04a9166d88716b869 I0423 19:50:43.685592 1 recorder.go:75] Recording events/openshift-ingress-canary with fingerprint=f6fa2f7560a668ca222dd03d58ddd6b2d79de7aa2ba3c032cde5cb49363d6498 I0423 19:50:43.685746 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-7qsv8 with fingerprint=8bbbc7f03219ebc290d163546a568e75a26fb987d5b339e254b43cf35737dabb I0423 19:50:43.685846 1 recorder.go:75] Recording config/pod/openshift-dns/dns-default-cfmp6 with fingerprint=2bfc3c6908f92b62d13fa3c0848f3700cc95da53242116e7da929533cea3b50a I0423 19:50:43.685960 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-6bdc985f4c-w29zk with fingerprint=b05c8aa9aa87e4952c346fc03e1da683e04723b9a6f0e81083458b25eb525dd9 I0423 19:50:43.686072 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-6bdc985f4c-zpklh with fingerprint=7bfd5a23ff122ee3f2bc14d4134b3db0cf7bd2e48e6687a9d29057764564eee3 I0423 19:50:43.686176 1 recorder.go:75] Recording config/pod/openshift-image-registry/image-registry-b9b695556-wr9bs with fingerprint=b4a94352166d946a3935e2647bd330b6a49cccc23394ed95c038423c6cbbeeb3 I0423 19:50:43.686186 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/logs/ingress-canary-bjjk2/serve-healthcheck-canary_current.log with fingerprint=4b791c799ffe72f70cd60b4ab385a1309543ee8f09e4f03de13730e2565624f2 I0423 19:50:43.686193 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/logs/ingress-canary-qwnb9/serve-healthcheck-canary_current.log with fingerprint=2923b839f42295b8934ca9498930f0b73f9261f0a699025839c4adc223eed4f1 I0423 19:50:43.686201 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/logs/ingress-canary-zw5rg/serve-healthcheck-canary_current.log with fingerprint=723a9d34d40bd825c865182c460346d0b39a8151911024640756fd6bce9043e4 I0423 19:50:43.686213 1 gather.go:177] gatherer "clusterconfig" function "operators_pods_and_events" took 4.658291615s to process 14 records W0423 19:50:44.056616 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. W0423 19:50:44.056637 1 gather_dvo_metrics.go:117] Unable to read metrics from endpoint "http://deployment-validation-operator-metrics.openshift-deployment-validation-operator.svc:8383": DVO metrics service was not available within the 5s timeout: context deadline exceeded I0423 19:50:44.056665 1 tasks_processing.go:74] worker 43 stopped. E0423 19:50:44.056675 1 gather.go:140] gatherer "clusterconfig" function "dvo_metrics" failed with the error: DVO metrics service was not available within the 5s timeout: context deadline exceeded I0423 19:50:44.056685 1 recorder.go:75] Recording config/dvo_metrics with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 W0423 19:50:44.056700 1 gather.go:155] issue recording gatherer "clusterconfig" function "dvo_metrics" result "config/dvo_metrics" because of the warning: warning: the record with the same fingerprint "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" was already recorded at path "config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt", recording another one with a different path "config/dvo_metrics" I0423 19:50:44.056712 1 gather.go:177] gatherer "clusterconfig" function "dvo_metrics" took 5.029854051s to process 1 records I0423 19:50:51.457549 1 tasks_processing.go:74] worker 1 stopped. I0423 19:50:51.457594 1 recorder.go:75] Recording config/installplans with fingerprint=f17dbfacc3bfddf27ca3b213b39495434cd4c4e9e3dbd69566ffb3845bbcf539 I0423 19:50:51.457608 1 gather.go:177] gatherer "clusterconfig" function "install_plans" took 12.432200956s to process 1 records I0423 19:50:52.021531 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 19:50:52.208057 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 19:50:52.231200 1 tasks_processing.go:74] worker 2 stopped. I0423 19:50:52.231457 1 recorder.go:75] Recording config/serviceaccounts with fingerprint=8e51bdad94ccef4d8b8d1dcd31ca57822f1858ab42bfebeef1b93869b7cdc5cb I0423 19:50:52.231473 1 gather.go:177] gatherer "clusterconfig" function "service_accounts" took 13.207179325s to process 1 records E0423 19:50:52.231529 1 periodic.go:254] "Unhandled Error" err="clusterconfig failed after 13.207s with: function \"pod_network_connectivity_checks\" failed with an error, function \"machines\" failed with an error, function \"machine_healthchecks\" failed with an error, function \"support_secret\" failed with an error, function \"machine_configs\" failed with an error, function \"config_maps\" failed with an error, function \"ingress_certificates\" failed with an error, function \"dvo_metrics\" failed with an error" I0423 19:50:52.232638 1 controllerstatus.go:89] name=periodic-clusterconfig healthy=false reason=PeriodicGatherFailed message=Source clusterconfig could not be retrieved: function "pod_network_connectivity_checks" failed with an error, function "machines" failed with an error, function "machine_healthchecks" failed with an error, function "support_secret" failed with an error, function "machine_configs" failed with an error, function "config_maps" failed with an error, function "ingress_certificates" failed with an error, function "dvo_metrics" failed with an error I0423 19:50:52.232687 1 periodic.go:216] Running workloads gatherer I0423 19:50:52.232700 1 tasks_processing.go:45] number of workers: 2 I0423 19:50:52.232709 1 tasks_processing.go:69] worker 1 listening for tasks. I0423 19:50:52.232714 1 tasks_processing.go:71] worker 1 working on helmchart_info task. I0423 19:50:52.232725 1 tasks_processing.go:69] worker 0 listening for tasks. I0423 19:50:52.232742 1 tasks_processing.go:71] worker 0 working on workload_info task. I0423 19:50:52.264525 1 tasks_processing.go:74] worker 1 stopped. I0423 19:50:52.264548 1 gather.go:177] gatherer "workloads" function "helmchart_info" took 31.799687ms to process 0 records I0423 19:50:52.290874 1 gather_workloads_info.go:278] Loaded pods in 0s, will wait 22s for image data I0423 19:50:52.300703 1 gather_workloads_info.go:387] No image sha256:730d1b6988025bef0daa3a9a5d8467ec4a26b0382cc52f91c3375b4590d3518a (11ms) I0423 19:50:52.311085 1 gather_workloads_info.go:387] No image sha256:a043239802b3eb8b323d285193d2527fad0ecec98ca91d188a3472a2fac8ae04 (10ms) I0423 19:50:52.321094 1 gather_workloads_info.go:387] No image sha256:90a8ffd9643ebb16a6a8c04bb38cf9ed58903e9d3bf836c68f399193db5edaf6 (10ms) I0423 19:50:52.330482 1 gather_workloads_info.go:387] No image sha256:e8b96d9318b3b8c9ed0afe4e6381f635c6b0c2f20772044ac68001ed49af2c87 (9ms) I0423 19:50:52.338495 1 gather_workloads_info.go:387] No image sha256:0a4dfb8d4c1b3849319d45b4c54dff26a7238a2c08fcaa121f93073e95ab12e8 (8ms) I0423 19:50:52.346606 1 gather_workloads_info.go:387] No image sha256:5808401268394502d335281ea1a294b07210461b986b58f91d7d1f29c0029c6d (8ms) I0423 19:50:52.354928 1 gather_workloads_info.go:387] No image sha256:5ac9c549d65fc1d8bc900773bebee43e9192bcec1bb5fa46afb4597230c16ac7 (8ms) I0423 19:50:52.363018 1 gather_workloads_info.go:387] No image sha256:084aa9b0f8a6d478549dc384d4e66da13ee9b25cc98531da861cc19dee2a9e8f (8ms) I0423 19:50:52.371020 1 gather_workloads_info.go:387] No image sha256:875c77e5d144f03fb91d8cee0259f6966683ca88d1bf818dbf4652c16b70312c (8ms) I0423 19:50:52.379309 1 gather_workloads_info.go:387] No image sha256:25c148fd380b1a9db3f6039d2e0eabc489a954921452391390ba9192b2325678 (8ms) I0423 19:50:52.405922 1 gather_workloads_info.go:387] No image sha256:e1ba458cf6f0b3606c90880da72db8ab99cd11040bae84baebe3ff2e0d1ea075 (27ms) I0423 19:50:52.409830 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 19:50:52.499063 1 gather_workloads_info.go:387] No image sha256:ce138e8cf5b96557b1864ef6f27c2608bdca59be2611804366cef7169c36291e (93ms) I0423 19:50:52.600106 1 gather_workloads_info.go:387] No image sha256:5c6d21c3f97366bc7ab57031cc027b67405a684bd804ce364ed5998b0685eaca (101ms) I0423 19:50:52.700545 1 gather_workloads_info.go:387] No image sha256:5aaea0419169e55832cc27acfe0fe3b9513a343d6bdf71d3da1575ed322245d0 (100ms) I0423 19:50:52.799455 1 gather_workloads_info.go:387] No image sha256:15677f0b70e6aa2dfaf088e45fc1a425c22bf6fda326b8116f87e88b6694dfab (99ms) I0423 19:50:52.899766 1 gather_workloads_info.go:387] No image sha256:000105ef5150e7079b90a613fb9e6193e2a6ef9b1908d2dce44f2395d4fd070f (100ms) I0423 19:50:53.005802 1 gather_workloads_info.go:387] No image sha256:89277d8d4560d71db88c2dcc67c992a24544ca21810920b609c1d49d53b4a287 (106ms) I0423 19:50:53.100489 1 gather_workloads_info.go:387] No image sha256:55b1db6038c5beaed54c626e3343b7a8589cc0be8dc41d1a66b4deab766ff520 (95ms) I0423 19:50:53.204735 1 gather_workloads_info.go:387] No image sha256:48e883a3932aea9457f55cd4628d21397d429d4828f42f95d1c903d9d5395bde (104ms) I0423 19:50:53.299772 1 gather_workloads_info.go:387] No image sha256:934d8e8c50f3c609b8eea80d1051111fe3d066fe8c65c79572072ae55fcb0a86 (95ms) I0423 19:50:53.399097 1 gather_workloads_info.go:387] No image sha256:673ebc8cc22c56c8f410e011b2fa950d28cf7b6420e17fdb6580d6cb10523384 (99ms) I0423 19:50:53.499668 1 gather_workloads_info.go:387] No image sha256:c15ca0c0ad60fe8757c2d5d1723fcdd7a1ed6c0251a90d22a7e6cae6811d01aa (101ms) I0423 19:50:53.604806 1 gather_workloads_info.go:387] No image sha256:d6fbe0075cbb12bfd287c973704fadf97154c7f73e370733d976a40835e9155a (105ms) I0423 19:50:53.699833 1 gather_workloads_info.go:387] No image sha256:ab60623bb32f7e75fca71ef65137731cae347a21c7a4091dfd583fa00732721c (95ms) I0423 19:50:53.804391 1 gather_workloads_info.go:387] No image sha256:30e597ec5d6bb96ff70a4f8688c748b659cd4fd5d73d222e8701821d236795c5 (105ms) I0423 19:50:53.900178 1 gather_workloads_info.go:387] No image sha256:08c5a78c8a5af04c549e2273aaf4bb452a75bf038d68aa9d01bb2aff66c30e90 (96ms) I0423 19:50:54.002583 1 gather_workloads_info.go:387] No image sha256:6814f5fced0ef219d06374011c68a11a3da788a764a00a69ade435466d9ee240 (102ms) I0423 19:50:54.098973 1 gather_workloads_info.go:387] No image sha256:4733236617781e3469ffffb15e4daaa1f14ea8e1c52b426a3787a4f1f2945424 (96ms) I0423 19:50:54.204686 1 gather_workloads_info.go:387] No image sha256:e3732e356ae2324565c74cae57d5d016917314fa293ddaa3a68ee9ae030c6f07 (106ms) I0423 19:50:54.299335 1 gather_workloads_info.go:387] No image sha256:637d41f067a5239096fc22b135181cda5113da833f1370e7a73965e83792e93a (95ms) I0423 19:50:54.402735 1 gather_workloads_info.go:387] No image sha256:5b6212b8f539f08e78417d8a4b7485ca0b4e7927cacd7b752742a28841bc8ccd (103ms) I0423 19:50:54.499273 1 gather_workloads_info.go:387] No image sha256:3bba1358d4a0ae878ff491c0c2cbfffe60649e110b40342b878fe8fa332f8858 (97ms) I0423 19:50:54.599111 1 gather_workloads_info.go:387] No image sha256:d64bea34bf3e1bb0b3a701c3ff14e66665afc1b050f28124ad7e6888eaec3a81 (100ms) I0423 19:50:54.599135 1 tasks_processing.go:74] worker 0 stopped. E0423 19:50:54.599144 1 gather.go:140] gatherer "workloads" function "workload_info" failed with the error: no running pods found for the insights-runtime-extractor statefulset I0423 19:50:54.599374 1 recorder.go:75] Recording config/workload_info with fingerprint=6d4012f8ba3fadcba1e197ee3d947ab59eacfab94c6052d09b508f475ecc9dca I0423 19:50:54.599389 1 gather.go:177] gatherer "workloads" function "workload_info" took 2.366385532s to process 1 records E0423 19:50:54.599412 1 periodic.go:254] "Unhandled Error" err="workloads failed after 2.366s with: function \"workload_info\" failed with an error" I0423 19:50:54.600514 1 controllerstatus.go:89] name=periodic-workloads healthy=false reason=PeriodicGatherFailed message=Source workloads could not be retrieved: function "workload_info" failed with an error I0423 19:50:54.600525 1 periodic.go:216] Running conditional gatherer I0423 19:50:54.607293 1 requests.go:294] Making HTTP GET request at: https://console.redhat.com/api/gathering/v2/4.21.9/gathering_rules I0423 19:50:54.613704 1 conditional_gatherer.go:107] Get "https://console.redhat.com/api/gathering/v2/4.21.9/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.129.0.10:36803->172.30.0.10:53: read: connection refused E0423 19:50:54.613994 1 conditional_gatherer.go:322] unable to update alerts cache: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0423 19:50:54.614049 1 conditional_gatherer.go:384] updating version cache for conditional gatherer I0423 19:50:54.622939 1 conditional_gatherer.go:392] cluster version is '4.21.9' E0423 19:50:54.622951 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 19:50:54.622956 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 19:50:54.622960 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 19:50:54.622963 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 19:50:54.622966 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 19:50:54.622969 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 19:50:54.622972 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 19:50:54.622974 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0423 19:50:54.622977 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing I0423 19:50:54.622990 1 tasks_processing.go:45] number of workers: 3 I0423 19:50:54.623000 1 tasks_processing.go:69] worker 2 listening for tasks. I0423 19:50:54.623005 1 tasks_processing.go:71] worker 2 working on conditional_gatherer_rules task. I0423 19:50:54.623010 1 tasks_processing.go:69] worker 0 listening for tasks. I0423 19:50:54.623020 1 tasks_processing.go:71] worker 0 working on remote_configuration task. I0423 19:50:54.623021 1 tasks_processing.go:69] worker 1 listening for tasks. I0423 19:50:54.623026 1 tasks_processing.go:71] worker 2 working on rapid_container_logs task. I0423 19:50:54.623030 1 tasks_processing.go:74] worker 1 stopped. I0423 19:50:54.623084 1 recorder.go:75] Recording insights-operator/conditional-gatherer-rules with fingerprint=7034af97e7e41c22e4b775abdd4b9066c8ebb19da33eb7f69f39bfd2eb5f6406 I0423 19:50:54.623097 1 gather.go:177] gatherer "conditional" function "conditional_gatherer_rules" took 833ns to process 1 records I0423 19:50:54.623126 1 recorder.go:75] Recording insights-operator/remote-configuration with fingerprint=0394430c431eec4d48bb1811a90918e95161d2282c59af26f2473613cc0959db I0423 19:50:54.623134 1 gather.go:177] gatherer "conditional" function "remote_configuration" took 993ns to process 1 records I0423 19:50:54.623139 1 tasks_processing.go:74] worker 0 stopped. I0423 19:50:54.623259 1 tasks_processing.go:74] worker 2 stopped. I0423 19:50:54.623272 1 gather.go:177] gatherer "conditional" function "rapid_container_logs" took 224.015µs to process 0 records I0423 19:50:54.623292 1 controllerstatus.go:89] name=periodic-conditional healthy=false reason=NotAvailable message=Get "https://console.redhat.com/api/gathering/v2/4.21.9/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.129.0.10:36803->172.30.0.10:53: read: connection refused I0423 19:50:54.623309 1 recorder.go:75] Recording insights-operator/remote-configuration.json with fingerprint=359de9c990c741675cec72fda96b5c3682221efdb4799f5eaa6e9805bcd3b5c1 I0423 19:50:54.645734 1 recorder.go:75] Recording insights-operator/gathers with fingerprint=1c5d733de2cea0e0dc003c7a9488561565c7c8daa217d66c93bea928f737fe86 I0423 19:50:54.645849 1 diskrecorder.go:70] Writing 110 records to /var/lib/insights-operator/insights-2026-04-23-195054.tar.gz I0423 19:50:54.654006 1 diskrecorder.go:51] Wrote 110 records to disk in 8ms I0423 19:50:54.654039 1 periodic.go:285] Gathering cluster info every 2h0m0s I0423 19:50:54.654054 1 periodic.go:286] Configuration is dataReporting: interval: 2h0m0s, uploadEndpoint: https://console.redhat.com/api/ingress/v1/upload, storagePath: /var/lib/insights-operator, downloadEndpoint: https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports, conditionalGathererEndpoint: https://console.redhat.com/api/gathering/v2/%s/gathering_rules, obfuscation: [] sca: disabled: false, endpoint: https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates, interval: 8h0m0s alerting: disabled: false clusterTransfer: endpoint: https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/, interval: 12h0m0s proxy: httpProxy: , httpsProxy: , noProxy: I0423 19:51:08.573663 1 configmapobserver.go:84] configmaps "insights-config" not found I0423 19:51:58.497816 1 observer_polling.go:111] Observed file "/var/run/configmaps/service-ca-bundle/service-ca.crt" has been created (hash="b3d7b3a1b83404e88251dd732b740d8575f3d0dd82a9fae61d7f2e8f2562cc47") W0423 19:51:58.497850 1 builder.go:160] Restart triggered because of file /var/run/configmaps/service-ca-bundle/service-ca.crt was created I0423 19:51:58.497900 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.crt" has been created (hash="d0a419f68da91a81f5d9d576a58c4966bcab08d44c496d6355069e13dc6818ee") I0423 19:51:58.497912 1 genericapiserver.go:548] "[graceful-termination] shutdown event" name="ShutdownInitiated"