W0320 09:56:37.676225 1 cmd.go:257] Using insecure, self-signed certificates I0320 09:56:38.417534 1 start.go:138] Unable to read service ca bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0320 09:56:38.417842 1 observer_polling.go:159] Starting file observer I0320 09:56:40.214348 1 operator.go:76] Starting insights-operator v0.0.0-master+$Format:%H$ I0320 09:56:40.214556 1 legacy_config.go:327] Current config: {"report":false,"storagePath":"/var/lib/insights-operator","interval":"2h","endpoint":"https://console.redhat.com/api/ingress/v1/upload","conditionalGathererEndpoint":"https://console.redhat.com/api/gathering/v2/%s/gathering_rules","pull_report":{"endpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports","delay":"60s","timeout":"3000s","min_retry":"30s"},"impersonate":"system:serviceaccount:openshift-insights:gather","enableGlobalObfuscation":false,"ocm":{"scaEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates","scaInterval":"8h","scaDisabled":false,"clusterTransferEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/","clusterTransferInterval":"12h"},"disableInsightsAlerts":false,"processingStatusEndpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status","reportEndpointTechPreview":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/report"} I0320 09:56:40.215063 1 secure_serving.go:57] Forcing use of http/1.1 only W0320 09:56:40.215084 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. W0320 09:56:40.215088 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. W0320 09:56:40.215092 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. W0320 09:56:40.215094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. W0320 09:56:40.215096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. W0320 09:56:40.215099 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. I0320 09:56:40.215166 1 simple_featuregate_reader.go:171] Starting feature-gate-detector I0320 09:56:40.218388 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-insights", Name:"insights-operator", UID:"7eb76287-3019-4d66-8819-cebd70b544d4", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BuildCSIVolumes", "CPMSMachineNamePrefix", "ConsolePluginContentSecurityPolicy", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNSInstall", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "HyperShiftOnlyDynamicResourceAllocation", "ImageStreamImportMode", "ImageVolume", "KMSv1", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "ManagedBootImagesAzure", "ManagedBootImagesvSphere", "MetricsCollectionProfiles", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "PinnedImages", "PreconfiguredUDNAddresses", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SigstoreImageVerification", "SigstoreImageVerificationPKI", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiDisk", "VSphereMultiNetworks", "VolumeAttributesClass"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSDualStackInstall", "AWSServiceLBNetworkSecurityGroup", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureDualStackInstall", "AzureMultiDisk", "BootImageSkewEnforcement", "BootcNodeManagement", "CBORServingAndStorage", "CRDCompatibilityRequirementOperator", "ClientsAllowCBOR", "ClientsPreferCBOR", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterAPIMachineManagement", "ClusterAPIMachineManagementVSphere", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "EtcdBackendQuota", "EventTTL", "EventedPLEG", "Example", "Example2", "ExternalSnapshotMetadata", "GCPClusterHostedDNS", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "GCPDualStackInstall", "ImageModeStatusReporting", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryptionProvider", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "ManagedBootImagesCPMS", "MaxUnavailableStatefulSet", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "MultiDiskSetup", "MutableCSINodeAllocatableCount", "MutatingAdmissionPolicy", "NewOLM", "NewOLMBoxCutterRuntime", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterInstall", "NutanixMultiSubnets", "OSStreams", "OVNObservability", "OnPremDNSRecords", "OpenShiftPodSecurityAdmission", "ProvisioningRequestAvailable", "SELinuxMount", "ShortCertRotation", "SignatureStores", "TranslateStreamCloseWebsocketRequests", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VolumeGroupSnapshot"}} I0320 09:56:40.218385 1 operator.go:141] FeatureGates initialized: knownFeatureGates=[AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BuildCSIVolumes CPMSMachineNamePrefix ConsolePluginContentSecurityPolicy ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GCPClusterHostedDNSInstall GatewayAPI GatewayAPIController HighlyAvailableArbiter HyperShiftOnlyDynamicResourceAllocation ImageStreamImportMode ImageVolume KMSv1 MachineConfigNodes ManagedBootImages ManagedBootImagesAWS ManagedBootImagesAzure ManagedBootImagesvSphere MetricsCollectionProfiles NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation PinnedImages PreconfiguredUDNAddresses ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SigstoreImageVerification SigstoreImageVerificationPKI StoragePerformantSecurityPolicy UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiDisk VSphereMultiNetworks VolumeAttributesClass AWSClusterHostedDNS AWSClusterHostedDNSInstall AWSDedicatedHosts AWSDualStackInstall AWSServiceLBNetworkSecurityGroup AutomatedEtcdBackup AzureClusterHostedDNSInstall AzureDedicatedHosts AzureDualStackInstall AzureMultiDisk BootImageSkewEnforcement BootcNodeManagement CBORServingAndStorage CRDCompatibilityRequirementOperator ClientsAllowCBOR ClientsPreferCBOR ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterAPIMachineManagement ClusterAPIMachineManagementVSphere ClusterMonitoringConfig ClusterVersionOperatorConfiguration DNSNameResolver DualReplica DyanmicServiceEndpointIBMCloud EtcdBackendQuota EventTTL EventedPLEG Example Example2 ExternalSnapshotMetadata GCPClusterHostedDNS GCPCustomAPIEndpoints GCPCustomAPIEndpointsInstall GCPDualStackInstall ImageModeStatusReporting IngressControllerDynamicConfigurationManager InsightsConfig InsightsOnDemandDataGather IrreconcilableMachineConfig KMSEncryptionProvider MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController ManagedBootImagesCPMS MaxUnavailableStatefulSet MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure MultiDiskSetup MutableCSINodeAllocatableCount MutatingAdmissionPolicy NewOLM NewOLMBoxCutterRuntime NewOLMCatalogdAPIV1Metas NewOLMOwnSingleNamespace NewOLMPreflightPermissionChecks NewOLMWebhookProviderOpenshiftServiceCA NoRegistryClusterInstall NutanixMultiSubnets OSStreams OVNObservability OnPremDNSRecords OpenShiftPodSecurityAdmission ProvisioningRequestAvailable SELinuxMount ShortCertRotation SignatureStores TranslateStreamCloseWebsocketRequests VSphereConfigurableMaxAllowedBlockVolumesPerNode VSphereHostVMGroupZonal VSphereMixedNodeEnv VolumeGroupSnapshot] I0320 09:56:40.220546 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController I0320 09:56:40.220558 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController I0320 09:56:40.220562 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0320 09:56:40.220572 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0320 09:56:40.220581 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0320 09:56:40.220586 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0320 09:56:40.220923 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/tmp/serving-cert-495997436/tls.crt::/tmp/serving-cert-495997436/tls.key" I0320 09:56:40.221181 1 secure_serving.go:213] Serving securely on [::]:8443 I0320 09:56:40.221208 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" W0320 09:56:40.223923 1 configmapobserver.go:64] Cannot get the configuration config map: configmaps "insights-config" not found. Default configuration is used. I0320 09:56:40.223992 1 secretconfigobserver.go:216] Legacy configuration set: enabled=false endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=false reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0320 09:56:40.224464 1 base_controller.go:76] Waiting for caches to sync for ConfigController I0320 09:56:40.230177 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0320 09:56:40.230194 1 secretconfigobserver.go:204] Legacy configuration updated: enabled=true endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=true reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0320 09:56:40.234786 1 secretconfigobserver.go:119] support secret does not exist I0320 09:56:40.239680 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0320 09:56:40.244488 1 secretconfigobserver.go:119] support secret does not exist I0320 09:56:40.247700 1 recorder.go:176] Pruning old reports every 4h17m57s, max age is 288h0m0s I0320 09:56:40.252708 1 periodic.go:216] Running clusterconfig gatherer I0320 09:56:40.252723 1 controllerstatus.go:80] name=insightsreport healthy=true reason= message= I0320 09:56:40.252730 1 insightsreport.go:296] Starting report retriever I0320 09:56:40.252737 1 insightsreport.go:298] Insights analysis reports will be downloaded from the https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports endpoint with a delay of 1m0s I0320 09:56:40.252765 1 tasks_processing.go:45] number of workers: 32 I0320 09:56:40.252791 1 tasks_processing.go:69] worker 31 listening for tasks. I0320 09:56:40.252797 1 tasks_processing.go:71] worker 31 working on olm_operators task. I0320 09:56:40.252818 1 tasks_processing.go:69] worker 14 listening for tasks. I0320 09:56:40.252830 1 tasks_processing.go:69] worker 7 listening for tasks. I0320 09:56:40.252829 1 tasks_processing.go:69] worker 22 listening for tasks. I0320 09:56:40.252836 1 tasks_processing.go:69] worker 6 listening for tasks. I0320 09:56:40.252838 1 tasks_processing.go:69] worker 8 listening for tasks. I0320 09:56:40.252706 1 controllerstatus.go:80] name=insightsuploader healthy=true reason= message= I0320 09:56:40.252845 1 tasks_processing.go:69] worker 9 listening for tasks. I0320 09:56:40.252846 1 tasks_processing.go:69] worker 19 listening for tasks. I0320 09:56:40.252850 1 insightsuploader.go:86] Reporting status periodically to https://console.redhat.com/api/ingress/v1/upload every 2h0m0s, starting in 1m30s I0320 09:56:40.252854 1 tasks_processing.go:69] worker 20 listening for tasks. I0320 09:56:40.252857 1 tasks_processing.go:69] worker 18 listening for tasks. I0320 09:56:40.252863 1 tasks_processing.go:69] worker 21 listening for tasks. I0320 09:56:40.252859 1 tasks_processing.go:69] worker 10 listening for tasks. I0320 09:56:40.252865 1 tasks_processing.go:69] worker 11 listening for tasks. I0320 09:56:40.252870 1 tasks_processing.go:69] worker 17 listening for tasks. I0320 09:56:40.252873 1 tasks_processing.go:69] worker 12 listening for tasks. I0320 09:56:40.252867 1 tasks_processing.go:69] worker 0 listening for tasks. I0320 09:56:40.252878 1 tasks_processing.go:69] worker 13 listening for tasks. I0320 09:56:40.252875 1 tasks_processing.go:69] worker 1 listening for tasks. I0320 09:56:40.252883 1 tasks_processing.go:69] worker 2 listening for tasks. I0320 09:56:40.252886 1 tasks_processing.go:69] worker 24 listening for tasks. I0320 09:56:40.252886 1 tasks_processing.go:69] worker 15 listening for tasks. I0320 09:56:40.252891 1 tasks_processing.go:69] worker 4 listening for tasks. I0320 09:56:40.252894 1 tasks_processing.go:69] worker 26 listening for tasks. I0320 09:56:40.252880 1 tasks_processing.go:69] worker 16 listening for tasks. I0320 09:56:40.252894 1 tasks_processing.go:69] worker 23 listening for tasks. I0320 09:56:40.252904 1 tasks_processing.go:69] worker 29 listening for tasks. I0320 09:56:40.252905 1 tasks_processing.go:69] worker 28 listening for tasks. I0320 09:56:40.252906 1 tasks_processing.go:69] worker 30 listening for tasks. I0320 09:56:40.252899 1 tasks_processing.go:69] worker 25 listening for tasks. I0320 09:56:40.252894 1 tasks_processing.go:69] worker 5 listening for tasks. I0320 09:56:40.252896 1 tasks_processing.go:69] worker 3 listening for tasks. I0320 09:56:40.252904 1 tasks_processing.go:69] worker 27 listening for tasks. I0320 09:56:40.252974 1 tasks_processing.go:71] worker 12 working on nodenetworkstates task. I0320 09:56:40.252980 1 tasks_processing.go:71] worker 14 working on storage_classes task. I0320 09:56:40.252984 1 tasks_processing.go:71] worker 13 working on openshift_logging task. I0320 09:56:40.253000 1 tasks_processing.go:71] worker 8 working on jaegers task. I0320 09:56:40.253013 1 tasks_processing.go:71] worker 22 working on overlapping_namespace_uids task. I0320 09:56:40.253023 1 tasks_processing.go:71] worker 7 working on schedulers task. I0320 09:56:40.253049 1 tasks_processing.go:71] worker 19 working on container_images task. I0320 09:56:40.252975 1 tasks_processing.go:71] worker 18 working on ingress task. I0320 09:56:40.253191 1 tasks_processing.go:71] worker 9 working on sap_config task. I0320 09:56:40.253191 1 tasks_processing.go:71] worker 6 working on machine_config_pools task. I0320 09:56:40.253341 1 tasks_processing.go:71] worker 20 working on nodenetworkconfigurationpolicies task. I0320 09:56:40.253406 1 tasks_processing.go:71] worker 25 working on openshift_machine_api_events task. I0320 09:56:40.252980 1 tasks_processing.go:71] worker 0 working on sap_datahubs task. I0320 09:56:40.253406 1 tasks_processing.go:71] worker 21 working on machine_configs task. I0320 09:56:40.253349 1 tasks_processing.go:71] worker 1 working on support_secret task. I0320 09:56:40.253356 1 tasks_processing.go:71] worker 2 working on qemu_kubevirt_launcher_logs task. I0320 09:56:40.253373 1 tasks_processing.go:71] worker 24 working on silenced_alerts task. W0320 09:56:40.253833 1 gather_silenced_alerts.go:38] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0320 09:56:40.253848 1 tasks_processing.go:71] worker 24 working on install_plans task. I0320 09:56:40.253377 1 tasks_processing.go:71] worker 15 working on container_runtime_configs task. I0320 09:56:40.253886 1 gather.go:177] gatherer "clusterconfig" function "silenced_alerts" took 32.899µs to process 0 records I0320 09:56:40.253381 1 tasks_processing.go:71] worker 4 working on metrics task. W0320 09:56:40.253987 1 gather_most_recent_metrics.go:64] Unable to load metrics client, no metrics will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0320 09:56:40.253999 1 tasks_processing.go:71] worker 4 working on mutating_webhook_configurations task. I0320 09:56:40.253385 1 tasks_processing.go:71] worker 26 working on dvo_metrics task. I0320 09:56:40.254062 1 gather.go:177] gatherer "clusterconfig" function "metrics" took 26.318µs to process 0 records I0320 09:56:40.253388 1 tasks_processing.go:71] worker 16 working on number_of_pods_and_netnamespaces_with_sdn_annotations task. I0320 09:56:40.253392 1 tasks_processing.go:71] worker 23 working on image_registries task. I0320 09:56:40.253395 1 tasks_processing.go:71] worker 29 working on machines task. I0320 09:56:40.253400 1 tasks_processing.go:71] worker 28 working on image task. I0320 09:56:40.253401 1 tasks_processing.go:71] worker 11 working on nodes task. I0320 09:56:40.253403 1 tasks_processing.go:71] worker 30 working on image_pruners task. I0320 09:56:40.253408 1 tasks_processing.go:71] worker 10 working on lokistack task. I0320 09:56:40.252977 1 tasks_processing.go:71] worker 27 working on oauths task. I0320 09:56:40.253415 1 tasks_processing.go:71] worker 3 working on cost_management_metrics_configs task. I0320 09:56:40.253415 1 tasks_processing.go:71] worker 17 working on networks task. I0320 09:56:40.253418 1 tasks_processing.go:71] worker 5 working on active_alerts task. W0320 09:56:40.255030 1 gather_active_alerts.go:54] Unable to load alerts client, no alerts will be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0320 09:56:40.255040 1 tasks_processing.go:71] worker 5 working on version task. I0320 09:56:40.255057 1 gather.go:177] gatherer "clusterconfig" function "active_alerts" took 25.718µs to process 0 records I0320 09:56:40.257626 1 tasks_processing.go:71] worker 13 working on tsdb_status task. I0320 09:56:40.257636 1 gather.go:177] gatherer "clusterconfig" function "openshift_logging" took 4.628754ms to process 0 records W0320 09:56:40.257652 1 gather_prometheus_tsdb_status.go:38] Unable to load metrics client, tsdb status cannot be collected: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0320 09:56:40.257658 1 tasks_processing.go:71] worker 13 working on feature_gates task. I0320 09:56:40.257712 1 gather.go:177] gatherer "clusterconfig" function "tsdb_status" took 22.489µs to process 0 records I0320 09:56:40.258749 1 tasks_processing.go:71] worker 9 working on openstack_dataplanenodesets task. I0320 09:56:40.258756 1 gather.go:177] gatherer "clusterconfig" function "sap_config" took 5.544667ms to process 0 records I0320 09:56:40.259370 1 tasks_processing.go:71] worker 12 working on machine_healthchecks task. I0320 09:56:40.259378 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkstates" took 6.382013ms to process 0 records I0320 09:56:40.263267 1 tasks_processing.go:71] worker 15 working on crds task. I0320 09:56:40.263333 1 gather.go:177] gatherer "clusterconfig" function "container_runtime_configs" took 9.395129ms to process 0 records I0320 09:56:40.263346 1 gather.go:177] gatherer "clusterconfig" function "machine_config_pools" took 10.082579ms to process 0 records I0320 09:56:40.263351 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanenodesets" took 4.548081ms to process 0 records I0320 09:56:40.263358 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkconfigurationpolicies" took 9.904144ms to process 0 records I0320 09:56:40.263397 1 tasks_processing.go:71] worker 9 working on aggregated_monitoring_cr_names task. I0320 09:56:40.263407 1 tasks_processing.go:71] worker 20 working on ceph_cluster task. I0320 09:56:40.263449 1 tasks_processing.go:71] worker 6 working on node_logs task. I0320 09:56:40.263469 1 tasks_processing.go:71] worker 29 working on clusterroles task. E0320 09:56:40.263482 1 gather.go:140] gatherer "clusterconfig" function "machines" failed with the error: machines.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machines" in API group "machine.openshift.io" at the cluster scope I0320 09:56:40.263676 1 gather.go:177] gatherer "clusterconfig" function "machines" took 9.111261ms to process 0 records I0320 09:56:40.263684 1 gather.go:177] gatherer "clusterconfig" function "olm_operators" took 10.833466ms to process 0 records I0320 09:56:40.263691 1 tasks_processing.go:71] worker 31 working on node_features task. I0320 09:56:40.264048 1 tasks_processing.go:71] worker 12 working on sap_pods task. E0320 09:56:40.264062 1 gather.go:140] gatherer "clusterconfig" function "machine_healthchecks" failed with the error: machinehealthchecks.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machinehealthchecks" in API group "machine.openshift.io" at the cluster scope I0320 09:56:40.264073 1 gather.go:177] gatherer "clusterconfig" function "machine_healthchecks" took 4.661977ms to process 0 records I0320 09:56:40.264184 1 controller.go:129] Initializing last reported time to 0001-01-01T00:00:00Z I0320 09:56:40.264199 1 controller.go:254] Source periodic-workloads *controllerstatus.Simple is not ready I0320 09:56:40.264204 1 controller.go:254] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0320 09:56:40.264209 1 controller.go:254] Source periodic-conditional *controllerstatus.Simple is not ready I0320 09:56:40.264391 1 controller.go:531] The operator is still being initialized I0320 09:56:40.264408 1 controller.go:554] The operator is healthy I0320 09:56:40.264597 1 tasks_processing.go:71] worker 0 working on openstack_version task. I0320 09:56:40.264756 1 gather.go:177] gatherer "clusterconfig" function "sap_datahubs" took 11.11774ms to process 0 records I0320 09:56:40.265130 1 gather.go:177] gatherer "clusterconfig" function "cost_management_metrics_configs" took 10.191213ms to process 0 records I0320 09:56:40.265111 1 tasks_processing.go:71] worker 3 working on pod_network_connectivity_checks task. I0320 09:56:40.266393 1 tasks_processing.go:71] worker 10 working on infrastructures task. I0320 09:56:40.266401 1 gather.go:177] gatherer "clusterconfig" function "lokistack" took 11.629395ms to process 0 records I0320 09:56:40.266414 1 gather.go:177] gatherer "clusterconfig" function "jaegers" took 13.370105ms to process 0 records I0320 09:56:40.266434 1 tasks_processing.go:71] worker 14 working on monitoring_persistent_volumes task. I0320 09:56:40.266442 1 tasks_processing.go:71] worker 8 working on pdbs task. I0320 09:56:40.266544 1 recorder.go:75] Recording config/storage/storageclasses/gp2-csi with fingerprint=6a2452fca8e7422d8ab1725ca02e8e374db3d6917f15e5fe896200befad417e6 I0320 09:56:40.266571 1 recorder.go:75] Recording config/storage/storageclasses/gp3-csi with fingerprint=8fae3781d3dbede1d7bb567553dc86281f84da654a40c5dfd99d1f08eaf96426 I0320 09:56:40.266583 1 gather.go:177] gatherer "clusterconfig" function "storage_classes" took 13.435144ms to process 2 records I0320 09:56:40.266684 1 tasks_processing.go:71] worker 18 working on cluster_apiserver task. I0320 09:56:40.266942 1 recorder.go:75] Recording config/ingress with fingerprint=c4f6e5af70fd6e15cedf16186c519841a9f205b4a21cae5d02b3baebec77389d I0320 09:56:40.266957 1 gather.go:177] gatherer "clusterconfig" function "ingress" took 13.49399ms to process 1 records I0320 09:56:40.269471 1 tasks_processing.go:71] worker 1 working on service_accounts task. E0320 09:56:40.269518 1 gather.go:140] gatherer "clusterconfig" function "support_secret" failed with the error: secrets "support" not found I0320 09:56:40.269533 1 gather.go:177] gatherer "clusterconfig" function "support_secret" took 15.830784ms to process 0 records I0320 09:56:40.269707 1 tasks_processing.go:71] worker 27 working on machine_sets task. I0320 09:56:40.270000 1 recorder.go:75] Recording config/oauth with fingerprint=9f815493daaefd6f82bdc53d89130d0186c26fec9660fe5a51ef54eb95490a25 I0320 09:56:40.270015 1 gather.go:177] gatherer "clusterconfig" function "oauths" took 14.84916ms to process 1 records I0320 09:56:40.270101 1 tasks_processing.go:71] worker 13 working on storage_cluster task. I0320 09:56:40.270119 1 recorder.go:75] Recording config/featuregate with fingerprint=283a6f0ab3325ef91b1d1b76a9f66597a219701cd40e98faf486cc477b9b37ad I0320 09:56:40.270132 1 gather.go:177] gatherer "clusterconfig" function "feature_gates" took 12.127182ms to process 1 records I0320 09:56:40.270238 1 tasks_processing.go:71] worker 23 working on openstack_controlplanes task. E0320 09:56:40.270553 1 gather_node_features.go:86] GatherNodeFeatures: NodeFeatures resource not found in openshift-nfd namespace (may not be installed) I0320 09:56:40.270683 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/config/cluster with fingerprint=63d23eba5467b9dce96c5c6fce559baeaf201cd570e327720cb5a7a93611c98a I0320 09:56:40.270695 1 gather.go:177] gatherer "clusterconfig" function "image_registries" took 16.011763ms to process 1 records I0320 09:56:40.270705 1 gather.go:177] gatherer "clusterconfig" function "openstack_version" took 5.90832ms to process 0 records I0320 09:56:40.270708 1 gather.go:177] gatherer "clusterconfig" function "node_features" took 6.867124ms to process 0 records I0320 09:56:40.270715 1 tasks_processing.go:71] worker 0 working on operators_pods_and_events task. I0320 09:56:40.270776 1 tasks_processing.go:71] worker 31 working on certificate_signing_requests task. I0320 09:56:40.271567 1 tasks_processing.go:71] worker 21 working on validating_webhook_configurations task. E0320 09:56:40.271575 1 gather.go:140] gatherer "clusterconfig" function "machine_configs" failed with the error: getting MachineConfigPools failed: the server could not find the requested resource (get machineconfigpools.machineconfiguration.openshift.io) I0320 09:56:40.271601 1 recorder.go:75] Recording aggregated/unused_machine_configs_count with fingerprint=4bfc9fa984e5dfcd45848faaf05269de7619bf42edf9f781751af5ee05c1a499 I0320 09:56:40.271616 1 gather.go:177] gatherer "clusterconfig" function "machine_configs" took 18.051372ms to process 1 records I0320 09:56:40.277244 1 gather_logs.go:145] no pods in namespace were found I0320 09:56:40.277259 1 tasks_processing.go:71] worker 2 working on authentication task. I0320 09:56:40.277273 1 gather.go:177] gatherer "clusterconfig" function "qemu_kubevirt_launcher_logs" took 23.578469ms to process 0 records I0320 09:56:40.280996 1 tasks_processing.go:71] worker 17 working on openstack_dataplanedeployments task. I0320 09:56:40.281122 1 recorder.go:75] Recording config/network with fingerprint=04a2d3f5710e5b17bf35a0d7c43a8dcb41ca80e86ff1551a3e1c0d2bbcb39854 I0320 09:56:40.281135 1 gather.go:177] gatherer "clusterconfig" function "networks" took 25.996496ms to process 1 records I0320 09:56:40.281363 1 tasks_processing.go:71] worker 27 working on machine_autoscalers task. I0320 09:56:40.281364 1 gather.go:177] gatherer "clusterconfig" function "machine_sets" took 11.638958ms to process 0 records I0320 09:56:40.281676 1 tasks_processing.go:71] worker 20 working on operators task. I0320 09:56:40.281683 1 gather.go:177] gatherer "clusterconfig" function "ceph_cluster" took 18.256484ms to process 0 records I0320 09:56:40.281702 1 gather.go:177] gatherer "clusterconfig" function "storage_cluster" took 11.575304ms to process 0 records I0320 09:56:40.281710 1 tasks_processing.go:71] worker 13 working on config_maps task. I0320 09:56:40.281737 1 tasks_processing.go:71] worker 6 working on ingress_certificates task. I0320 09:56:40.281746 1 gather.go:177] gatherer "clusterconfig" function "node_logs" took 18.237541ms to process 0 records I0320 09:56:40.281887 1 gather.go:177] gatherer "clusterconfig" function "openshift_machine_api_events" took 28.305739ms to process 0 records I0320 09:56:40.281900 1 tasks_processing.go:71] worker 25 working on proxies task. I0320 09:56:40.282078 1 tasks_processing.go:74] worker 12 stopped. I0320 09:56:40.282091 1 gather.go:177] gatherer "clusterconfig" function "sap_pods" took 18.016833ms to process 0 records I0320 09:56:40.282208 1 tasks_processing.go:74] worker 28 stopped. I0320 09:56:40.282291 1 recorder.go:75] Recording config/image with fingerprint=ef43cacefe84cab4d1c2bed66b24da0f2a4c9572378928a2d9bad69fe4e2c233 I0320 09:56:40.282306 1 gather.go:177] gatherer "clusterconfig" function "image" took 27.758454ms to process 1 records I0320 09:56:40.282394 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/imagepruner/cluster with fingerprint=258ef3f0b59349a1dee1065ff2a3e13b6290ffb9b1aaf933aec0190534e256d1 I0320 09:56:40.282405 1 gather.go:177] gatherer "clusterconfig" function "image_pruners" took 27.527335ms to process 1 records I0320 09:56:40.282395 1 tasks_processing.go:74] worker 30 stopped. I0320 09:56:40.283163 1 tasks_processing.go:74] worker 4 stopped. I0320 09:56:40.283347 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/aws-pod-identity with fingerprint=409995b92aad76ab968d13f6813cd078f884431e1bec7c0533309408e534ec4b I0320 09:56:40.283432 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-podimagespec-mutation with fingerprint=205e3403aaacf663cae065ce6df9db776ae8d14035f220995b7fceda60a8605c I0320 09:56:40.283476 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-service-mutation with fingerprint=a6430717fd260f6e1019c3c240a2f674fea3fef1db5a78d98567997c506e7946 I0320 09:56:40.283486 1 gather.go:177] gatherer "clusterconfig" function "mutating_webhook_configurations" took 29.152123ms to process 3 records I0320 09:56:40.283747 1 tasks_processing.go:74] worker 7 stopped. I0320 09:56:40.283814 1 recorder.go:75] Recording config/schedulers/cluster with fingerprint=0111e20560b7f54e35156cec71ee4ed9dd39fc0b1bd51fd03f83d183af1d2d38 I0320 09:56:40.283825 1 gather.go:177] gatherer "clusterconfig" function "schedulers" took 30.714896ms to process 1 records I0320 09:56:40.283833 1 gather.go:177] gatherer "clusterconfig" function "openstack_controlplanes" took 13.517934ms to process 0 records I0320 09:56:40.283838 1 tasks_processing.go:74] worker 23 stopped. I0320 09:56:40.286731 1 tasks_processing.go:74] worker 3 stopped. E0320 09:56:40.286750 1 gather.go:140] gatherer "clusterconfig" function "pod_network_connectivity_checks" failed with the error: the server could not find the requested resource (get podnetworkconnectivitychecks.controlplane.operator.openshift.io) I0320 09:56:40.286760 1 gather.go:177] gatherer "clusterconfig" function "pod_network_connectivity_checks" took 21.522705ms to process 0 records I0320 09:56:40.287024 1 tasks_processing.go:74] worker 8 stopped. I0320 09:56:40.287118 1 recorder.go:75] Recording config/pdbs/openshift-image-registry/image-registry with fingerprint=f41c7a15f3d607a255f2c6dc9d21a5d6aa8f86b607ea0ce7f0f5561d82bad9c2 I0320 09:56:40.287144 1 recorder.go:75] Recording config/pdbs/openshift-ingress/router-default with fingerprint=94c2e57cfca938a8ec49b7da6a6f052779f047b7345d4b34b330fb2c64d5f97c I0320 09:56:40.287158 1 recorder.go:75] Recording config/pdbs/openshift-operator-lifecycle-manager/packageserver-pdb with fingerprint=35da203b991a4da1ff28f41463ce7c0f8688e5dd99beb8d8a057e3d454525e9a I0320 09:56:40.287167 1 gather.go:177] gatherer "clusterconfig" function "pdbs" took 20.572551ms to process 3 records I0320 09:56:40.287305 1 tasks_processing.go:74] worker 18 stopped. I0320 09:56:40.287483 1 recorder.go:75] Recording config/apiserver with fingerprint=3c0531c0bd0a1575138c3851c5ce08c8f1b5127c0eb8d83eb9dcf52801c5b7fc I0320 09:56:40.287526 1 gather.go:177] gatherer "clusterconfig" function "cluster_apiserver" took 20.610197ms to process 1 records I0320 09:56:40.287727 1 tasks_processing.go:74] worker 17 stopped. I0320 09:56:40.287774 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanedeployments" took 6.71655ms to process 0 records I0320 09:56:40.288099 1 tasks_processing.go:74] worker 11 stopped. I0320 09:56:40.288589 1 recorder.go:75] Recording config/node/ip-10-0-0-174.ec2.internal with fingerprint=e22a305be27c05c09a8168a0c01857d2674e0a8cb3e7b1fc37b300ba2a5bfc65 I0320 09:56:40.288649 1 recorder.go:75] Recording config/node/ip-10-0-1-80.ec2.internal with fingerprint=8c795f7fb80298e1b82e6dc0e57917908925636ae6daa51f812f1036b8480a5c I0320 09:56:40.288728 1 recorder.go:75] Recording config/node/ip-10-0-2-72.ec2.internal with fingerprint=8831b754f3426563ed441058ea0b950a63b5835ed66404b866d67cf38e019757 I0320 09:56:40.288744 1 gather.go:177] gatherer "clusterconfig" function "nodes" took 33.56174ms to process 3 records I0320 09:56:40.288765 1 tasks_processing.go:74] worker 21 stopped. I0320 09:56:40.288938 1 recorder.go:75] Recording config/validatingwebhookconfigurations/multus.openshift.io with fingerprint=b2e946804e70f0b136ecfb5bc6082074142816e3e68b6ed1569428210208656c I0320 09:56:40.289067 1 recorder.go:75] Recording config/validatingwebhookconfigurations/network-node-identity.openshift.io with fingerprint=3c4c8adbb9f1334535211675dd4ff0428d4088c7f7647731685c11ddee67eacc I0320 09:56:40.289103 1 recorder.go:75] Recording config/validatingwebhookconfigurations/performance-addon-operator with fingerprint=6f1e9039d7024f2b1a13829edc9e82fb31457efa290109f259f46759cbbfddd7 I0320 09:56:40.289199 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterrolebindings-validation with fingerprint=7d47a4677901283660843276235cb78fbc518ee9c747ab206c4f897b6f8b34fd I0320 09:56:40.289246 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterroles-validation with fingerprint=69dcd8693446bff183e87b64d0ed72b59c12fd30185c0c325e0cda8ccca6378e I0320 09:56:40.289283 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-ingress-config-validation with fingerprint=1605020dc7f2fcd4b99d95166c37af96aa1b6ef409a8684d5974036b5e287392 I0320 09:56:40.289339 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-network-operator-validation with fingerprint=6ac3ababc5cb73ad9931a9c1566f949fc9a96d3b347980f41d86514f1c2680ec I0320 09:56:40.289417 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-regular-user-validation with fingerprint=aea2f918916ea93426e863dbc2cbaee1b371cc4a44cf4edd60cec718f0d12296 I0320 09:56:40.289462 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-scc-validation with fingerprint=2c0e377674c565afd2da490669dd80de20166d6de400122231fdf264ba089acf I0320 09:56:40.289499 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-serviceaccount-validation with fingerprint=ec37fa532af0452b6c4ba57e8e28a4ab8fc37aef97b02746400589fe0467811b I0320 09:56:40.289537 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-techpreviewnoupgrade-validation with fingerprint=0b89564e8cebf9f0ce99d293a992f5496c6f665cb3f1b3c31ccb778b95b23c44 I0320 09:56:40.289549 1 gather.go:177] gatherer "clusterconfig" function "validating_webhook_configurations" took 17.178206ms to process 11 records I0320 09:56:40.289575 1 recorder.go:75] Recording config/namespaces_with_overlapping_uids with fingerprint=4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945 I0320 09:56:40.289585 1 gather.go:177] gatherer "clusterconfig" function "overlapping_namespace_uids" took 36.287531ms to process 1 records I0320 09:56:40.289593 1 tasks_processing.go:74] worker 22 stopped. I0320 09:56:40.290262 1 tasks_processing.go:74] worker 31 stopped. I0320 09:56:40.290276 1 gather.go:177] gatherer "clusterconfig" function "certificate_signing_requests" took 19.468805ms to process 0 records W0320 09:56:40.290831 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0320 09:56:40.291778 1 tasks_processing.go:74] worker 27 stopped. I0320 09:56:40.291788 1 gather.go:177] gatherer "clusterconfig" function "machine_autoscalers" took 10.384929ms to process 0 records I0320 09:56:40.296100 1 tasks_processing.go:74] worker 2 stopped. I0320 09:56:40.296328 1 recorder.go:75] Recording config/authentication with fingerprint=d1b0a28535c6481cf8eee37f05c8aa67f2697df1a0768d16f03ed9a329d9c8a9 I0320 09:56:40.296345 1 gather.go:177] gatherer "clusterconfig" function "authentication" took 18.830568ms to process 1 records I0320 09:56:40.299394 1 tasks_processing.go:74] worker 14 stopped. I0320 09:56:40.299409 1 gather.go:177] gatherer "clusterconfig" function "monitoring_persistent_volumes" took 32.949167ms to process 0 records I0320 09:56:40.299535 1 tasks_processing.go:74] worker 10 stopped. I0320 09:56:40.299635 1 sca.go:136] Pulling SCA certificates from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates. Next check is in 8h0m0s I0320 09:56:40.299652 1 cluster_transfer.go:83] checking the availability of cluster transfer. Next check is in 12h0m0s W0320 09:56:40.300109 1 operator.go:328] started I0320 09:56:40.300203 1 base_controller.go:76] Waiting for caches to sync for LoggingSyncer I0320 09:56:40.301655 1 recorder.go:75] Recording config/infrastructure with fingerprint=103e89c1e4e4635703e2d0cfac1371939c85e1cc7331ff8168ca28c1e55e436d I0320 09:56:40.301727 1 gather.go:177] gatherer "clusterconfig" function "infrastructures" took 33.130513ms to process 1 records I0320 09:56:40.302006 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/admin with fingerprint=75939e6dc60859bc176a783a5f29ac368bcf234be60f571d05f3b26256261f92 I0320 09:56:40.302178 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/edit with fingerprint=73d4e3ce20a69305fc85f27389d23362a8f67efa1bd50a7ebef329837e638f4a I0320 09:56:40.302209 1 gather.go:177] gatherer "clusterconfig" function "clusterroles" took 37.19383ms to process 2 records I0320 09:56:40.302346 1 recorder.go:75] Recording config/proxy with fingerprint=e89e9f6246cfdc85fb0b5d788e5bd1e97ad8cb93ba932a051c870ecc2d7dadb4 I0320 09:56:40.302390 1 gather.go:177] gatherer "clusterconfig" function "proxies" took 18.995476ms to process 1 records I0320 09:56:40.302590 1 tasks_processing.go:74] worker 29 stopped. I0320 09:56:40.302655 1 tasks_processing.go:74] worker 25 stopped. I0320 09:56:40.302734 1 tasks_processing.go:74] worker 15 stopped. I0320 09:56:40.304832 1 recorder.go:75] Recording config/crd/volumesnapshots.snapshot.storage.k8s.io with fingerprint=d0a0a70624a08ab11fafe119691f6de1fd799f507d317e53e05ed2c6e20b87ab I0320 09:56:40.305147 1 recorder.go:75] Recording config/crd/volumesnapshotcontents.snapshot.storage.k8s.io with fingerprint=6bd49c2bf4d1a2764f79a5021f32ea144ea97524aa1ccdd45832f96a6497eaa5 I0320 09:56:40.305162 1 gather.go:177] gatherer "clusterconfig" function "crds" took 38.491718ms to process 2 records I0320 09:56:40.305173 1 gather.go:177] gatherer "clusterconfig" function "aggregated_monitoring_cr_names" took 39.174107ms to process 0 records I0320 09:56:40.305180 1 tasks_processing.go:74] worker 9 stopped. I0320 09:56:40.312389 1 tasks_processing.go:74] worker 19 stopped. I0320 09:56:40.312410 1 prometheus_rules.go:88] Prometheus rules successfully created I0320 09:56:40.314041 1 recorder.go:75] Recording config/pod/openshift-ovn-kubernetes/ovnkube-node-8dgxz with fingerprint=edf0d26180840daabd1574f8f88bb856bb0e8232233f03a7c0c43ba6658c0cc5 I0320 09:56:40.314262 1 recorder.go:75] Recording config/pod/openshift-ovn-kubernetes/ovnkube-node-n7cth with fingerprint=021c1d8135889ec0707ad3674a4802c684ead5cdfa5de3338280408efc79a0e3 I0320 09:56:40.314314 1 recorder.go:75] Recording config/running_containers with fingerprint=205532389647563ed9c749a24c7b3e36149a104f5e490a55d802c8197bc4fe90 I0320 09:56:40.314340 1 gather.go:177] gatherer "clusterconfig" function "container_images" took 59.324895ms to process 3 records I0320 09:56:40.314351 1 gather.go:177] gatherer "clusterconfig" function "number_of_pods_and_netnamespaces_with_sdn_annotations" took 58.927412ms to process 0 records I0320 09:56:40.314358 1 tasks_processing.go:74] worker 16 stopped. I0320 09:56:40.315340 1 tasks_processing.go:74] worker 5 stopped. I0320 09:56:40.315567 1 recorder.go:75] Recording config/version with fingerprint=7a19c61e8e2f37a30c46155345fbb824208738e0fdd075719ea5c19c7885ce90 I0320 09:56:40.315582 1 recorder.go:75] Recording config/id with fingerprint=5c7fdb3df40049f7203f59e858310bc688b55cabaf82d7d5670beafb3c81f00a I0320 09:56:40.315588 1 gather.go:177] gatherer "clusterconfig" function "version" took 60.291772ms to process 2 records I0320 09:56:40.315668 1 controller.go:254] Source clusterTransferController *clustertransfer.Controller is not ready I0320 09:56:40.315684 1 controller.go:254] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0320 09:56:40.315688 1 controller.go:254] Source periodic-conditional *controllerstatus.Simple is not ready I0320 09:56:40.315691 1 controller.go:254] Source periodic-workloads *controllerstatus.Simple is not ready I0320 09:56:40.315694 1 controller.go:254] Source scaController *sca.Controller is not ready I0320 09:56:40.315709 1 controller.go:531] The operator is still being initialized I0320 09:56:40.315715 1 controller.go:554] The operator is healthy I0320 09:56:40.320613 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0320 09:56:40.320637 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController I0320 09:56:40.320668 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file E0320 09:56:40.322917 1 cluster_transfer.go:95] failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%2734494c97-6cd9-4ed8-b867-ecd8ceb844cf%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.130.0.9:48721->172.30.0.10:53: read: connection refused I0320 09:56:40.322928 1 controllerstatus.go:80] name=clusterTransferController healthy=true reason=Disconnected message=failed to pull cluster transfer: unable to retrieve cluster transfer data from https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/: Get "https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/?search=cluster_uuid+is+%2734494c97-6cd9-4ed8-b867-ecd8ceb844cf%27+and+status+is+%27accepted%27": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.130.0.9:48721->172.30.0.10:53: read: connection refused I0320 09:56:40.325003 1 base_controller.go:82] Caches are synced for ConfigController I0320 09:56:40.325014 1 base_controller.go:119] Starting #1 worker of ConfigController controller ... I0320 09:56:40.334474 1 configmapobserver.go:84] configmaps "insights-config" not found I0320 09:56:40.352946 1 tasks_processing.go:74] worker 13 stopped. E0320 09:56:40.352961 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "cluster-monitoring-config" not found E0320 09:56:40.352972 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "gateway-mode-config" not found E0320 09:56:40.352976 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "insights-config" not found I0320 09:56:40.352986 1 recorder.go:75] Recording config/configmaps/openshift-config/installer-images/images.json with fingerprint=c8a9805405aae7d04b7f03c09c51097942070877009b6daebd5573ef8127446e I0320 09:56:40.353020 1 recorder.go:75] Recording config/configmaps/openshift-config/kube-root-ca.crt/ca.crt with fingerprint=d476c7d3f5b104863f08f481b1264dcc68cc272ecefb0ecb709b18a6afab034d I0320 09:56:40.353028 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/invoker with fingerprint=76b482f683cd3ef9da02debac5b26080a5aeb06ff768ee5c21117514dff29d8a I0320 09:56:40.353032 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/version with fingerprint=3b70b9787487b548204d806ebebf90b0c2d460145e34cb33c0d8d3a7a011d907 I0320 09:56:40.353036 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 I0320 09:56:40.353073 1 recorder.go:75] Recording config/configmaps/openshift-config/rosa-brand-logo/rosa-brand-logo.svg with fingerprint=6ed8ca4dd7a8eee7249182bc006e9649ce84d76c551ddfaaa33e55d8c4cc1ed0 I0320 09:56:40.353080 1 recorder.go:75] Recording config/configmaps/kube-system/cluster-config-v1/install-config with fingerprint=ab3811c6b83fd7b8e920094cfa3080d1b4ee3c35ec4c8379437b21d27bd6608d I0320 09:56:40.353086 1 gather.go:177] gatherer "clusterconfig" function "config_maps" took 71.221596ms to process 7 records I0320 09:56:40.385928 1 requests.go:205] Asking for SCA certificate with "{"arch": ["x86_64"]}" payload W0320 09:56:40.389437 1 sca.go:161] Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.130.0.9:33208->172.30.0.10:53: read: connection refused I0320 09:56:40.389449 1 controllerstatus.go:80] name=scaController healthy=true reason=NonHTTPError message=Failed to pull SCA certs from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: unable to retrieve SCA certs data from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates: Post "https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates": dial tcp: lookup api.openshift.com on 172.30.0.10:53: read udp 10.130.0.9:33208->172.30.0.10:53: read: connection refused I0320 09:56:40.401027 1 base_controller.go:82] Caches are synced for LoggingSyncer I0320 09:56:40.401039 1 base_controller.go:119] Starting #1 worker of LoggingSyncer controller ... I0320 09:56:40.401889 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0320 09:56:40.420957 1 tasks_processing.go:74] worker 6 stopped. E0320 09:56:40.420974 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret 'router-certs-default' in namespace 'openshift-ingress': secrets "router-certs-default" not found E0320 09:56:40.420984 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret '2p5cer0igbj7c319rhdi0r3gc63kvskh-primary-cert-bundle-secret' in namespace 'openshift-ingress-operator': secrets "2p5cer0igbj7c319rhdi0r3gc63kvskh-primary-cert-bundle-secret" not found I0320 09:56:40.421067 1 recorder.go:75] Recording aggregated/ingress_controllers_certs with fingerprint=5cd73ffdab81b20d934982929ab352169eadd0c0b776f6d6807ffaf8970192be I0320 09:56:40.421084 1 gather.go:177] gatherer "clusterconfig" function "ingress_certificates" took 139.155157ms to process 1 records I0320 09:56:40.712186 1 gather_cluster_operator_pods_and_events.go:121] Found 18 pods with 21 containers I0320 09:56:40.712201 1 gather_cluster_operator_pods_and_events.go:235] Maximum buffer size: 1198372 bytes I0320 09:56:40.712213 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-26hcj pod in namespace openshift-dns (previous: false). I0320 09:56:40.919131 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-26hcj pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-26hcj\" is waiting to start: ContainerCreating" I0320 09:56:40.919147 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-26hcj\" is waiting to start: ContainerCreating" I0320 09:56:40.919156 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-26hcj pod in namespace openshift-dns (previous: false). I0320 09:56:41.118114 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-26hcj pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-26hcj\" is waiting to start: ContainerCreating" I0320 09:56:41.118129 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-26hcj\" is waiting to start: ContainerCreating" I0320 09:56:41.118139 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-6kvtj pod in namespace openshift-dns (previous: false). W0320 09:56:41.288530 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0320 09:56:41.356925 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-6kvtj pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-6kvtj\" is waiting to start: ContainerCreating" I0320 09:56:41.356944 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-6kvtj\" is waiting to start: ContainerCreating" I0320 09:56:41.356951 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-6kvtj pod in namespace openshift-dns (previous: false). I0320 09:56:41.549527 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-6kvtj pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-6kvtj\" is waiting to start: ContainerCreating" I0320 09:56:41.549543 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-6kvtj\" is waiting to start: ContainerCreating" I0320 09:56:41.549555 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns container dns-default-7c6hk pod in namespace openshift-dns (previous: false). I0320 09:56:41.744182 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-7c6hk pod in namespace openshift-dns for failing operator dns (previous: false): "container \"dns\" in pod \"dns-default-7c6hk\" is waiting to start: ContainerCreating" I0320 09:56:41.744219 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"dns\" in pod \"dns-default-7c6hk\" is waiting to start: ContainerCreating" I0320 09:56:41.744229 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container dns-default-7c6hk pod in namespace openshift-dns (previous: false). I0320 09:56:41.917086 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for dns-default-7c6hk pod in namespace openshift-dns for failing operator kube-rbac-proxy (previous: false): "container \"kube-rbac-proxy\" in pod \"dns-default-7c6hk\" is waiting to start: ContainerCreating" I0320 09:56:41.917103 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"kube-rbac-proxy\" in pod \"dns-default-7c6hk\" is waiting to start: ContainerCreating" I0320 09:56:41.917120 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-4ps6f pod in namespace openshift-dns (previous: false). I0320 09:56:41.929620 1 gather_cluster_operators.go:184] Unable to get configs.samples.operator.openshift.io resource due to: configs.samples.operator.openshift.io "cluster" not found I0320 09:56:42.125504 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0320 09:56:42.125520 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-ftfkj pod in namespace openshift-dns (previous: false). W0320 09:56:42.288226 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0320 09:56:42.319173 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0320 09:56:42.319187 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for dns-node-resolver container node-resolver-rgb95 pod in namespace openshift-dns (previous: false). I0320 09:56:42.516496 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0320 09:56:42.516513 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-6554496cf7-494b4 pod in namespace openshift-image-registry (previous: false). I0320 09:56:42.530632 1 tasks_processing.go:74] worker 20 stopped. I0320 09:56:42.530675 1 recorder.go:75] Recording config/clusteroperator/console with fingerprint=247d33007311311a56ffae123089e77e08b740a8334304710bcc935b7cbf89b3 I0320 09:56:42.530714 1 recorder.go:75] Recording config/clusteroperator/csi-snapshot-controller with fingerprint=e9cf61170e7f5d6b89ca0614ec20f7efb181920ed1d393abc58982059f3aa5b5 I0320 09:56:42.530745 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/csisnapshotcontroller/cluster with fingerprint=5adc514f4b63e2f1ecc68bf6f9c0af70c5eea04522a49524e102721b1c41f80e I0320 09:56:42.530768 1 recorder.go:75] Recording config/clusteroperator/dns with fingerprint=456e73d93d3bdb25b594e969856820995b0d75afe2bb9477791323b24c4b7226 I0320 09:56:42.530799 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/dns/default with fingerprint=9e7b4ce029030d3d8c3b49af92c556acdcc415000b40d3f969dbdc42c432b47f I0320 09:56:42.530823 1 recorder.go:75] Recording config/clusteroperator/image-registry with fingerprint=b8523563360dd53e2000b3a17623eba5ef1da2a6423cd6cd5b3a5d8263cfa8a8 I0320 09:56:42.530852 1 recorder.go:75] Recording config/clusteroperator/ingress with fingerprint=2bdfc138931a9629e8c49fe3108bb791fbce09cad01119d14bfc8c67186ebba4 I0320 09:56:42.530875 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/ingresscontroller/openshift-ingress-operator/default with fingerprint=2314684c649957905898110bbd7b3ccc20618914d1184b6a3a8ddfae93c66af8 I0320 09:56:42.530902 1 recorder.go:75] Recording config/clusteroperator/insights with fingerprint=c15e2aab146e456508206b9482079ecec97f69247e57d011942e1b433357d0aa I0320 09:56:42.530912 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/insightsoperator/cluster with fingerprint=e5ff11d57817f84a678f6fa9565af55bd1120227c16a21933637ab62675a6d70 I0320 09:56:42.530928 1 recorder.go:75] Recording config/clusteroperator/kube-apiserver with fingerprint=eb7c3b7594a9a663cb719582cb4a153f9e32b3359e5c99470060275752f31967 I0320 09:56:42.530937 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubeapiserver/cluster with fingerprint=51503bf0b784fcf65ea46bcaf1f72ac1a5c4d5dc211934f18f27871efed05762 I0320 09:56:42.530953 1 recorder.go:75] Recording config/clusteroperator/kube-controller-manager with fingerprint=c24fa8cb12f2e5ca1e78046159a044ace33626f732f18a196c5965f8b2b634a7 I0320 09:56:42.530964 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubecontrollermanager/cluster with fingerprint=ce90c0d4f367d7da085074268031798382ae7c54fdcb0a21f15a4818fe308c11 I0320 09:56:42.530979 1 recorder.go:75] Recording config/clusteroperator/kube-scheduler with fingerprint=a88c3a7e0d39d386a0c5818de799933125e02296c3fa06b125551f1b8574cd48 I0320 09:56:42.530986 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubescheduler/cluster with fingerprint=f2940fb9fd20c19951dfc295eb363b7fba0c505f5ae61f01967a063099e6b60a I0320 09:56:42.530998 1 recorder.go:75] Recording config/clusteroperator/kube-storage-version-migrator with fingerprint=1db4f6cf58748334fdb217fa1c4bbb21e1a3952268943e3ed9d85987de21d71a I0320 09:56:42.531007 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubestorageversionmigrator/cluster with fingerprint=9351181aa7e6ada41ef581ab31e13516c6b934cc95710154bafb2eb222cb58db I0320 09:56:42.531032 1 recorder.go:75] Recording config/clusteroperator/monitoring with fingerprint=ad00af109f58df143049216eef60b6991dde4b65c78557cf1a7779c3b8813ecb I0320 09:56:42.531163 1 recorder.go:75] Recording config/clusteroperator/network with fingerprint=69320d75a15798ea92fd8e36cde34027d2da8845f84a02b94844e4ddeeed9e28 I0320 09:56:42.531181 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/ovn with fingerprint=626a89d20e0deaed5b6dfb533acfe65f4bb1618bd200a703b62e60c5d16d94ab I0320 09:56:42.531190 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/signer with fingerprint=90410b16914712b85b3c4578716ad8c0ae072e688f4cd1e022bf76f20da3506d I0320 09:56:42.531211 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/network/cluster with fingerprint=a93d15eaecb455a0e40ecb2826eeecc1533899204ddd3c3921d15ab70af7ae75 I0320 09:56:42.531234 1 recorder.go:75] Recording config/clusteroperator/node-tuning with fingerprint=4f975cb520f474b567d88fdabb7d2e320395a84065fbe8c0f39343e6a1d1a638 I0320 09:56:42.531257 1 recorder.go:75] Recording config/clusteroperator/openshift-apiserver with fingerprint=f93a9fb444d89481519538244ade6068604e7781b6707f1e2531db6aaef944f9 I0320 09:56:42.531271 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftapiserver/cluster with fingerprint=e712e6cf27339b441e4ed1f4cde91dbde7e952698ba93407e4457db63a4a4c76 I0320 09:56:42.531289 1 recorder.go:75] Recording config/clusteroperator/openshift-controller-manager with fingerprint=2348d1c172190c01d6be3bbb33e1e272c4b228fb04a08a3a68a5f33e9f0bc0eb I0320 09:56:42.531298 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftcontrollermanager/cluster with fingerprint=d71a0f4672f9b45d9fc8293bf1687afc650fd28d32e2e30de27523fe7b4eadf7 I0320 09:56:42.531312 1 recorder.go:75] Recording config/clusteroperator/openshift-samples with fingerprint=08660f3194b651c704feade9a2df2502d7e07b4cff85fb1e43573b9836ed55cd I0320 09:56:42.531338 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager with fingerprint=a4554ad3de672f59bb4c2ab8bd8d7547d40289f6955297ae7cc868620fb6b95f I0320 09:56:42.531352 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-catalog with fingerprint=e812b226d33806137aee18ae5c14ad307dc5edce5fd6539f13d6fa6b0bfdfee4 I0320 09:56:42.531367 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-packageserver with fingerprint=42ca972c4bb17be2a323eedbc8bb090a404b6a250eb4689a79cee5504b0fcc56 I0320 09:56:42.531386 1 recorder.go:75] Recording config/clusteroperator/service-ca with fingerprint=94acc7fdd2af2572fe68a6bca3c8cf49dc87d80321aa5d920d6c6542d1313759 I0320 09:56:42.531395 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/serviceca/cluster with fingerprint=812f7edc2cdb30e61e7f2b29454357a40b1a507a4b0c2b7729193b67f0e3b4aa I0320 09:56:42.531419 1 recorder.go:75] Recording config/clusteroperator/storage with fingerprint=51e15fc1083669ba120a97ac757243a2834a22cafffe0cd34f72e4040199dc90 I0320 09:56:42.531436 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/clustercsidriver/ebs.csi.aws.com with fingerprint=510064d6f6bcced87ab5bd2ddaff3d0edd7f93f4a4f7af2641f29fc53ffab21e I0320 09:56:42.531443 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/storage/cluster with fingerprint=8e480f8c1ce1b39baac42d8ec780c57c2592929ae0c801b61ffad49ba13f33ad I0320 09:56:42.531450 1 gather.go:177] gatherer "clusterconfig" function "operators" took 2.248935051s to process 37 records I0320 09:56:42.716025 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-6554496cf7-494b4 pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-6554496cf7-494b4\" is waiting to start: ContainerCreating" I0320 09:56:42.716046 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-6554496cf7-494b4\" is waiting to start: ContainerCreating" I0320 09:56:42.716056 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-b8f67457d-25dkw pod in namespace openshift-image-registry (previous: false). I0320 09:56:42.918771 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-b8f67457d-25dkw pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-b8f67457d-25dkw\" is waiting to start: ContainerCreating" I0320 09:56:42.918792 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-b8f67457d-25dkw\" is waiting to start: ContainerCreating" I0320 09:56:42.918815 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for registry container image-registry-b8f67457d-bhg2d pod in namespace openshift-image-registry (previous: false). I0320 09:56:43.117627 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for image-registry-b8f67457d-bhg2d pod in namespace openshift-image-registry for failing operator registry (previous: false): "container \"registry\" in pod \"image-registry-b8f67457d-bhg2d\" is waiting to start: ContainerCreating" I0320 09:56:43.117675 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"registry\" in pod \"image-registry-b8f67457d-bhg2d\" is waiting to start: ContainerCreating" I0320 09:56:43.117691 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-fmcw2 pod in namespace openshift-image-registry (previous: false). W0320 09:56:43.290243 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0320 09:56:43.318687 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0320 09:56:43.318709 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-wnbw2 pod in namespace openshift-image-registry (previous: false). I0320 09:56:43.517716 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0320 09:56:43.517732 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-ca container node-ca-x4mmc pod in namespace openshift-image-registry (previous: false). I0320 09:56:43.716573 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0320 09:56:43.716591 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-5758fdcb8c-f9cdv pod in namespace openshift-ingress (previous: false). I0320 09:56:43.918909 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-5758fdcb8c-f9cdv pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-5758fdcb8c-f9cdv\" is waiting to start: ContainerCreating" I0320 09:56:43.918926 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-5758fdcb8c-f9cdv\" is waiting to start: ContainerCreating" I0320 09:56:43.918936 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-6dd445d79-67h89 pod in namespace openshift-ingress (previous: false). I0320 09:56:44.117953 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-6dd445d79-67h89 pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-6dd445d79-67h89\" is waiting to start: ContainerCreating" I0320 09:56:44.117973 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-6dd445d79-67h89\" is waiting to start: ContainerCreating" I0320 09:56:44.117991 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for router container router-default-6dd445d79-p7tzc pod in namespace openshift-ingress (previous: false). W0320 09:56:44.288428 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0320 09:56:44.316995 1 gather_cluster_operator_pods_and_events.go:408] Failed to fetch log for router-default-6dd445d79-p7tzc pod in namespace openshift-ingress for failing operator router (previous: false): "container \"router\" in pod \"router-default-6dd445d79-p7tzc\" is waiting to start: ContainerCreating" I0320 09:56:44.317010 1 gather_cluster_operator_pods_and_events.go:280] Error: "container \"router\" in pod \"router-default-6dd445d79-p7tzc\" is waiting to start: ContainerCreating" I0320 09:56:44.317020 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-gflcx pod in namespace openshift-ingress-canary (previous: false). I0320 09:56:44.517550 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-hv6dq pod in namespace openshift-ingress-canary (previous: false). I0320 09:56:44.719721 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for serve-healthcheck-canary container ingress-canary-vbsj8 pod in namespace openshift-ingress-canary (previous: false). I0320 09:56:44.924729 1 tasks_processing.go:74] worker 0 stopped. I0320 09:56:44.924818 1 recorder.go:75] Recording events/openshift-dns-operator with fingerprint=429d7c8fe51da3e27835113dbfbb4b38510f8b2e44be0702b935781587e3eb24 I0320 09:56:44.924868 1 recorder.go:75] Recording events/openshift-dns with fingerprint=6a193858d3f26eeda4d80fe850ec42e6e2f815d655f559125421d0619891740d I0320 09:56:44.924959 1 recorder.go:75] Recording events/openshift-image-registry with fingerprint=b5ac66daa4f75d38697dbfcbd6bac735f27ebefdf94f4c130ebe6b1537caf932 I0320 09:56:44.924987 1 recorder.go:75] Recording events/openshift-ingress-operator with fingerprint=79dcbbfd08779e92b2038fc0f279eb44f1996b4122154d00a29eb85abd108b88 I0320 09:56:44.925034 1 recorder.go:75] Recording events/openshift-ingress with fingerprint=8796937ffa3f4117fd3177fc68d07ef9810d4653c106f00e7c469086d59a6db3 I0320 09:56:44.925082 1 recorder.go:75] Recording events/openshift-ingress-canary with fingerprint=790187da2e1d338d5d6c562887f0ee7cfd54cbfd0b0c4e4ae4b974308fe87b83 I0320 09:56:44.925091 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/logs/ingress-canary-gflcx/serve-healthcheck-canary_current.log with fingerprint=ab6bfe37f8a51eff83882f2e58dd8e6e5c205a89dc4849a638c7a930cee6ebfd I0320 09:56:44.925095 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/logs/ingress-canary-hv6dq/serve-healthcheck-canary_current.log with fingerprint=db739d35982e74fafdf056e9d5249c38660f10496b81711cfc40965bdc90cf95 I0320 09:56:44.925099 1 recorder.go:75] Recording config/pod/openshift-ingress-canary/logs/ingress-canary-vbsj8/serve-healthcheck-canary_current.log with fingerprint=37b11d6465efaaf62b84254c4e91a496df4bb99c3e7a1abd06ca07595cb93d4f I0320 09:56:44.925106 1 gather.go:177] gatherer "clusterconfig" function "operators_pods_and_events" took 4.653996433s to process 9 records W0320 09:56:45.284930 1 gather_dvo_metrics.go:117] Unable to read metrics from endpoint "http://deployment-validation-operator-metrics.openshift-deployment-validation-operator.svc:8383": DVO metrics service was not available within the 5s timeout: context deadline exceeded I0320 09:56:45.284954 1 tasks_processing.go:74] worker 26 stopped. E0320 09:56:45.284964 1 gather.go:140] gatherer "clusterconfig" function "dvo_metrics" failed with the error: DVO metrics service was not available within the 5s timeout: context deadline exceeded I0320 09:56:45.284977 1 recorder.go:75] Recording config/dvo_metrics with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 W0320 09:56:45.284993 1 gather.go:155] issue recording gatherer "clusterconfig" function "dvo_metrics" result "config/dvo_metrics" because of the warning: warning: the record with the same fingerprint "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" was already recorded at path "config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt", recording another one with a different path "config/dvo_metrics" I0320 09:56:45.285003 1 gather.go:177] gatherer "clusterconfig" function "dvo_metrics" took 5.030932254s to process 1 records I0320 09:56:52.405866 1 configmapobserver.go:84] configmaps "insights-config" not found I0320 09:56:52.896015 1 tasks_processing.go:74] worker 24 stopped. I0320 09:56:52.896061 1 recorder.go:75] Recording config/installplans with fingerprint=7b887df561a3a9e6ef0dc672845aa5d56e348505006b7496d3a2f83892b0c95b I0320 09:56:52.896078 1 gather.go:177] gatherer "clusterconfig" function "install_plans" took 12.642149517s to process 1 records I0320 09:56:53.675485 1 tasks_processing.go:74] worker 1 stopped. I0320 09:56:53.675772 1 recorder.go:75] Recording config/serviceaccounts with fingerprint=a07873ad19c38a900d225cd4e761acaa82a2cca4c691307b4c7c64dc85e3e73c I0320 09:56:53.675790 1 gather.go:177] gatherer "clusterconfig" function "service_accounts" took 13.405991629s to process 1 records E0320 09:56:53.675856 1 periodic.go:254] "Unhandled Error" err="clusterconfig failed after 13.423s with: function \"machines\" failed with an error, function \"machine_healthchecks\" failed with an error, function \"support_secret\" failed with an error, function \"machine_configs\" failed with an error, function \"pod_network_connectivity_checks\" failed with an error, function \"config_maps\" failed with an error, function \"ingress_certificates\" failed with an error, function \"dvo_metrics\" failed with an error" I0320 09:56:53.676962 1 controllerstatus.go:89] name=periodic-clusterconfig healthy=false reason=PeriodicGatherFailed message=Source clusterconfig could not be retrieved: function "machines" failed with an error, function "machine_healthchecks" failed with an error, function "support_secret" failed with an error, function "machine_configs" failed with an error, function "pod_network_connectivity_checks" failed with an error, function "config_maps" failed with an error, function "ingress_certificates" failed with an error, function "dvo_metrics" failed with an error I0320 09:56:53.676976 1 periodic.go:216] Running workloads gatherer I0320 09:56:53.676992 1 tasks_processing.go:45] number of workers: 2 I0320 09:56:53.677000 1 tasks_processing.go:69] worker 1 listening for tasks. I0320 09:56:53.677006 1 tasks_processing.go:71] worker 1 working on workload_info task. I0320 09:56:53.677008 1 tasks_processing.go:69] worker 0 listening for tasks. I0320 09:56:53.677087 1 tasks_processing.go:71] worker 0 working on helmchart_info task. I0320 09:56:53.701796 1 gather_workloads_info.go:278] Loaded pods in 0s, will wait 22s for image data I0320 09:56:53.702946 1 tasks_processing.go:74] worker 0 stopped. I0320 09:56:53.702965 1 gather.go:177] gatherer "workloads" function "helmchart_info" took 25.836991ms to process 0 records I0320 09:56:53.711585 1 gather_workloads_info.go:387] No image sha256:3e889ea85c9c4748c548b84ae135a8adc88f42026ce8073538ca2fe7d89b8004 (10ms) I0320 09:56:53.720310 1 gather_workloads_info.go:387] No image sha256:1bf773bc4ec02bf6d68c82f32175366b01fb24f6da8c724b7621748692aec7dc (9ms) I0320 09:56:53.728777 1 gather_workloads_info.go:387] No image sha256:0c90ca6acbe2f4fdbd253ae22f5b3286247c6940321f7d2b9801aea21ef5269f (8ms) I0320 09:56:53.736988 1 gather_workloads_info.go:387] No image sha256:1aabd00f813e1afccdab16a845e6346288f5cf7e320fa5bb97c683702191c575 (8ms) I0320 09:56:53.745218 1 gather_workloads_info.go:387] No image sha256:07f14370b57046dc2ac49be933bb2248aaf5943e8fb56ce9703eaa7c673b999f (8ms) I0320 09:56:53.753309 1 gather_workloads_info.go:387] No image sha256:2a4f6a54cd5135c8121832236063419071fa5d00c10a051c197d3d01370e4d9c (8ms) I0320 09:56:53.761632 1 gather_workloads_info.go:387] No image sha256:0b2cc6aadb1abc1d35a55da0c9884eb9426d7ff7a97174daeb09bb1e1a23cfae (8ms) I0320 09:56:53.769647 1 gather_workloads_info.go:387] No image sha256:bbb5af28bbed69e8189825fdc6665f4e8dbcd0c0c83fa479943ea102fe9452d6 (8ms) I0320 09:56:53.777688 1 gather_workloads_info.go:387] No image sha256:db40a3d4e50016c645fdb4a1f389dfcdbc9febc10c174816d681abe5c14153cd (8ms) I0320 09:56:53.785575 1 gather_workloads_info.go:387] No image sha256:dd6a017092d57ab3a6663f3822a0fa0e3bce4365859d9522dd173405276cd49a (8ms) I0320 09:56:53.810200 1 gather_workloads_info.go:387] No image sha256:60648d27017709a420e43353f9a83d7c466017b3b771200672a4dbc3c3f7a3ae (25ms) I0320 09:56:53.911484 1 gather_workloads_info.go:387] No image sha256:1c325fa866a794c56a7b30ca9d4fe225606566651bb12a2ecce8f60cbddb2809 (101ms) I0320 09:56:54.011713 1 gather_workloads_info.go:387] No image sha256:25d9148ea34149935323231c64952679727b6f7f235febc3be1df63e0bf18c5e (100ms) I0320 09:56:54.110853 1 gather_workloads_info.go:387] No image sha256:7ea1ff613255c1ee67b34ec82b68ec6f5a4a6d283620a8e0ad142d4a31eb9211 (99ms) I0320 09:56:54.211653 1 gather_workloads_info.go:387] No image sha256:5efa227ddac82e56eddc03761e220af1f2d6a4b22ee00b2baf426523f26c0319 (101ms) I0320 09:56:54.311917 1 gather_workloads_info.go:387] No image sha256:11a864539ae77a94afcf8d2e1abf9168081c5081a1bc06088f4be5206cdc02df (100ms) I0320 09:56:54.411585 1 gather_workloads_info.go:387] No image sha256:7dd9584d291967387f2a1ade991ad3bf8565bb59f49b013217951399e8273498 (100ms) I0320 09:56:54.510946 1 gather_workloads_info.go:387] No image sha256:188e4e2d697e2ca3c7c16772c0ab12253f00bd8433d8691ed7058fe7bafbbbf4 (99ms) I0320 09:56:54.612075 1 gather_workloads_info.go:387] No image sha256:124f1b391b0f376a5deffd05b29ef0896ca0912e1f67b3a2051ed6678c54a57e (101ms) I0320 09:56:54.711473 1 gather_workloads_info.go:387] No image sha256:862d5befeba504d5058264cf65f16fc307eea4eb3914fbbbf40fbe61714b0958 (99ms) I0320 09:56:54.811514 1 gather_workloads_info.go:387] No image sha256:ab32e70fb149ae7470c941ca73bcfe9eba540bd5058a3d697a3ddf96597d0132 (100ms) I0320 09:56:54.910970 1 gather_workloads_info.go:387] No image sha256:c85dd52f4140690eef0d762834ab8b9af2faedbbd3bd9dd5616d4cd0bebd4740 (99ms) I0320 09:56:55.011968 1 gather_workloads_info.go:387] No image sha256:04c7a2673571e9069c934d86b4308c0f57b2715b45db500837b7708ae1f20ae7 (101ms) I0320 09:56:55.112059 1 gather_workloads_info.go:387] No image sha256:17e68da37c74bdc95a4eb43b51ce671027807ee610a4ee5237324a8c0dc9d044 (100ms) I0320 09:56:55.212057 1 gather_workloads_info.go:387] No image sha256:26e4f5b039a340bc1ae080fb9d92799700b57ae5df45120768f516682d6cc54b (100ms) I0320 09:56:55.311752 1 gather_workloads_info.go:387] No image sha256:c15ca0c0ad60fe8757c2d5d1723fcdd7a1ed6c0251a90d22a7e6cae6811d01aa (100ms) I0320 09:56:55.411590 1 gather_workloads_info.go:387] No image sha256:41e0dcd2f91df9234249f168b3fbe6d55cfdb3ebd2e598d5edfa47f905c046cb (100ms) I0320 09:56:55.449103 1 configmapobserver.go:84] configmaps "insights-config" not found I0320 09:56:55.511484 1 gather_workloads_info.go:387] No image sha256:04ee35aa6f87e954d386228808017cf2938c5a96a052059c6767078d5f3ad8d6 (100ms) I0320 09:56:55.612137 1 gather_workloads_info.go:387] No image sha256:50b7c5185c4a6e63145753415c92723c59136d14b951ea0f4f06739b00893329 (101ms) I0320 09:56:55.649645 1 configmapobserver.go:84] configmaps "insights-config" not found I0320 09:56:55.710997 1 gather_workloads_info.go:387] No image sha256:77d8ddf2957025a39616328b5cf2b012de6cedce63fb79eebc5270d7b3401574 (99ms) I0320 09:56:55.811800 1 gather_workloads_info.go:387] No image sha256:05376e7e2304d22dd866d8e25dd1e5ff8988b2232059fadfd12cb440af785635 (101ms) I0320 09:56:55.911392 1 gather_workloads_info.go:387] No image sha256:79fca57a4fd5ef00472a287218b84061b941905c61fcc55e89073dd93fa77648 (100ms) I0320 09:56:56.011711 1 gather_workloads_info.go:387] No image sha256:3e262d5207d5e2077e4a14f4fbddc70d4d2651dbaaf955147013feb58a18cddd (100ms) I0320 09:56:56.011743 1 tasks_processing.go:74] worker 1 stopped. E0320 09:56:56.011755 1 gather.go:140] gatherer "workloads" function "workload_info" failed with the error: no running pods found for the insights-runtime-extractor statefulset I0320 09:56:56.011990 1 recorder.go:75] Recording config/workload_info with fingerprint=ee2a83f50c93748654794553ce58459a1e69d021eb59e9eb54a129f64b293c30 I0320 09:56:56.012006 1 gather.go:177] gatherer "workloads" function "workload_info" took 2.334728289s to process 1 records E0320 09:56:56.012029 1 periodic.go:254] "Unhandled Error" err="workloads failed after 2.335s with: function \"workload_info\" failed with an error" I0320 09:56:56.013132 1 controllerstatus.go:89] name=periodic-workloads healthy=false reason=PeriodicGatherFailed message=Source workloads could not be retrieved: function "workload_info" failed with an error I0320 09:56:56.013146 1 periodic.go:216] Running conditional gatherer I0320 09:56:56.018639 1 requests.go:294] Making HTTP GET request at: https://console.redhat.com/api/gathering/v2/4.21.5/gathering_rules I0320 09:56:56.025771 1 conditional_gatherer.go:107] Get "https://console.redhat.com/api/gathering/v2/4.21.5/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.130.0.9:39388->172.30.0.10:53: read: connection refused E0320 09:56:56.026096 1 conditional_gatherer.go:322] unable to update alerts cache: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory I0320 09:56:56.026151 1 conditional_gatherer.go:384] updating version cache for conditional gatherer I0320 09:56:56.031839 1 conditional_gatherer.go:392] cluster version is '4.21.5' E0320 09:56:56.031851 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0320 09:56:56.031856 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0320 09:56:56.031859 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0320 09:56:56.031863 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0320 09:56:56.031866 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0320 09:56:56.031868 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0320 09:56:56.031871 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0320 09:56:56.031873 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0320 09:56:56.031876 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing I0320 09:56:56.031891 1 tasks_processing.go:45] number of workers: 3 I0320 09:56:56.031907 1 tasks_processing.go:69] worker 2 listening for tasks. I0320 09:56:56.031914 1 tasks_processing.go:71] worker 2 working on conditional_gatherer_rules task. I0320 09:56:56.031926 1 tasks_processing.go:69] worker 0 listening for tasks. I0320 09:56:56.031935 1 tasks_processing.go:69] worker 1 listening for tasks. I0320 09:56:56.031937 1 tasks_processing.go:71] worker 0 working on remote_configuration task. I0320 09:56:56.031942 1 tasks_processing.go:74] worker 1 stopped. I0320 09:56:56.031952 1 tasks_processing.go:71] worker 2 working on rapid_container_logs task. I0320 09:56:56.032018 1 recorder.go:75] Recording insights-operator/conditional-gatherer-rules with fingerprint=7034af97e7e41c22e4b775abdd4b9066c8ebb19da33eb7f69f39bfd2eb5f6406 I0320 09:56:56.032035 1 gather.go:177] gatherer "conditional" function "conditional_gatherer_rules" took 1.22µs to process 1 records I0320 09:56:56.032083 1 recorder.go:75] Recording insights-operator/remote-configuration with fingerprint=0394430c431eec4d48bb1811a90918e95161d2282c59af26f2473613cc0959db I0320 09:56:56.032094 1 gather.go:177] gatherer "conditional" function "remote_configuration" took 1.252µs to process 1 records I0320 09:56:56.032102 1 tasks_processing.go:74] worker 0 stopped. I0320 09:56:56.032271 1 tasks_processing.go:74] worker 2 stopped. I0320 09:56:56.032298 1 gather.go:177] gatherer "conditional" function "rapid_container_logs" took 301.577µs to process 0 records I0320 09:56:56.032352 1 controllerstatus.go:89] name=periodic-conditional healthy=false reason=NotAvailable message=Get "https://console.redhat.com/api/gathering/v2/4.21.5/gathering_rules": dial tcp: lookup console.redhat.com on 172.30.0.10:53: read udp 10.130.0.9:39388->172.30.0.10:53: read: connection refused I0320 09:56:56.032381 1 recorder.go:75] Recording insights-operator/remote-configuration.json with fingerprint=359de9c990c741675cec72fda96b5c3682221efdb4799f5eaa6e9805bcd3b5c1 I0320 09:56:56.060194 1 recorder.go:75] Recording insights-operator/gathers with fingerprint=81c62740f8f850296fa2c0aa5fb5008b2c820c91e7b24ec32112d3bb5fe0b43c I0320 09:56:56.060354 1 diskrecorder.go:70] Writing 106 records to /var/lib/insights-operator/insights-2026-03-20-095656.tar.gz I0320 09:56:56.066734 1 diskrecorder.go:51] Wrote 106 records to disk in 6ms I0320 09:56:56.066766 1 periodic.go:285] Gathering cluster info every 2h0m0s I0320 09:56:56.066788 1 periodic.go:286] Configuration is dataReporting: interval: 2h0m0s, uploadEndpoint: https://console.redhat.com/api/ingress/v1/upload, storagePath: /var/lib/insights-operator, downloadEndpoint: https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports, conditionalGathererEndpoint: https://console.redhat.com/api/gathering/v2/%s/gathering_rules, obfuscation: [] sca: disabled: false, endpoint: https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates, interval: 8h0m0s alerting: disabled: false clusterTransfer: endpoint: https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/, interval: 12h0m0s proxy: httpProxy: , httpsProxy: , noProxy: I0320 09:57:04.556573 1 configmapobserver.go:84] configmaps "insights-config" not found I0320 09:58:08.419357 1 observer_polling.go:111] Observed file "/var/run/secrets/serving-cert/tls.key" has been created (hash="f76871644c755581972314e83fb1eaad397e9747b6522e3f2454085ff1968f16") W0320 09:58:08.419388 1 builder.go:160] Restart triggered because of file /var/run/secrets/serving-cert/tls.key was created I0320 09:58:08.419460 1 genericapiserver.go:548] "[graceful-termination] shutdown event" name="ShutdownInitiated" I0320 09:58:08.419478 1 observer_polling.go:111] Observed file "/var/run/configmaps/service-ca-bundle/service-ca.crt" has been created (hash="4f23e4fdf96246337e8d3dce8c1ca8774e0e7135701604a4b70c190ca2489db0")