I0504 20:40:55.861528 1 cmd.go:253] Using service-serving-cert provided certificates I0504 20:40:55.862060 1 observer_polling.go:159] Starting file observer I0504 20:40:56.404750 1 operator.go:76] Starting insights-operator v0.0.0-master+$Format:%H$ I0504 20:40:56.404978 1 legacy_config.go:327] Current config: {"report":false,"storagePath":"/var/lib/insights-operator","interval":"2h","endpoint":"https://console.redhat.com/api/ingress/v1/upload","conditionalGathererEndpoint":"https://console.redhat.com/api/gathering/v2/%s/gathering_rules","pull_report":{"endpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports","delay":"60s","timeout":"3000s","min_retry":"30s"},"impersonate":"system:serviceaccount:openshift-insights:gather","enableGlobalObfuscation":false,"ocm":{"scaEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates","scaInterval":"8h","scaDisabled":false,"clusterTransferEndpoint":"https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/","clusterTransferInterval":"12h"},"disableInsightsAlerts":false,"processingStatusEndpoint":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status","reportEndpointTechPreview":"https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/report"} I0504 20:40:56.405748 1 simple_featuregate_reader.go:171] Starting feature-gate-detector I0504 20:40:56.405837 1 secure_serving.go:57] Forcing use of http/1.1 only W0504 20:40:56.405877 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. W0504 20:40:56.405885 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. W0504 20:40:56.405893 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. W0504 20:40:56.405899 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. W0504 20:40:56.405905 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. W0504 20:40:56.405910 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. I0504 20:40:56.411267 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-insights", Name:"insights-operator", UID:"ba5c066e-2163-4573-9d86-b503974dbc3f", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BuildCSIVolumes", "CPMSMachineNamePrefix", "ConsolePluginContentSecurityPolicy", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNSInstall", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "HyperShiftOnlyDynamicResourceAllocation", "ImageStreamImportMode", "ImageVolume", "KMSv1", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "ManagedBootImagesAzure", "ManagedBootImagesvSphere", "MetricsCollectionProfiles", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "PinnedImages", "PreconfiguredUDNAddresses", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SigstoreImageVerification", "SigstoreImageVerificationPKI", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiDisk", "VSphereMultiNetworks", "VolumeAttributesClass"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSDualStackInstall", "AWSServiceLBNetworkSecurityGroup", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureDualStackInstall", "AzureMultiDisk", "BootImageSkewEnforcement", "BootcNodeManagement", "CBORServingAndStorage", "CRDCompatibilityRequirementOperator", "ClientsAllowCBOR", "ClientsPreferCBOR", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterAPIMachineManagement", "ClusterAPIMachineManagementVSphere", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "EtcdBackendQuota", "EventTTL", "EventedPLEG", "Example", "Example2", "ExternalSnapshotMetadata", "GCPClusterHostedDNS", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "GCPDualStackInstall", "ImageModeStatusReporting", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryption", "KMSEncryptionProvider", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "ManagedBootImagesCPMS", "MaxUnavailableStatefulSet", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "MultiDiskSetup", "MutableCSINodeAllocatableCount", "MutatingAdmissionPolicy", "NewOLM", "NewOLMBoxCutterRuntime", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterInstall", "NutanixMultiSubnets", "OSStreams", "OVNObservability", "OnPremDNSRecords", "OpenShiftPodSecurityAdmission", "ProvisioningRequestAvailable", "SELinuxMount", "ShortCertRotation", "SignatureStores", "TranslateStreamCloseWebsocketRequests", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VolumeGroupSnapshot"}} I0504 20:40:56.411260 1 operator.go:141] FeatureGates initialized: knownFeatureGates=[AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BuildCSIVolumes CPMSMachineNamePrefix ConsolePluginContentSecurityPolicy ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GCPClusterHostedDNSInstall GatewayAPI GatewayAPIController HighlyAvailableArbiter HyperShiftOnlyDynamicResourceAllocation ImageStreamImportMode ImageVolume KMSv1 MachineConfigNodes ManagedBootImages ManagedBootImagesAWS ManagedBootImagesAzure ManagedBootImagesvSphere MetricsCollectionProfiles NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation PinnedImages PreconfiguredUDNAddresses ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SigstoreImageVerification SigstoreImageVerificationPKI StoragePerformantSecurityPolicy UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiDisk VSphereMultiNetworks VolumeAttributesClass AWSClusterHostedDNS AWSClusterHostedDNSInstall AWSDedicatedHosts AWSDualStackInstall AWSServiceLBNetworkSecurityGroup AutomatedEtcdBackup AzureClusterHostedDNSInstall AzureDedicatedHosts AzureDualStackInstall AzureMultiDisk BootImageSkewEnforcement BootcNodeManagement CBORServingAndStorage CRDCompatibilityRequirementOperator ClientsAllowCBOR ClientsPreferCBOR ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterAPIMachineManagement ClusterAPIMachineManagementVSphere ClusterMonitoringConfig ClusterVersionOperatorConfiguration DNSNameResolver DualReplica DyanmicServiceEndpointIBMCloud EtcdBackendQuota EventTTL EventedPLEG Example Example2 ExternalSnapshotMetadata GCPClusterHostedDNS GCPCustomAPIEndpoints GCPCustomAPIEndpointsInstall GCPDualStackInstall ImageModeStatusReporting IngressControllerDynamicConfigurationManager InsightsConfig InsightsOnDemandDataGather IrreconcilableMachineConfig KMSEncryption KMSEncryptionProvider MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController ManagedBootImagesCPMS MaxUnavailableStatefulSet MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure MultiDiskSetup MutableCSINodeAllocatableCount MutatingAdmissionPolicy NewOLM NewOLMBoxCutterRuntime NewOLMCatalogdAPIV1Metas NewOLMOwnSingleNamespace NewOLMPreflightPermissionChecks NewOLMWebhookProviderOpenshiftServiceCA NoRegistryClusterInstall NutanixMultiSubnets OSStreams OVNObservability OnPremDNSRecords OpenShiftPodSecurityAdmission ProvisioningRequestAvailable SELinuxMount ShortCertRotation SignatureStores TranslateStreamCloseWebsocketRequests VSphereConfigurableMaxAllowedBlockVolumesPerNode VSphereHostVMGroupZonal VSphereMixedNodeEnv VolumeGroupSnapshot] I0504 20:40:56.413480 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController I0504 20:40:56.413507 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController I0504 20:40:56.413515 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0504 20:40:56.413537 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0504 20:40:56.413541 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0504 20:40:56.413557 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0504 20:40:56.413870 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" I0504 20:40:56.414212 1 secure_serving.go:213] Serving securely on [::]:8443 I0504 20:40:56.414242 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" W0504 20:40:56.419444 1 configmapobserver.go:64] Cannot get the configuration config map: configmaps "insights-config" not found. Default configuration is used. I0504 20:40:56.419476 1 secretconfigobserver.go:216] Legacy configuration set: enabled=false endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=false reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0504 20:40:56.419591 1 base_controller.go:76] Waiting for caches to sync for ConfigController I0504 20:40:56.426239 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 20:40:56.426259 1 secretconfigobserver.go:204] Legacy configuration updated: enabled=true endpoint=https://console.redhat.com/api/ingress/v1/upload conditional_gatherer_endpoint=https://console.redhat.com/api/gathering/v2/%s/gathering_rules interval=2h0m0s token=true reportEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports initialPollingDelay=1m0s minRetryTime=30s pollingTimeout=50m0s processingStatusEndpoint=https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/request/%s/status I0504 20:40:56.432786 1 secretconfigobserver.go:119] support secret does not exist I0504 20:40:56.439102 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 20:40:56.447098 1 secretconfigobserver.go:119] support secret does not exist I0504 20:40:56.449455 1 recorder.go:176] Pruning old reports every 6h55m52s, max age is 288h0m0s I0504 20:40:56.456309 1 periodic.go:216] Running clusterconfig gatherer I0504 20:40:56.456314 1 controllerstatus.go:80] name=insightsuploader healthy=true reason= message= I0504 20:40:56.456336 1 insightsuploader.go:86] Reporting status periodically to https://console.redhat.com/api/ingress/v1/upload every 2h0m0s, starting in 1m30s I0504 20:40:56.456362 1 tasks_processing.go:45] number of workers: 64 I0504 20:40:56.456393 1 controllerstatus.go:80] name=insightsreport healthy=true reason= message= I0504 20:40:56.456406 1 tasks_processing.go:69] worker 0 listening for tasks. I0504 20:40:56.456398 1 tasks_processing.go:69] worker 2 listening for tasks. I0504 20:40:56.456410 1 insightsreport.go:296] Starting report retriever I0504 20:40:56.456422 1 tasks_processing.go:69] worker 1 listening for tasks. I0504 20:40:56.456425 1 insightsreport.go:298] Insights analysis reports will be downloaded from the https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports endpoint with a delay of 1m0s I0504 20:40:56.456430 1 tasks_processing.go:69] worker 13 listening for tasks. I0504 20:40:56.456433 1 tasks_processing.go:69] worker 7 listening for tasks. I0504 20:40:56.456442 1 tasks_processing.go:69] worker 18 listening for tasks. I0504 20:40:56.456440 1 tasks_processing.go:69] worker 3 listening for tasks. I0504 20:40:56.456449 1 tasks_processing.go:69] worker 4 listening for tasks. I0504 20:40:56.456449 1 tasks_processing.go:69] worker 8 listening for tasks. I0504 20:40:56.456456 1 tasks_processing.go:69] worker 9 listening for tasks. I0504 20:40:56.456458 1 tasks_processing.go:69] worker 14 listening for tasks. I0504 20:40:56.456464 1 tasks_processing.go:69] worker 5 listening for tasks. I0504 20:40:56.456467 1 tasks_processing.go:69] worker 10 listening for tasks. I0504 20:40:56.456471 1 tasks_processing.go:69] worker 17 listening for tasks. I0504 20:40:56.456475 1 tasks_processing.go:69] worker 6 listening for tasks. I0504 20:40:56.456472 1 tasks_processing.go:69] worker 21 listening for tasks. I0504 20:40:56.456481 1 tasks_processing.go:69] worker 11 listening for tasks. I0504 20:40:56.456482 1 tasks_processing.go:69] worker 16 listening for tasks. I0504 20:40:56.456489 1 tasks_processing.go:69] worker 12 listening for tasks. I0504 20:40:56.456493 1 tasks_processing.go:69] worker 20 listening for tasks. I0504 20:40:56.456482 1 tasks_processing.go:69] worker 15 listening for tasks. I0504 20:40:56.456502 1 tasks_processing.go:69] worker 22 listening for tasks. I0504 20:40:56.456503 1 tasks_processing.go:69] worker 24 listening for tasks. I0504 20:40:56.456484 1 tasks_processing.go:69] worker 19 listening for tasks. I0504 20:40:56.456500 1 tasks_processing.go:69] worker 23 listening for tasks. I0504 20:40:56.456575 1 tasks_processing.go:71] worker 23 working on crds task. I0504 20:40:56.456584 1 tasks_processing.go:71] worker 7 working on container_images task. I0504 20:40:56.456594 1 tasks_processing.go:69] worker 26 listening for tasks. I0504 20:40:56.456604 1 tasks_processing.go:71] worker 26 working on sap_datahubs task. I0504 20:40:56.456611 1 tasks_processing.go:69] worker 51 listening for tasks. I0504 20:40:56.456622 1 tasks_processing.go:69] worker 52 listening for tasks. I0504 20:40:56.456616 1 tasks_processing.go:71] worker 12 working on jaegers task. I0504 20:40:56.456627 1 tasks_processing.go:71] worker 22 working on qemu_kubevirt_launcher_logs task. I0504 20:40:56.456635 1 tasks_processing.go:71] worker 8 working on storage_cluster task. I0504 20:40:56.456640 1 tasks_processing.go:69] worker 54 listening for tasks. I0504 20:40:56.456649 1 tasks_processing.go:69] worker 55 listening for tasks. I0504 20:40:56.456584 1 tasks_processing.go:69] worker 48 listening for tasks. I0504 20:40:56.456576 1 tasks_processing.go:69] worker 47 listening for tasks. I0504 20:40:56.456578 1 tasks_processing.go:71] worker 17 working on mutating_webhook_configurations task. I0504 20:40:56.456662 1 tasks_processing.go:69] worker 62 listening for tasks. I0504 20:40:56.456661 1 tasks_processing.go:71] worker 24 working on openstack_dataplanenodesets task. I0504 20:40:56.456671 1 tasks_processing.go:69] worker 63 listening for tasks. I0504 20:40:56.456677 1 tasks_processing.go:71] worker 0 working on service_accounts task. I0504 20:40:56.456683 1 tasks_processing.go:69] worker 61 listening for tasks. I0504 20:40:56.456687 1 tasks_processing.go:71] worker 19 working on machine_configs task. I0504 20:40:56.456703 1 tasks_processing.go:69] worker 59 listening for tasks. I0504 20:40:56.456632 1 tasks_processing.go:69] worker 53 listening for tasks. I0504 20:40:56.456712 1 tasks_processing.go:71] worker 20 working on openstack_version task. I0504 20:40:56.456603 1 tasks_processing.go:69] worker 50 listening for tasks. I0504 20:40:56.456705 1 tasks_processing.go:71] worker 6 working on machine_config_pools task. I0504 20:40:56.457049 1 tasks_processing.go:71] worker 15 working on nodenetworkconfigurationpolicies task. I0504 20:40:56.457087 1 tasks_processing.go:71] worker 16 working on schedulers task. I0504 20:40:56.457181 1 tasks_processing.go:71] worker 5 working on clusterroles task. I0504 20:40:56.456577 1 tasks_processing.go:71] worker 13 working on feature_gates task. I0504 20:40:56.457284 1 tasks_processing.go:71] worker 9 working on install_plans task. I0504 20:40:56.457347 1 tasks_processing.go:69] worker 60 listening for tasks. I0504 20:40:56.457376 1 tasks_processing.go:71] worker 14 working on infrastructures task. I0504 20:40:56.457557 1 tasks_processing.go:71] worker 3 working on openstack_controlplanes task. I0504 20:40:56.457634 1 tasks_processing.go:69] worker 27 listening for tasks. I0504 20:40:56.457687 1 tasks_processing.go:69] worker 28 listening for tasks. I0504 20:40:56.457702 1 tasks_processing.go:69] worker 29 listening for tasks. I0504 20:40:56.457712 1 tasks_processing.go:69] worker 30 listening for tasks. I0504 20:40:56.456585 1 tasks_processing.go:69] worker 25 listening for tasks. I0504 20:40:56.457837 1 tasks_processing.go:69] worker 32 listening for tasks. I0504 20:40:56.456714 1 tasks_processing.go:71] worker 11 working on metrics task. I0504 20:40:56.457802 1 tasks_processing.go:69] worker 31 listening for tasks. I0504 20:40:56.457882 1 tasks_processing.go:71] worker 10 working on authentication task. I0504 20:40:56.456709 1 tasks_processing.go:71] worker 21 working on machines task. I0504 20:40:56.457990 1 tasks_processing.go:69] worker 35 listening for tasks. I0504 20:40:56.457999 1 tasks_processing.go:69] worker 36 listening for tasks. I0504 20:40:56.458030 1 tasks_processing.go:69] worker 58 listening for tasks. I0504 20:40:56.458036 1 tasks_processing.go:69] worker 37 listening for tasks. I0504 20:40:56.458087 1 tasks_processing.go:69] worker 38 listening for tasks. I0504 20:40:56.458121 1 tasks_processing.go:69] worker 45 listening for tasks. I0504 20:40:56.456656 1 tasks_processing.go:69] worker 56 listening for tasks. I0504 20:40:56.458089 1 tasks_processing.go:71] worker 4 working on storage_classes task. I0504 20:40:56.458221 1 tasks_processing.go:69] worker 39 listening for tasks. I0504 20:40:56.458229 1 tasks_processing.go:71] worker 39 working on cost_management_metrics_configs task. I0504 20:40:56.458243 1 tasks_processing.go:71] worker 1 working on machine_sets task. I0504 20:40:56.459174 1 tasks_processing.go:69] worker 43 listening for tasks. I0504 20:40:56.456594 1 tasks_processing.go:69] worker 49 listening for tasks. I0504 20:40:56.458180 1 tasks_processing.go:69] worker 46 listening for tasks. I0504 20:40:56.458193 1 tasks_processing.go:71] worker 38 working on node_features task. I0504 20:40:56.458203 1 tasks_processing.go:71] worker 45 working on ingress task. I0504 20:40:56.458211 1 tasks_processing.go:71] worker 56 working on ceph_cluster task. I0504 20:40:56.458268 1 tasks_processing.go:71] worker 53 working on overlapping_namespace_uids task. I0504 20:40:56.459273 1 tasks_processing.go:71] worker 46 working on support_secret task. I0504 20:40:56.459605 1 tasks_processing.go:71] worker 30 working on nodes task. I0504 20:40:56.459677 1 tasks_processing.go:71] worker 60 working on nodenetworkstates task. I0504 20:40:56.459726 1 tasks_processing.go:71] worker 35 working on image task. I0504 20:40:56.459779 1 tasks_processing.go:71] worker 31 working on proxies task. I0504 20:40:56.459182 1 tasks_processing.go:69] worker 44 listening for tasks. I0504 20:40:56.459190 1 tasks_processing.go:71] worker 51 working on openstack_dataplanedeployments task. I0504 20:40:56.459786 1 tasks_processing.go:71] worker 28 working on sap_pods task. I0504 20:40:56.459191 1 tasks_processing.go:69] worker 57 listening for tasks. I0504 20:40:56.457900 1 tasks_processing.go:69] worker 34 listening for tasks. I0504 20:40:56.459804 1 tasks_processing.go:71] worker 29 working on certificate_signing_requests task. I0504 20:40:56.460071 1 tasks_processing.go:71] worker 36 working on lokistack task. I0504 20:40:56.460080 1 tasks_processing.go:71] worker 58 working on ingress_certificates task. I0504 20:40:56.459199 1 tasks_processing.go:71] worker 32 working on oauths task. I0504 20:40:56.460181 1 tasks_processing.go:71] worker 27 working on openshift_logging task. I0504 20:40:56.457892 1 tasks_processing.go:69] worker 33 listening for tasks. I0504 20:40:56.458109 1 tasks_processing.go:71] worker 2 working on image_pruners task. I0504 20:40:56.460183 1 tasks_processing.go:71] worker 54 working on operators_pods_and_events task. I0504 20:40:56.460352 1 tasks_processing.go:71] worker 33 working on machine_healthchecks task. I0504 20:40:56.460529 1 tasks_processing.go:71] worker 44 working on active_alerts task. I0504 20:40:56.460691 1 tasks_processing.go:71] worker 43 working on openshift_machine_api_events task. I0504 20:40:56.460751 1 tasks_processing.go:71] worker 34 working on config_maps task. I0504 20:40:56.461475 1 tasks_processing.go:71] worker 37 working on sap_config task. I0504 20:40:56.461877 1 tasks_processing.go:71] worker 52 working on pdbs task. I0504 20:40:56.462105 1 tasks_processing.go:71] worker 18 working on networks task. I0504 20:40:56.462190 1 tasks_processing.go:71] worker 55 working on monitoring_persistent_volumes task. I0504 20:40:56.462363 1 tasks_processing.go:71] worker 48 working on version task. I0504 20:40:56.462409 1 tasks_processing.go:71] worker 62 working on olm_operators task. I0504 20:40:56.462403 1 tasks_processing.go:71] worker 47 working on image_registries task. I0504 20:40:56.462592 1 tasks_processing.go:71] worker 63 working on machine_autoscalers task. I0504 20:40:56.462773 1 tasks_processing.go:71] worker 61 working on container_runtime_configs task. I0504 20:40:56.462791 1 tasks_processing.go:71] worker 59 working on number_of_pods_and_netnamespaces_with_sdn_annotations task. I0504 20:40:56.463049 1 tasks_processing.go:71] worker 50 working on tsdb_status task. I0504 20:40:56.463282 1 tasks_processing.go:71] worker 25 working on aggregated_monitoring_cr_names task. I0504 20:40:56.462797 1 tasks_processing.go:69] worker 42 listening for tasks. I0504 20:40:56.463631 1 tasks_processing.go:71] worker 42 working on node_logs task. I0504 20:40:56.462806 1 tasks_processing.go:69] worker 40 listening for tasks. I0504 20:40:56.462813 1 tasks_processing.go:69] worker 41 listening for tasks. I0504 20:40:56.464112 1 tasks_processing.go:71] worker 57 working on pod_network_connectivity_checks task. I0504 20:40:56.464195 1 tasks_processing.go:71] worker 41 working on validating_webhook_configurations task. I0504 20:40:56.464657 1 tasks_processing.go:71] worker 40 working on dvo_metrics task. I0504 20:40:56.464682 1 tasks_processing.go:71] worker 49 working on silenced_alerts task. I0504 20:40:56.465556 1 tasks_processing.go:71] worker 26 working on operators task. I0504 20:40:56.465841 1 gather.go:177] gatherer "clusterconfig" function "sap_datahubs" took 8.942773ms to process 0 records I0504 20:40:56.465939 1 tasks_processing.go:71] worker 6 working on cluster_apiserver task. I0504 20:40:56.465951 1 gather.go:177] gatherer "clusterconfig" function "machine_config_pools" took 8.970874ms to process 0 records I0504 20:40:56.466223 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanenodesets" took 9.269956ms to process 0 records I0504 20:40:56.466262 1 gather.go:177] gatherer "clusterconfig" function "openstack_controlplanes" took 8.371449ms to process 0 records I0504 20:40:56.466292 1 tasks_processing.go:74] worker 3 stopped. I0504 20:40:56.466296 1 tasks_processing.go:74] worker 24 stopped. I0504 20:40:56.466387 1 controller.go:129] Initializing last reported time to 0001-01-01T00:00:00Z I0504 20:40:56.466408 1 controller.go:368] The initial operator extension status is healthy I0504 20:40:56.466418 1 controller.go:254] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0504 20:40:56.466425 1 controller.go:254] Source periodic-conditional *controllerstatus.Simple is not ready I0504 20:40:56.466430 1 controller.go:254] Source periodic-workloads *controllerstatus.Simple is not ready I0504 20:40:56.466454 1 controller.go:531] The operator is still being initialized I0504 20:40:56.466463 1 controller.go:554] The operator is healthy I0504 20:40:56.466575 1 sca.go:136] Pulling SCA certificates from https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates. Next check is in 8h0m0s I0504 20:40:56.466620 1 cluster_transfer.go:83] checking the availability of cluster transfer. Next check is in 12h0m0s W0504 20:40:56.466763 1 operator.go:328] started I0504 20:40:56.466792 1 base_controller.go:76] Waiting for caches to sync for LoggingSyncer I0504 20:40:56.467956 1 tasks_processing.go:74] worker 10 stopped. I0504 20:40:56.468261 1 recorder.go:75] Recording config/authentication with fingerprint=938ddf5d78cf9c635599c9c2aca98fde29001ab60eeaf4e305bf8f69f2283604 I0504 20:40:56.468280 1 gather.go:177] gatherer "clusterconfig" function "authentication" took 10.035982ms to process 1 records E0504 20:40:56.470540 1 gather_prometheus_tsdb_status.go:49] Unable to tsdb status: Get "https://prometheus-k8s.openshift-monitoring.svc:9091/api/v1/status/tsdb": dial tcp 172.30.161.151:9091: connect: connection refused I0504 20:40:56.470569 1 tasks_processing.go:74] worker 50 stopped. E0504 20:40:56.470562 1 gather_most_recent_metrics.go:87] Unable to retrieve most recent metrics: Get "https://prometheus-k8s.openshift-monitoring.svc:9091/federate?match%5B%5D=cluster_installer&match%5B%5D=namespace%3Acontainer_cpu_usage%3Asum&match%5B%5D=namespace%3Acontainer_memory_usage_bytes%3Asum&match%5B%5D=vsphere_node_hw_version_total&match%5B%5D=virt_platform&match%5B%5D=console_helm_installs_total&match%5B%5D=console_helm_upgrades_total&match%5B%5D=console_helm_uninstalls_total&match%5B%5D=openshift_apps_deploymentconfigs_strategy_total&match%5B%5D=etcd_server_slow_apply_total&match%5B%5D=etcd_server_slow_read_indexes_total&match%5B%5D=haproxy_exporter_server_threshold": dial tcp 172.30.161.151:9091: connect: connection refused E0504 20:40:56.470579 1 gather.go:140] gatherer "clusterconfig" function "tsdb_status" failed with the error: Get "https://prometheus-k8s.openshift-monitoring.svc:9091/api/v1/status/tsdb": dial tcp 172.30.161.151:9091: connect: connection refused I0504 20:40:56.470669 1 gather.go:177] gatherer "clusterconfig" function "tsdb_status" took 7.501584ms to process 0 records E0504 20:40:56.470691 1 gather.go:140] gatherer "clusterconfig" function "metrics" failed with the error: Get "https://prometheus-k8s.openshift-monitoring.svc:9091/federate?match%5B%5D=cluster_installer&match%5B%5D=namespace%3Acontainer_cpu_usage%3Asum&match%5B%5D=namespace%3Acontainer_memory_usage_bytes%3Asum&match%5B%5D=vsphere_node_hw_version_total&match%5B%5D=virt_platform&match%5B%5D=console_helm_installs_total&match%5B%5D=console_helm_upgrades_total&match%5B%5D=console_helm_uninstalls_total&match%5B%5D=openshift_apps_deploymentconfigs_strategy_total&match%5B%5D=etcd_server_slow_apply_total&match%5B%5D=etcd_server_slow_read_indexes_total&match%5B%5D=haproxy_exporter_server_threshold": dial tcp 172.30.161.151:9091: connect: connection refused I0504 20:40:56.470713 1 gather.go:177] gatherer "clusterconfig" function "metrics" took 12.372914ms to process 0 records I0504 20:40:56.470715 1 tasks_processing.go:74] worker 11 stopped. I0504 20:40:56.472970 1 gather_logs.go:145] no pods in namespace were found I0504 20:40:56.472988 1 tasks_processing.go:74] worker 22 stopped. I0504 20:40:56.473000 1 gather.go:177] gatherer "clusterconfig" function "qemu_kubevirt_launcher_logs" took 16.347561ms to process 0 records I0504 20:40:56.473325 1 tasks_processing.go:74] worker 16 stopped. I0504 20:40:56.473480 1 recorder.go:75] Recording config/schedulers/cluster with fingerprint=6e1737378b2c70d4702ec5a93dba2d2818ade2f9619a3763578532e7a10e4106 I0504 20:40:56.473525 1 gather.go:177] gatherer "clusterconfig" function "schedulers" took 16.206763ms to process 1 records I0504 20:40:56.473568 1 gather.go:177] gatherer "clusterconfig" function "cost_management_metrics_configs" took 15.250938ms to process 0 records I0504 20:40:56.473644 1 tasks_processing.go:74] worker 39 stopped. I0504 20:40:56.473727 1 tasks_processing.go:74] worker 13 stopped. I0504 20:40:56.473728 1 recorder.go:75] Recording config/featuregate with fingerprint=758e72ee72aa2e056bcabbccfd63b35f901629af178ff5bc9cf5ed46b1d68dcc I0504 20:40:56.473780 1 gather.go:177] gatherer "clusterconfig" function "feature_gates" took 16.220068ms to process 1 records E0504 20:40:56.473801 1 gather.go:140] gatherer "clusterconfig" function "machines" failed with the error: machines.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machines" in API group "machine.openshift.io" at the cluster scope I0504 20:40:56.473813 1 gather.go:177] gatherer "clusterconfig" function "machines" took 15.624271ms to process 0 records I0504 20:40:56.473882 1 tasks_processing.go:74] worker 21 stopped. I0504 20:40:56.473933 1 tasks_processing.go:74] worker 14 stopped. I0504 20:40:56.474614 1 recorder.go:75] Recording config/infrastructure with fingerprint=f3fc466be80ba8b2b0f307cc807f8338601963dfd78c0d7e78269afbc1091f41 I0504 20:40:56.474631 1 gather.go:177] gatherer "clusterconfig" function "infrastructures" took 16.351055ms to process 1 records I0504 20:40:56.476074 1 tasks_processing.go:74] worker 17 stopped. I0504 20:40:56.476644 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/aws-pod-identity with fingerprint=1cbb66b960bd0da1dfc2d09303e54de5fc6d1e49c33f5630c128d75b0f267d22 I0504 20:40:56.476708 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-podimagespec-mutation with fingerprint=d808e3b504b4e4df4285e4b4896b5625fadb3dc7cddab5537d1683b2ebe8fab5 I0504 20:40:56.476755 1 recorder.go:75] Recording config/mutatingwebhookconfigurations/sre-service-mutation with fingerprint=574c732273070d1ffef2ac10a5d439dbab5ca0f4898343d6f570eea5e119644b I0504 20:40:56.476774 1 gather.go:177] gatherer "clusterconfig" function "mutating_webhook_configurations" took 19.406571ms to process 3 records I0504 20:40:56.478812 1 tasks_processing.go:74] worker 45 stopped. I0504 20:40:56.478960 1 recorder.go:75] Recording config/ingress with fingerprint=7c5913397bf564043e4b90ac79857b342005cecfded7201a833f9aaf42fb4d93 I0504 20:40:56.478974 1 gather.go:177] gatherer "clusterconfig" function "ingress" took 19.502378ms to process 1 records I0504 20:40:56.481324 1 tasks_processing.go:74] worker 20 stopped. I0504 20:40:56.481344 1 gather.go:177] gatherer "clusterconfig" function "openstack_version" took 24.591533ms to process 0 records I0504 20:40:56.483511 1 tasks_processing.go:74] worker 43 stopped. I0504 20:40:56.483527 1 gather.go:177] gatherer "clusterconfig" function "openshift_machine_api_events" took 22.770517ms to process 0 records I0504 20:40:56.483780 1 controller.go:254] Source periodic-clusterconfig *controllerstatus.Simple is not ready I0504 20:40:56.483796 1 controller.go:254] Source periodic-conditional *controllerstatus.Simple is not ready I0504 20:40:56.483801 1 controller.go:254] Source periodic-workloads *controllerstatus.Simple is not ready I0504 20:40:56.483807 1 controller.go:254] Source scaController *sca.Controller is not ready I0504 20:40:56.483811 1 controller.go:254] Source clusterTransferController *clustertransfer.Controller is not ready I0504 20:40:56.483833 1 controller.go:531] The operator is still being initialized I0504 20:40:56.483844 1 controller.go:554] The operator is healthy I0504 20:40:56.484008 1 tasks_processing.go:74] worker 61 stopped. I0504 20:40:56.484026 1 gather.go:177] gatherer "clusterconfig" function "container_runtime_configs" took 21.178038ms to process 0 records I0504 20:40:56.485001 1 tasks_processing.go:74] worker 37 stopped. I0504 20:40:56.485017 1 gather.go:177] gatherer "clusterconfig" function "sap_config" took 23.147112ms to process 0 records I0504 20:40:56.485328 1 tasks_processing.go:74] worker 8 stopped. I0504 20:40:56.485347 1 gather.go:177] gatherer "clusterconfig" function "storage_cluster" took 28.682611ms to process 0 records I0504 20:40:56.485986 1 tasks_processing.go:74] worker 12 stopped. I0504 20:40:56.485999 1 gather.go:177] gatherer "clusterconfig" function "jaegers" took 29.348634ms to process 0 records I0504 20:40:56.486669 1 tasks_processing.go:74] worker 1 stopped. I0504 20:40:56.486687 1 gather.go:177] gatherer "clusterconfig" function "machine_sets" took 28.351049ms to process 0 records E0504 20:40:56.486702 1 gather.go:140] gatherer "clusterconfig" function "pod_network_connectivity_checks" failed with the error: the server could not find the requested resource (get podnetworkconnectivitychecks.controlplane.operator.openshift.io) I0504 20:40:56.486711 1 gather.go:177] gatherer "clusterconfig" function "pod_network_connectivity_checks" took 22.533456ms to process 0 records I0504 20:40:56.486721 1 tasks_processing.go:74] worker 57 stopped. E0504 20:40:56.488479 1 gather_node_features.go:86] GatherNodeFeatures: NodeFeatures resource not found in openshift-nfd namespace (may not be installed) I0504 20:40:56.488502 1 tasks_processing.go:74] worker 38 stopped. I0504 20:40:56.488516 1 gather.go:177] gatherer "clusterconfig" function "node_features" took 29.148785ms to process 0 records I0504 20:40:56.489090 1 tasks_processing.go:74] worker 6 stopped. I0504 20:40:56.489276 1 recorder.go:75] Recording config/apiserver with fingerprint=6c9f445a1a2ee677092a611138700c0d85c44dd887cc511feffed344843bf9c6 I0504 20:40:56.489301 1 gather.go:177] gatherer "clusterconfig" function "cluster_apiserver" took 23.091866ms to process 1 records I0504 20:40:56.489428 1 tasks_processing.go:74] worker 32 stopped. I0504 20:40:56.489758 1 recorder.go:75] Recording config/oauth with fingerprint=804eb5740a72b4eaf593a468a27c0bbcb512b6fde0fed92451653df97d45ee55 I0504 20:40:56.489776 1 gather.go:177] gatherer "clusterconfig" function "oauths" took 29.213141ms to process 1 records I0504 20:40:56.490513 1 tasks_processing.go:74] worker 31 stopped. I0504 20:40:56.490674 1 recorder.go:75] Recording config/proxy with fingerprint=8f4d5a762fe13ab0cb7d4dbf13e00a75e4179a19247a2db39bd698382c55c9a6 I0504 20:40:56.490694 1 gather.go:177] gatherer "clusterconfig" function "proxies" took 30.671006ms to process 1 records I0504 20:40:56.490957 1 tasks_processing.go:74] worker 19 stopped. E0504 20:40:56.490974 1 gather.go:140] gatherer "clusterconfig" function "machine_configs" failed with the error: getting MachineConfigPools failed: the server could not find the requested resource (get machineconfigpools.machineconfiguration.openshift.io) I0504 20:40:56.491004 1 recorder.go:75] Recording aggregated/unused_machine_configs_count with fingerprint=4bfc9fa984e5dfcd45848faaf05269de7619bf42edf9f781751af5ee05c1a499 I0504 20:40:56.491024 1 gather.go:177] gatherer "clusterconfig" function "machine_configs" took 34.251279ms to process 1 records I0504 20:40:56.491960 1 tasks_processing.go:74] worker 60 stopped. I0504 20:40:56.492018 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkstates" took 32.24927ms to process 0 records E0504 20:40:56.492040 1 gather.go:140] gatherer "clusterconfig" function "machine_healthchecks" failed with the error: machinehealthchecks.machine.openshift.io is forbidden: User "system:serviceaccount:openshift-insights:gather" cannot list resource "machinehealthchecks" in API group "machine.openshift.io" at the cluster scope I0504 20:40:56.492055 1 gather.go:177] gatherer "clusterconfig" function "machine_healthchecks" took 31.644686ms to process 0 records I0504 20:40:56.492066 1 tasks_processing.go:74] worker 33 stopped. I0504 20:40:56.492276 1 tasks_processing.go:74] worker 52 stopped. I0504 20:40:56.492413 1 recorder.go:75] Recording config/pdbs/openshift-console/console with fingerprint=8c29dbba09195a1d94dc981c0a54e121066b3c1b9510574630ff2256c351dfaf I0504 20:40:56.492442 1 recorder.go:75] Recording config/pdbs/openshift-console/downloads with fingerprint=3f1727f250007ad185ca400a364f752494c74a65acde5177357fd5f0c4f4cc99 I0504 20:40:56.492460 1 recorder.go:75] Recording config/pdbs/openshift-image-registry/image-registry with fingerprint=58b142e5d9731d4beebd82fb3da4e3e72a880fb4736ef81df36e6441be875082 I0504 20:40:56.492477 1 recorder.go:75] Recording config/pdbs/openshift-ingress/router-default with fingerprint=68557f9611a83b22b35e34565191b608f6542dd7d39fa57072a732d7ffda2484 I0504 20:40:56.492492 1 recorder.go:75] Recording config/pdbs/openshift-monitoring/alertmanager-main with fingerprint=bc4fabf46784a43030d8db07526ac65eda3fa535d4ad01494e6a748b6838997b I0504 20:40:56.492505 1 recorder.go:75] Recording config/pdbs/openshift-monitoring/metrics-server with fingerprint=cf48203ee2a41a1c5bf1fdfa3a27502e6bdd7828636bcc5f50a5f0c864d4c36e I0504 20:40:56.492518 1 recorder.go:75] Recording config/pdbs/openshift-monitoring/monitoring-plugin with fingerprint=6ac8923d8b05f9b22c22ff6af54c8ed45d3e91d800858dc34483b0c266213917 I0504 20:40:56.492531 1 recorder.go:75] Recording config/pdbs/openshift-monitoring/prometheus-k8s with fingerprint=03e98970b82d711090f166e7c4643f32c96e3873dbf66e61ff05627e526d0e61 I0504 20:40:56.492547 1 recorder.go:75] Recording config/pdbs/openshift-monitoring/prometheus-operator-admission-webhook with fingerprint=21f082adef2c702012d3a1f27d0c0f99118d7e2f30f6da13c0e32f0fcf8ed2c3 I0504 20:40:56.492567 1 recorder.go:75] Recording config/pdbs/openshift-monitoring/thanos-querier-pdb with fingerprint=1e8fcaf251994b1f297f1ac82de4bae83b77730e4c60a73ba0f362282bd4607f I0504 20:40:56.492590 1 recorder.go:75] Recording config/pdbs/openshift-operator-lifecycle-manager/packageserver-pdb with fingerprint=e16810b01ab1964bdd6f526fe4c480f3a90d1c8f4b3b211b4895935ae75c89eb I0504 20:40:56.492602 1 gather.go:177] gatherer "clusterconfig" function "pdbs" took 30.347593ms to process 11 records I0504 20:40:56.492615 1 gather.go:177] gatherer "clusterconfig" function "lokistack" took 32.243532ms to process 0 records I0504 20:40:56.492634 1 tasks_processing.go:74] worker 36 stopped. I0504 20:40:56.492707 1 tasks_processing.go:74] worker 30 stopped. I0504 20:40:56.492981 1 recorder.go:75] Recording config/node/ip-10-0-0-208.ec2.internal with fingerprint=c2d3aa89b1770b6cf5a0a0002b9f8276a77af1794de0b0f579098f8b17146a63 I0504 20:40:56.493062 1 recorder.go:75] Recording config/node/ip-10-0-1-101.ec2.internal with fingerprint=827d80663f864fa7b817e14a2a87bfa3e5884d94db704ab25ec513706d8e47a8 I0504 20:40:56.493190 1 recorder.go:75] Recording config/node/ip-10-0-2-58.ec2.internal with fingerprint=89868bf0fc7b561711bf01748e719e0356b78d46cb49df10968dd9470aa1313b I0504 20:40:56.493203 1 gather.go:177] gatherer "clusterconfig" function "nodes" took 32.712127ms to process 3 records I0504 20:40:56.494807 1 tasks_processing.go:74] worker 29 stopped. I0504 20:40:56.494826 1 gather.go:177] gatherer "clusterconfig" function "certificate_signing_requests" took 34.718995ms to process 0 records I0504 20:40:56.496392 1 tasks_processing.go:74] worker 28 stopped. I0504 20:40:56.496406 1 gather.go:177] gatherer "clusterconfig" function "sap_pods" took 36.38493ms to process 0 records I0504 20:40:56.496554 1 tasks_processing.go:74] worker 27 stopped. I0504 20:40:56.496567 1 gather.go:177] gatherer "clusterconfig" function "openshift_logging" took 36.283116ms to process 0 records I0504 20:40:56.496576 1 gather.go:177] gatherer "clusterconfig" function "openstack_dataplanedeployments" took 36.62035ms to process 0 records I0504 20:40:56.496584 1 tasks_processing.go:74] worker 51 stopped. I0504 20:40:56.496762 1 tasks_processing.go:74] worker 15 stopped. I0504 20:40:56.496775 1 gather.go:177] gatherer "clusterconfig" function "nodenetworkconfigurationpolicies" took 39.69037ms to process 0 records I0504 20:40:56.497436 1 tasks_processing.go:74] worker 23 stopped. I0504 20:40:56.498538 1 recorder.go:75] Recording config/crd/volumesnapshots.snapshot.storage.k8s.io with fingerprint=729438e7f59b5f5be62efa9deb1732a6c14e8fab8271e354c9fe44ddc508ac8b I0504 20:40:56.499083 1 recorder.go:75] Recording config/crd/volumesnapshotcontents.snapshot.storage.k8s.io with fingerprint=9fcbaeb282c6b01dae24734d6f9f587b2dec5c2d2105313e9c762da84b9440df I0504 20:40:56.499115 1 gather.go:177] gatherer "clusterconfig" function "crds" took 40.840508ms to process 2 records I0504 20:40:56.499142 1 gather.go:177] gatherer "clusterconfig" function "node_logs" took 34.429434ms to process 0 records I0504 20:40:56.499160 1 tasks_processing.go:74] worker 42 stopped. I0504 20:40:56.500289 1 tasks_processing.go:74] worker 35 stopped. I0504 20:40:56.500483 1 recorder.go:75] Recording config/image with fingerprint=a2c3c8707303a34d84769282b655e4f298b7db47dcc45d11cfdcb9e04414925c I0504 20:40:56.500659 1 gather.go:177] gatherer "clusterconfig" function "image" took 40.541049ms to process 1 records I0504 20:40:56.501024 1 gather.go:177] gatherer "clusterconfig" function "machine_autoscalers" took 37.766532ms to process 0 records I0504 20:40:56.501036 1 gather.go:177] gatherer "clusterconfig" function "ceph_cluster" took 41.075259ms to process 0 records I0504 20:40:56.501055 1 tasks_processing.go:74] worker 56 stopped. I0504 20:40:56.501064 1 tasks_processing.go:74] worker 63 stopped. I0504 20:40:56.501422 1 tasks_processing.go:74] worker 41 stopped. I0504 20:40:56.501549 1 recorder.go:75] Recording config/validatingwebhookconfigurations/alertmanagerconfigs.openshift.io with fingerprint=ab3463c0c1e2180f6a210c53ddb74c6027ea4011b2036135e6aebd03dd8ef32c I0504 20:40:56.501623 1 recorder.go:75] Recording config/validatingwebhookconfigurations/multus.openshift.io with fingerprint=04fe6e57fc9df6454bb88c04147648ddee668d1f7623be3dac8e502f4d891396 I0504 20:40:56.501701 1 recorder.go:75] Recording config/validatingwebhookconfigurations/network-node-identity.openshift.io with fingerprint=c8b3374d0654952cce4ca3c9ea5041151302848100c230961a9f1e3ec0822b14 I0504 20:40:56.501732 1 recorder.go:75] Recording config/validatingwebhookconfigurations/performance-addon-operator with fingerprint=fa6064a9296b820413bfd229840e6c6640279bf2d9221d6953eb6e26444c7083 I0504 20:40:56.501763 1 recorder.go:75] Recording config/validatingwebhookconfigurations/prometheusrules.openshift.io with fingerprint=507d90e295e8c072e9347181cb390a5300ace6b305951766090da094cabd9289 I0504 20:40:56.501788 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterrolebindings-validation with fingerprint=f6389a4561caa39a92e3b878a154918787a1dfee4f912789717381aba469416d I0504 20:40:56.501809 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-clusterroles-validation with fingerprint=91f8d93a7c3630b59897f712a8738e40f51fbfb05852c273d82dd2d1b52a77a6 I0504 20:40:56.501830 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-ingress-config-validation with fingerprint=afb559582766f4fcd07a72adc850d917dc967ad51e9e219887e1092524be7928 I0504 20:40:56.501853 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-network-operator-validation with fingerprint=e0a381241ff1218e0c213efb0e3989c1b476512879df0c0165bdf1eae1588914 I0504 20:40:56.501884 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-regular-user-validation with fingerprint=fdeb1d3627eb183623d282969422a3fff0fadedf1f686e7a9a0e5c398140932e I0504 20:40:56.501918 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-scc-validation with fingerprint=6655293c83872b6127f3967d473ef5b2bdd69577cd5c6b46bcf728937c0c690c I0504 20:40:56.501954 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-serviceaccount-validation with fingerprint=633025c82317889c149c4efa90db6702203f9c4ea8cfb2e5499b46e838fa313e I0504 20:40:56.501981 1 recorder.go:75] Recording config/validatingwebhookconfigurations/sre-techpreviewnoupgrade-validation with fingerprint=96371d34f306e8c56a40ccf8e64fc7b2f116229c767a62b6eac36c5b7c73f003 I0504 20:40:56.501987 1 gather.go:177] gatherer "clusterconfig" function "validating_webhook_configurations" took 37.178544ms to process 13 records I0504 20:40:56.504586 1 tasks_processing.go:74] worker 18 stopped. I0504 20:40:56.504781 1 recorder.go:75] Recording config/network with fingerprint=3badc857caa9570f4bd65fe5e21d64f1c7a1f6f6b50f526cf867be905e9e74f3 I0504 20:40:56.504797 1 gather.go:177] gatherer "clusterconfig" function "networks" took 42.398655ms to process 1 records I0504 20:40:56.505434 1 tasks_processing.go:74] worker 62 stopped. I0504 20:40:56.505524 1 recorder.go:75] Recording config/olm_operators with fingerprint=35fb8bbc0ff6378651d8aa15461f2c05360065487572a3010988b7ba8e3d6855 I0504 20:40:56.505541 1 gather.go:177] gatherer "clusterconfig" function "olm_operators" took 43.003396ms to process 1 records I0504 20:40:56.507329 1 tasks_processing.go:74] worker 5 stopped. I0504 20:40:56.507584 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/admin with fingerprint=738e67eda0c2ab835aa9752d51e4aa3389b2ba2703cb2afce2dd7585df711a2b I0504 20:40:56.507739 1 recorder.go:75] Recording cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/edit with fingerprint=51fe3a37e71a21930b5a23aa65afb80b5058869fde89aec839af71b8177527d5 I0504 20:40:56.507753 1 gather.go:177] gatherer "clusterconfig" function "clusterroles" took 50.128917ms to process 2 records I0504 20:40:56.508335 1 tasks_processing.go:74] worker 55 stopped. I0504 20:40:56.508358 1 gather.go:177] gatherer "clusterconfig" function "monitoring_persistent_volumes" took 46.082242ms to process 0 records W0504 20:40:56.511353 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0504 20:40:56.514243 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0504 20:40:56.514260 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0504 20:40:56.514278 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController I0504 20:40:56.515415 1 requests.go:205] Asking for SCA certificate with "{"arch": ["x86_64"]}" payload I0504 20:40:56.516603 1 tasks_processing.go:74] worker 46 stopped. E0504 20:40:56.516629 1 gather.go:140] gatherer "clusterconfig" function "support_secret" failed with the error: secrets "support" not found I0504 20:40:56.516675 1 gather.go:177] gatherer "clusterconfig" function "support_secret" took 57.265112ms to process 0 records I0504 20:40:56.519649 1 tasks_processing.go:74] worker 2 stopped. I0504 20:40:56.519659 1 base_controller.go:82] Caches are synced for ConfigController I0504 20:40:56.519736 1 base_controller.go:119] Starting #1 worker of ConfigController controller ... I0504 20:40:56.520010 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/imagepruner/cluster with fingerprint=68083718d2756e071dd008a54f09de12601e668ea138c9b66d9905d49121626e I0504 20:40:56.520035 1 gather.go:177] gatherer "clusterconfig" function "image_pruners" took 59.317468ms to process 1 records I0504 20:40:56.520166 1 recorder.go:75] Recording config/storage/storageclasses/gp2-csi with fingerprint=0a62ef59f92f88fbf84f58973147746b95718c51b3c3ddf18c010da88cbfc7f0 I0504 20:40:56.520188 1 recorder.go:75] Recording config/storage/storageclasses/gp3-csi with fingerprint=3b5c64d674f1b138a5768ddfc4ef9ea298b6a3872866b4d6e0f8263d25479d12 I0504 20:40:56.520195 1 gather.go:177] gatherer "clusterconfig" function "storage_classes" took 61.648369ms to process 2 records I0504 20:40:56.520206 1 tasks_processing.go:74] worker 4 stopped. I0504 20:40:56.520421 1 tasks_processing.go:74] worker 47 stopped. I0504 20:40:56.520949 1 recorder.go:75] Recording config/clusteroperator/imageregistry.operator.openshift.io/config/cluster with fingerprint=ec4f8c1b890dad5f132faa61ad704b39d7cfd8ccbdb6459db00fd2ac538a51a5 I0504 20:40:56.521015 1 gather.go:177] gatherer "clusterconfig" function "image_registries" took 57.961558ms to process 1 records I0504 20:40:56.521037 1 gather.go:177] gatherer "clusterconfig" function "aggregated_monitoring_cr_names" took 57.613063ms to process 0 records I0504 20:40:56.521050 1 tasks_processing.go:74] worker 25 stopped. I0504 20:40:56.521781 1 tasks_processing.go:74] worker 53 stopped. I0504 20:40:56.521820 1 recorder.go:75] Recording config/namespaces_with_overlapping_uids with fingerprint=4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945 I0504 20:40:56.521836 1 gather.go:177] gatherer "clusterconfig" function "overlapping_namespace_uids" took 62.539976ms to process 1 records I0504 20:40:56.528040 1 configmapobserver.go:84] configmaps "insights-config" not found I0504 20:40:56.531827 1 tasks_processing.go:74] worker 48 stopped. I0504 20:40:56.532263 1 recorder.go:75] Recording config/version with fingerprint=3f85c0ab8443ee3873e613210aa919560459225d6b97e7cdc7bc90a5c9d9d733 I0504 20:40:56.532282 1 recorder.go:75] Recording config/id with fingerprint=aa587084b6ddb64893536e3f93906a20ec5e66f274c2a6c475474beec8dd98db I0504 20:40:56.532292 1 gather.go:177] gatherer "clusterconfig" function "version" took 69.411382ms to process 2 records I0504 20:40:56.544649 1 tasks_processing.go:74] worker 59 stopped. I0504 20:40:56.544683 1 gather.go:177] gatherer "clusterconfig" function "number_of_pods_and_netnamespaces_with_sdn_annotations" took 81.61321ms to process 0 records I0504 20:40:56.545075 1 tasks_processing.go:74] worker 7 stopped. I0504 20:40:56.546787 1 recorder.go:75] Recording config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m with fingerprint=1a5f79d65b66d8df089efd659c0d73dc7a91e3bc4162d5ef630294818e08782b I0504 20:40:56.546890 1 recorder.go:75] Recording config/running_containers with fingerprint=55a90fad8b8fc5705d876731ebb92a1a1ff281e6285872b46c5d7726555ec828 I0504 20:40:56.546903 1 gather.go:177] gatherer "clusterconfig" function "container_images" took 88.468929ms to process 2 records E0504 20:40:56.546915 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "cluster-monitoring-config" not found E0504 20:40:56.546920 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "gateway-mode-config" not found E0504 20:40:56.546924 1 gather.go:140] gatherer "clusterconfig" function "config_maps" failed with the error: configmaps "insights-config" not found I0504 20:40:56.546932 1 recorder.go:75] Recording config/configmaps/openshift-config/installer-images/images.json with fingerprint=ad286723f58bdcfc37aeba3ec5b4110c08e1af59cd34d14b4bfaab02d18e4856 I0504 20:40:56.546939 1 tasks_processing.go:74] worker 34 stopped. I0504 20:40:56.546973 1 recorder.go:75] Recording config/configmaps/openshift-config/kube-root-ca.crt/ca.crt with fingerprint=d476c7d3f5b104863f08f481b1264dcc68cc272ecefb0ecb709b18a6afab034d I0504 20:40:56.546992 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/invoker with fingerprint=76b482f683cd3ef9da02debac5b26080a5aeb06ff768ee5c21117514dff29d8a I0504 20:40:56.547000 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-install/version with fingerprint=2ab5ab7b1b10d7fcf1197bb24dea7c90f400e4effc18c7356873209d54fdf84b I0504 20:40:56.547011 1 recorder.go:75] Recording config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt with fingerprint=34cdf5d36d45a9e3ee7c2be31546e3d937ae4ebf144ac07b0210c06a1610a94d I0504 20:40:56.547050 1 recorder.go:75] Recording config/configmaps/openshift-config/rosa-brand-logo/rosa-brand-logo.svg with fingerprint=6ed8ca4dd7a8eee7249182bc006e9649ce84d76c551ddfaaa33e55d8c4cc1ed0 I0504 20:40:56.547057 1 recorder.go:75] Recording config/configmaps/kube-system/cluster-config-v1/install-config with fingerprint=ab3811c6b83fd7b8e920094cfa3080d1b4ee3c35ec4c8379437b21d27bd6608d I0504 20:40:56.547062 1 gather.go:177] gatherer "clusterconfig" function "config_maps" took 84.754146ms to process 7 records I0504 20:40:56.567656 1 base_controller.go:82] Caches are synced for LoggingSyncer I0504 20:40:56.567675 1 base_controller.go:119] Starting #1 worker of LoggingSyncer controller ... I0504 20:40:56.573200 1 cluster_transfer.go:102] no available accepted cluster transfer I0504 20:40:56.573217 1 controllerstatus.go:80] name=clusterTransferController healthy=true reason=NoClusterTransfer message=no available cluster transfer I0504 20:40:56.608703 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0504 20:40:56.609372 1 gather_cluster_operator_pods_and_events.go:121] Found 28 pods with 73 containers I0504 20:40:56.609385 1 gather_cluster_operator_pods_and_events.go:235] Maximum buffer size: 344737 bytes I0504 20:40:56.609395 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for init-config-reloader container alertmanager-main-0 pod in namespace openshift-monitoring (previous: false). I0504 20:40:56.615289 1 gather_cluster_operators.go:184] Unable to get dnsrecords.ingress.operator.openshift.io resource due to: dnsrecords.ingress.operator.openshift.io "default" not found I0504 20:40:56.628870 1 tasks_processing.go:74] worker 58 stopped. E0504 20:40:56.628888 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret 'router-certs-default' in namespace 'openshift-ingress': secrets "router-certs-default" not found E0504 20:40:56.628895 1 gather.go:140] gatherer "clusterconfig" function "ingress_certificates" failed with the error: failed to get secret '2q3b3os70c667l7uhbfrt2sq5dfhc2b8-primary-cert-bundle-secret' in namespace 'openshift-ingress-operator': secrets "2q3b3os70c667l7uhbfrt2sq5dfhc2b8-primary-cert-bundle-secret" not found I0504 20:40:56.628969 1 recorder.go:75] Recording aggregated/ingress_controllers_certs with fingerprint=c8f9411fda97e6f875a77f1c8151cf104bf86bab7ed052cdd3030a0c8e943cf1 I0504 20:40:56.628990 1 gather.go:177] gatherer "clusterconfig" function "ingress_certificates" took 168.771723ms to process 1 records I0504 20:40:56.649317 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for alertmanager container alertmanager-main-0 pod in namespace openshift-monitoring (previous: false). I0504 20:40:56.663974 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for config-reloader container alertmanager-main-0 pod in namespace openshift-monitoring (previous: false). I0504 20:40:56.721358 1 sca.go:266] etc-pki-entitlement secret successfully created I0504 20:40:56.721378 1 controllerstatus.go:80] name=scaController healthy=true reason=Updated message=SCA certs successfully updated I0504 20:40:56.729494 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-web container alertmanager-main-0 pod in namespace openshift-monitoring (previous: false). I0504 20:40:56.926531 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container alertmanager-main-0 pod in namespace openshift-monitoring (previous: false). I0504 20:40:57.125739 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-metric container alertmanager-main-0 pod in namespace openshift-monitoring (previous: false). I0504 20:40:57.326247 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prom-label-proxy container alertmanager-main-0 pod in namespace openshift-monitoring (previous: false). W0504 20:40:57.511004 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0504 20:40:57.513220 1 tasks_processing.go:74] worker 49 stopped. I0504 20:40:57.513243 1 recorder.go:75] Recording config/silenced_alerts.json with fingerprint=37517e5f3dc66819f61f5a7bb8ace1921282415f10551d2defa5c3eb0985b570 I0504 20:40:57.513258 1 gather.go:177] gatherer "clusterconfig" function "silenced_alerts" took 1.048343713s to process 1 records I0504 20:40:57.513325 1 tasks_processing.go:74] worker 44 stopped. I0504 20:40:57.513382 1 recorder.go:75] Recording config/alerts with fingerprint=4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945 W0504 20:40:57.513403 1 gather.go:155] issue recording gatherer "clusterconfig" function "active_alerts" result "config/alerts.json" because of the warning: warning: the record with the same fingerprint "4f53cda18c2baa0c0354bb5f9a3ecbe5ed12ab4d8e11ba873c2f11161202b945" was already recorded at path "config/namespaces_with_overlapping_uids.json", recording another one with a different path "config/alerts.json" I0504 20:40:57.513417 1 gather.go:177] gatherer "clusterconfig" function "active_alerts" took 1.052741086s to process 1 records I0504 20:40:57.527834 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for init-config-reloader container alertmanager-main-1 pod in namespace openshift-monitoring (previous: false). I0504 20:40:57.726430 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for alertmanager container alertmanager-main-1 pod in namespace openshift-monitoring (previous: false). I0504 20:40:57.930230 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for config-reloader container alertmanager-main-1 pod in namespace openshift-monitoring (previous: false). I0504 20:40:58.124681 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-web container alertmanager-main-1 pod in namespace openshift-monitoring (previous: false). I0504 20:40:58.324602 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container alertmanager-main-1 pod in namespace openshift-monitoring (previous: false). W0504 20:40:58.511191 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0504 20:40:58.523842 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-metric container alertmanager-main-1 pod in namespace openshift-monitoring (previous: false). I0504 20:40:58.724344 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prom-label-proxy container alertmanager-main-1 pod in namespace openshift-monitoring (previous: false). I0504 20:40:58.923964 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for cluster-monitoring-operator container cluster-monitoring-operator-557b74df6-k2skl pod in namespace openshift-monitoring (previous: false). I0504 20:40:58.958049 1 tasks_processing.go:74] worker 26 stopped. I0504 20:40:58.958206 1 recorder.go:75] Recording config/clusteroperator/console with fingerprint=2737647a5d4dd8e0cbe5169d888c11bfd0162b1ed9db5b8c9a6a9f93383b68ad I0504 20:40:58.958267 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/console/cluster with fingerprint=be84fc75c87766f0f4d4e9fb1f9db618d008a21cc463e8239af81a1c296612d1 I0504 20:40:58.958313 1 recorder.go:75] Recording config/clusteroperator/csi-snapshot-controller with fingerprint=c0ea475bdca6cc2b2ebe49b7e76b8775917af87ad68e6b941788fcb4aeb0870e I0504 20:40:58.958332 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/csisnapshotcontroller/cluster with fingerprint=5adc514f4b63e2f1ecc68bf6f9c0af70c5eea04522a49524e102721b1c41f80e I0504 20:40:58.958362 1 recorder.go:75] Recording config/clusteroperator/dns with fingerprint=8622b91b4cdf8da2945ab56e3aee10157fffdf14e55ecaad2d3ccd5ad4852f98 I0504 20:40:58.958387 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/dns/default with fingerprint=9e7b4ce029030d3d8c3b49af92c556acdcc415000b40d3f969dbdc42c432b47f I0504 20:40:58.958420 1 recorder.go:75] Recording config/clusteroperator/image-registry with fingerprint=16360a491c3d0cda9ca7bb3778991b177793195f4363f02c67aed20e8d89ca90 I0504 20:40:58.958460 1 recorder.go:75] Recording config/clusteroperator/ingress with fingerprint=ef81e50342cf481c011ece5761554c57d168901c065a5f417c8db67ea8bc442d I0504 20:40:58.958494 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/ingresscontroller/openshift-ingress-operator/default with fingerprint=7c238efa9b3412103d7be215f7f1f11f56d7891ad5f8b01761ea861b480cc94b I0504 20:40:58.958533 1 recorder.go:75] Recording config/clusteroperator/insights with fingerprint=6bc85e55de5d6fdf8aae88a9e950b7155d133c81e1fc00e650ed23395ecaf5f6 I0504 20:40:58.958820 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/insightsoperator/cluster with fingerprint=f403da3b0b07babb139eca29df3b27085219871ad6b354226d7a1c2880b9a09c I0504 20:40:58.958852 1 recorder.go:75] Recording config/clusteroperator/kube-apiserver with fingerprint=a58b3ec0f01e059c505aba93cdd8b37f004e03ce9af6714b6e3001f745110bc4 I0504 20:40:58.958866 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubeapiserver/cluster with fingerprint=51503bf0b784fcf65ea46bcaf1f72ac1a5c4d5dc211934f18f27871efed05762 I0504 20:40:58.958891 1 recorder.go:75] Recording config/clusteroperator/kube-controller-manager with fingerprint=b9d613450c24d8cd8956a0704d8f39a14fe18cca81209fba0c3bcac883c6f34e I0504 20:40:58.958907 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubecontrollermanager/cluster with fingerprint=ce90c0d4f367d7da085074268031798382ae7c54fdcb0a21f15a4818fe308c11 I0504 20:40:58.958933 1 recorder.go:75] Recording config/clusteroperator/kube-scheduler with fingerprint=07d33f4888457b75fc9a8bc6c35d4df8eb5cb3ff302c981bc07e309207128246 I0504 20:40:58.958948 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubescheduler/cluster with fingerprint=f2940fb9fd20c19951dfc295eb363b7fba0c505f5ae61f01967a063099e6b60a I0504 20:40:58.958976 1 recorder.go:75] Recording config/clusteroperator/kube-storage-version-migrator with fingerprint=0e30c0f77eab8ffc3ef33d2d0d90a541ee90432bbd421c7c4c645b5d4d4dd202 I0504 20:40:58.958990 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/kubestorageversionmigrator/cluster with fingerprint=9351181aa7e6ada41ef581ab31e13516c6b934cc95710154bafb2eb222cb58db I0504 20:40:58.959019 1 recorder.go:75] Recording config/clusteroperator/monitoring with fingerprint=1ddc42a0eeb704a1905efeaaddc69fa1a1220d49ff759f79f15c9902534bb4ee I0504 20:40:58.959148 1 recorder.go:75] Recording config/clusteroperator/network with fingerprint=20713a1805a7cada77b3c8fe233b1f643346b5777b1ad2545289d75bce774b9f I0504 20:40:58.959164 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/ovn with fingerprint=626a89d20e0deaed5b6dfb533acfe65f4bb1618bd200a703b62e60c5d16d94ab I0504 20:40:58.959177 1 recorder.go:75] Recording config/clusteroperator/network.operator.openshift.io/operatorpki/openshift-ovn-kubernetes/signer with fingerprint=90410b16914712b85b3c4578716ad8c0ae072e688f4cd1e022bf76f20da3506d I0504 20:40:58.959212 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/network/cluster with fingerprint=a93d15eaecb455a0e40ecb2826eeecc1533899204ddd3c3921d15ab70af7ae75 I0504 20:40:58.959250 1 recorder.go:75] Recording config/clusteroperator/node-tuning with fingerprint=a0f51c1af207dfada9e87f0bc5775e199298c13dcd9f4463a96aa06f4b955c32 I0504 20:40:58.959296 1 recorder.go:75] Recording config/clusteroperator/openshift-apiserver with fingerprint=ee8143ec681462c3fc9a6a8ca6e00c6b1b3dc8aaa68cc7d71303b6651718fc76 I0504 20:40:58.959311 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftapiserver/cluster with fingerprint=e712e6cf27339b441e4ed1f4cde91dbde7e952698ba93407e4457db63a4a4c76 I0504 20:40:58.959339 1 recorder.go:75] Recording config/clusteroperator/openshift-controller-manager with fingerprint=d51712f3e63635898e53d6aa311dc59e4347f9dc0a9d3bd4d0b5840bddb3c525 I0504 20:40:58.959354 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/openshiftcontrollermanager/cluster with fingerprint=d71a0f4672f9b45d9fc8293bf1687afc650fd28d32e2e30de27523fe7b4eadf7 I0504 20:40:58.959382 1 recorder.go:75] Recording config/clusteroperator/openshift-samples with fingerprint=129e29347aa0500680d63cb8f77d3925e4ef65cb9ea93a9a98047d05f76caf4f I0504 20:40:58.959397 1 recorder.go:75] Recording config/clusteroperator/samples.operator.openshift.io/config/cluster with fingerprint=439a4284281b8dcef0621bb14ba23e2175a28a50613f86b33a171c49689474fd I0504 20:40:58.959421 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager with fingerprint=0c61380654eea56fb9ff536461bd2b906f8cb8e437157da55b76810c07ea653f I0504 20:40:58.959448 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-catalog with fingerprint=295f016207d3be94852a37129a27ec12f94383de998d5736eae382ea64d3b66f I0504 20:40:58.959478 1 recorder.go:75] Recording config/clusteroperator/operator-lifecycle-manager-packageserver with fingerprint=b448e3b8782b980743f0a40769807ba6e31df8c503a7e7bf203c7a8bfad58a52 I0504 20:40:58.959513 1 recorder.go:75] Recording config/clusteroperator/service-ca with fingerprint=fbe26ef3ffb24660032b27dd854943521d10648f759a805cca27a9bd6ef8db60 I0504 20:40:58.959527 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/serviceca/cluster with fingerprint=812f7edc2cdb30e61e7f2b29454357a40b1a507a4b0c2b7729193b67f0e3b4aa I0504 20:40:58.959569 1 recorder.go:75] Recording config/clusteroperator/storage with fingerprint=6a9186a6e5286c4a62f1279df1d2003f08259ed6b2efc22f72e93fdc07522cbf I0504 20:40:58.959597 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/clustercsidriver/ebs.csi.aws.com with fingerprint=510064d6f6bcced87ab5bd2ddaff3d0edd7f93f4a4f7af2641f29fc53ffab21e I0504 20:40:58.959611 1 recorder.go:75] Recording config/clusteroperator/operator.openshift.io/storage/cluster with fingerprint=8e480f8c1ce1b39baac42d8ec780c57c2592929ae0c801b61ffad49ba13f33ad I0504 20:40:58.959623 1 gather.go:177] gatherer "clusterconfig" function "operators" took 2.492430267s to process 39 records I0504 20:40:59.127793 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-state-metrics container kube-state-metrics-6dfffbb599-5x92w pod in namespace openshift-monitoring (previous: false). I0504 20:40:59.326736 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-main container kube-state-metrics-6dfffbb599-5x92w pod in namespace openshift-monitoring (previous: false). W0504 20:40:59.511626 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0504 20:40:59.527288 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-self container kube-state-metrics-6dfffbb599-5x92w pod in namespace openshift-monitoring (previous: false). I0504 20:40:59.725979 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for metrics-server container metrics-server-7bf585d689-9x26q pod in namespace openshift-monitoring (previous: false). I0504 20:40:59.932781 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for metrics-server container metrics-server-7bf585d689-txtxt pod in namespace openshift-monitoring (previous: false). I0504 20:41:00.123996 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for monitoring-plugin container monitoring-plugin-5cbb48bf6b-m6wqz pod in namespace openshift-monitoring (previous: false). I0504 20:41:00.326756 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for monitoring-plugin container monitoring-plugin-5cbb48bf6b-x59fv pod in namespace openshift-monitoring (previous: false). W0504 20:41:00.511536 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. I0504 20:41:00.524274 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for init-textfile container node-exporter-2bjm9 pod in namespace openshift-monitoring (previous: false). I0504 20:41:00.726046 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0504 20:41:00.726063 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-exporter container node-exporter-2bjm9 pod in namespace openshift-monitoring (previous: false). I0504 20:41:00.926516 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container node-exporter-2bjm9 pod in namespace openshift-monitoring (previous: false). I0504 20:41:01.135920 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for init-textfile container node-exporter-gttf2 pod in namespace openshift-monitoring (previous: false). I0504 20:41:01.324818 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0504 20:41:01.324834 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-exporter container node-exporter-gttf2 pod in namespace openshift-monitoring (previous: false). W0504 20:41:01.511097 1 gather_dvo_metrics.go:210] Failed to read the DVO metrics. Trying again. W0504 20:41:01.511136 1 gather_dvo_metrics.go:117] Unable to read metrics from endpoint "http://deployment-validation-operator-metrics.openshift-deployment-validation-operator.svc:8383": DVO metrics service was not available within the 5s timeout: context deadline exceeded I0504 20:41:01.511154 1 tasks_processing.go:74] worker 40 stopped. E0504 20:41:01.511169 1 gather.go:140] gatherer "clusterconfig" function "dvo_metrics" failed with the error: DVO metrics service was not available within the 5s timeout: context deadline exceeded I0504 20:41:01.511184 1 recorder.go:75] Recording config/dvo_metrics with fingerprint=e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 I0504 20:41:01.511198 1 gather.go:177] gatherer "clusterconfig" function "dvo_metrics" took 5.046445844s to process 1 records I0504 20:41:01.527267 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container node-exporter-gttf2 pod in namespace openshift-monitoring (previous: false). I0504 20:41:01.728121 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for init-textfile container node-exporter-krgvb pod in namespace openshift-monitoring (previous: false). I0504 20:41:01.926409 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0504 20:41:01.926428 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for node-exporter container node-exporter-krgvb pod in namespace openshift-monitoring (previous: false). I0504 20:41:02.128078 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container node-exporter-krgvb pod in namespace openshift-monitoring (previous: false). I0504 20:41:02.326284 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-main container openshift-state-metrics-b8587444b-g6p9n pod in namespace openshift-monitoring (previous: false). I0504 20:41:02.526506 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-self container openshift-state-metrics-b8587444b-g6p9n pod in namespace openshift-monitoring (previous: false). I0504 20:41:02.725936 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for openshift-state-metrics container openshift-state-metrics-b8587444b-g6p9n pod in namespace openshift-monitoring (previous: false). I0504 20:41:02.925701 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for init-config-reloader container prometheus-k8s-0 pod in namespace openshift-monitoring (previous: false). I0504 20:41:03.130099 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prometheus container prometheus-k8s-0 pod in namespace openshift-monitoring (previous: false). I0504 20:41:03.325944 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for config-reloader container prometheus-k8s-0 pod in namespace openshift-monitoring (previous: false). I0504 20:41:03.525094 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for thanos-sidecar container prometheus-k8s-0 pod in namespace openshift-monitoring (previous: false). I0504 20:41:03.725166 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-web container prometheus-k8s-0 pod in namespace openshift-monitoring (previous: false). I0504 20:41:03.924715 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container prometheus-k8s-0 pod in namespace openshift-monitoring (previous: false). I0504 20:41:04.125121 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-thanos container prometheus-k8s-0 pod in namespace openshift-monitoring (previous: false). I0504 20:41:04.324803 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for init-config-reloader container prometheus-k8s-1 pod in namespace openshift-monitoring (previous: false). I0504 20:41:04.525212 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prometheus container prometheus-k8s-1 pod in namespace openshift-monitoring (previous: false). I0504 20:41:04.728864 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for config-reloader container prometheus-k8s-1 pod in namespace openshift-monitoring (previous: false). I0504 20:41:04.926098 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for thanos-sidecar container prometheus-k8s-1 pod in namespace openshift-monitoring (previous: false). I0504 20:41:05.123184 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-web container prometheus-k8s-1 pod in namespace openshift-monitoring (previous: false). I0504 20:41:05.322588 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container prometheus-k8s-1 pod in namespace openshift-monitoring (previous: false). I0504 20:41:05.522891 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-thanos container prometheus-k8s-1 pod in namespace openshift-monitoring (previous: false). I0504 20:41:05.723011 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prometheus-operator container prometheus-operator-76b589d668-bbn8q pod in namespace openshift-monitoring (previous: false). I0504 20:41:05.924903 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container prometheus-operator-76b589d668-bbn8q pod in namespace openshift-monitoring (previous: false). I0504 20:41:06.122354 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prometheus-operator-admission-webhook container prometheus-operator-admission-webhook-59b79769fb-gfbb7 pod in namespace openshift-monitoring (previous: false). I0504 20:41:06.322684 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prometheus-operator-admission-webhook container prometheus-operator-admission-webhook-59b79769fb-pcdzn pod in namespace openshift-monitoring (previous: false). I0504 20:41:06.530648 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for thanos-query container thanos-querier-57d4668466-jt7dg pod in namespace openshift-monitoring (previous: false). I0504 20:41:06.722464 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-web container thanos-querier-57d4668466-jt7dg pod in namespace openshift-monitoring (previous: false). I0504 20:41:06.922850 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container thanos-querier-57d4668466-jt7dg pod in namespace openshift-monitoring (previous: false). I0504 20:41:07.122530 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prom-label-proxy container thanos-querier-57d4668466-jt7dg pod in namespace openshift-monitoring (previous: false). I0504 20:41:07.321677 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-rules container thanos-querier-57d4668466-jt7dg pod in namespace openshift-monitoring (previous: false). I0504 20:41:07.524028 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-metrics container thanos-querier-57d4668466-jt7dg pod in namespace openshift-monitoring (previous: false). I0504 20:41:07.723870 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for thanos-query container thanos-querier-57d4668466-xlt5t pod in namespace openshift-monitoring (previous: false). I0504 20:41:07.924772 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-web container thanos-querier-57d4668466-xlt5t pod in namespace openshift-monitoring (previous: false). I0504 20:41:08.122699 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy container thanos-querier-57d4668466-xlt5t pod in namespace openshift-monitoring (previous: false). I0504 20:41:08.322768 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for prom-label-proxy container thanos-querier-57d4668466-xlt5t pod in namespace openshift-monitoring (previous: false). I0504 20:41:08.522852 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-rules container thanos-querier-57d4668466-xlt5t pod in namespace openshift-monitoring (previous: false). I0504 20:41:08.723435 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for kube-rbac-proxy-metrics container thanos-querier-57d4668466-xlt5t pod in namespace openshift-monitoring (previous: false). I0504 20:41:08.922706 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for networking-console-plugin container networking-console-plugin-7f449dfbfd-b6t2k pod in namespace openshift-network-console (previous: false). I0504 20:41:09.123819 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for networking-console-plugin container networking-console-plugin-7f449dfbfd-zrqkw pod in namespace openshift-network-console (previous: false). I0504 20:41:09.279653 1 tasks_processing.go:74] worker 9 stopped. I0504 20:41:09.279697 1 recorder.go:75] Recording config/installplans with fingerprint=95dfd5f33a9a46199239158a0ec0183b629818eca24ae41f4d706d7a3d604aec I0504 20:41:09.279711 1 gather.go:177] gatherer "clusterconfig" function "install_plans" took 12.822314175s to process 1 records I0504 20:41:09.323638 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for console-operator container console-operator-66dbbb6468-5pf5m pod in namespace openshift-console-operator (previous: true). I0504 20:41:09.523802 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for console-operator container console-operator-66dbbb6468-5pf5m pod in namespace openshift-console-operator (previous: false). I0504 20:41:09.726397 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for console container console-5c56547d87-mztrn pod in namespace openshift-console (previous: false). I0504 20:41:09.923702 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for console container console-5c56547d87-z9tb6 pod in namespace openshift-console (previous: false). I0504 20:41:10.065217 1 tasks_processing.go:74] worker 0 stopped. I0504 20:41:10.065488 1 recorder.go:75] Recording config/serviceaccounts with fingerprint=60cf7c92813366cf47e5b7fc386c82c00b6c4f395901a511340826e94e3650c4 I0504 20:41:10.065507 1 gather.go:177] gatherer "clusterconfig" function "service_accounts" took 13.608524095s to process 1 records I0504 20:41:10.123850 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for console container console-746d468875-7cgwt pod in namespace openshift-console (previous: false). I0504 20:41:10.316738 1 gather_cluster_operator_pods_and_events.go:280] Error: "log buffer is empty" I0504 20:41:10.316755 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for console container console-746d468875-zzrsl pod in namespace openshift-console (previous: false). I0504 20:41:10.527703 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for download-server container downloads-5c54f4944f-7x4tz pod in namespace openshift-console (previous: false). I0504 20:41:10.728771 1 gather_cluster_operator_pods_and_events.go:365] Fetching logs for download-server container downloads-5c54f4944f-jcjp6 pod in namespace openshift-console (previous: false). I0504 20:41:10.925280 1 tasks_processing.go:74] worker 54 stopped. I0504 20:41:10.925933 1 recorder.go:75] Recording events/openshift-monitoring with fingerprint=e6a92dc17ad5037679a24011f00c26c6beae492a88c51a2536b29976566b0c3e I0504 20:41:10.925977 1 recorder.go:75] Recording events/openshift-network-console with fingerprint=02aa97f6712e3b613b6036747faf3d87fabb385a5b0a0cd8838883aa537e0227 I0504 20:41:10.926068 1 recorder.go:75] Recording events/openshift-console-operator with fingerprint=1c8768d8f934c1c5fb86811f2abe940f8df98bc4303d219eb4abb432d2bbda45 I0504 20:41:10.926146 1 recorder.go:75] Recording events/openshift-console with fingerprint=ca96c6d471a5a7e60ef9a9424cc154941a9dfcc8d703ace74383d3281cef46ff I0504 20:41:10.926157 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-0/init-config-reloader_current.log with fingerprint=60b8e55c8180a4c38275e024ce6cef7bdc8deac08d8f1c22eb2f42e8606ff6da I0504 20:41:10.926170 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-0/alertmanager_current.log with fingerprint=dbf3558d79f86bb1943b7ad69b809d23c159a4facd8a7febac67b42b23578196 I0504 20:41:10.926197 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-0/config-reloader_current.log with fingerprint=d74e34aa9a05d1d995065be786a7caf7375ff005e3097eb8616980fd17e020db I0504 20:41:10.926221 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-0/kube-rbac-proxy-web_current.log with fingerprint=f557cd9a29ee426a468e3bcbb815826c038b817ada5dc7bee17e4e81c4ee4511 I0504 20:41:10.926229 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-0/kube-rbac-proxy_current.log with fingerprint=3647bf7b56630f9ad8723661a1123afd18b79745ee2f74e6ac8e57a80b03b0ef I0504 20:41:10.926235 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-0/kube-rbac-proxy-metric_current.log with fingerprint=371951d905d8f8f02fba3a62e0c6e12a4cc629a24fdc6bf28061bcbca2212250 I0504 20:41:10.926240 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-0/prom-label-proxy_current.log with fingerprint=1b51bcf6185f22909b074c17b6fdf1eda0bcd0191bff368ae7e5572296f00c42 I0504 20:41:10.926246 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-1/init-config-reloader_current.log with fingerprint=55369a0bc4870913e24e985ea7d9dbb40f161cf1b37caec57a9446d1a4a1c708 I0504 20:41:10.926258 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-1/alertmanager_current.log with fingerprint=ea0e4f6b4a02c67002fc422bd582592d1ee87721a0dbefedecf46dbfd7cae038 I0504 20:41:10.926272 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-1/config-reloader_current.log with fingerprint=d5db909e826e146704f827161a82e719922fc6ee0abd04e69e5984f46bca4b9a I0504 20:41:10.926278 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-1/kube-rbac-proxy-web_current.log with fingerprint=e3cc66636b3a0c6348c2c03809fd78848b6c40b8550fea0e1b68e91e449938d1 I0504 20:41:10.926284 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-1/kube-rbac-proxy_current.log with fingerprint=80e83549087add68f8a4b5a9b9f8f53f206b9c7088b1e9e3c22ddeb96e066872 I0504 20:41:10.926291 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-1/kube-rbac-proxy-metric_current.log with fingerprint=655f0b0b7209704a7b134e68cd1a3d1184831330e38ac81bc85a6927ee92f255 I0504 20:41:10.926295 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/alertmanager-main-1/prom-label-proxy_current.log with fingerprint=192f3828f2c256fdddae3a504dd1d17627a40712322f2f9fac5a56d4f1fb7870 I0504 20:41:10.926413 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/cluster-monitoring-operator-557b74df6-k2skl/cluster-monitoring-operator_current.log with fingerprint=17494ef0298fca03e1c96e39dfcd89960bc8daa9a667d0c9d8fcd303d46226cf I0504 20:41:10.926438 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/kube-state-metrics-6dfffbb599-5x92w/kube-state-metrics_current.log with fingerprint=48b3fa189f1ccbcf326e6dee7daf72c1abf45b202e6bb38a4262a252c7b4dc2a I0504 20:41:10.926449 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/kube-state-metrics-6dfffbb599-5x92w/kube-rbac-proxy-main_current.log with fingerprint=f50b9744ebddde6374ddfa9bb7d7ad4b54c0eb63f0d040608a07e56f2b843eac I0504 20:41:10.926455 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/kube-state-metrics-6dfffbb599-5x92w/kube-rbac-proxy-self_current.log with fingerprint=79101bb5215017736678cc63a475a6cc57434a0dfc9cca92ab473ca744b34e00 I0504 20:41:10.926463 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/metrics-server-7bf585d689-9x26q/metrics-server_current.log with fingerprint=1b304923313cba5976ee18472161e6cc5acc58d8c722337eb36c98936640343c I0504 20:41:10.926470 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/metrics-server-7bf585d689-txtxt/metrics-server_current.log with fingerprint=6fd63662c088a143bdcf3c745b113461cd2373562d2bd79e01580c0f51b63c85 I0504 20:41:10.926481 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/monitoring-plugin-5cbb48bf6b-m6wqz/monitoring-plugin_current.log with fingerprint=d1d3f2ac9e64a016bb2162efee8b1151993140f9d622069249d9529823054f8d I0504 20:41:10.926493 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/monitoring-plugin-5cbb48bf6b-x59fv/monitoring-plugin_current.log with fingerprint=babfcd32d661cf2cb5fe30710962b12d9e3143a62194175024614afbde871b22 I0504 20:41:10.926519 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/node-exporter-2bjm9/node-exporter_current.log with fingerprint=b8b691603ced38beefd92a8eda231d93a0aaf975927ff2af5e5bfbce95d24009 I0504 20:41:10.926528 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/node-exporter-2bjm9/kube-rbac-proxy_current.log with fingerprint=71d0cf60653767e280ef2f1d0a528bcef96b3e746d8b12b96f680c9de2af0f6e I0504 20:41:10.926553 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/node-exporter-gttf2/node-exporter_current.log with fingerprint=7e56643dddc3f4a3b928e7f42b8fd9c18e03622039a20f4f8ad047021325573a I0504 20:41:10.926563 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/node-exporter-gttf2/kube-rbac-proxy_current.log with fingerprint=db694648ae5c6488801312ae01c1589bfa0105bf4b6b3dbf569cb454390b6219 I0504 20:41:10.926589 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/node-exporter-krgvb/node-exporter_current.log with fingerprint=73c279ed32a7c8315de4ad68a69f17de80fa83c03f378a04acacab78372753f4 I0504 20:41:10.926597 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/node-exporter-krgvb/kube-rbac-proxy_current.log with fingerprint=3ede484eaafa64f98c0f6d6770ff5b4160b20a0ca938401f2bf9b7071268e153 I0504 20:41:10.926603 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/openshift-state-metrics-b8587444b-g6p9n/kube-rbac-proxy-main_current.log with fingerprint=197a3acdf86ffd6b04fb66b7b073bfb59515306db045e706295b401d8c92e582 I0504 20:41:10.926609 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/openshift-state-metrics-b8587444b-g6p9n/kube-rbac-proxy-self_current.log with fingerprint=5eb107e7a2700f32dbf63246df2cc9587fb1e64fb8441671d56f66203909abe7 I0504 20:41:10.926619 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/openshift-state-metrics-b8587444b-g6p9n/openshift-state-metrics_current.log with fingerprint=31535f7986e94381c89baf0ebac86b54ab69b1a32bf892d8ad70469e12783687 I0504 20:41:10.926626 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-0/init-config-reloader_current.log with fingerprint=71a216da93753349067d0e5825cf6470ae713ca25927a717d9c1e8a39cffbb57 I0504 20:41:10.926697 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-0/prometheus_current.log with fingerprint=968c6b3c4b39ff5d07ee2c223576b034066c99716ff9578c2bb4a139254187c2 I0504 20:41:10.926711 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-0/config-reloader_current.log with fingerprint=3dc15548fb5cc5701339ecc32b26252f1ea6ae2ee0b84d9dac14b64a39ed13cf I0504 20:41:10.926722 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-0/thanos-sidecar_current.log with fingerprint=7d5dab6d0f929b129f07ea6e37f8e68a026ee6dba4f90809f8268beea3726d82 I0504 20:41:10.926729 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-0/kube-rbac-proxy-web_current.log with fingerprint=f17e0071302a8cca7567756dab41d9d477097163b024ab8455dd546af2b8907f I0504 20:41:10.926734 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-0/kube-rbac-proxy_current.log with fingerprint=651c6196c36f0a495ca50f3a1fa061bff2535a470411d9ede1cf05f1da537653 I0504 20:41:10.926742 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-0/kube-rbac-proxy-thanos_current.log with fingerprint=53ee760388631145319948683123c0d3a4a85927358631060d4d9a1882e2eb9b I0504 20:41:10.926748 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-1/init-config-reloader_current.log with fingerprint=9f5b948c151337d262bd5e594da698ca6cb00464f11a7f2f614ce0e0a5862600 I0504 20:41:10.926815 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-1/prometheus_current.log with fingerprint=5938a53dcb3b72cf07cdeda8e2b2480eebb6185d3045459c6884addc82df5b0d I0504 20:41:10.926829 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-1/config-reloader_current.log with fingerprint=a4cbe3762a11f35ac2eb9ee4fc4971259a8741b95d81fd9052fff04b3e864fcf I0504 20:41:10.926839 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-1/thanos-sidecar_current.log with fingerprint=ab58ffacb89993b8b488a1597394e17bd59917465e866ec27ced3aa01783e813 I0504 20:41:10.926846 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-1/kube-rbac-proxy-web_current.log with fingerprint=7be7850d4b9f58308f66222c9be512f63304891c85adf1c96bc96a3689540b98 I0504 20:41:10.926851 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-1/kube-rbac-proxy_current.log with fingerprint=d2c4ce9733b3a1565bcf13ba56299d36e2dfec392c5faf4bc5107ae72b4872ab I0504 20:41:10.926857 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-k8s-1/kube-rbac-proxy-thanos_current.log with fingerprint=ca18970515e8536a4884ba2ddf7af5d35f1a90fcaa6cb36d25afc15177cdb1c8 I0504 20:41:10.926935 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-operator-76b589d668-bbn8q/prometheus-operator_current.log with fingerprint=f9cd9a3919656d36f163502cbc70a4ef44154e206e65d59d8e8e3deae46eb9c2 I0504 20:41:10.926944 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-operator-76b589d668-bbn8q/kube-rbac-proxy_current.log with fingerprint=2b6de7030d289c207f44e28eda07fc58e5ecf12145ff5b349448b92371066890 I0504 20:41:10.926951 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-operator-admission-webhook-59b79769fb-gfbb7/prometheus-operator-admission-webhook_current.log with fingerprint=dddbb8b9ea8585e0275442be1831953196d3324b45030bd82f2912f1dbc226ff I0504 20:41:10.926960 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/prometheus-operator-admission-webhook-59b79769fb-pcdzn/prometheus-operator-admission-webhook_current.log with fingerprint=05acc301776ce9e0c0d6272a688d35c7dfb72ef5a9be2a4a3df37c018e739324 I0504 20:41:10.926972 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-jt7dg/thanos-query_current.log with fingerprint=6cd434f1e99a14ec6e488a79b5ec8e588dcb587eccaafa412279b47c4c4a1c2b I0504 20:41:10.926982 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-jt7dg/kube-rbac-proxy-web_current.log with fingerprint=3f1bf63348fc5df6e049f16328c8d7f372f8b3fab7cf9d83d6f1f0993bfc03c4 I0504 20:41:10.926988 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-jt7dg/kube-rbac-proxy_current.log with fingerprint=879fd849de0d8a41c67adb341c984339168bd08874e189a9cbfc27626c26d40f I0504 20:41:10.926994 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-jt7dg/prom-label-proxy_current.log with fingerprint=4c1b464585e9f6b30c898448a14938fddde2b01d193b37543b8bb4889d3a1aef I0504 20:41:10.927000 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-jt7dg/kube-rbac-proxy-rules_current.log with fingerprint=825ddbdcfd0a06af7fdd45b28881ca64ee115b81d07cb69e9c6314320df9e7e0 I0504 20:41:10.927008 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-jt7dg/kube-rbac-proxy-metrics_current.log with fingerprint=b5d7d2cc17afff7d3c0be182b173ccc3bae9a90e85859d9efa5aa23524ccf984 I0504 20:41:10.927019 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-xlt5t/thanos-query_current.log with fingerprint=cfddbaaea1624549d72ff4f05a4e24fd240812b42c1a6cc971a4c3a38d74ef09 I0504 20:41:10.927025 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-xlt5t/kube-rbac-proxy-web_current.log with fingerprint=c584672128e5fbf6771a866e256180fd67e0018aadd8c6e86d1948d396931a7f I0504 20:41:10.927031 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-xlt5t/kube-rbac-proxy_current.log with fingerprint=7c768d6eb7d9006646b5ea7eff0402135a0dffe07f9a172092b867815f6ba67b I0504 20:41:10.927035 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-xlt5t/prom-label-proxy_current.log with fingerprint=809dabe6a82846e72046470712c1c1e2619af48255d6550c97b4e01ee8b013f7 I0504 20:41:10.927041 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-xlt5t/kube-rbac-proxy-rules_current.log with fingerprint=9f9aed6aa86d905434f165bfe43f797059c124f409464b48ea5aa0bdd755e206 I0504 20:41:10.927049 1 recorder.go:75] Recording config/pod/openshift-monitoring/logs/thanos-querier-57d4668466-xlt5t/kube-rbac-proxy-metrics_current.log with fingerprint=4bb6bebd81c8236539263ced0ccf648504df5ed585010ee6cf5e6922d5af9bd3 I0504 20:41:10.927057 1 recorder.go:75] Recording config/pod/openshift-network-console/logs/networking-console-plugin-7f449dfbfd-b6t2k/networking-console-plugin_current.log with fingerprint=28c947f31634204707ea5caffc3a6b7357138e01f05822b74c935ea453fbb355 I0504 20:41:10.927063 1 recorder.go:75] Recording config/pod/openshift-network-console/logs/networking-console-plugin-7f449dfbfd-zrqkw/networking-console-plugin_current.log with fingerprint=8d0bd0ef6d1ecb2c81d46656d106cea3210560c8ba3668b27700858c8fb79b78 I0504 20:41:10.927208 1 recorder.go:75] Recording config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m with fingerprint=1a5f79d65b66d8df089efd659c0d73dc7a91e3bc4162d5ef630294818e08782b E0504 20:41:10.927225 1 gather.go:161] error recording gatherer "clusterconfig" function "operators_pods_and_events" result "config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m.json" because of the error: the record with the same name "config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m.json" was already recorded and had the fingerprint "1a5f79d65b66d8df089efd659c0d73dc7a91e3bc4162d5ef630294818e08782b", overwriting with the record having fingerprint "1a5f79d65b66d8df089efd659c0d73dc7a91e3bc4162d5ef630294818e08782b" W0504 20:41:10.927238 1 gather.go:155] issue recording gatherer "clusterconfig" function "operators_pods_and_events" result "config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m.json" because of the warning: warning: the record with the same fingerprint "1a5f79d65b66d8df089efd659c0d73dc7a91e3bc4162d5ef630294818e08782b" was already recorded at path "config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m.json", recording another one with a different path "config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m.json" I0504 20:41:10.927249 1 recorder.go:75] Recording config/pod/openshift-console-operator/logs/console-operator-66dbbb6468-5pf5m/console-operator_previous.log with fingerprint=4820e44122b1023aa92a0ae380ba5c29359c6dfe1be916971e67c820d0351ad7 I0504 20:41:10.927498 1 recorder.go:75] Recording config/pod/openshift-console-operator/logs/console-operator-66dbbb6468-5pf5m/console-operator_current.log with fingerprint=8ebd945c23d4a1c8b21001c83b8da4880c414dd35381fb9cbf44961bfaa3afda I0504 20:41:10.927513 1 recorder.go:75] Recording config/pod/openshift-console/logs/console-5c56547d87-mztrn/console_current.log with fingerprint=8111d89f8b25453c801aa5c10c1b8ac2c47b2ce288f546870036b8183cea4a3f I0504 20:41:10.927526 1 recorder.go:75] Recording config/pod/openshift-console/logs/console-5c56547d87-z9tb6/console_current.log with fingerprint=ee47788b1f2559d8cc60e3a05bca9608eb1c7ae76fb4a8222e2e8a782cc6e563 I0504 20:41:10.927542 1 recorder.go:75] Recording config/pod/openshift-console/logs/console-746d468875-zzrsl/console_current.log with fingerprint=dfab81c9e9775bee296f2d8e387749feb14af094c2823721a692062051884b89 I0504 20:41:10.927571 1 recorder.go:75] Recording config/pod/openshift-console/logs/downloads-5c54f4944f-7x4tz/download-server_current.log with fingerprint=03c5b176a1326fe2b46a3ef437e1af50214414fcc9bbdfd4d44e9be33e87526b I0504 20:41:10.927600 1 recorder.go:75] Recording config/pod/openshift-console/logs/downloads-5c54f4944f-jcjp6/download-server_current.log with fingerprint=90d0294eab2c10d01cfe7190cd811704b1b651ff35e2fde4477946bfa588687e I0504 20:41:10.927607 1 gather.go:177] gatherer "clusterconfig" function "operators_pods_and_events" took 14.464926678s to process 75 records E0504 20:41:10.927671 1 periodic.go:254] "Unhandled Error" err="clusterconfig failed after 14.471s with: function \"tsdb_status\" failed with an error, function \"metrics\" failed with an error, function \"machines\" failed with an error, function \"pod_network_connectivity_checks\" failed with an error, function \"machine_configs\" failed with an error, function \"machine_healthchecks\" failed with an error, function \"support_secret\" failed with an error, function \"config_maps\" failed with an error, function \"ingress_certificates\" failed with an error, function \"dvo_metrics\" failed with an error, unable to record function \"operators_pods_and_events\" record \"config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m.json\"" I0504 20:41:10.928778 1 controllerstatus.go:89] name=periodic-clusterconfig healthy=false reason=PeriodicGatherFailed message=Source clusterconfig could not be retrieved: function "tsdb_status" failed with an error, function "metrics" failed with an error, function "machines" failed with an error, function "pod_network_connectivity_checks" failed with an error, function "machine_configs" failed with an error, function "machine_healthchecks" failed with an error, function "support_secret" failed with an error, function "config_maps" failed with an error, function "ingress_certificates" failed with an error, function "dvo_metrics" failed with an error, unable to record function "operators_pods_and_events" record "config/pod/openshift-console-operator/console-operator-66dbbb6468-5pf5m.json" I0504 20:41:10.928791 1 periodic.go:216] Running workloads gatherer I0504 20:41:10.928807 1 tasks_processing.go:45] number of workers: 2 I0504 20:41:10.928813 1 tasks_processing.go:69] worker 1 listening for tasks. I0504 20:41:10.928818 1 tasks_processing.go:71] worker 1 working on workload_info task. I0504 20:41:10.928824 1 tasks_processing.go:69] worker 0 listening for tasks. I0504 20:41:10.928837 1 tasks_processing.go:71] worker 0 working on helmchart_info task. I0504 20:41:10.935501 1 gather_workloads_runtime_infos.go:63] Gathering workload runtime info for node ip-10-0-0-208.ec2.internal... I0504 20:41:10.935515 1 gather_workloads_runtime_infos.go:63] Gathering workload runtime info for node ip-10-0-2-58.ec2.internal... I0504 20:41:10.935503 1 gather_workloads_runtime_infos.go:63] Gathering workload runtime info for node ip-10-0-1-101.ec2.internal... I0504 20:41:10.964361 1 tasks_processing.go:74] worker 0 stopped. I0504 20:41:10.964378 1 gather.go:177] gatherer "workloads" function "helmchart_info" took 35.511316ms to process 0 records I0504 20:41:15.570991 1 gather_workloads_runtime_infos.go:84] Gathered workload runtime infos in 5s I0504 20:41:15.596029 1 gather_workloads_info.go:278] Loaded pods in 0s, will wait 24s for image data I0504 20:41:15.607264 1 gather_workloads_info.go:387] No image sha256:730d1b6988025bef0daa3a9a5d8467ec4a26b0382cc52f91c3375b4590d3518a (12ms) I0504 20:41:15.619616 1 gather_workloads_info.go:387] No image sha256:a043239802b3eb8b323d285193d2527fad0ecec98ca91d188a3472a2fac8ae04 (12ms) I0504 20:41:15.629811 1 gather_workloads_info.go:387] No image sha256:934d8e8c50f3c609b8eea80d1051111fe3d066fe8c65c79572072ae55fcb0a86 (10ms) I0504 20:41:15.638089 1 gather_workloads_info.go:387] No image sha256:3bba1358d4a0ae878ff491c0c2cbfffe60649e110b40342b878fe8fa332f8858 (8ms) I0504 20:41:15.647066 1 gather_workloads_info.go:387] No image sha256:5c6d21c3f97366bc7ab57031cc027b67405a684bd804ce364ed5998b0685eaca (9ms) I0504 20:41:15.655728 1 gather_workloads_info.go:398] Found image sha256:23204877f166455ce8f9de9b91a467eeda576f172d5f96124a68aa986c4775d6 (9ms) I0504 20:41:15.664471 1 gather_workloads_info.go:387] No image sha256:1eb6a6ee3c66ac15be6bc21bb6015c595c11c15b9dfd21bd9f5ce7f7b990ebaa (9ms) I0504 20:41:15.672927 1 gather_workloads_info.go:387] No image sha256:637d41f067a5239096fc22b135181cda5113da833f1370e7a73965e83792e93a (8ms) I0504 20:41:15.681886 1 gather_workloads_info.go:387] No image sha256:5ac9c549d65fc1d8bc900773bebee43e9192bcec1bb5fa46afb4597230c16ac7 (9ms) I0504 20:41:15.691241 1 gather_workloads_info.go:387] No image sha256:48e883a3932aea9457f55cd4628d21397d429d4828f42f95d1c903d9d5395bde (9ms) I0504 20:41:15.705049 1 gather_workloads_info.go:387] No image sha256:15928a5b803fa00781b408dbc4c3bfa7c5287b18584ba31f817def8a4430c725 (14ms) I0504 20:41:15.804306 1 gather_workloads_info.go:387] No image sha256:9ef6dc311b880a6442b7a3a19006ddd955c7da36850bf000f5c9b68fd0862dd4 (99ms) I0504 20:41:15.905169 1 gather_workloads_info.go:387] No image sha256:25c148fd380b1a9db3f6039d2e0eabc489a954921452391390ba9192b2325678 (101ms) I0504 20:41:16.005106 1 gather_workloads_info.go:387] No image sha256:875c77e5d144f03fb91d8cee0259f6966683ca88d1bf818dbf4652c16b70312c (100ms) I0504 20:41:16.104754 1 gather_workloads_info.go:387] No image sha256:3380fecebaeee139c1607fbcd9346cebc92806679568e2244a8905ef0f52aad2 (100ms) I0504 20:41:16.204707 1 gather_workloads_info.go:387] No image sha256:0acfe677ed58031f8381cb8a4f2d8917ec4cfa95485e2a46529dffae54c75ace (100ms) I0504 20:41:16.304427 1 gather_workloads_info.go:387] No image sha256:48f5559439bea030a6423652da448c25fff0c62ba293eb521843c1a8eb89ddb0 (100ms) I0504 20:41:16.405225 1 gather_workloads_info.go:387] No image sha256:55b1db6038c5beaed54c626e3343b7a8589cc0be8dc41d1a66b4deab766ff520 (101ms) I0504 20:41:16.507258 1 gather_workloads_info.go:398] Found image sha256:673ebc8cc22c56c8f410e011b2fa950d28cf7b6420e17fdb6580d6cb10523384 (102ms) I0504 20:41:16.605374 1 gather_workloads_info.go:387] No image sha256:0b2e6536a0ca3afc5e64a218087a6287c74b77a359075a385f1949ca5bd51b22 (98ms) I0504 20:41:16.704932 1 gather_workloads_info.go:387] No image sha256:15e1b5f763553e86d0230a751b0783e3b990438fd07cf96eed164bccb72ed564 (100ms) I0504 20:41:16.805170 1 gather_workloads_info.go:387] No image sha256:89277d8d4560d71db88c2dcc67c992a24544ca21810920b609c1d49d53b4a287 (100ms) I0504 20:41:16.905345 1 gather_workloads_info.go:387] No image sha256:5b6212b8f539f08e78417d8a4b7485ca0b4e7927cacd7b752742a28841bc8ccd (100ms) I0504 20:41:17.005243 1 gather_workloads_info.go:387] No image sha256:0a4dfb8d4c1b3849319d45b4c54dff26a7238a2c08fcaa121f93073e95ab12e8 (100ms) I0504 20:41:17.105596 1 gather_workloads_info.go:387] No image sha256:572ebf7c7fab1a518582f3ca46c866730423cce6e8a19f614dc15fd61bf0c65a (100ms) I0504 20:41:17.205797 1 gather_workloads_info.go:387] No image sha256:5940cc34b322325618929b1e0eb3864dae73d54c8f5b474e218dfb28123a61c7 (100ms) I0504 20:41:17.311245 1 gather_workloads_info.go:387] No image sha256:9d6d63dce784df5863e6a0ba4c11317a26c9d756d4e62d47759c97190a71cce5 (105ms) I0504 20:41:17.405003 1 gather_workloads_info.go:387] No image sha256:8f881f6f0498eaba860d711dc8e9dc066db8728f22abedb82e1dc9a8e1c30046 (94ms) I0504 20:41:17.505305 1 gather_workloads_info.go:387] No image sha256:52db05084f2a2e7d7bf4040e4865df88ec4074efbbe59a78d861be2c5ab2aa40 (100ms) I0504 20:41:17.605603 1 gather_workloads_info.go:387] No image sha256:5d6a2b8d2508c7da71192661e7c0dcf1335f48b6336a20189abffbe79eeff962 (100ms) I0504 20:41:17.705805 1 gather_workloads_info.go:387] No image sha256:bdeb1c455cffccdb6c90db238e9d371de06340b1d6105d47f66d87293f878ac1 (100ms) I0504 20:41:17.804836 1 gather_workloads_info.go:387] No image sha256:084aa9b0f8a6d478549dc384d4e66da13ee9b25cc98531da861cc19dee2a9e8f (99ms) I0504 20:41:17.905143 1 gather_workloads_info.go:387] No image sha256:ab60623bb32f7e75fca71ef65137731cae347a21c7a4091dfd583fa00732721c (100ms) I0504 20:41:18.006153 1 gather_workloads_info.go:387] No image sha256:d64bea34bf3e1bb0b3a701c3ff14e66665afc1b050f28124ad7e6888eaec3a81 (101ms) I0504 20:41:18.105643 1 gather_workloads_info.go:387] No image sha256:30e597ec5d6bb96ff70a4f8688c748b659cd4fd5d73d222e8701821d236795c5 (99ms) I0504 20:41:18.205597 1 gather_workloads_info.go:387] No image sha256:17aefdc2f4752c23dc8f0947bb01fc34635822eeb18f2d8a5eddf8211b84d450 (100ms) I0504 20:41:18.306143 1 gather_workloads_info.go:387] No image sha256:53ff8efdd8f563421ebc1859ab021a3bbc1306f82d4ff5d9c21bc60ec5f57898 (101ms) I0504 20:41:18.405284 1 gather_workloads_info.go:387] No image sha256:15677f0b70e6aa2dfaf088e45fc1a425c22bf6fda326b8116f87e88b6694dfab (99ms) I0504 20:41:18.505736 1 gather_workloads_info.go:387] No image sha256:5808401268394502d335281ea1a294b07210461b986b58f91d7d1f29c0029c6d (100ms) I0504 20:41:18.604589 1 gather_workloads_info.go:387] No image sha256:5aaea0419169e55832cc27acfe0fe3b9513a343d6bdf71d3da1575ed322245d0 (99ms) I0504 20:41:18.705859 1 gather_workloads_info.go:387] No image sha256:000105ef5150e7079b90a613fb9e6193e2a6ef9b1908d2dce44f2395d4fd070f (101ms) I0504 20:41:18.805152 1 gather_workloads_info.go:387] No image sha256:5ff204630794311b0b37fa7b197b933ad85d76a481bf7cdb3bcbada08f0cdcbf (99ms) I0504 20:41:18.905456 1 gather_workloads_info.go:387] No image sha256:08c5a78c8a5af04c549e2273aaf4bb452a75bf038d68aa9d01bb2aff66c30e90 (100ms) I0504 20:41:19.005695 1 gather_workloads_info.go:387] No image sha256:6814f5fced0ef219d06374011c68a11a3da788a764a00a69ade435466d9ee240 (100ms) I0504 20:41:19.105373 1 gather_workloads_info.go:387] No image sha256:6c6d412ccf0c3d6a14645a79df6c4658c0f8725d7954cb83c3d2f0116499345a (100ms) I0504 20:41:19.205912 1 gather_workloads_info.go:387] No image sha256:ae7d3453fd734ecc865e5f9bb16f29244ebffe6291b77e1d4e496f71eb053174 (101ms) I0504 20:41:19.205947 1 tasks_processing.go:74] worker 1 stopped. I0504 20:41:19.206319 1 recorder.go:75] Recording config/workload_info with fingerprint=6fdb359537f4f99c4e529bcdeee1ce57f0f592d2a56af96817dd7d551b496ee7 I0504 20:41:19.206337 1 gather.go:177] gatherer "workloads" function "workload_info" took 8.277122131s to process 1 records I0504 20:41:19.206356 1 periodic.go:263] Periodic gather workloads completed in 8.277s I0504 20:41:19.206365 1 controllerstatus.go:80] name=periodic-workloads healthy=true reason= message= I0504 20:41:19.206371 1 periodic.go:216] Running conditional gatherer I0504 20:41:19.212764 1 requests.go:294] Making HTTP GET request at: https://console.redhat.com/api/gathering/v2/4.21.9/gathering_rules I0504 20:41:19.358423 1 conditional_gatherer.go:338] updating alerts cache for conditional gatherer E0504 20:41:19.362296 1 conditional_gatherer.go:324] unable to update alerts cache: Get "https://prometheus-k8s.openshift-monitoring.svc:9091/api/v1/query?match%5B%5D=ALERTS%7Balertstate%3D%22firing%22%7D&query=ALERTS": dial tcp 172.30.161.151:9091: connect: connection refused I0504 20:41:19.362373 1 conditional_gatherer.go:384] updating version cache for conditional gatherer I0504 20:41:19.367818 1 conditional_gatherer.go:392] cluster version is '4.21.9' E0504 20:41:19.367837 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367844 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367848 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367851 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367854 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367858 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367860 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367863 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367866 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing E0504 20:41:19.367870 1 conditional_gatherer.go:209] error checking conditions for a gathering rule: alerts cache is missing I0504 20:41:19.367884 1 tasks_processing.go:45] number of workers: 3 I0504 20:41:19.367898 1 tasks_processing.go:69] worker 2 listening for tasks. I0504 20:41:19.367902 1 tasks_processing.go:71] worker 2 working on conditional_gatherer_rules task. I0504 20:41:19.367911 1 tasks_processing.go:69] worker 0 listening for tasks. I0504 20:41:19.367926 1 tasks_processing.go:69] worker 1 listening for tasks. I0504 20:41:19.367934 1 tasks_processing.go:71] worker 0 working on remote_configuration task. I0504 20:41:19.367936 1 tasks_processing.go:74] worker 1 stopped. I0504 20:41:19.367951 1 tasks_processing.go:71] worker 2 working on rapid_container_logs task. I0504 20:41:19.367988 1 recorder.go:75] Recording insights-operator/conditional-gatherer-rules with fingerprint=9fafdcbf373f1f367f1d69551ea59114592cce1681189be0f3dde774e0530661 I0504 20:41:19.368001 1 gather.go:177] gatherer "conditional" function "conditional_gatherer_rules" took 991ns to process 1 records I0504 20:41:19.368034 1 recorder.go:75] Recording insights-operator/remote-configuration with fingerprint=5f681cc1442447fd2c1534f876836dff54caf3c545b582a5a8aae894f9203c5d I0504 20:41:19.368043 1 gather.go:177] gatherer "conditional" function "remote_configuration" took 1.275µs to process 1 records I0504 20:41:19.368048 1 tasks_processing.go:74] worker 0 stopped. I0504 20:41:19.368248 1 tasks_processing.go:74] worker 2 stopped. I0504 20:41:19.368264 1 gather.go:177] gatherer "conditional" function "rapid_container_logs" took 275.509µs to process 0 records I0504 20:41:19.368287 1 controllerstatus.go:80] name=periodic-conditional healthy=true reason=Succeeded message= I0504 20:41:19.368310 1 recorder.go:75] Recording insights-operator/remote-configuration.json with fingerprint=363fa5c8fa2573b18bc685faeca9ca9f021a8f80760895b7701f54e3dcd5c85f I0504 20:41:19.394949 1 recorder.go:75] Recording insights-operator/gathers with fingerprint=e4ca52c58dffa4aea09e16361113810c96c096b4aab23c02ebcecd7efd3d3ab0 I0504 20:41:19.395089 1 diskrecorder.go:70] Writing 185 records to /var/lib/insights-operator/insights-2026-05-04-204119.tar.gz I0504 20:41:19.407574 1 diskrecorder.go:51] Wrote 185 records to disk in 12ms I0504 20:41:19.407613 1 periodic.go:285] Gathering cluster info every 2h0m0s I0504 20:41:19.407631 1 periodic.go:286] Configuration is dataReporting: interval: 2h0m0s, uploadEndpoint: https://console.redhat.com/api/ingress/v1/upload, storagePath: /var/lib/insights-operator, downloadEndpoint: https://console.redhat.com/api/insights-results-aggregator/v2/cluster/%s/reports, conditionalGathererEndpoint: https://console.redhat.com/api/gathering/v2/%s/gathering_rules, obfuscation: [] sca: disabled: false, endpoint: https://api.openshift.com/api/accounts_mgmt/v1/entitlement_certificates, interval: 8h0m0s alerting: disabled: false clusterTransfer: endpoint: https://api.openshift.com/api/accounts_mgmt/v1/cluster_transfers/, interval: 12h0m0s proxy: httpProxy: , httpsProxy: , noProxy: I0504 20:42:26.456880 1 diskrecorder.go:223] Found files to send: insights-2026-05-04-204119.tar.gz I0504 20:42:26.456911 1 insightsuploader.go:150] Checking archives to upload periodically every 15m13.830955885s I0504 20:42:26.456920 1 insightsuploader.go:165] Uploading latest report since 0001-01-01T00:00:00Z I0504 20:42:26.466716 1 requests.go:46] Uploading application/vnd.redhat.openshift.periodic to https://console.redhat.com/api/ingress/v1/upload I0504 20:42:26.756434 1 requests.go:87] Successfully reported id=2026-05-04T20:42:26Z x-rh-insights-request-id=0edb1546d6094507aa3ed69947085842, wrote=92998 I0504 20:42:26.756494 1 insightsuploader.go:187] Uploaded report successfully in 299.567363ms I0504 20:42:26.756516 1 controller.go:129] Initializing last reported time to 2026-05-04T20:42:26Z I0504 20:42:26.756558 1 insightsreport.go:304] Archive uploaded, starting pulling report... I0504 20:42:26.756568 1 insightsreport.go:215] Starting retrieving report from Smart Proxy I0504 20:42:26.756577 1 insightsreport.go:221] Initial delay for pulling: 1m0s I0504 20:42:26.762152 1 controller.go:554] The operator is healthy I0504 20:42:56.474730 1 controller.go:554] The operator is healthy I0504 20:43:28.801053 1 insightsreport.go:137] Pulling report from smart-proxy I0504 20:43:28.801082 1 insightsreport.go:149] Retrieving report I0504 20:43:28.807873 1 requests.go:110] Retrieving report for cluster: a3144f2c-a465-4047-b648-d84385256ec9 I0504 20:43:28.807887 1 requests.go:111] Endpoint: https://console.redhat.com/api/insights-results-aggregator/v2/cluster/a3144f2c-a465-4047-b648-d84385256ec9/reports I0504 20:43:28.810975 1 requests.go:121] Retrieving report from https://console.redhat.com/api/insights-results-aggregator/v2/cluster/a3144f2c-a465-4047-b648-d84385256ec9/reports I0504 20:43:29.138967 1 insightsreport.go:184] Report retrieved I0504 20:43:29.164468 1 insightsreport.go:239] Report retrieved correctly I0504 20:44:56.476432 1 controller.go:554] The operator is healthy I0504 20:45:56.447542 1 secretconfigobserver.go:136] Refreshing configuration from cluster pull secret I0504 20:45:56.454450 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 20:45:56.454471 1 secretconfigobserver.go:162] Refreshing configuration from cluster support secret I0504 20:45:56.458667 1 secretconfigobserver.go:119] support secret does not exist I0504 20:46:56.569915 1 controller.go:554] The operator is healthy I0504 20:48:56.475588 1 controller.go:554] The operator is healthy I0504 20:50:56.456527 1 configmapobserver.go:84] configmaps "insights-config" not found I0504 20:50:56.459593 1 secretconfigobserver.go:136] Refreshing configuration from cluster pull secret I0504 20:50:56.575828 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 20:50:56.575852 1 secretconfigobserver.go:162] Refreshing configuration from cluster support secret I0504 20:50:56.576009 1 controller.go:554] The operator is healthy I0504 20:50:56.619696 1 secretconfigobserver.go:119] support secret does not exist I0504 20:52:56.473208 1 controller.go:554] The operator is healthy I0504 20:54:56.473021 1 controller.go:554] The operator is healthy I0504 20:55:56.620230 1 secretconfigobserver.go:136] Refreshing configuration from cluster pull secret I0504 20:55:56.626939 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 20:55:56.626958 1 secretconfigobserver.go:162] Refreshing configuration from cluster support secret I0504 20:55:56.630973 1 secretconfigobserver.go:119] support secret does not exist I0504 20:56:56.472724 1 controller.go:554] The operator is healthy I0504 20:57:45.589021 1 insightsuploader.go:146] Nothing to report since 2026-05-04T20:42:26Z I0504 20:58:56.473427 1 controller.go:554] The operator is healthy I0504 21:00:56.436866 1 configmapobserver.go:84] configmaps "insights-config" not found I0504 21:00:56.440946 1 configmapobserver.go:84] configmaps "insights-config" not found I0504 21:00:56.472280 1 controller.go:554] The operator is healthy I0504 21:00:56.631547 1 secretconfigobserver.go:136] Refreshing configuration from cluster pull secret I0504 21:00:56.636106 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 21:00:56.636136 1 secretconfigobserver.go:162] Refreshing configuration from cluster support secret I0504 21:00:56.640072 1 secretconfigobserver.go:119] support secret does not exist I0504 21:02:56.473811 1 controller.go:554] The operator is healthy I0504 21:04:56.474159 1 controller.go:554] The operator is healthy I0504 21:05:56.641019 1 secretconfigobserver.go:136] Refreshing configuration from cluster pull secret I0504 21:05:56.646874 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 21:05:56.646893 1 secretconfigobserver.go:162] Refreshing configuration from cluster support secret I0504 21:05:56.651032 1 secretconfigobserver.go:119] support secret does not exist I0504 21:06:56.473998 1 controller.go:554] The operator is healthy I0504 21:08:56.473551 1 controller.go:554] The operator is healthy I0504 21:10:56.437214 1 configmapobserver.go:84] configmaps "insights-config" not found I0504 21:10:56.473086 1 controller.go:554] The operator is healthy I0504 21:10:56.651510 1 secretconfigobserver.go:136] Refreshing configuration from cluster pull secret I0504 21:10:56.656284 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 21:10:56.656309 1 secretconfigobserver.go:162] Refreshing configuration from cluster support secret I0504 21:10:56.660405 1 secretconfigobserver.go:119] support secret does not exist I0504 21:12:56.485109 1 controller.go:554] The operator is healthy I0504 21:13:53.920883 1 insightsuploader.go:146] Nothing to report since 2026-05-04T20:42:26Z I0504 21:14:56.473612 1 controller.go:554] The operator is healthy I0504 21:15:56.660937 1 secretconfigobserver.go:136] Refreshing configuration from cluster pull secret I0504 21:15:56.666013 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 21:15:56.666031 1 secretconfigobserver.go:162] Refreshing configuration from cluster support secret I0504 21:15:56.670265 1 secretconfigobserver.go:119] support secret does not exist I0504 21:16:56.473970 1 controller.go:554] The operator is healthy I0504 21:18:56.472795 1 controller.go:554] The operator is healthy I0504 21:20:56.437814 1 configmapobserver.go:84] configmaps "insights-config" not found I0504 21:20:56.472554 1 controller.go:554] The operator is healthy I0504 21:20:56.670965 1 secretconfigobserver.go:136] Refreshing configuration from cluster pull secret I0504 21:20:56.675735 1 secretconfigobserver.go:249] Found cloud.openshift.com token I0504 21:20:56.675752 1 secretconfigobserver.go:162] Refreshing configuration from cluster support secret I0504 21:20:56.679790 1 secretconfigobserver.go:119] support secret does not exist