W0420 11:41:07.694337 1 client_config.go:667] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work. I0420 11:41:07.694731 1 envvar.go:172] "Feature gate default state" feature="ClientsAllowCBOR" enabled=false I0420 11:41:07.694747 1 envvar.go:172] "Feature gate default state" feature="ClientsPreferCBOR" enabled=false I0420 11:41:07.694751 1 envvar.go:172] "Feature gate default state" feature="InformerResourceVersion" enabled=false I0420 11:41:07.694755 1 envvar.go:172] "Feature gate default state" feature="WatchListClient" enabled=false I0420 11:41:07.720336 1 simple_featuregate_reader.go:171] Starting feature-gate-detector I0420 11:41:07.723628 1 reflector.go:376] Caches populated for *v1.FeatureGate from github.com/openshift/client-go/config/informers/externalversions/factory.go:125 I0420 11:41:07.723847 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BuildCSIVolumes", "CPMSMachineNamePrefix", "ConsolePluginContentSecurityPolicy", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "ImageVolume", "IngressControllerLBSubnetsAWS", "KMSv1", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "MetricsCollectionProfiles", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "PinnedImages", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SetEIPForNLBIngressController", "SigstoreImageVerification", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiDisk", "VSphereMultiNetworks"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSServiceLBNetworkSecurityGroup", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureMultiDisk", "BootImageSkewEnforcement", "BootcNodeManagement", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "Example2", "ExternalSnapshotMetadata", "GCPClusterHostedDNS", "GCPClusterHostedDNSInstall", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "ImageModeStatusReporting", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryptionProvider", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "ManagedBootImagesAzure", "ManagedBootImagesvSphere", "MaxUnavailableStatefulSet", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "MultiDiskSetup", "MutatingAdmissionPolicy", "NewOLM", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterOperations", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PreconfiguredUDNAddresses", "SELinuxMount", "ShortCertRotation", "SignatureStores", "SigstoreImageVerificationPKI", "TranslateStreamCloseWebsocketRequests", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VolumeAttributesClass", "VolumeGroupSnapshot"}} I0420 11:41:07.724112 1 reflector.go:376] Caches populated for *v1.ClusterVersion from github.com/openshift/client-go/config/informers/externalversions/factory.go:125 I0420 11:41:07.724830 1 dynamic_serving_content.go:116] "Loaded a new cert/key pair" name="serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key" I0420 11:41:07.734132 1 reflector.go:376] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/informers/factory.go:160 I0420 11:41:07.734235 1 reflector.go:376] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.735471 1 reflector.go:376] Caches populated for *v1.Secret from k8s.io/client-go/informers/factory.go:160 I0420 11:41:07.735513 1 reflector.go:376] Caches populated for *v1.Secret from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.736253 1 reflector.go:376] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.736394 1 reflector.go:376] Caches populated for *v1.APIServer from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.736486 1 reflector.go:376] Caches populated for *v1.Console from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.736746 1 reflector.go:376] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.738296 1 reflector.go:376] Caches populated for *v1.ClusterOperator from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.738378 1 operator.go:647] Triggering an update due to a change in ClusterOperator.config.openshift.io/ingress I0420 11:41:07.738621 1 reflector.go:376] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:566 I0420 11:41:07.739066 1 reflector.go:376] Caches populated for *v1.ClusterVersion from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.739428 1 reflector.go:376] Caches populated for *v1.Infrastructure from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.740213 1 operator.go:647] Triggering an update due to a change in *v1.APIServer/cluster I0420 11:41:07.743509 1 operator.go:647] Triggering an update due to a change in *v1.Infrastructure/cluster I0420 11:41:07.743729 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:07.774920 1 reflector.go:376] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:569 I0420 11:41:07.827791 1 operator.go:598] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. I0420 11:41:07.827821 1 rule_controller.go:113] Starting alerting rules controller I0420 11:41:07.827837 1 base_controller.go:76] Waiting for caches to sync for OpenShiftMonitoringTelemeterClientCertRequester W0420 11:41:07.827850 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:41:07.827853 1 base_controller.go:82] Caches are synced for OpenShiftMonitoringTelemeterClientCertRequester I0420 11:41:07.827850 1 relabel_controller.go:117] Starting alert relabel config controller I0420 11:41:07.827843 1 shared_informer.go:313] Waiting for caches to sync for AlertingRule controller I0420 11:41:07.827874 1 shared_informer.go:313] Waiting for caches to sync for AlertRelabelConfig controller I0420 11:41:07.827860 1 base_controller.go:119] Starting #1 worker of OpenShiftMonitoringTelemeterClientCertRequester controller ... I0420 11:41:07.827877 1 base_controller.go:76] Waiting for caches to sync for OpenShiftMonitoringMetricsServerClientCertRequester I0420 11:41:07.827898 1 base_controller.go:82] Caches are synced for OpenShiftMonitoringMetricsServerClientCertRequester I0420 11:41:07.827906 1 base_controller.go:119] Starting #1 worker of OpenShiftMonitoringMetricsServerClientCertRequester controller ... I0420 11:41:07.827820 1 base_controller.go:76] Waiting for caches to sync for OpenShiftMonitoringClientCertRequester I0420 11:41:07.827939 1 base_controller.go:82] Caches are synced for OpenShiftMonitoringClientCertRequester I0420 11:41:07.827954 1 base_controller.go:119] Starting #1 worker of OpenShiftMonitoringClientCertRequester controller ... I0420 11:41:07.831299 1 reflector.go:376] Caches populated for *v1.Secret from github.com/openshift/cluster-monitoring-operator/pkg/alert/relabel_controller.go:122 I0420 11:41:07.831951 1 reflector.go:376] Caches populated for *v1.PrometheusRule from github.com/openshift/cluster-monitoring-operator/pkg/alert/rule_controller.go:117 I0420 11:41:07.832315 1 reflector.go:376] Caches populated for *v1.AlertRelabelConfig from github.com/openshift/cluster-monitoring-operator/pkg/alert/relabel_controller.go:121 I0420 11:41:07.832468 1 reflector.go:376] Caches populated for *v1.AlertingRule from github.com/openshift/cluster-monitoring-operator/pkg/alert/rule_controller.go:118 I0420 11:41:07.833512 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-server-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.833524 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.833608 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NoValidCertificateFound' No valid client certificate for OpenShiftMonitoringMetricsServerClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates I0420 11:41:07.833620 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NoValidCertificateFound' No valid client certificate for OpenShiftMonitoringTelemeterClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates I0420 11:41:07.834484 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.834607 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'NoValidCertificateFound' No valid client certificate for OpenShiftMonitoringClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates I0420 11:41:07.844146 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'CSRCreated' A csr "system:openshift:openshift-monitoring-qgpqr" is created for OpenShiftMonitoringTelemeterClientCertRequester I0420 11:41:07.847231 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'CSRCreated' A csr "system:openshift:openshift-monitoring-w99kz" is created for OpenShiftMonitoringMetricsServerClientCertRequester I0420 11:41:07.851368 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'CSRCreated' A csr "system:openshift:openshift-monitoring-4zcb6" is created for OpenShiftMonitoringClientCertRequester I0420 11:41:07.859644 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.859877 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-server-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.860742 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.864102 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.866538 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-server-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.866662 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.872722 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.875635 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.875736 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-server-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.877608 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.887640 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:41:07.891775 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-server-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.891779 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.892601 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.899968 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.900675 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.900698 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-server-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.910851 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.911204 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-server-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.911314 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.922528 1 tasks.go:49] processing task group 1 of 3 I0420 11:41:07.922560 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:41:07.922591 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:41:07.927083 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.928157 1 shared_informer.go:320] Caches are synced for AlertRelabelConfig controller I0420 11:41:07.928163 1 shared_informer.go:320] Caches are synced for AlertingRule controller I0420 11:41:07.938978 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-server-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.939795 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for metrics-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.939803 1 annotations.go:31] Updating "openshift.io/owning-component" annotation for federate-client-certs/openshift-monitoring, diff: string( - "", + "Monitoring", ) I0420 11:41:07.957980 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/prometheus-operator -n openshift-monitoring because it was missing I0420 11:41:07.973790 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/alert-relabel-configs -n openshift-monitoring because it was missing I0420 11:41:07.976760 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClientCertificateCreated' A new client certificate for OpenShiftMonitoringClientCertRequester is available I0420 11:41:07.979386 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClientCertificateCreated' A new client certificate for OpenShiftMonitoringMetricsServerClientCertRequester is available I0420 11:41:07.979468 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClientCertificateCreated' A new client certificate for OpenShiftMonitoringTelemeterClientCertRequester is available I0420 11:41:07.980106 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/metrics-server-client-certs I0420 11:41:07.980470 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/federate-client-certs I0420 11:41:07.980525 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/metrics-client-certs I0420 11:41:08.032305 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/metrics-client-ca -n openshift-monitoring because it was missing I0420 11:41:08.067369 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:41:08.137016 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/prometheus-operator because it was missing I0420 11:41:08.213277 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/prometheus-operator because it was missing I0420 11:41:08.233457 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing I0420 11:41:08.296762 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing I0420 11:41:08.325326 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'PodDisruptionBudgetCreated' Created PodDisruptionBudget.policy/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing I0420 11:41:08.555315 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController I0420 11:41:08.556830 1 maxinflight.go:139] "Initialized nonMutatingChan" len=400 I0420 11:41:08.556846 1 maxinflight.go:145] "Initialized mutatingChan" len=200 I0420 11:41:08.556857 1 maxinflight.go:116] "Set denominator for readonly requests" limit=400 I0420 11:41:08.556861 1 maxinflight.go:120] "Set denominator for mutating requests" limit=200 I0420 11:41:08.559672 1 secure_serving.go:57] Forcing use of http/1.1 only I0420 11:41:08.559688 1 genericapiserver.go:535] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete I0420 11:41:08.564309 1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController I0420 11:41:08.564323 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController I0420 11:41:08.564322 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" I0420 11:41:08.564330 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" I0420 11:41:08.564336 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0420 11:41:08.564342 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0420 11:41:08.564593 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key" I0420 11:41:08.564692 1 tlsconfig.go:203] "Loaded serving cert" certName="serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key" certDetail="\"*.cluster-monitoring-operator.openshift-monitoring.svc\" [serving] validServingFor=[*.cluster-monitoring-operator.openshift-monitoring.svc,*.cluster-monitoring-operator.openshift-monitoring.svc.cluster.local,cluster-monitoring-operator.openshift-monitoring.svc,cluster-monitoring-operator.openshift-monitoring.svc.cluster.local] issuer=\"openshift-service-serving-signer@1776685239\" (2026-04-20 11:40:50 +0000 UTC to 2028-04-19 11:40:51 +0000 UTC (now=2026-04-20 11:41:08.564666251 +0000 UTC))" I0420 11:41:08.564865 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1776685268\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1776685267\" (2026-04-20 10:41:07 +0000 UTC to 2027-04-20 10:41:07 +0000 UTC (now=2026-04-20 11:41:08.564850544 +0000 UTC))" I0420 11:41:08.564885 1 secure_serving.go:213] Serving securely on [::]:8443 I0420 11:41:08.564906 1 genericapiserver.go:685] [graceful-termination] waiting for shutdown to be initiated I0420 11:41:08.564921 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" I0420 11:41:08.567720 1 reflector.go:376] Caches populated for *v1.ConfigMap from k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209 I0420 11:41:08.567735 1 reflector.go:376] Caches populated for *v1.ConfigMap from k8s.io/apiserver/pkg/authentication/request/headerrequest/requestheader_controller.go:183 I0420 11:41:08.567724 1 reflector.go:376] Caches populated for *v1.ConfigMap from k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209 I0420 11:41:08.664616 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file I0420 11:41:08.664634 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file I0420 11:41:08.664669 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController I0420 11:41:08.664800 1 tlsconfig.go:181] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2026-04-20 11:28:55 +0000 UTC to 2036-04-17 11:28:55 +0000 UTC (now=2026-04-20 11:41:08.664776204 +0000 UTC))" I0420 11:41:08.665000 1 tlsconfig.go:203] "Loaded serving cert" certName="serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key" certDetail="\"*.cluster-monitoring-operator.openshift-monitoring.svc\" [serving] validServingFor=[*.cluster-monitoring-operator.openshift-monitoring.svc,*.cluster-monitoring-operator.openshift-monitoring.svc.cluster.local,cluster-monitoring-operator.openshift-monitoring.svc,cluster-monitoring-operator.openshift-monitoring.svc.cluster.local] issuer=\"openshift-service-serving-signer@1776685239\" (2026-04-20 11:40:50 +0000 UTC to 2028-04-19 11:40:51 +0000 UTC (now=2026-04-20 11:41:08.664984333 +0000 UTC))" I0420 11:41:08.665163 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1776685268\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1776685267\" (2026-04-20 10:41:07 +0000 UTC to 2027-04-20 10:41:07 +0000 UTC (now=2026-04-20 11:41:08.665151031 +0000 UTC))" I0420 11:41:08.665263 1 tlsconfig.go:181] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-control-plane-signer\" [] issuer=\"\" (2026-04-20 11:28:58 +0000 UTC to 2036-04-17 11:28:58 +0000 UTC (now=2026-04-20 11:41:08.665253284 +0000 UTC))" I0420 11:41:08.665291 1 tlsconfig.go:181] "Loaded client CA" index=1 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-apiserver-to-kubelet-signer\" [] issuer=\"\" (2026-04-20 11:29:03 +0000 UTC to 2036-04-17 11:29:03 +0000 UTC (now=2026-04-20 11:41:08.665282427 +0000 UTC))" I0420 11:41:08.665309 1 tlsconfig.go:181] "Loaded client CA" index=2 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"admin-kubeconfig-signer\" [] issuer=\"\" (2026-04-20 11:29:06 +0000 UTC to 2036-04-17 11:29:06 +0000 UTC (now=2026-04-20 11:41:08.665300438 +0000 UTC))" I0420 11:41:08.665324 1 tlsconfig.go:181] "Loaded client CA" index=3 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"hcco-signer\" [] issuer=\"\" (2026-04-20 11:29:09 +0000 UTC to 2036-04-17 11:29:09 +0000 UTC (now=2026-04-20 11:41:08.665316853 +0000 UTC))" I0420 11:41:08.665340 1 tlsconfig.go:181] "Loaded client CA" index=4 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-csr-signer\" [] issuer=\"\" (2026-04-20 11:29:12 +0000 UTC to 2036-04-17 11:29:12 +0000 UTC (now=2026-04-20 11:41:08.665332868 +0000 UTC))" I0420 11:41:08.665358 1 tlsconfig.go:181] "Loaded client CA" index=5 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"ocm-production-2pprssu02fngo91nqt802gka1s7j10lq-kx-ad82caabda_customer-system-admin-signer@1776684690\" [] issuer=\"\" (2026-04-20 11:31:30 +0000 UTC to 2026-04-27 11:31:31 +0000 UTC (now=2026-04-20 11:41:08.665348419 +0000 UTC))" I0420 11:41:08.665374 1 tlsconfig.go:181] "Loaded client CA" index=6 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"ocm-production-2pprssu02fngo91nqt802gka1s7j10lq-kx-ad82caabda_sre-system-admin-signer@1776684690\" [] issuer=\"\" (2026-04-20 11:31:33 +0000 UTC to 2026-04-27 11:31:34 +0000 UTC (now=2026-04-20 11:41:08.665365516 +0000 UTC))" I0420 11:41:08.665388 1 tlsconfig.go:181] "Loaded client CA" index=7 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2026-04-20 11:28:55 +0000 UTC to 2036-04-17 11:28:55 +0000 UTC (now=2026-04-20 11:41:08.665381492 +0000 UTC))" I0420 11:41:08.665550 1 tlsconfig.go:203] "Loaded serving cert" certName="serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key" certDetail="\"*.cluster-monitoring-operator.openshift-monitoring.svc\" [serving] validServingFor=[*.cluster-monitoring-operator.openshift-monitoring.svc,*.cluster-monitoring-operator.openshift-monitoring.svc.cluster.local,cluster-monitoring-operator.openshift-monitoring.svc,cluster-monitoring-operator.openshift-monitoring.svc.cluster.local] issuer=\"openshift-service-serving-signer@1776685239\" (2026-04-20 11:40:50 +0000 UTC to 2028-04-19 11:40:51 +0000 UTC (now=2026-04-20 11:41:08.665538004 +0000 UTC))" I0420 11:41:08.665715 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1776685268\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1776685267\" (2026-04-20 10:41:07 +0000 UTC to 2027-04-20 10:41:07 +0000 UTC (now=2026-04-20 11:41:08.665704292 +0000 UTC))" I0420 11:41:11.387566 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ValidatingWebhookConfigurationCreated' Created ValidatingWebhookConfiguration.admissionregistration.k8s.io/prometheusrules.openshift.io because it was missing I0420 11:41:11.402978 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ValidatingWebhookConfigurationCreated' Created ValidatingWebhookConfiguration.admissionregistration.k8s.io/alertmanagerconfigs.openshift.io because it was missing I0420 11:41:11.425318 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/prometheus-operator -n openshift-monitoring because it was missing I0420 11:41:11.579765 1 operator.go:647] Triggering an update due to a change in *v1.Console/cluster I0420 11:41:15.560824 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:41:15.560861 1 tasks.go:49] processing task group 2 of 3 I0420 11:41:15.560900 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:41:15.560905 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:41:15.560919 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:41:15.560944 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:41:15.560939 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:41:15.560953 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:41:15.560964 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:41:15.560966 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:41:15.560970 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:41:15.560987 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:41:15.560980 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:41:15.560989 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:41:15.561003 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:41:15.561012 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:41:15.561015 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:41:15.579244 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/kube-state-metrics -n openshift-monitoring because it was missing I0420 11:41:15.585265 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/openshift-state-metrics -n openshift-monitoring because it was missing I0420 11:41:15.608282 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/thanos-querier -n openshift-monitoring because it was missing I0420 11:41:15.612725 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/telemeter-client -n openshift-monitoring because it was missing I0420 11:41:15.612755 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/node-exporter -n openshift-monitoring because it was missing I0420 11:41:15.634987 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/kubelet-serving-ca-bundle -n openshift-monitoring because it was missing I0420 11:41:15.652101 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:41:15.662280 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:41:15.685753 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:41:15.710034 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/monitoring-rules-edit because it was missing I0420 11:41:15.716119 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/prometheus-k8s because it was missing I0420 11:41:15.760792 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/kube-state-metrics because it was missing I0420 11:41:15.765318 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/telemeter-client because it was missing I0420 11:41:15.765339 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/openshift-state-metrics because it was missing I0420 11:41:15.765345 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/node-exporter because it was missing I0420 11:41:15.828826 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/monitoring-rules-view because it was missing I0420 11:41:15.832183 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/prometheus-k8s because it was missing I0420 11:41:15.846403 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/metrics-server -n openshift-monitoring because it was missing I0420 11:41:15.902901 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/openshift-state-metrics because it was missing I0420 11:41:15.906738 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/telemeter-client because it was missing I0420 11:41:15.913632 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/node-exporter because it was missing I0420 11:41:15.923003 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/kube-state-metrics because it was missing W0420 11:41:15.923356 1 tasks.go:73] task 8 of 15: UpdatingTelemeterClient failed: reconciling Telemeter client cluster monitoring view ClusterRoleBinding failed: clusterrolebindings.rbac.authorization.k8s.io "cluster-monitoring-view" not found I0420 11:41:15.923373 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Warning' reason: 'ClusterRoleBindingCreateFailed' Failed to create ClusterRoleBinding.rbac.authorization.k8s.io/telemeter-client-view: clusterrolebindings.rbac.authorization.k8s.io "cluster-monitoring-view" not found I0420 11:41:15.923695 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/openshift-state-metrics -n openshift-monitoring because it was missing I0420 11:41:15.952705 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/node-exporter -n openshift-monitoring because it was missing I0420 11:41:15.971331 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/kube-state-metrics -n openshift-monitoring because it was missing I0420 11:41:15.979444 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/node-exporter-accelerators-collector-config -n openshift-monitoring because it was missing I0420 11:41:15.990519 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/system:metrics-server because it was missing I0420 11:41:15.997319 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/kube-state-metrics-custom-resource-state-configmap -n openshift-monitoring because it was missing I0420 11:41:15.999180 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/monitoring-edit because it was missing I0420 11:41:16.173304 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/system:metrics-server because it was missing I0420 11:41:16.186721 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/alert-routing-edit because it was missing I0420 11:41:16.260193 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/metrics-server:system:auth-delegator because it was missing I0420 11:41:16.265167 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/cluster-monitoring-view because it was missing I0420 11:41:16.338243 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/system:aggregated-metrics-reader because it was missing I0420 11:41:16.353229 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleBindingCreated' Created RoleBinding.rbac.authorization.k8s.io/metrics-server-auth-reader -n kube-system because it was missing I0420 11:41:16.405430 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/metrics-server -n openshift-monitoring because it was missing I0420 11:41:16.461591 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/pod-metrics-reader because it was missing I0420 11:41:16.544502 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/user-workload-monitoring-config-edit -n openshift-user-workload-monitoring because it was missing I0420 11:41:16.651334 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-api-reader -n openshift-user-workload-monitoring because it was missing I0420 11:41:16.651386 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'PodDisruptionBudgetCreated' Created PodDisruptionBudget.policy/alertmanager-main -n openshift-monitoring because it was missing I0420 11:41:16.727947 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/thanos-querier-kube-rbac-proxy-web -n openshift-monitoring because it was missing I0420 11:41:16.741183 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/thanos-querier -n openshift-monitoring because it was missing I0420 11:41:16.758712 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/alertmanager-main because it was missing I0420 11:41:16.772485 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-api-writer -n openshift-user-workload-monitoring because it was missing I0420 11:41:16.867914 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleCreated' Created ClusterRole.rbac.authorization.k8s.io/thanos-querier because it was missing I0420 11:41:16.867944 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/alertmanager-main because it was missing I0420 11:41:16.883387 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-edit -n openshift-monitoring because it was missing I0420 11:41:16.888042 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/alertmanager-main -n openshift-monitoring because it was missing I0420 11:41:16.952277 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/alertmanager-main -n openshift-monitoring because it was missing I0420 11:41:16.975739 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/alertmanager-trusted-ca-bundle I0420 11:41:16.975960 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/alertmanager-trusted-ca-bundle -n openshift-monitoring because it was missing I0420 11:41:17.012199 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/thanos-querier because it was missing I0420 11:41:17.035184 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-view -n openshift-monitoring because it was missing I0420 11:41:17.129719 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/cluster-monitoring-metrics-api -n openshift-monitoring because it was missing I0420 11:41:17.295841 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/alertmanager-trusted-ca-bundle I0420 11:41:17.727755 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/kube-rbac-proxy -n openshift-monitoring because it was missing I0420 11:41:17.744558 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/prometheus-k8s-kube-rbac-proxy-web -n openshift-monitoring because it was missing I0420 11:41:17.762245 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/prometheus-k8s -n openshift-monitoring because it was missing I0420 11:41:17.783611 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:41:17.783634 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/grpc-tls -n openshift-monitoring because it was missing I0420 11:41:17.784457 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/grpc-tls I0420 11:41:17.866248 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleBindingCreated' Created RoleBinding.rbac.authorization.k8s.io/alertmanager-prometheusk8s -n openshift-monitoring because it was missing I0420 11:41:17.962617 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/prometheus-k8s-config -n openshift-monitoring because it was missing I0420 11:41:18.035842 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/thanos-querier-grpc-tls-1hbu3nvtgg9jl -n openshift-monitoring because it was missing I0420 11:41:18.066158 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/prometheus-k8s -n default because it was missing I0420 11:41:18.265507 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/prometheus-k8s -n kube-system because it was missing I0420 11:41:18.377294 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-monitoring because it was missing I0420 11:41:18.467538 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleCreated' Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-user-workload-monitoring because it was missing I0420 11:41:18.571136 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleBindingCreated' Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n default because it was missing I0420 11:41:18.678927 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleBindingCreated' Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n kube-system because it was missing I0420 11:41:18.696844 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:41:18.713161 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:41:18.780457 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleBindingCreated' Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-monitoring because it was missing I0420 11:41:18.880661 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleBindingCreated' Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-user-workload-monitoring because it was missing I0420 11:41:18.987561 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RoleBindingCreated' Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s-config -n openshift-monitoring because it was missing I0420 11:41:19.105322 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/prometheus-k8s -n openshift-monitoring because it was missing I0420 11:41:19.123538 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:41:19.126134 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:41:19.126812 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/prometheus-k8s-thanos-sidecar -n openshift-monitoring because it was missing I0420 11:41:19.442401 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/metrics-server-audit-profiles -n openshift-monitoring because it was missing I0420 11:41:20.028457 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:41:20.468709 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/metrics-server-6e6promqaqbtn -n openshift-monitoring because it was missing I0420 11:41:20.645140 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceAccountCreated' Created ServiceAccount/monitoring-plugin -n openshift-monitoring because it was missing I0420 11:41:20.690603 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/monitoring-plugin -n openshift-monitoring because it was missing I0420 11:41:22.162854 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/prometheus-k8s-grpc-tls-a1tuo93nkhsfe -n openshift-monitoring because it was missing I0420 11:41:22.179549 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'PodDisruptionBudgetCreated' Created PodDisruptionBudget.policy/prometheus-k8s -n openshift-monitoring because it was missing I0420 11:41:22.206214 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/prometheus-trusted-ca-bundle -n openshift-monitoring because it was missing I0420 11:41:22.217776 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/prometheus-k8s-additional-alertmanager-configs -n openshift-monitoring because it was missing I0420 11:41:24.775558 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:41:24.775605 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'PodDisruptionBudgetCreated' Created PodDisruptionBudget.policy/monitoring-plugin -n openshift-monitoring because it was missing I0420 11:41:29.140779 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'PodDisruptionBudgetCreated' Created PodDisruptionBudget.policy/thanos-querier-pdb -n openshift-monitoring because it was missing I0420 11:41:29.181965 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:41:37.396312 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:37.632535 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:37.654115 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:45.544133 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'PodDisruptionBudgetCreated' Created PodDisruptionBudget.policy/metrics-server -n openshift-monitoring because it was missing I0420 11:41:45.557736 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:41:45.557806 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'APIServiceCreated' Created APIService.apiregistration.k8s.io/v1beta1.metrics.k8s.io because it was missing I0420 11:41:53.404618 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:53.438238 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:53.458541 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:53.498769 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:53.532843 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:54.404683 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:41:57.075372 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:42:10.417275 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:42:11.423282 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:42:32.282062 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:42:32.282116 1 operator.go:875] 1 reconciliation(s) failed, 3 more attempt(s) will be made before reporting failures. E0420 11:42:32.282125 1 operator.go:706] Syncing "openshift-monitoring/cluster-monitoring-config" failed E0420 11:42:32.282144 1 operator.go:707] "Unhandled Error" err="sync \"openshift-monitoring/cluster-monitoring-config\" failed: cluster monitoring update failed (reason: UpdatingTelemeterClientFailed)" logger="UnhandledError" W0420 11:42:32.283255 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:42:32.314324 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:42:32.319676 1 tasks.go:49] processing task group 1 of 3 I0420 11:42:32.319701 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:42:32.319724 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:42:32.334274 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:42:33.384363 1 admissionregistration.go:144] ValidatingWebhookConfiguration "/prometheusrules.openshift.io" changes: {"webhooks":[{"admissionReviewVersions":["v1"],"clientConfig":{"caBundle":"LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURVVENDQWptZ0F3SUJBZ0lJQnBHSFc1cEFMRkl3RFFZSktvWklodmNOQVFFTEJRQXdOakUwTURJR0ExVUUKQXd3cmIzQmxibk5vYVdaMExYTmxjblpwWTJVdGMyVnlkbWx1WnkxemFXZHVaWEpBTVRjM05qWTROVEl6T1RBZQpGdzB5TmpBME1qQXhNVFF3TXpoYUZ3MHlPREEyTVRneE1UUXdNemxhTURZeE5EQXlCZ05WQkFNTUsyOXdaVzV6CmFHbG1kQzF6WlhKMmFXTmxMWE5sY25acGJtY3RjMmxuYm1WeVFERTNOelkyT0RVeU16a3dnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRRE9rQytCanBxcXdMSk16Q3Q2SWdOSXBWWG1zRHNBMjVBQwpFU1g4RVRqTk5YNjhnS2ZLVG9BOEF4VnZxOUxGWW10T2k5MVdCQVFwcitWRDN3VXVaRkFyMWx3MVBIWEVNQXJYCkxwSFlZTkd2NTUvV1pGTURWeUQ4K3lGT3pPNFBDRVhmMkRJTk44UzArNFlPdENGVS92NUQ3dlQ3cTJjZGEyL1oKcnNsb0hhbXBlcko5czhJTm0zbTdUZU9MK1VmRk9aU3lJa2l0OG9KN1dVWWIwTkpNeXNZWkpqRE5qalFRcENBeQpjYWxqZmgyMllaSkdMOHpqZU9NMEJoYlpBczJndVFRZHJrRTdUcmhzRHdsOWVpYjdjSjU4UWVsYy9lOWZvOG9KCldKcmlMcW40TkNXbmg4V2dTZStGSjcwc1JwK2doREVxWVUyc0s2Zk5GWUVXamFFWStlU3hBZ01CQUFHall6QmgKTUE0R0ExVWREd0VCL3dRRUF3SUNwREFQQmdOVkhSTUJBZjhFQlRBREFRSC9NQjBHQTFVZERnUVdCQlFGZWtZTwpNalBKSmlybHpVd1NkUkFvbXdnR3NEQWZCZ05WSFNNRUdEQVdnQlFGZWtZT01qUEpKaXJselV3U2RSQW9td2dHCnNEQU5CZ2txaGtpRzl3MEJBUXNGQUFPQ0FRRUFhMGo3Ry9CTmI5TXk1ckNWY0IyK1YvSWtVOEErL3l4NCtXRnAKbTl2Q25uSUxOSUYwQXlSRi8vNGpySndUQStMbXUxc1NMRmdNejhaeGhCNnZabjN6VFdjbFBZaXdUb2ZkeUFEcgpqZVkrQlZTaHVOS3pLVkRsQ1lJbWsvM3I0SnNQWVY3MXJ4MjZCbjNseGN2Umt0djlGbm8zNmFsUHdGLzFYbGNYCk5waWVUVGFpa08xK2FTNEN5MStYbTl0cjE1TWI0eEw2UlhJQ1F2RFRKVVpGKys0T0dCbkhvZ1pZZmpCRWREd3kKTGZSclBIWlp6OVFmTHpmanV3cjlQR3F4QlE5Y0Vzbzl4bDdINnkyeWFTTVpRMmJYMEN1c3BtbzI0cXN1M21tWgp2Skx3S1VxM1hFd3licG5xRW1pY2FMdGVwZmtUd1BYNkNack80c0hMNnpPTDl5N3I3QT09Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K","service":{"name":"prometheus-operator-admission-webhook","namespace":"openshift-monitoring","path":"/admission-prometheusrules/validate","port":8443}},"failurePolicy":"Ignore","name":"prometheusrules.openshift.io","rules":[{"apiGroups":["monitoring.coreos.com"],"apiVersions":["v1"],"operations":["CREATE","UPDATE"],"resources":["prometheusrules"],"scope":"Namespaced"}],"sideEffects":"None","timeoutSeconds":5}]} I0420 11:42:33.391428 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ValidatingWebhookConfigurationUpdated' Updated ValidatingWebhookConfiguration.admissionregistration.k8s.io/prometheusrules.openshift.io because it changed I0420 11:42:33.398981 1 admissionregistration.go:144] ValidatingWebhookConfiguration "/alertmanagerconfigs.openshift.io" changes: {"webhooks":[{"admissionReviewVersions":["v1"],"clientConfig":{"caBundle":"LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURVVENDQWptZ0F3SUJBZ0lJQnBHSFc1cEFMRkl3RFFZSktvWklodmNOQVFFTEJRQXdOakUwTURJR0ExVUUKQXd3cmIzQmxibk5vYVdaMExYTmxjblpwWTJVdGMyVnlkbWx1WnkxemFXZHVaWEpBTVRjM05qWTROVEl6T1RBZQpGdzB5TmpBME1qQXhNVFF3TXpoYUZ3MHlPREEyTVRneE1UUXdNemxhTURZeE5EQXlCZ05WQkFNTUsyOXdaVzV6CmFHbG1kQzF6WlhKMmFXTmxMWE5sY25acGJtY3RjMmxuYm1WeVFERTNOelkyT0RVeU16a3dnZ0VpTUEwR0NTcUcKU0liM0RRRUJBUVVBQTRJQkR3QXdnZ0VLQW9JQkFRRE9rQytCanBxcXdMSk16Q3Q2SWdOSXBWWG1zRHNBMjVBQwpFU1g4RVRqTk5YNjhnS2ZLVG9BOEF4VnZxOUxGWW10T2k5MVdCQVFwcitWRDN3VXVaRkFyMWx3MVBIWEVNQXJYCkxwSFlZTkd2NTUvV1pGTURWeUQ4K3lGT3pPNFBDRVhmMkRJTk44UzArNFlPdENGVS92NUQ3dlQ3cTJjZGEyL1oKcnNsb0hhbXBlcko5czhJTm0zbTdUZU9MK1VmRk9aU3lJa2l0OG9KN1dVWWIwTkpNeXNZWkpqRE5qalFRcENBeQpjYWxqZmgyMllaSkdMOHpqZU9NMEJoYlpBczJndVFRZHJrRTdUcmhzRHdsOWVpYjdjSjU4UWVsYy9lOWZvOG9KCldKcmlMcW40TkNXbmg4V2dTZStGSjcwc1JwK2doREVxWVUyc0s2Zk5GWUVXamFFWStlU3hBZ01CQUFHall6QmgKTUE0R0ExVWREd0VCL3dRRUF3SUNwREFQQmdOVkhSTUJBZjhFQlRBREFRSC9NQjBHQTFVZERnUVdCQlFGZWtZTwpNalBKSmlybHpVd1NkUkFvbXdnR3NEQWZCZ05WSFNNRUdEQVdnQlFGZWtZT01qUEpKaXJselV3U2RSQW9td2dHCnNEQU5CZ2txaGtpRzl3MEJBUXNGQUFPQ0FRRUFhMGo3Ry9CTmI5TXk1ckNWY0IyK1YvSWtVOEErL3l4NCtXRnAKbTl2Q25uSUxOSUYwQXlSRi8vNGpySndUQStMbXUxc1NMRmdNejhaeGhCNnZabjN6VFdjbFBZaXdUb2ZkeUFEcgpqZVkrQlZTaHVOS3pLVkRsQ1lJbWsvM3I0SnNQWVY3MXJ4MjZCbjNseGN2Umt0djlGbm8zNmFsUHdGLzFYbGNYCk5waWVUVGFpa08xK2FTNEN5MStYbTl0cjE1TWI0eEw2UlhJQ1F2RFRKVVpGKys0T0dCbkhvZ1pZZmpCRWREd3kKTGZSclBIWlp6OVFmTHpmanV3cjlQR3F4QlE5Y0Vzbzl4bDdINnkyeWFTTVpRMmJYMEN1c3BtbzI0cXN1M21tWgp2Skx3S1VxM1hFd3licG5xRW1pY2FMdGVwZmtUd1BYNkNack80c0hMNnpPTDl5N3I3QT09Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K","service":{"name":"prometheus-operator-admission-webhook","namespace":"openshift-monitoring","path":"/admission-alertmanagerconfigs/validate","port":8443}},"failurePolicy":"Ignore","name":"alertmanagerconfigs.openshift.io","rules":[{"apiGroups":["monitoring.coreos.com"],"apiVersions":["v1alpha1"],"operations":["CREATE","UPDATE"],"resources":["alertmanagerconfigs"],"scope":"Namespaced"}],"sideEffects":"None","timeoutSeconds":5}]} I0420 11:42:33.408833 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ValidatingWebhookConfigurationUpdated' Updated ValidatingWebhookConfiguration.admissionregistration.k8s.io/alertmanagerconfigs.openshift.io because it changed I0420 11:42:35.532713 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:42:35.532738 1 tasks.go:49] processing task group 2 of 3 I0420 11:42:35.532762 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:42:35.532775 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:42:35.532786 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:42:35.532807 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:42:35.532823 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:42:35.532836 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:42:35.532890 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:42:35.532907 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:42:35.532907 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:42:35.532939 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:42:35.532944 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:42:35.532984 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:42:35.532776 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:42:35.533054 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:42:35.533105 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:42:35.592357 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:42:35.613379 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:42:35.622916 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:42:35.624944 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ClusterRoleBindingCreated' Created ClusterRoleBinding.rbac.authorization.k8s.io/telemeter-client-view because it was missing I0420 11:42:35.640763 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ServiceCreated' Created Service/telemeter-client -n openshift-monitoring because it was missing I0420 11:42:35.662165 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/telemeter-client -n openshift-monitoring because it was missing I0420 11:42:35.675726 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/telemeter-client-kube-rbac-proxy-config -n openshift-monitoring because it was missing I0420 11:42:35.686157 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:42:35.687555 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/telemeter-trusted-ca-bundle I0420 11:42:35.746464 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/telemeter-trusted-ca-bundle I0420 11:42:36.606533 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:42:36.640554 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:42:36.644274 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:42:36.656781 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:42:36.677289 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:42:38.697207 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:42:40.676436 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:42:40.735989 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/telemeter-trusted-ca-bundle-8i12ta5c71j38 -n openshift-monitoring because it was missing I0420 11:42:41.596913 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:42:43.847507 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 11:43:46.716795 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:44:50.937174 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:44:50.937207 1 tasks.go:49] processing task group 3 of 3 I0420 11:44:50.937222 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 11:44:50.996221 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 11:44:50.996313 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"97014fda-e4b1-47d5-8cb3-77a4c59706bd", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/monitoring-shared-config -n openshift-config-managed because it was missing I0420 11:44:51.002078 1 operator.go:851] Updating ClusterOperator status to done. W0420 11:44:51.045082 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:44:51.095469 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:44:51.101908 1 tasks.go:49] processing task group 1 of 3 I0420 11:44:51.101930 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:44:51.101950 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:44:51.116534 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:44:54.258728 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:44:54.258752 1 tasks.go:49] processing task group 2 of 3 I0420 11:44:54.258779 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:44:54.258789 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:44:54.258803 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:44:54.258813 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:44:54.258830 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:44:54.258850 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:44:54.258868 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:44:54.258879 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:44:54.258895 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:44:54.258917 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:44:54.258938 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:44:54.258964 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:44:54.258983 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:44:54.259001 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:44:54.258794 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:44:54.306165 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:44:54.331503 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:44:54.347673 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:44:54.393062 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:44:55.331493 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:44:55.355353 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:44:55.355419 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:44:55.372169 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:44:55.396783 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:44:57.426635 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:44:59.389128 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:45:00.317698 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:45:00.401518 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 11:45:05.409945 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:45:05.680161 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:45:09.702223 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:45:09.702251 1 tasks.go:49] processing task group 3 of 3 I0420 11:45:09.702269 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 11:45:09.749792 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 11:45:09.755521 1 operator.go:851] Updating ClusterOperator status to done. W0420 11:45:09.765463 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:45:09.795383 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:45:09.800962 1 tasks.go:49] processing task group 1 of 3 I0420 11:45:09.800985 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:45:09.800996 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:45:09.817011 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:45:12.943834 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:45:12.943858 1 tasks.go:49] processing task group 2 of 3 I0420 11:45:12.943882 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:45:12.943890 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:45:12.943898 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:45:12.943929 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:45:12.943943 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:45:12.943966 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:45:12.943977 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:45:12.943997 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:45:12.944027 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:45:12.944051 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:45:12.944056 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:45:12.944078 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:45:12.944093 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:45:12.944118 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:45:12.944027 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:45:12.996555 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:45:13.018881 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:45:13.024160 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:45:13.087369 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:45:14.015084 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:45:14.038356 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:45:14.049113 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:45:14.068630 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:45:14.070328 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:45:16.111262 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:45:18.084101 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:45:19.013141 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:45:19.095392 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 11:45:24.102688 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:45:28.282084 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:45:28.282108 1 tasks.go:49] processing task group 3 of 3 I0420 11:45:28.282125 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 11:45:28.324612 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 11:45:28.329760 1 operator.go:851] Updating ClusterOperator status to done. I0420 11:46:07.828888 1 operator.go:609] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. W0420 11:46:07.828923 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:46:07.866533 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:46:07.871379 1 tasks.go:49] processing task group 1 of 3 I0420 11:46:07.871408 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:46:07.871425 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:46:07.883810 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:46:11.035085 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:46:11.035115 1 tasks.go:49] processing task group 2 of 3 I0420 11:46:11.035143 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:46:11.035153 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:46:11.035176 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:46:11.035203 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:46:11.035224 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:46:11.035251 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:46:11.035264 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:46:11.035290 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:46:11.035312 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:46:11.035332 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:46:11.035339 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:46:11.035364 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:46:11.035378 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:46:11.035408 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:46:11.035427 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:46:11.103658 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:46:11.119949 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:46:11.123737 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:46:11.185007 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:46:12.114249 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:46:12.149519 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:46:12.152085 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:46:12.164103 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:46:12.186470 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:46:14.220391 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:46:16.180326 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:46:17.103813 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:46:17.200931 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 11:46:22.204985 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:46:26.398513 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:46:26.398542 1 tasks.go:49] processing task group 3 of 3 I0420 11:46:26.398559 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 11:46:26.438634 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 11:46:26.443812 1 operator.go:851] Updating ClusterOperator status to done. I0420 11:47:35.769114 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version W0420 11:47:35.769145 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:47:35.803447 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:47:35.809766 1 tasks.go:49] processing task group 1 of 3 I0420 11:47:35.809801 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:47:35.809813 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:47:35.827483 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:47:38.967936 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:47:38.967962 1 tasks.go:49] processing task group 2 of 3 I0420 11:47:38.967989 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:47:38.968001 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:47:38.968011 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:47:38.968031 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:47:38.968048 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:47:38.968076 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:47:38.968088 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:47:38.968097 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:47:38.968115 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:47:38.968139 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:47:38.968164 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:47:38.968184 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:47:38.968196 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:47:38.968222 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:47:38.967999 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:47:39.022902 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:47:39.041495 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:47:39.049200 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:47:39.108483 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:47:40.041997 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:47:40.065885 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:47:40.076514 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:47:40.099180 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:47:40.109836 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:47:42.124471 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:47:44.116497 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:47:45.030723 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:47:45.116811 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 11:47:50.135611 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:47:54.290709 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:47:54.290738 1 tasks.go:49] processing task group 3 of 3 I0420 11:47:54.290752 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 11:47:54.338423 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 11:47:54.344684 1 operator.go:851] Updating ClusterOperator status to done. I0420 11:51:07.828814 1 operator.go:609] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. W0420 11:51:07.828880 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:51:07.871656 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:51:07.876655 1 tasks.go:49] processing task group 1 of 3 I0420 11:51:07.876687 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:51:07.876703 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:51:07.893208 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:51:11.923785 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:51:11.923817 1 tasks.go:49] processing task group 2 of 3 I0420 11:51:11.923847 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:51:11.923853 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:51:11.923871 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:51:11.923890 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:51:11.923913 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:51:11.923940 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:51:11.923960 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:51:11.923981 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:51:11.923996 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:51:11.924034 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:51:11.924038 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:51:11.924086 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:51:11.924082 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:51:11.924092 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:51:11.924012 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:51:12.373115 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:51:12.537807 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:51:12.636702 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:51:13.105152 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:51:13.586676 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:51:13.674402 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:51:13.729277 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:51:13.789303 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:51:14.110906 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:51:15.910609 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:51:17.711459 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:51:18.697192 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:51:18.852779 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 11:51:23.828741 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:51:28.216126 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:51:28.216155 1 tasks.go:49] processing task group 3 of 3 I0420 11:51:28.216176 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 11:51:28.273709 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 11:51:28.280322 1 operator.go:851] Updating ClusterOperator status to done. I0420 11:56:07.736563 1 operator.go:647] Triggering an update due to a change in *v1.APIServer/cluster W0420 11:56:07.736627 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:56:07.736626 1 operator.go:647] Triggering an update due to a change in *v1.Console/cluster I0420 11:56:07.736817 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/grpc-tls I0420 11:56:07.736838 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/metrics-client-certs I0420 11:56:07.736910 1 operator.go:678] Triggering an update due to ConfigMap or Secret: kube-system/extension-apiserver-authentication I0420 11:56:07.736950 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/metrics-server-client-certs I0420 11:56:07.736963 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/federate-client-certs I0420 11:56:07.739189 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/telemeter-trusted-ca-bundle I0420 11:56:07.739211 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/alertmanager-trusted-ca-bundle I0420 11:56:07.739214 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 11:56:07.740193 1 operator.go:647] Triggering an update due to a change in *v1.Infrastructure/cluster I0420 11:56:07.769433 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:56:07.775210 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-config-managed/kubelet-serving-ca I0420 11:56:07.776108 1 tasks.go:49] processing task group 1 of 3 I0420 11:56:07.776136 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:56:07.776148 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:56:07.790969 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:56:07.828527 1 operator.go:609] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. I0420 11:56:10.957085 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:56:10.957109 1 tasks.go:49] processing task group 2 of 3 I0420 11:56:10.957139 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:56:10.957152 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:56:10.957162 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:56:10.957174 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:56:10.957185 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:56:10.957208 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:56:10.957239 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:56:10.957259 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:56:10.957288 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:56:10.957303 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:56:10.957330 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:56:10.957329 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:56:10.957350 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:56:10.957394 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:56:10.957376 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:56:11.045340 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:56:11.090122 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:56:11.102826 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:56:11.159262 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:56:12.087943 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:56:12.089914 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:56:12.092053 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:56:12.095504 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:56:12.139243 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:56:14.168520 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:56:16.138939 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:56:17.040756 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:56:17.151529 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 11:56:22.143823 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:56:26.331357 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:56:26.331386 1 tasks.go:49] processing task group 3 of 3 I0420 11:56:26.331395 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 11:56:26.372843 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 11:56:26.378866 1 operator.go:851] Updating ClusterOperator status to done. W0420 11:56:26.389055 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 11:56:26.417643 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 11:56:26.422625 1 tasks.go:49] processing task group 1 of 3 I0420 11:56:26.422651 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 11:56:26.422664 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:56:26.436219 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 11:56:29.564097 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 11:56:29.564122 1 tasks.go:49] processing task group 2 of 3 I0420 11:56:29.564142 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:56:29.564148 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 11:56:29.564158 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:56:29.564182 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 11:56:29.564202 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 11:56:29.564217 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:56:29.564239 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 11:56:29.564262 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 11:56:29.564291 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 11:56:29.564308 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:56:29.564337 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:56:29.564344 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 11:56:29.564361 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 11:56:29.564359 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 11:56:29.564407 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:56:29.611952 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 11:56:29.636324 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 11:56:29.645871 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 11:56:29.699550 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 11:56:30.637919 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 11:56:30.663523 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 11:56:30.665174 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 11:56:30.686423 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 11:56:30.694228 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 11:56:32.733208 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 11:56:34.693769 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 11:56:35.625642 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 11:56:35.711849 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 11:56:40.726084 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 11:56:44.902112 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 11:56:44.902141 1 tasks.go:49] processing task group 3 of 3 I0420 11:56:44.902166 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 11:56:44.944390 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 11:56:44.950408 1 operator.go:851] Updating ClusterOperator status to done. I0420 12:01:07.828197 1 operator.go:609] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. W0420 12:01:07.828261 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 12:01:07.868825 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 12:01:07.873914 1 tasks.go:49] processing task group 1 of 3 I0420 12:01:07.873952 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 12:01:07.873965 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 12:01:07.890228 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 12:01:11.044714 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 12:01:11.044739 1 tasks.go:49] processing task group 2 of 3 I0420 12:01:11.044763 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 12:01:11.044766 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 12:01:11.044781 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 12:01:11.044811 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 12:01:11.044832 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 12:01:11.044857 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 12:01:11.044868 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 12:01:11.044887 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 12:01:11.044913 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 12:01:11.044951 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 12:01:11.044946 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 12:01:11.044964 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 12:01:11.044982 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 12:01:11.045008 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 12:01:11.045023 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 12:01:11.101285 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 12:01:11.130134 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 12:01:11.132721 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 12:01:11.194915 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 12:01:12.132685 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 12:01:12.164361 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 12:01:12.164822 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 12:01:12.174988 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 12:01:12.194823 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 12:01:14.222073 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 12:01:16.183981 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 12:01:17.105289 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 12:01:17.200077 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 12:01:22.234702 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 12:01:26.390924 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 12:01:26.390950 1 tasks.go:49] processing task group 3 of 3 I0420 12:01:26.390968 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 12:01:26.431741 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 12:01:26.438153 1 operator.go:851] Updating ClusterOperator status to done. I0420 12:06:07.828617 1 operator.go:609] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. W0420 12:06:07.828679 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 12:06:07.861764 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 12:06:07.866776 1 tasks.go:49] processing task group 1 of 3 I0420 12:06:07.866805 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 12:06:07.866817 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 12:06:07.879212 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 12:06:11.034088 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 12:06:11.034115 1 tasks.go:49] processing task group 2 of 3 I0420 12:06:11.034145 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 12:06:11.034153 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 12:06:11.034170 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 12:06:11.034191 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 12:06:11.034225 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 12:06:11.034241 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 12:06:11.034251 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 12:06:11.034277 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 12:06:11.034289 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 12:06:11.034320 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 12:06:11.034347 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 12:06:11.034374 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 12:06:11.034396 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 12:06:11.034423 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 12:06:11.034442 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 12:06:11.100417 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 12:06:11.119394 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 12:06:11.125095 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 12:06:11.189833 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 12:06:12.120861 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 12:06:12.144731 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 12:06:12.146381 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 12:06:12.165090 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 12:06:12.186211 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 12:06:14.212413 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 12:06:16.183215 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 12:06:17.097163 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 12:06:17.196244 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 12:06:22.214157 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 12:06:26.381081 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 12:06:26.381109 1 tasks.go:49] processing task group 3 of 3 I0420 12:06:26.381128 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 12:06:26.422141 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 12:06:26.428020 1 operator.go:851] Updating ClusterOperator status to done. I0420 12:11:07.736850 1 operator.go:647] Triggering an update due to a change in *v1.APIServer/cluster I0420 12:11:07.736903 1 operator.go:647] Triggering an update due to a change in *v1.Console/cluster I0420 12:11:07.736976 1 operator.go:678] Triggering an update due to ConfigMap or Secret: kube-system/extension-apiserver-authentication W0420 12:11:07.736977 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 12:11:07.737079 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/metrics-server-client-certs I0420 12:11:07.737118 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/federate-client-certs I0420 12:11:07.737229 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/grpc-tls I0420 12:11:07.737289 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/metrics-client-certs I0420 12:11:07.739343 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version I0420 12:11:07.739430 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/alertmanager-trusted-ca-bundle I0420 12:11:07.739471 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-monitoring/telemeter-trusted-ca-bundle I0420 12:11:07.740651 1 operator.go:647] Triggering an update due to a change in *v1.Infrastructure/cluster I0420 12:11:07.770197 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 12:11:07.775413 1 operator.go:678] Triggering an update due to ConfigMap or Secret: openshift-config-managed/kubelet-serving-ca I0420 12:11:07.777444 1 tasks.go:49] processing task group 1 of 3 I0420 12:11:07.777481 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 12:11:07.777504 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 12:11:07.792958 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 12:11:07.828141 1 operator.go:609] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. I0420 12:11:10.940016 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 12:11:10.940043 1 tasks.go:49] processing task group 2 of 3 I0420 12:11:10.940080 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 12:11:10.940081 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 12:11:10.940089 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 12:11:10.940104 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 12:11:10.940125 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 12:11:10.940144 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 12:11:10.940158 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 12:11:10.940175 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 12:11:10.940201 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 12:11:10.940289 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 12:11:10.940303 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 12:11:10.940321 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 12:11:10.940403 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 12:11:10.940420 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 12:11:10.940295 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 12:11:11.021233 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 12:11:11.045165 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 12:11:11.061468 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 12:11:11.117621 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 12:11:12.038200 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 12:11:12.068112 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 12:11:12.069274 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 12:11:12.072518 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 12:11:12.126376 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 12:11:14.135753 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 12:11:16.105639 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 12:11:17.008193 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 12:11:17.124124 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 12:11:22.113372 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 12:11:26.301168 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 12:11:26.301200 1 tasks.go:49] processing task group 3 of 3 I0420 12:11:26.301224 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 12:11:26.343991 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 12:11:26.349753 1 operator.go:851] Updating ClusterOperator status to done. W0420 12:11:26.360381 1 operator.go:974] No Cluster Monitoring ConfigMap was found. Using defaults. I0420 12:11:26.390655 1 operator.go:830] Updating ClusterOperator status to InProgress. I0420 12:11:26.395800 1 tasks.go:49] processing task group 1 of 3 I0420 12:11:26.395843 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator I0420 12:11:26.395848 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA I0420 12:11:26.411439 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA I0420 12:11:29.563988 1 tasks.go:76] ran task 2 of 2: UpdatingPrometheusOperator I0420 12:11:29.564012 1 tasks.go:49] processing task group 2 of 3 I0420 12:11:29.564041 1 tasks.go:70] running task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 12:11:29.564045 1 tasks.go:70] running task 3 of 15: UpdatingAlertmanager I0420 12:11:29.564051 1 tasks.go:70] running task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 12:11:29.564071 1 tasks.go:70] running task 9 of 15: UpdatingThanosQuerier I0420 12:11:29.564102 1 tasks.go:70] running task 6 of 15: UpdatingOpenshiftStateMetrics I0420 12:11:29.564115 1 tasks.go:70] running task 10 of 15: UpdatingControlPlaneComponents I0420 12:11:29.564125 1 tasks.go:70] running task 4 of 15: UpdatingNodeExporter I0420 12:11:29.564133 1 tasks.go:70] running task 13 of 15: UpdatingUserWorkloadPrometheus I0420 12:11:29.564147 1 tasks.go:70] running task 11 of 15: UpdatingConsolePluginComponents I0420 12:11:29.564164 1 tasks.go:70] running task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 12:11:29.564183 1 tasks.go:70] running task 7 of 15: UpdatingMetricsServer I0420 12:11:29.564080 1 tasks.go:70] running task 2 of 15: UpdatingPrometheus I0420 12:11:29.564194 1 tasks.go:70] running task 8 of 15: UpdatingTelemeterClient I0420 12:11:29.564100 1 tasks.go:70] running task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 12:11:29.564220 1 tasks.go:70] running task 5 of 15: UpdatingKubeStateMetrics I0420 12:11:29.642886 1 tasks.go:76] ran task 12 of 15: UpdatingUserWorkloadPrometheusOperator I0420 12:11:29.670950 1 tasks.go:76] ran task 14 of 15: UpdatingUserWorkloadAlertmanager I0420 12:11:29.693146 1 tasks.go:76] ran task 10 of 15: UpdatingControlPlaneComponents I0420 12:11:29.771047 1 tasks.go:76] ran task 1 of 15: UpdatingClusterMonitoringOperatorDeps I0420 12:11:30.656313 1 tasks.go:76] ran task 6 of 15: UpdatingOpenshiftStateMetrics I0420 12:11:30.681366 1 tasks.go:76] ran task 15 of 15: UpdatingUserWorkloadThanosRuler I0420 12:11:30.693607 1 tasks.go:76] ran task 5 of 15: UpdatingKubeStateMetrics I0420 12:11:30.707944 1 tasks.go:76] ran task 13 of 15: UpdatingUserWorkloadPrometheus I0420 12:11:30.727590 1 tasks.go:76] ran task 4 of 15: UpdatingNodeExporter I0420 12:11:32.747431 1 tasks.go:76] ran task 9 of 15: UpdatingThanosQuerier I0420 12:11:34.716595 1 tasks.go:76] ran task 7 of 15: UpdatingMetricsServer I0420 12:11:35.626143 1 tasks.go:76] ran task 11 of 15: UpdatingConsolePluginComponents I0420 12:11:35.746380 1 tasks.go:76] ran task 8 of 15: UpdatingTelemeterClient I0420 12:11:40.738425 1 tasks.go:76] ran task 3 of 15: UpdatingAlertmanager I0420 12:11:44.925416 1 tasks.go:76] ran task 2 of 15: UpdatingPrometheus I0420 12:11:44.925455 1 tasks.go:49] processing task group 3 of 3 I0420 12:11:44.925473 1 tasks.go:70] running task 1 of 1: UpdatingConfigurationSharing I0420 12:11:44.966711 1 tasks.go:76] ran task 1 of 1: UpdatingConfigurationSharing I0420 12:11:44.972640 1 operator.go:851] Updating ClusterOperator status to done.