ts=2026-05-06T20:13:28.996605977Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:218 msg="Starting Prometheus Operator" version="(version=0.85.0, branch=HEAD, revision=2740c4be8)" build_context="(go=go1.24.6 (Red Hat 1.24.6-1.el9_6) X:strictfipsruntime, platform=linux/amd64, user=root, date=20251208-13:40:21, tags=strictfipsruntime)" feature_gates="PrometheusAgentDaemonSet=false,PrometheusShardRetentionPolicy=false,PrometheusTopologySharding=false,StatusForConfigurationResources=false" ts=2026-05-06T20:13:28.99664622Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:219 msg="Operator's configuration" watch_referenced_objects_in_all_namespaces=true controller_id=openshift-monitoring/prometheus-operator enable_config_reloader_probes=false ts=2026-05-06T20:13:28.996888715Z level=info caller=/go/src/github.com/coreos/prometheus-operator/internal/goruntime/cpu.go:27 msg="Leaving GOMAXPROCS=16: CPU quota undefined" ts=2026-05-06T20:13:28.996942129Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:233 msg="Namespaces filtering configuration " config="{allow_list=\"\",deny_list=\"\",prometheus_allow_list=\"openshift-monitoring\",alertmanager_allow_list=\"openshift-monitoring\",alertmanagerconfig_allow_list=\"\",thanosruler_allow_list=\"openshift-monitoring\"}" ts=2026-05-06T20:13:29.012767622Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:274 msg="connection established" kubernetes_version=1.33.6 ts=2026-05-06T20:13:29.027296675Z level=warn caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:81 msg="resource \"scrapeconfigs\" (group: \"monitoring.coreos.com/v1alpha1\") not installed in the cluster" ts=2026-05-06T20:13:29.027328402Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:359 msg="Kubernetes API capabilities" endpointslices=true ts=2026-05-06T20:13:29.058276238Z level=warn caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:81 msg="resource \"prometheusagents\" (group: \"monitoring.coreos.com/v1alpha1\") not installed in the cluster" ts=2026-05-06T20:13:29.117665643Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-05-06T20:13:29.117635458Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/kubelet/controller.go:214 msg="Starting controller" component=kubelet_endpoints kubelet_object=kube-system/kubelet ts=2026-05-06T20:13:29.117727473Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/server/server.go:293 msg="starting insecure server" address=127.0.0.1:8080 ts=2026-05-06T20:13:29.117674802Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-05-06T20:13:29.118135953Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.164135881Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:13:29.172870539Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:13:29.218016472Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-05-06T20:13:29.218083451Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-05-06T20:13:29.218095189Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-05-06T20:13:29.218109646Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-05-06T20:13:29.218118589Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-05-06T20:13:29.218133405Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-05-06T20:13:29.21814326Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-05-06T20:13:29.21815873Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-05-06T20:13:29.218170541Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-05-06T20:13:29.218187721Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-05-06T20:13:29.218197564Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-05-06T20:13:29.218210728Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/thanos/operator.go:318 msg="successfully synced all caches" component=thanos-controller ts=2026-05-06T20:13:29.21830735Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-05-06T20:13:29.218360362Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-05-06T20:13:29.218370603Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-05-06T20:13:29.218385187Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-05-06T20:13:29.218394833Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-05-06T20:13:29.218408128Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-05-06T20:13:29.218417556Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-05-06T20:13:29.218434836Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-05-06T20:13:29.218445148Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-05-06T20:13:29.218458419Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-05-06T20:13:29.218470376Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-05-06T20:13:29.218479585Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-05-06T20:13:29.218485376Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-05-06T20:13:29.218492677Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:367 msg="successfully synced all caches" component=alertmanager-controller ts=2026-05-06T20:13:29.219663315Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.21970107Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.219731195Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.219743538Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.219750816Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.219759982Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.219765703Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.219773771Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.219779091Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.219788105Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.219814978Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.219824886Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.2198305Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.219837961Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.219843741Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.219851613Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.219857031Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.21986489Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-05-06T20:13:29.219880565Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-05-06T20:13:29.219891885Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:439 msg="successfully synced all caches" component=prometheus-controller ts=2026-05-06T20:13:32.900382952Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:32.992904401Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:722 msg="StatefulSet not found" component=alertmanager-controller key=openshift-monitoring/alertmanager-main ts=2026-05-06T20:13:33.007541327Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:33.007628601Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:722 msg="StatefulSet not found" component=alertmanager-controller key=openshift-monitoring/alertmanager-main ts=2026-05-06T20:13:33.162332112Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:33.357563419Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:33.614966544Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:33.72678759Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:34.308166962Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:34.929688435Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:35.10823452Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:35.223120473Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:35.750077839Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:36.371542837Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:36.486358042Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:36.777352463Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:37.155180601Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:37.255449474Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:37.983133679Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:38.06496906Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:38.08164949Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:38.229601937Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:38.322575272Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:38.366155159Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:38.473272983Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:38.777911445Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:38.996776656Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:41.295254041Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:41.430120743Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:41.516822797Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:41.55960598Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:47.096809833Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:47.241020583Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:47.296782104Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:47.417025505Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:47.498209061Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:47.55194711Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:54.759816048Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:13:54.884190986Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:54.968308387Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:13:55.009367533Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:14:02.825237261Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:14:02.980413868Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:14:03.054283061Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:14:12.929608008Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:14:13.055705228Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:14:13.139608973Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:14:13.180675962Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:14:52.615583163Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:14:52.734784606Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:14:52.873451293Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:14:56.806937786Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:14:57.030971055Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:14:57.251883811Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:16:29.171076715Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:16:29.17795781Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:19:29.151864885Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:19:29.157935823Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:21:51.771788439Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:22:19.90126568Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:22:20.045808341Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:22:20.100175885Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:22:20.247182912Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:22:20.298473861Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:22:20.446330951Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:22:20.496691687Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:22:29.143325336Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:22:29.163410508Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:23:44.183786838Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:23:44.338221988Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:23:44.405605174Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:25:29.145686024Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:25:29.151971554Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:28:29.147837401Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:28:29.155492146Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:31:29.166771924Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:31:29.173367219Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:34:29.16803945Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:34:29.174490126Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:37:29.167525615Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:37:29.1743958Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:40:15.629895499Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Probe err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.6299147Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1alpha1.AlertmanagerConfig err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629895609Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629891962Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PrometheusRule err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629934647Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629950197Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PodMonitor err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629959719Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.62995928Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:571 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629967162Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629944196Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.ServiceMonitor err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629979428Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PrometheusRule err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.629992393Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:486 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630000448Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.ThanosRuler err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630015996Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630021214Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/thanos/operator.go:370 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.63002407Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630023645Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:573 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630030039Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/thanos/operator.go:368 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630040006Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:484 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630031152Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630044243Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Prometheus err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630050134Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Alertmanager err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:15.630020485Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-05-06T20:40:25.091615294Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:40:25.091614484Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:40:25.200094902Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:40:25.292571642Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:40:29.053327985Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:40:29.053335695Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:40:29.135505463Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-05-06T20:40:29.147585414Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:40:29.155142671Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:40:29.270416477Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-05-06T20:43:29.143096736Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:43:29.149165938Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:46:29.141475122Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-05-06T20:46:29.146838015Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice"