ts=2026-03-19T13:57:24.015732249Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:260 msg="Starting Prometheus Operator" version="(version=0.87.1, branch=HEAD, revision=87123f32)" build_context="(go=go1.24.13 (Red Hat 1.24.13-1.el9_6) X:strictfipsruntime, platform=linux/amd64, user=root, date=20260228-04:20:52, tags=strictfipsruntime)" feature_gates="PrometheusAgentDaemonSet=false,PrometheusShardRetentionPolicy=false,PrometheusTopologySharding=false,StatusForConfigurationResources=false" ts=2026-03-19T13:57:24.015840781Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:261 msg="Operator's configuration" watch_referenced_objects_in_all_namespaces=true controller_id=openshift-monitoring/prometheus-operator enable_config_reloader_probes=false ts=2026-03-19T13:57:24.016455154Z level=info caller=/go/src/github.com/coreos/prometheus-operator/internal/goruntime/cpu.go:27 msg="Leaving GOMAXPROCS=8: CPU quota undefined" ts=2026-03-19T13:57:24.016498566Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:275 msg="Namespaces filtering configuration " config="{allow_list=\"\",deny_list=\"\",prometheus_allow_list=\"openshift-monitoring\",alertmanager_allow_list=\"openshift-monitoring\",alertmanagerconfig_allow_list=\"\",thanosruler_allow_list=\"openshift-monitoring\"}" ts=2026-03-19T13:57:24.034615405Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:316 msg="connection established" kubernetes_version=1.34.4 ts=2026-03-19T13:57:24.04702344Z level=warn caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:82 msg="resource \"scrapeconfigs\" (group: \"monitoring.coreos.com/v1alpha1\") not installed in the cluster" ts=2026-03-19T13:57:24.047051237Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:401 msg="Kubernetes API capabilities" endpointslices=true ts=2026-03-19T13:57:24.073846571Z level=warn caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:82 msg="resource \"prometheusagents\" (group: \"monitoring.coreos.com/v1alpha1\") not installed in the cluster" ts=2026-03-19T13:57:24.136372851Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/server/server.go:293 msg="starting insecure server" address=127.0.0.1:8080 ts=2026-03-19T13:57:24.13641265Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/kubelet/controller.go:214 msg="Starting controller" component=kubelet_endpoints kubelet_object=kube-system/kubelet ts=2026-03-19T13:57:24.136465619Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.136456678Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=thanos ts=2026-03-19T13:57:24.136460971Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=alertmanager ts=2026-03-19T13:57:24.187387068Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T13:57:24.194045816Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T13:57:24.239244322Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239330781Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.239344824Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239360438Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.239370084Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239382819Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.239391109Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.23940604Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.239415091Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239428265Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.23944237Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239461995Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.239474853Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239488158Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.23949705Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239509337Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.239518795Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239530939Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=prometheus ts=2026-03-19T13:57:24.23953862Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=prometheus ts=2026-03-19T13:57:24.239552052Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:458 msg="successfully synced all caches" component=prometheus-controller ts=2026-03-19T13:57:24.241582672Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=thanos ts=2026-03-19T13:57:24.241589155Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=alertmanager ts=2026-03-19T13:57:24.241644179Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=thanos ts=2026-03-19T13:57:24.241660409Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=thanos ts=2026-03-19T13:57:24.241687268Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=thanos ts=2026-03-19T13:57:24.241707422Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=thanos ts=2026-03-19T13:57:24.241646512Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=alertmanager ts=2026-03-19T13:57:24.241764108Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=thanos ts=2026-03-19T13:57:24.24177797Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=thanos ts=2026-03-19T13:57:24.241796006Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=thanos ts=2026-03-19T13:57:24.241883957Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=thanos ts=2026-03-19T13:57:24.241799281Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=alertmanager ts=2026-03-19T13:57:24.24198255Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=alertmanager ts=2026-03-19T13:57:24.242000901Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=thanos ts=2026-03-19T13:57:24.242017742Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=alertmanager ts=2026-03-19T13:57:24.242024533Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=thanos ts=2026-03-19T13:57:24.242041557Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=alertmanager ts=2026-03-19T13:57:24.242046855Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/thanos/operator.go:339 msg="successfully synced all caches" component=thanos-controller ts=2026-03-19T13:57:24.242048585Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=alertmanager ts=2026-03-19T13:57:24.242068968Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=alertmanager ts=2026-03-19T13:57:24.242092723Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=alertmanager ts=2026-03-19T13:57:24.242108962Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=alertmanager ts=2026-03-19T13:57:24.242121686Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=alertmanager ts=2026-03-19T13:57:24.242140387Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:349 msg="Waiting for caches to sync" controller=alertmanager ts=2026-03-19T13:57:24.24215094Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:356 msg="Caches are synced" controller=alertmanager ts=2026-03-19T13:57:24.24216205Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:376 msg="successfully synced all caches" component=alertmanager-controller ts=2026-03-19T13:57:28.087220052Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:28.153753332Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:730 msg="StatefulSet not found" component=alertmanager-controller key=openshift-monitoring/alertmanager-main ts=2026-03-19T13:57:28.167194684Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:28.167388383Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:730 msg="StatefulSet not found" component=alertmanager-controller key=openshift-monitoring/alertmanager-main ts=2026-03-19T13:57:28.262325204Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:28.338219737Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:28.458364934Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:28.832617692Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:28.892446631Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:28.949929393Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:29.315395758Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:30.024697833Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:30.100322619Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:30.296635306Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:30.565057139Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:31.5877669Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:31.758623818Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:31.830137971Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:32.004178887Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:33.069511882Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:33.13344983Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:33.150133036Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:33.208095254Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:33.265862573Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:33.320112392Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:33.354923296Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:33.409594187Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:33.538548507Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:33.605194899Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:33.88391096Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:34.418972213Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:34.56042997Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:34.635678283Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:34.689277798Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:34.90791744Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:35.037617008Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:35.102834163Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:35.131067248Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:35.328387862Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:35.455198743Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:35.529848839Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:35.555132357Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:45.117153566Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:45.305096545Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:45.352652698Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:45.536745907Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:45.630426091Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:45.657034786Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:59.632917234Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:57:59.801537181Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:57:59.839732546Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:58:08.112270886Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:58:08.225253555Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:58:08.28731105Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:58:08.318083314Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:58:08.455344166Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:58:08.495673865Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:58:08.794013291Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:58:08.904106752Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:58:08.971869493Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:58:09.022785888Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:00.466346924Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PodMonitor err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.46647183Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466492032Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Probe err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466493007Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PrometheusRule err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.46649874Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466342771Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Alertmanager err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466371398Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466378531Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1alpha1.AlertmanagerConfig err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466534397Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/thanos/operator.go:389 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466540912Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:590 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.46634333Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466378464Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:493 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466407426Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466556464Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466424587Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.ServiceMonitor err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466418289Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:592 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466428673Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/thanos/operator.go:391 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466434295Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PrometheusRule err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.46645676Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.ThanosRuler err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.46645722Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.46645659Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Prometheus err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466465072Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:00.466498098Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:568 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:495 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-03-19T13:59:06.480158793Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:06.480186715Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:06.542661406Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:06.654159971Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:06.695874021Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:06.832644157Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:07.481000345Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:07.554592055Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:07.666845371Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:11.540341635Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:11.540389428Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:11.736298288Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:11.935912412Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:12.155256194Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:14.651704344Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:14.76793891Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:14.835555701Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:14.878804275Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T13:59:15.050490989Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T13:59:15.100730912Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:00:24.158002547Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:00:24.162665005Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:03:24.16114429Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:03:24.165390899Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:06:24.159026452Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:06:24.164369303Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:07:07.359937937Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:07:34.045315066Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:07:34.186295696Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T14:07:34.228241437Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:07:34.387088085Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T14:07:34.425270734Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:07:34.582094313Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T14:07:34.632182435Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:08:52.88184513Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:08:53.020524154Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T14:08:53.061459221Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:08:53.203040922Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:611 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-03-19T14:08:53.245606812Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:854 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-03-19T14:09:24.155848958Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:09:24.160450681Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:12:24.155348911Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-03-19T14:12:24.159906807Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice"