ts=2026-04-10T16:44:20.482486126Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:218 msg="Starting Prometheus Operator" version="(version=0.85.0, branch=HEAD, revision=2740c4be8)" build_context="(go=go1.24.6 (Red Hat 1.24.6-1.el9_6) X:strictfipsruntime, platform=linux/amd64, user=root, date=20251208-13:40:21, tags=strictfipsruntime)" feature_gates="PrometheusAgentDaemonSet=false,PrometheusShardRetentionPolicy=false,PrometheusTopologySharding=false,StatusForConfigurationResources=false" ts=2026-04-10T16:44:20.482546853Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:219 msg="Operator's configuration" watch_referenced_objects_in_all_namespaces=true controller_id=openshift-monitoring/prometheus-operator enable_config_reloader_probes=false ts=2026-04-10T16:44:20.482852526Z level=info caller=/go/src/github.com/coreos/prometheus-operator/internal/goruntime/cpu.go:27 msg="Leaving GOMAXPROCS=16: CPU quota undefined" ts=2026-04-10T16:44:20.482911711Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:233 msg="Namespaces filtering configuration " config="{allow_list=\"\",deny_list=\"\",prometheus_allow_list=\"openshift-monitoring\",alertmanager_allow_list=\"openshift-monitoring\",alertmanagerconfig_allow_list=\"\",thanosruler_allow_list=\"openshift-monitoring\"}" ts=2026-04-10T16:44:20.497565931Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:274 msg="connection established" kubernetes_version=1.33.6 ts=2026-04-10T16:44:20.512289243Z level=warn caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:81 msg="resource \"scrapeconfigs\" (group: \"monitoring.coreos.com/v1alpha1\") not installed in the cluster" ts=2026-04-10T16:44:20.512322161Z level=info caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:359 msg="Kubernetes API capabilities" endpointslices=true ts=2026-04-10T16:44:20.542440865Z level=warn caller=/go/src/github.com/coreos/prometheus-operator/cmd/operator/main.go:81 msg="resource \"prometheusagents\" (group: \"monitoring.coreos.com/v1alpha1\") not installed in the cluster" ts=2026-04-10T16:44:20.595675511Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/kubelet/controller.go:214 msg="Starting controller" component=kubelet_endpoints kubelet_object=kube-system/kubelet ts=2026-04-10T16:44:20.595705666Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.595708671Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-04-10T16:44:20.595749089Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-04-10T16:44:20.595738859Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/server/server.go:293 msg="starting insecure server" address=127.0.0.1:8080 ts=2026-04-10T16:44:20.632575066Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:44:20.639931412Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:44:20.696069886Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696081676Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-04-10T16:44:20.696132167Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-04-10T16:44:20.696140731Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-04-10T16:44:20.696143441Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.696149518Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-04-10T16:44:20.696153276Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696155639Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-04-10T16:44:20.696163713Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.696168845Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-04-10T16:44:20.696175634Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-04-10T16:44:20.696185279Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-04-10T16:44:20.696169803Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.69619076Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-04-10T16:44:20.696215101Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=thanos ts=2026-04-10T16:44:20.696221309Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=thanos ts=2026-04-10T16:44:20.696217221Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.696232624Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696230466Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/thanos/operator.go:318 msg="successfully synced all caches" component=thanos-controller ts=2026-04-10T16:44:20.696242941Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.696596883Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696649685Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.696663443Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696689404Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.696698791Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696712406Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.696735998Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696749847Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.696758925Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696786238Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=prometheus ts=2026-04-10T16:44:20.69679486Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=prometheus ts=2026-04-10T16:44:20.696805128Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:439 msg="successfully synced all caches" component=prometheus-controller ts=2026-04-10T16:44:20.697897112Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-04-10T16:44:20.697935486Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-04-10T16:44:20.697943934Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-04-10T16:44:20.697954039Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-04-10T16:44:20.697959434Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-04-10T16:44:20.697967879Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-04-10T16:44:20.697973173Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-04-10T16:44:20.697982615Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-04-10T16:44:20.697992682Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-04-10T16:44:20.698008543Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-04-10T16:44:20.698017368Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-04-10T16:44:20.698027845Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:350 msg="Waiting for caches to sync" controller=alertmanager ts=2026-04-10T16:44:20.69803664Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/shared_informer.go:357 msg="Caches are synced" controller=alertmanager ts=2026-04-10T16:44:20.698044619Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:367 msg="successfully synced all caches" component=alertmanager-controller ts=2026-04-10T16:44:23.920017237Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:23.985695365Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:722 msg="StatefulSet not found" component=alertmanager-controller key=openshift-monitoring/alertmanager-main ts=2026-04-10T16:44:24.006172314Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:24.101675613Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:24.180505263Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:24.306658409Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:24.711224729Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:24.777432638Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:24.838331168Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:24.950946786Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:25.97568514Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:26.118359837Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:26.396155617Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:27.418633037Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:27.676988888Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:27.799807592Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:27.969741203Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:28.961308274Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:29.02452241Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:29.032670189Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:44:29.157208055Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:29.216337915Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:29.257212683Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:44:29.277163421Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:29.377878837Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:29.478848997Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:44:29.643172687Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:44:45.028915401Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:44:45.165567422Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:45.207394129Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:44:45.314922268Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:45.374763709Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:45.426502997Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:44:59.460436371Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:44:59.595657595Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:44:59.638955936Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:45:13.947570656Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:45:14.053118951Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:45:14.119552422Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:45:14.150820214Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:45:42.40070555Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:45:42.400716904Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:45:42.542756604Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:45:42.627722826Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:45:47.474748414Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:45:47.474751902Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:45:51.550423506Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:45:51.6862028Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:45:51.751990612Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:45:51.806519908Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:45:51.80814045Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:47:20.622972287Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:47:20.627740786Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:50:20.618163597Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:50:20.623476906Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:52:12.182584654Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:52:45.559996336Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:52:45.904511163Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:52:46.079897534Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:52:46.228011448Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:52:46.29420401Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:53:20.620297065Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:53:20.626142976Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:54:12.711537837Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:54:12.844441654Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T16:54:12.889398547Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T16:56:20.617742499Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:56:20.623875458Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:59:20.633322011Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T16:59:20.638359436Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T17:02:20.63139133Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T17:02:20.636229372Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T17:05:20.635714166Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T17:05:20.640595204Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T17:05:30.133765587Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T17:05:30.134103214Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T17:05:30.260381394Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T17:05:30.308200214Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T17:05:35.122660625Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T17:05:35.122673224Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T17:05:35.237207855Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:603 msg="sync alertmanager" component=alertmanager-controller key=openshift-monitoring/main ts=2026-04-10T17:05:35.310335323Z level=info caller=/go/src/github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:832 msg="sync prometheus" component=prometheus-controller key=openshift-monitoring/k8s ts=2026-04-10T17:08:20.619528181Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T17:08:20.624321089Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/rest/warnings.go:110 msg="Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice" ts=2026-04-10T17:10:30.562470063Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Prometheus err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562497552Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562511591Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562522383Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562520963Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562544183Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.ThanosRuler err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562549907Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/thanos/operator.go:368 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.56254518Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/thanos/operator.go:370 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562480564Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562570135Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PodMonitor err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562565968Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:486 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562570753Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562476737Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/alertmanager/operator.go:484 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562584972Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.StatefulSet err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562578993Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Alertmanager err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562590993Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PrometheusRule err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562612157Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1alpha1.AlertmanagerConfig err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562625726Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.ServiceMonitor err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562622106Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.Probe err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562624596Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PrometheusRule err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562624348Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:571 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562640767Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/prometheus/server/operator.go:573 type=*v1.Namespace err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding" ts=2026-04-10T17:10:30.562640922Z level=info caller=/go/src/github.com/coreos/prometheus-operator/vendor/k8s.io/client-go/tools/cache/reflector.go:556 msg="Warning: watch ended with error" reflector=github.com/coreos/prometheus-operator/pkg/informers/informers.go:137 type=*v1.PartialObjectMetadata err="an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding"