I0213 18:21:17.135329 1 serving.go:386] Generated self-signed cert in-memory I0213 18:21:17.337276 1 controllermanager.go:185] "Starting" version="v1.32.5" I0213 18:21:17.337298 1 controllermanager.go:187] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" I0213 18:21:17.338651 1 dynamic_cafile_content.go:161] "Starting controller" name="request-header::/etc/kubernetes/pki/front-proxy-ca.crt" I0213 18:21:17.338668 1 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" I0213 18:21:17.338869 1 secure_serving.go:213] Serving securely on 127.0.0.1:10257 I0213 18:21:17.338955 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" I0213 18:21:17.339078 1 leaderelection.go:257] attempting to acquire leader lease kube-system/kube-controller-manager... E0213 18:21:18.584504 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: leases.coordination.k8s.io "kube-controller-manager" is forbidden: User "system:kube-controller-manager" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-system" I0213 18:21:21.289305 1 leaderelection.go:271] successfully acquired lease kube-system/kube-controller-manager I0213 18:21:21.289383 1 event.go:389] "Event occurred" object="kube-system/kube-controller-manager" fieldPath="" kind="Lease" apiVersion="coordination.k8s.io/v1" type="Normal" reason="LeaderElection" message="kind-mapt-control-plane_6b032b3b-ddd8-43a2-bc77-d702da5dcc4e became leader" I0213 18:21:21.291658 1 controllermanager.go:765] "Started controller" controller="serviceaccount-token-controller" I0213 18:21:21.291670 1 shared_informer.go:313] Waiting for caches to sync for tokens I0213 18:21:21.310657 1 controllermanager.go:765] "Started controller" controller="token-cleaner-controller" I0213 18:21:21.310677 1 controllermanager.go:723] "Skipping a cloud provider controller" controller="cloud-node-lifecycle-controller" I0213 18:21:21.310683 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="kube-apiserver-serving-clustertrustbundle-publisher-controller" requiredFeatureGates=["ClusterTrustBundle"] I0213 18:21:21.310724 1 controllermanager.go:743] "Warning: skipping controller" controller="storage-version-migrator-controller" I0213 18:21:21.310795 1 tokencleaner.go:117] "Starting token cleaner controller" logger="token-cleaner-controller" I0213 18:21:21.310822 1 shared_informer.go:313] Waiting for caches to sync for token_cleaner I0213 18:21:21.310831 1 shared_informer.go:320] Caches are synced for token_cleaner I0213 18:21:21.328967 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="rolebindings.rbac.authorization.k8s.io" I0213 18:21:21.329006 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="limitranges" I0213 18:21:21.329018 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="poddisruptionbudgets.policy" I0213 18:21:21.329028 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="roles.rbac.authorization.k8s.io" I0213 18:21:21.329045 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="networkpolicies.networking.k8s.io" I0213 18:21:21.329062 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="controllerrevisions.apps" I0213 18:21:21.329083 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="daemonsets.apps" I0213 18:21:21.329116 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="cronjobs.batch" I0213 18:21:21.329150 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="ingresses.networking.k8s.io" I0213 18:21:21.329169 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="podtemplates" W0213 18:21:21.329199 1 shared_informer.go:597] resyncPeriod 13h47m18.785082006s is smaller than resyncCheckPeriod 17h46m1.244146155s and the informer has already started. Changing it to 17h46m1.244146155s I0213 18:21:21.329241 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="serviceaccounts" I0213 18:21:21.329280 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="csistoragecapacities.storage.k8s.io" I0213 18:21:21.329326 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="leases.coordination.k8s.io" I0213 18:21:21.329363 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="statefulsets.apps" I0213 18:21:21.329383 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="jobs.batch" I0213 18:21:21.329398 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="endpointslices.discovery.k8s.io" I0213 18:21:21.329421 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="endpoints" I0213 18:21:21.329432 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="deployments.apps" I0213 18:21:21.329449 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="replicasets.apps" I0213 18:21:21.329467 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="horizontalpodautoscalers.autoscaling" I0213 18:21:21.329507 1 controllermanager.go:765] "Started controller" controller="resourcequota-controller" I0213 18:21:21.329548 1 resource_quota_controller.go:300] "Starting resource quota controller" logger="resourcequota-controller" I0213 18:21:21.329570 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:21:21.329590 1 resource_quota_monitor.go:308] "QuotaMonitor running" logger="resourcequota-controller" I0213 18:21:21.338964 1 controllermanager.go:765] "Started controller" controller="disruption-controller" I0213 18:21:21.339062 1 disruption.go:452] "Sending events to api server." logger="disruption-controller" I0213 18:21:21.339096 1 disruption.go:463] "Starting disruption controller" logger="disruption-controller" I0213 18:21:21.339104 1 shared_informer.go:313] Waiting for caches to sync for disruption I0213 18:21:21.346054 1 controllermanager.go:765] "Started controller" controller="cronjob-controller" I0213 18:21:21.346193 1 cronjob_controllerv2.go:145] "Starting cronjob controller v2" logger="cronjob-controller" I0213 18:21:21.346214 1 shared_informer.go:313] Waiting for caches to sync for cronjob I0213 18:21:21.353159 1 controllermanager.go:765] "Started controller" controller="bootstrap-signer-controller" I0213 18:21:21.353280 1 shared_informer.go:313] Waiting for caches to sync for bootstrap_signer I0213 18:21:21.360200 1 controllermanager.go:765] "Started controller" controller="legacy-serviceaccount-token-cleaner-controller" I0213 18:21:21.360286 1 legacy_serviceaccount_token_cleaner.go:103] "Starting legacy service account token cleaner controller" logger="legacy-serviceaccount-token-cleaner-controller" I0213 18:21:21.360308 1 shared_informer.go:313] Waiting for caches to sync for legacy-service-account-token-cleaner I0213 18:21:21.367418 1 controllermanager.go:765] "Started controller" controller="endpointslice-mirroring-controller" I0213 18:21:21.367626 1 endpointslicemirroring_controller.go:227] "Starting EndpointSliceMirroring controller" logger="endpointslice-mirroring-controller" I0213 18:21:21.367653 1 shared_informer.go:313] Waiting for caches to sync for endpoint_slice_mirroring I0213 18:21:21.374748 1 controllermanager.go:765] "Started controller" controller="serviceaccount-controller" I0213 18:21:21.374834 1 serviceaccounts_controller.go:114] "Starting service account controller" logger="serviceaccount-controller" I0213 18:21:21.374856 1 shared_informer.go:313] Waiting for caches to sync for service account I0213 18:21:21.392378 1 shared_informer.go:320] Caches are synced for tokens I0213 18:21:21.395154 1 controllermanager.go:765] "Started controller" controller="daemonset-controller" I0213 18:21:21.395223 1 daemon_controller.go:294] "Starting daemon sets controller" logger="daemonset-controller" I0213 18:21:21.395233 1 shared_informer.go:313] Waiting for caches to sync for daemon sets I0213 18:21:21.691434 1 controllermanager.go:765] "Started controller" controller="horizontal-pod-autoscaler-controller" I0213 18:21:21.691455 1 controllermanager.go:723] "Skipping a cloud provider controller" controller="node-route-controller" I0213 18:21:21.691477 1 horizontal.go:201] "Starting HPA controller" logger="horizontal-pod-autoscaler-controller" I0213 18:21:21.691486 1 shared_informer.go:313] Waiting for caches to sync for HPA I0213 18:21:21.845512 1 controllermanager.go:765] "Started controller" controller="taint-eviction-controller" I0213 18:21:21.845579 1 taint_eviction.go:281] "Starting" logger="taint-eviction-controller" controller="taint-eviction-controller" I0213 18:21:21.845622 1 taint_eviction.go:287] "Sending events to api server" logger="taint-eviction-controller" I0213 18:21:21.845660 1 shared_informer.go:313] Waiting for caches to sync for taint-eviction-controller I0213 18:21:21.995706 1 controllermanager.go:765] "Started controller" controller="endpointslice-controller" I0213 18:21:21.995792 1 endpointslice_controller.go:281] "Starting endpoint slice controller" logger="endpointslice-controller" I0213 18:21:21.995815 1 shared_informer.go:313] Waiting for caches to sync for endpoint_slice I0213 18:21:22.241214 1 garbagecollector.go:144] "Starting controller" logger="garbage-collector-controller" controller="garbagecollector" I0213 18:21:22.241240 1 controllermanager.go:765] "Started controller" controller="garbage-collector-controller" I0213 18:21:22.241248 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 18:21:22.241274 1 graph_builder.go:351] "Running" logger="garbage-collector-controller" component="GraphBuilder" I0213 18:21:22.495981 1 controllermanager.go:765] "Started controller" controller="statefulset-controller" I0213 18:21:22.496063 1 stateful_set.go:166] "Starting stateful set controller" logger="statefulset-controller" I0213 18:21:22.496081 1 shared_informer.go:313] Waiting for caches to sync for stateful set I0213 18:21:22.647316 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-signing-controller" name="csrsigning-kubelet-serving" I0213 18:21:22.647342 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrsigning-kubelet-serving I0213 18:21:22.647347 1 dynamic_serving_content.go:135] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0213 18:21:22.647636 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-signing-controller" name="csrsigning-kubelet-client" I0213 18:21:22.647667 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrsigning-kubelet-client I0213 18:21:22.647709 1 dynamic_serving_content.go:135] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0213 18:21:22.647926 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-signing-controller" name="csrsigning-kube-apiserver-client" I0213 18:21:22.647951 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrsigning-kube-apiserver-client I0213 18:21:22.647965 1 dynamic_serving_content.go:135] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0213 18:21:22.648148 1 controllermanager.go:765] "Started controller" controller="certificatesigningrequest-signing-controller" I0213 18:21:22.648195 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-signing-controller" name="csrsigning-legacy-unknown" I0213 18:21:22.648203 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrsigning-legacy-unknown I0213 18:21:22.648232 1 dynamic_serving_content.go:135] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0213 18:21:22.692718 1 controllermanager.go:765] "Started controller" controller="certificatesigningrequest-approving-controller" I0213 18:21:22.692740 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-approving-controller" name="csrapproving" I0213 18:21:22.692766 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrapproving I0213 18:21:22.846504 1 range_allocator.go:112] "No Secondary Service CIDR provided. Skipping filtering out secondary service addresses" logger="node-ipam-controller" I0213 18:21:22.846537 1 controllermanager.go:765] "Started controller" controller="node-ipam-controller" I0213 18:21:22.846681 1 node_ipam_controller.go:141] "Starting ipam controller" logger="node-ipam-controller" I0213 18:21:22.846704 1 shared_informer.go:313] Waiting for caches to sync for node I0213 18:21:23.001804 1 controllermanager.go:765] "Started controller" controller="persistentvolume-expander-controller" I0213 18:21:23.001846 1 expand_controller.go:329] "Starting expand controller" logger="persistentvolume-expander-controller" I0213 18:21:23.001861 1 shared_informer.go:313] Waiting for caches to sync for expand I0213 18:21:23.145933 1 controllermanager.go:765] "Started controller" controller="persistentvolume-protection-controller" I0213 18:21:23.145988 1 pv_protection_controller.go:81] "Starting PV protection controller" logger="persistentvolume-protection-controller" I0213 18:21:23.146001 1 shared_informer.go:313] Waiting for caches to sync for PV protection I0213 18:21:23.295147 1 controllermanager.go:765] "Started controller" controller="endpoints-controller" I0213 18:21:23.295214 1 endpoints_controller.go:182] "Starting endpoint controller" logger="endpoints-controller" I0213 18:21:23.295222 1 shared_informer.go:313] Waiting for caches to sync for endpoint I0213 18:21:23.445910 1 controllermanager.go:765] "Started controller" controller="deployment-controller" I0213 18:21:23.445992 1 deployment_controller.go:173] "Starting controller" logger="deployment-controller" controller="deployment" I0213 18:21:23.446004 1 shared_informer.go:313] Waiting for caches to sync for deployment I0213 18:21:23.596141 1 controllermanager.go:765] "Started controller" controller="ttl-after-finished-controller" I0213 18:21:23.596201 1 ttlafterfinished_controller.go:112] "Starting TTL after finished controller" logger="ttl-after-finished-controller" I0213 18:21:23.596213 1 shared_informer.go:313] Waiting for caches to sync for TTL after finished I0213 18:21:23.746123 1 controllermanager.go:765] "Started controller" controller="ttl-controller" I0213 18:21:23.746144 1 ttl_controller.go:127] "Starting TTL controller" logger="ttl-controller" I0213 18:21:23.746158 1 shared_informer.go:313] Waiting for caches to sync for TTL I0213 18:21:23.895934 1 controllermanager.go:765] "Started controller" controller="persistentvolume-attach-detach-controller" I0213 18:21:23.896046 1 attach_detach_controller.go:338] "Starting attach detach controller" logger="persistentvolume-attach-detach-controller" I0213 18:21:23.896071 1 shared_informer.go:313] Waiting for caches to sync for attach detach I0213 18:21:24.045882 1 controllermanager.go:765] "Started controller" controller="root-ca-certificate-publisher-controller" I0213 18:21:24.045904 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="resourceclaim-controller" requiredFeatureGates=["DynamicResourceAllocation"] I0213 18:21:24.045943 1 publisher.go:107] "Starting root CA cert publisher controller" logger="root-ca-certificate-publisher-controller" I0213 18:21:24.045953 1 shared_informer.go:313] Waiting for caches to sync for crt configmap I0213 18:21:24.196845 1 controllermanager.go:765] "Started controller" controller="pod-garbage-collector-controller" I0213 18:21:24.196965 1 gc_controller.go:99] "Starting GC controller" logger="pod-garbage-collector-controller" I0213 18:21:24.196991 1 shared_informer.go:313] Waiting for caches to sync for GC I0213 18:21:24.346451 1 controllermanager.go:765] "Started controller" controller="job-controller" I0213 18:21:24.346523 1 job_controller.go:243] "Starting job controller" logger="job-controller" I0213 18:21:24.346534 1 shared_informer.go:313] Waiting for caches to sync for job I0213 18:21:24.595879 1 controllermanager.go:765] "Started controller" controller="namespace-controller" I0213 18:21:24.595908 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="volumeattributesclass-protection-controller" requiredFeatureGates=["VolumeAttributesClass"] I0213 18:21:24.595972 1 namespace_controller.go:202] "Starting namespace controller" logger="namespace-controller" I0213 18:21:24.595994 1 shared_informer.go:313] Waiting for caches to sync for namespace I0213 18:21:24.746191 1 controllermanager.go:765] "Started controller" controller="persistentvolume-binder-controller" I0213 18:21:24.746297 1 pv_controller_base.go:308] "Starting persistent volume controller" logger="persistentvolume-binder-controller" I0213 18:21:24.746318 1 shared_informer.go:313] Waiting for caches to sync for persistent volume I0213 18:21:24.897543 1 controllermanager.go:765] "Started controller" controller="clusterrole-aggregation-controller" I0213 18:21:24.897623 1 clusterroleaggregation_controller.go:194] "Starting ClusterRoleAggregator controller" logger="clusterrole-aggregation-controller" I0213 18:21:24.897632 1 shared_informer.go:313] Waiting for caches to sync for ClusterRoleAggregator I0213 18:21:25.046278 1 controllermanager.go:765] "Started controller" controller="persistentvolumeclaim-protection-controller" I0213 18:21:25.046310 1 pvc_protection_controller.go:168] "Starting PVC protection controller" logger="persistentvolumeclaim-protection-controller" I0213 18:21:25.046328 1 shared_informer.go:313] Waiting for caches to sync for PVC protection I0213 18:21:25.196251 1 controllermanager.go:765] "Started controller" controller="ephemeral-volume-controller" I0213 18:21:25.196457 1 controller.go:173] "Starting ephemeral volume controller" logger="ephemeral-volume-controller" I0213 18:21:25.196484 1 shared_informer.go:313] Waiting for caches to sync for ephemeral I0213 18:21:25.391838 1 controllermanager.go:765] "Started controller" controller="validatingadmissionpolicy-status-controller" I0213 18:21:25.391865 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="service-cidr-controller" requiredFeatureGates=["MultiCIDRServiceAllocator"] I0213 18:21:25.391911 1 shared_informer.go:313] Waiting for caches to sync for validatingadmissionpolicy-status I0213 18:21:25.545352 1 controllermanager.go:765] "Started controller" controller="replicationcontroller-controller" I0213 18:21:25.545447 1 replica_set.go:217] "Starting controller" logger="replicationcontroller-controller" name="replicationcontroller" I0213 18:21:25.545457 1 shared_informer.go:313] Waiting for caches to sync for ReplicationController I0213 18:21:25.695751 1 controllermanager.go:765] "Started controller" controller="replicaset-controller" I0213 18:21:25.695773 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="selinux-warning-controller" requiredFeatureGates=["SELinuxChangePolicy"] I0213 18:21:25.695863 1 replica_set.go:217] "Starting controller" logger="replicaset-controller" name="replicaset" I0213 18:21:25.695880 1 shared_informer.go:313] Waiting for caches to sync for ReplicaSet I0213 18:21:25.741938 1 node_lifecycle_controller.go:432] "Controller will reconcile labels" logger="node-lifecycle-controller" I0213 18:21:25.741987 1 controllermanager.go:765] "Started controller" controller="node-lifecycle-controller" I0213 18:21:25.741993 1 controllermanager.go:723] "Skipping a cloud provider controller" controller="service-lb-controller" I0213 18:21:25.741999 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="storageversion-garbage-collector-controller" requiredFeatureGates=["APIServerIdentity","StorageVersionAPI"] I0213 18:21:25.742071 1 node_lifecycle_controller.go:466] "Sending events to api server" logger="node-lifecycle-controller" I0213 18:21:25.742122 1 node_lifecycle_controller.go:477] "Starting node controller" logger="node-lifecycle-controller" I0213 18:21:25.742139 1 shared_informer.go:313] Waiting for caches to sync for taint I0213 18:21:25.791890 1 controllermanager.go:765] "Started controller" controller="certificatesigningrequest-cleaner-controller" I0213 18:21:25.791915 1 cleaner.go:83] "Starting CSR cleaner controller" logger="certificatesigningrequest-cleaner-controller" I0213 18:21:25.796856 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:21:25.801174 1 actual_state_of_world.go:541] "Failed to update statusUpdateNeeded field in actual state of world" logger="persistentvolume-attach-detach-controller" err="Failed to set statusUpdateNeeded to needed true, because nodeName=\"kind-mapt-control-plane\" does not exist" I0213 18:21:25.801895 1 shared_informer.go:320] Caches are synced for expand I0213 18:21:25.805666 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 18:21:25.830210 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:21:25.839704 1 shared_informer.go:320] Caches are synced for disruption I0213 18:21:25.841992 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:21:25.842010 1 garbagecollector.go:154] "Garbage collector: all resource monitors have synced" logger="garbage-collector-controller" I0213 18:21:25.842023 1 garbagecollector.go:157] "Proceeding to collect garbage" logger="garbage-collector-controller" I0213 18:21:25.842308 1 shared_informer.go:320] Caches are synced for taint I0213 18:21:25.842371 1 node_lifecycle_controller.go:1234] "Initializing eviction metric for zone" logger="node-lifecycle-controller" zone="" I0213 18:21:25.842526 1 node_lifecycle_controller.go:886] "Missing timestamp for Node. Assuming now as a timestamp" logger="node-lifecycle-controller" node="kind-mapt-control-plane" I0213 18:21:25.842568 1 node_lifecycle_controller.go:1038] "Controller detected that all Nodes are not-Ready. Entering master disruption mode" logger="node-lifecycle-controller" I0213 18:21:25.845842 1 shared_informer.go:320] Caches are synced for ReplicationController I0213 18:21:25.845873 1 shared_informer.go:320] Caches are synced for taint-eviction-controller I0213 18:21:25.846094 1 shared_informer.go:320] Caches are synced for PV protection I0213 18:21:25.846183 1 shared_informer.go:320] Caches are synced for crt configmap I0213 18:21:25.846204 1 shared_informer.go:320] Caches are synced for deployment I0213 18:21:25.846293 1 shared_informer.go:320] Caches are synced for cronjob I0213 18:21:25.846318 1 shared_informer.go:320] Caches are synced for TTL I0213 18:21:25.846344 1 shared_informer.go:320] Caches are synced for persistent volume I0213 18:21:25.846363 1 shared_informer.go:320] Caches are synced for PVC protection I0213 18:21:25.846721 1 shared_informer.go:320] Caches are synced for job I0213 18:21:25.846798 1 shared_informer.go:320] Caches are synced for node I0213 18:21:25.846878 1 range_allocator.go:177] "Sending events to api server" logger="node-ipam-controller" I0213 18:21:25.846926 1 range_allocator.go:183] "Starting range CIDR allocator" logger="node-ipam-controller" I0213 18:21:25.846939 1 shared_informer.go:313] Waiting for caches to sync for cidrallocator I0213 18:21:25.846947 1 shared_informer.go:320] Caches are synced for cidrallocator I0213 18:21:25.847382 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-kubelet-serving I0213 18:21:25.847934 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-kubelet-client I0213 18:21:25.847997 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-kube-apiserver-client I0213 18:21:25.849120 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-legacy-unknown I0213 18:21:25.853464 1 shared_informer.go:320] Caches are synced for bootstrap_signer I0213 18:21:25.857285 1 range_allocator.go:428] "Set node PodCIDR" logger="node-ipam-controller" node="kind-mapt-control-plane" podCIDRs=["10.244.0.0/24"] I0213 18:21:25.857314 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:21:25.857339 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:21:25.861278 1 shared_informer.go:320] Caches are synced for legacy-service-account-token-cleaner I0213 18:21:25.868568 1 shared_informer.go:320] Caches are synced for endpoint_slice_mirroring I0213 18:21:25.875974 1 shared_informer.go:320] Caches are synced for service account I0213 18:21:25.892542 1 shared_informer.go:320] Caches are synced for validatingadmissionpolicy-status I0213 18:21:25.892570 1 shared_informer.go:320] Caches are synced for HPA I0213 18:21:25.893770 1 shared_informer.go:320] Caches are synced for certificate-csrapproving I0213 18:21:25.896050 1 shared_informer.go:320] Caches are synced for namespace I0213 18:21:25.896074 1 shared_informer.go:320] Caches are synced for daemon sets I0213 18:21:25.896093 1 shared_informer.go:320] Caches are synced for endpoint I0213 18:21:25.896113 1 shared_informer.go:320] Caches are synced for attach detach I0213 18:21:25.896219 1 shared_informer.go:320] Caches are synced for stateful set I0213 18:21:25.896247 1 shared_informer.go:320] Caches are synced for endpoint_slice I0213 18:21:25.896264 1 shared_informer.go:320] Caches are synced for TTL after finished I0213 18:21:25.896289 1 shared_informer.go:320] Caches are synced for ReplicaSet I0213 18:21:25.896527 1 shared_informer.go:320] Caches are synced for ephemeral I0213 18:21:25.897502 1 shared_informer.go:320] Caches are synced for GC I0213 18:21:25.897550 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:21:25.897991 1 shared_informer.go:320] Caches are synced for ClusterRoleAggregator I0213 18:21:25.905762 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:21:26.500355 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:21:27.009670 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="810.195527ms" I0213 18:21:27.019815 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="10.086623ms" I0213 18:21:27.020002 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="55.091µs" I0213 18:21:27.020204 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="820.837825ms" I0213 18:21:27.029280 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="9.022523ms" I0213 18:21:27.029397 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="43.55µs" I0213 18:21:27.040043 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="40.75µs" I0213 18:21:41.700142 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:21:41.708013 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:21:41.714088 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="72.131µs" I0213 18:21:41.717501 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="49.41µs" I0213 18:21:41.728953 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="45.251µs" I0213 18:21:41.740412 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="55.75µs" I0213 18:21:41.749836 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="57.391µs" I0213 18:21:45.170017 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="50.411µs" I0213 18:21:45.189655 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="6.747282ms" I0213 18:21:45.189734 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="29.35µs" I0213 18:21:45.210045 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="9.121454ms" I0213 18:21:45.210135 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="57.401µs" I0213 18:21:45.221303 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="7.443489ms" I0213 18:21:45.221377 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="37.15µs" I0213 18:21:45.845286 1 node_lifecycle_controller.go:1057] "Controller detected that some Nodes are Ready. Exiting master disruption mode" logger="node-lifecycle-controller" I0213 18:21:52.147129 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:22:02.389004 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:22:22.526398 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:24:14.807545 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:24:45.100938 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:25:05.529189 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:25:15.613437 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:25:18.682575 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-7459f86744" duration="24.1473ms" I0213 18:25:18.689570 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-7459f86744" duration="6.917295ms" I0213 18:25:18.689705 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-7459f86744" duration="59.811µs" I0213 18:25:18.693182 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-7459f86744" duration="37.851µs" I0213 18:25:18.725125 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-cainjector-55949d4888" duration="17.645062ms" I0213 18:25:18.731844 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-cainjector-55949d4888" duration="6.67198ms" I0213 18:25:18.731936 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-cainjector-55949d4888" duration="45.721µs" I0213 18:25:18.738483 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-cainjector-55949d4888" duration="41.81µs" I0213 18:25:18.775320 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-webhook-68cbf75c9c" duration="21.12731ms" I0213 18:25:18.784731 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-webhook-68cbf75c9c" duration="9.291894ms" I0213 18:25:18.784833 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-webhook-68cbf75c9c" duration="58.181µs" I0213 18:25:18.788517 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-webhook-68cbf75c9c" duration="56.991µs" I0213 18:25:18.791715 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="0s" I0213 18:25:18.803373 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:18.812069 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:18.812101 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:18.824855 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:21.111915 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="test-pvc-ns" I0213 18:25:25.113295 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-cainjector-55949d4888" duration="9.866103ms" I0213 18:25:25.113388 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-cainjector-55949d4888" duration="47.051µs" I0213 18:25:25.923406 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="issuers.cert-manager.io" I0213 18:25:25.923449 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="orders.acme.cert-manager.io" I0213 18:25:25.923469 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="certificates.cert-manager.io" I0213 18:25:25.923493 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="challenges.acme.cert-manager.io" I0213 18:25:25.923520 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="certificaterequests.cert-manager.io" I0213 18:25:25.923580 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:25:25.932939 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 18:25:25.951718 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:25:27.023765 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:25:27.033924 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:25:27.119319 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-7459f86744" duration="9.464506ms" I0213 18:25:27.119398 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-7459f86744" duration="37.051µs" I0213 18:25:35.128906 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:36.136882 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:36.237253 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:25:37.133923 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-webhook-68cbf75c9c" duration="61.061µs" I0213 18:25:43.146352 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-webhook-68cbf75c9c" duration="9.314865ms" I0213 18:25:43.146460 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/cert-manager-webhook-68cbf75c9c" duration="49.23µs" I0213 18:25:44.076041 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="17.910107ms" I0213 18:25:44.083287 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="7.20848ms" I0213 18:25:44.083411 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="56.591µs" I0213 18:25:44.090869 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="43.011µs" I0213 18:25:45.148585 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:46.157343 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:46.286362 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:46.501379 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:25:47.164016 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:47.167459 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="46.461µs" I0213 18:25:47.171995 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="1s" I0213 18:25:48.162566 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="59.851µs" I0213 18:25:50.166717 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="58.681µs" I0213 18:25:56.780460 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:25:57.040088 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 18:25:57.141052 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:25:58.183015 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="9.67577ms" I0213 18:25:58.183114 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="cert-manager/trust-manager-865b9c84ff" duration="54.201µs" I0213 18:26:00.137965 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-864c79545c" duration="22.508453ms" I0213 18:26:00.145426 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-864c79545c" duration="7.407043ms" I0213 18:26:00.145515 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-864c79545c" duration="37.371µs" I0213 18:26:00.148818 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-864c79545c" duration="36.92µs" I0213 18:26:00.183942 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-webhook-b678db645" duration="17.900096ms" I0213 18:26:00.190438 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-webhook-b678db645" duration="6.447767ms" I0213 18:26:00.190527 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-webhook-b678db645" duration="47.441µs" I0213 18:26:00.197275 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-webhook-b678db645" duration="39.11µs" I0213 18:26:04.204494 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-webhook-b678db645" duration="9.424286ms" I0213 18:26:04.204568 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-webhook-b678db645" duration="33.48µs" I0213 18:26:04.218782 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-864c79545c" duration="10.633616ms" I0213 18:26:04.218865 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-operator/tekton-operator-864c79545c" duration="33.87µs" E0213 18:26:15.382687 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:15.388860 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:15.399550 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:15.420001 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:15.460451 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:15.541406 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:15.702371 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:16.023027 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:16.663906 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:17.944046 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:18.194472 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:18.199916 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:18.210417 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:18.231519 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:18.272507 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:18.353435 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:18.514679 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:18.834975 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:19.009934 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:19.015214 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:19.025777 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:19.046810 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:19.087139 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:19.167302 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:19.327930 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:19.475635 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:19.648963 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:19.804295 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" E0213 18:26:19.810630 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" E0213 18:26:19.821924 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" E0213 18:26:19.843821 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" E0213 18:26:19.886144 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" E0213 18:26:19.905070 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 067e0bdc-50b1-4541-ada0-c83ffa2f90f7]" E0213 18:26:19.911513 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 067e0bdc-50b1-4541-ada0-c83ffa2f90f7]" E0213 18:26:19.922853 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 067e0bdc-50b1-4541-ada0-c83ffa2f90f7]" E0213 18:26:19.944386 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 067e0bdc-50b1-4541-ada0-c83ffa2f90f7]" E0213 18:26:19.967818 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" E0213 18:26:19.986266 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 067e0bdc-50b1-4541-ada0-c83ffa2f90f7]" E0213 18:26:20.005104 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-events-controller, uid: c97eee71-5bcb-4b1b-bd89-79db77a9d810]" E0213 18:26:20.011516 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-events-controller, uid: c97eee71-5bcb-4b1b-bd89-79db77a9d810]" E0213 18:26:20.022846 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-events-controller, uid: c97eee71-5bcb-4b1b-bd89-79db77a9d810]" E0213 18:26:20.044048 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-events-controller, uid: c97eee71-5bcb-4b1b-bd89-79db77a9d810]" E0213 18:26:20.067825 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 067e0bdc-50b1-4541-ada0-c83ffa2f90f7]" E0213 18:26:20.091466 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-events-controller, uid: c97eee71-5bcb-4b1b-bd89-79db77a9d810]" E0213 18:26:20.105264 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-leader-election, uid: 49064618-8066-4080-bb39-0695b1ec7b49]" E0213 18:26:20.111734 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-leader-election, uid: 49064618-8066-4080-bb39-0695b1ec7b49]" E0213 18:26:20.123055 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-leader-election, uid: 49064618-8066-4080-bb39-0695b1ec7b49]" E0213 18:26:20.129335 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" E0213 18:26:20.144575 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-leader-election, uid: 49064618-8066-4080-bb39-0695b1ec7b49]" E0213 18:26:20.173332 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-events-controller, uid: c97eee71-5bcb-4b1b-bd89-79db77a9d810]" E0213 18:26:20.186580 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-leader-election, uid: 49064618-8066-4080-bb39-0695b1ec7b49]" E0213 18:26:20.209302 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-info, uid: fb63b518-cb56-4e79-ba5c-353877fff76d]" E0213 18:26:20.217583 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-info, uid: fb63b518-cb56-4e79-ba5c-353877fff76d]" E0213 18:26:20.219245 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: b0fdc9bd-b9b5-47d1-8597-afebbe56d1c6]" E0213 18:26:20.225945 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: b0fdc9bd-b9b5-47d1-8597-afebbe56d1c6]" E0213 18:26:20.229518 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-info, uid: fb63b518-cb56-4e79-ba5c-353877fff76d]" E0213 18:26:20.229699 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 067e0bdc-50b1-4541-ada0-c83ffa2f90f7]" E0213 18:26:20.238283 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: b0fdc9bd-b9b5-47d1-8597-afebbe56d1c6]" E0213 18:26:20.251104 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-info, uid: fb63b518-cb56-4e79-ba5c-353877fff76d]" E0213 18:26:20.260494 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: b0fdc9bd-b9b5-47d1-8597-afebbe56d1c6]" E0213 18:26:20.267716 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-leader-election, uid: 49064618-8066-4080-bb39-0695b1ec7b49]" E0213 18:26:20.290038 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonInstallerSet, namespace: , name: pipeline-main-static-b6fd8, uid: cb26cd7b-014c-438b-ba5a-65d739ff709d]" E0213 18:26:20.292836 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-info, uid: fb63b518-cb56-4e79-ba5c-353877fff76d]" E0213 18:26:20.305261 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: b0fdc9bd-b9b5-47d1-8597-afebbe56d1c6]" E0213 18:26:20.334923 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-events-controller, uid: c97eee71-5bcb-4b1b-bd89-79db77a9d810]" E0213 18:26:20.374239 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-info, uid: fb63b518-cb56-4e79-ba5c-353877fff76d]" E0213 18:26:20.381790 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 22526c94-ee31-41e4-a212-554842ddc511]" E0213 18:26:20.387243 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: b0fdc9bd-b9b5-47d1-8597-afebbe56d1c6]" E0213 18:26:20.388695 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 22526c94-ee31-41e4-a212-554842ddc511]" E0213 18:26:20.388701 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-events-controller, uid: 7025e36a-c916-4aa6-9c4c-eed069e0f85f]" E0213 18:26:20.398119 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-events-controller, uid: 7025e36a-c916-4aa6-9c4c-eed069e0f85f]" E0213 18:26:20.402085 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 22526c94-ee31-41e4-a212-554842ddc511]" E0213 18:26:20.412593 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-events-controller, uid: 7025e36a-c916-4aa6-9c4c-eed069e0f85f]" E0213 18:26:20.424041 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 22526c94-ee31-41e4-a212-554842ddc511]" E0213 18:26:20.429513 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-leader-election, uid: 49064618-8066-4080-bb39-0695b1ec7b49]" E0213 18:26:20.433864 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-events-controller, uid: 7025e36a-c916-4aa6-9c4c-eed069e0f85f]" E0213 18:26:20.451475 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" E0213 18:26:20.465773 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 22526c94-ee31-41e4-a212-554842ddc511]" E0213 18:26:20.475083 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-events-controller, uid: 7025e36a-c916-4aa6-9c4c-eed069e0f85f]" E0213 18:26:20.504770 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonConfig." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonConfig, namespace: , name: config, uid: b20e5c33-5f32-4e8f-984c-5556026588d0]" E0213 18:26:20.536624 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-info, uid: fb63b518-cb56-4e79-ba5c-353877fff76d]" E0213 18:26:20.548321 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 22526c94-ee31-41e4-a212-554842ddc511]" E0213 18:26:20.549274 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: b0fdc9bd-b9b5-47d1-8597-afebbe56d1c6]" E0213 18:26:20.551460 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 067e0bdc-50b1-4541-ada0-c83ffa2f90f7]" E0213 18:26:20.556944 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-events-controller, uid: 7025e36a-c916-4aa6-9c4c-eed069e0f85f]" E0213 18:26:20.656467 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-events-controller, uid: c97eee71-5bcb-4b1b-bd89-79db77a9d810]" E0213 18:26:20.710588 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 22526c94-ee31-41e4-a212-554842ddc511]" E0213 18:26:20.719117 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-events-controller, uid: 7025e36a-c916-4aa6-9c4c-eed069e0f85f]" E0213 18:26:20.751018 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-leader-election, uid: 49064618-8066-4080-bb39-0695b1ec7b49]" E0213 18:26:20.756273 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonPipeline." logger="garbage-collector-controller" item="[operator.tekton.dev/v1alpha1/TektonPipeline, namespace: , name: pipeline, uid: 31545ebd-c4c5-4ec3-897d-4b9c34a5993e]" E0213 18:26:20.781738 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/Secret, namespace: tekton-pipelines, name: webhook-certs, uid: 7cbf3759-9652-4be1-bb9e-4a057b1a58ef]" E0213 18:26:20.788143 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/Secret, namespace: tekton-pipelines, name: webhook-certs, uid: 7cbf3759-9652-4be1-bb9e-4a057b1a58ef]" E0213 18:26:20.799782 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/Secret, namespace: tekton-pipelines, name: webhook-certs, uid: 7cbf3759-9652-4be1-bb9e-4a057b1a58ef]" E0213 18:26:20.821296 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/Secret, namespace: tekton-pipelines, name: webhook-certs, uid: 7cbf3759-9652-4be1-bb9e-4a057b1a58ef]" E0213 18:26:20.858474 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-info, uid: fb63b518-cb56-4e79-ba5c-353877fff76d]" E0213 18:26:20.863251 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/Secret, namespace: tekton-pipelines, name: webhook-certs, uid: 7cbf3759-9652-4be1-bb9e-4a057b1a58ef]" E0213 18:26:20.871132 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: b0fdc9bd-b9b5-47d1-8597-afebbe56d1c6]" E0213 18:26:20.887104 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-defaults, uid: dad113a6-4def-41ff-b7f5-e90a063b2c14]" E0213 18:26:20.894008 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-defaults, uid: dad113a6-4def-41ff-b7f5-e90a063b2c14]" E0213 18:26:20.894140 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-events, uid: 1ec00d0c-e509-46b8-8da3-95e72526caec]" E0213 18:26:20.900757 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: feature-flags, uid: bf242595-0063-4d02-9cd7-3473b2e8b0d7]" E0213 18:26:20.900800 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-events, uid: 1ec00d0c-e509-46b8-8da3-95e72526caec]" E0213 18:26:20.906131 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-defaults, uid: dad113a6-4def-41ff-b7f5-e90a063b2c14]" E0213 18:26:20.907110 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: feature-flags, uid: bf242595-0063-4d02-9cd7-3473b2e8b0d7]" E0213 18:26:20.928850 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-events, uid: 1ec00d0c-e509-46b8-8da3-95e72526caec]" E0213 18:26:20.954789 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: feature-flags, uid: bf242595-0063-4d02-9cd7-3473b2e8b0d7]" E0213 18:26:20.979373 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-defaults, uid: dad113a6-4def-41ff-b7f5-e90a063b2c14]" E0213 18:26:21.004893 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/Secret, namespace: tekton-pipelines, name: webhook-certs, uid: 7cbf3759-9652-4be1-bb9e-4a057b1a58ef]" E0213 18:26:21.029307 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-events, uid: 1ec00d0c-e509-46b8-8da3-95e72526caec]" E0213 18:26:21.056999 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: feature-flags, uid: bf242595-0063-4d02-9cd7-3473b2e8b0d7]" E0213 18:26:21.079889 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: pipelines-info, uid: f1148d87-ffdd-4e5f-bbdf-6dc80c489bc6]" E0213 18:26:21.103952 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-leader-election-controller, uid: a3d8b562-44f6-4460-9162-1394d4a76c87]" E0213 18:26:21.130136 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-defaults, uid: dad113a6-4def-41ff-b7f5-e90a063b2c14]" E0213 18:26:21.155381 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-pipelines-webhook, uid: 22526c94-ee31-41e4-a212-554842ddc511]" E0213 18:26:21.179583 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ServiceAccount, namespace: tekton-pipelines, name: tekton-events-controller, uid: 7025e36a-c916-4aa6-9c4c-eed069e0f85f]" E0213 18:26:21.204988 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-events, uid: 1ec00d0c-e509-46b8-8da3-95e72526caec]" E0213 18:26:21.229460 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: config-leader-election-events, uid: 172e8a3c-6e03-4aa3-8a16-3a64b0bc9038]" E0213 18:26:21.255001 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[v1/ConfigMap, namespace: tekton-pipelines, name: pipelines-info, uid: f1148d87-ffdd-4e5f-bbdf-6dc80c489bc6]" E0213 18:26:21.279977 1 garbagecollector.go:358] "error syncing item" err="unable to get REST mapping for operator.tekton.dev/v1alpha1/TektonInstallerSet." logger="garbage-collector-controller" item="[rbac.authorization.k8s.io/v1/Role, namespace: tekton-pipelines, name: tekton-pipelines-controller, uid: 83f3fe17-c005-4c65-b343-c0d3ba741849]" I0213 18:26:27.031255 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="verificationpolicies.tekton.dev" I0213 18:26:27.031293 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="tasks.tekton.dev" I0213 18:26:27.031318 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="pipelineruns.tekton.dev" I0213 18:26:27.031341 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="stepactions.tekton.dev" I0213 18:26:27.031371 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="resolutionrequests.resolution.tekton.dev" I0213 18:26:27.031383 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="taskruns.tekton.dev" I0213 18:26:27.031399 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="customruns.tekton.dev" I0213 18:26:27.031423 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="pipelines.tekton.dev" I0213 18:26:27.031504 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:26:27.150593 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 18:26:27.400217 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:26:27.552511 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-5dfff97774" duration="21.972619ms" I0213 18:26:27.571101 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-5dfff97774" duration="18.516814ms" I0213 18:26:27.571209 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-5dfff97774" duration="64.431µs" I0213 18:26:27.647170 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-f58b9fbcd" duration="18.89822ms" I0213 18:26:27.659024 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-f58b9fbcd" duration="11.797428ms" I0213 18:26:27.659150 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-f58b9fbcd" duration="72.781µs" I0213 18:26:27.747525 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="18.462003ms" I0213 18:26:27.755306 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="7.732113ms" I0213 18:26:27.755446 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="65.681µs" I0213 18:26:27.759055 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="63.311µs" I0213 18:26:27.847088 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-65f6889cdf" duration="18.272481ms" I0213 18:26:27.856825 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-65f6889cdf" duration="9.686294ms" I0213 18:26:27.856917 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-65f6889cdf" duration="62.241µs" I0213 18:26:27.860916 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-65f6889cdf" duration="46.71µs" I0213 18:26:27.953568 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-operator-proxy-webhook-64c9b98f89" duration="21.777466ms" I0213 18:26:27.965935 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-operator-proxy-webhook-64c9b98f89" duration="12.276676ms" I0213 18:26:27.966038 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-operator-proxy-webhook-64c9b98f89" duration="53.04µs" I0213 18:26:27.971020 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-operator-proxy-webhook-64c9b98f89" duration="67.761µs" I0213 18:26:28.232650 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:26:28.251451 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:26:30.258150 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-5dfff97774" duration="74.472µs" I0213 18:26:31.262739 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-f58b9fbcd" duration="66.011µs" I0213 18:26:33.275085 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="8.893951ms" I0213 18:26:33.275192 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="57.911µs" I0213 18:26:34.282115 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-65f6889cdf" duration="178.433µs" I0213 18:26:36.281766 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-operator-proxy-webhook-64c9b98f89" duration="5.875453ms" I0213 18:26:36.281847 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-operator-proxy-webhook-64c9b98f89" duration="33.68µs" I0213 18:26:37.677046 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:26:37.831736 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:26:37.841123 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:26:38.298102 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:26:38.307647 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:26:41.273012 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-5dfff97774" duration="10.032759ms" I0213 18:26:41.273117 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-5dfff97774" duration="61.971µs" I0213 18:26:43.279204 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-f58b9fbcd" duration="9.730534ms" I0213 18:26:43.279303 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-f58b9fbcd" duration="55.201µs" I0213 18:26:46.294866 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-65f6889cdf" duration="15.314293ms" I0213 18:26:46.294976 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-65f6889cdf" duration="63.691µs" W0213 18:26:52.848427 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:26:52.849079 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:26:53.310567 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:26:53.311211 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:26:53.437246 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-controller-8444b7b678" duration="16.99717ms" I0213 18:26:53.446356 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-controller-8444b7b678" duration="9.058894ms" I0213 18:26:53.446463 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-controller-8444b7b678" duration="60.851µs" I0213 18:26:53.453814 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-controller-8444b7b678" duration="53.531µs" I0213 18:26:53.539267 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-webhook-f64478468" duration="18.674266ms" I0213 18:26:53.548002 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-webhook-f64478468" duration="8.685158ms" I0213 18:26:53.548112 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-webhook-f64478468" duration="63.441µs" I0213 18:26:53.551992 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-webhook-f64478468" duration="70.012µs" I0213 18:26:53.636839 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-core-interceptors-7cc7cd5547" duration="17.564029ms" I0213 18:26:53.657034 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-core-interceptors-7cc7cd5547" duration="20.073949ms" I0213 18:26:53.657144 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-core-interceptors-7cc7cd5547" duration="58.581µs" I0213 18:26:56.337560 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-controller-8444b7b678" duration="9.867707ms" I0213 18:26:56.337665 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-controller-8444b7b678" duration="64.971µs" I0213 18:26:58.130655 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:26:58.238978 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="triggertemplates.triggers.tekton.dev" I0213 18:26:58.239023 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="eventlisteners.triggers.tekton.dev" I0213 18:26:58.239036 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="triggerbindings.triggers.tekton.dev" I0213 18:26:58.239055 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="interceptors.triggers.tekton.dev" I0213 18:26:58.239140 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="triggers.triggers.tekton.dev" I0213 18:26:58.239337 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:26:58.260998 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 18:26:58.339683 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:26:58.343664 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-webhook-f64478468" duration="10.209823ms" I0213 18:26:58.343765 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-webhook-f64478468" duration="52.29µs" I0213 18:26:58.361514 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:26:59.346951 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-core-interceptors-7cc7cd5547" duration="101.702µs" W0213 18:27:07.853649 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:27:07.854358 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:27:08.317463 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:27:08.318092 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:27:11.353058 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-core-interceptors-7cc7cd5547" duration="9.542661ms" I0213 18:27:11.353183 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-triggers-core-interceptors-7cc7cd5547" duration="80.191µs" I0213 18:27:13.139831 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="22.703041ms" I0213 18:27:13.155809 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="15.921863ms" I0213 18:27:13.155913 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="56.421µs" I0213 18:27:13.159436 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="87.742µs" I0213 18:27:16.380711 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="6.391882ms" I0213 18:27:16.380832 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="83.261µs" I0213 18:27:18.277186 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:27:18.650927 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-api-546b75cb88" duration="17.722532ms" I0213 18:27:18.660325 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-api-546b75cb88" duration="9.352409ms" I0213 18:27:18.660441 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-api-546b75cb88" duration="71.751µs" I0213 18:27:18.665558 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-api-546b75cb88" duration="47.73µs" I0213 18:27:18.716328 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-retention-policy-agent-68c6b7cf5d" duration="23.3136ms" I0213 18:27:18.725768 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-retention-policy-agent-68c6b7cf5d" duration="9.386829ms" I0213 18:27:18.725866 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-retention-policy-agent-68c6b7cf5d" duration="62.651µs" I0213 18:27:18.729911 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-retention-policy-agent-68c6b7cf5d" duration="59.121µs" I0213 18:27:18.812199 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-watcher-6c9b786b7" duration="18.510184ms" I0213 18:27:18.824934 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-watcher-6c9b786b7" duration="12.692772ms" I0213 18:27:18.825109 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-watcher-6c9b786b7" duration="49.571µs" I0213 18:27:18.825165 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-watcher-6c9b786b7" duration="29.42µs" I0213 18:27:18.832548 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-watcher-6c9b786b7" duration="109.952µs" I0213 18:27:22.400111 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-api-546b75cb88" duration="74.611µs" W0213 18:27:22.862404 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:27:22.863100 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:27:23.322404 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:27:23.323054 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:27:24.418510 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-retention-policy-agent-68c6b7cf5d" duration="9.040063ms" I0213 18:27:24.418689 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-retention-policy-agent-68c6b7cf5d" duration="114.392µs" I0213 18:27:26.427297 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-watcher-6c9b786b7" duration="9.309044ms" I0213 18:27:26.427418 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-watcher-6c9b786b7" duration="76.002µs" I0213 18:27:28.632188 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:27:37.868217 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:27:37.868982 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:27:38.327158 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:27:38.327840 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:27:38.781195 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:27:48.975382 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-api-546b75cb88" duration="67.641µs" I0213 18:27:48.998510 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-api-546b75cb88" duration="9.457406ms" I0213 18:27:48.998639 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-results-api-546b75cb88" duration="81.781µs" W0213 18:27:52.877500 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:27:52.878166 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:27:53.162378 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="21.160065ms" I0213 18:27:53.171269 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="8.824847ms" I0213 18:27:53.171382 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="66.401µs" I0213 18:27:53.175496 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="64.851µs" W0213 18:27:53.332316 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:27:53.332945 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:27:55.488093 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="77.201µs" I0213 18:27:55.509864 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="9.69012ms" I0213 18:27:55.509952 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="45.24µs" I0213 18:27:58.351510 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="extensions.dashboard.tekton.dev" I0213 18:27:58.351580 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:27:58.373310 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 18:27:58.373356 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:27:58.452131 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:27:59.389239 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:28:00.392805 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-7c567d6b77" duration="17.32887ms" I0213 18:28:00.402877 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-7c567d6b77" duration="10.005224ms" I0213 18:28:00.402964 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-7c567d6b77" duration="53.971µs" I0213 18:28:00.407171 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-7c567d6b77" duration="62.6µs" I0213 18:28:00.410412 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-controller-5cff8fdbb9" duration="24.815128ms" I0213 18:28:00.420406 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-controller-5cff8fdbb9" duration="9.922593ms" I0213 18:28:00.420507 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-controller-5cff8fdbb9" duration="55.221µs" I0213 18:28:00.431323 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-controller-5cff8fdbb9" duration="59.791µs" I0213 18:28:00.454659 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-watcher-6f9fd4cbb5" duration="15.529214ms" I0213 18:28:00.464986 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-watcher-6f9fd4cbb5" duration="10.266958ms" I0213 18:28:00.465129 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-watcher-6f9fd4cbb5" duration="98.321µs" I0213 18:28:00.473012 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-watcher-6f9fd4cbb5" duration="71.641µs" I0213 18:28:00.506318 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-webhook-66c64db47f" duration="18.337004ms" I0213 18:28:00.513381 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-webhook-66c64db47f" duration="7.019671ms" I0213 18:28:00.513459 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-webhook-66c64db47f" duration="48.511µs" I0213 18:28:00.513498 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-webhook-66c64db47f" duration="26.741µs" I0213 18:28:00.517136 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-webhook-66c64db47f" duration="35.101µs" I0213 18:28:01.308476 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="0s" I0213 18:28:01.316085 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:01.327110 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:01.327137 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:01.343739 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:01.523393 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-7c567d6b77" duration="9.197802ms" I0213 18:28:01.523515 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-7c567d6b77" duration="70.821µs" I0213 18:28:01.553361 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="20.340923ms" I0213 18:28:01.561406 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="7.988785ms" I0213 18:28:01.561517 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="61.371µs" I0213 18:28:01.951527 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="dex/dex-58b5779f4" duration="47.478534ms" E0213 18:28:01.981148 1 garbagecollector.go:360] "Unhandled Error" err="error syncing item &garbagecollector.node{identity:garbagecollector.objectReference{OwnerReference:v1.OwnerReference{APIVersion:\"rbac.authorization.k8s.io/v1\", Kind:\"ClusterRole\", Name:\"tekton-dashboard-backend-view\", UID:\"6d86a841-2659-4583-b3f9-68542b32d955\", Controller:(*bool)(nil), BlockOwnerDeletion:(*bool)(nil)}, Namespace:\"\"}, dependentsLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:1}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, dependents:map[*garbagecollector.node]struct {}{}, deletingDependents:false, deletingDependentsLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, beingDeleted:false, beingDeletedLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, virtual:false, virtualLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, owners:[]v1.OwnerReference{v1.OwnerReference{APIVersion:\"operator.tekton.dev/v1alpha1\", Kind:\"TektonInstallerSet\", Name:\"dashboard-main-static-qmxfv\", UID:\"ef783610-3445-43b0-87d0-c22f06366122\", Controller:(*bool)(0xc0032a7ad0), BlockOwnerDeletion:(*bool)(0xc0032a7ad1)}}}: clusterroles.rbac.authorization.k8s.io \"tekton-dashboard-backend-view\" not found" logger="UnhandledError" I0213 18:28:01.986192 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="dex/dex-58b5779f4" duration="34.610469ms" I0213 18:28:01.986265 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="dex/dex-58b5779f4" duration="27.14µs" I0213 18:28:02.019088 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="70.601µs" E0213 18:28:02.058592 1 garbagecollector.go:360] "Unhandled Error" err="error syncing item &garbagecollector.node{identity:garbagecollector.objectReference{OwnerReference:v1.OwnerReference{APIVersion:\"rbac.authorization.k8s.io/v1\", Kind:\"ClusterRole\", Name:\"tekton-dashboard-tenant-view\", UID:\"252c5fce-a768-4c91-95a8-fa390f3d4de2\", Controller:(*bool)(nil), BlockOwnerDeletion:(*bool)(nil)}, Namespace:\"\"}, dependentsLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:1}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, dependents:map[*garbagecollector.node]struct {}{}, deletingDependents:false, deletingDependentsLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, beingDeleted:false, beingDeletedLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, virtual:false, virtualLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, owners:[]v1.OwnerReference{v1.OwnerReference{APIVersion:\"operator.tekton.dev/v1alpha1\", Kind:\"TektonInstallerSet\", Name:\"dashboard-main-static-qmxfv\", UID:\"ef783610-3445-43b0-87d0-c22f06366122\", Controller:(*bool)(0xc0032a7900), BlockOwnerDeletion:(*bool)(0xc0032a7901)}}}: clusterroles.rbac.authorization.k8s.io \"tekton-dashboard-tenant-view\" not found" logger="UnhandledError" E0213 18:28:02.202859 1 garbagecollector.go:360] "Unhandled Error" err="error syncing item &garbagecollector.node{identity:garbagecollector.objectReference{OwnerReference:v1.OwnerReference{APIVersion:\"v1\", Kind:\"ServiceAccount\", Name:\"tekton-dashboard\", UID:\"55c100d1-915a-42ae-b939-b36c7e2c96e6\", Controller:(*bool)(nil), BlockOwnerDeletion:(*bool)(nil)}, Namespace:\"tekton-pipelines\"}, dependentsLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:1}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, dependents:map[*garbagecollector.node]struct {}{}, deletingDependents:false, deletingDependentsLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, beingDeleted:false, beingDeletedLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, virtual:false, virtualLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:atomic.Int32{_:atomic.noCopy{}, v:0}, readerWait:atomic.Int32{_:atomic.noCopy{}, v:0}}, owners:[]v1.OwnerReference{v1.OwnerReference{APIVersion:\"operator.tekton.dev/v1alpha1\", Kind:\"TektonInstallerSet\", Name:\"dashboard-main-static-qmxfv\", UID:\"ef783610-3445-43b0-87d0-c22f06366122\", Controller:(*bool)(0xc004081207), BlockOwnerDeletion:(*bool)(0xc004081208)}}}: serviceaccounts \"tekton-dashboard\" not found" logger="UnhandledError" E0213 18:28:02.270045 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:28:02.512308 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="5.988596ms" I0213 18:28:02.512408 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="52.9µs" I0213 18:28:02.523809 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="73.861µs" I0213 18:28:02.532343 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-remote-resolvers-dbc8f9b75" duration="58.151µs" I0213 18:28:02.626306 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kind-registry/registry-5b599f9d9d" duration="21.799284ms" I0213 18:28:02.633453 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kind-registry/registry-5b599f9d9d" duration="7.097573ms" I0213 18:28:02.633555 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kind-registry/registry-5b599f9d9d" duration="53.721µs" I0213 18:28:02.636983 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kind-registry/registry-5b599f9d9d" duration="49.84µs" I0213 18:28:02.768907 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="9.878013ms" I0213 18:28:02.769037 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="87.722µs" W0213 18:28:03.089532 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:28:03.090537 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:28:03.091297 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:28:03.091329 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:28:03.139897 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="71.711µs" I0213 18:28:03.542253 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="74.761µs" I0213 18:28:03.542489 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-56b4589864" duration="23.63441ms" I0213 18:28:03.548666 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="181.162µs" I0213 18:28:03.551987 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-56b4589864" duration="9.463216ms" I0213 18:28:03.552100 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-56b4589864" duration="72.931µs" I0213 18:28:03.559727 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-56b4589864" duration="54.561µs" I0213 18:28:03.568094 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-dashboard-56b8f9f7d4" duration="7.731µs" I0213 18:28:04.530365 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-controller-5cff8fdbb9" duration="61.951µs" I0213 18:28:04.554942 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-controller-5cff8fdbb9" duration="11.12189ms" I0213 18:28:04.555014 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-controller-5cff8fdbb9" duration="42.441µs" I0213 18:28:04.566591 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-56b4589864" duration="9.202672ms" I0213 18:28:04.566749 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-56b4589864" duration="82.211µs" I0213 18:28:04.596859 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="21.692623ms" I0213 18:28:04.607246 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="10.323309ms" I0213 18:28:04.607549 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="70.931µs" I0213 18:28:04.935295 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="77.992µs" I0213 18:28:05.540213 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="59.251µs" I0213 18:28:05.549030 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-chains-controller-67b58566c9" duration="53.041µs" W0213 18:28:06.165827 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:28:06.166503 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:28:06.167165 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:28:06.167190 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:28:06.550839 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-watcher-6f9fd4cbb5" duration="50.491µs" I0213 18:28:07.559118 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-webhook-66c64db47f" duration="6.123238ms" I0213 18:28:07.559301 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-webhook-66c64db47f" duration="129.112µs" W0213 18:28:07.883248 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:28:07.883952 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:28:08.338642 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:28:08.339273 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:28:09.584405 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:28:11.408441 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:28:11.409048 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:28:11.409673 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:28:11.409701 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:28:18.566817 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-watcher-6f9fd4cbb5" duration="9.801122ms" I0213 18:28:18.566911 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="pipelines-as-code/pipelines-as-code-watcher-6f9fd4cbb5" duration="44.221µs" W0213 18:28:18.980853 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:28:18.981496 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:28:18.982114 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:28:18.982141 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:28:22.893797 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:28:22.894580 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:28:23.344953 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:28:23.345744 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:28:23.604422 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:24.612680 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:26.616316 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="dex/dex-58b5779f4" duration="41.241µs" I0213 18:28:27.627645 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kind-registry/registry-5b599f9d9d" duration="9.212003ms" I0213 18:28:27.627729 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kind-registry/registry-5b599f9d9d" duration="41.071µs" I0213 18:28:27.887539 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="21.587751ms" I0213 18:28:27.904739 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="17.141487ms" I0213 18:28:27.904838 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="54.25µs" I0213 18:28:28.382065 1 shared_informer.go:313] Waiting for caches to sync for garbage collector W0213 18:28:28.459079 1 shared_informer.go:597] resyncPeriod 20h19m4.120317207s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459158 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="signingkeies.dex.coreos.com" W0213 18:28:28.459192 1 shared_informer.go:597] resyncPeriod 20h30m10.211455725s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459207 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="devicetokens.dex.coreos.com" W0213 18:28:28.459223 1 shared_informer.go:597] resyncPeriod 13h26m59.854450487s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459243 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="oauth2clients.dex.coreos.com" W0213 18:28:28.459252 1 shared_informer.go:597] resyncPeriod 18h49m47.162636661s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459262 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="authrequests.dex.coreos.com" W0213 18:28:28.459285 1 shared_informer.go:597] resyncPeriod 16h0m3.039383697s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459300 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="passwords.dex.coreos.com" W0213 18:28:28.459319 1 shared_informer.go:597] resyncPeriod 20h21m9.005966168s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459332 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="devicerequests.dex.coreos.com" W0213 18:28:28.459345 1 shared_informer.go:597] resyncPeriod 17h59m27.516359908s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459359 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="offlinesessionses.dex.coreos.com" W0213 18:28:28.459370 1 shared_informer.go:597] resyncPeriod 13h0m40.80822697s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459387 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="connectors.dex.coreos.com" W0213 18:28:28.459404 1 shared_informer.go:597] resyncPeriod 12h26m25.550345844s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459418 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="refreshtokens.dex.coreos.com" I0213 18:28:28.459447 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="authcodes.dex.coreos.com" W0213 18:28:28.459455 1 shared_informer.go:597] resyncPeriod 16h3m37.803328129s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:28.459473 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="repositories.pipelinesascode.tekton.dev" I0213 18:28:28.459566 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:28:29.482364 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:28:29.560507 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:28:29.624231 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:29.730785 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:28:30.632070 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:32.388859 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="24.457451ms" I0213 18:28:32.411456 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-background-controller-5f9f9b856f" duration="22.397621ms" I0213 18:28:32.415087 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="26.167035ms" I0213 18:28:32.415263 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="53.491µs" I0213 18:28:32.415327 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="36.271µs" I0213 18:28:32.425972 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-cleanup-controller-bfd46c7d6" duration="7.241734ms" I0213 18:28:32.426044 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-cleanup-controller-bfd46c7d6" duration="39.99µs" I0213 18:28:32.429636 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-background-controller-5f9f9b856f" duration="18.094599ms" I0213 18:28:32.429720 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-background-controller-5f9f9b856f" duration="42.731µs" I0213 18:28:32.451179 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-reports-controller-7dcdc4cc4d" duration="8.777316ms" I0213 18:28:32.451886 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-reports-controller-7dcdc4cc4d" duration="60.931µs" I0213 18:28:32.645381 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:33.653785 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:34.668810 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="13.348431ms" I0213 18:28:34.668883 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="29.3µs" W0213 18:28:36.594205 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:28:36.594836 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:28:36.595464 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:28:36.595495 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:28:36.665497 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="62.871µs" I0213 18:28:37.631990 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="dex/dex-58b5779f4" duration="10.224887ms" I0213 18:28:37.632087 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="dex/dex-58b5779f4" duration="57.461µs" W0213 18:28:37.899894 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:28:37.900574 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:28:38.350107 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:28:38.350738 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:28:38.677417 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="68.931µs" I0213 18:28:38.696092 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-background-controller-5f9f9b856f" duration="9.232023ms" I0213 18:28:38.696227 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-background-controller-5f9f9b856f" duration="79.982µs" I0213 18:28:41.686905 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="94.601µs" I0213 18:28:50.708252 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:50.719400 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="71.441µs" I0213 18:28:50.746522 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="14.667361ms" I0213 18:28:50.746672 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kyverno/kyverno-admission-controller-877795fc7" duration="73.331µs" I0213 18:28:51.711489 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:51.715220 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:51.863315 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:52.718833 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" I0213 18:28:52.731026 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="tekton-pipelines/tekton-chains-signing-secret" delay="1s" W0213 18:28:52.909038 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:28:52.909769 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:28:53.355030 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:28:53.355638 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:28:59.493510 1 shared_informer.go:313] Waiting for caches to sync for garbage collector W0213 18:28:59.566217 1 shared_informer.go:597] resyncPeriod 12h53m42.660495703s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566280 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="components.appstudio.redhat.com" W0213 18:28:59.566298 1 shared_informer.go:597] resyncPeriod 16h34m30.532128378s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566309 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="componentdetectionqueries.appstudio.redhat.com" W0213 18:28:59.566349 1 shared_informer.go:597] resyncPeriod 20h30m38.441921839s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566365 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="snapshotenvironmentbindings.appstudio.redhat.com" W0213 18:28:59.566386 1 shared_informer.go:597] resyncPeriod 21h48m0.165160301s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566395 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="promotionruns.appstudio.redhat.com" W0213 18:28:59.566422 1 shared_informer.go:597] resyncPeriod 13h49m56.144469797s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566436 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="snapshots.appstudio.redhat.com" W0213 18:28:59.566451 1 shared_informer.go:597] resyncPeriod 20h31m28.680816982s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566473 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="namespacedgeneratingpolicies.policies.kyverno.io" W0213 18:28:59.566490 1 shared_informer.go:597] resyncPeriod 20h1m58.772490782s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566508 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="updaterequests.kyverno.io" I0213 18:28:59.566547 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="namespaceddeletingpolicies.policies.kyverno.io" W0213 18:28:59.566558 1 shared_informer.go:597] resyncPeriod 15h45m44.607442037s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566573 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="namespacedvalidatingpolicies.policies.kyverno.io" W0213 18:28:59.566631 1 shared_informer.go:597] resyncPeriod 19h51m33.287351944s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566661 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="cleanuppolicies.kyverno.io" W0213 18:28:59.566678 1 shared_informer.go:597] resyncPeriod 13h3m10.022075225s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566689 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="namespacedmutatingpolicies.policies.kyverno.io" W0213 18:28:59.566702 1 shared_informer.go:597] resyncPeriod 20h59m5.608204606s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566718 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="policyexceptions.policies.kyverno.io" W0213 18:28:59.566739 1 shared_informer.go:597] resyncPeriod 20h7m4.937577298s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566753 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="namespacedimagevalidatingpolicies.policies.kyverno.io" W0213 18:28:59.566766 1 shared_informer.go:597] resyncPeriod 12h11m41.994575378s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566777 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="policyreports.wgpolicyk8s.io" W0213 18:28:59.566787 1 shared_informer.go:597] resyncPeriod 17h48m59.103749099s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566825 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="environments.appstudio.redhat.com" W0213 18:28:59.566838 1 shared_informer.go:597] resyncPeriod 20h30m41.880142814s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566851 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="applications.appstudio.redhat.com" W0213 18:28:59.566860 1 shared_informer.go:597] resyncPeriod 20h34m57.619795058s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566870 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="deploymenttargets.appstudio.redhat.com" I0213 18:28:59.566882 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="policyexceptions.kyverno.io" W0213 18:28:59.566889 1 shared_informer.go:597] resyncPeriod 16h14m16.748508877s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566903 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="policies.kyverno.io" W0213 18:28:59.566911 1 shared_informer.go:597] resyncPeriod 18h57m0.022098669s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566920 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="ephemeralreports.reports.kyverno.io" W0213 18:28:59.566945 1 shared_informer.go:597] resyncPeriod 19h29m4.370993273s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:28:59.566959 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="deploymenttargetclaims.appstudio.redhat.com" I0213 18:28:59.567188 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:29:00.339889 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:29:00.694676 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:29:00.767872 1 shared_informer.go:320] Caches are synced for resource quota W0213 18:29:07.923060 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:29:07.924039 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:29:08.360942 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:29:08.361631 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:29:08.723696 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="release-service/release-service-controller-manager-dfc8b9975" duration="27.120819ms" I0213 18:29:08.734122 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="release-service/release-service-controller-manager-dfc8b9975" duration="10.299658ms" I0213 18:29:08.734267 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="release-service/release-service-controller-manager-dfc8b9975" duration="53.191µs" I0213 18:29:08.742544 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="release-service/release-service-controller-manager-dfc8b9975" duration="52.961µs" I0213 18:29:10.476551 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:29:14.773204 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="release-service/release-service-controller-manager-dfc8b9975" duration="62.281µs" I0213 18:29:15.096514 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="build-service/build-service-controller-manager-7955844657" duration="22.494083ms" I0213 18:29:15.105664 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="build-service/build-service-controller-manager-7955844657" duration="9.101621ms" I0213 18:29:15.105740 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="build-service/build-service-controller-manager-7955844657" duration="35.36µs" I0213 18:29:15.105784 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="build-service/build-service-controller-manager-7955844657" duration="21.411µs" I0213 18:29:15.113214 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="build-service/build-service-controller-manager-7955844657" duration="53.191µs" I0213 18:29:18.783356 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="build-service/build-service-controller-manager-7955844657" duration="48.491µs" I0213 18:29:20.834972 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:29:22.926862 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="integration-service/integration-service-controller-manager-f5668ffc8" duration="24.546032ms" I0213 18:29:22.935727 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="integration-service/integration-service-controller-manager-f5668ffc8" duration="8.720315ms" I0213 18:29:22.935821 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="integration-service/integration-service-controller-manager-f5668ffc8" duration="38.35µs" W0213 18:29:22.936486 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:29:22.937284 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:29:22.944486 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="integration-service/integration-service-controller-manager-f5668ffc8" duration="47.37µs" W0213 18:29:23.365716 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:29:23.366357 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:29:23.781982 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="namespace-lister/namespace-lister-584d4574c4" duration="21.911724ms" I0213 18:29:23.799838 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="namespace-lister/namespace-lister-584d4574c4" duration="17.799395ms" I0213 18:29:23.814347 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="namespace-lister/namespace-lister-584d4574c4" duration="14.464477ms" I0213 18:29:23.814445 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="namespace-lister/namespace-lister-584d4574c4" duration="53.2µs" W0213 18:29:23.930422 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:29:23.931220 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:29:23.931914 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:29:23.931942 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:29:24.885587 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="24.500361ms" I0213 18:29:24.905121 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="19.458259ms" I0213 18:29:24.905219 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="51.161µs" I0213 18:29:25.787351 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="release-service/release-service-controller-manager-dfc8b9975" duration="11.269132ms" I0213 18:29:25.787439 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="release-service/release-service-controller-manager-dfc8b9975" duration="42.301µs" I0213 18:29:28.813182 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="integration-service/integration-service-controller-manager-f5668ffc8" duration="43.141µs" I0213 18:29:29.796692 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="build-service/build-service-controller-manager-7955844657" duration="9.713819ms" I0213 18:29:29.796771 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="build-service/build-service-controller-manager-7955844657" duration="36.521µs" I0213 18:29:29.817216 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="namespace-lister/namespace-lister-584d4574c4" duration="75.791µs" I0213 18:29:30.706035 1 shared_informer.go:313] Waiting for caches to sync for garbage collector W0213 18:29:30.773757 1 shared_informer.go:597] resyncPeriod 15h42m51.186131608s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:29:30.773807 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="componentgroups.appstudio.redhat.com" W0213 18:29:30.773862 1 shared_informer.go:597] resyncPeriod 19h39m10.786805683s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:29:30.773890 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="releases.appstudio.redhat.com" W0213 18:29:30.773910 1 shared_informer.go:597] resyncPeriod 18h55m9.750955167s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:29:30.773920 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="internalservicesconfigs.appstudio.redhat.com" W0213 18:29:30.773938 1 shared_informer.go:597] resyncPeriod 15h50m50.232715705s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:29:30.773949 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="releaseplans.appstudio.redhat.com" W0213 18:29:30.773958 1 shared_informer.go:597] resyncPeriod 14h48m21.454255708s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:29:30.773976 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="enterprisecontractpolicies.appstudio.redhat.com" W0213 18:29:30.773992 1 shared_informer.go:597] resyncPeriod 15h44m57.182561545s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:29:30.774001 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="releaseplanadmissions.appstudio.redhat.com" W0213 18:29:30.774014 1 shared_informer.go:597] resyncPeriod 20h48m36.266528879s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:29:30.774024 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="integrationtestscenarios.appstudio.redhat.com" I0213 18:29:30.774046 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="internalrequests.appstudio.redhat.com" W0213 18:29:30.774057 1 shared_informer.go:597] resyncPeriod 16h26m16.536122838s is smaller than resyncCheckPeriod 21h53m47.801145158s and the informer has already started. Changing it to 21h53m47.801145158s I0213 18:29:30.774069 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="releaseserviceconfigs.appstudio.redhat.com" I0213 18:29:30.774184 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:29:30.830067 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="namespace-lister/namespace-lister-584d4574c4" duration="7.01046ms" I0213 18:29:30.830164 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="namespace-lister/namespace-lister-584d4574c4" duration="52.081µs" I0213 18:29:31.071547 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:29:31.774489 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:29:31.806357 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:29:37.845771 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="53.081µs" W0213 18:29:37.941881 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:29:37.942505 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:29:38.370595 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:29:38.371245 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:29:39.827433 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="integration-service/integration-service-controller-manager-f5668ffc8" duration="9.112173ms" I0213 18:29:39.827656 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="integration-service/integration-service-controller-manager-f5668ffc8" duration="72.501µs" I0213 18:29:39.853324 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="58.441µs" I0213 18:29:45.872538 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="58.241µs" I0213 18:29:46.876969 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="60.941µs" I0213 18:29:51.518115 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:29:52.951858 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:29:52.952499 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:29:53.376385 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:29:53.377050 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:29:54.909665 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="90.051µs" W0213 18:30:07.957709 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:30:07.958371 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:30:08.076754 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:30:08.077416 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:30:08.078045 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:30:08.078071 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:30:08.381045 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:30:08.381931 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:30:11.692655 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:30:21.758344 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:30:22.968266 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:30:22.968969 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:30:23.386592 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:30:23.387243 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:30:25.929183 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="9.484228ms" I0213 18:30:25.929277 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="48.97µs" W0213 18:30:37.974176 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:30:37.974854 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:30:38.391799 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:30:38.392382 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:30:39.126482 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:30:39.127137 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:30:39.127752 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:30:39.127778 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:30:42.321880 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:30:47.039819 1 job_controller.go:598] "enqueueing job" logger="job-controller" key="cert-manager/cert-manager-startupapicheck" delay="0s" W0213 18:30:52.984713 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:30:52.985378 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:30:53.396269 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:30:53.396923 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:30:55.938929 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="8.983581ms" I0213 18:30:55.939030 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="konflux-ui/proxy-777bd9fd66" duration="51.201µs" I0213 18:31:01.287233 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="image-controller/image-controller-controller-manager-d76778956" duration="31.819617ms" I0213 18:31:01.294239 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="image-controller/image-controller-controller-manager-d76778956" duration="6.936883ms" I0213 18:31:01.294324 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="image-controller/image-controller-controller-manager-d76778956" duration="39.52µs" I0213 18:31:01.298114 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="image-controller/image-controller-controller-manager-d76778956" duration="38.001µs" I0213 18:31:01.790455 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="imagerepositories.appstudio.redhat.com" I0213 18:31:01.790545 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 18:31:01.825466 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 18:31:02.826036 1 shared_informer.go:320] Caches are synced for garbage collector I0213 18:31:02.890856 1 shared_informer.go:320] Caches are synced for resource quota I0213 18:31:02.916887 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:31:05.088552 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="image-controller/image-controller-controller-manager-d76778956" duration="45.62µs" W0213 18:31:07.989697 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:31:07.990332 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:31:08.400457 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:31:08.401074 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:31:13.072488 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:31:13.180558 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:31:13.181272 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:31:13.181883 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:31:13.181915 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:31:16.103775 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="image-controller/image-controller-controller-manager-d76778956" duration="10.192197ms" I0213 18:31:16.103964 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="image-controller/image-controller-controller-manager-d76778956" duration="64.491µs" I0213 18:31:16.681790 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="24.148054ms" I0213 18:31:16.689676 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="7.833735ms" I0213 18:31:16.689747 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="31.07µs" I0213 18:31:16.693096 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="41.5µs" I0213 18:31:19.135957 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="6.827712ms" I0213 18:31:19.136031 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="35.081µs" I0213 18:31:19.161593 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="18.246194ms" I0213 18:31:19.170235 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="7.545411ms" I0213 18:31:19.170316 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="32.181µs" I0213 18:31:19.792350 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="46.461µs" I0213 18:31:20.138745 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="45.641µs" I0213 18:31:20.144085 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-58b95dcb75" duration="45.55µs" W0213 18:31:22.999808 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:31:23.000484 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:31:23.405239 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:31:23.405877 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:31:23.420208 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:31:38.006121 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:31:38.006845 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:31:38.409833 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:31:38.410431 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:31:53.016096 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:31:53.016753 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:31:53.414764 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:31:53.415420 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:32:04.407316 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:32:07.275288 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:32:07.275898 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:32:07.276508 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:32:07.276536 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:32:08.021578 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:32:08.022268 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:32:08.420048 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:32:08.420685 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:32:14.597848 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:32:18.291156 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="43.06µs" W0213 18:32:23.031377 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:32:23.032020 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:32:23.425152 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:32:23.425763 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:32:38.036839 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:32:38.037478 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:32:38.430218 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:32:38.430865 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:32:48.397279 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:32:48.398180 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:32:48.398868 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:32:48.398897 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:32:53.047364 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:32:53.048062 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:32:53.435412 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:32:53.436057 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:32:55.065958 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:33:05.216542 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:33:08.052143 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:33:08.052826 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:33:08.441078 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:33:08.441765 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:33:22.258500 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:33:22.259201 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:33:22.259838 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:33:22.259863 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:33:23.062115 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:33:23.062794 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:33:23.451407 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:33:23.452059 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:33:38.067428 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:33:38.068074 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:33:38.456172 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:33:38.456776 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:33:48.532174 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="51.71µs" W0213 18:33:53.077537 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:33:53.078212 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:33:53.461096 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:33:53.461734 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:33:56.285858 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:34:06.661087 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:34:08.082712 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:34:08.083355 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:34:08.465491 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:34:08.466087 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:34:18.551878 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:34:18.552721 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:34:18.553292 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:34:18.553318 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:34:23.093370 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:34:23.094056 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:34:23.470219 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:34:23.470879 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:34:38.098370 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:34:38.099079 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:34:38.475251 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:34:38.475860 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:34:53.108634 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:34:53.109341 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:34:53.480532 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:34:53.481218 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:35:02.437695 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:35:02.438327 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:35:02.438959 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:35:02.438984 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:35:08.120571 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:35:08.121260 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:35:08.485317 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:35:08.485965 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:35:23.130452 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:35:23.131125 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:35:23.489743 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:35:23.490353 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:35:28.420756 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:35:38.135184 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:35:38.135892 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:35:38.495020 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:35:38.495693 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:35:38.510954 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:35:39.577553 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:35:39.578181 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:35:39.579085 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:35:39.579127 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:35:48.595330 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:35:53.145518 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:35:53.146187 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:35:53.500563 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:35:53.501199 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:36:08.151115 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:36:08.151803 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:36:08.505929 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:36:08.506569 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:36:18.928193 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="53.121µs" W0213 18:36:23.162028 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:36:23.162744 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:36:23.510714 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:36:23.511336 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:36:24.402288 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:36:24.403083 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:36:24.403753 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:36:24.403782 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:36:38.167412 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:36:38.168116 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:36:38.515677 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:36:38.516303 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:36:49.884164 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:36:53.178457 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:36:53.179138 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:36:53.520459 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:36:53.521120 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:37:00.154430 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:37:08.183562 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:37:08.184338 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:37:08.524987 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:37:08.525587 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:37:19.096053 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="47.12µs" W0213 18:37:21.657636 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:37:21.658241 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:37:21.658849 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:37:21.658874 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:37:23.193843 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:37:23.194590 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:37:23.530218 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:37:23.530832 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:37:30.624733 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:37:38.201340 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:37:38.202048 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:37:38.534843 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:37:38.535428 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:37:40.994069 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:37:53.211923 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:37:53.212591 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:37:53.540780 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:37:53.541397 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:37:54.388972 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:37:54.389765 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:37:54.390437 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:37:54.390467 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:38:08.217918 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:38:08.218571 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:38:08.545733 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:38:08.546449 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:38:12.057165 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:38:22.205730 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:38:23.227795 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:38:23.228422 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:38:23.550304 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:38:23.550922 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:38:25.217566 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:38:25.218250 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:38:25.218862 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:38:25.218888 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:38:32.267206 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:38:38.233479 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:38:38.234362 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:38:38.555089 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:38:38.555750 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:38:42.443986 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:38:44.000575 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="plan-and-admission-managed" I0213 18:38:44.395519 1 stateful_set.go:466] "StatefulSet has been deleted" logger="statefulset-controller" key="tenant-dev-jydu/affinity-assistant-aecc0e3c43" I0213 18:38:44.909964 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="neg-rp-managed" I0213 18:38:45.019089 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="block-rp-managed" I0213 18:38:50.984830 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="block-rp-dev-frap" I0213 18:38:51.181560 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="neg-rp-dev-ipmc" I0213 18:38:52.479476 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:38:53.251405 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:38:53.252328 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:38:53.560362 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:38:53.561080 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:39:02.515778 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:39:08.257008 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:39:08.257787 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:39:08.564647 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:39:08.565240 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:39:12.810454 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:39:19.498594 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="62.301µs" W0213 18:39:20.751922 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:39:20.752565 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:39:20.753204 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:39:20.753229 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:39:22.882737 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:39:23.267570 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:39:23.268348 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:39:23.569163 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:39:23.569838 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:39:33.164703 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:39:38.273318 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:39:38.274006 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:39:38.574184 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:39:38.574840 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:39:43.296309 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:39:53.287253 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:39:53.287936 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:39:53.579699 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:39:53.580262 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:40:08.293226 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:40:08.293912 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:40:08.585134 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:40:08.585776 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:40:10.483626 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:40:10.484235 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:40:10.484850 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:40:10.484876 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" I0213 18:40:13.751389 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:40:23.303659 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:40:23.304437 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:40:23.589698 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:40:23.590298 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:40:24.153971 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:40:26.879654 1 stateful_set.go:466] "StatefulSet has been deleted" logger="statefulset-controller" key="happy-path-managed/affinity-assistant-6832ab5c15" I0213 18:40:34.264663 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:40:38.309150 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:40:38.309838 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:40:38.597295 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:40:38.598058 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:40:48.036507 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:40:48.037147 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:40:48.037711 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:40:48.037742 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:40:53.319033 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:40:53.319700 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:40:53.602277 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:40:53.602981 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:41:08.324023 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:41:08.324784 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:41:08.607544 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:41:08.608199 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:41:18.973138 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="58.041µs" W0213 18:41:23.334680 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:41:23.335409 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:41:23.612460 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:41:23.613096 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:41:29.802385 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="tenant-dev-jydu" W0213 18:41:38.340839 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:41:38.341540 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:41:38.616896 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:41:38.617524 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:41:43.278974 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:41:43.279591 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:41:43.280193 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:41:43.280218 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:41:53.351784 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:41:53.352437 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:41:53.622013 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:41:53.622593 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:41:55.577257 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 18:42:05.623049 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:42:08.357568 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:42:08.358211 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:42:08.629428 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:42:08.630053 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:42:23.367627 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:42:23.368299 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:42:23.633864 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:42:23.634423 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:42:38.133730 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:42:38.134367 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:42:38.134998 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:42:38.135024 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:42:38.373364 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:42:38.374045 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:42:38.638229 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:42:38.638845 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:42:48.228526 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="7.464268ms" I0213 18:42:48.228649 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="smee-client/gosmee-client-6b649c8699" duration="81.831µs" W0213 18:42:53.383846 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:42:53.384472 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:42:53.642541 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:42:53.643146 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:43:06.444231 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:43:08.388688 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:43:08.389317 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 18:43:08.647490 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:43:08.648118 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:43:12.446661 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="happy-path-managed" I0213 18:43:16.849001 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 18:43:18.641183 1 reflector.go:362] The watchlist request ended with an error, falling back to the standard LIST/WATCH semantics because making progress is better than deadlocking, err = the server could not find the requested resource E0213 18:43:18.641823 1 metadata.go:231] "The watchlist request ended with an error, falling back to the standard LIST semantics" err="the server could not find the requested resource" resource="dashboard.tekton.dev/v1alpha1, Resource=extensions" W0213 18:43:18.642370 1 reflector.go:569] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0213 18:43:18.642396 1 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError" W0213 18:43:23.398953 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:43:23.399663 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 18:43:23.483730 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="happy-path-ymrv" W0213 18:43:23.653115 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 18:43:23.653737 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-operator-proxy-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError"