I0213 13:37:33.434330 1 serving.go:386] Generated self-signed cert in-memory I0213 13:37:34.017182 1 controllermanager.go:185] "Starting" version="v1.32.5" I0213 13:37:34.017200 1 controllermanager.go:187] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" I0213 13:37:34.018240 1 dynamic_cafile_content.go:161] "Starting controller" name="request-header::/etc/kubernetes/pki/front-proxy-ca.crt" I0213 13:37:34.018247 1 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" I0213 13:37:34.018300 1 secure_serving.go:213] Serving securely on 127.0.0.1:10257 I0213 13:37:34.018325 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" I0213 13:37:34.018485 1 leaderelection.go:257] attempting to acquire leader lease kube-system/kube-controller-manager... E0213 13:37:34.940124 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: leases.coordination.k8s.io "kube-controller-manager" is forbidden: User "system:kube-controller-manager" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-system" I0213 13:37:38.854485 1 leaderelection.go:271] successfully acquired lease kube-system/kube-controller-manager I0213 13:37:38.854516 1 event.go:389] "Event occurred" object="kube-system/kube-controller-manager" fieldPath="" kind="Lease" apiVersion="coordination.k8s.io/v1" type="Normal" reason="LeaderElection" message="kind-mapt-control-plane_f2729534-e8e0-4d66-afdd-5cdfcb759d5b became leader" I0213 13:37:38.856537 1 controllermanager.go:765] "Started controller" controller="serviceaccount-token-controller" I0213 13:37:38.856594 1 shared_informer.go:313] Waiting for caches to sync for tokens I0213 13:37:38.868804 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-signing-controller" name="csrsigning-kubelet-serving" I0213 13:37:38.868830 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrsigning-kubelet-serving I0213 13:37:38.868873 1 dynamic_serving_content.go:135] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0213 13:37:38.869095 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-signing-controller" name="csrsigning-kubelet-client" I0213 13:37:38.869111 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrsigning-kubelet-client I0213 13:37:38.869112 1 dynamic_serving_content.go:135] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0213 13:37:38.869214 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-signing-controller" name="csrsigning-kube-apiserver-client" I0213 13:37:38.869230 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrsigning-kube-apiserver-client I0213 13:37:38.869244 1 dynamic_serving_content.go:135] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0213 13:37:38.869352 1 controllermanager.go:765] "Started controller" controller="certificatesigningrequest-signing-controller" I0213 13:37:38.869364 1 controllermanager.go:723] "Skipping a cloud provider controller" controller="cloud-node-lifecycle-controller" I0213 13:37:38.869451 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-signing-controller" name="csrsigning-legacy-unknown" I0213 13:37:38.869464 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrsigning-legacy-unknown I0213 13:37:38.869478 1 dynamic_serving_content.go:135] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0213 13:37:38.875999 1 controllermanager.go:765] "Started controller" controller="persistentvolume-attach-detach-controller" I0213 13:37:38.876071 1 attach_detach_controller.go:338] "Starting attach detach controller" logger="persistentvolume-attach-detach-controller" I0213 13:37:38.876087 1 shared_informer.go:313] Waiting for caches to sync for attach detach I0213 13:37:38.882315 1 controllermanager.go:765] "Started controller" controller="legacy-serviceaccount-token-cleaner-controller" I0213 13:37:38.882336 1 legacy_serviceaccount_token_cleaner.go:103] "Starting legacy service account token cleaner controller" logger="legacy-serviceaccount-token-cleaner-controller" I0213 13:37:38.882347 1 shared_informer.go:313] Waiting for caches to sync for legacy-service-account-token-cleaner I0213 13:37:38.883467 1 controllermanager.go:765] "Started controller" controller="certificatesigningrequest-approving-controller" I0213 13:37:38.883565 1 certificate_controller.go:120] "Starting certificate controller" logger="certificatesigningrequest-approving-controller" name="csrapproving" I0213 13:37:38.883578 1 shared_informer.go:313] Waiting for caches to sync for certificate-csrapproving I0213 13:37:38.890183 1 controllermanager.go:765] "Started controller" controller="persistentvolumeclaim-protection-controller" I0213 13:37:38.890209 1 pvc_protection_controller.go:168] "Starting PVC protection controller" logger="persistentvolumeclaim-protection-controller" I0213 13:37:38.890219 1 shared_informer.go:313] Waiting for caches to sync for PVC protection I0213 13:37:38.898646 1 controllermanager.go:765] "Started controller" controller="persistentvolume-protection-controller" I0213 13:37:38.898657 1 pv_protection_controller.go:81] "Starting PV protection controller" logger="persistentvolume-protection-controller" I0213 13:37:38.898663 1 controllermanager.go:743] "Warning: skipping controller" controller="storage-version-migrator-controller" I0213 13:37:38.898665 1 shared_informer.go:313] Waiting for caches to sync for PV protection I0213 13:37:38.907961 1 controllermanager.go:765] "Started controller" controller="replicationcontroller-controller" I0213 13:37:38.908086 1 replica_set.go:217] "Starting controller" logger="replicationcontroller-controller" name="replicationcontroller" I0213 13:37:38.908098 1 shared_informer.go:313] Waiting for caches to sync for ReplicationController I0213 13:37:38.916791 1 garbagecollector.go:144] "Starting controller" logger="garbage-collector-controller" controller="garbagecollector" I0213 13:37:38.916806 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 13:37:38.916822 1 controllermanager.go:765] "Started controller" controller="garbage-collector-controller" I0213 13:37:38.916870 1 graph_builder.go:351] "Running" logger="garbage-collector-controller" component="GraphBuilder" I0213 13:37:38.925643 1 controllermanager.go:765] "Started controller" controller="job-controller" I0213 13:37:38.926198 1 job_controller.go:243] "Starting job controller" logger="job-controller" I0213 13:37:38.926217 1 shared_informer.go:313] Waiting for caches to sync for job I0213 13:37:38.957224 1 shared_informer.go:320] Caches are synced for tokens I0213 13:37:39.157092 1 controllermanager.go:765] "Started controller" controller="horizontal-pod-autoscaler-controller" I0213 13:37:39.157146 1 horizontal.go:201] "Starting HPA controller" logger="horizontal-pod-autoscaler-controller" I0213 13:37:39.157158 1 shared_informer.go:313] Waiting for caches to sync for HPA I0213 13:37:39.207285 1 controllermanager.go:765] "Started controller" controller="certificatesigningrequest-cleaner-controller" I0213 13:37:39.207332 1 cleaner.go:83] "Starting CSR cleaner controller" logger="certificatesigningrequest-cleaner-controller" I0213 13:37:39.364317 1 controllermanager.go:765] "Started controller" controller="ttl-after-finished-controller" I0213 13:37:39.364342 1 ttlafterfinished_controller.go:112] "Starting TTL after finished controller" logger="ttl-after-finished-controller" I0213 13:37:39.364356 1 shared_informer.go:313] Waiting for caches to sync for TTL after finished I0213 13:37:39.513529 1 controllermanager.go:765] "Started controller" controller="ephemeral-volume-controller" I0213 13:37:39.513661 1 controller.go:173] "Starting ephemeral volume controller" logger="ephemeral-volume-controller" I0213 13:37:39.513676 1 shared_informer.go:313] Waiting for caches to sync for ephemeral I0213 13:37:39.707576 1 controllermanager.go:765] "Started controller" controller="validatingadmissionpolicy-status-controller" I0213 13:37:39.707628 1 shared_informer.go:313] Waiting for caches to sync for validatingadmissionpolicy-status I0213 13:37:39.860608 1 controllermanager.go:765] "Started controller" controller="pod-garbage-collector-controller" I0213 13:37:39.860669 1 gc_controller.go:99] "Starting GC controller" logger="pod-garbage-collector-controller" I0213 13:37:39.860681 1 shared_informer.go:313] Waiting for caches to sync for GC I0213 13:37:40.162526 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="poddisruptionbudgets.policy" I0213 13:37:40.162565 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="limitranges" I0213 13:37:40.162608 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="statefulsets.apps" I0213 13:37:40.162631 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="rolebindings.rbac.authorization.k8s.io" I0213 13:37:40.162656 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="endpoints" I0213 13:37:40.162695 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="networkpolicies.networking.k8s.io" I0213 13:37:40.162720 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="replicasets.apps" I0213 13:37:40.162740 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="deployments.apps" I0213 13:37:40.162757 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="jobs.batch" I0213 13:37:40.162806 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="controllerrevisions.apps" I0213 13:37:40.162826 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="horizontalpodautoscalers.autoscaling" I0213 13:37:40.162870 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="endpointslices.discovery.k8s.io" I0213 13:37:40.162915 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="podtemplates" I0213 13:37:40.162926 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="daemonsets.apps" I0213 13:37:40.162938 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="csistoragecapacities.storage.k8s.io" I0213 13:37:40.162956 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="serviceaccounts" I0213 13:37:40.162966 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="ingresses.networking.k8s.io" I0213 13:37:40.162976 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="roles.rbac.authorization.k8s.io" I0213 13:37:40.162992 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="cronjobs.batch" I0213 13:37:40.163005 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="leases.coordination.k8s.io" I0213 13:37:40.163018 1 controllermanager.go:765] "Started controller" controller="resourcequota-controller" I0213 13:37:40.163031 1 resource_quota_controller.go:300] "Starting resource quota controller" logger="resourcequota-controller" I0213 13:37:40.163044 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 13:37:40.163064 1 resource_quota_monitor.go:308] "QuotaMonitor running" logger="resourcequota-controller" I0213 13:37:40.311222 1 controllermanager.go:765] "Started controller" controller="daemonset-controller" I0213 13:37:40.311288 1 daemon_controller.go:294] "Starting daemon sets controller" logger="daemonset-controller" I0213 13:37:40.311304 1 shared_informer.go:313] Waiting for caches to sync for daemon sets I0213 13:37:40.461357 1 controllermanager.go:765] "Started controller" controller="replicaset-controller" I0213 13:37:40.461377 1 controllermanager.go:723] "Skipping a cloud provider controller" controller="service-lb-controller" I0213 13:37:40.461470 1 replica_set.go:217] "Starting controller" logger="replicaset-controller" name="replicaset" I0213 13:37:40.461487 1 shared_informer.go:313] Waiting for caches to sync for ReplicaSet I0213 13:37:40.611049 1 controllermanager.go:765] "Started controller" controller="persistentvolume-binder-controller" I0213 13:37:40.611069 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="volumeattributesclass-protection-controller" requiredFeatureGates=["VolumeAttributesClass"] I0213 13:37:40.611173 1 pv_controller_base.go:308] "Starting persistent volume controller" logger="persistentvolume-binder-controller" I0213 13:37:40.611191 1 shared_informer.go:313] Waiting for caches to sync for persistent volume I0213 13:37:40.760945 1 controllermanager.go:765] "Started controller" controller="root-ca-certificate-publisher-controller" I0213 13:37:40.760993 1 publisher.go:107] "Starting root CA cert publisher controller" logger="root-ca-certificate-publisher-controller" I0213 13:37:40.761001 1 shared_informer.go:313] Waiting for caches to sync for crt configmap I0213 13:37:40.910172 1 controllermanager.go:765] "Started controller" controller="endpoints-controller" I0213 13:37:40.910190 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="resourceclaim-controller" requiredFeatureGates=["DynamicResourceAllocation"] I0213 13:37:40.910202 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="selinux-warning-controller" requiredFeatureGates=["SELinuxChangePolicy"] I0213 13:37:40.910208 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="storageversion-garbage-collector-controller" requiredFeatureGates=["APIServerIdentity","StorageVersionAPI"] I0213 13:37:40.910276 1 endpoints_controller.go:182] "Starting endpoint controller" logger="endpoints-controller" I0213 13:37:40.910291 1 shared_informer.go:313] Waiting for caches to sync for endpoint I0213 13:37:41.060939 1 controllermanager.go:765] "Started controller" controller="deployment-controller" I0213 13:37:41.061020 1 deployment_controller.go:173] "Starting controller" logger="deployment-controller" controller="deployment" I0213 13:37:41.061039 1 shared_informer.go:313] Waiting for caches to sync for deployment I0213 13:37:41.211400 1 controllermanager.go:765] "Started controller" controller="cronjob-controller" I0213 13:37:41.211516 1 cronjob_controllerv2.go:145] "Starting cronjob controller v2" logger="cronjob-controller" I0213 13:37:41.211537 1 shared_informer.go:313] Waiting for caches to sync for cronjob I0213 13:37:41.460343 1 range_allocator.go:112] "No Secondary Service CIDR provided. Skipping filtering out secondary service addresses" logger="node-ipam-controller" I0213 13:37:41.460391 1 controllermanager.go:765] "Started controller" controller="node-ipam-controller" I0213 13:37:41.460497 1 node_ipam_controller.go:141] "Starting ipam controller" logger="node-ipam-controller" I0213 13:37:41.460515 1 shared_informer.go:313] Waiting for caches to sync for node I0213 13:37:41.610172 1 controllermanager.go:765] "Started controller" controller="clusterrole-aggregation-controller" I0213 13:37:41.610213 1 clusterroleaggregation_controller.go:194] "Starting ClusterRoleAggregator controller" logger="clusterrole-aggregation-controller" I0213 13:37:41.610219 1 shared_informer.go:313] Waiting for caches to sync for ClusterRoleAggregator I0213 13:37:41.760366 1 controllermanager.go:765] "Started controller" controller="serviceaccount-controller" I0213 13:37:41.760401 1 serviceaccounts_controller.go:114] "Starting service account controller" logger="serviceaccount-controller" I0213 13:37:41.760407 1 shared_informer.go:313] Waiting for caches to sync for service account I0213 13:37:41.910876 1 controllermanager.go:765] "Started controller" controller="statefulset-controller" I0213 13:37:41.910984 1 stateful_set.go:166] "Starting stateful set controller" logger="statefulset-controller" I0213 13:37:41.911003 1 shared_informer.go:313] Waiting for caches to sync for stateful set I0213 13:37:41.957446 1 node_lifecycle_controller.go:432] "Controller will reconcile labels" logger="node-lifecycle-controller" I0213 13:37:41.957486 1 controllermanager.go:765] "Started controller" controller="node-lifecycle-controller" I0213 13:37:41.957491 1 controllermanager.go:723] "Skipping a cloud provider controller" controller="node-route-controller" I0213 13:37:41.957497 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="service-cidr-controller" requiredFeatureGates=["MultiCIDRServiceAllocator"] I0213 13:37:41.957563 1 node_lifecycle_controller.go:466] "Sending events to api server" logger="node-lifecycle-controller" I0213 13:37:41.957597 1 node_lifecycle_controller.go:477] "Starting node controller" logger="node-lifecycle-controller" I0213 13:37:41.957603 1 shared_informer.go:313] Waiting for caches to sync for taint I0213 13:37:42.156326 1 controllermanager.go:765] "Started controller" controller="disruption-controller" I0213 13:37:42.156363 1 disruption.go:452] "Sending events to api server." logger="disruption-controller" I0213 13:37:42.156405 1 disruption.go:463] "Starting disruption controller" logger="disruption-controller" I0213 13:37:42.156416 1 shared_informer.go:313] Waiting for caches to sync for disruption I0213 13:37:42.310190 1 controllermanager.go:765] "Started controller" controller="endpointslice-mirroring-controller" I0213 13:37:42.310294 1 endpointslicemirroring_controller.go:227] "Starting EndpointSliceMirroring controller" logger="endpointslice-mirroring-controller" I0213 13:37:42.310307 1 shared_informer.go:313] Waiting for caches to sync for endpoint_slice_mirroring I0213 13:37:42.459994 1 controllermanager.go:765] "Started controller" controller="bootstrap-signer-controller" I0213 13:37:42.460030 1 shared_informer.go:313] Waiting for caches to sync for bootstrap_signer I0213 13:37:42.610020 1 controllermanager.go:765] "Started controller" controller="token-cleaner-controller" I0213 13:37:42.610036 1 tokencleaner.go:117] "Starting token cleaner controller" logger="token-cleaner-controller" I0213 13:37:42.610046 1 shared_informer.go:313] Waiting for caches to sync for token_cleaner I0213 13:37:42.610054 1 shared_informer.go:320] Caches are synced for token_cleaner I0213 13:37:42.656723 1 controllermanager.go:765] "Started controller" controller="taint-eviction-controller" I0213 13:37:42.656752 1 taint_eviction.go:281] "Starting" logger="taint-eviction-controller" controller="taint-eviction-controller" I0213 13:37:42.656776 1 taint_eviction.go:287] "Sending events to api server" logger="taint-eviction-controller" I0213 13:37:42.656813 1 shared_informer.go:313] Waiting for caches to sync for taint-eviction-controller I0213 13:37:42.810516 1 controllermanager.go:765] "Started controller" controller="endpointslice-controller" I0213 13:37:42.810589 1 endpointslice_controller.go:281] "Starting endpoint slice controller" logger="endpointslice-controller" I0213 13:37:42.810603 1 shared_informer.go:313] Waiting for caches to sync for endpoint_slice I0213 13:37:42.959900 1 controllermanager.go:765] "Started controller" controller="ttl-controller" I0213 13:37:42.959917 1 ttl_controller.go:127] "Starting TTL controller" logger="ttl-controller" I0213 13:37:42.959926 1 shared_informer.go:313] Waiting for caches to sync for TTL I0213 13:37:43.109627 1 controllermanager.go:765] "Started controller" controller="persistentvolume-expander-controller" I0213 13:37:43.109642 1 controllermanager.go:717] "Controller is disabled by a feature gate" controller="kube-apiserver-serving-clustertrustbundle-publisher-controller" requiredFeatureGates=["ClusterTrustBundle"] I0213 13:37:43.109673 1 expand_controller.go:329] "Starting expand controller" logger="persistentvolume-expander-controller" I0213 13:37:43.109680 1 shared_informer.go:313] Waiting for caches to sync for expand I0213 13:37:43.363454 1 controllermanager.go:765] "Started controller" controller="namespace-controller" I0213 13:37:43.363545 1 namespace_controller.go:202] "Starting namespace controller" logger="namespace-controller" I0213 13:37:43.363560 1 shared_informer.go:313] Waiting for caches to sync for namespace I0213 13:37:43.366261 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 13:37:43.369778 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-kubelet-client I0213 13:37:43.370682 1 actual_state_of_world.go:541] "Failed to update statusUpdateNeeded field in actual state of world" logger="persistentvolume-attach-detach-controller" err="Failed to set statusUpdateNeeded to needed true, because nodeName=\"kind-mapt-control-plane\" does not exist" I0213 13:37:43.374189 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 13:37:43.377785 1 shared_informer.go:320] Caches are synced for attach detach I0213 13:37:43.382917 1 shared_informer.go:320] Caches are synced for legacy-service-account-token-cleaner I0213 13:37:43.384061 1 shared_informer.go:320] Caches are synced for certificate-csrapproving I0213 13:37:43.390271 1 shared_informer.go:320] Caches are synced for PVC protection I0213 13:37:43.399571 1 shared_informer.go:320] Caches are synced for PV protection I0213 13:37:43.408185 1 shared_informer.go:320] Caches are synced for ReplicationController I0213 13:37:43.408202 1 shared_informer.go:320] Caches are synced for validatingadmissionpolicy-status I0213 13:37:43.410331 1 shared_informer.go:320] Caches are synced for expand I0213 13:37:43.410351 1 shared_informer.go:320] Caches are synced for endpoint I0213 13:37:43.410382 1 shared_informer.go:320] Caches are synced for ClusterRoleAggregator I0213 13:37:43.410441 1 shared_informer.go:320] Caches are synced for endpoint_slice_mirroring I0213 13:37:43.410795 1 shared_informer.go:320] Caches are synced for endpoint_slice I0213 13:37:43.411099 1 shared_informer.go:320] Caches are synced for stateful set I0213 13:37:43.411402 1 shared_informer.go:320] Caches are synced for daemon sets I0213 13:37:43.411417 1 shared_informer.go:320] Caches are synced for persistent volume I0213 13:37:43.412585 1 shared_informer.go:320] Caches are synced for cronjob I0213 13:37:43.413729 1 shared_informer.go:320] Caches are synced for ephemeral I0213 13:37:43.416992 1 shared_informer.go:320] Caches are synced for garbage collector I0213 13:37:43.417008 1 garbagecollector.go:154] "Garbage collector: all resource monitors have synced" logger="garbage-collector-controller" I0213 13:37:43.417014 1 garbagecollector.go:157] "Proceeding to collect garbage" logger="garbage-collector-controller" I0213 13:37:43.426397 1 shared_informer.go:320] Caches are synced for job I0213 13:37:43.456813 1 shared_informer.go:320] Caches are synced for disruption I0213 13:37:43.456875 1 shared_informer.go:320] Caches are synced for taint-eviction-controller I0213 13:37:43.458058 1 shared_informer.go:320] Caches are synced for taint I0213 13:37:43.458087 1 shared_informer.go:320] Caches are synced for HPA I0213 13:37:43.458132 1 node_lifecycle_controller.go:1234] "Initializing eviction metric for zone" logger="node-lifecycle-controller" zone="" I0213 13:37:43.458202 1 node_lifecycle_controller.go:886] "Missing timestamp for Node. Assuming now as a timestamp" logger="node-lifecycle-controller" node="kind-mapt-control-plane" I0213 13:37:43.458240 1 node_lifecycle_controller.go:1038] "Controller detected that all Nodes are not-Ready. Entering master disruption mode" logger="node-lifecycle-controller" I0213 13:37:43.460354 1 shared_informer.go:320] Caches are synced for bootstrap_signer I0213 13:37:43.460377 1 shared_informer.go:320] Caches are synced for TTL I0213 13:37:43.460428 1 shared_informer.go:320] Caches are synced for service account I0213 13:37:43.460575 1 shared_informer.go:320] Caches are synced for node I0213 13:37:43.460619 1 range_allocator.go:177] "Sending events to api server" logger="node-ipam-controller" I0213 13:37:43.460643 1 range_allocator.go:183] "Starting range CIDR allocator" logger="node-ipam-controller" I0213 13:37:43.460648 1 shared_informer.go:313] Waiting for caches to sync for cidrallocator I0213 13:37:43.460652 1 shared_informer.go:320] Caches are synced for cidrallocator I0213 13:37:43.460855 1 shared_informer.go:320] Caches are synced for GC I0213 13:37:43.461035 1 shared_informer.go:320] Caches are synced for crt configmap I0213 13:37:43.461155 1 shared_informer.go:320] Caches are synced for deployment I0213 13:37:43.461515 1 shared_informer.go:320] Caches are synced for ReplicaSet I0213 13:37:43.463732 1 shared_informer.go:320] Caches are synced for namespace I0213 13:37:43.463768 1 shared_informer.go:320] Caches are synced for resource quota I0213 13:37:43.464914 1 shared_informer.go:320] Caches are synced for TTL after finished I0213 13:37:43.467133 1 shared_informer.go:320] Caches are synced for resource quota I0213 13:37:43.469516 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-legacy-unknown I0213 13:37:43.469567 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-kube-apiserver-client I0213 13:37:43.469652 1 shared_informer.go:320] Caches are synced for certificate-csrsigning-kubelet-serving I0213 13:37:43.469714 1 range_allocator.go:428] "Set node PodCIDR" logger="node-ipam-controller" node="kind-mapt-control-plane" podCIDRs=["10.244.0.0/24"] I0213 13:37:43.469738 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:37:43.469760 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:37:43.474959 1 shared_informer.go:320] Caches are synced for garbage collector I0213 13:37:44.364021 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:37:44.577349 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="114.484703ms" I0213 13:37:44.582499 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="119.745356ms" I0213 13:37:44.586163 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="8.763217ms" I0213 13:37:44.586232 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="30.93µs" I0213 13:37:44.590451 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="7.915186ms" I0213 13:37:44.595218 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="48.46µs" I0213 13:37:44.601766 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="11.283259ms" I0213 13:37:44.601821 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="31.26µs" I0213 13:37:58.814019 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:37:58.821316 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:37:58.827429 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="42.6µs" I0213 13:37:58.830153 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="109.86µs" I0213 13:37:58.830201 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="24.92µs" I0213 13:37:58.840928 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="35.47µs" I0213 13:37:58.850590 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="92.04µs" I0213 13:37:58.858155 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="34.42µs" I0213 13:38:03.461576 1 node_lifecycle_controller.go:1057] "Controller detected that some Nodes are Ready. Exiting master disruption mode" logger="node-lifecycle-controller" I0213 13:38:04.018074 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="61.32µs" I0213 13:38:04.033277 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="6.044195ms" I0213 13:38:04.033370 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-7dc846544d" duration="53.76µs" I0213 13:38:04.042649 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="6.254525ms" I0213 13:38:04.042725 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="56.4µs" I0213 13:38:08.129404 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:38:15.034572 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="48.4µs" I0213 13:38:15.051448 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="8.219746ms" I0213 13:38:15.051512 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/coredns-668d6bf9bc" duration="37.48µs" I0213 13:38:28.567379 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:39:19.783522 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:39:30.050640 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:39:40.144591 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:39:50.344953 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:40:21.133923 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:40:31.359074 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:40:41.744631 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:40:51.891940 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:41:12.104471 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:41:22.281396 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:41:32.463797 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:41:42.807823 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:42:13.499744 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="internalrequests.appstudio.redhat.com" I0213 13:42:13.499789 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="internalservicesconfigs.appstudio.redhat.com" I0213 13:42:13.499862 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 13:42:13.501949 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 13:42:14.600204 1 shared_informer.go:320] Caches are synced for resource quota I0213 13:42:14.602558 1 shared_informer.go:320] Caches are synced for garbage collector I0213 13:42:24.079009 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:42:55.170480 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:43:35.788480 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:43:55.933379 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:44:05.172919 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-868956fb6c" duration="20.56255ms" I0213 13:44:05.192691 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-868956fb6c" duration="19.733113ms" I0213 13:44:05.192807 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-868956fb6c" duration="65.581µs" I0213 13:44:05.192915 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-868956fb6c" duration="80.222µs" I0213 13:44:05.193005 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-868956fb6c" duration="37.051µs" I0213 13:44:05.270600 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-bcd5b75f7" duration="16.918953ms" I0213 13:44:05.288149 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-bcd5b75f7" duration="17.506695ms" I0213 13:44:05.288233 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-bcd5b75f7" duration="47.871µs" I0213 13:44:06.074372 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines-resolvers/tekton-pipelines-remote-resolvers-84d77d7555" duration="27.458669ms" I0213 13:44:06.085103 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines-resolvers/tekton-pipelines-remote-resolvers-84d77d7555" duration="10.697209ms" I0213 13:44:06.085184 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines-resolvers/tekton-pipelines-remote-resolvers-84d77d7555" duration="49.121µs" I0213 13:44:06.089570 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines-resolvers/tekton-pipelines-remote-resolvers-84d77d7555" duration="33.471µs" I0213 13:44:06.216314 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-b974bd4b" duration="20.221263ms" I0213 13:44:06.226756 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-b974bd4b" duration="10.408223ms" I0213 13:44:06.226863 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-b974bd4b" duration="69.542µs" I0213 13:44:06.230363 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-b974bd4b" duration="48.591µs" I0213 13:44:07.627651 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-868956fb6c" duration="74.781µs" I0213 13:44:08.632224 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-bcd5b75f7" duration="69.102µs" I0213 13:44:09.642118 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines-resolvers/tekton-pipelines-remote-resolvers-84d77d7555" duration="8.338769ms" I0213 13:44:09.642212 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines-resolvers/tekton-pipelines-remote-resolvers-84d77d7555" duration="61.022µs" I0213 13:44:10.986537 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-b974bd4b" duration="76.322µs" I0213 13:44:14.614596 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="stepactions.tekton.dev" I0213 13:44:14.614699 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="pipelineruns.tekton.dev" I0213 13:44:14.614720 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="tasks.tekton.dev" I0213 13:44:14.614752 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="customruns.tekton.dev" I0213 13:44:14.614786 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="taskruns.tekton.dev" I0213 13:44:14.614892 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="verificationpolicies.tekton.dev" I0213 13:44:14.614932 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="resolutionrequests.resolution.tekton.dev" I0213 13:44:14.614963 1 resource_quota_monitor.go:227] "QuotaMonitor created object count evaluator" logger="resourcequota-controller" resource="pipelines.tekton.dev" I0213 13:44:14.615069 1 shared_informer.go:313] Waiting for caches to sync for resource quota I0213 13:44:14.618731 1 shared_informer.go:313] Waiting for caches to sync for garbage collector I0213 13:44:15.715395 1 shared_informer.go:320] Caches are synced for resource quota I0213 13:44:15.719710 1 shared_informer.go:320] Caches are synced for garbage collector I0213 13:44:16.423395 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:44:18.639611 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-868956fb6c" duration="6.631772ms" I0213 13:44:18.639677 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-controller-868956fb6c" duration="38.251µs" I0213 13:44:20.652061 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-bcd5b75f7" duration="7.774237ms" I0213 13:44:20.652157 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-events-controller-bcd5b75f7" duration="59.792µs" W0213 13:44:21.152177 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:44:21.300755 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:44:21.668119 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-b974bd4b" duration="18.316413ms" I0213 13:44:21.668214 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="tekton-pipelines/tekton-pipelines-webhook-b974bd4b" duration="57.802µs" I0213 13:44:27.012099 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:44:27.719015 1 stateful_set.go:466] "StatefulSet has been deleted" logger="statefulset-controller" key="default/affinity-assistant-1f5b5fa6e6" W0213 13:44:36.304790 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:44:36.305366 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:44:47.484228 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:44:51.311777 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:44:51.312363 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:44:57.504993 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:45:06.317081 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:45:06.317701 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:45:07.888484 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:45:21.324582 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:45:21.325231 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:45:36.330101 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:45:36.330655 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:45:51.337944 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:45:51.338618 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:46:06.343051 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:46:06.343623 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:46:21.351784 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:46:21.352374 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:46:36.356876 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:46:36.357387 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:46:39.770432 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:46:50.124434 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:46:51.365275 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:46:51.365930 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:47:00.527114 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:47:06.370936 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:47:06.371484 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:47:21.377677 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:47:21.378274 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:47:36.384070 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:47:36.384619 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:47:51.391497 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:47:51.392091 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:47:51.673355 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:48:02.075151 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:48:06.396615 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:48:06.397248 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:48:21.403998 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:48:21.404639 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:48:36.409791 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:48:36.410457 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:48:42.609777 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:48:51.417825 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:48:51.418453 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:48:52.630083 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:49:06.423065 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:49:06.423680 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" W0213 13:49:21.430251 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:49:21.430792 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError" I0213 13:49:23.417448 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" I0213 13:49:33.483349 1 range_allocator.go:247] "Successfully synced" logger="node-ipam-controller" key="kind-mapt-control-plane" W0213 13:49:36.434649 1 type.go:183] The watchlist request for pods ended with an error, falling back to the standard LIST semantics, err = the server could not find the requested resource (get pods.metrics.k8s.io) E0213 13:49:36.435312 1 horizontal.go:275] "Unhandled Error" err="failed to compute desired number of replicas based on listed metrics for Deployment/tekton-pipelines/tekton-pipelines-webhook: invalid metrics (1 invalid out of 1), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: the server could not find the requested resource (get pods.metrics.k8s.io)" logger="UnhandledError"