+ . /ovnkube-lib/ovnkube-lib.sh ++ set -x ++ K8S_NODE=ip-10-0-0-158.ec2.internal ++ [[ -n ip-10-0-0-158.ec2.internal ]] ++ [[ -f /env/ip-10-0-0-158.ec2.internal ]] ++ northd_pidfile=/var/run/ovn/ovn-northd.pid ++ controller_pidfile=/var/run/ovn/ovn-controller.pid ++ controller_logfile=/var/log/ovn/acl-audit-log.log ++ vswitch_dbsock=/var/run/openvswitch/db.sock ++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid ++ nbdb_sock=/var/run/ovn/ovnnb_db.sock ++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl ++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid ++ sbdb_sock=/var/run/ovn/ovnsb_db.sock ++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl + start-ovnkube-node 4 29103 29105 + local log_level=4 + local metrics_port=29103 + local ovn_metrics_port=29105 + ovn_advertised_udn_isolation_mode_flag= + openflow_probe_flag= + [[ 3 -ne 3 ]] + ovs-vsctl br-exists br-ex + add_garp_drop_flow br-ex + local bridge=br-ex + local cookie=0x0305 + local priority=499 ++ ovs-vsctl list-ports br-ex + for port_name in $(ovs-vsctl list-ports "$bridge") + [[ ens5 == *to-br-int ]] + cni-bin-copy + . /host/etc/os-release ++ NAME='Red Hat Enterprise Linux CoreOS' ++ VERSION='9.6.20260401-0 (Plow)' ++ ID=rhel ++ ID_LIKE=fedora ++ VERSION_ID=9.6 ++ PLATFORM_ID=platform:el9 ++ PRETTY_NAME='Red Hat Enterprise Linux CoreOS 9.6.20260401-0 (Plow)' ++ ANSI_COLOR='0;31' ++ LOGO=fedora-logo-icon ++ CPE_NAME=cpe:/o:redhat:enterprise_linux:9::baseos ++ HOME_URL=https://www.redhat.com/ ++ DOCUMENTATION_URL=https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/9 ++ BUG_REPORT_URL=https://issues.redhat.com/ ++ REDHAT_BUGZILLA_PRODUCT='Red Hat Enterprise Linux 9' ++ REDHAT_BUGZILLA_PRODUCT_VERSION=9.6 ++ REDHAT_SUPPORT_PRODUCT='Red Hat Enterprise Linux' ++ REDHAT_SUPPORT_PRODUCT_VERSION=9.6 ++ OSTREE_VERSION=9.6.20260401-0 ++ VARIANT=CoreOS ++ VARIANT_ID=coreos ++ OPENSHIFT_VERSION=4.21 + rhelmajor= + case "${ID}" in ++ echo 9.6 ++ cut -f 1 -d . + rhelmajor=9 + sourcedir=/usr/libexec/cni/ + case "${rhelmajor}" in + sourcedir=/usr/libexec/cni/rhel9 + cp -f /usr/libexec/cni/rhel9/ovn-k8s-cni-overlay /cni-bin-dir/ ++ date '+%m%d %H:%M:%S.%N' + echo 'I0514 16:08:33.213200257 - disable conntrack on geneve port' I0514 16:08:33.213200257 - disable conntrack on geneve port + iptables -t raw -A PREROUTING -p udp --dport 6081 -j NOTRACK + iptables -t raw -A OUTPUT -p udp --dport 6081 -j NOTRACK + ip6tables -t raw -A PREROUTING -p udp --dport 6081 -j NOTRACK + ip6tables -t raw -A OUTPUT -p udp --dport 6081 -j NOTRACK ++ date '+%m%d %H:%M:%S.%N' I0514 16:08:33.239649097 - starting ovnkube-node + echo 'I0514 16:08:33.239649097 - starting ovnkube-node' + egress_features_enable_flag='--enable-egress-ip=true --enable-egress-firewall=true --enable-egress-qos=true --enable-egress-service=true' + init_ovnkube_controller='--init-ovnkube-controller ip-10-0-0-158.ec2.internal' + multi_external_gateway_enable_flag=--enable-multi-external-gateway=true + gateway_interface=br-ex + enable_multicast_flag=--enable-multicast + OVN_NODE_MODE=full + '[' full == dpu-host ']' + '[' shared == shared ']' + gateway_mode_flags='--gateway-mode shared --gateway-interface br-ex' + export_network_flows_flags= + [[ -n '' ]] + [[ -n '' ]] + [[ -n '' ]] + [[ -n '' ]] + [[ -n '' ]] + [[ -n '' ]] + gw_interface_flag= + '[' -d /sys/class/net/br-ex1 ']' + node_mgmt_port_netdev_flags= + [[ -n '' ]] + [[ -n '' ]] + multi_network_enabled_flag= + [[ true == \t\r\u\e ]] + [[ full != \d\p\u\-\h\o\s\t ]] + multi_network_enabled_flag=--enable-multi-network + network_segmentation_enabled_flag= + [[ true == \t\r\u\e ]] + [[ full != \d\p\u\-\h\o\s\t ]] + multi_network_enabled_flag=--enable-multi-network + network_segmentation_enabled_flag=--enable-network-segmentation + route_advertisements_enable_flag= + [[ false == \t\r\u\e ]] + preconfigured_udn_addresses_enable_flag= + [[ true == \t\r\u\e ]] + preconfigured_udn_addresses_enable_flag=--enable-preconfigured-udn-addresses + network_observability_enabled_flag= + [[ false == \t\r\u\e ]] + multi_network_policy_enabled_flag= + [[ false == \t\r\u\e ]] + admin_network_policy_enabled_flag= + [[ true == \t\r\u\e ]] + [[ full != \d\p\u\-\h\o\s\t ]] + admin_network_policy_enabled_flag=--enable-admin-network-policy + dns_name_resolver_enabled_flag= + [[ false == \t\r\u\e ]] + ip_forwarding_flag= + '[' '' == Global ']' + ip_forwarding_flag=--disable-forwarding + sysctl -w net.ipv4.ip_forward=0 net.ipv4.ip_forward = 0 + sysctl -w net.ipv6.conf.all.forwarding=0 net.ipv6.conf.all.forwarding = 0 + [[ '' != '' ]] + [[ '' != '' ]] + NETWORK_NODE_IDENTITY_ENABLE= + [[ true == \t\r\u\e ]] + NETWORK_NODE_IDENTITY_ENABLE=' --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h ' + ovn_v4_join_subnet_opt= + [[ '' != '' ]] + ovn_v6_join_subnet_opt= + [[ '' != '' ]] + ovn_v4_masquerade_subnet_opt= + [[ 169.254.0.0/17 != '' ]] + ovn_v4_masquerade_subnet_opt='--gateway-v4-masquerade-subnet 169.254.0.0/17' + ovn_v6_masquerade_subnet_opt= + [[ fd69::/112 != '' ]] + ovn_v6_masquerade_subnet_opt='--gateway-v6-masquerade-subnet fd69::/112' + ovn_v4_transit_switch_subnet_opt= + [[ '' != '' ]] + ovn_v6_transit_switch_subnet_opt= + [[ '' != '' ]] + exec /usr/bin/ovnkube --init-ovnkube-controller ip-10-0-0-158.ec2.internal --init-node ip-10-0-0-158.ec2.internal --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode shared --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-preconfigured-udn-addresses --enable-admin-network-policy --enable-multicast --zone ip-10-0-0-158.ec2.internal --enable-interconnect --acl-logging-rate-limit 20 --disable-forwarding --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112 --enable-egress-ip=true --enable-egress-firewall=true --enable-egress-qos=true --enable-egress-service=true --enable-multi-external-gateway=true I0514 16:08:33.284378 3780 config.go:2726] Parsed config file /run/ovnkube-config/ovnkube.conf I0514 16:08:33.284454 3780 config.go:2727] Parsed config: {Default:{MTU:8901 RoutableMTU:0 ConntrackZone:64000 HostMasqConntrackZone:0 OVNMasqConntrackZone:0 HostNodePortConntrackZone:0 ReassemblyConntrackZone:0 EncapType:geneve EncapIP: EffectiveEncapIP: EncapPort:6081 InactivityProbe:100000 OpenFlowProbe:0 OfctrlWaitBeforeClear:0 MonitorAll:true OVSDBTxnTimeout:1m40s LFlowCacheEnable:true LFlowCacheLimit:0 LFlowCacheLimitKb:1048576 RawClusterSubnets:10.128.0.0/14/23 ClusterSubnets:[] EnableUDPAggregation:true Zone:global RawUDNAllowedDefaultServices:default/kubernetes,openshift-dns/dns-default UDNAllowedDefaultServices:[] Transport:geneve} Logging:{File: CNIFile: LibovsdbFile: Level:4 LogFileMaxSize:100 LogFileMaxBackups:5 LogFileMaxAge:5 ACLLoggingRateLimit:20} Monitoring:{RawNetFlowTargets: RawSFlowTargets: RawIPFIXTargets: NetFlowTargets:[] SFlowTargets:[] IPFIXTargets:[]} IPFIX:{Sampling:400 CacheActiveTimeout:60 CacheMaxFlows:0} CNI:{ConfDir:/etc/cni/net.d Plugin:ovn-k8s-cni-overlay} OVNKubernetesFeature:{EnableAdminNetworkPolicy:false EnableEgressIP:false EgressIPReachabiltyTotalTimeout:1 EnableEgressFirewall:false EnableEgressQoS:false EnableEgressService:false EgressIPNodeHealthCheckPort:9107 EnableMultiNetwork:false EnableNetworkSegmentation:true EnableNetworkConnect:false EnablePreconfiguredUDNAddresses:true EnableRouteAdvertisements:false EnableEVPN:false EnableMultiNetworkPolicy:false EnableStatelessNetPol:false EnableInterconnect:false EnableMultiExternalGateway:false EnablePersistentIPs:false EnableDNSNameResolver:false EnableServiceTemplateSupport:false EnableObservability:false EnableNetworkQoS:false AdvertisedUDNIsolationMode:strict EnableDynamicUDNAllocation:false UDNDeletionGracePeriod:2m0s} Kubernetes:{BootstrapKubeconfig: CertDir: CertDuration:10m0s Kubeconfig: CACert: CACertData: CAData:[] APIServer:https://api.kx-60ed1223ed.hypershift.local:443 Token: TokenFile: CompatServiceCIDR: RawServiceCIDRs:172.30.0.0/16 ServiceCIDRs:[] OVNConfigNamespace:openshift-ovn-kubernetes OVNEmptyLbEvents:false RawNoHostSubnetNodes: NoHostSubnetNodes: HostNetworkNamespace:openshift-host-network DisableRequestedChassis:false PlatformType:AWS HealthzBindAddress:0.0.0.0:10256 CompatMetricsBindAddress: CompatOVNMetricsBindAddress: CompatMetricsEnablePprof:false DNSServiceNamespace:openshift-dns DNSServiceName:dns-default} Metrics:{BindAddress: OVNMetricsBindAddress: ExportOVSMetrics:false EnablePprof:false NodeServerPrivKey: NodeServerCert: EnableConfigDuration:false EnableScaleMetrics:false} OvnNorth:{Address: PrivKey: Cert: CACert: CertCommonName: Scheme: ElectionTimer:0 northbound:false RunDir:/var/run/ovn/ DbLocation:/etc/ovn/ovnnb_db.db exec:} OvnSouth:{Address: PrivKey: Cert: CACert: CertCommonName: Scheme: ElectionTimer:0 northbound:false RunDir:/var/run/ovn/ DbLocation:/etc/ovn/ovnsb_db.db exec:} Gateway:{Mode:shared Interface: GatewayAcceleratedInterface: EgressGWInterface: NextHop: VLANID:0 NodeportEnable:true DisableSNATMultipleGWs:false V4JoinSubnet:100.64.0.0/16 V6JoinSubnet:fd98::/64 V4MasqueradeSubnet:169.254.169.0/29 V6MasqueradeSubnet:fd69::/125 MasqueradeIPs:{V4OVNMasqueradeIP:169.254.169.1 V6OVNMasqueradeIP:fd69::1 V4HostMasqueradeIP:169.254.169.2 V6HostMasqueradeIP:fd69::2 V4HostETPLocalMasqueradeIP:169.254.169.3 V6HostETPLocalMasqueradeIP:fd69::3 V4DummyNextHopMasqueradeIP:169.254.169.4 V6DummyNextHopMasqueradeIP:fd69::4 V4OVNServiceHairpinMasqueradeIP:169.254.169.5 V6OVNServiceHairpinMasqueradeIP:fd69::5} DisablePacketMTUCheck:false RouterSubnet: SingleNode:false DisableForwarding:false AllowNoUplink:false EphemeralPortRange:} MasterHA:{ElectionLeaseDuration:137 ElectionRenewDeadline:107 ElectionRetryPeriod:26} ClusterMgrHA:{ElectionLeaseDuration:137 ElectionRenewDeadline:107 ElectionRetryPeriod:26} HybridOverlay:{Enabled:false RawClusterSubnets: ClusterSubnets:[] VXLANPort:4789} OvnKubeNode:{Mode:full MgmtPortNetdev: MgmtPortDPResourceName:} ClusterManager:{V4TransitSubnet:100.88.0.0/16 V6TransitSubnet:fd97::/64} OvsPaths:{RunDir:/var/run/openvswitch/} NoOverlay:{OutboundSNAT: Routing:} ManagedBGP:{ASNumber:64512 Topology:}} I0514 16:08:33.285845 3780 kube.go:431] Waiting for certificate I0514 16:08:33.285881 3780 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client" I0514 16:08:33.285915 3780 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client" I0514 16:08:33.286727 3780 cert_rotation.go:141] "Starting client certificate rotation controller" logger="tls-transport-cache" I0514 16:08:33.319954 3780 reflector.go:358] "Starting reflector" logger="kubernetes.io/kube-apiserver-client" type="*v1.CertificateSigningRequest" resyncPeriod="0s" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" I0514 16:08:33.319975 3780 reflector.go:404] "Listing and watching" logger="kubernetes.io/kube-apiserver-client" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" I0514 16:08:33.323898 3780 reflector.go:436] "Caches populated" logger="kubernetes.io/kube-apiserver-client" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" I0514 16:08:33.323980 3780 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client" csr="csr-6hhjw" I0514 16:08:33.327842 3780 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client" csr="csr-6hhjw" I0514 16:08:33.327912 3780 reflector.go:364] "Stopping reflector" logger="kubernetes.io/kube-apiserver-client" type="*v1.CertificateSigningRequest" resyncPeriod="0s" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" I0514 16:08:34.285964 3780 kube.go:438] Certificate found I0514 16:08:34.286963 3780 cert_rotation.go:141] "Starting client certificate rotation controller" logger="tls-transport-cache" I0514 16:08:34.287972 3780 metrics.go:532] Starting metrics server at address "127.0.0.1:29103" I0514 16:08:34.335448 3780 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client" expiration="2026-05-15 16:03:33 +0000 UTC" deadline="2026-05-15 11:46:02.754286838 +0000 UTC" I0514 16:08:34.335509 3780 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client" sleep="19h37m28.418804163s" I0514 16:08:34.355053 3780 metrics.go:164] OVN Kube log file not specified in config, therefore not starting the log file metric monitor I0514 16:08:34.372033 3780 metrics.go:512] Create OVN Metrics Server on address: 127.0.0.1:29105 I0514 16:08:34.372489 3780 server.go:101] MetricServer registers OVS metrics I0514 16:08:34.372811 3780 server.go:105] MetricServer registers OVN DB metrics I0514 16:08:34.374324 3780 node_controller_manager.go:338] Starting the node network controller manager, Mode: full I0514 16:08:34.374393 3780 factory.go:561] Starting watch factory I0514 16:08:34.374499 3780 reflector.go:358] "Starting reflector" type="*v1.NetworkPolicy" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.374525 3780 reflector.go:404] "Listing and watching" type="*v1.NetworkPolicy" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.374530 3780 reflector.go:358] "Starting reflector" type="*v1.EndpointSlice" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.374500 3780 reflector.go:358] "Starting reflector" type="*v1.Node" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.374546 3780 reflector.go:404] "Listing and watching" type="*v1.EndpointSlice" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.374551 3780 reflector.go:404] "Listing and watching" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.374726 3780 reflector.go:358] "Starting reflector" type="*v1.Namespace" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.374751 3780 reflector.go:404] "Listing and watching" type="*v1.Namespace" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.375177 3780 reflector.go:358] "Starting reflector" type="*v1.Pod" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.375198 3780 reflector.go:404] "Listing and watching" type="*v1.Pod" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.376699 3780 reflector.go:358] "Starting reflector" type="*v1.Service" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.376721 3780 reflector.go:404] "Listing and watching" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.377075 3780 ovn_db.go:330] /var/run/openvswitch/ovnnb_db.sock getting info failed: stat /var/run/openvswitch/ovnnb_db.sock: no such file or directory I0514 16:08:34.377122 3780 ovn_db.go:327] ovnnb_db.sock found at /var/run/ovn/ I0514 16:08:34.379112 3780 controller_manager.go:370] Starting the ovnkube controller I0514 16:08:34.379130 3780 controller_manager.go:375] Waiting up to 5m0s for NBDB zone to match: ip-10-0-0-158.ec2.internal I0514 16:08:34.379233 3780 controller_manager.go:395] NBDB zone sync took: 88.285µs I0514 16:08:34.379244 3780 factory.go:561] Starting watch factory I0514 16:08:34.391271 3780 ovn_db.go:388] Found OVN NB DB: &{0x1e9e6e0 /etc/ovn/ovnnb_db.db OVN_Northbound 0} I0514 16:08:34.391295 3780 ovn_db.go:395] Found OVN SB DB: &{0x1e9e9a0 /etc/ovn/ovnsb_db.db OVN_Southbound 0} I0514 16:08:34.393099 3780 reflector.go:436] "Caches populated" type="*v1.NetworkPolicy" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.394393 3780 reflector.go:436] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.394779 3780 reflector.go:436] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.396039 3780 ovn_db.go:410] Found db is standalone, don't register db_cluster metrics I0514 16:08:34.396071 3780 server.go:109] MetricServer registers OVN Controller metrics I0514 16:08:34.397021 3780 reflector.go:436] "Caches populated" type="*v1.EndpointSlice" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.398182 3780 reflector.go:436] "Caches populated" type="*v1.Namespace" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.401204 3780 server.go:113] MetricServer registers OVN Northd metrics I0514 16:08:34.405818 3780 metrics.go:519] OVN Metrics Server starts to run ... I0514 16:08:34.406881 3780 reflector.go:436] "Caches populated" type="*v1.Pod" reflector="k8s.io/client-go/informers/factory.go:160" I0514 16:08:34.480211 3780 factory.go:1977] *v1.NetworkPolicy informer cache synced successfully I0514 16:08:34.480243 3780 factory.go:1977] *v1.Namespace informer cache synced successfully I0514 16:08:34.480249 3780 factory.go:1977] *v1.Node informer cache synced successfully I0514 16:08:34.480254 3780 factory.go:1977] *v1.Service informer cache synced successfully I0514 16:08:34.480258 3780 factory.go:1977] *v1.EndpointSlice informer cache synced successfully I0514 16:08:34.480261 3780 factory.go:1977] *v1.Pod informer cache synced successfully I0514 16:08:34.480380 3780 reflector.go:358] "Starting reflector" type="*v1alpha1.BaselineAdminNetworkPolicy" resyncPeriod="0s" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0514 16:08:34.480422 3780 reflector.go:404] "Listing and watching" type="*v1alpha1.BaselineAdminNetworkPolicy" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0514 16:08:34.480516 3780 reflector.go:358] "Starting reflector" type="*v1alpha1.AdminNetworkPolicy" resyncPeriod="0s" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0514 16:08:34.480549 3780 reflector.go:404] "Listing and watching" type="*v1alpha1.AdminNetworkPolicy" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0514 16:08:34.484373 3780 reflector.go:436] "Caches populated" type="*v1alpha1.BaselineAdminNetworkPolicy" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0514 16:08:34.484381 3780 reflector.go:436] "Caches populated" type="*v1alpha1.AdminNetworkPolicy" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0514 16:08:34.484523 3780 factory.go:1977] *v1.Service informer cache synced successfully I0514 16:08:34.484535 3780 factory.go:1977] *v1.EndpointSlice informer cache synced successfully I0514 16:08:34.484538 3780 factory.go:1977] *v1.Pod informer cache synced successfully I0514 16:08:34.484542 3780 factory.go:1977] *v1.NetworkPolicy informer cache synced successfully I0514 16:08:34.484545 3780 factory.go:1977] *v1.Namespace informer cache synced successfully I0514 16:08:34.484548 3780 factory.go:1977] *v1.Node informer cache synced successfully I0514 16:08:34.486199 3780 factory.go:1977] *v1alpha1.AdminNetworkPolicy informer cache synced successfully I0514 16:08:34.486210 3780 factory.go:1977] *v1alpha1.BaselineAdminNetworkPolicy informer cache synced successfully I0514 16:08:34.486287 3780 reflector.go:358] "Starting reflector" type="*v1.EgressIP" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.486307 3780 reflector.go:404] "Listing and watching" type="*v1.EgressIP" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.489480 3780 reflector.go:436] "Caches populated" type="*v1.EgressIP" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.489588 3780 factory.go:1977] *v1alpha1.AdminNetworkPolicy informer cache synced successfully I0514 16:08:34.489599 3780 factory.go:1977] *v1alpha1.BaselineAdminNetworkPolicy informer cache synced successfully I0514 16:08:34.491474 3780 factory.go:1977] *v1.EgressIP informer cache synced successfully I0514 16:08:34.491540 3780 reflector.go:358] "Starting reflector" type="*v1.EgressFirewall" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.491552 3780 reflector.go:404] "Listing and watching" type="*v1.EgressFirewall" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.494750 3780 reflector.go:436] "Caches populated" type="*v1.EgressFirewall" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.494751 3780 factory.go:1977] *v1.EgressIP informer cache synced successfully I0514 16:08:34.496664 3780 factory.go:1977] *v1.EgressFirewall informer cache synced successfully I0514 16:08:34.496739 3780 reflector.go:358] "Starting reflector" type="*v1.EgressQoS" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.496755 3780 reflector.go:404] "Listing and watching" type="*v1.EgressQoS" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.499857 3780 factory.go:1977] *v1.EgressFirewall informer cache synced successfully I0514 16:08:34.499972 3780 reflector.go:436] "Caches populated" type="*v1.EgressQoS" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.501982 3780 factory.go:1977] *v1.EgressQoS informer cache synced successfully I0514 16:08:34.502054 3780 reflector.go:358] "Starting reflector" type="*v1.EgressService" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.502070 3780 reflector.go:404] "Listing and watching" type="*v1.EgressService" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.504924 3780 factory.go:1977] *v1.EgressQoS informer cache synced successfully I0514 16:08:34.505027 3780 reflector.go:436] "Caches populated" type="*v1.EgressService" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.507025 3780 factory.go:1977] *v1.EgressService informer cache synced successfully I0514 16:08:34.507101 3780 reflector.go:358] "Starting reflector" type="*v1.AdminPolicyBasedExternalRoute" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.507117 3780 reflector.go:404] "Listing and watching" type="*v1.AdminPolicyBasedExternalRoute" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.509976 3780 factory.go:1977] *v1.EgressService informer cache synced successfully I0514 16:08:34.510330 3780 reflector.go:436] "Caches populated" type="*v1.AdminPolicyBasedExternalRoute" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.512292 3780 factory.go:1977] *v1.AdminPolicyBasedExternalRoute informer cache synced successfully I0514 16:08:34.512470 3780 reflector.go:358] "Starting reflector" type="*v1.NetworkAttachmentDefinition" resyncPeriod="0s" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" I0514 16:08:34.512488 3780 reflector.go:404] "Listing and watching" type="*v1.NetworkAttachmentDefinition" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" I0514 16:08:34.515169 3780 factory.go:1977] *v1.AdminPolicyBasedExternalRoute informer cache synced successfully I0514 16:08:34.515808 3780 reflector.go:436] "Caches populated" type="*v1.NetworkAttachmentDefinition" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" I0514 16:08:34.517728 3780 factory.go:1977] *v1.NetworkAttachmentDefinition informer cache synced successfully I0514 16:08:34.517788 3780 reflector.go:358] "Starting reflector" type="*v1.ClusterUserDefinedNetwork" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.517800 3780 reflector.go:404] "Listing and watching" type="*v1.ClusterUserDefinedNetwork" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.518250 3780 reflector.go:358] "Starting reflector" type="*v1.UserDefinedNetwork" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.518267 3780 reflector.go:404] "Listing and watching" type="*v1.UserDefinedNetwork" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.520647 3780 factory.go:1977] *v1.NetworkAttachmentDefinition informer cache synced successfully I0514 16:08:34.521078 3780 reflector.go:436] "Caches populated" type="*v1.ClusterUserDefinedNetwork" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.521496 3780 reflector.go:436] "Caches populated" type="*v1.UserDefinedNetwork" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0514 16:08:34.523525 3780 factory.go:1977] *v1.UserDefinedNetwork informer cache synced successfully I0514 16:08:34.523536 3780 factory.go:1977] *v1.ClusterUserDefinedNetwork informer cache synced successfully I0514 16:08:34.523543 3780 factory.go:680] Watch Factory start up complete, took: 149.153731ms I0514 16:08:34.523611 3780 default_node_network_controller.go:171] Enable node proxy healthz server on 0.0.0.0:10256 I0514 16:08:34.525812 3780 factory.go:1977] *v1.UserDefinedNetwork informer cache synced successfully I0514 16:08:34.525821 3780 factory.go:1977] *v1.ClusterUserDefinedNetwork informer cache synced successfully I0514 16:08:34.525828 3780 factory.go:680] Watch Factory start up complete, took: 146.583677ms I0514 16:08:34.525834 3780 controller_manager.go:406] Waiting up to 5m0s for a node to have "ip-10-0-0-158.ec2.internal" zone I0514 16:08:34.568814 3780 default_node_network_controller.go:710] Initializing the default node network controller I0514 16:08:34.650725 3780 udn_isolation.go:101] Starting UDN host isolation manager I0514 16:08:34.656876 3780 udn_isolation.go:112] Found kubelet cgroup path: system.slice/kubelet.service I0514 16:08:34.729147 3780 controller.go:133] Adding controller udn-host-isolation-manager event handlers I0514 16:08:34.729304 3780 shared_informer.go:349] "Waiting for caches to sync" controller="udn-host-isolation-manager" I0514 16:08:34.729319 3780 shared_informer.go:356] "Caches are synced" controller="udn-host-isolation-manager" I0514 16:08:34.760082 3780 controller.go:157] Starting controller udn-host-isolation-manager with 1 workers I0514 16:08:34.760137 3780 default_node_network_controller.go:809] Node ip-10-0-0-158.ec2.internal ready for ovn initialization with subnet 10.129.0.0/23 I0514 16:08:34.765692 3780 kube.go:131] Setting annotations map[k8s.ovn.org/node-encap-ips:["10.0.0.158"] k8s.ovn.org/zone-name:ip-10-0-0-158.ec2.internal] on node ip-10-0-0-158.ec2.internal I0514 16:08:34.788504 3780 gateway_init.go:234] Initializing Gateway Functionality for Gateway PreStart I0514 16:08:34.794730 3780 helper_linux.go:93] Provided gateway interface "br-ex", found as index: 5 I0514 16:08:34.794912 3780 helper_linux.go:118] Found default gateway interface br-ex 10.0.0.1 I0514 16:08:34.795103 3780 gateway_init.go:263] Preparing Gateway I0514 16:08:34.795117 3780 gateway_shared_intf.go:1687] Creating new gateway I0514 16:08:34.877719 3780 iptables.go:108] Creating table: filter chain: FORWARD I0514 16:08:34.883607 3780 iptables.go:108] Creating table: filter chain: OUTPUT I0514 16:08:34.890170 3780 gateway_shared_intf.go:1816] Gateway Creation Complete I0514 16:08:34.890186 3780 kube.go:131] Setting annotations map[k8s.ovn.org/gateway-mtu-support: k8s.ovn.org/l3-gateway-config:{"default":{"mode":"shared","bridge-id":"br-ex","interface-id":"br-ex_ip-10-0-0-158.ec2.internal","mac-address":"12:78:fc:1e:a6:77","ip-addresses":["10.0.0.158/24"],"ip-address":"10.0.0.158/24","next-hops":["10.0.0.1"],"next-hop":"10.0.0.1","node-port-enable":"true","vlan-id":"0"}} k8s.ovn.org/node-chassis-id:206f44f8-e769-4521-ad97-6e1d777bbf92 k8s.ovn.org/node-encap-ips:["10.0.0.158"] k8s.ovn.org/node-primary-ifaddr:{"ipv4":"10.0.0.158/24"} k8s.ovn.org/zone-name:ip-10-0-0-158.ec2.internal] on node ip-10-0-0-158.ec2.internal I0514 16:08:34.915024 3780 default_node_network_controller.go:519] Node connection status = connected I0514 16:08:35.026700 3780 controller_manager.go:429] Waiting for node in zone sync took: 500.853521ms I0514 16:08:35.033509 3780 controller_manager.go:611] Switching to transit router for layer2 networks I0514 16:08:35.033522 3780 kube.go:131] Setting annotations map[k8s.ovn.org/layer2-topology-version:2.0] on node ip-10-0-0-158.ec2.internal I0514 16:08:35.051803 3780 metrics.go:164] OVN Kube log file not specified in config, therefore not starting the log file metric monitor I0514 16:08:35.058333 3780 controller_manager.go:290] SCTP support detected in OVN I0514 16:08:35.058914 3780 services_controller.go:69] Creating services controller for network=default I0514 16:08:35.058998 3780 udn_enabled_svc.go:80] Waiting for service informer to sync I0514 16:08:35.059124 3780 controller.go:133] Adding controller [zone-nad-controller NAD controller] event handlers I0514 16:08:35.059154 3780 shared_informer.go:349] "Waiting for caches to sync" controller="[zone-nad-controller NAD controller]" I0514 16:08:35.059223 3780 shared_informer.go:356] "Caches are synced" controller="[zone-nad-controller NAD controller]" I0514 16:08:35.059258 3780 nad_controller.go:541] [zone-nad-controller NAD controller]: 1 NADs are missing the network ID annotation, fetching from nodes I0514 16:08:35.059429 3780 udn_enabled_svc.go:95] Performing full resync I0514 16:08:35.059610 3780 controller.go:157] Starting controller [zone-nad-controller NAD controller] with 1 workers I0514 16:08:35.059684 3780 network_controller.go:373] [zone-nad-controller network controller]: syncing all networks I0514 16:08:35.059706 3780 nad_controller.go:580] [zone-nad-controller NAD controller]: finished syncing NAD openshift-ovn-kubernetes/default, took 62.015µs I0514 16:08:35.059733 3780 network_controller.go:402] [zone-nad-controller network controller]: finished syncing network default, took 38.051µs I0514 16:08:35.059743 3780 network_controller.go:384] [zone-nad-controller network controller]: finished syncing all networks. Time taken: 60.243µs I0514 16:08:35.059749 3780 controller.go:157] Starting controller [zone-nad-controller network controller] with 1 workers I0514 16:08:35.059750 3780 udn_enabled_svc.go:99] Waiting for handler to sync I0514 16:08:35.059755 3780 nad_controller.go:379] [zone-nad-controller NAD controller]: started I0514 16:08:35.059761 3780 udn_enabled_svc.go:104] Starting worker I0514 16:08:35.059763 3780 default_network_controller.go:325] Starting the default network controller I0514 16:08:35.059857 3780 network_controller.go:402] [zone-nad-controller network controller]: finished syncing network default, took 13.858µs I0514 16:08:35.069435 3780 default_network_controller.go:395] Cleaning External Gateway ECMP routes I0514 16:08:35.069487 3780 repair.go:33] Syncing exgw routes took 41.026µs I0514 16:08:35.069511 3780 default_network_controller.go:414] Starting all the Watchers... I0514 16:08:35.069657 3780 namespace.go:98] [openshift-console] adding namespace I0514 16:08:35.069665 3780 namespace.go:98] [kube-system] adding namespace I0514 16:08:35.069677 3780 namespace.go:98] [openshift-dns] adding namespace I0514 16:08:35.069679 3780 namespace.go:98] [openshift-host-network] adding namespace I0514 16:08:35.069659 3780 namespace.go:98] [openshift] adding namespace I0514 16:08:35.069677 3780 namespace.go:98] [default] adding namespace I0514 16:08:35.069680 3780 namespace.go:98] [openshift-deployment-validation-operator] adding namespace I0514 16:08:35.069693 3780 namespace.go:98] [openshift-network-node-identity] adding namespace I0514 16:08:35.069694 3780 namespace.go:98] [openshift-cloud-credential-operator] adding namespace I0514 16:08:35.069697 3780 namespace.go:98] [openshift-insights] adding namespace I0514 16:08:35.069696 3780 namespace.go:98] [openshift-operators] adding namespace I0514 16:08:35.069705 3780 namespace.go:98] [openshift-monitoring] adding namespace I0514 16:08:35.069708 3780 namespace.go:98] [openshift-kube-controller-manager-operator] adding namespace I0514 16:08:35.069714 3780 namespace.go:98] [openshift-must-gather-operator] adding namespace I0514 16:08:35.069659 3780 namespace.go:98] [openshift-kube-controller-manager] adding namespace I0514 16:08:35.070175 3780 namespace.go:102] [openshift-console] adding namespace took 505.298µs I0514 16:08:35.070191 3780 namespace.go:98] [openshift-kube-scheduler] adding namespace I0514 16:08:35.070627 3780 namespace.go:102] [openshift-kube-scheduler] adding namespace took 431.46µs I0514 16:08:35.070639 3780 namespace.go:98] [openshift-user-workload-monitoring] adding namespace W0514 16:08:35.071213 3780 base_network_controller_namespace.go:410] Failed to get IPs for pod kube-system/global-pull-secret-syncer-6htzf: pod kube-system/global-pull-secret-syncer-6htzf: no pod IPs found W0514 16:08:35.071243 3780 base_network_controller_namespace.go:410] Failed to get IPs for pod kube-system/global-pull-secret-syncer-7mwkt: pod kube-system/global-pull-secret-syncer-7mwkt: no pod IPs found I0514 16:08:35.071228 3780 namespace.go:102] [openshift-user-workload-monitoring] adding namespace took 582.85µs I0514 16:08:35.071316 3780 namespace.go:98] [openshift-backplane-srep-ro] adding namespace I0514 16:08:35.071983 3780 namespace.go:102] [kube-system] adding namespace took 2.309801ms I0514 16:08:35.071994 3780 namespace.go:98] [kube-node-lease] adding namespace I0514 16:08:35.072580 3780 namespace.go:102] [openshift-dns] adding namespace took 2.893102ms I0514 16:08:35.072590 3780 namespace.go:98] [openshift-controller-manager-operator] adding namespace I0514 16:08:35.073200 3780 namespace.go:102] [openshift-host-network] adding namespace took 3.507039ms I0514 16:08:35.073221 3780 namespace.go:98] [openshift-machine-config-operator] adding namespace I0514 16:08:35.073794 3780 namespace.go:102] [openshift] adding namespace took 4.099244ms I0514 16:08:35.073812 3780 namespace.go:98] [openshift-config] adding namespace I0514 16:08:35.074370 3780 namespace.go:102] [default] adding namespace took 4.67348ms I0514 16:08:35.074388 3780 namespace.go:98] [openshift-network-console] adding namespace I0514 16:08:35.074920 3780 namespace.go:102] [openshift-deployment-validation-operator] adding namespace took 5.222566ms I0514 16:08:35.074933 3780 namespace.go:98] [openshift-backplane-srep] adding namespace I0514 16:08:35.075389 3780 namespace.go:102] [openshift-network-node-identity] adding namespace took 5.690977ms I0514 16:08:35.075400 3780 namespace.go:98] [openshift-backplane-mcs-tier-two] adding namespace I0514 16:08:35.075843 3780 namespace.go:102] [openshift-cloud-credential-operator] adding namespace took 6.137916ms I0514 16:08:35.075861 3780 namespace.go:98] [openshift-cluster-samples-operator] adding namespace I0514 16:08:35.076446 3780 namespace.go:102] [openshift-insights] adding namespace took 6.737581ms I0514 16:08:35.076464 3780 namespace.go:98] [openshift-service-ca-operator] adding namespace I0514 16:08:35.077111 3780 namespace.go:102] [openshift-operators] adding namespace took 7.402132ms I0514 16:08:35.077126 3780 namespace.go:98] [openshift-backplane-managed-scripts] adding namespace I0514 16:08:35.078068 3780 namespace.go:102] [openshift-monitoring] adding namespace took 8.352906ms I0514 16:08:35.078083 3780 namespace.go:98] [openshift-console-operator] adding namespace I0514 16:08:35.078728 3780 namespace.go:102] [openshift-kube-controller-manager-operator] adding namespace took 9.008253ms I0514 16:08:35.078745 3780 namespace.go:98] [openshift-backplane-cee] adding namespace I0514 16:08:35.079323 3780 namespace.go:102] [openshift-must-gather-operator] adding namespace took 9.600934ms I0514 16:08:35.079336 3780 namespace.go:98] [openshift-backplane-mobb] adding namespace I0514 16:08:35.080023 3780 namespace.go:102] [openshift-kube-controller-manager] adding namespace took 10.25397ms I0514 16:08:35.080043 3780 namespace.go:98] [openshift-operators-redhat] adding namespace I0514 16:08:35.080485 3780 namespace.go:102] [openshift-backplane-srep-ro] adding namespace took 9.160917ms I0514 16:08:35.080502 3780 namespace.go:98] [openshift-config-managed] adding namespace I0514 16:08:35.081070 3780 namespace.go:102] [kube-node-lease] adding namespace took 9.069663ms I0514 16:08:35.081085 3780 namespace.go:98] [openshift-machine-api] adding namespace I0514 16:08:35.081678 3780 namespace.go:102] [openshift-controller-manager-operator] adding namespace took 9.080047ms I0514 16:08:35.081697 3780 namespace.go:98] [openshift-marketplace] adding namespace I0514 16:08:35.082191 3780 namespace.go:102] [openshift-machine-config-operator] adding namespace took 8.946281ms I0514 16:08:35.082206 3780 namespace.go:98] [openshift-customer-monitoring] adding namespace I0514 16:08:35.082861 3780 namespace.go:102] [openshift-config] adding namespace took 9.043816ms I0514 16:08:35.082872 3780 namespace.go:98] [openshift-backplane-lpsre] adding namespace I0514 16:08:35.083342 3780 namespace.go:102] [openshift-network-console] adding namespace took 8.947115ms I0514 16:08:35.083355 3780 namespace.go:98] [kube-public] adding namespace I0514 16:08:35.083940 3780 namespace.go:102] [openshift-backplane-srep] adding namespace took 9.001682ms I0514 16:08:35.083954 3780 namespace.go:98] [openshift-cluster-node-tuning-operator] adding namespace I0514 16:08:35.084353 3780 namespace.go:102] [openshift-backplane-mcs-tier-two] adding namespace took 8.949084ms I0514 16:08:35.084364 3780 namespace.go:98] [openshift-cluster-storage-operator] adding namespace I0514 16:08:35.084803 3780 namespace.go:102] [openshift-cluster-samples-operator] adding namespace took 8.937999ms I0514 16:08:35.084813 3780 namespace.go:98] [openshift-controller-manager] adding namespace I0514 16:08:35.085257 3780 namespace.go:102] [openshift-service-ca-operator] adding namespace took 8.786288ms I0514 16:08:35.085271 3780 namespace.go:98] [openshift-image-registry] adding namespace I0514 16:08:35.085660 3780 namespace.go:102] [openshift-backplane-managed-scripts] adding namespace took 8.530058ms I0514 16:08:35.085669 3780 namespace.go:98] [openshift-ingress-canary] adding namespace I0514 16:08:35.086158 3780 namespace.go:102] [openshift-console-operator] adding namespace took 8.07025ms I0514 16:08:35.086172 3780 namespace.go:98] [open-cluster-management-agent-addon] adding namespace I0514 16:08:35.086663 3780 namespace.go:102] [openshift-backplane-cee] adding namespace took 7.911077ms I0514 16:08:35.086676 3780 namespace.go:98] [openshift-etcd] adding namespace I0514 16:08:35.087117 3780 namespace.go:102] [openshift-backplane-mobb] adding namespace took 7.775936ms I0514 16:08:35.087130 3780 namespace.go:98] [openshift-apiserver-operator] adding namespace I0514 16:08:35.087523 3780 namespace.go:102] [openshift-operators-redhat] adding namespace took 7.474304ms I0514 16:08:35.087536 3780 namespace.go:98] [openshift-network-diagnostics] adding namespace I0514 16:08:35.087924 3780 namespace.go:102] [openshift-config-managed] adding namespace took 7.416663ms I0514 16:08:35.088335 3780 namespace.go:102] [openshift-machine-api] adding namespace took 7.24517ms I0514 16:08:35.088348 3780 namespace.go:98] [openshift-ovn-kubernetes] adding namespace I0514 16:08:35.088779 3780 namespace.go:102] [openshift-marketplace] adding namespace took 7.07554ms I0514 16:08:35.088796 3780 namespace.go:98] [openshift-backplane-ai-agent] adding namespace I0514 16:08:35.089301 3780 namespace.go:102] [openshift-customer-monitoring] adding namespace took 7.090142ms I0514 16:08:35.089315 3780 namespace.go:98] [openshift-apiserver] adding namespace I0514 16:08:35.089774 3780 namespace.go:102] [openshift-backplane-lpsre] adding namespace took 6.89544ms I0514 16:08:35.089787 3780 namespace.go:98] [openshift-kube-apiserver] adding namespace I0514 16:08:35.090244 3780 namespace.go:102] [kube-public] adding namespace took 6.884888ms I0514 16:08:35.090257 3780 namespace.go:98] [openshift-authentication-operator] adding namespace I0514 16:08:35.090603 3780 namespace.go:102] [openshift-cluster-node-tuning-operator] adding namespace took 6.643576ms I0514 16:08:35.090619 3780 namespace.go:98] [openshift-config-operator] adding namespace I0514 16:08:35.091096 3780 namespace.go:102] [openshift-cluster-storage-operator] adding namespace took 6.727904ms I0514 16:08:35.091109 3780 namespace.go:98] [openshift-kube-scheduler-operator] adding namespace I0514 16:08:35.091522 3780 namespace.go:102] [openshift-controller-manager] adding namespace took 6.705221ms I0514 16:08:35.091531 3780 namespace.go:98] [openshift-cluster-version] adding namespace I0514 16:08:35.091958 3780 namespace.go:102] [openshift-image-registry] adding namespace took 6.683311ms I0514 16:08:35.091967 3780 namespace.go:98] [openshift-multus] adding namespace I0514 16:08:35.092302 3780 namespace.go:102] [openshift-ingress-canary] adding namespace took 6.628397ms I0514 16:08:35.092310 3780 namespace.go:98] [openshift-operator-lifecycle-manager] adding namespace I0514 16:08:35.092782 3780 namespace.go:102] [open-cluster-management-agent-addon] adding namespace took 6.603477ms I0514 16:08:35.092796 3780 namespace.go:98] [openshift-ingress-operator] adding namespace I0514 16:08:35.093302 3780 namespace.go:102] [openshift-etcd] adding namespace took 6.621107ms I0514 16:08:35.093312 3780 namespace.go:98] [openshift-cluster-machine-approver] adding namespace W0514 16:08:35.093879 3780 base_network_controller_namespace.go:410] Failed to get IPs for pod openshift-network-diagnostics/network-check-target-gljng: pod openshift-network-diagnostics/network-check-target-gljng: no pod IPs found W0514 16:08:35.093898 3780 base_network_controller_namespace.go:410] Failed to get IPs for pod openshift-network-diagnostics/network-check-target-hmxmt: pod openshift-network-diagnostics/network-check-target-hmxmt: no pod IPs found I0514 16:08:35.093927 3780 namespace.go:102] [openshift-apiserver-operator] adding namespace took 6.790987ms I0514 16:08:35.093943 3780 namespace.go:98] [openshift-backplane-tam] adding namespace I0514 16:08:35.094502 3780 namespace.go:102] [openshift-network-diagnostics] adding namespace took 6.959777ms I0514 16:08:35.094518 3780 namespace.go:98] [openshift-route-controller-manager] adding namespace I0514 16:08:35.095169 3780 namespace.go:102] [openshift-ovn-kubernetes] adding namespace took 6.8166ms I0514 16:08:35.095181 3780 namespace.go:98] [openshift-dns-operator] adding namespace I0514 16:08:35.095735 3780 namespace.go:102] [openshift-backplane-ai-agent] adding namespace took 6.934435ms I0514 16:08:35.095746 3780 namespace.go:98] [openshift-network-operator] adding namespace I0514 16:08:35.096322 3780 namespace.go:102] [openshift-apiserver] adding namespace took 7.002028ms I0514 16:08:35.096336 3780 namespace.go:98] [openshift-console-user-settings] adding namespace I0514 16:08:35.096757 3780 namespace.go:102] [openshift-kube-apiserver] adding namespace took 6.965738ms I0514 16:08:35.096767 3780 namespace.go:98] [openshift-backplane-cse] adding namespace I0514 16:08:35.097275 3780 namespace.go:102] [openshift-authentication-operator] adding namespace took 7.010948ms I0514 16:08:35.097293 3780 namespace.go:98] [openshift-cloud-network-config-controller] adding namespace I0514 16:08:35.097778 3780 namespace.go:102] [openshift-config-operator] adding namespace took 7.151887ms I0514 16:08:35.097793 3780 namespace.go:98] [openshift-infra] adding namespace I0514 16:08:35.098348 3780 namespace.go:102] [openshift-kube-scheduler-operator] adding namespace took 7.233382ms I0514 16:08:35.098363 3780 namespace.go:98] [openshift-kube-apiserver-operator] adding namespace W0514 16:08:35.098802 3780 base_network_controller_namespace.go:410] Failed to get IPs for pod openshift-multus/network-metrics-daemon-ht2xn: pod openshift-multus/network-metrics-daemon-ht2xn: no pod IPs found W0514 16:08:35.098814 3780 base_network_controller_namespace.go:410] Failed to get IPs for pod openshift-multus/network-metrics-daemon-zcs8m: pod openshift-multus/network-metrics-daemon-zcs8m: no pod IPs found I0514 16:08:35.098835 3780 namespace.go:102] [openshift-cluster-version] adding namespace took 7.295278ms I0514 16:08:35.099268 3780 namespace.go:102] [openshift-multus] adding namespace took 7.296474ms I0514 16:08:35.099281 3780 namespace.go:98] [openshift-authentication] adding namespace I0514 16:08:35.100110 3780 namespace.go:102] [openshift-operator-lifecycle-manager] adding namespace took 7.795002ms I0514 16:08:35.100123 3780 namespace.go:98] [dedicated-admin] adding namespace I0514 16:08:35.100542 3780 namespace.go:102] [openshift-ingress-operator] adding namespace took 7.741373ms I0514 16:08:35.100555 3780 namespace.go:98] [open-cluster-management-2q9q434hk8vgse7t1urllt227aefs26c] adding namespace I0514 16:08:35.100936 3780 namespace.go:102] [openshift-cluster-machine-approver] adding namespace took 7.619734ms I0514 16:08:35.100948 3780 namespace.go:98] [openshift-node] adding namespace I0514 16:08:35.101319 3780 namespace.go:102] [openshift-backplane-tam] adding namespace took 7.371083ms I0514 16:08:35.101330 3780 namespace.go:98] [openshift-logging] adding namespace I0514 16:08:35.101875 3780 namespace.go:102] [openshift-route-controller-manager] adding namespace took 7.351419ms I0514 16:08:35.101886 3780 namespace.go:98] [openshift-backplane] adding namespace I0514 16:08:35.102314 3780 namespace.go:102] [openshift-dns-operator] adding namespace took 7.126106ms I0514 16:08:35.102332 3780 namespace.go:98] [openshift-backplane-csm] adding namespace I0514 16:08:35.102820 3780 namespace.go:102] [openshift-network-operator] adding namespace took 7.067793ms I0514 16:08:35.102831 3780 namespace.go:98] [openshift-kube-storage-version-migrator-operator] adding namespace I0514 16:08:35.103279 3780 namespace.go:102] [openshift-console-user-settings] adding namespace took 6.937979ms I0514 16:08:35.103292 3780 namespace.go:98] [openshift-package-operator] adding namespace I0514 16:08:35.103638 3780 namespace.go:102] [openshift-backplane-cse] adding namespace took 6.866452ms I0514 16:08:35.104074 3780 namespace.go:102] [openshift-cloud-network-config-controller] adding namespace took 6.775917ms I0514 16:08:35.104085 3780 namespace.go:98] [openshift-ingress] adding namespace I0514 16:08:35.104472 3780 namespace.go:102] [openshift-infra] adding namespace took 6.674862ms I0514 16:08:35.104845 3780 namespace.go:102] [openshift-kube-apiserver-operator] adding namespace took 6.47748ms I0514 16:08:35.105212 3780 namespace.go:102] [openshift-authentication] adding namespace took 5.928129ms I0514 16:08:35.105653 3780 namespace.go:102] [dedicated-admin] adding namespace took 5.524935ms I0514 16:08:35.105664 3780 namespace.go:98] [openshift-cluster-csi-drivers] adding namespace I0514 16:08:35.106049 3780 namespace.go:102] [open-cluster-management-2q9q434hk8vgse7t1urllt227aefs26c] adding namespace took 5.48999ms I0514 16:08:35.106449 3780 namespace.go:102] [openshift-node] adding namespace took 5.497201ms I0514 16:08:35.106835 3780 namespace.go:102] [openshift-logging] adding namespace took 5.501425ms I0514 16:08:35.107209 3780 namespace.go:102] [openshift-backplane] adding namespace took 5.317602ms I0514 16:08:35.107221 3780 namespace.go:98] [openshift-cloud-controller-manager] adding namespace I0514 16:08:35.107635 3780 namespace.go:102] [openshift-backplane-csm] adding namespace took 5.297706ms I0514 16:08:35.108015 3780 namespace.go:102] [openshift-kube-storage-version-migrator-operator] adding namespace took 5.178542ms I0514 16:08:35.108430 3780 namespace.go:102] [openshift-package-operator] adding namespace took 5.133786ms I0514 16:08:35.108915 3780 namespace.go:102] [openshift-ingress] adding namespace took 4.82448ms I0514 16:08:35.109312 3780 namespace.go:102] [openshift-cluster-csi-drivers] adding namespace took 3.642801ms I0514 16:08:35.109686 3780 namespace.go:102] [openshift-cloud-controller-manager] adding namespace took 2.459981ms W0514 16:08:35.109758 3780 master.go:471] Unable to parse remote node ip-10-0-2-139.ec2.internal chassis-id annotation. Chassis may be removed during sync I0514 16:08:35.109837 3780 master.go:548] Adding or Updating local node "ip-10-0-0-158.ec2.internal" for network "default" I0514 16:08:35.109895 3780 obj_retry.go:578] Failed to create *v1.Node ip-10-0-2-139.ec2.internal, error: adding or updating remote node chassis ip-10-0-2-139.ec2.internal failed, err - failed to create or update chassis to remote for remote node ip-10-0-2-139.ec2.internal, error: failed to parse node chassis-id for node - ip-10-0-2-139.ec2.internal, error: suppressed error logged: k8s.ovn.org/node-chassis-id annotation not found for node ip-10-0-2-139.ec2.internal I0514 16:08:35.110754 3780 zone_ic_handler.go:195] Time taken to create transit switch: 428.399µs I0514 16:08:35.110766 3780 zone_ic_handler.go:277] Creating interconnect resources for remote zone node ip-10-0-1-122.ec2.internal for the network default I0514 16:08:35.111646 3780 zone_ic_handler.go:282] Creating Interconnect resources for node "ip-10-0-1-122.ec2.internal" on network "default" took: 1.329536ms W0514 16:08:35.117736 3780 gateway.go:207] Unable to fetch podIPs for pod kube-system/global-pull-secret-syncer-6htzf: pod kube-system/global-pull-secret-syncer-6htzf: no pod IPs found W0514 16:08:35.117758 3780 gateway.go:207] Unable to fetch podIPs for pod openshift-network-diagnostics/network-check-target-hmxmt: pod openshift-network-diagnostics/network-check-target-hmxmt: no pod IPs found W0514 16:08:35.117770 3780 gateway.go:207] Unable to fetch podIPs for pod openshift-multus/network-metrics-daemon-ht2xn: pod openshift-multus/network-metrics-daemon-ht2xn: no pod IPs found I0514 16:08:35.118666 3780 zone_ic_handler.go:203] Creating interconnect resources for local zone node ip-10-0-0-158.ec2.internal for the network default I0514 16:08:35.118860 3780 zone_ic_handler.go:195] Time taken to create transit switch: 181.453µs I0514 16:08:35.119611 3780 services_controller.go:192] Starting controller ovn-lb-controller for network=default I0514 16:08:35.119633 3780 services_controller.go:200] Waiting for node tracker handler to sync for network=default I0514 16:08:35.119650 3780 shared_informer.go:349] "Waiting for caches to sync" controller="node-tracker-controller" I0514 16:08:35.119652 3780 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-1-122.ec2.internal in network "default" I0514 16:08:35.119719 3780 node_tracker.go:169] Node ip-10-0-1-122.ec2.internal switch + router changed, syncing services in network "default" I0514 16:08:35.119737 3780 services_controller.go:574] Full service sync requested for network=default I0514 16:08:35.119745 3780 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-2-139.ec2.internal in network "default" I0514 16:08:35.119754 3780 node_tracker.go:235] Node ip-10-0-2-139.ec2.internal has invalid / no gateway config: k8s.ovn.org/l3-gateway-config annotation not found for node "ip-10-0-2-139.ec2.internal" I0514 16:08:35.119764 3780 node_tracker.go:169] Node ip-10-0-2-139.ec2.internal switch + router changed, syncing services in network "default" I0514 16:08:35.119768 3780 services_controller.go:574] Full service sync requested for network=default I0514 16:08:35.119773 3780 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-0-158.ec2.internal in network "default" I0514 16:08:35.119795 3780 node_tracker.go:169] Node ip-10-0-0-158.ec2.internal switch + router changed, syncing services in network "default" I0514 16:08:35.119802 3780 services_controller.go:574] Full service sync requested for network=default I0514 16:08:35.220088 3780 shared_informer.go:356] "Caches are synced" controller="node-tracker-controller" I0514 16:08:35.220104 3780 services_controller.go:208] Setting up event handlers for services for network=default I0514 16:08:35.220156 3780 services_controller.go:218] Setting up event handlers for endpoint slices for network=default I0514 16:08:35.220200 3780 services_controller.go:233] Waiting for service and endpoint handlers to sync for network=default I0514 16:08:35.220210 3780 shared_informer.go:349] "Waiting for caches to sync" controller="ovn-lb-controller" I0514 16:08:35.320364 3780 shared_informer.go:356] "Caches are synced" controller="ovn-lb-controller" I0514 16:08:35.320381 3780 repair.go:60] Starting repairing loop for services I0514 16:08:35.320466 3780 repair.go:131] Deleted 0 stale service LBs I0514 16:08:35.320483 3780 repair.go:137] Deleted 0 stale Chassis Template Vars I0514 16:08:35.320500 3780 repair.go:62] Finished repairing loop for services: 121.175µs I0514 16:08:35.320523 3780 services_controller.go:363] Controller cache of 0 load balancers initialized for 0 services for network=default I0514 16:08:35.320533 3780 services_controller.go:254] Starting workers for network=default I0514 16:08:35.320871 3780 obj_retry.go:578] Failed to create *v1.Pod kube-system/global-pull-secret-syncer-7mwkt, error: failed to obtain IPs to add remote pod kube-system/global-pull-secret-syncer-7mwkt: suppressed error logged: pod kube-system/global-pull-secret-syncer-7mwkt: no pod IPs found I0514 16:08:35.320970 3780 base_network_controller_pods.go:487] [default/openshift-multus/network-metrics-daemon-ht2xn] creating logical port openshift-multus_network-metrics-daemon-ht2xn for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:35.321029 3780 base_network_controller_pods.go:487] [default/openshift-network-diagnostics/network-check-target-hmxmt] creating logical port openshift-network-diagnostics_network-check-target-hmxmt for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:35.320972 3780 base_network_controller_pods.go:487] [default/kube-system/global-pull-secret-syncer-6htzf] creating logical port kube-system_global-pull-secret-syncer-6htzf for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:35.321098 3780 kube.go:256] Updating pod openshift-multus/network-metrics-daemon-ht2xn I0514 16:08:35.321114 3780 kube.go:256] Updating pod openshift-network-diagnostics/network-check-target-hmxmt I0514 16:08:35.321141 3780 kube.go:256] Updating pod kube-system/global-pull-secret-syncer-6htzf I0514 16:08:35.321031 3780 obj_retry.go:578] Failed to create *v1.Pod openshift-network-diagnostics/network-check-target-gljng, error: failed to obtain IPs to add remote pod openshift-network-diagnostics/network-check-target-gljng: suppressed error logged: pod openshift-network-diagnostics/network-check-target-gljng: no pod IPs found I0514 16:08:35.321203 3780 obj_retry.go:578] Failed to create *v1.Pod openshift-multus/network-metrics-daemon-zcs8m, error: failed to obtain IPs to add remote pod openshift-multus/network-metrics-daemon-zcs8m: suppressed error logged: pod openshift-multus/network-metrics-daemon-zcs8m: no pod IPs found I0514 16:08:35.340258 3780 pod.go:62] [kube-system/global-pull-secret-syncer-6htzf] pod update took 19.139579ms I0514 16:08:35.340274 3780 base_network_controller_pods.go:951] [default/kube-system/global-pull-secret-syncer-6htzf] addLogicalPort annotation time took 19.156807ms I0514 16:08:35.341270 3780 pods.go:271] [kube-system/global-pull-secret-syncer-6htzf] addLogicalPort took 20.30991ms, libovsdb time 725.841µs I0514 16:08:35.342377 3780 pod.go:62] [openshift-network-diagnostics/network-check-target-hmxmt] pod update took 21.299007ms I0514 16:08:35.342393 3780 base_network_controller_pods.go:951] [default/openshift-network-diagnostics/network-check-target-hmxmt] addLogicalPort annotation time took 21.315876ms I0514 16:08:35.342398 3780 pod.go:62] [openshift-multus/network-metrics-daemon-ht2xn] pod update took 21.355189ms I0514 16:08:35.342431 3780 base_network_controller_pods.go:951] [default/openshift-multus/network-metrics-daemon-ht2xn] addLogicalPort annotation time took 21.388678ms I0514 16:08:35.343309 3780 pods.go:271] [openshift-network-diagnostics/network-check-target-hmxmt] addLogicalPort took 22.284933ms, libovsdb time 681.276µs I0514 16:08:35.343575 3780 pods.go:271] [openshift-multus/network-metrics-daemon-ht2xn] addLogicalPort took 22.610972ms, libovsdb time 827.082µs I0514 16:08:35.344129 3780 admin_network_policy_controller.go:227] Starting controller default-network-controller I0514 16:08:35.344156 3780 shared_informer.go:349] "Waiting for caches to sync" controller="default-network-controller" I0514 16:08:35.344169 3780 shared_informer.go:356] "Caches are synced" controller="default-network-controller" I0514 16:08:35.344176 3780 admin_network_policy_controller.go:237] Repairing Admin Network Policies I0514 16:08:35.344380 3780 repair.go:29] Repairing admin network policies took 196.761µs I0514 16:08:35.344557 3780 repair.go:92] Repairing baseline admin network policies took 164.535µs I0514 16:08:35.345045 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/allow-ingress-webhook for network default I0514 16:08:35.345062 3780 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/allow-ingress-webhook added to peer address sets [] I0514 16:08:35.347230 3780 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/allow-ingress-webhook resources completed, update namespace loglevel I0514 16:08:35.347250 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-egress-to-api-server for network default I0514 16:08:35.347259 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-egress-to-api-server added to peer address sets [] I0514 16:08:35.349290 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-egress-to-api-server resources completed, update namespace loglevel I0514 16:08:35.349308 3780 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/catalog-operator for network default I0514 16:08:35.349319 3780 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/catalog-operator added to peer address sets [] I0514 16:08:35.352448 3780 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/catalog-operator resources completed, update namespace loglevel I0514 16:08:35.352466 3780 base_network_controller_policy.go:1179] Adding network policy openshift-operators/default-allow-all for network default I0514 16:08:35.352474 3780 base_network_controller_policy.go:1054] Policy openshift-operators/default-allow-all added to peer address sets [] I0514 16:08:35.354165 3780 base_network_controller_policy.go:1232] Create network policy openshift-operators/default-allow-all resources completed, update namespace loglevel I0514 16:08:35.354183 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-ingress-to-operator-metrics-range for network default I0514 16:08:35.354189 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-ingress-to-operator-metrics-range added to peer address sets [] I0514 16:08:35.354653 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-ingress-to-operator-metrics-range resources completed, update namespace loglevel I0514 16:08:35.354676 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-version/default-deny for network default I0514 16:08:35.354686 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-version/default-deny added to peer address sets [] I0514 16:08:35.356183 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-version/default-deny resources completed, update namespace loglevel I0514 16:08:35.356201 3780 base_network_controller_policy.go:1179] Adding network policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-grpc-server for network default I0514 16:08:35.356209 3780 base_network_controller_policy.go:1054] Policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-grpc-server added to peer address sets [] I0514 16:08:35.357983 3780 base_network_controller_policy.go:1232] Create network policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-grpc-server resources completed, update namespace loglevel I0514 16:08:35.357999 3780 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/default-deny-all-traffic for network default I0514 16:08:35.358005 3780 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/default-deny-all-traffic added to peer address sets [] I0514 16:08:35.358304 3780 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/default-deny-all-traffic resources completed, update namespace loglevel I0514 16:08:35.358322 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/allow-egress for network default I0514 16:08:35.358333 3780 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/allow-egress added to peer address sets [] I0514 16:08:35.359141 3780 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/allow-egress resources completed, update namespace loglevel I0514 16:08:35.359157 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/default-deny for network default I0514 16:08:35.359163 3780 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/default-deny added to peer address sets [] I0514 16:08:35.359488 3780 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/default-deny resources completed, update namespace loglevel I0514 16:08:35.359504 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-to-dns for network default I0514 16:08:35.360448 3780 pod_selector_address_set.go:213] Created shared address set for pod selector LS{ML:{kubernetes.io/metadata.name: openshift-dns,},}_LS{ML:{dns.operator.openshift.io/daemonset-dns: default,},} I0514 16:08:35.360462 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-to-dns added to peer address sets [LS{ML:{kubernetes.io/metadata.name: openshift-dns,},}_LS{ML:{dns.operator.openshift.io/daemonset-dns: default,},}] I0514 16:08:35.361424 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-to-dns resources completed, update namespace loglevel I0514 16:08:35.361446 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/default-deny for network default I0514 16:08:35.361453 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/default-deny added to peer address sets [] I0514 16:08:35.362917 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/default-deny resources completed, update namespace loglevel I0514 16:08:35.362938 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/allow-ingress-to-operator-metrics for network default I0514 16:08:35.362947 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/allow-ingress-to-operator-metrics added to peer address sets [] I0514 16:08:35.363515 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/allow-ingress-to-operator-metrics resources completed, update namespace loglevel I0514 16:08:35.363532 3780 base_network_controller_policy.go:1179] Adding network policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-unpack-bundles for network default I0514 16:08:35.363541 3780 base_network_controller_policy.go:1054] Policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-unpack-bundles added to peer address sets [] I0514 16:08:35.365023 3780 base_network_controller_policy.go:1232] Create network policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-unpack-bundles resources completed, update namespace loglevel I0514 16:08:35.365049 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/allow-egress-to-api-server for network default I0514 16:08:35.365061 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/allow-egress-to-api-server added to peer address sets [] I0514 16:08:35.365668 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/allow-egress-to-api-server resources completed, update namespace loglevel I0514 16:08:35.365683 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-all-egress for network default I0514 16:08:35.365691 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-all-egress added to peer address sets [] I0514 16:08:35.366321 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-all-egress resources completed, update namespace loglevel I0514 16:08:35.366339 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-ingress-to-metrics-range for network default I0514 16:08:35.366347 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-ingress-to-metrics-range added to peer address sets [] I0514 16:08:35.366867 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-ingress-to-metrics-range resources completed, update namespace loglevel I0514 16:08:35.366884 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/allow-all-egress for network default I0514 16:08:35.366892 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/allow-all-egress added to peer address sets [] I0514 16:08:35.367495 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/allow-all-egress resources completed, update namespace loglevel I0514 16:08:35.367513 3780 base_network_controller_policy.go:1179] Adding network policy openshift-deployment-validation-operator/allow-from-openshift-olm for network default I0514 16:08:35.367520 3780 base_network_controller_policy.go:1054] Policy openshift-deployment-validation-operator/allow-from-openshift-olm added to peer address sets [] I0514 16:08:35.368066 3780 base_network_controller_policy.go:1232] Create network policy openshift-deployment-validation-operator/allow-from-openshift-olm resources completed, update namespace loglevel I0514 16:08:35.368082 3780 base_network_controller_policy.go:1179] Adding network policy openshift-marketplace/default-deny-all for network default I0514 16:08:35.368090 3780 base_network_controller_policy.go:1054] Policy openshift-marketplace/default-deny-all added to peer address sets [] I0514 16:08:35.369553 3780 base_network_controller_policy.go:1232] Create network policy openshift-marketplace/default-deny-all resources completed, update namespace loglevel I0514 16:08:35.369571 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/allow-ingress-pprof for network default I0514 16:08:35.369578 3780 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/allow-ingress-pprof added to peer address sets [] I0514 16:08:35.370131 3780 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/allow-ingress-pprof resources completed, update namespace loglevel I0514 16:08:35.370154 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/allow-to-dns for network default I0514 16:08:35.370184 3780 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/allow-to-dns added to peer address sets [LS{ML:{kubernetes.io/metadata.name: openshift-dns,},}_LS{ML:{dns.operator.openshift.io/daemonset-dns: default,},}] I0514 16:08:35.371067 3780 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/allow-to-dns resources completed, update namespace loglevel I0514 16:08:35.371085 3780 base_network_controller_policy.go:1179] Adding network policy openshift-deployment-validation-operator/allow-from-openshift-insights for network default I0514 16:08:35.371093 3780 base_network_controller_policy.go:1054] Policy openshift-deployment-validation-operator/allow-from-openshift-insights added to peer address sets [] I0514 16:08:35.372195 3780 base_network_controller_policy.go:1232] Create network policy openshift-deployment-validation-operator/allow-from-openshift-insights resources completed, update namespace loglevel I0514 16:08:35.372210 3780 base_network_controller_policy.go:1179] Adding network policy openshift-marketplace/unpack-bundles for network default I0514 16:08:35.372219 3780 base_network_controller_policy.go:1054] Policy openshift-marketplace/unpack-bundles added to peer address sets [] I0514 16:08:35.373213 3780 base_network_controller_policy.go:1232] Create network policy openshift-marketplace/unpack-bundles resources completed, update namespace loglevel I0514 16:08:35.373236 3780 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/olm-operator for network default I0514 16:08:35.373247 3780 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/olm-operator added to peer address sets [] I0514 16:08:35.374526 3780 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/olm-operator resources completed, update namespace loglevel I0514 16:08:35.374545 3780 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/allow-ingress-metrics for network default I0514 16:08:35.374553 3780 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/allow-ingress-metrics added to peer address sets [] I0514 16:08:35.375113 3780 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/allow-ingress-metrics resources completed, update namespace loglevel I0514 16:08:35.375135 3780 base_network_controller_policy.go:1179] Adding network policy openshift-marketplace/marketplace-operator for network default I0514 16:08:35.375149 3780 base_network_controller_policy.go:1054] Policy openshift-marketplace/marketplace-operator added to peer address sets [] I0514 16:08:35.376352 3780 base_network_controller_policy.go:1232] Create network policy openshift-marketplace/marketplace-operator resources completed, update namespace loglevel I0514 16:08:35.376371 3780 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/package-server-manager for network default I0514 16:08:35.376384 3780 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/package-server-manager added to peer address sets [] I0514 16:08:35.377633 3780 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/package-server-manager resources completed, update namespace loglevel I0514 16:08:35.377652 3780 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/packageserver for network default I0514 16:08:35.377664 3780 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/packageserver added to peer address sets [] I0514 16:08:35.379148 3780 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/packageserver resources completed, update namespace loglevel I0514 16:08:35.385904 3780 egressfirewall.go:374] Starting EgressFirewall controller I0514 16:08:35.385942 3780 controller.go:133] Adding controller egress-firewall-controller event handlers I0514 16:08:35.385943 3780 controller.go:133] Adding controller egress-firewall-controller-node event handlers I0514 16:08:35.385971 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egress-firewall-controller" I0514 16:08:35.385977 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egress-firewall-controller-node" I0514 16:08:35.385987 3780 shared_informer.go:356] "Caches are synced" controller="egress-firewall-controller" I0514 16:08:35.385991 3780 shared_informer.go:356] "Caches are synced" controller="egress-firewall-controller-node" I0514 16:08:35.386240 3780 controller.go:157] Starting controller egress-firewall-controller with 1 workers I0514 16:08:35.386256 3780 controller.go:157] Starting controller egress-firewall-controller-node with 1 workers I0514 16:08:35.386264 3780 controller.go:157] Starting controller egress-firewall-controller-NAD with 1 workers I0514 16:08:35.386276 3780 egressqos.go:193] Setting up event handlers for EgressQoS I0514 16:08:35.386288 3780 egressfirewall.go:723] Syncing node "ip-10-0-0-158.ec2.internal" for egress firewall I0514 16:08:35.386297 3780 egressfirewall.go:723] Syncing node "ip-10-0-1-122.ec2.internal" for egress firewall I0514 16:08:35.386301 3780 egressfirewall.go:723] Syncing node "ip-10-0-2-139.ec2.internal" for egress firewall I0514 16:08:35.386422 3780 egressqos.go:245] Starting EgressQoS Controller I0514 16:08:35.386437 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressqosnodes" I0514 16:08:35.386446 3780 shared_informer.go:356] "Caches are synced" controller="egressqosnodes" I0514 16:08:35.386454 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressqospods" I0514 16:08:35.386461 3780 shared_informer.go:356] "Caches are synced" controller="egressqospods" I0514 16:08:35.386469 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressqos" I0514 16:08:35.386478 3780 shared_informer.go:356] "Caches are synced" controller="egressqos" I0514 16:08:35.386484 3780 egressqos.go:259] Repairing EgressQoSes I0514 16:08:35.386489 3780 egressqos.go:399] Starting repairing loop for egressqos I0514 16:08:35.386571 3780 egressqos.go:401] Finished repairing loop for egressqos: 82.095µs I0514 16:08:35.386586 3780 egressservice_zone.go:132] Setting up event handlers for Egress Services I0514 16:08:35.386614 3780 egressqos.go:1007] Processing sync for EgressQoS node ip-10-0-0-158.ec2.internal I0514 16:08:35.386668 3780 egressqos.go:1010] Finished syncing EgressQoS node ip-10-0-0-158.ec2.internal : 56.466µs I0514 16:08:35.386701 3780 egressservice_zone.go:210] Starting Egress Services Controller I0514 16:08:35.386713 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices" I0514 16:08:35.386721 3780 shared_informer.go:356] "Caches are synced" controller="egressservices" I0514 16:08:35.386732 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_services" I0514 16:08:35.386741 3780 shared_informer.go:356] "Caches are synced" controller="egressservices_services" I0514 16:08:35.386751 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_endpointslices" I0514 16:08:35.386758 3780 shared_informer.go:356] "Caches are synced" controller="egressservices_endpointslices" I0514 16:08:35.386766 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_nodes" I0514 16:08:35.386773 3780 shared_informer.go:356] "Caches are synced" controller="egressservices_nodes" I0514 16:08:35.386778 3780 egressservice_zone.go:228] Repairing Egress Services I0514 16:08:35.387042 3780 master_controller.go:88] Starting Admin Policy Based Route Controller I0514 16:08:35.387050 3780 external_controller.go:278] Starting Admin Policy Based Route Controller I0514 16:08:35.387070 3780 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-0-158.ec2.internal I0514 16:08:35.387083 3780 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-0-158.ec2.internal: 15.643µs I0514 16:08:35.387093 3780 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-1-122.ec2.internal I0514 16:08:35.387106 3780 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-1-122.ec2.internal: 14.243µs I0514 16:08:35.387114 3780 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-2-139.ec2.internal I0514 16:08:35.387120 3780 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-2-139.ec2.internal: 6.356µs I0514 16:08:35.387217 3780 default_network_controller.go:585] Completing all the Watchers took 317.697822ms I0514 16:08:35.387231 3780 default_network_controller.go:589] Starting unidling controllers I0514 16:08:35.387248 3780 unidle.go:46] Registering OVN SB ControllerEvent handler I0514 16:08:35.387263 3780 unidle.go:63] Populating Initial ContollerEvent events I0514 16:08:35.387279 3780 unidle.go:79] Setting up event handlers for services I0514 16:08:35.387332 3780 ovnkube.go:545] Waiting for OVN northbound database changes to be processed by ovn-controller I0514 16:08:35.416379 3780 default_node_network_controller.go:519] Node connection status = connected I0514 16:08:35.915519 3780 default_node_network_controller.go:519] Node connection status = connected I0514 16:08:36.415564 3780 default_node_network_controller.go:519] Node connection status = connected I0514 16:08:36.916205 3780 default_node_network_controller.go:519] Node connection status = connected I0514 16:08:37.187946 3780 ovnkube.go:550] Finished waiting for OVN northbound database changes to be processed by ovn-controller I0514 16:08:37.325371 3780 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-2-139.ec2.internal in network "default" I0514 16:08:37.325401 3780 node_tracker.go:235] Node ip-10-0-2-139.ec2.internal has invalid / no gateway config: k8s.ovn.org/l3-gateway-config annotation not found for node "ip-10-0-2-139.ec2.internal" I0514 16:08:37.325438 3780 node_tracker.go:169] Node ip-10-0-2-139.ec2.internal switch + router changed, syncing services in network "default" I0514 16:08:37.325452 3780 services_controller.go:574] Full service sync requested for network=default I0514 16:08:37.325494 3780 default_network_controller.go:1025] Node "ip-10-0-2-139.ec2.internal" in remote zone "ip-10-0-2-139.ec2.internal", network "default", needs interconnect zone sync up. Zone cluster changed: false I0514 16:08:37.362134 3780 default_network_controller.go:1025] Node "ip-10-0-2-139.ec2.internal" in remote zone "ip-10-0-2-139.ec2.internal", network "default", needs interconnect zone sync up. Zone cluster changed: false I0514 16:08:37.416121 3780 default_node_network_controller.go:519] Node connection status = connected I0514 16:08:37.433769 3780 bridgeconfig.go:537] Gateway is ready I0514 16:08:37.433790 3780 gateway_shared_intf.go:1733] Creating Gateway Openflow Manager I0514 16:08:37.445950 3780 node_ip_handler_linux.go:247] Node primary address changed to 10.0.0.158. Updating OVN encap IP. I0514 16:08:37.451641 3780 node_ip_handler_linux.go:542] Will not update encap IP 10.0.0.158 - it is already configured I0514 16:08:37.451657 3780 node_ip_handler_linux.go:485] Node address changed to map[10.0.0.158/24:{} 172.20.0.1/32:{}]. Updating annotations. I0514 16:08:37.452099 3780 kube.go:131] Setting annotations map[k8s.ovn.org/host-cidrs:["10.0.0.158/24","172.20.0.1/32"] k8s.ovn.org/l3-gateway-config:{"default":{"mode":"shared","bridge-id":"br-ex","interface-id":"br-ex_ip-10-0-0-158.ec2.internal","mac-address":"12:78:fc:1e:a6:77","ip-addresses":["10.0.0.158/24"],"ip-address":"10.0.0.158/24","next-hops":["10.0.0.1"],"next-hop":"10.0.0.1","node-port-enable":"true","vlan-id":"0"}} k8s.ovn.org/node-chassis-id:206f44f8-e769-4521-ad97-6e1d777bbf92 k8s.ovn.org/node-primary-ifaddr:{"ipv4":"10.0.0.158/24"}] on node ip-10-0-0-158.ec2.internal I0514 16:08:37.471807 3780 egressfirewall.go:723] Syncing node "ip-10-0-0-158.ec2.internal" for egress firewall I0514 16:08:37.471872 3780 master.go:548] Adding or Updating local node "ip-10-0-0-158.ec2.internal" for network "default" I0514 16:08:37.471843 3780 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-0-158.ec2.internal I0514 16:08:37.471905 3780 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-0-158.ec2.internal: 63.875µs I0514 16:08:37.471874 3780 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-0-158.ec2.internal in network "default" I0514 16:08:37.472036 3780 node_tracker.go:169] Node ip-10-0-0-158.ec2.internal switch + router changed, syncing services in network "default" I0514 16:08:37.472061 3780 services_controller.go:574] Full service sync requested for network=default I0514 16:08:37.472120 3780 gateway_shared_intf.go:2023] Setting OVN Masquerade route with source: 10.0.0.158 I0514 16:08:37.472233 3780 kube.go:131] Setting annotations map[k8s.ovn.org/node-masquerade-subnet:{"ipv4":"169.254.0.0/17","ipv6":"fd69::/112"}] on node ip-10-0-0-158.ec2.internal I0514 16:08:37.476921 3780 default_network_controller.go:1084] Egress IP detected IP address change for node ip-10-0-0-158.ec2.internal. Updating no re-route policies I0514 16:08:37.477011 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-98qgx I0514 16:08:37.477019 3780 obj_retry.go:334] Retry object setup: *v1.Pod kube-system/global-pull-secret-syncer-6htzf I0514 16:08:37.477018 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-0-158.ec2.internal I0514 16:08:37.477037 3780 obj_retry.go:398] Adding new object: *v1.Pod kube-system/global-pull-secret-syncer-6htzf I0514 16:08:37.477035 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-zcs8m I0514 16:08:37.477042 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jd49s I0514 16:08:37.477048 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-dns/node-resolver-srctw I0514 16:08:37.477048 3780 obj_retry.go:334] Retry object setup: *v1.Pod kube-system/konnectivity-agent-hfkzv I0514 16:08:37.477052 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jd49s I0514 16:08:37.477054 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-dns/node-resolver-srctw I0514 16:08:37.477062 3780 obj_retry.go:398] Adding new object: *v1.Pod kube-system/konnectivity-agent-hfkzv I0514 16:08:37.477063 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-cluster-node-tuning-operator/tuned-7slk4 I0514 16:08:37.477070 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bcl9q I0514 16:08:37.477082 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-cluster-node-tuning-operator/tuned-7slk4 I0514 16:08:37.477084 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-dns/node-resolver-srctw after 0 failed attempt(s) I0514 16:08:37.477087 3780 obj_retry.go:334] Retry object setup: *v1.Pod kube-system/global-pull-secret-syncer-7mwkt I0514 16:08:37.477088 3780 obj_retry.go:418] Retry successful for *v1.Pod kube-system/konnectivity-agent-hfkzv after 0 failed attempt(s) I0514 16:08:37.477088 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-hmxmt I0514 16:08:37.477096 3780 obj_retry.go:398] Adding new object: *v1.Pod kube-system/global-pull-secret-syncer-7mwkt I0514 16:08:37.477095 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-multus/multus-d58vb I0514 16:08:37.477098 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-image-registry/node-ca-tjnkr I0514 16:08:37.477103 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-cluster-node-tuning-operator/tuned-7slk4 after 0 failed attempt(s) I0514 16:08:37.477106 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-hmxmt I0514 16:08:37.477128 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-ht2xn I0514 16:08:37.477138 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-gljng I0514 16:08:37.477136 3780 obj_retry.go:409] Retry add failed for *v1.Pod kube-system/global-pull-secret-syncer-7mwkt, will try again later: failed to obtain IPs to add remote pod kube-system/global-pull-secret-syncer-7mwkt: suppressed error logged: pod kube-system/global-pull-secret-syncer-7mwkt: no pod IPs found I0514 16:08:37.477141 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-ht2xn I0514 16:08:37.477145 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-gljng I0514 16:08:37.477150 3780 base_network_controller_pods.go:487] [default/openshift-network-diagnostics/network-check-target-hmxmt] creating logical port openshift-network-diagnostics_network-check-target-hmxmt for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:37.477170 3780 base_network_controller_pods.go:487] [default/openshift-multus/network-metrics-daemon-ht2xn] creating logical port openshift-multus_network-metrics-daemon-ht2xn for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:37.477172 3780 obj_retry.go:409] Retry add failed for *v1.Pod openshift-network-diagnostics/network-check-target-gljng, will try again later: failed to obtain IPs to add remote pod openshift-network-diagnostics/network-check-target-gljng: suppressed error logged: pod openshift-network-diagnostics/network-check-target-gljng: no pod IPs found I0514 16:08:37.477031 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-98qgx I0514 16:08:37.477198 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-98qgx after 0 failed attempt(s) I0514 16:08:37.477039 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-0-158.ec2.internal I0514 16:08:37.477224 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-0-158.ec2.internal after 0 failed attempt(s) I0514 16:08:37.477072 3780 obj_retry.go:334] Retry object setup: *v1.Pod kube-system/kube-apiserver-proxy-ip-10-0-0-158.ec2.internal I0514 16:08:37.477086 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bcl9q I0514 16:08:37.477253 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-bcl9q after 0 failed attempt(s) I0514 16:08:37.477052 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-zcs8m I0514 16:08:37.477294 3780 obj_retry.go:409] Retry add failed for *v1.Pod openshift-multus/network-metrics-daemon-zcs8m, will try again later: failed to obtain IPs to add remote pod openshift-multus/network-metrics-daemon-zcs8m: suppressed error logged: pod openshift-multus/network-metrics-daemon-zcs8m: no pod IPs found I0514 16:08:37.477094 3780 base_network_controller_pods.go:487] [default/kube-system/global-pull-secret-syncer-6htzf] creating logical port kube-system_global-pull-secret-syncer-6htzf for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:37.477095 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-jd49s after 0 failed attempt(s) I0514 16:08:37.477112 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-multus/multus-d58vb I0514 16:08:37.477111 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-image-registry/node-ca-tjnkr I0514 16:08:37.477429 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-multus/multus-d58vb after 0 failed attempt(s) I0514 16:08:37.477455 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-image-registry/node-ca-tjnkr after 0 failed attempt(s) I0514 16:08:37.477115 3780 obj_retry.go:334] Retry object setup: *v1.Pod openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-9pm2v I0514 16:08:37.477479 3780 obj_retry.go:398] Adding new object: *v1.Pod openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-9pm2v I0514 16:08:37.477499 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-9pm2v after 0 failed attempt(s) I0514 16:08:37.477233 3780 obj_retry.go:398] Adding new object: *v1.Pod kube-system/kube-apiserver-proxy-ip-10-0-0-158.ec2.internal I0514 16:08:37.477536 3780 obj_retry.go:418] Retry successful for *v1.Pod kube-system/kube-apiserver-proxy-ip-10-0-0-158.ec2.internal after 0 failed attempt(s) I0514 16:08:37.477918 3780 pods.go:271] [openshift-network-diagnostics/network-check-target-hmxmt] addLogicalPort took 773.378µs, libovsdb time 295.07µs I0514 16:08:37.477929 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-target-hmxmt after 0 failed attempt(s) I0514 16:08:37.478127 3780 pods.go:271] [openshift-multus/network-metrics-daemon-ht2xn] addLogicalPort took 960.095µs, libovsdb time 338.489µs I0514 16:08:37.478141 3780 obj_retry.go:418] Retry successful for *v1.Pod openshift-multus/network-metrics-daemon-ht2xn after 0 failed attempt(s) I0514 16:08:37.478239 3780 pods.go:271] [kube-system/global-pull-secret-syncer-6htzf] addLogicalPort took 1.161102ms, libovsdb time 262.573µs I0514 16:08:37.478248 3780 obj_retry.go:418] Retry successful for *v1.Pod kube-system/global-pull-secret-syncer-6htzf after 0 failed attempt(s) I0514 16:08:37.490498 3780 gateway_shared_intf.go:1799] Creating Gateway Node Port Watcher I0514 16:08:37.502769 3780 iptables.go:108] Creating table: mangle chain: OUTPUT I0514 16:08:37.506156 3780 iptables.go:108] Creating table: nat chain: OUTPUT I0514 16:08:37.562577 3780 iptables.go:108] Creating table: nat chain: PREROUTING I0514 16:08:37.566394 3780 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-2-139.ec2.internal in network "default" I0514 16:08:37.566470 3780 node_tracker.go:169] Node ip-10-0-2-139.ec2.internal switch + router changed, syncing services in network "default" I0514 16:08:37.566487 3780 services_controller.go:574] Full service sync requested for network=default I0514 16:08:37.566529 3780 default_network_controller.go:1025] Node "ip-10-0-2-139.ec2.internal" in remote zone "ip-10-0-2-139.ec2.internal", network "default", needs interconnect zone sync up. Zone cluster changed: false I0514 16:08:37.567582 3780 zone_ic_handler.go:195] Time taken to create transit switch: 316.44µs I0514 16:08:37.567601 3780 zone_ic_handler.go:277] Creating interconnect resources for remote zone node ip-10-0-2-139.ec2.internal for the network default I0514 16:08:37.568766 3780 zone_ic_handler.go:282] Creating Interconnect resources for node "ip-10-0-2-139.ec2.internal" on network "default" took: 1.513073ms I0514 16:08:37.622452 3780 iptables.go:108] Creating table: filter chain: FORWARD I0514 16:08:37.639591 3780 gateway_shared_intf.go:2124] Ensuring IP Neighbor entry for: 169.254.0.1 W0514 16:08:37.639833 3780 gateway_shared_intf.go:2130] Failed to remove IP neighbor entry for ip 169.254.0.1, on iface br-ex: failed to delete neighbour entry 169.254.0.1 : no such file or directory I0514 16:08:37.639884 3780 gateway_shared_intf.go:2124] Ensuring IP Neighbor entry for: 169.254.0.4 W0514 16:08:37.639950 3780 gateway_shared_intf.go:2130] Failed to remove IP neighbor entry for ip 169.254.0.4, on iface br-ex: failed to delete neighbour entry 169.254.0.4 : no such file or directory I0514 16:08:37.640001 3780 gateway_init.go:325] Gateway and management port readiness took 2.728800883s I0514 16:08:37.640069 3780 controller.go:133] Adding controller [node-nad-controller NAD controller] event handlers I0514 16:08:37.640117 3780 shared_informer.go:349] "Waiting for caches to sync" controller="[node-nad-controller NAD controller]" I0514 16:08:37.640131 3780 shared_informer.go:356] "Caches are synced" controller="[node-nad-controller NAD controller]" I0514 16:08:37.640140 3780 nad_controller.go:541] [node-nad-controller NAD controller]: 1 NADs are missing the network ID annotation, fetching from nodes I0514 16:08:37.640224 3780 controller.go:157] Starting controller [node-nad-controller NAD controller] with 1 workers I0514 16:08:37.640321 3780 nad_controller.go:580] [node-nad-controller NAD controller]: finished syncing NAD openshift-ovn-kubernetes/default, took 63.084µs I0514 16:08:37.640637 3780 network_controller.go:373] [node-nad-controller network controller]: syncing all networks I0514 16:08:37.640668 3780 network_controller.go:402] [node-nad-controller network controller]: finished syncing network default, took 20.684µs I0514 16:08:37.640677 3780 network_controller.go:384] [node-nad-controller network controller]: finished syncing all networks. Time taken: 41.065µs I0514 16:08:37.640683 3780 controller.go:157] Starting controller [node-nad-controller network controller] with 1 workers I0514 16:08:37.640692 3780 nad_controller.go:379] [node-nad-controller NAD controller]: started I0514 16:08:37.640697 3780 default_node_network_controller.go:899] Starting the default node network controller I0514 16:08:37.640702 3780 gateway_init.go:335] Initializing Gateway Functionality for gateway Start I0514 16:08:37.640730 3780 network_controller.go:402] [node-nad-controller network controller]: finished syncing network default, took 16.909µs I0514 16:08:37.640916 3780 default_node_network_controller.go:1460] MTU (9001) of network interface br-ex is big enough to deal with Geneve header overhead (sum 8959). I0514 16:08:37.640925 3780 default_node_network_controller.go:935] Waiting for gateway and management port readiness... I0514 16:08:37.640950 3780 obj_retry_gateway.go:33] [newRetryFrameworkNodeWithParameters] g.watchFactory=&{0xc000c2a550 0xc0002dd3b0 0xc0002dd490 0xc0002dd500 0xc0002dd570 0xc0002dd650 0xc0002dd6c0 0xc000c7c230 0xc0002dd730 0xc0002dd7a0 0xc000c7c3c0 0xc00657f500 0xc0002dd810 map[0x2ff9080:0xc0044240c0 0x2ff95c0:0xc004424240 0x2ffa2e0:0xc0030872c0 0x2ffa580:0xc004424060 0x2ffad60:0xc004424120 0x2ffb000:0xc004424180 0x2ffb2a0:0xc0044241e0 0x2ffba80:0xc004424360 0x2ffbd20:0xc0044243c0 0x3026980:0xc0006a2c60 0x3026d20:0xc0006a2cc0 0x3027800:0xc000c48fc0 0x3029160:0xc001b74000 0x30298a0:0xc00258c240 0x302a720:0xc001b74060] 0xc00016a770 0} I0514 16:08:37.641031 3780 gateway.go:151] Starting gateway service sync I0514 16:08:37.641527 3780 gateway_iptables.go:474] Recreating iptables rules for table: nat, chain: OVN-KUBE-ITP I0514 16:08:37.643718 3780 gateway_iptables.go:474] Recreating iptables rules for table: nat, chain: OVN-KUBE-NODEPORT I0514 16:08:37.662859 3780 gateway_iptables.go:474] Recreating iptables rules for table: nat, chain: OVN-KUBE-EXTERNALIP I0514 16:08:37.664034 3780 gateway_iptables.go:474] Recreating iptables rules for table: nat, chain: OVN-KUBE-ETP I0514 16:08:37.665295 3780 gateway_iptables.go:474] Recreating iptables rules for table: mangle, chain: OVN-KUBE-ITP I0514 16:08:37.699935 3780 gateway.go:168] Gateway service sync done. Time taken: 58.883071ms I0514 16:08:37.700269 3780 healthcheck.go:145] Opening healthcheck "openshift-ingress/router-default" on port 30597 I0514 16:08:37.700703 3780 obj_retry_gateway.go:33] [newRetryFrameworkNodeWithParameters] g.watchFactory=&{0xc000c2a550 0xc0002dd3b0 0xc0002dd490 0xc0002dd500 0xc0002dd570 0xc0002dd650 0xc0002dd6c0 0xc000c7c230 0xc0002dd730 0xc0002dd7a0 0xc000c7c3c0 0xc00657f500 0xc0002dd810 map[0x2ff9080:0xc0044240c0 0x2ff95c0:0xc004424240 0x2ffa2e0:0xc0030872c0 0x2ffa580:0xc004424060 0x2ffad60:0xc004424120 0x2ffb000:0xc004424180 0x2ffb2a0:0xc0044241e0 0x2ffba80:0xc004424360 0x2ffbd20:0xc0044243c0 0x3026980:0xc0006a2c60 0x3026d20:0xc0006a2cc0 0x3027800:0xc000c48fc0 0x3029160:0xc001b74000 0x30298a0:0xc00258c240 0x302a720:0xc001b74060] 0xc00016a770 0} I0514 16:08:37.701107 3780 obj_retry_gateway.go:33] [newRetryFrameworkNodeWithParameters] g.watchFactory=&{0xc000c2a550 0xc0002dd3b0 0xc0002dd490 0xc0002dd500 0xc0002dd570 0xc0002dd650 0xc0002dd6c0 0xc000c7c230 0xc0002dd730 0xc0002dd7a0 0xc000c7c3c0 0xc00657f500 0xc0002dd810 map[0x2ff9080:0xc0044240c0 0x2ff95c0:0xc004424240 0x2ffa2e0:0xc0030872c0 0x2ffa580:0xc004424060 0x2ffad60:0xc004424120 0x2ffb000:0xc004424180 0x2ffb2a0:0xc0044241e0 0x2ffba80:0xc004424360 0x2ffbd20:0xc0044243c0 0x3026980:0xc0006a2c60 0x3026d20:0xc0006a2cc0 0x3027800:0xc000c48fc0 0x3029160:0xc001b74000 0x30298a0:0xc00258c240 0x302a720:0xc001b74060] 0xc00016a770 0} I0514 16:08:37.701141 3780 gateway.go:532] Reconciling gateway with updates I0514 16:08:37.706450 3780 gateway.go:339] Spawning Conntrack Rule Check Thread I0514 16:08:37.706487 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress-operator/metrics I0514 16:08:37.706501 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics I0514 16:08:37.706507 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics I0514 16:08:37.706527 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics I0514 16:08:37.706530 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics I0514 16:08:37.706531 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-dns/dns-default I0514 16:08:37.706533 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/packageserver I0514 16:08:37.706530 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-console-operator/metrics I0514 16:08:37.706542 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-dns/dns-default I0514 16:08:37.706547 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/packageserver I0514 16:08:37.706549 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-service-ca-operator/metrics I0514 16:08:37.706553 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator I0514 16:08:37.706553 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-console-operator/metrics I0514 16:08:37.706563 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-server I0514 16:08:37.706538 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress/router-internal-default I0514 16:08:37.706577 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-server I0514 16:08:37.706579 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller I0514 16:08:37.706576 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-diagnostics/network-check-target I0514 16:08:37.706514 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress-operator/metrics I0514 16:08:37.706588 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics I0514 16:08:37.706590 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller I0514 16:08:37.706592 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-diagnostics/network-check-target I0514 16:08:37.706599 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog I0514 16:08:37.706602 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-image-registry/image-registry I0514 16:08:37.706612 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog I0514 16:08:37.706622 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-apiserver-operator/metrics I0514 16:08:37.706621 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio I0514 16:08:37.706616 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver I0514 16:08:37.706631 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-apiserver-operator/metrics I0514 16:08:37.706566 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-service-ca-operator/metrics I0514 16:08:37.706633 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio I0514 16:08:37.706639 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller after 0 failed attempt(s) I0514 16:08:37.706645 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver I0514 16:08:37.706650 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-server after 0 failed attempt(s) I0514 16:08:37.706650 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-diagnostics/network-check-target after 0 failed attempt(s) I0514 16:08:37.706650 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/packageserver after 0 failed attempt(s) I0514 16:08:37.706657 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver after 0 failed attempt(s) I0514 16:08:37.706493 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift-oauth-apiserver I0514 16:08:37.706635 3780 default_node_network_controller.go:944] Gateway and management port readiness took 65.701388ms I0514 16:08:37.706669 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics I0514 16:08:37.706673 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift-oauth-apiserver I0514 16:08:37.706681 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics I0514 16:08:37.706582 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress/router-internal-default I0514 16:08:37.706619 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-dns-operator/metrics I0514 16:08:37.706512 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics I0514 16:08:37.706737 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-dns-operator/metrics I0514 16:08:37.706615 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-image-registry/image-registry I0514 16:08:37.706602 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics I0514 16:08:37.706768 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift-apiserver I0514 16:08:37.706782 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift-apiserver I0514 16:08:37.706653 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog after 0 failed attempt(s) I0514 16:08:37.706798 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-version/cluster-version-operator I0514 16:08:37.706565 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator I0514 16:08:37.706821 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics I0514 16:08:37.706584 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-config-operator/metrics I0514 16:08:37.706833 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator after 0 failed attempt(s) I0514 16:08:37.706839 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-config-operator/metrics I0514 16:08:37.706840 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-insights/metrics I0514 16:08:37.706856 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-insights/metrics I0514 16:08:37.706858 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/kubernetes I0514 16:08:37.706866 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress-canary/ingress-canary I0514 16:08:37.706867 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics I0514 16:08:37.706865 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics I0514 16:08:37.706870 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service I0514 16:08:37.706878 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress-canary/ingress-canary I0514 16:08:37.706878 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics I0514 16:08:37.706882 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics I0514 16:08:37.706888 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service I0514 16:08:37.706886 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-multus/network-metrics-service I0514 16:08:37.706894 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon I0514 16:08:37.706902 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-multus/network-metrics-service I0514 16:08:37.706905 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon I0514 16:08:37.706911 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-multus/network-metrics-service after 0 failed attempt(s) I0514 16:08:37.706857 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-diagnostics/network-check-source I0514 16:08:37.706933 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-diagnostics/network-check-source I0514 16:08:37.706944 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-diagnostics/network-check-source after 0 failed attempt(s) I0514 16:08:37.706826 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node I0514 16:08:37.706960 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node I0514 16:08:37.706976 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node after 0 failed attempt(s) I0514 16:08:37.706829 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-version/cluster-version-operator I0514 16:08:37.707014 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-operator/metrics I0514 16:08:37.706741 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics I0514 16:08:37.707077 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-operator/metrics I0514 16:08:37.707090 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-operator/metrics after 0 failed attempt(s) I0514 16:08:37.706873 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/kubernetes I0514 16:08:37.706635 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress-operator/metrics after 0 failed attempt(s) I0514 16:08:37.706882 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics I0514 16:08:37.706840 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-console/networking-console-plugin I0514 16:08:37.707172 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-console/networking-console-plugin I0514 16:08:37.706752 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-apiserver/check-endpoints I0514 16:08:37.707215 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-apiserver/check-endpoints I0514 16:08:37.707028 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-authentication-operator/metrics I0514 16:08:37.707236 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-authentication-operator/metrics I0514 16:08:37.707038 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-samples-operator/metrics I0514 16:08:37.707337 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-samples-operator/metrics I0514 16:08:37.707345 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-samples-operator/metrics after 0 failed attempt(s) I0514 16:08:37.706827 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress/router-default I0514 16:08:37.707365 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress/router-default I0514 16:08:37.707049 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift I0514 16:08:37.707384 3780 port_claim.go:71] Svc openshift-ingress/router-default: port 30846 is already open, no action needed I0514 16:08:37.707386 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift I0514 16:08:37.707391 3780 port_claim.go:71] Svc openshift-ingress/router-default: port 30515 is already open, no action needed I0514 16:08:37.707398 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift after 0 failed attempt(s) I0514 16:08:37.707044 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator I0514 16:08:37.707424 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator I0514 16:08:37.707428 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator after 0 failed attempt(s) I0514 16:08:37.708122 3780 node_ip_handler_linux.go:160] Node IP manager is running W0514 16:08:37.745748 3780 port_linux.go:245] Missing or unable to find route entry for subnet 10.128.0.0/14 via gateway 10.129.0.1 on link ovn-k8s-mp0 with MTU: 0 W0514 16:08:37.746005 3780 port_linux.go:245] Missing or unable to find route entry for subnet 169.254.0.3/32 via gateway 10.129.0.1 on link ovn-k8s-mp0 with MTU: 0 W0514 16:08:37.746180 3780 port_linux.go:261] Missing arp entry for MAC/IP binding (0a:58:0a:81:00:01/10.129.0.1) on link ovn-k8s-mp0 I0514 16:08:37.756241 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-dns/dns-default after 0 failed attempt(s) I0514 16:08:37.756281 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-service-ca-operator/metrics after 0 failed attempt(s) I0514 16:08:37.756302 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics after 0 failed attempt(s) I0514 16:08:37.756313 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics after 0 failed attempt(s) I0514 16:08:37.756324 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-apiserver-operator/metrics after 0 failed attempt(s) I0514 16:08:37.756333 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-console-operator/metrics after 0 failed attempt(s) I0514 16:08:37.756339 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio after 0 failed attempt(s) I0514 16:08:37.756345 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics after 0 failed attempt(s) I0514 16:08:37.756352 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift-oauth-apiserver after 0 failed attempt(s) I0514 16:08:37.756358 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress/router-internal-default after 0 failed attempt(s) I0514 16:08:37.756364 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-dns-operator/metrics after 0 failed attempt(s) I0514 16:08:37.756369 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics after 0 failed attempt(s) I0514 16:08:37.756375 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift-apiserver after 0 failed attempt(s) I0514 16:08:37.756381 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics after 0 failed attempt(s) I0514 16:08:37.756391 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics after 0 failed attempt(s) I0514 16:08:37.756397 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-image-registry/image-registry after 0 failed attempt(s) I0514 16:08:37.756403 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service after 0 failed attempt(s) I0514 16:08:37.756426 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress-canary/ingress-canary after 0 failed attempt(s) I0514 16:08:37.756452 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-version/cluster-version-operator after 0 failed attempt(s) I0514 16:08:37.756464 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-insights/metrics after 0 failed attempt(s) I0514 16:08:37.756472 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics after 0 failed attempt(s) I0514 16:08:37.797518 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/kubernetes after 0 failed attempt(s) I0514 16:08:37.797559 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon after 0 failed attempt(s) I0514 16:08:37.797568 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-config-operator/metrics after 0 failed attempt(s) I0514 16:08:37.797574 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics after 0 failed attempt(s) I0514 16:08:37.797579 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-console/networking-console-plugin after 0 failed attempt(s) I0514 16:08:37.797586 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-apiserver/check-endpoints after 0 failed attempt(s) I0514 16:08:37.797591 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-authentication-operator/metrics after 0 failed attempt(s) I0514 16:08:37.837925 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress/router-default after 0 failed attempt(s) I0514 16:08:39.126720 3780 egressfirewall.go:723] Syncing node "ip-10-0-2-139.ec2.internal" for egress firewall I0514 16:08:39.126788 3780 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-2-139.ec2.internal I0514 16:08:39.126804 3780 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-2-139.ec2.internal: 20.598µs I0514 16:08:39.126804 3780 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-2-139.ec2.internal in network "default" I0514 16:08:39.126807 3780 default_network_controller.go:1084] Egress IP detected IP address change for node ip-10-0-2-139.ec2.internal. Updating no re-route policies I0514 16:08:39.126855 3780 node_tracker.go:169] Node ip-10-0-2-139.ec2.internal switch + router changed, syncing services in network "default" I0514 16:08:39.126872 3780 services_controller.go:574] Full service sync requested for network=default I0514 16:08:39.407554 3780 default_node_network_controller.go:1380] Starting node controller node sync W0514 16:08:39.407582 3780 egressip_healthcheck.go:76] Health checking using insecure connection I0514 16:08:39.407635 3780 egressip_healthcheck.go:109] Starting Egress IP Health Server on 10.129.0.2:9107 I0514 16:08:39.418147 3780 default_node_network_controller.go:1422] Node controller node sync done. Time taken: 10.574575ms I0514 16:08:39.418252 3780 default_node_network_controller.go:1293] Adding remote node "ip-10-0-1-122.ec2.internal", IP: 10.0.1.122 to PMTUD blocking rules I0514 16:08:39.418265 3780 default_node_network_controller.go:1293] Adding remote node "ip-10-0-1-122.ec2.internal", IP: 172.20.0.1 to PMTUD blocking rules I0514 16:08:39.418278 3780 default_node_network_controller.go:1293] Adding remote node "ip-10-0-2-139.ec2.internal", IP: 10.0.2.139 to PMTUD blocking rules I0514 16:08:39.418294 3780 default_node_network_controller.go:1293] Adding remote node "ip-10-0-2-139.ec2.internal", IP: 172.20.0.1 to PMTUD blocking rules I0514 16:08:39.428844 3780 healthcheck_node.go:124] "Starting node proxy healthz server" address="0.0.0.0:10256" I0514 16:08:39.429123 3780 egressservice_node.go:93] Setting up event handlers for Egress Services I0514 16:08:39.429250 3780 egressservice_node.go:183] Starting Egress Services Controller I0514 16:08:39.429268 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices" I0514 16:08:39.429279 3780 shared_informer.go:356] "Caches are synced" controller="egressservices" I0514 16:08:39.429285 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_services" I0514 16:08:39.429290 3780 shared_informer.go:356] "Caches are synced" controller="egressservices_services" I0514 16:08:39.429312 3780 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_endpointslices" I0514 16:08:39.429322 3780 shared_informer.go:356] "Caches are synced" controller="egressservices_endpointslices" I0514 16:08:39.429326 3780 egressservice_node.go:197] Repairing Egress Services I0514 16:08:39.438997 3780 node_controller.go:43] Starting Admin Policy Based Route Node Controller I0514 16:08:39.439007 3780 external_controller.go:278] Starting Admin Policy Based Route Controller I0514 16:08:39.439243 3780 default_node_network_controller.go:1077] Egress IP for secondary host network is disabled I0514 16:08:39.439263 3780 link_network_manager.go:119] Link manager is running I0514 16:08:39.439274 3780 default_node_network_controller.go:1098] Default node network controller initialized and ready. I0514 16:08:39.439526 3780 ovspinning_linux.go:50] OVS CPU affinity pinning disabled I0514 16:08:39.439803 3780 vrf_manager.go:128] VRF manager is running I0514 16:08:39.439891 3780 node_controller_manager.go:437] Removing flows to drop GARP I0514 16:08:39.439903 3780 gateway.go:532] Reconciling gateway with updates I0514 16:08:39.462311 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver I0514 16:08:39.462332 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver I0514 16:08:39.462340 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver after 0 failed attempt(s) I0514 16:08:39.462349 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-diagnostics/network-check-source I0514 16:08:39.462353 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-diagnostics/network-check-source I0514 16:08:39.462357 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-diagnostics/network-check-source after 0 failed attempt(s) I0514 16:08:39.462365 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/packageserver I0514 16:08:39.462362 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-insights/metrics I0514 16:08:39.462376 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio I0514 16:08:39.462377 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics I0514 16:08:39.462369 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/packageserver I0514 16:08:39.462384 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-insights/metrics I0514 16:08:39.462388 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress-canary/ingress-canary I0514 16:08:39.462398 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics I0514 16:08:39.462398 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics I0514 16:08:39.462390 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics I0514 16:08:39.462423 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift-oauth-apiserver I0514 16:08:39.462427 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics I0514 16:08:39.462430 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift-oauth-apiserver I0514 16:08:39.462431 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node I0514 16:08:39.462434 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator I0514 16:08:39.462447 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node I0514 16:08:39.462446 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics I0514 16:08:39.462376 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-samples-operator/metrics I0514 16:08:39.462431 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-console/networking-console-plugin I0514 16:08:39.462461 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/packageserver after 0 failed attempt(s) I0514 16:08:39.462462 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-dns-operator/metrics I0514 16:08:39.462465 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog I0514 16:08:39.462471 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics after 0 failed attempt(s) I0514 16:08:39.462470 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics I0514 16:08:39.462476 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog I0514 16:08:39.462476 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-console-operator/metrics I0514 16:08:39.462478 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator I0514 16:08:39.462483 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress/router-internal-default I0514 16:08:39.462488 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon I0514 16:08:39.462488 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift-oauth-apiserver after 0 failed attempt(s) I0514 16:08:39.462491 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress/router-default I0514 16:08:39.462496 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon I0514 16:08:39.462496 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-config-operator/metrics I0514 16:08:39.462500 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress/router-internal-default I0514 16:08:39.462502 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress/router-default I0514 16:08:39.462504 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-apiserver-operator/metrics I0514 16:08:39.462506 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics I0514 16:08:39.462506 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-version/cluster-version-operator I0514 16:08:39.462513 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-apiserver-operator/metrics I0514 16:08:39.462515 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog after 0 failed attempt(s) I0514 16:08:39.462519 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-version/cluster-version-operator I0514 16:08:39.462501 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller I0514 16:08:39.462525 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-dns/dns-default I0514 16:08:39.462519 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics I0514 16:08:39.462532 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-dns/dns-default I0514 16:08:39.462541 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics I0514 16:08:39.462543 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon after 0 failed attempt(s) I0514 16:08:39.462546 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-apiserver-operator/metrics after 0 failed attempt(s) I0514 16:08:39.462546 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-version/cluster-version-operator after 0 failed attempt(s) I0514 16:08:39.462364 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-image-registry/image-registry I0514 16:08:39.462558 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-image-registry/image-registry I0514 16:08:39.462566 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress/router-internal-default after 0 failed attempt(s) I0514 16:08:39.462530 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller I0514 16:08:39.462584 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-image-registry/image-registry after 0 failed attempt(s) I0514 16:08:39.462595 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics after 0 failed attempt(s) I0514 16:08:39.462517 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics I0514 16:08:39.462401 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress-canary/ingress-canary I0514 16:08:39.462484 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift-apiserver I0514 16:08:39.462655 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift-apiserver I0514 16:08:39.462467 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-service-ca-operator/metrics I0514 16:08:39.462674 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-service-ca-operator/metrics I0514 16:08:39.462480 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics I0514 16:08:39.462462 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics I0514 16:08:39.462387 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-apiserver/check-endpoints I0514 16:08:39.462451 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-diagnostics/network-check-target I0514 16:08:39.462729 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-apiserver/check-endpoints I0514 16:08:39.462734 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-diagnostics/network-check-target I0514 16:08:39.462456 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node after 0 failed attempt(s) I0514 16:08:39.462417 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics I0514 16:08:39.462463 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/kubernetes I0514 16:08:39.462788 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/kubernetes I0514 16:08:39.462467 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-console/networking-console-plugin I0514 16:08:39.462451 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator I0514 16:08:39.462831 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator after 0 failed attempt(s) I0514 16:08:39.462472 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-dns-operator/metrics I0514 16:08:39.462479 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics after 0 failed attempt(s) I0514 16:08:39.462475 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-operator/metrics I0514 16:08:39.462858 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-operator/metrics I0514 16:08:39.462863 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-operator/metrics after 0 failed attempt(s) I0514 16:08:39.462477 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress-operator/metrics I0514 16:08:39.462869 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress-operator/metrics I0514 16:08:39.462484 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift I0514 16:08:39.462879 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift I0514 16:08:39.462885 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift after 0 failed attempt(s) I0514 16:08:39.462490 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-console-operator/metrics I0514 16:08:39.462491 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator I0514 16:08:39.462918 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator after 0 failed attempt(s) I0514 16:08:39.462491 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-server I0514 16:08:39.462926 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-server I0514 16:08:39.462382 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio I0514 16:08:39.462494 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-multus/network-metrics-service I0514 16:08:39.462980 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-multus/network-metrics-service I0514 16:08:39.462985 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-multus/network-metrics-service after 0 failed attempt(s) I0514 16:08:39.462493 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-authentication-operator/metrics I0514 16:08:39.462991 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-authentication-operator/metrics I0514 16:08:39.462501 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service I0514 16:08:39.463003 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service I0514 16:08:39.462507 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-config-operator/metrics I0514 16:08:39.462506 3780 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics I0514 16:08:39.463068 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics I0514 16:08:39.462507 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-insights/metrics after 0 failed attempt(s) I0514 16:08:39.462518 3780 port_claim.go:71] Svc openshift-ingress/router-default: port 30846 is already open, no action needed I0514 16:08:39.463100 3780 port_claim.go:71] Svc openshift-ingress/router-default: port 30515 is already open, no action needed I0514 16:08:39.462465 3780 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-samples-operator/metrics I0514 16:08:39.463133 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-samples-operator/metrics after 0 failed attempt(s) I0514 16:08:39.486618 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-dns/dns-default after 0 failed attempt(s) I0514 16:08:39.486635 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller after 0 failed attempt(s) I0514 16:08:39.486641 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics after 0 failed attempt(s) I0514 16:08:39.486646 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress-canary/ingress-canary after 0 failed attempt(s) I0514 16:08:39.486651 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift-apiserver after 0 failed attempt(s) I0514 16:08:39.486656 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-service-ca-operator/metrics after 0 failed attempt(s) I0514 16:08:39.486661 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics after 0 failed attempt(s) I0514 16:08:39.486666 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics after 0 failed attempt(s) I0514 16:08:39.486670 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-apiserver/check-endpoints after 0 failed attempt(s) I0514 16:08:39.486675 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-diagnostics/network-check-target after 0 failed attempt(s) I0514 16:08:39.486679 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics after 0 failed attempt(s) I0514 16:08:39.525009 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/kubernetes after 0 failed attempt(s) I0514 16:08:39.525028 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-console/networking-console-plugin after 0 failed attempt(s) I0514 16:08:39.525034 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-dns-operator/metrics after 0 failed attempt(s) I0514 16:08:39.525039 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress-operator/metrics after 0 failed attempt(s) I0514 16:08:39.525045 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-console-operator/metrics after 0 failed attempt(s) I0514 16:08:39.525050 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-server after 0 failed attempt(s) I0514 16:08:39.525056 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio after 0 failed attempt(s) I0514 16:08:39.525060 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-authentication-operator/metrics after 0 failed attempt(s) I0514 16:08:39.525066 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service after 0 failed attempt(s) I0514 16:08:39.525071 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-config-operator/metrics after 0 failed attempt(s) I0514 16:08:39.525077 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics after 0 failed attempt(s) I0514 16:08:39.562169 3780 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress/router-default after 0 failed attempt(s) W0514 16:08:41.454809 3780 util.go:870] Failed to build global endpoints for port TCP/grpc: empty IP address endpoints for service openshift-deployment-validation-operator/deployment-validation-operator-catalog I0514 16:08:41.462762 3780 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-0-158.ec2.internal I0514 16:08:41.462789 3780 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-0-158.ec2.internal: 31.07µs W0514 16:08:41.490608 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console-operator/metrics I0514 16:08:41.549716 3780 base_network_controller_pods.go:487] [default/openshift-image-registry/image-registry-84c97879cc-rjbmv] creating logical port openshift-image-registry_image-registry-84c97879cc-rjbmv for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:41.549837 3780 kube.go:256] Updating pod openshift-image-registry/image-registry-84c97879cc-rjbmv I0514 16:08:41.565233 3780 base_network_controller_pods.go:487] [default/openshift-ingress/router-default-6d969849c5-4l44k] creating logical port openshift-ingress_router-default-6d969849c5-4l44k for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:41.565307 3780 kube.go:256] Updating pod openshift-ingress/router-default-6d969849c5-4l44k I0514 16:08:41.636167 3780 pod.go:62] [openshift-image-registry/image-registry-84c97879cc-rjbmv] pod update took 86.354924ms I0514 16:08:41.636184 3780 base_network_controller_pods.go:951] [default/openshift-image-registry/image-registry-84c97879cc-rjbmv] addLogicalPort annotation time took 86.375852ms I0514 16:08:41.637285 3780 pods.go:271] [openshift-image-registry/image-registry-84c97879cc-rjbmv] addLogicalPort took 87.576922ms, libovsdb time 877.297µs I0514 16:08:41.648584 3780 pod.go:62] [openshift-ingress/router-default-6d969849c5-4l44k] pod update took 83.300785ms I0514 16:08:41.648603 3780 base_network_controller_pods.go:951] [default/openshift-ingress/router-default-6d969849c5-4l44k] addLogicalPort annotation time took 83.321226ms I0514 16:08:41.649500 3780 pods.go:271] [openshift-ingress/router-default-6d969849c5-4l44k] addLogicalPort took 84.276372ms, libovsdb time 663.501µs I0514 16:08:41.696892 3780 base_network_controller_pods.go:487] [default/openshift-ingress-canary/ingress-canary-lqpjv] creating logical port openshift-ingress-canary_ingress-canary-lqpjv for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:41.696981 3780 kube.go:256] Updating pod openshift-ingress-canary/ingress-canary-lqpjv I0514 16:08:41.699346 3780 base_network_controller_pods.go:487] [default/openshift-dns/dns-default-d9llh] creating logical port openshift-dns_dns-default-d9llh for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:41.699431 3780 kube.go:256] Updating pod openshift-dns/dns-default-d9llh I0514 16:08:41.718396 3780 pod.go:62] [openshift-ingress-canary/ingress-canary-lqpjv] pod update took 21.442001ms I0514 16:08:41.718430 3780 base_network_controller_pods.go:951] [default/openshift-ingress-canary/ingress-canary-lqpjv] addLogicalPort annotation time took 21.477615ms I0514 16:08:41.718522 3780 pod.go:62] [openshift-dns/dns-default-d9llh] pod update took 19.129656ms I0514 16:08:41.718538 3780 base_network_controller_pods.go:951] [default/openshift-dns/dns-default-d9llh] addLogicalPort annotation time took 19.148699ms I0514 16:08:41.719565 3780 pods.go:271] [openshift-ingress-canary/ingress-canary-lqpjv] addLogicalPort took 22.69797ms, libovsdb time 898.687µs I0514 16:08:41.719993 3780 pods.go:271] [openshift-dns/dns-default-d9llh] addLogicalPort took 20.656879ms, libovsdb time 1.251411ms I0514 16:08:42.090918 3780 cni.go:452] [openshift-ingress-canary/ingress-canary-lqpjv 9c57e85dec52cdc316a699131ed2967ed172aeb2fab0255c4e3bab433b7e54b3 network default NAD default NAD key ] ADD starting CNI request [openshift-ingress-canary/ingress-canary-lqpjv 9c57e85dec52cdc316a699131ed2967ed172aeb2fab0255c4e3bab433b7e54b3 network default NAD default NAD key ] I0514 16:08:42.120729 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-ingress-canary, podName: ingress-canary-lqpjv, hostIfaceName: 9c57e85dec52cdc, network: default, NAD default, SandboxID: "9c57e85dec52cdc316a699131ed2967ed172aeb2fab0255c4e3bab433b7e54b3", PCI device ID: , UID: "709c00d1-92be-4ffe-b138-cb2e284a8c48", MAC: 0a:58:0a:81:00:08, IPs: [10.129.0.8/23] I0514 16:08:42.175700 3780 cni.go:473] [openshift-ingress-canary/ingress-canary-lqpjv 9c57e85dec52cdc316a699131ed2967ed172aeb2fab0255c4e3bab433b7e54b3 network default NAD default NAD key default] ADD finished CNI request [openshift-ingress-canary/ingress-canary-lqpjv 9c57e85dec52cdc316a699131ed2967ed172aeb2fab0255c4e3bab433b7e54b3 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ae:91:15:e7:3b:11\",\"name\":\"9c57e85dec52cdc\"},{\"mac\":\"0a:58:0a:81:00:08\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/38a4628d-1ed2-4a8f-8db4-be4debaaa401\"}],\"ips\":[{\"address\":\"10.129.0.8/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:08:42.226556 3780 namespace.go:98] [openshift-kube-storage-version-migrator] adding namespace I0514 16:08:42.227317 3780 namespace.go:102] [openshift-kube-storage-version-migrator] adding namespace took 743.963µs I0514 16:08:42.251430 3780 namespace.go:142] [openshift-kube-storage-version-migrator] updating namespace I0514 16:08:42.792779 3780 namespace.go:98] [openshift-service-ca] adding namespace I0514 16:08:42.793739 3780 namespace.go:102] [openshift-service-ca] adding namespace took 935.554µs I0514 16:08:42.811635 3780 namespace.go:142] [openshift-service-ca] updating namespace I0514 16:08:42.821712 3780 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-2-139.ec2.internal I0514 16:08:42.821741 3780 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-2-139.ec2.internal: 34.944µs I0514 16:08:44.462714 3780 base_network_controller_pods.go:487] [default/openshift-service-ca/service-ca-d9cb48497-jc2fv] creating logical port openshift-service-ca_service-ca-d9cb48497-jc2fv for pod on switch ip-10-0-0-158.ec2.internal I0514 16:08:44.462840 3780 kube.go:256] Updating pod openshift-service-ca/service-ca-d9cb48497-jc2fv I0514 16:08:44.478728 3780 pod.go:62] [openshift-service-ca/service-ca-d9cb48497-jc2fv] pod update took 15.916419ms I0514 16:08:44.478747 3780 base_network_controller_pods.go:951] [default/openshift-service-ca/service-ca-d9cb48497-jc2fv] addLogicalPort annotation time took 15.939549ms I0514 16:08:44.479931 3780 pods.go:271] [openshift-service-ca/service-ca-d9cb48497-jc2fv] addLogicalPort took 17.239595ms, libovsdb time 914.649µs I0514 16:08:44.872261 3780 cni.go:452] [openshift-service-ca/service-ca-d9cb48497-jc2fv d9f854169c1557ab5546b50bc1a6ff512e80647fbaf1ca9480144ee4a77b34fe network default NAD default NAD key ] ADD starting CNI request [openshift-service-ca/service-ca-d9cb48497-jc2fv d9f854169c1557ab5546b50bc1a6ff512e80647fbaf1ca9480144ee4a77b34fe network default NAD default NAD key ] I0514 16:08:44.881915 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-service-ca, podName: service-ca-d9cb48497-jc2fv, hostIfaceName: d9f854169c1557a, network: default, NAD default, SandboxID: "d9f854169c1557ab5546b50bc1a6ff512e80647fbaf1ca9480144ee4a77b34fe", PCI device ID: , UID: "c29c6df1-f553-464e-a118-824023fa8707", MAC: 0a:58:0a:81:00:0a, IPs: [10.129.0.10/23] I0514 16:08:44.925826 3780 cni.go:473] [openshift-service-ca/service-ca-d9cb48497-jc2fv d9f854169c1557ab5546b50bc1a6ff512e80647fbaf1ca9480144ee4a77b34fe network default NAD default NAD key default] ADD finished CNI request [openshift-service-ca/service-ca-d9cb48497-jc2fv d9f854169c1557ab5546b50bc1a6ff512e80647fbaf1ca9480144ee4a77b34fe network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"66:fc:91:c0:7e:5e\",\"name\":\"d9f854169c1557a\"},{\"mac\":\"0a:58:0a:81:00:0a\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/56252357-f423-4419-9ff2-ad8c15c17bce\"}],\"ips\":[{\"address\":\"10.129.0.10/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:08:52.262426 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console-operator/metrics I0514 16:08:57.537251 3780 cni.go:452] [openshift-image-registry/image-registry-84c97879cc-rjbmv 173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825 network default NAD default NAD key ] ADD starting CNI request [openshift-image-registry/image-registry-84c97879cc-rjbmv 173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825 network default NAD default NAD key ] I0514 16:08:57.547828 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-image-registry, podName: image-registry-84c97879cc-rjbmv, hostIfaceName: 173437752a140f9, network: default, NAD default, SandboxID: "173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825", PCI device ID: , UID: "2b25c638-1447-4fa2-86e1-32895593a404", MAC: 0a:58:0a:81:00:06, IPs: [10.129.0.6/23] I0514 16:08:57.591222 3780 cni.go:473] [openshift-image-registry/image-registry-84c97879cc-rjbmv 173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825 network default NAD default NAD key default] ADD finished CNI request [openshift-image-registry/image-registry-84c97879cc-rjbmv 173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"b2:c8:27:33:c6:7a\",\"name\":\"173437752a140f9\"},{\"mac\":\"0a:58:0a:81:00:06\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/4a062ede-9b5c-4724-9f3b-2705d0d9d7b7\"}],\"ips\":[{\"address\":\"10.129.0.6/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:08:57.724245 3780 cni.go:452] [openshift-dns/dns-default-d9llh f8520757e842878a9e65eb920e7239f8c4bfe88aa6605c1465065159e206b380 network default NAD default NAD key ] ADD starting CNI request [openshift-dns/dns-default-d9llh f8520757e842878a9e65eb920e7239f8c4bfe88aa6605c1465065159e206b380 network default NAD default NAD key ] I0514 16:08:57.734351 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-dns, podName: dns-default-d9llh, hostIfaceName: f8520757e842878, network: default, NAD default, SandboxID: "f8520757e842878a9e65eb920e7239f8c4bfe88aa6605c1465065159e206b380", PCI device ID: , UID: "af6296b7-ea61-4314-956b-2ebe4b154a48", MAC: 0a:58:0a:81:00:09, IPs: [10.129.0.9/23] I0514 16:08:57.777373 3780 cni.go:473] [openshift-dns/dns-default-d9llh f8520757e842878a9e65eb920e7239f8c4bfe88aa6605c1465065159e206b380 network default NAD default NAD key default] ADD finished CNI request [openshift-dns/dns-default-d9llh f8520757e842878a9e65eb920e7239f8c4bfe88aa6605c1465065159e206b380 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"5a:75:65:35:1d:80\",\"name\":\"f8520757e842878\"},{\"mac\":\"0a:58:0a:81:00:09\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/9ab6b0f6-0659-4a4f-b15d-35e55a42e703\"}],\"ips\":[{\"address\":\"10.129.0.9/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:09:04.718325 3780 util.go:870] Failed to build global endpoints for port TCP/5000-tcp: empty IP address endpoints for service openshift-image-registry/image-registry W0514 16:09:04.723446 3780 util.go:870] Failed to build global endpoints for port TCP/: empty IP address endpoints for service openshift-network-diagnostics/network-check-target W0514 16:09:04.727319 3780 util.go:870] Failed to build global endpoints for port UDP/dns: empty IP address endpoints for service openshift-dns/dns-default W0514 16:09:04.727331 3780 util.go:870] Failed to build global endpoints for port TCP/dns-tcp: empty IP address endpoints for service openshift-dns/dns-default W0514 16:09:04.727336 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-dns/dns-default I0514 16:09:04.794325 3780 base_network_controller_pods.go:487] [default/openshift-console/downloads-5c54f4944f-xmqqq] creating logical port openshift-console_downloads-5c54f4944f-xmqqq for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:04.794445 3780 kube.go:256] Updating pod openshift-console/downloads-5c54f4944f-xmqqq W0514 16:09:04.824539 3780 util.go:870] Failed to build global endpoints for port TCP/5000-tcp: empty IP address endpoints for service openshift-image-registry/image-registry I0514 16:09:04.828012 3780 pod.go:62] [openshift-console/downloads-5c54f4944f-xmqqq] pod update took 33.614825ms I0514 16:09:04.828030 3780 base_network_controller_pods.go:951] [default/openshift-console/downloads-5c54f4944f-xmqqq] addLogicalPort annotation time took 33.637941ms I0514 16:09:04.829100 3780 pods.go:271] [openshift-console/downloads-5c54f4944f-xmqqq] addLogicalPort took 34.796671ms, libovsdb time 790.725µs I0514 16:09:04.858959 3780 base_network_controller_pods.go:487] [default/openshift-insights/insights-runtime-extractor-54mws] creating logical port openshift-insights_insights-runtime-extractor-54mws for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:04.859041 3780 kube.go:256] Updating pod openshift-insights/insights-runtime-extractor-54mws I0514 16:09:04.910293 3780 pod.go:62] [openshift-insights/insights-runtime-extractor-54mws] pod update took 51.278094ms I0514 16:09:04.910314 3780 base_network_controller_pods.go:951] [default/openshift-insights/insights-runtime-extractor-54mws] addLogicalPort annotation time took 51.306619ms I0514 16:09:04.911504 3780 pods.go:271] [openshift-insights/insights-runtime-extractor-54mws] addLogicalPort took 52.552501ms, libovsdb time 688.272µs I0514 16:09:05.153217 3780 cni.go:452] [openshift-console/downloads-5c54f4944f-xmqqq 92150e7df39ba24e7e45377bc121984cea4bf38c0f6130f9fe6f54b5536acc04 network default NAD default NAD key ] ADD starting CNI request [openshift-console/downloads-5c54f4944f-xmqqq 92150e7df39ba24e7e45377bc121984cea4bf38c0f6130f9fe6f54b5536acc04 network default NAD default NAD key ] I0514 16:09:05.163174 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-console, podName: downloads-5c54f4944f-xmqqq, hostIfaceName: 92150e7df39ba24, network: default, NAD default, SandboxID: "92150e7df39ba24e7e45377bc121984cea4bf38c0f6130f9fe6f54b5536acc04", PCI device ID: , UID: "6f69e6fd-cca1-4a5e-9c45-76d9f2a186dc", MAC: 0a:58:0a:81:00:0b, IPs: [10.129.0.11/23] I0514 16:09:05.209054 3780 cni.go:473] [openshift-console/downloads-5c54f4944f-xmqqq 92150e7df39ba24e7e45377bc121984cea4bf38c0f6130f9fe6f54b5536acc04 network default NAD default NAD key default] ADD finished CNI request [openshift-console/downloads-5c54f4944f-xmqqq 92150e7df39ba24e7e45377bc121984cea4bf38c0f6130f9fe6f54b5536acc04 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"2a:9b:93:28:75:1d\",\"name\":\"92150e7df39ba24\"},{\"mac\":\"0a:58:0a:81:00:0b\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/fffbb2ec-a09b-49e3-ae1b-9a58f94dc96e\"}],\"ips\":[{\"address\":\"10.129.0.11/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:05.216482 3780 cni.go:452] [openshift-insights/insights-runtime-extractor-54mws 48d0fa130b1620c638cdaa6bec4582575c7cc70adc16e594ea418a90a4b883f2 network default NAD default NAD key ] ADD starting CNI request [openshift-insights/insights-runtime-extractor-54mws 48d0fa130b1620c638cdaa6bec4582575c7cc70adc16e594ea418a90a4b883f2 network default NAD default NAD key ] I0514 16:09:05.225944 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-insights, podName: insights-runtime-extractor-54mws, hostIfaceName: 48d0fa130b1620c, network: default, NAD default, SandboxID: "48d0fa130b1620c638cdaa6bec4582575c7cc70adc16e594ea418a90a4b883f2", PCI device ID: , UID: "ef17b656-a9b5-4204-9715-0292b22891af", MAC: 0a:58:0a:81:00:0c, IPs: [10.129.0.12/23] I0514 16:09:05.269647 3780 cni.go:473] [openshift-insights/insights-runtime-extractor-54mws 48d0fa130b1620c638cdaa6bec4582575c7cc70adc16e594ea418a90a4b883f2 network default NAD default NAD key default] ADD finished CNI request [openshift-insights/insights-runtime-extractor-54mws 48d0fa130b1620c638cdaa6bec4582575c7cc70adc16e594ea418a90a4b883f2 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"a2:8e:f7:27:d8:f1\",\"name\":\"48d0fa130b1620c\"},{\"mac\":\"0a:58:0a:81:00:0c\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/63db4af9-6523-428a-a189-bd2ec88fead3\"}],\"ips\":[{\"address\":\"10.129.0.12/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:06.455146 3780 obj_retry.go:498] Detected object openshift-image-registry/image-registry-84c97879cc-vz7dt of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:06.455833 3780 obj_retry.go:498] Detected object openshift-image-registry/image-registry-84c97879cc-vz7dt of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:06.472766 3780 obj_retry.go:492] Detected object openshift-image-registry/image-registry-84c97879cc-vz7dt of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:06.472787 3780 obj_retry.go:492] Detected object openshift-image-registry/image-registry-84c97879cc-vz7dt of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:06.477653 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-image-registry/image-registry-84c97879cc-vz7dt I0514 16:09:06.477674 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-image-registry/image-registry-84c97879cc-vz7dt W0514 16:09:07.447524 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service openshift-ingress/router-internal-default W0514 16:09:07.447541 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-ingress/router-internal-default W0514 16:09:07.447547 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-ingress/router-internal-default W0514 16:09:07.450544 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service openshift-ingress/router-default W0514 16:09:07.450554 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-ingress/router-default W0514 16:09:07.477587 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service openshift-ingress/router-default W0514 16:09:07.477600 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-ingress/router-default W0514 16:09:07.477697 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service openshift-ingress/router-internal-default W0514 16:09:07.477712 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-ingress/router-internal-default W0514 16:09:07.477718 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-ingress/router-internal-default W0514 16:09:07.495861 3780 util.go:870] Failed to build global endpoints for port TCP/5000-tcp: empty IP address endpoints for service openshift-image-registry/image-registry I0514 16:09:08.647593 3780 base_network_controller_pods.go:487] [default/openshift-console/console-bf589dbdc-kf6q4] creating logical port openshift-console_console-bf589dbdc-kf6q4 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:08.647714 3780 kube.go:256] Updating pod openshift-console/console-bf589dbdc-kf6q4 I0514 16:09:08.666271 3780 pod.go:62] [openshift-console/console-bf589dbdc-kf6q4] pod update took 18.606203ms I0514 16:09:08.666289 3780 base_network_controller_pods.go:951] [default/openshift-console/console-bf589dbdc-kf6q4] addLogicalPort annotation time took 18.630458ms I0514 16:09:08.667538 3780 pods.go:271] [openshift-console/console-bf589dbdc-kf6q4] addLogicalPort took 19.995862ms, libovsdb time 954.407µs I0514 16:09:09.052888 3780 cni.go:452] [openshift-console/console-bf589dbdc-kf6q4 dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62 network default NAD default NAD key ] ADD starting CNI request [openshift-console/console-bf589dbdc-kf6q4 dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62 network default NAD default NAD key ] I0514 16:09:09.063689 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-console, podName: console-bf589dbdc-kf6q4, hostIfaceName: dfc6fd367226bbc, network: default, NAD default, SandboxID: "dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62", PCI device ID: , UID: "f815d46e-ddcb-4a21-81c8-accf84b45643", MAC: 0a:58:0a:81:00:0d, IPs: [10.129.0.13/23] I0514 16:09:09.112105 3780 cni.go:473] [openshift-console/console-bf589dbdc-kf6q4 dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62 network default NAD default NAD key default] ADD finished CNI request [openshift-console/console-bf589dbdc-kf6q4 dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"0e:7c:60:04:e6:1e\",\"name\":\"dfc6fd367226bbc\"},{\"mac\":\"0a:58:0a:81:00:0d\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/d6b90be9-ac2c-40bb-b53c-d7209c903d77\"}],\"ips\":[{\"address\":\"10.129.0.13/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:10.323647 3780 base_network_controller_pods.go:487] [default/openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26] creating logical port openshift-monitoring_prometheus-operator-admission-webhook-59b79769fb-vpf26 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:10.323726 3780 kube.go:256] Updating pod openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26 I0514 16:09:10.341218 3780 pod.go:62] [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26] pod update took 17.518733ms I0514 16:09:10.341238 3780 base_network_controller_pods.go:951] [default/openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26] addLogicalPort annotation time took 17.543171ms I0514 16:09:10.342479 3780 pods.go:271] [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26] addLogicalPort took 18.845812ms, libovsdb time 816.958µs W0514 16:09:10.538207 3780 util.go:870] Failed to build global endpoints for port UDP/dns: empty IP address endpoints for service openshift-dns/dns-default W0514 16:09:10.538225 3780 util.go:870] Failed to build global endpoints for port TCP/dns-tcp: empty IP address endpoints for service openshift-dns/dns-default W0514 16:09:10.538231 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-dns/dns-default W0514 16:09:11.264450 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 5353: empty IP address endpoints for service openshift-dns/dns-default W0514 16:09:11.264475 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9154: empty IP address endpoints for service openshift-dns/dns-default W0514 16:09:11.264480 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 5353: empty IP address endpoints for service openshift-dns/dns-default I0514 16:09:11.287853 3780 cni.go:452] [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26 ec9e2a317972a5c6cf88844bbb805bbba088b244294b4b01591ad0cfcdedabae network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26 ec9e2a317972a5c6cf88844bbb805bbba088b244294b4b01591ad0cfcdedabae network default NAD default NAD key ] I0514 16:09:11.297275 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: prometheus-operator-admission-webhook-59b79769fb-vpf26, hostIfaceName: ec9e2a317972a5c, network: default, NAD default, SandboxID: "ec9e2a317972a5c6cf88844bbb805bbba088b244294b4b01591ad0cfcdedabae", PCI device ID: , UID: "297e059f-9433-48c4-88ce-60a1a93e3ae1", MAC: 0a:58:0a:81:00:0e, IPs: [10.129.0.14/23] I0514 16:09:11.344158 3780 cni.go:473] [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26 ec9e2a317972a5c6cf88844bbb805bbba088b244294b4b01591ad0cfcdedabae network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-vpf26 ec9e2a317972a5c6cf88844bbb805bbba088b244294b4b01591ad0cfcdedabae network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"e2:f0:12:48:09:a2\",\"name\":\"ec9e2a317972a5c\"},{\"mac\":\"0a:58:0a:81:00:0e\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/5eb70609-3f2f-46eb-8672-47f3e091ec4b\"}],\"ips\":[{\"address\":\"10.129.0.14/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:09:12.801692 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console/console W0514 16:09:13.275654 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console/console I0514 16:09:13.790598 3780 cni.go:452] [openshift-ingress/router-default-6d969849c5-4l44k e336c99a36e62cc14f334157bb8f2d7bca3037ec5272274123b3a9254e9661f9 network default NAD default NAD key ] ADD starting CNI request [openshift-ingress/router-default-6d969849c5-4l44k e336c99a36e62cc14f334157bb8f2d7bca3037ec5272274123b3a9254e9661f9 network default NAD default NAD key ] I0514 16:09:13.800819 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-ingress, podName: router-default-6d969849c5-4l44k, hostIfaceName: e336c99a36e62cc, network: default, NAD default, SandboxID: "e336c99a36e62cc14f334157bb8f2d7bca3037ec5272274123b3a9254e9661f9", PCI device ID: , UID: "1b8face3-cfe1-48e2-8e65-063cbaef24d3", MAC: 0a:58:0a:81:00:07, IPs: [10.129.0.7/23] I0514 16:09:13.845925 3780 cni.go:473] [openshift-ingress/router-default-6d969849c5-4l44k e336c99a36e62cc14f334157bb8f2d7bca3037ec5272274123b3a9254e9661f9 network default NAD default NAD key default] ADD finished CNI request [openshift-ingress/router-default-6d969849c5-4l44k e336c99a36e62cc14f334157bb8f2d7bca3037ec5272274123b3a9254e9661f9 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ee:4b:f2:0c:fe:07\",\"name\":\"e336c99a36e62cc\"},{\"mac\":\"0a:58:0a:81:00:07\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/af822549-aa35-4210-ab82-8181317257fc\"}],\"ips\":[{\"address\":\"10.129.0.7/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:09:14.276441 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-monitoring/prometheus-operator-admission-webhook W0514 16:09:14.302152 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-monitoring/prometheus-operator-admission-webhook W0514 16:09:14.336446 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 443: empty IP address endpoints for service openshift-ingress/router-default W0514 16:09:14.336467 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 80: empty IP address endpoints for service openshift-ingress/router-default I0514 16:09:14.892140 3780 cni.go:452] [kube-system/global-pull-secret-syncer-6htzf 5b68e1b7dc3340168e5f477b2229e57ed3ef6e796103a0daf4bd62d1001ecb8a network default NAD default NAD key ] ADD starting CNI request [kube-system/global-pull-secret-syncer-6htzf 5b68e1b7dc3340168e5f477b2229e57ed3ef6e796103a0daf4bd62d1001ecb8a network default NAD default NAD key ] I0514 16:09:14.892453 3780 cni.go:452] [openshift-multus/network-metrics-daemon-ht2xn d698ddfd5078e47eabc966c2cbb5adfd1ab9d43f9010b41dd06229c6e5cc1cd3 network default NAD default NAD key ] ADD starting CNI request [openshift-multus/network-metrics-daemon-ht2xn d698ddfd5078e47eabc966c2cbb5adfd1ab9d43f9010b41dd06229c6e5cc1cd3 network default NAD default NAD key ] I0514 16:09:14.903666 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-multus, podName: network-metrics-daemon-ht2xn, hostIfaceName: d698ddfd5078e47, network: default, NAD default, SandboxID: "d698ddfd5078e47eabc966c2cbb5adfd1ab9d43f9010b41dd06229c6e5cc1cd3", PCI device ID: , UID: "ed5c54e1-c901-4d82-9657-03e08d3fd052", MAC: 0a:58:0a:81:00:03, IPs: [10.129.0.3/23] I0514 16:09:14.903831 3780 helper_linux.go:556] ConfigureOVS: namespace: kube-system, podName: global-pull-secret-syncer-6htzf, hostIfaceName: 5b68e1b7dc33401, network: default, NAD default, SandboxID: "5b68e1b7dc3340168e5f477b2229e57ed3ef6e796103a0daf4bd62d1001ecb8a", PCI device ID: , UID: "0b6cbe28-d058-4907-b12d-320a50e2f9e3", MAC: 0a:58:0a:81:00:05, IPs: [10.129.0.5/23] I0514 16:09:14.951566 3780 cni.go:473] [kube-system/global-pull-secret-syncer-6htzf 5b68e1b7dc3340168e5f477b2229e57ed3ef6e796103a0daf4bd62d1001ecb8a network default NAD default NAD key default] ADD finished CNI request [kube-system/global-pull-secret-syncer-6htzf 5b68e1b7dc3340168e5f477b2229e57ed3ef6e796103a0daf4bd62d1001ecb8a network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"66:1f:6f:85:78:bf\",\"name\":\"5b68e1b7dc33401\"},{\"mac\":\"0a:58:0a:81:00:05\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/b9cdc1b8-449c-4205-a6d0-82296db4d899\"}],\"ips\":[{\"address\":\"10.129.0.5/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:14.969979 3780 cni.go:473] [openshift-multus/network-metrics-daemon-ht2xn d698ddfd5078e47eabc966c2cbb5adfd1ab9d43f9010b41dd06229c6e5cc1cd3 network default NAD default NAD key default] ADD finished CNI request [openshift-multus/network-metrics-daemon-ht2xn d698ddfd5078e47eabc966c2cbb5adfd1ab9d43f9010b41dd06229c6e5cc1cd3 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"8a:7d:6e:22:06:c9\",\"name\":\"d698ddfd5078e47\"},{\"mac\":\"0a:58:0a:81:00:03\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/c310cbc4-9b77-45e4-8ab3-338998dde211\"}],\"ips\":[{\"address\":\"10.129.0.3/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:15.195367 3780 cni.go:452] [openshift-network-diagnostics/network-check-target-hmxmt 9add08c665acde4d651971f40cbf4423a39df19303039e9c667e9fd1de2716f1 network default NAD default NAD key ] ADD starting CNI request [openshift-network-diagnostics/network-check-target-hmxmt 9add08c665acde4d651971f40cbf4423a39df19303039e9c667e9fd1de2716f1 network default NAD default NAD key ] I0514 16:09:15.205728 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-network-diagnostics, podName: network-check-target-hmxmt, hostIfaceName: 9add08c665acde4, network: default, NAD default, SandboxID: "9add08c665acde4d651971f40cbf4423a39df19303039e9c667e9fd1de2716f1", PCI device ID: , UID: "bd6e883a-2feb-4969-88e4-434840ba5282", MAC: 0a:58:0a:81:00:04, IPs: [10.129.0.4/23] I0514 16:09:15.254114 3780 cni.go:473] [openshift-network-diagnostics/network-check-target-hmxmt 9add08c665acde4d651971f40cbf4423a39df19303039e9c667e9fd1de2716f1 network default NAD default NAD key default] ADD finished CNI request [openshift-network-diagnostics/network-check-target-hmxmt 9add08c665acde4d651971f40cbf4423a39df19303039e9c667e9fd1de2716f1 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"fe:0a:95:61:61:90\",\"name\":\"9add08c665acde4\"},{\"mac\":\"0a:58:0a:81:00:04\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/c4523da6-1433-4145-acad-7e1e0d912bc7\"}],\"ips\":[{\"address\":\"10.129.0.4/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:09:15.293703 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 80: empty IP address endpoints for service openshift-ingress/router-default W0514 16:09:15.293722 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 443: empty IP address endpoints for service openshift-ingress/router-default I0514 16:09:15.366211 3780 iptables.go:108] Creating table: nat chain: OVN-KUBE-NODEPORT W0514 16:09:15.378186 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 1936: empty IP address endpoints for service openshift-ingress/router-internal-default W0514 16:09:15.378206 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 443: empty IP address endpoints for service openshift-ingress/router-internal-default W0514 16:09:15.378214 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 80: empty IP address endpoints for service openshift-ingress/router-internal-default I0514 16:09:16.398828 3780 base_network_controller_pods.go:487] [default/openshift-monitoring/prometheus-operator-76b589d668-x9wrb] creating logical port openshift-monitoring_prometheus-operator-76b589d668-x9wrb for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:16.398941 3780 kube.go:256] Updating pod openshift-monitoring/prometheus-operator-76b589d668-x9wrb I0514 16:09:16.415446 3780 pod.go:62] [openshift-monitoring/prometheus-operator-76b589d668-x9wrb] pod update took 16.53214ms I0514 16:09:16.415475 3780 base_network_controller_pods.go:951] [default/openshift-monitoring/prometheus-operator-76b589d668-x9wrb] addLogicalPort annotation time took 16.567845ms I0514 16:09:16.416871 3780 pods.go:271] [openshift-monitoring/prometheus-operator-76b589d668-x9wrb] addLogicalPort took 18.066637ms, libovsdb time 996.915µs W0514 16:09:18.439628 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console/console W0514 16:09:19.228471 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 5000: empty IP address endpoints for service openshift-image-registry/image-registry I0514 16:09:19.694801 3780 cni.go:452] [openshift-monitoring/prometheus-operator-76b589d668-x9wrb e6be08b0682b788434d91e6303ace2078bb68657d740ac156b143e737bc718f8 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/prometheus-operator-76b589d668-x9wrb e6be08b0682b788434d91e6303ace2078bb68657d740ac156b143e737bc718f8 network default NAD default NAD key ] I0514 16:09:19.705488 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: prometheus-operator-76b589d668-x9wrb, hostIfaceName: e6be08b0682b788, network: default, NAD default, SandboxID: "e6be08b0682b788434d91e6303ace2078bb68657d740ac156b143e737bc718f8", PCI device ID: , UID: "3560fbff-c3ac-4185-bd07-a88af5d843f3", MAC: 0a:58:0a:81:00:0f, IPs: [10.129.0.15/23] I0514 16:09:19.756597 3780 cni.go:473] [openshift-monitoring/prometheus-operator-76b589d668-x9wrb e6be08b0682b788434d91e6303ace2078bb68657d740ac156b143e737bc718f8 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/prometheus-operator-76b589d668-x9wrb e6be08b0682b788434d91e6303ace2078bb68657d740ac156b143e737bc718f8 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"b2:c9:23:46:6a:d9\",\"name\":\"e6be08b0682b788\"},{\"mac\":\"0a:58:0a:81:00:0f\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/3312b6ff-6a5b-4c88-aded-fca03ff2570d\"}],\"ips\":[{\"address\":\"10.129.0.15/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:09:19.840489 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console W0514 16:09:19.840521 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console W0514 16:09:20.331092 3780 util.go:870] Failed to build global endpoints for port TCP/: empty IP address endpoints for service openshift-network-diagnostics/network-check-target W0514 16:09:23.727317 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service openshift-network-diagnostics/network-check-target W0514 16:09:23.727340 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service openshift-network-diagnostics/network-check-target W0514 16:09:28.371287 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service openshift-console/downloads W0514 16:09:28.397678 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service openshift-console/downloads W0514 16:09:29.599657 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console W0514 16:09:29.599686 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console W0514 16:09:29.656492 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console W0514 16:09:29.656518 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console I0514 16:09:30.389422 3780 obj_retry.go:498] Detected object openshift-image-registry/image-registry-8674795574-vh8hq of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:30.390066 3780 obj_retry.go:498] Detected object openshift-image-registry/image-registry-8674795574-vh8hq of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:30.910668 3780 obj_retry.go:492] Detected object openshift-image-registry/image-registry-8674795574-vh8hq of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:30.910689 3780 obj_retry.go:492] Detected object openshift-image-registry/image-registry-8674795574-vh8hq of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:30.918243 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-image-registry/image-registry-8674795574-vh8hq I0514 16:09:30.918268 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-image-registry/image-registry-8674795574-vh8hq I0514 16:09:31.064824 3780 base_network_controller_pods.go:487] [default/openshift-monitoring/alertmanager-main-0] creating logical port openshift-monitoring_alertmanager-main-0 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:31.065017 3780 kube.go:256] Updating pod openshift-monitoring/alertmanager-main-0 I0514 16:09:31.097436 3780 pod.go:62] [openshift-monitoring/alertmanager-main-0] pod update took 32.441317ms I0514 16:09:31.097457 3780 base_network_controller_pods.go:951] [default/openshift-monitoring/alertmanager-main-0] addLogicalPort annotation time took 32.496575ms I0514 16:09:31.098805 3780 pods.go:271] [openshift-monitoring/alertmanager-main-0] addLogicalPort took 34.00128ms, libovsdb time 1.013527ms I0514 16:09:32.018671 3780 obj_retry.go:498] Detected object openshift-deployment-validation-operator/99b07c45bd4393a793940f31bf2887fd12a636d02cd8e5e017ce352a0dmmnmv of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:32.019275 3780 obj_retry.go:498] Detected object openshift-deployment-validation-operator/99b07c45bd4393a793940f31bf2887fd12a636d02cd8e5e017ce352a0dmmnmv of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:32.019313 3780 obj_retry.go:498] Detected object openshift-deployment-validation-operator/99b07c45bd4393a793940f31bf2887fd12a636d02cd8e5e017ce352a0dmmnmv of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:09:32.019329 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-deployment-validation-operator/allow-from-openshift-insights to delete 1 local pods... I0514 16:09:32.019349 3780 obj_retry.go:498] Detected object openshift-deployment-validation-operator/99b07c45bd4393a793940f31bf2887fd12a636d02cd8e5e017ce352a0dmmnmv of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:09:32.019355 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-deployment-validation-operator/deployment-validation-operator-catalog-unpack-bundles to delete 1 local pods... I0514 16:09:32.029225 3780 cni.go:452] [openshift-monitoring/alertmanager-main-0 5aa24178e014ea1cb51908ce0380e90ec76a1eeba951e69c39c1f27cdfb2dfaa network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/alertmanager-main-0 5aa24178e014ea1cb51908ce0380e90ec76a1eeba951e69c39c1f27cdfb2dfaa network default NAD default NAD key ] I0514 16:09:32.039811 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: alertmanager-main-0, hostIfaceName: 5aa24178e014ea1, network: default, NAD default, SandboxID: "5aa24178e014ea1cb51908ce0380e90ec76a1eeba951e69c39c1f27cdfb2dfaa", PCI device ID: , UID: "33d309a1-b402-4405-810e-876847d0e41d", MAC: 0a:58:0a:81:00:10, IPs: [10.129.0.16/23] I0514 16:09:32.091470 3780 cni.go:473] [openshift-monitoring/alertmanager-main-0 5aa24178e014ea1cb51908ce0380e90ec76a1eeba951e69c39c1f27cdfb2dfaa network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/alertmanager-main-0 5aa24178e014ea1cb51908ce0380e90ec76a1eeba951e69c39c1f27cdfb2dfaa network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"72:e1:21:d1:33:c6\",\"name\":\"5aa24178e014ea1\"},{\"mac\":\"0a:58:0a:81:00:10\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/0fcad20e-9ff5-417f-9507-78b50d2cc078\"}],\"ips\":[{\"address\":\"10.129.0.16/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:32.986136 3780 base_network_controller_pods.go:487] [default/openshift-monitoring/thanos-querier-795bc75756-cpqx9] creating logical port openshift-monitoring_thanos-querier-795bc75756-cpqx9 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:32.986274 3780 kube.go:256] Updating pod openshift-monitoring/thanos-querier-795bc75756-cpqx9 I0514 16:09:33.021035 3780 pod.go:62] [openshift-monitoring/thanos-querier-795bc75756-cpqx9] pod update took 34.817323ms I0514 16:09:33.021059 3780 base_network_controller_pods.go:951] [default/openshift-monitoring/thanos-querier-795bc75756-cpqx9] addLogicalPort annotation time took 34.845515ms I0514 16:09:33.022393 3780 pods.go:271] [openshift-monitoring/thanos-querier-795bc75756-cpqx9] addLogicalPort took 36.27496ms, libovsdb time 1.009548ms I0514 16:09:33.056593 3780 obj_retry.go:492] Detected object openshift-deployment-validation-operator/99b07c45bd4393a793940f31bf2887fd12a636d02cd8e5e017ce352a0dmmnmv of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:33.056614 3780 obj_retry.go:492] Detected object openshift-deployment-validation-operator/99b07c45bd4393a793940f31bf2887fd12a636d02cd8e5e017ce352a0dmmnmv of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:33.056630 3780 obj_retry.go:492] Detected object openshift-deployment-validation-operator/99b07c45bd4393a793940f31bf2887fd12a636d02cd8e5e017ce352a0dmmnmv of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:33.056637 3780 obj_retry.go:492] Detected object openshift-deployment-validation-operator/99b07c45bd4393a793940f31bf2887fd12a636d02cd8e5e017ce352a0dmmnmv of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:33.365894 3780 cni.go:452] [openshift-monitoring/thanos-querier-795bc75756-cpqx9 6d7dab16e686c6a03f8ca192ca2a0a9b690528143349d410b72a6f8146f4d8a6 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/thanos-querier-795bc75756-cpqx9 6d7dab16e686c6a03f8ca192ca2a0a9b690528143349d410b72a6f8146f4d8a6 network default NAD default NAD key ] I0514 16:09:33.376437 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: thanos-querier-795bc75756-cpqx9, hostIfaceName: 6d7dab16e686c6a, network: default, NAD default, SandboxID: "6d7dab16e686c6a03f8ca192ca2a0a9b690528143349d410b72a6f8146f4d8a6", PCI device ID: , UID: "d1e74246-32f1-45c2-aab1-7c6b53e1e0a9", MAC: 0a:58:0a:81:00:11, IPs: [10.129.0.17/23] W0514 16:09:33.422430 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service openshift-network-diagnostics/network-check-target W0514 16:09:33.422462 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service openshift-network-diagnostics/network-check-target I0514 16:09:33.429263 3780 cni.go:473] [openshift-monitoring/thanos-querier-795bc75756-cpqx9 6d7dab16e686c6a03f8ca192ca2a0a9b690528143349d410b72a6f8146f4d8a6 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/thanos-querier-795bc75756-cpqx9 6d7dab16e686c6a03f8ca192ca2a0a9b690528143349d410b72a6f8146f4d8a6 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"8a:d7:7b:9b:49:b3\",\"name\":\"6d7dab16e686c6a\"},{\"mac\":\"0a:58:0a:81:00:11\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/b2db1fb3-225f-4ef3-a523-d7d6eab3fd95\"}],\"ips\":[{\"address\":\"10.129.0.17/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:09:34.400241 3780 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0514 16:09:34.400329 3780 util.go:870] Failed to build global endpoints for port TCP/tenancy: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0514 16:09:34.400341 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/alertmanager-main I0514 16:09:34.624068 3780 base_network_controller_pods.go:487] [default/openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76] creating logical port openshift-monitoring_monitoring-plugin-5cbb48bf6b-mlj76 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:34.624195 3780 kube.go:256] Updating pod openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76 I0514 16:09:34.645487 3780 pod.go:62] [openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76] pod update took 21.3157ms I0514 16:09:34.645525 3780 base_network_controller_pods.go:951] [default/openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76] addLogicalPort annotation time took 21.359005ms I0514 16:09:34.646852 3780 pods.go:271] [openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76] addLogicalPort took 22.798739ms, libovsdb time 932.621µs W0514 16:09:34.707307 3780 base_network_controller_namespace.go:458] Unable to remove remote zone pod's openshift-console/console-856dcf8fc8-b7plv IP address from the namespace address-set, err: pod openshift-console/console-856dcf8fc8-b7plv: no pod IPs found I0514 16:09:34.725214 3780 base_network_controller_pods.go:487] [default/openshift-monitoring/metrics-server-945fdb585-d4vhv] creating logical port openshift-monitoring_metrics-server-945fdb585-d4vhv for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:34.725335 3780 kube.go:256] Updating pod openshift-monitoring/metrics-server-945fdb585-d4vhv I0514 16:09:34.752543 3780 pod.go:62] [openshift-monitoring/metrics-server-945fdb585-d4vhv] pod update took 27.229079ms I0514 16:09:34.752570 3780 base_network_controller_pods.go:951] [default/openshift-monitoring/metrics-server-945fdb585-d4vhv] addLogicalPort annotation time took 27.260592ms I0514 16:09:34.753859 3780 pods.go:271] [openshift-monitoring/metrics-server-945fdb585-d4vhv] addLogicalPort took 28.662117ms, libovsdb time 911.051µs W0514 16:09:34.955194 3780 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0514 16:09:34.955265 3780 util.go:870] Failed to build global endpoints for port TCP/tenancy: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0514 16:09:34.955281 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/alertmanager-main I0514 16:09:35.086340 3780 cni.go:452] [openshift-monitoring/metrics-server-945fdb585-d4vhv cb41ecf8692efecbffcfa3736e44877011c4ac126a03f1c82c00021ea9330a81 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/metrics-server-945fdb585-d4vhv cb41ecf8692efecbffcfa3736e44877011c4ac126a03f1c82c00021ea9330a81 network default NAD default NAD key ] I0514 16:09:35.095895 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: metrics-server-945fdb585-d4vhv, hostIfaceName: cb41ecf8692efec, network: default, NAD default, SandboxID: "cb41ecf8692efecbffcfa3736e44877011c4ac126a03f1c82c00021ea9330a81", PCI device ID: , UID: "02111b25-ff29-41ca-bdc0-1f4e08d31d3a", MAC: 0a:58:0a:81:00:13, IPs: [10.129.0.19/23] I0514 16:09:35.145460 3780 cni.go:473] [openshift-monitoring/metrics-server-945fdb585-d4vhv cb41ecf8692efecbffcfa3736e44877011c4ac126a03f1c82c00021ea9330a81 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/metrics-server-945fdb585-d4vhv cb41ecf8692efecbffcfa3736e44877011c4ac126a03f1c82c00021ea9330a81 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"8a:ba:ad:a0:b7:1b\",\"name\":\"cb41ecf8692efec\"},{\"mac\":\"0a:58:0a:81:00:13\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/a770f26f-12da-4e1c-b158-7e5a93472bc6\"}],\"ips\":[{\"address\":\"10.129.0.19/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:35.388810 3780 base_network_controller_pods.go:487] [default/openshift-monitoring/telemeter-client-5458ddb65-6x7n9] creating logical port openshift-monitoring_telemeter-client-5458ddb65-6x7n9 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:35.388947 3780 kube.go:256] Updating pod openshift-monitoring/telemeter-client-5458ddb65-6x7n9 I0514 16:09:35.413385 3780 pod.go:62] [openshift-monitoring/telemeter-client-5458ddb65-6x7n9] pod update took 24.470344ms I0514 16:09:35.413424 3780 base_network_controller_pods.go:951] [default/openshift-monitoring/telemeter-client-5458ddb65-6x7n9] addLogicalPort annotation time took 24.496639ms I0514 16:09:35.414700 3780 pods.go:271] [openshift-monitoring/telemeter-client-5458ddb65-6x7n9] addLogicalPort took 25.907925ms, libovsdb time 948.722µs I0514 16:09:36.054349 3780 cni.go:452] [openshift-monitoring/telemeter-client-5458ddb65-6x7n9 1d9c372907ba019c976f7cb06f359d763a1ecef18fcbb2680e776a3138acfd8e network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/telemeter-client-5458ddb65-6x7n9 1d9c372907ba019c976f7cb06f359d763a1ecef18fcbb2680e776a3138acfd8e network default NAD default NAD key ] I0514 16:09:36.055693 3780 cni.go:452] [openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76 1e0124c2bed76b87e12ac6a7d7e1a656af5f087c0063a7b3d06045b570649d95 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76 1e0124c2bed76b87e12ac6a7d7e1a656af5f087c0063a7b3d06045b570649d95 network default NAD default NAD key ] I0514 16:09:36.068336 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: telemeter-client-5458ddb65-6x7n9, hostIfaceName: 1d9c372907ba019, network: default, NAD default, SandboxID: "1d9c372907ba019c976f7cb06f359d763a1ecef18fcbb2680e776a3138acfd8e", PCI device ID: , UID: "39a982f6-e3ea-4ef5-905b-cef5d5b7c83d", MAC: 0a:58:0a:81:00:14, IPs: [10.129.0.20/23] I0514 16:09:36.068528 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: monitoring-plugin-5cbb48bf6b-mlj76, hostIfaceName: 1e0124c2bed76b8, network: default, NAD default, SandboxID: "1e0124c2bed76b87e12ac6a7d7e1a656af5f087c0063a7b3d06045b570649d95", PCI device ID: , UID: "4c77e7c6-9711-44d3-a1a9-2824045669bf", MAC: 0a:58:0a:81:00:12, IPs: [10.129.0.18/23] I0514 16:09:36.138030 3780 cni.go:473] [openshift-monitoring/telemeter-client-5458ddb65-6x7n9 1d9c372907ba019c976f7cb06f359d763a1ecef18fcbb2680e776a3138acfd8e network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/telemeter-client-5458ddb65-6x7n9 1d9c372907ba019c976f7cb06f359d763a1ecef18fcbb2680e776a3138acfd8e network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"22:71:43:16:99:91\",\"name\":\"1d9c372907ba019\"},{\"mac\":\"0a:58:0a:81:00:14\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/3cbfdbf7-76a5-4cc4-9921-b396be321fc7\"}],\"ips\":[{\"address\":\"10.129.0.20/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:36.158941 3780 cni.go:473] [openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76 1e0124c2bed76b87e12ac6a7d7e1a656af5f087c0063a7b3d06045b570649d95 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/monitoring-plugin-5cbb48bf6b-mlj76 1e0124c2bed76b87e12ac6a7d7e1a656af5f087c0063a7b3d06045b570649d95 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"72:d9:c3:5c:7d:83\",\"name\":\"1e0124c2bed76b8\"},{\"mac\":\"0a:58:0a:81:00:12\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/defb2fab-cd42-44b1-b3e1-5ccf67e8c797\"}],\"ips\":[{\"address\":\"10.129.0.18/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:36.433811 3780 base_network_controller_pods.go:487] [default/openshift-monitoring/prometheus-k8s-1] creating logical port openshift-monitoring_prometheus-k8s-1 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:36.433940 3780 kube.go:256] Updating pod openshift-monitoring/prometheus-k8s-1 I0514 16:09:36.462361 3780 pod.go:62] [openshift-monitoring/prometheus-k8s-1] pod update took 28.467787ms I0514 16:09:36.462392 3780 base_network_controller_pods.go:951] [default/openshift-monitoring/prometheus-k8s-1] addLogicalPort annotation time took 28.504839ms I0514 16:09:36.464017 3780 pods.go:271] [openshift-monitoring/prometheus-k8s-1] addLogicalPort took 30.221632ms, libovsdb time 1.116942ms I0514 16:09:36.635128 3780 namespace.go:142] [openshift-deployment-validation-operator] updating namespace I0514 16:09:36.649163 3780 namespace.go:142] [openshift-deployment-validation-operator] updating namespace I0514 16:09:36.816070 3780 cni.go:452] [openshift-monitoring/prometheus-k8s-1 e524f3e0ac0263cc12a2f94a2f87c7139389e3a279c9f69b10c9f7f91517c8c8 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/prometheus-k8s-1 e524f3e0ac0263cc12a2f94a2f87c7139389e3a279c9f69b10c9f7f91517c8c8 network default NAD default NAD key ] I0514 16:09:36.825954 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: prometheus-k8s-1, hostIfaceName: e524f3e0ac0263c, network: default, NAD default, SandboxID: "e524f3e0ac0263cc12a2f94a2f87c7139389e3a279c9f69b10c9f7f91517c8c8", PCI device ID: , UID: "f86ddebe-8cdb-499b-ba7b-f8ec244b5d45", MAC: 0a:58:0a:81:00:15, IPs: [10.129.0.21/23] I0514 16:09:36.876131 3780 cni.go:473] [openshift-monitoring/prometheus-k8s-1 e524f3e0ac0263cc12a2f94a2f87c7139389e3a279c9f69b10c9f7f91517c8c8 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/prometheus-k8s-1 e524f3e0ac0263cc12a2f94a2f87c7139389e3a279c9f69b10c9f7f91517c8c8 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"c6:84:83:73:c9:37\",\"name\":\"e524f3e0ac0263c\"},{\"mac\":\"0a:58:0a:81:00:15\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/f826d328-01bd-4046-8da7-98c7b007176d\"}],\"ips\":[{\"address\":\"10.129.0.21/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:09:37.955832 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-monitoring/metrics-server W0514 16:09:38.451093 3780 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:09:38.451106 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:09:38.475101 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-monitoring/metrics-server W0514 16:09:38.613010 3780 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:09:38.613041 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:09:38.969524 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-monitoring/monitoring-plugin W0514 16:09:39.477000 3780 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:39.477018 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:39.477025 3780 util.go:870] Failed to build global endpoints for port TCP/tenancy: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:39.477031 3780 util.go:870] Failed to build global endpoints for port TCP/tenancy-rules: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:39.991220 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:39.991235 3780 util.go:870] Failed to build global endpoints for port TCP/tenancy: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:39.991240 3780 util.go:870] Failed to build global endpoints for port TCP/tenancy-rules: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:39.991246 3780 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:41.483759 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:41.483785 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9094: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:41.483792 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:41.483797 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9093: empty IP address endpoints for service openshift-monitoring/thanos-querier W0514 16:09:41.729586 3780 base_network_controller_namespace.go:458] Unable to remove remote zone pod's openshift-console/console-685d749d99-lxh2s IP address from the namespace address-set, err: pod openshift-console/console-685d749d99-lxh2s: no pod IPs found W0514 16:09:41.729644 3780 base_network_controller_namespace.go:458] Unable to remove remote zone pod's openshift-console/console-685d749d99-pqlzm IP address from the namespace address-set, err: pod openshift-console/console-685d749d99-pqlzm: no pod IPs found W0514 16:09:43.010270 3780 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-deployment-validation-operator/deployment-validation-operator-metrics I0514 16:09:43.623960 3780 cni.go:452] [openshift-console/console-bf589dbdc-kf6q4 dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62 network default NAD default NAD key ] DEL starting CNI request [openshift-console/console-bf589dbdc-kf6q4 dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62 network default NAD default NAD key ] I0514 16:09:43.695208 3780 cni.go:473] [openshift-console/console-bf589dbdc-kf6q4 dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62 network default NAD default NAD key default] DEL finished CNI request [openshift-console/console-bf589dbdc-kf6q4 dfc6fd367226bbc306cf9a188e663d2615daf9aa791f5488be8ba598372f3d62 network default NAD default NAD key default], result "{}", err I0514 16:09:43.739835 3780 obj_retry.go:498] Detected object openshift-console/console-bf589dbdc-kf6q4 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:43.739872 3780 pods.go:176] Deleting pod: openshift-console/console-bf589dbdc-kf6q4 I0514 16:09:43.741446 3780 pods.go:236] Attempting to release IPs for pod: openshift-console/console-bf589dbdc-kf6q4, ips: 10.129.0.13 I0514 16:09:43.741472 3780 obj_retry.go:498] Detected object openshift-console/console-bf589dbdc-kf6q4 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:43.750914 3780 base_network_controller_pods.go:487] [default/openshift-console/console-5ddf55bfc7-5dcwk] creating logical port openshift-console_console-5ddf55bfc7-5dcwk for pod on switch ip-10-0-0-158.ec2.internal I0514 16:09:43.751024 3780 kube.go:256] Updating pod openshift-console/console-5ddf55bfc7-5dcwk I0514 16:09:43.767572 3780 pod.go:62] [openshift-console/console-5ddf55bfc7-5dcwk] pod update took 16.575818ms I0514 16:09:43.767592 3780 base_network_controller_pods.go:951] [default/openshift-console/console-5ddf55bfc7-5dcwk] addLogicalPort annotation time took 16.598822ms I0514 16:09:43.768718 3780 pods.go:271] [openshift-console/console-5ddf55bfc7-5dcwk] addLogicalPort took 17.816844ms, libovsdb time 823.031µs I0514 16:09:44.149633 3780 cni.go:452] [openshift-console/console-5ddf55bfc7-5dcwk d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab network default NAD default NAD key ] ADD starting CNI request [openshift-console/console-5ddf55bfc7-5dcwk d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab network default NAD default NAD key ] I0514 16:09:44.159005 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-console, podName: console-5ddf55bfc7-5dcwk, hostIfaceName: d57252abe4a6962, network: default, NAD default, SandboxID: "d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab", PCI device ID: , UID: "03cc8832-5cc2-4e5e-8031-d0caecc22208", MAC: 0a:58:0a:81:00:16, IPs: [10.129.0.22/23] I0514 16:09:44.206820 3780 cni.go:473] [openshift-console/console-5ddf55bfc7-5dcwk d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab network default NAD default NAD key default] ADD finished CNI request [openshift-console/console-5ddf55bfc7-5dcwk d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"2e:76:cd:06:47:cd\",\"name\":\"d57252abe4a6962\"},{\"mac\":\"0a:58:0a:81:00:16\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/710d8fe4-9856-48d5-8d6c-08b12f2bb905\"}],\"ips\":[{\"address\":\"10.129.0.22/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:09:44.478139 3780 obj_retry.go:492] Detected object openshift-console/console-bf589dbdc-kf6q4 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:44.478170 3780 obj_retry.go:492] Detected object openshift-console/console-bf589dbdc-kf6q4 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:44.482396 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-bf589dbdc-kf6q4 I0514 16:09:44.482454 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-bf589dbdc-kf6q4 W0514 16:09:46.638012 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-insights/metrics W0514 16:09:51.320772 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service openshift-network-diagnostics/network-check-target W0514 16:09:52.664600 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-kube-storage-version-migrator-operator/metrics I0514 16:09:53.703613 3780 cni.go:452] [openshift-image-registry/image-registry-84c97879cc-rjbmv 173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825 network default NAD default NAD key ] DEL starting CNI request [openshift-image-registry/image-registry-84c97879cc-rjbmv 173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825 network default NAD default NAD key ] I0514 16:09:53.774598 3780 cni.go:473] [openshift-image-registry/image-registry-84c97879cc-rjbmv 173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825 network default NAD default NAD key default] DEL finished CNI request [openshift-image-registry/image-registry-84c97879cc-rjbmv 173437752a140f99b3f04bc36fa1f3c4fb929374e2dde35c68237991be8e5825 network default NAD default NAD key default], result "{}", err I0514 16:09:53.846651 3780 obj_retry.go:498] Detected object openshift-image-registry/image-registry-84c97879cc-rjbmv of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:53.846680 3780 pods.go:176] Deleting pod: openshift-image-registry/image-registry-84c97879cc-rjbmv I0514 16:09:53.847982 3780 pods.go:236] Attempting to release IPs for pod: openshift-image-registry/image-registry-84c97879cc-rjbmv, ips: 10.129.0.6 I0514 16:09:53.848006 3780 obj_retry.go:498] Detected object openshift-image-registry/image-registry-84c97879cc-rjbmv of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it W0514 16:09:54.519074 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console I0514 16:09:54.531895 3780 obj_retry.go:492] Detected object openshift-image-registry/image-registry-84c97879cc-rjbmv of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:54.531933 3780 obj_retry.go:492] Detected object openshift-image-registry/image-registry-84c97879cc-rjbmv of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:54.541267 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-image-registry/image-registry-84c97879cc-rjbmv I0514 16:09:54.541292 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-image-registry/image-registry-84c97879cc-rjbmv I0514 16:09:54.980432 3780 obj_retry.go:498] Detected object openshift-console/console-bf589dbdc-2dcdv of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:54.980956 3780 obj_retry.go:498] Detected object openshift-console/console-bf589dbdc-2dcdv of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:09:55.068225 3780 obj_retry.go:492] Detected object openshift-console/console-bf589dbdc-2dcdv of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:55.068246 3780 obj_retry.go:492] Detected object openshift-console/console-bf589dbdc-2dcdv of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:09:55.074980 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-bf589dbdc-2dcdv I0514 16:09:55.075001 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-bf589dbdc-2dcdv W0514 16:09:58.980076 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 10250: empty IP address endpoints for service openshift-monitoring/metrics-server W0514 16:09:58.980100 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 10250: empty IP address endpoints for service openshift-monitoring/metrics-server W0514 16:09:59.443021 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 10250: empty IP address endpoints for service openshift-monitoring/metrics-server I0514 16:10:00.193135 3780 base_network_controller_pods.go:487] [default/openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8] creating logical port openshift-backplane_osd-delete-backplane-serviceaccounts-29646250-8h5z8 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:10:00.193271 3780 kube.go:256] Updating pod openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 I0514 16:10:00.208029 3780 pod.go:62] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8] pod update took 14.785207ms I0514 16:10:00.208049 3780 base_network_controller_pods.go:951] [default/openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8] addLogicalPort annotation time took 14.809121ms I0514 16:10:00.209230 3780 pods.go:271] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8] addLogicalPort took 16.113492ms, libovsdb time 875.004µs I0514 16:10:00.596802 3780 cni.go:452] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1 network default NAD default NAD key ] ADD starting CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1 network default NAD default NAD key ] I0514 16:10:00.605888 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-backplane, podName: osd-delete-backplane-serviceaccounts-29646250-8h5z8, hostIfaceName: 360ff73e16268a3, network: default, NAD default, SandboxID: "360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1", PCI device ID: , UID: "a9160e68-233e-4bd3-8bc2-14233406f94d", MAC: 0a:58:0a:81:00:17, IPs: [10.129.0.23/23] I0514 16:10:00.653375 3780 cni.go:473] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1 network default NAD default NAD key default] ADD finished CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ea:8e:36:85:57:4a\",\"name\":\"360ff73e16268a3\"},{\"mac\":\"0a:58:0a:81:00:17\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/c83a9976-b6c0-41d4-b250-adbc07714ed7\"}],\"ips\":[{\"address\":\"10.129.0.23/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:10:02.079078 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9095: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0514 16:10:02.079097 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0514 16:10:02.079104 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9097: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0514 16:10:06.721753 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-service-ca-operator/metrics I0514 16:10:19.860565 3780 obj_retry.go:498] Detected object openshift-console/console-856dcf8fc8-zkm56 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:10:19.861084 3780 obj_retry.go:498] Detected object openshift-console/console-856dcf8fc8-zkm56 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:10:20.784231 3780 obj_retry.go:492] Detected object openshift-console/console-856dcf8fc8-zkm56 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:10:20.784259 3780 obj_retry.go:492] Detected object openshift-console/console-856dcf8fc8-zkm56 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:10:20.789200 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-856dcf8fc8-zkm56 I0514 16:10:20.789225 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-856dcf8fc8-zkm56 I0514 16:10:24.355201 3780 obj_retry.go:498] Detected object openshift-ingress/router-default-6d5d57f4bf-g7mz6 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:10:24.355757 3780 obj_retry.go:498] Detected object openshift-ingress/router-default-6d5d57f4bf-g7mz6 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:10:24.799247 3780 obj_retry.go:492] Detected object openshift-ingress/router-default-6d5d57f4bf-g7mz6 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:10:24.799270 3780 obj_retry.go:492] Detected object openshift-ingress/router-default-6d5d57f4bf-g7mz6 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:10:24.805516 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-ingress/router-default-6d5d57f4bf-g7mz6 I0514 16:10:24.805539 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-ingress/router-default-6d5d57f4bf-g7mz6 W0514 16:10:36.894865 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:10:36.894887 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:10:36.894900 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:10:36.894907 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:10:37.710637 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0514 16:10:37.710684 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/prometheus-k8s I0514 16:10:44.719315 3780 cni.go:452] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1 network default NAD default NAD key ] DEL starting CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1 network default NAD default NAD key ] I0514 16:10:44.800006 3780 cni.go:473] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1 network default NAD default NAD key default] DEL finished CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 360ff73e16268a35e086ac05f5e9233a69c0dae9b0d61bf13dbbd296b57d5ea1 network default NAD default NAD key default], result "{}", err I0514 16:10:44.842508 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:10:44.842536 3780 pods.go:176] Deleting pod: openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 I0514 16:10:44.843764 3780 pods.go:236] Attempting to release IPs for pod: openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8, ips: 10.129.0.23 I0514 16:10:44.843794 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:10:45.729314 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:10:45.729338 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:11:14.626804 3780 cni.go:452] [openshift-console/console-5ddf55bfc7-5dcwk d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab network default NAD default NAD key ] DEL starting CNI request [openshift-console/console-5ddf55bfc7-5dcwk d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab network default NAD default NAD key ] I0514 16:11:14.705292 3780 cni.go:473] [openshift-console/console-5ddf55bfc7-5dcwk d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab network default NAD default NAD key default] DEL finished CNI request [openshift-console/console-5ddf55bfc7-5dcwk d57252abe4a69629030cb7d223b51fddc2175c3b5f1822a69e2c8c3ce2210bab network default NAD default NAD key default], result "{}", err I0514 16:11:14.752368 3780 obj_retry.go:498] Detected object openshift-console/console-5ddf55bfc7-5dcwk of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:11:14.752423 3780 pods.go:176] Deleting pod: openshift-console/console-5ddf55bfc7-5dcwk I0514 16:11:14.753784 3780 pods.go:236] Attempting to release IPs for pod: openshift-console/console-5ddf55bfc7-5dcwk, ips: 10.129.0.22 I0514 16:11:14.753817 3780 obj_retry.go:498] Detected object openshift-console/console-5ddf55bfc7-5dcwk of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:11:14.761373 3780 base_network_controller_pods.go:487] [default/openshift-console/console-5798b4f98f-db8m7] creating logical port openshift-console_console-5798b4f98f-db8m7 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:11:14.761530 3780 kube.go:256] Updating pod openshift-console/console-5798b4f98f-db8m7 I0514 16:11:14.775964 3780 pod.go:62] [openshift-console/console-5798b4f98f-db8m7] pod update took 14.467948ms I0514 16:11:14.775983 3780 base_network_controller_pods.go:951] [default/openshift-console/console-5798b4f98f-db8m7] addLogicalPort annotation time took 14.490655ms I0514 16:11:14.777094 3780 pods.go:271] [openshift-console/console-5798b4f98f-db8m7] addLogicalPort took 15.734997ms, libovsdb time 804.229µs I0514 16:11:15.119944 3780 cni.go:452] [openshift-console/console-5798b4f98f-db8m7 968e38dbc7f0acd430244cebcad3a24f47e9cb99015c64f0e48424d74e46b500 network default NAD default NAD key ] ADD starting CNI request [openshift-console/console-5798b4f98f-db8m7 968e38dbc7f0acd430244cebcad3a24f47e9cb99015c64f0e48424d74e46b500 network default NAD default NAD key ] I0514 16:11:15.124234 3780 obj_retry.go:492] Detected object openshift-console/console-5ddf55bfc7-5dcwk of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:11:15.124260 3780 obj_retry.go:492] Detected object openshift-console/console-5ddf55bfc7-5dcwk of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:11:15.129494 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-console, podName: console-5798b4f98f-db8m7, hostIfaceName: 968e38dbc7f0acd, network: default, NAD default, SandboxID: "968e38dbc7f0acd430244cebcad3a24f47e9cb99015c64f0e48424d74e46b500", PCI device ID: , UID: "c58395dc-2b2d-4654-b20d-8d5ba4cb43e3", MAC: 0a:58:0a:81:00:18, IPs: [10.129.0.24/23] I0514 16:11:15.130494 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-5ddf55bfc7-5dcwk I0514 16:11:15.130519 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-5ddf55bfc7-5dcwk I0514 16:11:15.177053 3780 cni.go:473] [openshift-console/console-5798b4f98f-db8m7 968e38dbc7f0acd430244cebcad3a24f47e9cb99015c64f0e48424d74e46b500 network default NAD default NAD key default] ADD finished CNI request [openshift-console/console-5798b4f98f-db8m7 968e38dbc7f0acd430244cebcad3a24f47e9cb99015c64f0e48424d74e46b500 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ae:c5:59:ca:a4:d6\",\"name\":\"968e38dbc7f0acd\"},{\"mac\":\"0a:58:0a:81:00:18\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/e98946bb-36df-4f6c-9e1f-656d09549f9f\"}],\"ips\":[{\"address\":\"10.129.0.24/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:11:25.268946 3780 obj_retry.go:498] Detected object openshift-console/console-5ddf55bfc7-rsjkj of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:11:25.269603 3780 obj_retry.go:498] Detected object openshift-console/console-5ddf55bfc7-rsjkj of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:11:25.416334 3780 obj_retry.go:492] Detected object openshift-console/console-5ddf55bfc7-rsjkj of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:11:25.416355 3780 obj_retry.go:492] Detected object openshift-console/console-5ddf55bfc7-rsjkj of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:11:25.422129 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-5ddf55bfc7-rsjkj I0514 16:11:25.422155 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-5ddf55bfc7-rsjkj W0514 16:11:25.890959 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console I0514 16:12:42.888657 3780 namespace.go:142] [openshift-kube-storage-version-migrator] updating namespace I0514 16:13:35.401078 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=122 I0514 16:13:37.487658 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=5 I0514 16:14:07.519189 3780 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=5 I0514 16:14:29.488201 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=7 I0514 16:14:36.396311 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=6 I0514 16:15:33.513903 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=6 I0514 16:15:46.508761 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=9 I0514 16:15:47.341574 3780 namespace.go:98] [tsf] adding namespace I0514 16:15:47.342575 3780 namespace.go:102] [tsf] adding namespace took 977.94µs I0514 16:15:47.359922 3780 namespace.go:142] [tsf] updating namespace I0514 16:15:47.375399 3780 namespace.go:142] [tsf] updating namespace I0514 16:15:51.223953 3780 namespace.go:98] [cert-manager-operator] adding namespace I0514 16:15:51.224066 3780 namespace.go:98] [tsf-tpa] adding namespace I0514 16:15:51.224844 3780 namespace.go:102] [cert-manager-operator] adding namespace took 869.623µs I0514 16:15:51.225474 3780 namespace.go:102] [tsf-tpa] adding namespace took 1.388981ms I0514 16:15:51.228364 3780 namespace.go:98] [rhbk-operator] adding namespace I0514 16:15:51.228501 3780 namespace.go:98] [konflux-ui] adding namespace I0514 16:15:51.228978 3780 namespace.go:102] [rhbk-operator] adding namespace took 600.396µs I0514 16:15:51.229698 3780 namespace.go:102] [konflux-ui] adding namespace took 1.181908ms I0514 16:15:51.232322 3780 namespace.go:98] [rhtpa-operator] adding namespace I0514 16:15:51.232366 3780 namespace.go:98] [tsf-tas] adding namespace I0514 16:15:51.232929 3780 namespace.go:102] [rhtpa-operator] adding namespace took 594.004µs I0514 16:15:51.232952 3780 namespace.go:98] [konflux-operator] adding namespace I0514 16:15:51.233546 3780 namespace.go:102] [konflux-operator] adding namespace took 585.992µs I0514 16:15:51.234142 3780 namespace.go:102] [tsf-tas] adding namespace took 1.758381ms I0514 16:15:51.234332 3780 namespace.go:98] [tsf-keycloak] adding namespace I0514 16:15:51.234914 3780 namespace.go:102] [tsf-keycloak] adding namespace took 567.765µs I0514 16:15:51.250923 3780 namespace.go:142] [cert-manager-operator] updating namespace I0514 16:15:51.332237 3780 namespace.go:142] [cert-manager-operator] updating namespace I0514 16:15:51.348848 3780 namespace.go:142] [rhbk-operator] updating namespace I0514 16:15:51.363024 3780 namespace.go:142] [tsf-tpa] updating namespace I0514 16:15:51.363100 3780 namespace.go:142] [rhtpa-operator] updating namespace I0514 16:15:51.397128 3780 namespace.go:142] [konflux-ui] updating namespace I0514 16:15:51.418111 3780 namespace.go:142] [konflux-operator] updating namespace I0514 16:15:51.418275 3780 namespace.go:142] [tsf-tas] updating namespace I0514 16:15:51.418326 3780 namespace.go:142] [tsf-keycloak] updating namespace I0514 16:15:51.444147 3780 namespace.go:142] [cert-manager-operator] updating namespace I0514 16:15:51.493032 3780 namespace.go:142] [rhbk-operator] updating namespace I0514 16:15:51.546068 3780 namespace.go:142] [rhbk-operator] updating namespace I0514 16:15:51.550619 3780 namespace.go:142] [konflux-ui] updating namespace I0514 16:15:51.569117 3780 namespace.go:142] [konflux-ui] updating namespace I0514 16:15:51.625793 3780 namespace.go:142] [rhtpa-operator] updating namespace I0514 16:15:51.648932 3780 namespace.go:142] [rhtpa-operator] updating namespace I0514 16:15:51.667961 3780 namespace.go:142] [tsf-tas] updating namespace I0514 16:15:51.693188 3780 namespace.go:142] [tsf-tas] updating namespace I0514 16:15:51.739324 3780 namespace.go:142] [konflux-operator] updating namespace I0514 16:15:51.762623 3780 namespace.go:142] [konflux-operator] updating namespace I0514 16:15:51.777907 3780 namespace.go:142] [tsf-keycloak] updating namespace I0514 16:15:51.830009 3780 namespace.go:142] [tsf-keycloak] updating namespace I0514 16:15:51.863262 3780 namespace.go:142] [tsf-tpa] updating namespace I0514 16:15:51.885047 3780 namespace.go:142] [tsf-tpa] updating namespace I0514 16:15:52.837904 3780 namespace.go:142] [tsf-tpa] updating namespace I0514 16:15:52.840543 3780 namespace.go:142] [tsf-keycloak] updating namespace I0514 16:15:54.401780 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=53 I0514 16:15:58.697729 3780 obj_retry.go:498] Detected object openshift-marketplace/79f871b6353a8442b4f467c124caf7fa84891a4baa69c2582c3b4a7b51nz2vc of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:15:58.698269 3780 obj_retry.go:498] Detected object openshift-marketplace/79f871b6353a8442b4f467c124caf7fa84891a4baa69c2582c3b4a7b51nz2vc of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:15:58.698295 3780 obj_retry.go:498] Detected object openshift-marketplace/79f871b6353a8442b4f467c124caf7fa84891a4baa69c2582c3b4a7b51nz2vc of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:15:58.698304 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0514 16:15:58.698315 3780 obj_retry.go:498] Detected object openshift-marketplace/79f871b6353a8442b4f467c124caf7fa84891a4baa69c2582c3b4a7b51nz2vc of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:15:58.698320 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0514 16:15:59.725213 3780 obj_retry.go:492] Detected object openshift-marketplace/79f871b6353a8442b4f467c124caf7fa84891a4baa69c2582c3b4a7b51nz2vc of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:15:59.725235 3780 obj_retry.go:492] Detected object openshift-marketplace/79f871b6353a8442b4f467c124caf7fa84891a4baa69c2582c3b4a7b51nz2vc of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:15:59.725251 3780 obj_retry.go:492] Detected object openshift-marketplace/79f871b6353a8442b4f467c124caf7fa84891a4baa69c2582c3b4a7b51nz2vc of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:15:59.725264 3780 obj_retry.go:492] Detected object openshift-marketplace/79f871b6353a8442b4f467c124caf7fa84891a4baa69c2582c3b4a7b51nz2vc of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:01.848523 3780 obj_retry.go:498] Detected object openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dl66v4 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:01.849085 3780 obj_retry.go:498] Detected object openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dl66v4 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:01.849113 3780 obj_retry.go:498] Detected object openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dl66v4 of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:01.849121 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0514 16:16:01.849131 3780 obj_retry.go:498] Detected object openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dl66v4 of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:01.849136 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0514 16:16:02.875044 3780 obj_retry.go:492] Detected object openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dl66v4 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:02.875067 3780 obj_retry.go:492] Detected object openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dl66v4 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:02.875082 3780 obj_retry.go:492] Detected object openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dl66v4 of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:02.875094 3780 obj_retry.go:492] Detected object openshift-marketplace/2bb52b92bd31ddf2ebbc335370ac517be50e61a93c2fe375393413c19dl66v4 of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:05.783554 3780 base_network_controller_pods.go:487] [default/tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c] creating logical port tsf-tpa_rhtpa-operator-controller-manager-586cbb496b-xrj9c for pod on switch ip-10-0-0-158.ec2.internal I0514 16:16:05.783677 3780 kube.go:256] Updating pod tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c I0514 16:16:05.801177 3780 pod.go:62] [tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c] pod update took 17.539434ms I0514 16:16:05.801196 3780 base_network_controller_pods.go:951] [default/tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c] addLogicalPort annotation time took 17.562476ms I0514 16:16:05.802363 3780 pods.go:271] [tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c] addLogicalPort took 18.82573ms, libovsdb time 851.275µs I0514 16:16:06.143394 3780 cni.go:452] [tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c dc474bdb3caf1ddf938c139bed5aae50e7b0dd2fba976c01e6781825b18d7237 network default NAD default NAD key ] ADD starting CNI request [tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c dc474bdb3caf1ddf938c139bed5aae50e7b0dd2fba976c01e6781825b18d7237 network default NAD default NAD key ] I0514 16:16:06.152469 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tpa, podName: rhtpa-operator-controller-manager-586cbb496b-xrj9c, hostIfaceName: dc474bdb3caf1dd, network: default, NAD default, SandboxID: "dc474bdb3caf1ddf938c139bed5aae50e7b0dd2fba976c01e6781825b18d7237", PCI device ID: , UID: "2b3e6e5c-2b9f-4f01-b928-352462409a57", MAC: 0a:58:0a:81:00:19, IPs: [10.129.0.25/23] I0514 16:16:06.199896 3780 cni.go:473] [tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c dc474bdb3caf1ddf938c139bed5aae50e7b0dd2fba976c01e6781825b18d7237 network default NAD default NAD key default] ADD finished CNI request [tsf-tpa/rhtpa-operator-controller-manager-586cbb496b-xrj9c dc474bdb3caf1ddf938c139bed5aae50e7b0dd2fba976c01e6781825b18d7237 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"12:1a:96:4d:be:8b\",\"name\":\"dc474bdb3caf1dd\"},{\"mac\":\"0a:58:0a:81:00:19\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/8179f6dc-ca9f-4081-8029-c2af1f0f901e\"}],\"ips\":[{\"address\":\"10.129.0.25/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:16:06.746146 3780 obj_retry.go:498] Detected object openshift-marketplace/5aa5e4cbc2f6ac9107b7dd5bfb7c8725ca307e639be07100d6806658feq25zs of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:06.746646 3780 obj_retry.go:498] Detected object openshift-marketplace/5aa5e4cbc2f6ac9107b7dd5bfb7c8725ca307e639be07100d6806658feq25zs of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:06.746677 3780 obj_retry.go:498] Detected object openshift-marketplace/5aa5e4cbc2f6ac9107b7dd5bfb7c8725ca307e639be07100d6806658feq25zs of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:06.746685 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0514 16:16:06.746695 3780 obj_retry.go:498] Detected object openshift-marketplace/5aa5e4cbc2f6ac9107b7dd5bfb7c8725ca307e639be07100d6806658feq25zs of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:06.746700 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0514 16:16:07.643938 3780 obj_retry.go:492] Detected object openshift-marketplace/5aa5e4cbc2f6ac9107b7dd5bfb7c8725ca307e639be07100d6806658feq25zs of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:07.643961 3780 obj_retry.go:492] Detected object openshift-marketplace/5aa5e4cbc2f6ac9107b7dd5bfb7c8725ca307e639be07100d6806658feq25zs of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:07.643973 3780 obj_retry.go:492] Detected object openshift-marketplace/5aa5e4cbc2f6ac9107b7dd5bfb7c8725ca307e639be07100d6806658feq25zs of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:07.643984 3780 obj_retry.go:492] Detected object openshift-marketplace/5aa5e4cbc2f6ac9107b7dd5bfb7c8725ca307e639be07100d6806658feq25zs of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:07.765367 3780 obj_retry.go:498] Detected object openshift-marketplace/5bfc73bccd2cef20bb9afe324d112069530a83adee9f41e094fba66f6emj9t6 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:07.765969 3780 obj_retry.go:498] Detected object openshift-marketplace/5bfc73bccd2cef20bb9afe324d112069530a83adee9f41e094fba66f6emj9t6 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:07.765998 3780 obj_retry.go:498] Detected object openshift-marketplace/5bfc73bccd2cef20bb9afe324d112069530a83adee9f41e094fba66f6emj9t6 of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:07.766006 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0514 16:16:07.766017 3780 obj_retry.go:498] Detected object openshift-marketplace/5bfc73bccd2cef20bb9afe324d112069530a83adee9f41e094fba66f6emj9t6 of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:07.766021 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0514 16:16:08.795689 3780 obj_retry.go:492] Detected object openshift-marketplace/5bfc73bccd2cef20bb9afe324d112069530a83adee9f41e094fba66f6emj9t6 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:08.795710 3780 obj_retry.go:492] Detected object openshift-marketplace/5bfc73bccd2cef20bb9afe324d112069530a83adee9f41e094fba66f6emj9t6 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:08.795722 3780 obj_retry.go:492] Detected object openshift-marketplace/5bfc73bccd2cef20bb9afe324d112069530a83adee9f41e094fba66f6emj9t6 of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:08.795737 3780 obj_retry.go:492] Detected object openshift-marketplace/5bfc73bccd2cef20bb9afe324d112069530a83adee9f41e094fba66f6emj9t6 of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed W0514 16:16:08.996789 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service I0514 16:16:10.012496 3780 obj_retry.go:498] Detected object openshift-marketplace/09e84d3fdfac1305b8f399db3dfe1b7c79d582fe1356304a82b56f71d77cmxw of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:10.013027 3780 obj_retry.go:498] Detected object openshift-marketplace/09e84d3fdfac1305b8f399db3dfe1b7c79d582fe1356304a82b56f71d77cmxw of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:10.013058 3780 obj_retry.go:498] Detected object openshift-marketplace/09e84d3fdfac1305b8f399db3dfe1b7c79d582fe1356304a82b56f71d77cmxw of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:10.013069 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0514 16:16:10.013082 3780 obj_retry.go:498] Detected object openshift-marketplace/09e84d3fdfac1305b8f399db3dfe1b7c79d582fe1356304a82b56f71d77cmxw of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:10.013086 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0514 16:16:10.213858 3780 obj_retry.go:498] Detected object openshift-marketplace/03dabd08ddba15f575168653b704f77a8a8f37b7d692dad9ab1e8c2136zcf9s of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:10.214360 3780 obj_retry.go:498] Detected object openshift-marketplace/03dabd08ddba15f575168653b704f77a8a8f37b7d692dad9ab1e8c2136zcf9s of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:16:10.214383 3780 obj_retry.go:498] Detected object openshift-marketplace/03dabd08ddba15f575168653b704f77a8a8f37b7d692dad9ab1e8c2136zcf9s of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:10.214393 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0514 16:16:10.214432 3780 obj_retry.go:498] Detected object openshift-marketplace/03dabd08ddba15f575168653b704f77a8a8f37b7d692dad9ab1e8c2136zcf9s of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0514 16:16:10.214441 3780 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... W0514 16:16:10.227071 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service W0514 16:16:10.274664 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service I0514 16:16:11.055049 3780 obj_retry.go:492] Detected object openshift-marketplace/09e84d3fdfac1305b8f399db3dfe1b7c79d582fe1356304a82b56f71d77cmxw of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:11.055078 3780 obj_retry.go:492] Detected object openshift-marketplace/09e84d3fdfac1305b8f399db3dfe1b7c79d582fe1356304a82b56f71d77cmxw of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:11.055091 3780 obj_retry.go:492] Detected object openshift-marketplace/09e84d3fdfac1305b8f399db3dfe1b7c79d582fe1356304a82b56f71d77cmxw of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:11.055100 3780 obj_retry.go:492] Detected object openshift-marketplace/09e84d3fdfac1305b8f399db3dfe1b7c79d582fe1356304a82b56f71d77cmxw of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:11.257258 3780 obj_retry.go:492] Detected object openshift-marketplace/03dabd08ddba15f575168653b704f77a8a8f37b7d692dad9ab1e8c2136zcf9s of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:11.257281 3780 obj_retry.go:492] Detected object openshift-marketplace/03dabd08ddba15f575168653b704f77a8a8f37b7d692dad9ab1e8c2136zcf9s of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:11.257297 3780 obj_retry.go:492] Detected object openshift-marketplace/03dabd08ddba15f575168653b704f77a8a8f37b7d692dad9ab1e8c2136zcf9s of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:16:11.257309 3780 obj_retry.go:492] Detected object openshift-marketplace/03dabd08ddba15f575168653b704f77a8a8f37b7d692dad9ab1e8c2136zcf9s of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed W0514 16:16:13.463927 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service W0514 16:16:13.473229 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service W0514 16:16:13.480703 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service W0514 16:16:13.492683 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service I0514 16:16:16.979224 3780 namespace.go:98] [cert-manager] adding namespace I0514 16:16:16.980080 3780 namespace.go:102] [cert-manager] adding namespace took 826.592µs I0514 16:16:17.000046 3780 namespace.go:142] [cert-manager] updating namespace I0514 16:16:17.023001 3780 namespace.go:142] [cert-manager] updating namespace I0514 16:16:18.525098 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=9 I0514 16:16:19.902089 3780 base_network_controller_pods.go:487] [default/rhbk-operator/rhbk-operator-69889d554d-wl9qt] creating logical port rhbk-operator_rhbk-operator-69889d554d-wl9qt for pod on switch ip-10-0-0-158.ec2.internal I0514 16:16:19.902197 3780 kube.go:256] Updating pod rhbk-operator/rhbk-operator-69889d554d-wl9qt I0514 16:16:19.922107 3780 pod.go:62] [rhbk-operator/rhbk-operator-69889d554d-wl9qt] pod update took 19.938095ms I0514 16:16:19.922127 3780 base_network_controller_pods.go:951] [default/rhbk-operator/rhbk-operator-69889d554d-wl9qt] addLogicalPort annotation time took 19.962762ms I0514 16:16:19.923439 3780 pods.go:271] [rhbk-operator/rhbk-operator-69889d554d-wl9qt] addLogicalPort took 21.366024ms, libovsdb time 947.983µs W0514 16:16:20.004099 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8443: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service I0514 16:16:20.261093 3780 cni.go:452] [rhbk-operator/rhbk-operator-69889d554d-wl9qt cb00312a4ea8f7a355aa732e1ebd872a034e2f6314601a61622895d83d12902f network default NAD default NAD key ] ADD starting CNI request [rhbk-operator/rhbk-operator-69889d554d-wl9qt cb00312a4ea8f7a355aa732e1ebd872a034e2f6314601a61622895d83d12902f network default NAD default NAD key ] I0514 16:16:20.269937 3780 helper_linux.go:556] ConfigureOVS: namespace: rhbk-operator, podName: rhbk-operator-69889d554d-wl9qt, hostIfaceName: cb00312a4ea8f7a, network: default, NAD default, SandboxID: "cb00312a4ea8f7a355aa732e1ebd872a034e2f6314601a61622895d83d12902f", PCI device ID: , UID: "5c939d6e-f4d2-4733-a4d0-f13e48b59b02", MAC: 0a:58:0a:81:00:1a, IPs: [10.129.0.26/23] I0514 16:16:20.317526 3780 cni.go:473] [rhbk-operator/rhbk-operator-69889d554d-wl9qt cb00312a4ea8f7a355aa732e1ebd872a034e2f6314601a61622895d83d12902f network default NAD default NAD key default] ADD finished CNI request [rhbk-operator/rhbk-operator-69889d554d-wl9qt cb00312a4ea8f7a355aa732e1ebd872a034e2f6314601a61622895d83d12902f network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"aa:a0:50:9d:a3:eb\",\"name\":\"cb00312a4ea8f7a\"},{\"mac\":\"0a:58:0a:81:00:1a\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/95d42434-7f96-4916-96f7-739d10558f40\"}],\"ips\":[{\"address\":\"10.129.0.26/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:16:26.236975 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service cert-manager/cert-manager-webhook W0514 16:16:26.236994 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service cert-manager/cert-manager-webhook I0514 16:16:30.704386 3780 namespace.go:142] [openshift-operators] updating namespace I0514 16:16:30.760079 3780 namespace.go:142] [openshift-operators] updating namespace W0514 16:16:37.301146 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service konflux-operator/konflux-operator-controller-manager-metrics-service I0514 16:16:38.877071 3780 namespace.go:142] [openshift-operators] updating namespace I0514 16:16:39.205534 3780 base_network_controller_pods.go:487] [default/openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx] creating logical port openshift-operators_openshift-pipelines-operator-6bdf6c8794-w9nzx for pod on switch ip-10-0-0-158.ec2.internal I0514 16:16:39.205662 3780 kube.go:256] Updating pod openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx I0514 16:16:39.262131 3780 pod.go:62] [openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx] pod update took 56.493204ms I0514 16:16:39.262152 3780 base_network_controller_pods.go:951] [default/openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx] addLogicalPort annotation time took 56.520009ms I0514 16:16:39.263307 3780 pods.go:271] [openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx] addLogicalPort took 57.787469ms, libovsdb time 804.681µs I0514 16:16:39.264111 3780 base_network_controller_pods.go:487] [default/openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5] creating logical port openshift-operators_tekton-operator-webhook-5864bd5898-dpfs5 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:16:39.264213 3780 kube.go:256] Updating pod openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5 I0514 16:16:39.295040 3780 pod.go:62] [openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5] pod update took 30.844831ms I0514 16:16:39.295063 3780 base_network_controller_pods.go:951] [default/openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5] addLogicalPort annotation time took 30.872171ms I0514 16:16:39.296116 3780 pods.go:271] [openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5] addLogicalPort took 32.017417ms, libovsdb time 736.545µs I0514 16:16:39.582622 3780 cni.go:452] [openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx 2cbf1428ce2f8a8b9d102ae7f251afee192efac8c5f959c108effaa1b57c1ad4 network default NAD default NAD key ] ADD starting CNI request [openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx 2cbf1428ce2f8a8b9d102ae7f251afee192efac8c5f959c108effaa1b57c1ad4 network default NAD default NAD key ] I0514 16:16:39.591547 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-operators, podName: openshift-pipelines-operator-6bdf6c8794-w9nzx, hostIfaceName: 2cbf1428ce2f8a8, network: default, NAD default, SandboxID: "2cbf1428ce2f8a8b9d102ae7f251afee192efac8c5f959c108effaa1b57c1ad4", PCI device ID: , UID: "612cc5f0-309a-4176-ae55-22ca70d8aaac", MAC: 0a:58:0a:81:00:1b, IPs: [10.129.0.27/23] I0514 16:16:39.641563 3780 cni.go:473] [openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx 2cbf1428ce2f8a8b9d102ae7f251afee192efac8c5f959c108effaa1b57c1ad4 network default NAD default NAD key default] ADD finished CNI request [openshift-operators/openshift-pipelines-operator-6bdf6c8794-w9nzx 2cbf1428ce2f8a8b9d102ae7f251afee192efac8c5f959c108effaa1b57c1ad4 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"82:a0:a8:14:92:e8\",\"name\":\"2cbf1428ce2f8a8\"},{\"mac\":\"0a:58:0a:81:00:1b\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/9f0b1840-b9d0-4168-8cee-d4756521ca62\"}],\"ips\":[{\"address\":\"10.129.0.27/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:16:39.926288 3780 cni.go:452] [openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5 a0a076c42f75072e20eb53722443a9e1be6cc6b9660dbd71d1ddb70c5cc2afe5 network default NAD default NAD key ] ADD starting CNI request [openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5 a0a076c42f75072e20eb53722443a9e1be6cc6b9660dbd71d1ddb70c5cc2afe5 network default NAD default NAD key ] I0514 16:16:39.935621 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-operators, podName: tekton-operator-webhook-5864bd5898-dpfs5, hostIfaceName: a0a076c42f75072, network: default, NAD default, SandboxID: "a0a076c42f75072e20eb53722443a9e1be6cc6b9660dbd71d1ddb70c5cc2afe5", PCI device ID: , UID: "60a62d55-de21-41a2-8209-513f19d0a172", MAC: 0a:58:0a:81:00:1c, IPs: [10.129.0.28/23] I0514 16:16:39.982334 3780 cni.go:473] [openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5 a0a076c42f75072e20eb53722443a9e1be6cc6b9660dbd71d1ddb70c5cc2afe5 network default NAD default NAD key default] ADD finished CNI request [openshift-operators/tekton-operator-webhook-5864bd5898-dpfs5 a0a076c42f75072e20eb53722443a9e1be6cc6b9660dbd71d1ddb70c5cc2afe5 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"52:06:04:5f:6d:5b\",\"name\":\"a0a076c42f75072\"},{\"mac\":\"0a:58:0a:81:00:1c\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/d4b174ec-6287-47e1-91b7-6e73f7e5a9fb\"}],\"ips\":[{\"address\":\"10.129.0.28/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:16:43.976282 3780 base_network_controller_pods.go:487] [default/openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4] creating logical port openshift-operators_rhtas-operator-controller-manager-68888cb75-9tfn4 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:16:43.976386 3780 kube.go:256] Updating pod openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4 I0514 16:16:43.999249 3780 pod.go:62] [openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4] pod update took 22.887395ms I0514 16:16:43.999275 3780 base_network_controller_pods.go:951] [default/openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4] addLogicalPort annotation time took 22.914698ms I0514 16:16:44.000541 3780 pods.go:271] [openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4] addLogicalPort took 24.273332ms, libovsdb time 863.87µs I0514 16:16:44.334370 3780 cni.go:452] [openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4 15758ec62e1aa72d73969257041474d4dbfc85bcf1efbc9411562ec7dc8a77ca network default NAD default NAD key ] ADD starting CNI request [openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4 15758ec62e1aa72d73969257041474d4dbfc85bcf1efbc9411562ec7dc8a77ca network default NAD default NAD key ] I0514 16:16:44.345092 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-operators, podName: rhtas-operator-controller-manager-68888cb75-9tfn4, hostIfaceName: 15758ec62e1aa72, network: default, NAD default, SandboxID: "15758ec62e1aa72d73969257041474d4dbfc85bcf1efbc9411562ec7dc8a77ca", PCI device ID: , UID: "3cb151dc-c048-40a7-9819-d2069ef19cb3", MAC: 0a:58:0a:81:00:1d, IPs: [10.129.0.29/23] I0514 16:16:44.393253 3780 cni.go:473] [openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4 15758ec62e1aa72d73969257041474d4dbfc85bcf1efbc9411562ec7dc8a77ca network default NAD default NAD key default] ADD finished CNI request [openshift-operators/rhtas-operator-controller-manager-68888cb75-9tfn4 15758ec62e1aa72d73969257041474d4dbfc85bcf1efbc9411562ec7dc8a77ca network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ae:8c:2a:7e:15:b7\",\"name\":\"15758ec62e1aa72\"},{\"mac\":\"0a:58:0a:81:00:1d\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/84c20fd0-ccd6-4039-bc3b-d2df968b2a00\"}],\"ips\":[{\"address\":\"10.129.0.29/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:16:46.380105 3780 namespace.go:98] [trusted-artifact-signer] adding namespace I0514 16:16:46.380895 3780 namespace.go:102] [trusted-artifact-signer] adding namespace took 769.311µs I0514 16:16:46.406012 3780 namespace.go:142] [trusted-artifact-signer] updating namespace I0514 16:16:46.427237 3780 namespace.go:142] [trusted-artifact-signer] updating namespace W0514 16:16:47.204837 3780 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-operators/rhtas-operator-controller-manager-metrics-service I0514 16:16:54.647845 3780 namespace.go:98] [openshift-pipelines] adding namespace I0514 16:16:54.648749 3780 namespace.go:102] [openshift-pipelines] adding namespace took 880.186µs I0514 16:16:54.662506 3780 namespace.go:142] [openshift-pipelines] updating namespace I0514 16:16:55.396878 3780 namespace.go:142] [cert-manager] updating namespace I0514 16:16:55.600491 3780 namespace.go:142] [cert-manager-operator] updating namespace I0514 16:16:56.001163 3780 namespace.go:142] [konflux-operator] updating namespace I0514 16:16:56.204077 3780 namespace.go:142] [konflux-ui] updating namespace I0514 16:16:56.410712 3780 namespace.go:142] [open-cluster-management-2q9q434hk8vgse7t1urllt227aefs26c] updating namespace I0514 16:16:56.805525 3780 namespace.go:142] [rhbk-operator] updating namespace I0514 16:16:57.036476 3780 namespace.go:142] [rhtpa-operator] updating namespace I0514 16:16:57.209639 3780 namespace.go:142] [trusted-artifact-signer] updating namespace I0514 16:16:57.434184 3780 namespace.go:142] [tsf] updating namespace I0514 16:16:57.632654 3780 namespace.go:142] [tsf-keycloak] updating namespace I0514 16:16:57.835326 3780 namespace.go:142] [tsf-tas] updating namespace I0514 16:16:58.037189 3780 namespace.go:142] [tsf-tpa] updating namespace I0514 16:16:58.173166 3780 namespace.go:142] [cert-manager] updating namespace I0514 16:16:58.184264 3780 namespace.go:142] [cert-manager-operator] updating namespace I0514 16:16:58.195231 3780 namespace.go:142] [default] updating namespace I0514 16:16:58.208523 3780 namespace.go:142] [konflux-operator] updating namespace W0514 16:16:58.208607 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service openshift-operators/rhtas-operator-controller-manager-metrics-service I0514 16:16:58.218778 3780 namespace.go:142] [konflux-ui] updating namespace I0514 16:16:58.232212 3780 namespace.go:142] [open-cluster-management-2q9q434hk8vgse7t1urllt227aefs26c] updating namespace I0514 16:16:58.244196 3780 namespace.go:142] [openshift] updating namespace I0514 16:16:58.255928 3780 namespace.go:142] [rhbk-operator] updating namespace I0514 16:16:58.267013 3780 namespace.go:142] [rhtpa-operator] updating namespace I0514 16:16:58.277111 3780 namespace.go:142] [trusted-artifact-signer] updating namespace I0514 16:16:58.288290 3780 namespace.go:142] [tsf] updating namespace I0514 16:16:58.298439 3780 namespace.go:142] [tsf-keycloak] updating namespace I0514 16:16:58.308152 3780 namespace.go:142] [tsf-tas] updating namespace I0514 16:16:58.320085 3780 namespace.go:142] [tsf-tpa] updating namespace I0514 16:16:58.420044 3780 namespace.go:142] [cert-manager] updating namespace I0514 16:16:58.520481 3780 namespace.go:142] [cert-manager-operator] updating namespace I0514 16:16:58.625046 3780 namespace.go:142] [default] updating namespace I0514 16:16:58.730254 3780 namespace.go:142] [konflux-operator] updating namespace I0514 16:16:58.819690 3780 namespace.go:142] [konflux-ui] updating namespace I0514 16:16:58.920153 3780 namespace.go:142] [open-cluster-management-2q9q434hk8vgse7t1urllt227aefs26c] updating namespace I0514 16:16:59.027165 3780 namespace.go:142] [openshift] updating namespace I0514 16:16:59.136018 3780 namespace.go:142] [rhbk-operator] updating namespace I0514 16:16:59.222486 3780 namespace.go:142] [rhtpa-operator] updating namespace I0514 16:16:59.334266 3780 namespace.go:142] [trusted-artifact-signer] updating namespace I0514 16:16:59.421092 3780 namespace.go:142] [tsf] updating namespace I0514 16:16:59.521993 3780 namespace.go:142] [tsf-keycloak] updating namespace I0514 16:16:59.619362 3780 namespace.go:142] [tsf-tas] updating namespace I0514 16:16:59.726177 3780 namespace.go:142] [tsf-tpa] updating namespace I0514 16:17:00.278019 3780 namespace.go:142] [openshift-pipelines] updating namespace I0514 16:17:07.087580 3780 obj_retry.go:498] Detected object tsf/test-tsf-subscriptions of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:17:07.088118 3780 obj_retry.go:498] Detected object tsf/test-tsf-subscriptions of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:17:07.108677 3780 obj_retry.go:492] Detected object tsf/test-tsf-subscriptions of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:17:07.108697 3780 obj_retry.go:492] Detected object tsf/test-tsf-subscriptions of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:17:07.119827 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf/test-tsf-subscriptions I0514 16:17:07.119852 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf/test-tsf-subscriptions I0514 16:17:08.673165 3780 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq] creating logical port openshift-pipelines_tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq for pod on switch ip-10-0-0-158.ec2.internal I0514 16:17:08.673262 3780 kube.go:256] Updating pod openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq I0514 16:17:08.688341 3780 pod.go:62] [openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq] pod update took 15.101931ms I0514 16:17:08.688366 3780 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq] addLogicalPort annotation time took 15.131461ms I0514 16:17:08.689601 3780 pods.go:271] [openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq] addLogicalPort took 16.453988ms, libovsdb time 814.8µs I0514 16:17:09.048063 3780 cni.go:452] [openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq 7ad415ed7d418ed3778d5cf92c8a29d82dbda12b8417cfdbd092bc9158ab7222 network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq 7ad415ed7d418ed3778d5cf92c8a29d82dbda12b8417cfdbd092bc9158ab7222 network default NAD default NAD key ] I0514 16:17:09.057534 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq, hostIfaceName: 7ad415ed7d418ed, network: default, NAD default, SandboxID: "7ad415ed7d418ed3778d5cf92c8a29d82dbda12b8417cfdbd092bc9158ab7222", PCI device ID: , UID: "6598058c-a80f-497f-a32f-e58163e9a0b1", MAC: 0a:58:0a:81:00:1e, IPs: [10.129.0.30/23] I0514 16:17:09.109584 3780 cni.go:473] [openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq 7ad415ed7d418ed3778d5cf92c8a29d82dbda12b8417cfdbd092bc9158ab7222 network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-pipelines-remote-resolvers-85f5bb4dfb-bkhrq 7ad415ed7d418ed3778d5cf92c8a29d82dbda12b8417cfdbd092bc9158ab7222 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"da:c5:61:27:dc:5c\",\"name\":\"7ad415ed7d418ed\"},{\"mac\":\"0a:58:0a:81:00:1e\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/0d258882-5b6f-4688-9d6c-ce1a67774300\"}],\"ips\":[{\"address\":\"10.129.0.30/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:17:12.458680 3780 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0514 16:17:12.458698 3780 util.go:870] Failed to build global endpoints for port TCP/probes: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0514 16:17:12.458703 3780 util.go:870] Failed to build global endpoints for port TCP/https-webhook: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0514 16:17:12.458709 3780 util.go:870] Failed to build global endpoints for port TCP/http-profiling: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0514 16:17:14.014344 3780 util.go:870] Failed to build global endpoints for port TCP/probes: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-controller W0514 16:17:14.014363 3780 util.go:870] Failed to build global endpoints for port TCP/http-profiling: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-controller W0514 16:17:14.014368 3780 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-controller W0514 16:17:14.043460 3780 util.go:870] Failed to build global endpoints for port TCP/probes: empty IP address endpoints for service openshift-pipelines/tekton-events-controller W0514 16:17:14.043480 3780 util.go:870] Failed to build global endpoints for port TCP/http-profiling: empty IP address endpoints for service openshift-pipelines/tekton-events-controller W0514 16:17:14.043486 3780 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/tekton-events-controller W0514 16:17:17.026370 3780 util.go:870] Failed to build global endpoints for port TCP/data: empty IP address endpoints for service tsf-keycloak/keycloak-pgsql W0514 16:17:19.044644 3780 util.go:870] Failed to build global endpoints for port TCP/data: empty IP address endpoints for service tsf-tpa/tpa-pgsql I0514 16:17:34.983844 3780 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s] creating logical port openshift-pipelines_tekton-triggers-controller-785c845664-fpk4s for pod on switch ip-10-0-0-158.ec2.internal I0514 16:17:34.983934 3780 kube.go:256] Updating pod openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s I0514 16:17:35.000955 3780 pod.go:62] [openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s] pod update took 17.04647ms I0514 16:17:35.000976 3780 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s] addLogicalPort annotation time took 17.072925ms I0514 16:17:35.005252 3780 pods.go:271] [openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s] addLogicalPort took 21.424319ms, libovsdb time 3.865884ms I0514 16:17:35.079727 3780 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l] creating logical port openshift-pipelines_tekton-triggers-webhook-6b5b45569f-jck8l for pod on switch ip-10-0-0-158.ec2.internal I0514 16:17:35.079813 3780 kube.go:256] Updating pod openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l I0514 16:17:35.098563 3780 pod.go:62] [openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l] pod update took 18.771087ms I0514 16:17:35.098585 3780 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l] addLogicalPort annotation time took 18.798699ms I0514 16:17:35.099841 3780 pods.go:271] [openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l] addLogicalPort took 20.126668ms, libovsdb time 919.726µs I0514 16:17:35.346068 3780 cni.go:452] [openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s b022387f1a74f5cb46a19c743e81bd8e916eefeadbbff003943435bdd662d1bf network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s b022387f1a74f5cb46a19c743e81bd8e916eefeadbbff003943435bdd662d1bf network default NAD default NAD key ] I0514 16:17:35.355939 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-triggers-controller-785c845664-fpk4s, hostIfaceName: b022387f1a74f5c, network: default, NAD default, SandboxID: "b022387f1a74f5cb46a19c743e81bd8e916eefeadbbff003943435bdd662d1bf", PCI device ID: , UID: "cb416d9d-2b3f-4c72-986f-e1dad0f9fca5", MAC: 0a:58:0a:81:00:1f, IPs: [10.129.0.31/23] I0514 16:17:35.407399 3780 cni.go:473] [openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s b022387f1a74f5cb46a19c743e81bd8e916eefeadbbff003943435bdd662d1bf network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-triggers-controller-785c845664-fpk4s b022387f1a74f5cb46a19c743e81bd8e916eefeadbbff003943435bdd662d1bf network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"a2:77:72:d8:db:c8\",\"name\":\"b022387f1a74f5c\"},{\"mac\":\"0a:58:0a:81:00:1f\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/f23d8399-98c2-4036-9b94-dbad5d85c2c4\"}],\"ips\":[{\"address\":\"10.129.0.31/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:17:35.440095 3780 cni.go:452] [openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l 91a572b7cb1413710ba451a55d889fd8da69f88fc5d693ac951f4449fcf30812 network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l 91a572b7cb1413710ba451a55d889fd8da69f88fc5d693ac951f4449fcf30812 network default NAD default NAD key ] I0514 16:17:35.450022 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-triggers-webhook-6b5b45569f-jck8l, hostIfaceName: 91a572b7cb14137, network: default, NAD default, SandboxID: "91a572b7cb1413710ba451a55d889fd8da69f88fc5d693ac951f4449fcf30812", PCI device ID: , UID: "dff17eeb-c4a2-490a-ada9-21bd8334fc4e", MAC: 0a:58:0a:81:00:20, IPs: [10.129.0.32/23] I0514 16:17:35.496960 3780 cni.go:473] [openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l 91a572b7cb1413710ba451a55d889fd8da69f88fc5d693ac951f4449fcf30812 network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-triggers-webhook-6b5b45569f-jck8l 91a572b7cb1413710ba451a55d889fd8da69f88fc5d693ac951f4449fcf30812 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"fa:2b:af:6a:fb:af\",\"name\":\"91a572b7cb14137\"},{\"mac\":\"0a:58:0a:81:00:20\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/c239746a-e102-479e-b947-ca71e51dac10\"}],\"ips\":[{\"address\":\"10.129.0.32/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:17:39.577337 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-pipelines/tekton-triggers-core-interceptors I0514 16:17:42.312761 3780 obj_retry.go:498] Detected object tsf/test-tsf-infrastructure of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:17:42.313262 3780 obj_retry.go:498] Detected object tsf/test-tsf-infrastructure of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:17:42.333315 3780 obj_retry.go:492] Detected object tsf/test-tsf-infrastructure of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:17:42.333335 3780 obj_retry.go:492] Detected object tsf/test-tsf-infrastructure of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:17:42.339703 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf/test-tsf-infrastructure I0514 16:17:42.339725 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf/test-tsf-infrastructure I0514 16:17:42.409982 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=762 I0514 16:17:44.250848 3780 base_network_controller_policy.go:1179] Adding network policy tsf-keycloak/keycloak-network-policy for network default I0514 16:17:44.251731 3780 pod_selector_address_set.go:213] Created shared address set for pod selector tsf-keycloak_LS{ML:{app: keycloak,app.kubernetes.io/instance: keycloak,app.kubernetes.io/managed-by: keycloak-operator,},} I0514 16:17:44.251755 3780 base_network_controller_policy.go:1054] Policy tsf-keycloak/keycloak-network-policy added to peer address sets [tsf-keycloak_LS{ML:{app: keycloak,app.kubernetes.io/instance: keycloak,app.kubernetes.io/managed-by: keycloak-operator,},}] I0514 16:17:44.254116 3780 base_network_controller_policy.go:1232] Create network policy tsf-keycloak/keycloak-network-policy resources completed, update namespace loglevel I0514 16:17:44.497743 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=10 W0514 16:17:51.238481 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-keycloak/keycloak W0514 16:17:51.238499 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-keycloak/keycloak W0514 16:17:51.238592 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-keycloak/keycloak-service W0514 16:17:51.238604 3780 util.go:870] Failed to build global endpoints for port TCP/management: empty IP address endpoints for service tsf-keycloak/keycloak-service I0514 16:17:57.681704 3780 iptables.go:108] Creating table: nat chain: OVN-KUBE-NODEPORT I0514 16:18:00.398542 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=113 I0514 16:18:00.761315 3780 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-results-postgres-0] creating logical port openshift-pipelines_tekton-results-postgres-0 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:18:00.761431 3780 kube.go:256] Updating pod openshift-pipelines/tekton-results-postgres-0 I0514 16:18:00.774337 3780 pod.go:62] [openshift-pipelines/tekton-results-postgres-0] pod update took 12.953245ms I0514 16:18:00.774358 3780 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-results-postgres-0] addLogicalPort annotation time took 12.977993ms I0514 16:18:00.775696 3780 pods.go:271] [openshift-pipelines/tekton-results-postgres-0] addLogicalPort took 14.398988ms, libovsdb time 905.729µs I0514 16:18:05.525495 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=10 I0514 16:18:10.733613 3780 cni.go:452] [openshift-pipelines/tekton-results-postgres-0 5da8826dc68ed8563e5cd585803392ab098aa838508ee5366819123bc57011ae network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-results-postgres-0 5da8826dc68ed8563e5cd585803392ab098aa838508ee5366819123bc57011ae network default NAD default NAD key ] I0514 16:18:10.742810 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-results-postgres-0, hostIfaceName: 5da8826dc68ed85, network: default, NAD default, SandboxID: "5da8826dc68ed8563e5cd585803392ab098aa838508ee5366819123bc57011ae", PCI device ID: , UID: "e2499280-6a96-466e-b7bc-197bbabf6976", MAC: 0a:58:0a:81:00:21, IPs: [10.129.0.33/23] I0514 16:18:10.792225 3780 cni.go:473] [openshift-pipelines/tekton-results-postgres-0 5da8826dc68ed8563e5cd585803392ab098aa838508ee5366819123bc57011ae network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-results-postgres-0 5da8826dc68ed8563e5cd585803392ab098aa838508ee5366819123bc57011ae network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"c6:ad:e2:44:8f:f0\",\"name\":\"5da8826dc68ed85\"},{\"mac\":\"0a:58:0a:81:00:21\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/41303de8-f25a-43b2-8f83-fbc0e7675e18\"}],\"ips\":[{\"address\":\"10.129.0.33/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:18:14.399189 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=96 I0514 16:18:14.774991 3780 namespace.go:142] [openshift-pipelines] updating namespace I0514 16:18:16.654360 3780 iptables.go:108] Creating table: nat chain: OVN-KUBE-NODEPORT I0514 16:18:18.084440 3780 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56] creating logical port openshift-pipelines_tekton-results-retention-policy-agent-585694b54-2lv56 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:18:18.084534 3780 kube.go:256] Updating pod openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56 I0514 16:18:18.099365 3780 pod.go:62] [openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56] pod update took 14.851914ms I0514 16:18:18.099387 3780 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56] addLogicalPort annotation time took 14.878ms I0514 16:18:18.100595 3780 pods.go:271] [openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56] addLogicalPort took 16.191299ms, libovsdb time 818.042µs I0514 16:18:18.442420 3780 cni.go:452] [openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56 5b6d3511dff3275016c3f2de80d75bc84e5931e5301802cdc38daaf777b948ed network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56 5b6d3511dff3275016c3f2de80d75bc84e5931e5301802cdc38daaf777b948ed network default NAD default NAD key ] I0514 16:18:18.451634 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-results-retention-policy-agent-585694b54-2lv56, hostIfaceName: 5b6d3511dff3275, network: default, NAD default, SandboxID: "5b6d3511dff3275016c3f2de80d75bc84e5931e5301802cdc38daaf777b948ed", PCI device ID: , UID: "ec00f45b-a954-4fbf-80d1-312738cea3f3", MAC: 0a:58:0a:81:00:22, IPs: [10.129.0.34/23] I0514 16:18:18.501348 3780 cni.go:473] [openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56 5b6d3511dff3275016c3f2de80d75bc84e5931e5301802cdc38daaf777b948ed network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-results-retention-policy-agent-585694b54-2lv56 5b6d3511dff3275016c3f2de80d75bc84e5931e5301802cdc38daaf777b948ed network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ba:39:99:81:dc:11\",\"name\":\"5b6d3511dff3275\"},{\"mac\":\"0a:58:0a:81:00:22\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/a843cf63-05a1-4a4f-8f35-c91a558ccbf3\"}],\"ips\":[{\"address\":\"10.129.0.34/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:18:18.502987 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=11 W0514 16:18:23.790863 3780 util.go:870] Failed to build global endpoints for port TCP/prometheus: empty IP address endpoints for service openshift-pipelines/tekton-results-api-service W0514 16:18:23.790884 3780 util.go:870] Failed to build global endpoints for port TCP/profiling: empty IP address endpoints for service openshift-pipelines/tekton-results-api-service W0514 16:18:23.790889 3780 util.go:870] Failed to build global endpoints for port TCP/server: empty IP address endpoints for service openshift-pipelines/tekton-results-api-service I0514 16:18:24.492441 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=10 I0514 16:18:34.660861 3780 obj_retry.go:498] Detected object tsf/test-tsf-iam of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:18:34.661371 3780 obj_retry.go:498] Detected object tsf/test-tsf-iam of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:18:34.702270 3780 obj_retry.go:492] Detected object tsf/test-tsf-iam of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:18:34.702291 3780 obj_retry.go:492] Detected object tsf/test-tsf-iam of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:18:34.708352 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf/test-tsf-iam I0514 16:18:34.708370 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf/test-tsf-iam I0514 16:18:36.901964 3780 base_network_controller_pods.go:487] [default/tsf-tas/segment-backup-installation-zxmkm-l6rwt] creating logical port tsf-tas_segment-backup-installation-zxmkm-l6rwt for pod on switch ip-10-0-0-158.ec2.internal I0514 16:18:36.902051 3780 kube.go:256] Updating pod tsf-tas/segment-backup-installation-zxmkm-l6rwt I0514 16:18:36.917043 3780 pod.go:62] [tsf-tas/segment-backup-installation-zxmkm-l6rwt] pod update took 15.015105ms I0514 16:18:36.917067 3780 base_network_controller_pods.go:951] [default/tsf-tas/segment-backup-installation-zxmkm-l6rwt] addLogicalPort annotation time took 15.043429ms I0514 16:18:36.918323 3780 pods.go:271] [tsf-tas/segment-backup-installation-zxmkm-l6rwt] addLogicalPort took 16.374063ms, libovsdb time 849.474µs I0514 16:18:37.262347 3780 cni.go:452] [tsf-tas/segment-backup-installation-zxmkm-l6rwt 69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660 network default NAD default NAD key ] ADD starting CNI request [tsf-tas/segment-backup-installation-zxmkm-l6rwt 69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660 network default NAD default NAD key ] I0514 16:18:37.272276 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: segment-backup-installation-zxmkm-l6rwt, hostIfaceName: 69bb605749ce89b, network: default, NAD default, SandboxID: "69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660", PCI device ID: , UID: "8d6d2b30-de6d-49e2-aecb-487fa08641eb", MAC: 0a:58:0a:81:00:23, IPs: [10.129.0.35/23] I0514 16:18:37.325031 3780 cni.go:473] [tsf-tas/segment-backup-installation-zxmkm-l6rwt 69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660 network default NAD default NAD key default] ADD finished CNI request [tsf-tas/segment-backup-installation-zxmkm-l6rwt 69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"f2:82:d5:d6:d1:66\",\"name\":\"69bb605749ce89b\"},{\"mac\":\"0a:58:0a:81:00:23\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/1066f3dd-ba56-48a3-b20b-e672d64ae745\"}],\"ips\":[{\"address\":\"10.129.0.35/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:18:37.455592 3780 base_network_controller_pods.go:487] [default/tsf-tas/fulcio-server-79bd59c648-h5xkz] creating logical port tsf-tas_fulcio-server-79bd59c648-h5xkz for pod on switch ip-10-0-0-158.ec2.internal I0514 16:18:37.455698 3780 kube.go:256] Updating pod tsf-tas/fulcio-server-79bd59c648-h5xkz I0514 16:18:37.472643 3780 pod.go:62] [tsf-tas/fulcio-server-79bd59c648-h5xkz] pod update took 16.970618ms I0514 16:18:37.472667 3780 base_network_controller_pods.go:951] [default/tsf-tas/fulcio-server-79bd59c648-h5xkz] addLogicalPort annotation time took 16.999796ms I0514 16:18:37.474193 3780 pods.go:271] [tsf-tas/fulcio-server-79bd59c648-h5xkz] addLogicalPort took 18.620098ms, libovsdb time 917.486µs I0514 16:18:37.784326 3780 base_network_controller_pods.go:487] [default/tsf-tas/ctlog-createtree-job-whldx-5tmb7] creating logical port tsf-tas_ctlog-createtree-job-whldx-5tmb7 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:18:37.784441 3780 kube.go:256] Updating pod tsf-tas/ctlog-createtree-job-whldx-5tmb7 I0514 16:18:37.805391 3780 pod.go:62] [tsf-tas/ctlog-createtree-job-whldx-5tmb7] pod update took 20.98461ms I0514 16:18:37.805446 3780 base_network_controller_pods.go:951] [default/tsf-tas/ctlog-createtree-job-whldx-5tmb7] addLogicalPort annotation time took 21.044114ms I0514 16:18:37.806909 3780 pods.go:271] [tsf-tas/ctlog-createtree-job-whldx-5tmb7] addLogicalPort took 22.604537ms, libovsdb time 942.169µs I0514 16:18:37.820060 3780 cni.go:452] [tsf-tas/fulcio-server-79bd59c648-h5xkz 146bbd8ebdfe04689de77ed20942ec44fb3906bda5e32950f38e4a3fcf8c2789 network default NAD default NAD key ] ADD starting CNI request [tsf-tas/fulcio-server-79bd59c648-h5xkz 146bbd8ebdfe04689de77ed20942ec44fb3906bda5e32950f38e4a3fcf8c2789 network default NAD default NAD key ] I0514 16:18:37.831298 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: fulcio-server-79bd59c648-h5xkz, hostIfaceName: 146bbd8ebdfe046, network: default, NAD default, SandboxID: "146bbd8ebdfe04689de77ed20942ec44fb3906bda5e32950f38e4a3fcf8c2789", PCI device ID: , UID: "f4a24b2e-052a-4cfe-9fe4-f289385a0c84", MAC: 0a:58:0a:81:00:24, IPs: [10.129.0.36/23] I0514 16:18:37.890229 3780 cni.go:473] [tsf-tas/fulcio-server-79bd59c648-h5xkz 146bbd8ebdfe04689de77ed20942ec44fb3906bda5e32950f38e4a3fcf8c2789 network default NAD default NAD key default] ADD finished CNI request [tsf-tas/fulcio-server-79bd59c648-h5xkz 146bbd8ebdfe04689de77ed20942ec44fb3906bda5e32950f38e4a3fcf8c2789 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"d2:4b:88:a1:11:fc\",\"name\":\"146bbd8ebdfe046\"},{\"mac\":\"0a:58:0a:81:00:24\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/8b1ee9e2-c144-4e21-95d3-5bcec9e642a5\"}],\"ips\":[{\"address\":\"10.129.0.36/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:18:38.146433 3780 cni.go:452] [tsf-tas/ctlog-createtree-job-whldx-5tmb7 6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c network default NAD default NAD key ] ADD starting CNI request [tsf-tas/ctlog-createtree-job-whldx-5tmb7 6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c network default NAD default NAD key ] I0514 16:18:38.165086 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: ctlog-createtree-job-whldx-5tmb7, hostIfaceName: 6e09d1c18a488ff, network: default, NAD default, SandboxID: "6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c", PCI device ID: , UID: "43984060-d3a9-4011-8541-ddcbd53e7e80", MAC: 0a:58:0a:81:00:25, IPs: [10.129.0.37/23] I0514 16:18:38.225236 3780 cni.go:473] [tsf-tas/ctlog-createtree-job-whldx-5tmb7 6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c network default NAD default NAD key default] ADD finished CNI request [tsf-tas/ctlog-createtree-job-whldx-5tmb7 6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"f6:ed:9d:9b:43:67\",\"name\":\"6e09d1c18a488ff\"},{\"mac\":\"0a:58:0a:81:00:25\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/eb1de965-aac2-4988-a539-553e399af24d\"}],\"ips\":[{\"address\":\"10.129.0.37/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:18:39.741284 3780 util.go:870] Failed to build global endpoints for port TCP/grpc: empty IP address endpoints for service tsf-tas/fulcio-server W0514 16:18:39.741311 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/fulcio-server W0514 16:18:39.770599 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 5554: empty IP address endpoints for service tsf-tas/fulcio-server W0514 16:18:39.770624 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 5555: empty IP address endpoints for service tsf-tas/fulcio-server I0514 16:18:46.541351 3780 obj_retry.go:498] Detected object tsf-keycloak/tsf-iam-lnsvg of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:18:46.541966 3780 obj_retry.go:498] Detected object tsf-keycloak/tsf-iam-lnsvg of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:18:47.568285 3780 obj_retry.go:492] Detected object tsf-keycloak/tsf-iam-lnsvg of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:18:47.568309 3780 obj_retry.go:492] Detected object tsf-keycloak/tsf-iam-lnsvg of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed W0514 16:18:47.611172 3780 util.go:870] Failed to build global endpoints for port TCP/grpc: empty IP address endpoints for service tsf-tas/trillian-logsigner I0514 16:18:50.799568 3780 cni.go:452] [tsf-tas/segment-backup-installation-zxmkm-l6rwt 69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660 network default NAD default NAD key ] DEL starting CNI request [tsf-tas/segment-backup-installation-zxmkm-l6rwt 69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660 network default NAD default NAD key ] I0514 16:18:50.892423 3780 cni.go:473] [tsf-tas/segment-backup-installation-zxmkm-l6rwt 69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660 network default NAD default NAD key default] DEL finished CNI request [tsf-tas/segment-backup-installation-zxmkm-l6rwt 69bb605749ce89b7dc1c2a5cd830d99eaed77f806b431afe6fc59bb2ae27f660 network default NAD default NAD key default], result "{}", err W0514 16:18:50.930297 3780 util.go:870] Failed to build global endpoints for port TCP/grpc: empty IP address endpoints for service tsf-tas/trillian-logserver I0514 16:18:50.951253 3780 obj_retry.go:498] Detected object tsf-tas/segment-backup-installation-zxmkm-l6rwt of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:18:50.951290 3780 pods.go:176] Deleting pod: tsf-tas/segment-backup-installation-zxmkm-l6rwt I0514 16:18:50.952532 3780 pods.go:236] Attempting to release IPs for pod: tsf-tas/segment-backup-installation-zxmkm-l6rwt, ips: 10.129.0.35 I0514 16:18:50.952557 3780 obj_retry.go:498] Detected object tsf-tas/segment-backup-installation-zxmkm-l6rwt of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:18:51.989635 3780 obj_retry.go:492] Detected object tsf-tas/segment-backup-installation-zxmkm-l6rwt of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:18:51.989658 3780 obj_retry.go:492] Detected object tsf-tas/segment-backup-installation-zxmkm-l6rwt of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed W0514 16:18:53.947498 3780 util.go:870] Failed to build global endpoints for port TCP/trillian-mysql: empty IP address endpoints for service tsf-tas/trillian-mysql I0514 16:18:55.436622 3780 base_network_controller_pods.go:487] [default/openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5] creating logical port openshift-pipelines_pipelines-as-code-controller-79b6fbbc5c-gn6x5 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:18:55.436714 3780 kube.go:256] Updating pod openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5 I0514 16:18:55.455257 3780 pod.go:62] [openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5] pod update took 18.568162ms I0514 16:18:55.455278 3780 base_network_controller_pods.go:951] [default/openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5] addLogicalPort annotation time took 18.594277ms I0514 16:18:55.456595 3780 pods.go:271] [openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5] addLogicalPort took 20.008656ms, libovsdb time 937.133µs I0514 16:18:55.802855 3780 cni.go:452] [openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5 c3ac4d7bd9ec63a76a64daf0b569f3259ae0345ce3050e8339c2138aec7713ce network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5 c3ac4d7bd9ec63a76a64daf0b569f3259ae0345ce3050e8339c2138aec7713ce network default NAD default NAD key ] I0514 16:18:55.812028 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: pipelines-as-code-controller-79b6fbbc5c-gn6x5, hostIfaceName: c3ac4d7bd9ec63a, network: default, NAD default, SandboxID: "c3ac4d7bd9ec63a76a64daf0b569f3259ae0345ce3050e8339c2138aec7713ce", PCI device ID: , UID: "ccfbd2c7-ff5a-45f9-b16f-659003276b54", MAC: 0a:58:0a:81:00:26, IPs: [10.129.0.38/23] I0514 16:18:55.863751 3780 cni.go:473] [openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5 c3ac4d7bd9ec63a76a64daf0b569f3259ae0345ce3050e8339c2138aec7713ce network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/pipelines-as-code-controller-79b6fbbc5c-gn6x5 c3ac4d7bd9ec63a76a64daf0b569f3259ae0345ce3050e8339c2138aec7713ce network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ae:4f:22:34:5d:91\",\"name\":\"c3ac4d7bd9ec63a\"},{\"mac\":\"0a:58:0a:81:00:26\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/55d101f2-4680-4fd4-9591-3cd8c13d9eb1\"}],\"ips\":[{\"address\":\"10.129.0.38/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:18:58.842140 3780 util.go:870] Failed to build global endpoints for port TCP/http-listener: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-controller W0514 16:18:58.842156 3780 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-controller W0514 16:18:58.869485 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8082: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-controller W0514 16:18:58.869502 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 9090: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-controller W0514 16:19:00.700551 3780 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-watcher I0514 16:19:25.983832 3780 obj_retry.go:498] Detected object tsf-tas/test-tsf-tas of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:25.984362 3780 obj_retry.go:498] Detected object tsf-tas/test-tsf-tas of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:26.006786 3780 obj_retry.go:492] Detected object tsf-tas/test-tsf-tas of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:26.006805 3780 obj_retry.go:492] Detected object tsf-tas/test-tsf-tas of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:26.012878 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tas/test-tsf-tas I0514 16:19:26.012902 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tas/test-tsf-tas I0514 16:19:31.019269 3780 obj_retry.go:498] Detected object tsf-tas/rekor-createtree-job-z85wt-9qzhj of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:31.019850 3780 obj_retry.go:498] Detected object tsf-tas/rekor-createtree-job-z85wt-9qzhj of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:31.895792 3780 obj_retry.go:492] Detected object tsf-tas/rekor-createtree-job-z85wt-9qzhj of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:31.895816 3780 obj_retry.go:492] Detected object tsf-tas/rekor-createtree-job-z85wt-9qzhj of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:34.355685 3780 base_network_controller_pods.go:487] [default/tsf-tas/rekor-search-ui-68b96df54c-mf6tr] creating logical port tsf-tas_rekor-search-ui-68b96df54c-mf6tr for pod on switch ip-10-0-0-158.ec2.internal I0514 16:19:34.355805 3780 kube.go:256] Updating pod tsf-tas/rekor-search-ui-68b96df54c-mf6tr I0514 16:19:34.369538 3780 pod.go:62] [tsf-tas/rekor-search-ui-68b96df54c-mf6tr] pod update took 13.77333ms I0514 16:19:34.369624 3780 base_network_controller_pods.go:951] [default/tsf-tas/rekor-search-ui-68b96df54c-mf6tr] addLogicalPort annotation time took 13.861873ms I0514 16:19:34.371039 3780 pods.go:271] [tsf-tas/rekor-search-ui-68b96df54c-mf6tr] addLogicalPort took 15.373853ms, libovsdb time 904.98µs I0514 16:19:34.715817 3780 cni.go:452] [tsf-tas/rekor-search-ui-68b96df54c-mf6tr fe61767cbb2d1a64b8b9f0dc3501f17307606405d0b260f4cd8be93742a18a6c network default NAD default NAD key ] ADD starting CNI request [tsf-tas/rekor-search-ui-68b96df54c-mf6tr fe61767cbb2d1a64b8b9f0dc3501f17307606405d0b260f4cd8be93742a18a6c network default NAD default NAD key ] I0514 16:19:34.725247 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: rekor-search-ui-68b96df54c-mf6tr, hostIfaceName: fe61767cbb2d1a6, network: default, NAD default, SandboxID: "fe61767cbb2d1a64b8b9f0dc3501f17307606405d0b260f4cd8be93742a18a6c", PCI device ID: , UID: "a8fb1753-bc30-4077-bdf3-2c24f512b0b9", MAC: 0a:58:0a:81:00:27, IPs: [10.129.0.39/23] I0514 16:19:34.776584 3780 cni.go:473] [tsf-tas/rekor-search-ui-68b96df54c-mf6tr fe61767cbb2d1a64b8b9f0dc3501f17307606405d0b260f4cd8be93742a18a6c network default NAD default NAD key default] ADD finished CNI request [tsf-tas/rekor-search-ui-68b96df54c-mf6tr fe61767cbb2d1a64b8b9f0dc3501f17307606405d0b260f4cd8be93742a18a6c network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"c2:f5:b7:a3:67:94\",\"name\":\"fe61767cbb2d1a6\"},{\"mac\":\"0a:58:0a:81:00:27\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/0ff73571-8ea0-43a3-9165-f420215c250c\"}],\"ips\":[{\"address\":\"10.129.0.39/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:19:35.032204 3780 cni.go:452] [tsf-tas/ctlog-createtree-job-whldx-5tmb7 6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c network default NAD default NAD key ] DEL starting CNI request [tsf-tas/ctlog-createtree-job-whldx-5tmb7 6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c network default NAD default NAD key ] I0514 16:19:35.108735 3780 cni.go:473] [tsf-tas/ctlog-createtree-job-whldx-5tmb7 6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c network default NAD default NAD key default] DEL finished CNI request [tsf-tas/ctlog-createtree-job-whldx-5tmb7 6e09d1c18a488ffe0c804c233ddd745c0aea6f01a3cd5445bc50002ce1fe2d9c network default NAD default NAD key default], result "{}", err I0514 16:19:35.153499 3780 obj_retry.go:498] Detected object tsf-tas/ctlog-createtree-job-whldx-5tmb7 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:35.153533 3780 pods.go:176] Deleting pod: tsf-tas/ctlog-createtree-job-whldx-5tmb7 I0514 16:19:35.154969 3780 pods.go:236] Attempting to release IPs for pod: tsf-tas/ctlog-createtree-job-whldx-5tmb7, ips: 10.129.0.37 I0514 16:19:35.154999 3780 obj_retry.go:498] Detected object tsf-tas/ctlog-createtree-job-whldx-5tmb7 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:35.294120 3780 obj_retry.go:498] Detected object tsf/patch-tekton-config-5hprk of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:35.294688 3780 obj_retry.go:498] Detected object tsf/patch-tekton-config-5hprk of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:36.050643 3780 obj_retry.go:492] Detected object tsf-tas/ctlog-createtree-job-whldx-5tmb7 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:36.050667 3780 obj_retry.go:492] Detected object tsf-tas/ctlog-createtree-job-whldx-5tmb7 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:36.317448 3780 obj_retry.go:492] Detected object tsf/patch-tekton-config-5hprk of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:36.317468 3780 obj_retry.go:492] Detected object tsf/patch-tekton-config-5hprk of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:36.863377 3780 base_network_controller_pods.go:487] [default/tsf-tas/rekor-server-5849bd9599-jvnjf] creating logical port tsf-tas_rekor-server-5849bd9599-jvnjf for pod on switch ip-10-0-0-158.ec2.internal I0514 16:19:36.863517 3780 kube.go:256] Updating pod tsf-tas/rekor-server-5849bd9599-jvnjf I0514 16:19:36.878349 3780 pod.go:62] [tsf-tas/rekor-server-5849bd9599-jvnjf] pod update took 14.866616ms I0514 16:19:36.878375 3780 base_network_controller_pods.go:951] [default/tsf-tas/rekor-server-5849bd9599-jvnjf] addLogicalPort annotation time took 14.898671ms I0514 16:19:36.879744 3780 pods.go:271] [tsf-tas/rekor-server-5849bd9599-jvnjf] addLogicalPort took 16.389087ms, libovsdb time 939.926µs W0514 16:19:39.057865 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/rekor-search-ui W0514 16:19:39.934462 3780 util.go:870] Failed to build global endpoints for port TCP/resp: empty IP address endpoints for service tsf-tas/rekor-redis W0514 16:19:41.952194 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/ctlog I0514 16:19:42.623525 3780 cni.go:452] [tsf-tas/rekor-server-5849bd9599-jvnjf d53152978c1f9f96d3b9bb3e15a89187c034e0472b03751dd2aebd95eb38db4e network default NAD default NAD key ] ADD starting CNI request [tsf-tas/rekor-server-5849bd9599-jvnjf d53152978c1f9f96d3b9bb3e15a89187c034e0472b03751dd2aebd95eb38db4e network default NAD default NAD key ] I0514 16:19:42.632865 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: rekor-server-5849bd9599-jvnjf, hostIfaceName: d53152978c1f9f9, network: default, NAD default, SandboxID: "d53152978c1f9f96d3b9bb3e15a89187c034e0472b03751dd2aebd95eb38db4e", PCI device ID: , UID: "66647c43-caea-4ead-acb3-6eceb9cbad12", MAC: 0a:58:0a:81:00:28, IPs: [10.129.0.40/23] I0514 16:19:42.684396 3780 cni.go:473] [tsf-tas/rekor-server-5849bd9599-jvnjf d53152978c1f9f96d3b9bb3e15a89187c034e0472b03751dd2aebd95eb38db4e network default NAD default NAD key default] ADD finished CNI request [tsf-tas/rekor-server-5849bd9599-jvnjf d53152978c1f9f96d3b9bb3e15a89187c034e0472b03751dd2aebd95eb38db4e network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"b6:8f:c9:cc:19:bb\",\"name\":\"d53152978c1f9f9\"},{\"mac\":\"0a:58:0a:81:00:28\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/029522a6-42cd-4e34-b480-d5f6d8f7cf4e\"}],\"ips\":[{\"address\":\"10.129.0.40/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:19:45.090425 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/rekor-server I0514 16:19:47.523249 3780 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=6 W0514 16:19:51.053957 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 3000: empty IP address endpoints for service tsf-tas/rekor-search-ui I0514 16:19:54.385285 3780 obj_retry.go:498] Detected object tsf/tsf-tekton-configuration-m58zp of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:54.385768 3780 obj_retry.go:498] Detected object tsf/tsf-tekton-configuration-m58zp of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:19:55.272937 3780 obj_retry.go:492] Detected object tsf/tsf-tekton-configuration-m58zp of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:55.272964 3780 obj_retry.go:492] Detected object tsf/tsf-tekton-configuration-m58zp of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:55.312239 3780 obj_retry.go:492] Detected object tsf/patch-tekton-config-5hprk of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:55.312258 3780 obj_retry.go:492] Detected object tsf/patch-tekton-config-5hprk of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:19:55.318479 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf/patch-tekton-config-5hprk I0514 16:19:55.318502 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf/patch-tekton-config-5hprk W0514 16:19:56.117543 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 3000: empty IP address endpoints for service tsf-tas/rekor-server I0514 16:20:00.171901 3780 base_network_controller_pods.go:487] [default/openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b] creating logical port openshift-backplane_osd-delete-backplane-serviceaccounts-29646260-c4c5b for pod on switch ip-10-0-0-158.ec2.internal I0514 16:20:00.171992 3780 kube.go:256] Updating pod openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b I0514 16:20:00.188065 3780 pod.go:62] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b] pod update took 16.099743ms I0514 16:20:00.188085 3780 base_network_controller_pods.go:951] [default/openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b] addLogicalPort annotation time took 16.124556ms I0514 16:20:00.189385 3780 pods.go:271] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b] addLogicalPort took 17.498752ms, libovsdb time 856.606µs I0514 16:20:00.211453 3780 obj_retry.go:498] Detected object tsf-tpa/pre-install-tsf-tpa of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:00.211913 3780 obj_retry.go:498] Detected object tsf-tpa/pre-install-tsf-tpa of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:00.233550 3780 obj_retry.go:492] Detected object tsf-tpa/pre-install-tsf-tpa of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:00.233567 3780 obj_retry.go:492] Detected object tsf-tpa/pre-install-tsf-tpa of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:00.243681 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/pre-install-tsf-tpa I0514 16:20:00.243701 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/pre-install-tsf-tpa I0514 16:20:00.530934 3780 cni.go:452] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b 877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077 network default NAD default NAD key ] ADD starting CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b 877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077 network default NAD default NAD key ] I0514 16:20:00.540039 3780 helper_linux.go:556] ConfigureOVS: namespace: openshift-backplane, podName: osd-delete-backplane-serviceaccounts-29646260-c4c5b, hostIfaceName: 877021f96eb130c, network: default, NAD default, SandboxID: "877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077", PCI device ID: , UID: "388d70a4-ad7d-4172-8a28-4fd7c3637f40", MAC: 0a:58:0a:81:00:29, IPs: [10.129.0.41/23] I0514 16:20:00.591208 3780 cni.go:473] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b 877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077 network default NAD default NAD key default] ADD finished CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b 877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"32:1c:75:42:ed:24\",\"name\":\"877021f96eb130c\"},{\"mac\":\"0a:58:0a:81:00:29\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/f43b196f-a7e1-4dc3-8ca2-318eb485975d\"}],\"ips\":[{\"address\":\"10.129.0.41/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:20:00.842334 3780 base_network_controller_pods.go:487] [default/tsf-tas/tuf-repository-init-7mq9q-777mt] creating logical port tsf-tas_tuf-repository-init-7mq9q-777mt for pod on switch ip-10-0-0-158.ec2.internal I0514 16:20:00.842448 3780 kube.go:256] Updating pod tsf-tas/tuf-repository-init-7mq9q-777mt I0514 16:20:00.857962 3780 pod.go:62] [tsf-tas/tuf-repository-init-7mq9q-777mt] pod update took 15.552504ms I0514 16:20:00.857983 3780 base_network_controller_pods.go:951] [default/tsf-tas/tuf-repository-init-7mq9q-777mt] addLogicalPort annotation time took 15.578922ms I0514 16:20:00.859302 3780 pods.go:271] [tsf-tas/tuf-repository-init-7mq9q-777mt] addLogicalPort took 16.985152ms, libovsdb time 857.93µs I0514 16:20:03.901356 3780 cni.go:452] [tsf-tas/tuf-repository-init-7mq9q-777mt c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36 network default NAD default NAD key ] ADD starting CNI request [tsf-tas/tuf-repository-init-7mq9q-777mt c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36 network default NAD default NAD key ] I0514 16:20:03.911372 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: tuf-repository-init-7mq9q-777mt, hostIfaceName: c5aeff2829580fb, network: default, NAD default, SandboxID: "c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36", PCI device ID: , UID: "427d571e-9ddb-48ed-8eff-538d7f68c470", MAC: 0a:58:0a:81:00:2a, IPs: [10.129.0.42/23] I0514 16:20:03.960767 3780 cni.go:473] [tsf-tas/tuf-repository-init-7mq9q-777mt c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36 network default NAD default NAD key default] ADD finished CNI request [tsf-tas/tuf-repository-init-7mq9q-777mt c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"26:83:24:d9:90:5a\",\"name\":\"c5aeff2829580fb\"},{\"mac\":\"0a:58:0a:81:00:2a\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/fb8a559f-b39b-4c37-b72e-ae69c1922589\"}],\"ips\":[{\"address\":\"10.129.0.42/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:20:04.193658 3780 cni.go:452] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b 877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077 network default NAD default NAD key ] DEL starting CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b 877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077 network default NAD default NAD key ] I0514 16:20:04.291236 3780 cni.go:473] [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b 877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077 network default NAD default NAD key default] DEL finished CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b 877021f96eb130cc44244a823b9f2be1427ba5f2d521b7aeadc68b4dda075077 network default NAD default NAD key default], result "{}", err I0514 16:20:04.343789 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:04.343823 3780 pods.go:176] Deleting pod: openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b I0514 16:20:04.345140 3780 pods.go:236] Attempting to release IPs for pod: openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b, ips: 10.129.0.41 I0514 16:20:04.345164 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:05.206255 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:05.206278 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:07.644658 3780 obj_retry.go:498] Detected object tsf-tpa/create-db-4wz6p of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:07.645252 3780 obj_retry.go:498] Detected object tsf-tpa/create-db-4wz6p of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:08.668473 3780 obj_retry.go:492] Detected object tsf-tpa/create-db-4wz6p of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:08.668498 3780 obj_retry.go:492] Detected object tsf-tpa/create-db-4wz6p of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:11.235270 3780 cni.go:452] [tsf-tas/tuf-repository-init-7mq9q-777mt c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36 network default NAD default NAD key ] DEL starting CNI request [tsf-tas/tuf-repository-init-7mq9q-777mt c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36 network default NAD default NAD key ] I0514 16:20:11.312069 3780 cni.go:473] [tsf-tas/tuf-repository-init-7mq9q-777mt c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36 network default NAD default NAD key default] DEL finished CNI request [tsf-tas/tuf-repository-init-7mq9q-777mt c5aeff2829580fbe3df447ec0bb761705800dca12880e55eb6154019b9643d36 network default NAD default NAD key default], result "{}", err I0514 16:20:11.373574 3780 obj_retry.go:498] Detected object tsf-tas/tuf-repository-init-7mq9q-777mt of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:11.373603 3780 pods.go:176] Deleting pod: tsf-tas/tuf-repository-init-7mq9q-777mt I0514 16:20:11.375108 3780 pods.go:236] Attempting to release IPs for pod: tsf-tas/tuf-repository-init-7mq9q-777mt, ips: 10.129.0.42 I0514 16:20:11.375137 3780 obj_retry.go:498] Detected object tsf-tas/tuf-repository-init-7mq9q-777mt of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:11.488046 3780 obj_retry.go:498] Detected object tsf-tpa/migrate-db-vxvdb of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:11.488514 3780 obj_retry.go:498] Detected object tsf-tpa/migrate-db-vxvdb of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:12.260744 3780 obj_retry.go:492] Detected object tsf-tas/tuf-repository-init-7mq9q-777mt of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:12.260770 3780 obj_retry.go:492] Detected object tsf-tas/tuf-repository-init-7mq9q-777mt of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:12.509375 3780 obj_retry.go:492] Detected object tsf-tpa/migrate-db-vxvdb of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:12.509398 3780 obj_retry.go:492] Detected object tsf-tpa/migrate-db-vxvdb of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:12.947041 3780 base_network_controller_pods.go:487] [default/tsf-tas/tuf-97fc8ff76-f6zd7] creating logical port tsf-tas_tuf-97fc8ff76-f6zd7 for pod on switch ip-10-0-0-158.ec2.internal I0514 16:20:12.947136 3780 kube.go:256] Updating pod tsf-tas/tuf-97fc8ff76-f6zd7 I0514 16:20:12.961242 3780 pod.go:62] [tsf-tas/tuf-97fc8ff76-f6zd7] pod update took 14.125531ms I0514 16:20:12.961269 3780 base_network_controller_pods.go:951] [default/tsf-tas/tuf-97fc8ff76-f6zd7] addLogicalPort annotation time took 14.156146ms I0514 16:20:12.962626 3780 pods.go:271] [tsf-tas/tuf-97fc8ff76-f6zd7] addLogicalPort took 15.60804ms, libovsdb time 926.73µs I0514 16:20:14.515093 3780 obj_retry.go:498] Detected object tsf-tpa/create-importers-rrp6f of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:14.515623 3780 obj_retry.go:498] Detected object tsf-tpa/create-importers-rrp6f of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:15.537558 3780 obj_retry.go:492] Detected object tsf-tpa/create-importers-rrp6f of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:15.537582 3780 obj_retry.go:492] Detected object tsf-tpa/create-importers-rrp6f of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:15.650567 3780 base_network_controller_pods.go:487] [default/tsf-tpa/server-7bb877f9-5l4sq] creating logical port tsf-tpa_server-7bb877f9-5l4sq for pod on switch ip-10-0-0-158.ec2.internal I0514 16:20:15.650698 3780 kube.go:256] Updating pod tsf-tpa/server-7bb877f9-5l4sq I0514 16:20:15.665763 3780 pod.go:62] [tsf-tpa/server-7bb877f9-5l4sq] pod update took 15.08806ms I0514 16:20:15.665787 3780 base_network_controller_pods.go:951] [default/tsf-tpa/server-7bb877f9-5l4sq] addLogicalPort annotation time took 15.116998ms I0514 16:20:15.667187 3780 pods.go:271] [tsf-tpa/server-7bb877f9-5l4sq] addLogicalPort took 16.636466ms, libovsdb time 1.014746ms I0514 16:20:17.023675 3780 base_network_controller_pods.go:487] [default/tsf-tpa/importer-86bcdf477b-8tnlj] creating logical port tsf-tpa_importer-86bcdf477b-8tnlj for pod on switch ip-10-0-0-158.ec2.internal I0514 16:20:17.023822 3780 kube.go:256] Updating pod tsf-tpa/importer-86bcdf477b-8tnlj I0514 16:20:17.040028 3780 pod.go:62] [tsf-tpa/importer-86bcdf477b-8tnlj] pod update took 16.237451ms I0514 16:20:17.040049 3780 base_network_controller_pods.go:951] [default/tsf-tpa/importer-86bcdf477b-8tnlj] addLogicalPort annotation time took 16.263833ms I0514 16:20:17.041458 3780 pods.go:271] [tsf-tpa/importer-86bcdf477b-8tnlj] addLogicalPort took 17.8006ms, libovsdb time 973.998µs I0514 16:20:22.909458 3780 cni.go:452] [tsf-tas/tuf-97fc8ff76-f6zd7 ca2ecc961d63fff8f6a4fb50ac4499702f4531c786e6cab09dd6bb44872d75cd network default NAD default NAD key ] ADD starting CNI request [tsf-tas/tuf-97fc8ff76-f6zd7 ca2ecc961d63fff8f6a4fb50ac4499702f4531c786e6cab09dd6bb44872d75cd network default NAD default NAD key ] I0514 16:20:22.918911 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: tuf-97fc8ff76-f6zd7, hostIfaceName: ca2ecc961d63fff, network: default, NAD default, SandboxID: "ca2ecc961d63fff8f6a4fb50ac4499702f4531c786e6cab09dd6bb44872d75cd", PCI device ID: , UID: "601a6af4-2954-4928-9672-32a9f7691a89", MAC: 0a:58:0a:81:00:2b, IPs: [10.129.0.43/23] I0514 16:20:22.971431 3780 cni.go:473] [tsf-tas/tuf-97fc8ff76-f6zd7 ca2ecc961d63fff8f6a4fb50ac4499702f4531c786e6cab09dd6bb44872d75cd network default NAD default NAD key default] ADD finished CNI request [tsf-tas/tuf-97fc8ff76-f6zd7 ca2ecc961d63fff8f6a4fb50ac4499702f4531c786e6cab09dd6bb44872d75cd network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"c2:9e:4c:c8:19:48\",\"name\":\"ca2ecc961d63fff\"},{\"mac\":\"0a:58:0a:81:00:2b\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/faaecfe2-377f-4e21-bfd1-614251d272dc\"}],\"ips\":[{\"address\":\"10.129.0.43/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:20:23.994393 3780 cni.go:452] [tsf-tpa/importer-86bcdf477b-8tnlj e752ba65eb30b6840fe173fa06de0f390de362f4deec2f319f7b8a1d91842e60 network default NAD default NAD key ] ADD starting CNI request [tsf-tpa/importer-86bcdf477b-8tnlj e752ba65eb30b6840fe173fa06de0f390de362f4deec2f319f7b8a1d91842e60 network default NAD default NAD key ] I0514 16:20:24.004239 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tpa, podName: importer-86bcdf477b-8tnlj, hostIfaceName: e752ba65eb30b68, network: default, NAD default, SandboxID: "e752ba65eb30b6840fe173fa06de0f390de362f4deec2f319f7b8a1d91842e60", PCI device ID: , UID: "6e6acd72-c5ac-4d59-ab8d-ec4ebc04a81a", MAC: 0a:58:0a:81:00:2d, IPs: [10.129.0.45/23] I0514 16:20:24.054314 3780 cni.go:473] [tsf-tpa/importer-86bcdf477b-8tnlj e752ba65eb30b6840fe173fa06de0f390de362f4deec2f319f7b8a1d91842e60 network default NAD default NAD key default] ADD finished CNI request [tsf-tpa/importer-86bcdf477b-8tnlj e752ba65eb30b6840fe173fa06de0f390de362f4deec2f319f7b8a1d91842e60 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"4a:c3:a3:68:50:5a\",\"name\":\"e752ba65eb30b68\"},{\"mac\":\"0a:58:0a:81:00:2d\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/02b56e9c-c648-4e23-9a48-78c33249d55b\"}],\"ips\":[{\"address\":\"10.129.0.45/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err I0514 16:20:24.117124 3780 cni.go:452] [tsf-tpa/server-7bb877f9-5l4sq 80879236a674ee169ab8dffa1144482389aed1a01b811b88f5f6092bb0284b51 network default NAD default NAD key ] ADD starting CNI request [tsf-tpa/server-7bb877f9-5l4sq 80879236a674ee169ab8dffa1144482389aed1a01b811b88f5f6092bb0284b51 network default NAD default NAD key ] I0514 16:20:24.126547 3780 helper_linux.go:556] ConfigureOVS: namespace: tsf-tpa, podName: server-7bb877f9-5l4sq, hostIfaceName: 80879236a674ee1, network: default, NAD default, SandboxID: "80879236a674ee169ab8dffa1144482389aed1a01b811b88f5f6092bb0284b51", PCI device ID: , UID: "0f4a6a73-8366-4857-b461-715ac5c1ba1a", MAC: 0a:58:0a:81:00:2c, IPs: [10.129.0.44/23] I0514 16:20:24.176186 3780 cni.go:473] [tsf-tpa/server-7bb877f9-5l4sq 80879236a674ee169ab8dffa1144482389aed1a01b811b88f5f6092bb0284b51 network default NAD default NAD key default] ADD finished CNI request [tsf-tpa/server-7bb877f9-5l4sq 80879236a674ee169ab8dffa1144482389aed1a01b811b88f5f6092bb0284b51 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"be:5d:e2:4e:82:4f\",\"name\":\"80879236a674ee1\"},{\"mac\":\"0a:58:0a:81:00:2c\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/80d366cf-9d85-4757-9052-2a4f7c5ab976\"}],\"ips\":[{\"address\":\"10.129.0.44/23\",\"gateway\":\"10.129.0.1\",\"interface\":1}]}", err W0514 16:20:25.316559 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/tuf W0514 16:20:26.320695 3780 util.go:870] Failed to build global endpoints for port TCP/endpoint: empty IP address endpoints for service tsf-tpa/server W0514 16:20:36.324936 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service tsf-tas/tuf W0514 16:20:37.361652 3780 util.go:959] Failed to build node endpoints for node ip-10-0-0-158.ec2.internal port 8080: empty IP address endpoints for service tsf-tpa/server I0514 16:20:54.530311 3780 obj_retry.go:498] Detected object tsf-tpa/test-tsf-tpa of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:54.530862 3780 obj_retry.go:498] Detected object tsf-tpa/test-tsf-tpa of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:20:54.577496 3780 obj_retry.go:492] Detected object tsf-tpa/test-tsf-tpa of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:54.577513 3780 obj_retry.go:492] Detected object tsf-tpa/test-tsf-tpa of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:20:54.585045 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/test-tsf-tpa I0514 16:20:54.585073 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/test-tsf-tpa I0514 16:20:55.396163 3780 namespace.go:98] [default-tenant] adding namespace I0514 16:20:55.397140 3780 namespace.go:102] [default-tenant] adding namespace took 954.416µs I0514 16:20:55.410181 3780 namespace.go:98] [integration-service] adding namespace I0514 16:20:55.410320 3780 namespace.go:98] [build-service] adding namespace I0514 16:20:55.411005 3780 namespace.go:102] [integration-service] adding namespace took 807.203µs I0514 16:20:55.411037 3780 namespace.go:98] [image-controller] adding namespace I0514 16:20:55.411627 3780 namespace.go:102] [image-controller] adding namespace took 582.492µs I0514 16:20:55.412258 3780 namespace.go:102] [build-service] adding namespace took 1.923974ms I0514 16:20:55.432798 3780 namespace.go:142] [default-tenant] updating namespace I0514 16:20:55.465607 3780 namespace.go:142] [default-tenant] updating namespace I0514 16:20:55.479996 3780 namespace.go:142] [integration-service] updating namespace I0514 16:20:55.504431 3780 namespace.go:142] [integration-service] updating namespace I0514 16:20:55.507225 3780 namespace.go:142] [image-controller] updating namespace I0514 16:20:55.522094 3780 namespace.go:142] [image-controller] updating namespace I0514 16:20:55.535271 3780 namespace.go:142] [build-service] updating namespace I0514 16:20:55.550311 3780 namespace.go:142] [build-service] updating namespace I0514 16:20:55.645259 3780 namespace.go:142] [build-service] updating namespace I0514 16:20:55.657034 3780 namespace.go:142] [integration-service] updating namespace I0514 16:20:55.704736 3780 namespace.go:98] [release-service] adding namespace I0514 16:20:55.705660 3780 namespace.go:102] [release-service] adding namespace took 908.835µs I0514 16:20:55.734779 3780 namespace.go:142] [konflux-ui] updating namespace I0514 16:20:55.755028 3780 namespace.go:142] [release-service] updating namespace I0514 16:20:55.800365 3780 namespace.go:142] [release-service] updating namespace I0514 16:20:55.820085 3780 namespace.go:98] [konflux-info] adding namespace I0514 16:20:55.820896 3780 namespace.go:102] [konflux-info] adding namespace took 796.275µs I0514 16:20:55.921966 3780 namespace.go:142] [konflux-info] updating namespace I0514 16:20:55.922121 3780 namespace.go:98] [namespace-lister] adding namespace I0514 16:20:55.922955 3780 namespace.go:102] [namespace-lister] adding namespace took 820.218µs I0514 16:20:55.962331 3780 namespace.go:98] [enterprise-contract-service] adding namespace I0514 16:20:55.963134 3780 namespace.go:102] [enterprise-contract-service] adding namespace took 784.516µs I0514 16:20:55.983328 3780 namespace.go:142] [image-controller] updating namespace I0514 16:20:55.983335 3780 namespace.go:142] [konflux-info] updating namespace I0514 16:20:55.986269 3780 namespace.go:142] [namespace-lister] updating namespace I0514 16:20:56.027385 3780 namespace.go:142] [namespace-lister] updating namespace I0514 16:20:56.047429 3780 namespace.go:142] [enterprise-contract-service] updating namespace I0514 16:20:56.137215 3780 namespace.go:142] [enterprise-contract-service] updating namespace I0514 16:20:56.148347 3780 namespace.go:142] [default-tenant] updating namespace I0514 16:20:56.237373 3780 namespace.go:98] [segment-bridge] adding namespace I0514 16:20:56.238261 3780 namespace.go:102] [segment-bridge] adding namespace took 871.235µs I0514 16:20:56.327370 3780 namespace.go:142] [segment-bridge] updating namespace I0514 16:20:56.339883 3780 namespace.go:98] [konflux-cli] adding namespace I0514 16:20:56.340645 3780 namespace.go:102] [konflux-cli] adding namespace took 746.898µs I0514 16:20:56.418304 3780 namespace.go:142] [konflux-cli] updating namespace I0514 16:20:56.445484 3780 namespace.go:142] [segment-bridge] updating namespace I0514 16:20:56.712252 3780 namespace.go:142] [konflux-cli] updating namespace I0514 16:20:57.015364 3780 base_network_controller_policy.go:1179] Adding network policy namespace-lister/namespace-lister-allow-from-konfluxui for network default I0514 16:20:57.016008 3780 pod_selector_address_set.go:213] Created shared address set for pod selector namespace-lister_LS{ML:{app: proxy,},} I0514 16:20:57.016029 3780 base_network_controller_policy.go:1054] Policy namespace-lister/namespace-lister-allow-from-konfluxui added to peer address sets [namespace-lister_LS{ML:{app: proxy,},}] I0514 16:20:57.018819 3780 base_network_controller_policy.go:1232] Create network policy namespace-lister/namespace-lister-allow-from-konfluxui resources completed, update namespace loglevel I0514 16:20:57.363754 3780 base_network_controller_policy.go:1179] Adding network policy namespace-lister/namespace-lister-allow-to-apiserver for network default I0514 16:20:57.363777 3780 base_network_controller_policy.go:1054] Policy namespace-lister/namespace-lister-allow-to-apiserver added to peer address sets [] I0514 16:20:57.364598 3780 base_network_controller_policy.go:1232] Create network policy namespace-lister/namespace-lister-allow-to-apiserver resources completed, update namespace loglevel I0514 16:20:58.547179 3780 namespace.go:142] [build-service] updating namespace I0514 16:20:59.210087 3780 namespace.go:142] [default-tenant] updating namespace I0514 16:20:59.490584 3780 namespace.go:142] [enterprise-contract-service] updating namespace I0514 16:20:59.702252 3780 namespace.go:142] [image-controller] updating namespace I0514 16:20:59.886459 3780 namespace.go:142] [integration-service] updating namespace I0514 16:21:00.093257 3780 namespace.go:142] [konflux-info] updating namespace I0514 16:21:00.285244 3780 namespace.go:142] [namespace-lister] updating namespace I0514 16:21:00.481419 3780 namespace.go:142] [release-service] updating namespace I0514 16:21:00.599607 3780 namespace.go:142] [build-service] updating namespace I0514 16:21:00.619301 3780 namespace.go:142] [default-tenant] updating namespace I0514 16:21:00.627329 3780 namespace.go:142] [enterprise-contract-service] updating namespace I0514 16:21:00.638336 3780 namespace.go:142] [image-controller] updating namespace I0514 16:21:00.645362 3780 namespace.go:142] [integration-service] updating namespace I0514 16:21:00.660195 3780 namespace.go:142] [konflux-info] updating namespace I0514 16:21:00.678214 3780 namespace.go:142] [namespace-lister] updating namespace I0514 16:21:00.684386 3780 namespace.go:142] [release-service] updating namespace I0514 16:21:00.721162 3780 namespace.go:142] [build-service] updating namespace I0514 16:21:00.808250 3780 namespace.go:142] [default-tenant] updating namespace I0514 16:21:00.905778 3780 namespace.go:142] [enterprise-contract-service] updating namespace I0514 16:21:01.010825 3780 namespace.go:142] [image-controller] updating namespace I0514 16:21:01.103383 3780 namespace.go:142] [integration-service] updating namespace I0514 16:21:01.208981 3780 namespace.go:142] [konflux-info] updating namespace I0514 16:21:01.309702 3780 namespace.go:142] [namespace-lister] updating namespace I0514 16:21:01.402391 3780 namespace.go:142] [release-service] updating namespace I0514 16:21:02.741638 3780 namespace.go:142] [konflux-cli] updating namespace I0514 16:21:02.941880 3780 namespace.go:142] [segment-bridge] updating namespace I0514 16:21:03.063100 3780 namespace.go:142] [konflux-cli] updating namespace I0514 16:21:03.079985 3780 namespace.go:142] [segment-bridge] updating namespace I0514 16:21:03.119176 3780 namespace.go:142] [konflux-cli] updating namespace I0514 16:21:03.185284 3780 namespace.go:142] [segment-bridge] updating namespace W0514 16:21:06.653787 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service build-service/build-service-controller-manager-metrics-service W0514 16:21:06.694615 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service namespace-lister/namespace-lister W0514 16:21:06.713858 3780 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service image-controller/image-controller-controller-manager-metrics-service W0514 16:21:10.488873 3780 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service konflux-ui/proxy W0514 16:21:10.488891 3780 util.go:870] Failed to build global endpoints for port TCP/web-tls: empty IP address endpoints for service konflux-ui/proxy W0514 16:21:13.722459 3780 util.go:870] Failed to build global endpoints for port TCP/: empty IP address endpoints for service release-service/release-service-webhook-service W0514 16:21:13.723582 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service release-service/release-service-controller-manager-metrics-service W0514 16:21:13.744079 3780 util.go:870] Failed to build global endpoints for port TCP/: empty IP address endpoints for service integration-service/integration-service-webhook-service W0514 16:21:13.745596 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service integration-service/integration-service-controller-manager-metrics-service W0514 16:21:13.775137 3780 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service konflux-ui/proxy W0514 16:21:13.775151 3780 util.go:870] Failed to build global endpoints for port TCP/web-tls: empty IP address endpoints for service konflux-ui/proxy W0514 16:21:13.798993 3780 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service integration-service/integration-service-controller-manager-metrics-service W0514 16:21:13.800039 3780 util.go:870] Failed to build global endpoints for port TCP/: empty IP address endpoints for service integration-service/integration-service-webhook-service W0514 16:21:13.816123 3780 util.go:870] Failed to build global endpoints for port TCP/dex: empty IP address endpoints for service konflux-ui/dex I0514 16:21:25.086990 3780 obj_retry.go:498] Detected object integration-service/integration-service-controller-manager-7b666bf6d8-7xlhz of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:21:25.087498 3780 obj_retry.go:498] Detected object integration-service/integration-service-controller-manager-7b666bf6d8-7xlhz of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:21:25.812311 3780 obj_retry.go:492] Detected object integration-service/integration-service-controller-manager-7b666bf6d8-7xlhz of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:21:25.812338 3780 obj_retry.go:492] Detected object integration-service/integration-service-controller-manager-7b666bf6d8-7xlhz of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:21:25.818829 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod integration-service/integration-service-controller-manager-7b666bf6d8-7xlhz I0514 16:21:25.818851 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod integration-service/integration-service-controller-manager-7b666bf6d8-7xlhz I0514 16:21:41.512912 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=5 I0514 16:22:28.344298 3780 obj_retry.go:498] Detected object konflux-ui/proxy-85d555f44b-wgmfg of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:22:28.344956 3780 obj_retry.go:498] Detected object konflux-ui/proxy-85d555f44b-wgmfg of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:22:29.174707 3780 obj_retry.go:492] Detected object konflux-ui/proxy-85d555f44b-wgmfg of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:22:29.174736 3780 obj_retry.go:492] Detected object konflux-ui/proxy-85d555f44b-wgmfg of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:22:29.183589 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod konflux-ui/proxy-85d555f44b-wgmfg I0514 16:22:29.183613 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod konflux-ui/proxy-85d555f44b-wgmfg I0514 16:22:30.327074 3780 obj_retry.go:498] Detected object konflux-ui/test-tsf-konflux of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:22:30.327638 3780 obj_retry.go:498] Detected object konflux-ui/test-tsf-konflux of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:22:30.353208 3780 obj_retry.go:492] Detected object konflux-ui/test-tsf-konflux of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:22:30.353230 3780 obj_retry.go:492] Detected object konflux-ui/test-tsf-konflux of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:22:30.360314 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod konflux-ui/test-tsf-konflux I0514 16:22:30.360331 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod konflux-ui/test-tsf-konflux I0514 16:22:34.491930 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=9 I0514 16:22:58.726639 3780 namespace.go:98] [managed-tenant-kicrm] adding namespace I0514 16:22:58.727731 3780 namespace.go:102] [managed-tenant-kicrm] adding namespace took 1.066083ms I0514 16:22:58.753562 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:22:58.776533 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:22:59.856325 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:22:59.990605 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:23:00.088828 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:23:07.565581 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=8 I0514 16:23:14.404521 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=160 I0514 16:23:23.400242 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=12 I0514 16:23:40.492560 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=10 I0514 16:23:52.413869 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=398 I0514 16:24:01.402909 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=46 I0514 16:24:06.496558 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=5 I0514 16:25:02.405978 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=135 I0514 16:25:16.517475 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=10 I0514 16:25:39.509751 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=8 I0514 16:26:14.527371 3780 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=8 I0514 16:26:27.529935 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=9 I0514 16:27:18.502199 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=10 I0514 16:27:32.403221 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=47 I0514 16:28:21.408493 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=7 I0514 16:29:16.496080 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=7 I0514 16:29:24.517019 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=8 I0514 16:30:04.725109 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646270-th9w6 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:30:04.725717 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646270-th9w6 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:30:05.599833 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646270-th9w6 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:30:05.599878 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646270-th9w6 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:30:07.496099 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=6 I0514 16:30:08.045552 3780 obj_retry.go:492] Detected object tsf-tpa/create-db-4wz6p of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:30:08.045573 3780 obj_retry.go:492] Detected object tsf-tpa/create-db-4wz6p of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:30:08.050127 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/create-db-4wz6p I0514 16:30:08.050139 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/create-db-4wz6p I0514 16:30:12.028309 3780 obj_retry.go:492] Detected object tsf-tpa/migrate-db-vxvdb of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:30:12.028354 3780 obj_retry.go:492] Detected object tsf-tpa/migrate-db-vxvdb of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:30:12.032661 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/migrate-db-vxvdb I0514 16:30:12.032686 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/migrate-db-vxvdb I0514 16:30:15.029622 3780 obj_retry.go:492] Detected object tsf-tpa/create-importers-rrp6f of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:30:15.029644 3780 obj_retry.go:492] Detected object tsf-tpa/create-importers-rrp6f of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:30:15.035507 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/create-importers-rrp6f I0514 16:30:15.035531 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/create-importers-rrp6f I0514 16:30:20.404091 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=8 I0514 16:30:29.569838 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=9 I0514 16:31:02.500195 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=6 W0514 16:31:30.059925 3780 duration.go:461] Config duration recorder: measurement expired for pod/openshift-console/console-5ddf55bfc7-rsjkj I0514 16:31:48.418116 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=24 I0514 16:32:16.521481 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=7 I0514 16:32:27.513210 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=7 I0514 16:32:34.505986 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=7 I0514 16:32:36.407433 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=42 I0514 16:32:48.533525 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=7 I0514 16:33:22.919794 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-pull-request-2t2v4-init-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:33:22.920404 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-pull-request-2t2v4-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:33:23.938128 3780 obj_retry.go:492] Detected object default-tenant/tsf-comp-jiwxt-on-pull-request-2t2v4-init-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:33:23.938152 3780 obj_retry.go:492] Detected object default-tenant/tsf-comp-jiwxt-on-pull-request-2t2v4-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:33:23.945160 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod default-tenant/tsf-comp-jiwxt-on-pull-request-2t2v4-init-pod I0514 16:33:23.945189 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod default-tenant/tsf-comp-jiwxt-on-pull-request-2t2v4-init-pod I0514 16:33:31.887664 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-init-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:33:31.888235 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:33:46.998542 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-clone-repository-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:33:46.999101 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-clone-repository-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:33:49.409274 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=9 I0514 16:35:00.472897 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-prefetch-dependencies-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:35:00.473471 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-prefetch-dependencies-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:35:17.520767 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=6 I0514 16:36:02.531146 3780 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=10 I0514 16:37:18.406961 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=10 I0514 16:37:24.524883 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=6 I0514 16:37:34.412591 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=11 I0514 16:37:42.424640 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-build-container-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:37:42.425144 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-build-container-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:37:49.503762 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=8 I0514 16:38:13.614769 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-build-image-index-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:38:13.615326 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-build-image-index-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:38:26.499816 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=10 I0514 16:38:32.499482 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=9 I0514 16:38:38.422089 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=133 I0514 16:38:39.411028 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=37 I0514 16:38:44.573398 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=9 I0514 16:38:50.408621 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=8 I0514 16:39:04.937217 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-deprecated-base-image-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:04.937831 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-deprecated-base-image-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:08.413421 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=7 I0514 16:39:20.509686 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=7 I0514 16:39:36.797300 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-sast-unicode-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:36.797822 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-sast-unicode-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:38.404477 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-sast-shell-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:38.405007 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-sast-shell-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:38.435645 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-rpms-signature-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:38.436157 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-rpms-signature-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:48.496260 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-tpa-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:48.496815 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-tpa-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:52.235112 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-clamav-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:52.235651 3780 obj_retry.go:498] Detected object default-tenant/tsf-comp-jiwxt-on-push-xfhhc-clamav-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:39:56.536856 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=8 I0514 16:39:57.516763 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=8 I0514 16:40:04.633791 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646280-t5phl of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:40:04.634321 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646280-t5phl of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:40:05.400745 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646280-t5phl of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:40:05.400769 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646280-t5phl of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:40:05.443928 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:40:05.443948 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:40:05.447105 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 I0514 16:40:05.447131 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-backplane/osd-delete-backplane-serviceaccounts-29646250-8h5z8 I0514 16:40:12.604175 3780 obj_retry.go:498] Detected object default-tenant/my-integration-test-saks-c4k7k-collect-keyless-params-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:40:12.604787 3780 obj_retry.go:498] Detected object default-tenant/my-integration-test-saks-c4k7k-collect-keyless-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it W0514 16:40:15.059927 3780 duration.go:461] Config duration recorder: measurement expired for pod/tsf-tas/tuf-repository-init-7mq9q-777mt I0514 16:40:31.910062 3780 obj_retry.go:498] Detected object default-tenant/my-integration-test-saks-c4k7k-verify-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:40:31.910661 3780 obj_retry.go:498] Detected object default-tenant/my-integration-test-saks-c4k7k-verify-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:08.084126 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-verify-access-to-resources-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:08.084733 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-verify-access-to-resources-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:17.191076 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-data-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:17.191795 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-data-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:28.338337 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-check-data-keys-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:28.338897 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-check-data-keys-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:29.365673 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-signing-params-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:29.366194 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-signing-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:38.970171 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-task-params-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:38.970710 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-task-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:39.004895 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-registry-token-secret-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:39.005318 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-registry-token-secret-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:41.441982 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-reduce-snapshot-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:41.442613 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-reduce-snapshot-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:54.588593 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-apply-mapping-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:41:54.589237 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-apply-mapping-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:42:04.701250 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-filter-already-released-images-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:42:04.701955 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-filter-already-released-images-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:42:10.885485 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-tpa-params-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:42:10.886076 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-collect-tpa-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:42:41.202107 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-verify-conforma-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:42:41.202685 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-verify-conforma-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:43:11.395590 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-push-snapshot-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:43:11.396121 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-push-snapshot-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:43:39.534423 3780 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=8 I0514 16:43:52.524425 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=9 I0514 16:43:58.845012 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-sign-image-cosign-keyless-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:43:58.845667 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-sign-image-cosign-keyless-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:43:59.528110 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=6 I0514 16:44:27.149938 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-process-component-sbom-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:44:27.150491 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-process-component-sbom-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:44:32.576667 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=7 I0514 16:44:34.434352 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/affinity-assistant-63365b456a-0 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:44:34.434960 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/affinity-assistant-63365b456a-0 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:44:35.034747 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/affinity-assistant-63365b456a-0 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:35.034776 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/affinity-assistant-63365b456a-0 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:35.042068 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/affinity-assistant-63365b456a-0 I0514 16:44:35.042093 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/affinity-assistant-63365b456a-0 I0514 16:44:35.111754 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-update-cr-status-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:44:35.112250 3780 obj_retry.go:498] Detected object managed-tenant-kicrm/managed-bqpr7-update-cr-status-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:44:39.823725 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:44:41.503580 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=8 I0514 16:44:44.861665 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-apply-mapping-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.861690 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-apply-mapping-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.870590 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-apply-mapping-pod I0514 16:44:44.870616 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-apply-mapping-pod I0514 16:44:44.879008 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-check-data-keys-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.879024 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-check-data-keys-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.886379 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-check-data-keys-pod I0514 16:44:44.886396 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-check-data-keys-pod I0514 16:44:44.896086 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-data-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.896103 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-data-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.904201 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-collect-data-pod I0514 16:44:44.904214 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-collect-data-pod I0514 16:44:44.913164 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-registry-token-secret-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.913186 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-registry-token-secret-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.919439 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-collect-registry-token-secret-pod I0514 16:44:44.919463 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-collect-registry-token-secret-pod I0514 16:44:44.928160 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-signing-params-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.928175 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-signing-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.934119 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-collect-signing-params-pod I0514 16:44:44.934140 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-collect-signing-params-pod I0514 16:44:44.942389 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-task-params-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.942420 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-task-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.949024 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-collect-task-params-pod I0514 16:44:44.949037 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-collect-task-params-pod I0514 16:44:44.957685 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-tpa-params-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.957699 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-collect-tpa-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.964182 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-collect-tpa-params-pod I0514 16:44:44.964206 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-collect-tpa-params-pod I0514 16:44:44.973364 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-filter-already-released-images-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.973381 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-filter-already-released-images-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.980484 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-filter-already-released-images-pod I0514 16:44:44.980504 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-filter-already-released-images-pod I0514 16:44:44.988818 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-process-component-sbom-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.988835 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-process-component-sbom-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:44.995843 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-process-component-sbom-pod I0514 16:44:44.995864 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-process-component-sbom-pod I0514 16:44:45.005675 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-push-snapshot-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.005691 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-push-snapshot-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.022002 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-push-snapshot-pod I0514 16:44:45.022020 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-push-snapshot-pod I0514 16:44:45.030521 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-reduce-snapshot-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.030537 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-reduce-snapshot-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.038185 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-reduce-snapshot-pod I0514 16:44:45.038206 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-reduce-snapshot-pod I0514 16:44:45.047002 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-sign-image-cosign-keyless-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.047017 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-sign-image-cosign-keyless-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.054507 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-sign-image-cosign-keyless-pod I0514 16:44:45.054522 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-sign-image-cosign-keyless-pod I0514 16:44:45.062954 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-update-cr-status-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.062971 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-update-cr-status-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.070516 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-update-cr-status-pod I0514 16:44:45.070537 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-update-cr-status-pod I0514 16:44:45.080329 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-verify-access-to-resources-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.080348 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-verify-access-to-resources-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.086440 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-verify-access-to-resources-pod I0514 16:44:45.086460 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-verify-access-to-resources-pod I0514 16:44:45.109863 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-verify-conforma-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.109876 3780 obj_retry.go:492] Detected object managed-tenant-kicrm/managed-bqpr7-verify-conforma-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:45.120597 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod managed-tenant-kicrm/managed-bqpr7-verify-conforma-pod I0514 16:44:45.120616 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod managed-tenant-kicrm/managed-bqpr7-verify-conforma-pod I0514 16:44:50.760013 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:44:51.071187 3780 obj_retry.go:492] Detected object default-tenant/my-integration-test-saks-c4k7k-collect-keyless-params-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:51.071211 3780 obj_retry.go:492] Detected object default-tenant/my-integration-test-saks-c4k7k-collect-keyless-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:51.075813 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod default-tenant/my-integration-test-saks-c4k7k-collect-keyless-params-pod I0514 16:44:51.075836 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod default-tenant/my-integration-test-saks-c4k7k-collect-keyless-params-pod I0514 16:44:51.883794 3780 obj_retry.go:492] Detected object default-tenant/my-integration-test-saks-c4k7k-verify-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:51.883815 3780 obj_retry.go:492] Detected object default-tenant/my-integration-test-saks-c4k7k-verify-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:44:51.890710 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod default-tenant/my-integration-test-saks-c4k7k-verify-pod I0514 16:44:51.890728 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod default-tenant/my-integration-test-saks-c4k7k-verify-pod I0514 16:44:57.259267 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:45:03.914120 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:45:10.336648 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:45:15.503490 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=7 I0514 16:45:17.131173 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:45:20.415940 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=7 I0514 16:45:28.416841 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=13 I0514 16:45:49.414460 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=49 I0514 16:45:52.314674 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:45:53.520374 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=5 I0514 16:46:09.514143 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=7 I0514 16:46:14.412342 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=9 I0514 16:46:21.411537 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=9 I0514 16:46:37.299636 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:47:02.815224 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:47:17.507339 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=11 I0514 16:47:18.868467 3780 namespace.go:142] [managed-tenant-kicrm] updating namespace I0514 16:47:18.886680 3780 namespace.go:317] [managed-tenant-kicrm] deleting namespace I0514 16:47:35.425212 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=307 I0514 16:47:54.218668 3780 obj_retry.go:498] Detected object openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-gtrnm of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:47:54.219260 3780 obj_retry.go:498] Detected object openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-gtrnm of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:47:54.562052 3780 obj_retry.go:492] Detected object openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-gtrnm of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:47:54.562072 3780 obj_retry.go:492] Detected object openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-gtrnm of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:47:54.569098 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-gtrnm I0514 16:47:54.569123 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-gtrnm I0514 16:48:34.541961 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=9 I0514 16:48:36.346269 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-pull-request-88pbd-init-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:48:36.346844 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-pull-request-88pbd-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:48:37.196616 3780 obj_retry.go:492] Detected object default-tenant/java-quarkus-177877728-on-pull-request-88pbd-init-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:48:37.196637 3780 obj_retry.go:492] Detected object default-tenant/java-quarkus-177877728-on-pull-request-88pbd-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:48:37.203088 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod default-tenant/java-quarkus-177877728-on-pull-request-88pbd-init-pod I0514 16:48:37.203120 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod default-tenant/java-quarkus-177877728-on-pull-request-88pbd-init-pod I0514 16:48:46.463619 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-init-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:48:46.464172 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:48:55.553226 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-clone-repository-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:48:55.553749 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-clone-repository-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:50:04.034322 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646290-xbwnr of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:50:04.034908 3780 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646290-xbwnr of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:50:04.901402 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646290-xbwnr of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:50:04.901447 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646290-xbwnr of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:50:04.946628 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:50:04.946645 3780 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:50:04.949934 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b I0514 16:50:04.949957 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-backplane/osd-delete-backplane-serviceaccounts-29646260-c4c5b I0514 16:50:08.364122 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-prefetch-dependencies-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:50:08.364710 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-prefetch-dependencies-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:50:47.579886 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=8 I0514 16:50:54.528163 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=7 I0514 16:51:07.507909 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=6 I0514 16:51:52.418472 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=35 I0514 16:51:53.416312 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=7 I0514 16:52:01.523997 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=7 I0514 16:52:05.531689 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=9 I0514 16:52:10.081458 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-build-container-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:10.082027 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-build-container-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:12.420970 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=13 I0514 16:52:14.419585 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=12 I0514 16:52:20.518290 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=7 I0514 16:52:40.453971 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-build-image-index-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:40.454557 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-build-image-index-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:46.511237 3780 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=6 I0514 16:52:47.394101 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-sast-unicode-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:47.394658 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-sast-unicode-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:53.428898 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=166 I0514 16:52:53.462671 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-sast-shell-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:53.463224 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-sast-shell-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:54.463021 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-tpa-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:54.463565 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-tpa-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:58.511132 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-rpms-signature-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:52:58.511659 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-rpms-signature-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:53:03.415349 3780 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=8 I0514 16:53:24.507187 3780 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=10 I0514 16:53:30.538376 3780 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=10 I0514 16:53:34.804041 3780 obj_retry.go:498] Detected object default-tenant/javacbdae7ecbcf7550bb14c7b632bd9849b1c4656d395b3c2e97e6543d-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:53:34.804599 3780 obj_retry.go:498] Detected object default-tenant/javacbdae7ecbcf7550bb14c7b632bd9849b1c4656d395b3c2e97e6543d-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:53:40.852137 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-clamav-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:53:40.852733 3780 obj_retry.go:498] Detected object default-tenant/java-quarkus-177877728-on-push-kwfrz-clamav-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:53:50.938617 3780 obj_retry.go:498] Detected object default-tenant/test-appfe9c927c882c4eacf1bb77e1d81f6ce3ffaa170e32b48ddebd6-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:53:50.939307 3780 obj_retry.go:498] Detected object default-tenant/test-appfe9c927c882c4eacf1bb77e1d81f6ce3ffaa170e32b48ddebd6-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:54:09.180448 3780 obj_retry.go:498] Detected object default-tenant/test-app-177877728-enterprise-contract-czqlc-verify-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:54:09.181009 3780 obj_retry.go:498] Detected object default-tenant/test-app-177877728-enterprise-contract-czqlc-verify-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:54:39.905351 3780 obj_retry.go:498] Detected object openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-6xfln of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0514 16:54:39.905927 3780 obj_retry.go:498] Detected object openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-6xfln of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0514 16:54:40.045549 3780 obj_retry.go:492] Detected object openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-6xfln of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:54:40.045570 3780 obj_retry.go:492] Detected object openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-6xfln of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0514 16:54:40.051697 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-6xfln I0514 16:54:40.051740 3780 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-pipelines/tekton-pipelines-webhook-7b6d6f4c94-6xfln