+ . /ovnkube-lib/ovnkube-lib.sh ++ set -x ++ K8S_NODE=ip-10-0-2-180.ec2.internal ++ [[ -n ip-10-0-2-180.ec2.internal ]] ++ [[ -f /env/ip-10-0-2-180.ec2.internal ]] ++ northd_pidfile=/var/run/ovn/ovn-northd.pid ++ controller_pidfile=/var/run/ovn/ovn-controller.pid ++ controller_logfile=/var/log/ovn/acl-audit-log.log ++ vswitch_dbsock=/var/run/openvswitch/db.sock ++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid ++ nbdb_sock=/var/run/ovn/ovnnb_db.sock ++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl ++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid ++ sbdb_sock=/var/run/ovn/ovnsb_db.sock ++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl + start-ovnkube-node 4 29103 29105 + local log_level=4 + local metrics_port=29103 + local ovn_metrics_port=29105 + ovn_advertised_udn_isolation_mode_flag= + openflow_probe_flag= + [[ 3 -ne 3 ]] + ovs-vsctl br-exists br-ex + add_garp_drop_flow br-ex + local bridge=br-ex + local cookie=0x0305 + local priority=499 ++ ovs-vsctl list-ports br-ex + for port_name in $(ovs-vsctl list-ports "$bridge") + [[ ens5 == *to-br-int ]] + cni-bin-copy + . /host/etc/os-release ++ NAME='Red Hat Enterprise Linux CoreOS' ++ VERSION='9.6.20260401-0 (Plow)' ++ ID=rhel ++ ID_LIKE=fedora ++ VERSION_ID=9.6 ++ PLATFORM_ID=platform:el9 ++ PRETTY_NAME='Red Hat Enterprise Linux CoreOS 9.6.20260401-0 (Plow)' ++ ANSI_COLOR='0;31' ++ LOGO=fedora-logo-icon ++ CPE_NAME=cpe:/o:redhat:enterprise_linux:9::baseos ++ HOME_URL=https://www.redhat.com/ ++ DOCUMENTATION_URL=https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/9 ++ BUG_REPORT_URL=https://issues.redhat.com/ ++ REDHAT_BUGZILLA_PRODUCT='Red Hat Enterprise Linux 9' ++ REDHAT_BUGZILLA_PRODUCT_VERSION=9.6 ++ REDHAT_SUPPORT_PRODUCT='Red Hat Enterprise Linux' ++ REDHAT_SUPPORT_PRODUCT_VERSION=9.6 ++ OSTREE_VERSION=9.6.20260401-0 ++ VARIANT=CoreOS ++ VARIANT_ID=coreos ++ OPENSHIFT_VERSION=4.21 + rhelmajor= + case "${ID}" in ++ echo 9.6 ++ cut -f 1 -d . + rhelmajor=9 + sourcedir=/usr/libexec/cni/ + case "${rhelmajor}" in + sourcedir=/usr/libexec/cni/rhel9 + cp -f /usr/libexec/cni/rhel9/ovn-k8s-cni-overlay /cni-bin-dir/ ++ date '+%m%d %H:%M:%S.%N' I0419 18:45:04.879449264 - disable conntrack on geneve port + echo 'I0419 18:45:04.879449264 - disable conntrack on geneve port' + iptables -t raw -A PREROUTING -p udp --dport 6081 -j NOTRACK + iptables -t raw -A OUTPUT -p udp --dport 6081 -j NOTRACK + ip6tables -t raw -A PREROUTING -p udp --dport 6081 -j NOTRACK + ip6tables -t raw -A OUTPUT -p udp --dport 6081 -j NOTRACK ++ date '+%m%d %H:%M:%S.%N' + echo 'I0419 18:45:04.917196193 - starting ovnkube-node' I0419 18:45:04.917196193 - starting ovnkube-node + egress_features_enable_flag='--enable-egress-ip=true --enable-egress-firewall=true --enable-egress-qos=true --enable-egress-service=true' + init_ovnkube_controller='--init-ovnkube-controller ip-10-0-2-180.ec2.internal' + multi_external_gateway_enable_flag=--enable-multi-external-gateway=true + gateway_interface=br-ex + enable_multicast_flag=--enable-multicast + OVN_NODE_MODE=full + '[' full == dpu-host ']' + '[' shared == shared ']' + gateway_mode_flags='--gateway-mode shared --gateway-interface br-ex' + export_network_flows_flags= + [[ -n '' ]] + [[ -n '' ]] + [[ -n '' ]] + [[ -n '' ]] + [[ -n '' ]] + [[ -n '' ]] + gw_interface_flag= + '[' -d /sys/class/net/br-ex1 ']' + node_mgmt_port_netdev_flags= + [[ -n '' ]] + [[ -n '' ]] + multi_network_enabled_flag= + [[ true == \t\r\u\e ]] + [[ full != \d\p\u\-\h\o\s\t ]] + multi_network_enabled_flag=--enable-multi-network + network_segmentation_enabled_flag= + [[ true == \t\r\u\e ]] + [[ full != \d\p\u\-\h\o\s\t ]] + multi_network_enabled_flag=--enable-multi-network + network_segmentation_enabled_flag=--enable-network-segmentation + route_advertisements_enable_flag= + [[ false == \t\r\u\e ]] + preconfigured_udn_addresses_enable_flag= + [[ true == \t\r\u\e ]] + preconfigured_udn_addresses_enable_flag=--enable-preconfigured-udn-addresses + network_observability_enabled_flag= + [[ false == \t\r\u\e ]] + multi_network_policy_enabled_flag= + [[ false == \t\r\u\e ]] + admin_network_policy_enabled_flag= + [[ true == \t\r\u\e ]] + [[ full != \d\p\u\-\h\o\s\t ]] + admin_network_policy_enabled_flag=--enable-admin-network-policy + dns_name_resolver_enabled_flag= + [[ false == \t\r\u\e ]] + ip_forwarding_flag= + '[' '' == Global ']' + ip_forwarding_flag=--disable-forwarding + sysctl -w net.ipv4.ip_forward=0 net.ipv4.ip_forward = 0 + sysctl -w net.ipv6.conf.all.forwarding=0 net.ipv6.conf.all.forwarding = 0 + [[ '' != '' ]] + [[ '' != '' ]] + NETWORK_NODE_IDENTITY_ENABLE= + [[ true == \t\r\u\e ]] + NETWORK_NODE_IDENTITY_ENABLE=' --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h ' + ovn_v4_join_subnet_opt= + [[ '' != '' ]] + ovn_v6_join_subnet_opt= + [[ '' != '' ]] + ovn_v4_masquerade_subnet_opt= + [[ 169.254.0.0/17 != '' ]] + ovn_v4_masquerade_subnet_opt='--gateway-v4-masquerade-subnet 169.254.0.0/17' + ovn_v6_masquerade_subnet_opt= + [[ fd69::/112 != '' ]] + ovn_v6_masquerade_subnet_opt='--gateway-v6-masquerade-subnet fd69::/112' + ovn_v4_transit_switch_subnet_opt= + [[ '' != '' ]] + ovn_v6_transit_switch_subnet_opt= + [[ '' != '' ]] + exec /usr/bin/ovnkube --init-ovnkube-controller ip-10-0-2-180.ec2.internal --init-node ip-10-0-2-180.ec2.internal --config-file=/run/ovnkube-config/ovnkube.conf --ovn-empty-lb-events --loglevel 4 --inactivity-probe=180000 --gateway-mode shared --gateway-interface br-ex --metrics-bind-address 127.0.0.1:29103 --ovn-metrics-bind-address 127.0.0.1:29105 --metrics-enable-pprof --metrics-enable-config-duration --export-ovs-metrics --disable-snat-multiple-gws --enable-multi-network --enable-network-segmentation --enable-preconfigured-udn-addresses --enable-admin-network-policy --enable-multicast --zone ip-10-0-2-180.ec2.internal --enable-interconnect --acl-logging-rate-limit 20 --disable-forwarding --bootstrap-kubeconfig=/var/lib/kubelet/kubeconfig --cert-dir=/etc/ovn/ovnkube-node-certs --cert-duration=24h --gateway-v4-masquerade-subnet 169.254.0.0/17 --gateway-v6-masquerade-subnet fd69::/112 --enable-egress-ip=true --enable-egress-firewall=true --enable-egress-qos=true --enable-egress-service=true --enable-multi-external-gateway=true I0419 18:45:04.971994 3645 config.go:2726] Parsed config file /run/ovnkube-config/ovnkube.conf I0419 18:45:04.972055 3645 config.go:2727] Parsed config: {Default:{MTU:8901 RoutableMTU:0 ConntrackZone:64000 HostMasqConntrackZone:0 OVNMasqConntrackZone:0 HostNodePortConntrackZone:0 ReassemblyConntrackZone:0 EncapType:geneve EncapIP: EffectiveEncapIP: EncapPort:6081 InactivityProbe:100000 OpenFlowProbe:0 OfctrlWaitBeforeClear:0 MonitorAll:true OVSDBTxnTimeout:1m40s LFlowCacheEnable:true LFlowCacheLimit:0 LFlowCacheLimitKb:1048576 RawClusterSubnets:10.128.0.0/14/23 ClusterSubnets:[] EnableUDPAggregation:true Zone:global RawUDNAllowedDefaultServices:default/kubernetes,openshift-dns/dns-default UDNAllowedDefaultServices:[] Transport:geneve} Logging:{File: CNIFile: LibovsdbFile: Level:4 LogFileMaxSize:100 LogFileMaxBackups:5 LogFileMaxAge:5 ACLLoggingRateLimit:20} Monitoring:{RawNetFlowTargets: RawSFlowTargets: RawIPFIXTargets: NetFlowTargets:[] SFlowTargets:[] IPFIXTargets:[]} IPFIX:{Sampling:400 CacheActiveTimeout:60 CacheMaxFlows:0} CNI:{ConfDir:/etc/cni/net.d Plugin:ovn-k8s-cni-overlay} OVNKubernetesFeature:{EnableAdminNetworkPolicy:false EnableEgressIP:false EgressIPReachabiltyTotalTimeout:1 EnableEgressFirewall:false EnableEgressQoS:false EnableEgressService:false EgressIPNodeHealthCheckPort:9107 EnableMultiNetwork:false EnableNetworkSegmentation:true EnableNetworkConnect:false EnablePreconfiguredUDNAddresses:true EnableRouteAdvertisements:false EnableEVPN:false EnableMultiNetworkPolicy:false EnableStatelessNetPol:false EnableInterconnect:false EnableMultiExternalGateway:false EnablePersistentIPs:false EnableDNSNameResolver:false EnableServiceTemplateSupport:false EnableObservability:false EnableNetworkQoS:false AdvertisedUDNIsolationMode:strict EnableDynamicUDNAllocation:false UDNDeletionGracePeriod:2m0s} Kubernetes:{BootstrapKubeconfig: CertDir: CertDuration:10m0s Kubeconfig: CACert: CACertData: CAData:[] APIServer:https://api.kx-d04b7a4899.hypershift.local:443 Token: TokenFile: CompatServiceCIDR: RawServiceCIDRs:172.30.0.0/16 ServiceCIDRs:[] OVNConfigNamespace:openshift-ovn-kubernetes OVNEmptyLbEvents:false RawNoHostSubnetNodes: NoHostSubnetNodes: HostNetworkNamespace:openshift-host-network DisableRequestedChassis:false PlatformType:AWS HealthzBindAddress:0.0.0.0:10256 CompatMetricsBindAddress: CompatOVNMetricsBindAddress: CompatMetricsEnablePprof:false DNSServiceNamespace:openshift-dns DNSServiceName:dns-default} Metrics:{BindAddress: OVNMetricsBindAddress: ExportOVSMetrics:false EnablePprof:false NodeServerPrivKey: NodeServerCert: EnableConfigDuration:false EnableScaleMetrics:false} OvnNorth:{Address: PrivKey: Cert: CACert: CertCommonName: Scheme: ElectionTimer:0 northbound:false RunDir:/var/run/ovn/ DbLocation:/etc/ovn/ovnnb_db.db exec:} OvnSouth:{Address: PrivKey: Cert: CACert: CertCommonName: Scheme: ElectionTimer:0 northbound:false RunDir:/var/run/ovn/ DbLocation:/etc/ovn/ovnsb_db.db exec:} Gateway:{Mode:shared Interface: GatewayAcceleratedInterface: EgressGWInterface: NextHop: VLANID:0 NodeportEnable:true DisableSNATMultipleGWs:false V4JoinSubnet:100.64.0.0/16 V6JoinSubnet:fd98::/64 V4MasqueradeSubnet:169.254.169.0/29 V6MasqueradeSubnet:fd69::/125 MasqueradeIPs:{V4OVNMasqueradeIP:169.254.169.1 V6OVNMasqueradeIP:fd69::1 V4HostMasqueradeIP:169.254.169.2 V6HostMasqueradeIP:fd69::2 V4HostETPLocalMasqueradeIP:169.254.169.3 V6HostETPLocalMasqueradeIP:fd69::3 V4DummyNextHopMasqueradeIP:169.254.169.4 V6DummyNextHopMasqueradeIP:fd69::4 V4OVNServiceHairpinMasqueradeIP:169.254.169.5 V6OVNServiceHairpinMasqueradeIP:fd69::5} DisablePacketMTUCheck:false RouterSubnet: SingleNode:false DisableForwarding:false AllowNoUplink:false EphemeralPortRange:} MasterHA:{ElectionLeaseDuration:137 ElectionRenewDeadline:107 ElectionRetryPeriod:26} ClusterMgrHA:{ElectionLeaseDuration:137 ElectionRenewDeadline:107 ElectionRetryPeriod:26} HybridOverlay:{Enabled:false RawClusterSubnets: ClusterSubnets:[] VXLANPort:4789} OvnKubeNode:{Mode:full MgmtPortNetdev: MgmtPortDPResourceName:} ClusterManager:{V4TransitSubnet:100.88.0.0/16 V6TransitSubnet:fd97::/64} OvsPaths:{RunDir:/var/run/openvswitch/} NoOverlay:{OutboundSNAT: Routing:} ManagedBGP:{ASNumber:64512 Topology:}} I0419 18:45:04.973847 3645 kube.go:431] Waiting for certificate I0419 18:45:04.973893 3645 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client" I0419 18:45:04.973927 3645 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client" I0419 18:45:04.974726 3645 cert_rotation.go:141] "Starting client certificate rotation controller" logger="tls-transport-cache" I0419 18:45:05.002473 3645 reflector.go:358] "Starting reflector" logger="kubernetes.io/kube-apiserver-client" type="*v1.CertificateSigningRequest" resyncPeriod="0s" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" I0419 18:45:05.002504 3645 reflector.go:404] "Listing and watching" logger="kubernetes.io/kube-apiserver-client" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" I0419 18:45:05.005961 3645 reflector.go:436] "Caches populated" logger="kubernetes.io/kube-apiserver-client" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" I0419 18:45:05.006050 3645 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client" csr="csr-zhsnx" I0419 18:45:05.012298 3645 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client" csr="csr-zhsnx" I0419 18:45:05.012377 3645 reflector.go:364] "Stopping reflector" logger="kubernetes.io/kube-apiserver-client" type="*v1.CertificateSigningRequest" resyncPeriod="0s" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" I0419 18:45:05.974071 3645 kube.go:438] Certificate found I0419 18:45:05.974623 3645 cert_rotation.go:141] "Starting client certificate rotation controller" logger="tls-transport-cache" I0419 18:45:05.975529 3645 metrics.go:532] Starting metrics server at address "127.0.0.1:29103" I0419 18:45:06.039822 3645 metrics.go:164] OVN Kube log file not specified in config, therefore not starting the log file metric monitor I0419 18:45:06.057712 3645 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client" expiration="2026-04-20 18:40:05 +0000 UTC" deadline="2026-04-20 13:42:19.437306 +0000 UTC" I0419 18:45:06.057792 3645 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client" sleep="18h57m13.379539841s" I0419 18:45:06.064451 3645 metrics.go:512] Create OVN Metrics Server on address: 127.0.0.1:29105 I0419 18:45:06.064537 3645 server.go:101] MetricServer registers OVS metrics I0419 18:45:06.064566 3645 node_controller_manager.go:338] Starting the node network controller manager, Mode: full I0419 18:45:06.064624 3645 factory.go:561] Starting watch factory I0419 18:45:06.064723 3645 reflector.go:358] "Starting reflector" type="*v1.Node" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.064742 3645 reflector.go:404] "Listing and watching" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.064829 3645 reflector.go:358] "Starting reflector" type="*v1.Service" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.064842 3645 reflector.go:358] "Starting reflector" type="*v1.NetworkPolicy" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.064859 3645 reflector.go:404] "Listing and watching" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.064866 3645 reflector.go:358] "Starting reflector" type="*v1.Namespace" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.064885 3645 reflector.go:404] "Listing and watching" type="*v1.NetworkPolicy" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.064887 3645 reflector.go:404] "Listing and watching" type="*v1.Namespace" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.064917 3645 server.go:105] MetricServer registers OVN DB metrics I0419 18:45:06.065124 3645 reflector.go:358] "Starting reflector" type="*v1.Pod" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.065194 3645 reflector.go:404] "Listing and watching" type="*v1.Pod" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.065133 3645 reflector.go:358] "Starting reflector" type="*v1.EndpointSlice" resyncPeriod="0s" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.065252 3645 reflector.go:404] "Listing and watching" type="*v1.EndpointSlice" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.067662 3645 controller_manager.go:370] Starting the ovnkube controller I0419 18:45:06.067674 3645 controller_manager.go:375] Waiting up to 5m0s for NBDB zone to match: ip-10-0-2-180.ec2.internal I0419 18:45:06.067746 3645 controller_manager.go:395] NBDB zone sync took: 65.47µs I0419 18:45:06.067756 3645 factory.go:561] Starting watch factory I0419 18:45:06.070905 3645 ovn_db.go:330] /var/run/openvswitch/ovnnb_db.sock getting info failed: stat /var/run/openvswitch/ovnnb_db.sock: no such file or directory I0419 18:45:06.070929 3645 ovn_db.go:327] ovnnb_db.sock found at /var/run/ovn/ I0419 18:45:06.083874 3645 reflector.go:436] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.083926 3645 reflector.go:436] "Caches populated" type="*v1.NetworkPolicy" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.084467 3645 reflector.go:436] "Caches populated" type="*v1.EndpointSlice" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.084754 3645 reflector.go:436] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.085450 3645 ovn_db.go:388] Found OVN NB DB: &{0x1e9e6e0 /etc/ovn/ovnnb_db.db OVN_Northbound 0} I0419 18:45:06.085515 3645 ovn_db.go:395] Found OVN SB DB: &{0x1e9e9a0 /etc/ovn/ovnsb_db.db OVN_Southbound 0} I0419 18:45:06.089571 3645 reflector.go:436] "Caches populated" type="*v1.Namespace" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.090243 3645 ovn_db.go:410] Found db is standalone, don't register db_cluster metrics I0419 18:45:06.090283 3645 server.go:109] MetricServer registers OVN Controller metrics I0419 18:45:06.095618 3645 server.go:113] MetricServer registers OVN Northd metrics I0419 18:45:06.099241 3645 reflector.go:436] "Caches populated" type="*v1.Pod" reflector="k8s.io/client-go/informers/factory.go:160" I0419 18:45:06.102920 3645 metrics.go:519] OVN Metrics Server starts to run ... I0419 18:45:06.170516 3645 factory.go:1977] *v1.Node informer cache synced successfully I0419 18:45:06.170562 3645 factory.go:1977] *v1.Service informer cache synced successfully I0419 18:45:06.170566 3645 factory.go:1977] *v1.EndpointSlice informer cache synced successfully I0419 18:45:06.170569 3645 factory.go:1977] *v1.Pod informer cache synced successfully I0419 18:45:06.170572 3645 factory.go:1977] *v1.NetworkPolicy informer cache synced successfully I0419 18:45:06.170575 3645 factory.go:1977] *v1.Namespace informer cache synced successfully I0419 18:45:06.170667 3645 reflector.go:358] "Starting reflector" type="*v1alpha1.BaselineAdminNetworkPolicy" resyncPeriod="0s" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0419 18:45:06.170676 3645 reflector.go:358] "Starting reflector" type="*v1alpha1.AdminNetworkPolicy" resyncPeriod="0s" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0419 18:45:06.170686 3645 reflector.go:404] "Listing and watching" type="*v1alpha1.BaselineAdminNetworkPolicy" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0419 18:45:06.170695 3645 reflector.go:404] "Listing and watching" type="*v1alpha1.AdminNetworkPolicy" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0419 18:45:06.173442 3645 factory.go:1977] *v1.Node informer cache synced successfully I0419 18:45:06.173456 3645 factory.go:1977] *v1.Service informer cache synced successfully I0419 18:45:06.173459 3645 factory.go:1977] *v1.EndpointSlice informer cache synced successfully I0419 18:45:06.173463 3645 factory.go:1977] *v1.Pod informer cache synced successfully I0419 18:45:06.173466 3645 factory.go:1977] *v1.NetworkPolicy informer cache synced successfully I0419 18:45:06.173469 3645 factory.go:1977] *v1.Namespace informer cache synced successfully I0419 18:45:06.173732 3645 reflector.go:436] "Caches populated" type="*v1alpha1.BaselineAdminNetworkPolicy" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0419 18:45:06.173733 3645 reflector.go:436] "Caches populated" type="*v1alpha1.AdminNetworkPolicy" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" I0419 18:45:06.176108 3645 factory.go:1977] *v1alpha1.BaselineAdminNetworkPolicy informer cache synced successfully I0419 18:45:06.176118 3645 factory.go:1977] *v1alpha1.AdminNetworkPolicy informer cache synced successfully I0419 18:45:06.176179 3645 reflector.go:358] "Starting reflector" type="*v1.EgressIP" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.176193 3645 reflector.go:404] "Listing and watching" type="*v1.EgressIP" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.178492 3645 factory.go:1977] *v1alpha1.AdminNetworkPolicy informer cache synced successfully I0419 18:45:06.178502 3645 factory.go:1977] *v1alpha1.BaselineAdminNetworkPolicy informer cache synced successfully I0419 18:45:06.178671 3645 reflector.go:436] "Caches populated" type="*v1.EgressIP" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.181943 3645 factory.go:1977] *v1.EgressIP informer cache synced successfully I0419 18:45:06.182002 3645 reflector.go:358] "Starting reflector" type="*v1.EgressFirewall" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.182014 3645 reflector.go:404] "Listing and watching" type="*v1.EgressFirewall" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.184228 3645 factory.go:1977] *v1.EgressIP informer cache synced successfully I0419 18:45:06.184462 3645 reflector.go:436] "Caches populated" type="*v1.EgressFirewall" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.187527 3645 factory.go:1977] *v1.EgressFirewall informer cache synced successfully I0419 18:45:06.187582 3645 reflector.go:358] "Starting reflector" type="*v1.EgressQoS" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.187592 3645 reflector.go:404] "Listing and watching" type="*v1.EgressQoS" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.189739 3645 factory.go:1977] *v1.EgressFirewall informer cache synced successfully I0419 18:45:06.189895 3645 reflector.go:436] "Caches populated" type="*v1.EgressQoS" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.193037 3645 factory.go:1977] *v1.EgressQoS informer cache synced successfully I0419 18:45:06.193091 3645 reflector.go:358] "Starting reflector" type="*v1.EgressService" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.193102 3645 reflector.go:404] "Listing and watching" type="*v1.EgressService" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.195231 3645 factory.go:1977] *v1.EgressQoS informer cache synced successfully I0419 18:45:06.195377 3645 reflector.go:436] "Caches populated" type="*v1.EgressService" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.198564 3645 factory.go:1977] *v1.EgressService informer cache synced successfully I0419 18:45:06.198614 3645 reflector.go:358] "Starting reflector" type="*v1.AdminPolicyBasedExternalRoute" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.198625 3645 reflector.go:404] "Listing and watching" type="*v1.AdminPolicyBasedExternalRoute" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.200794 3645 factory.go:1977] *v1.EgressService informer cache synced successfully I0419 18:45:06.201013 3645 reflector.go:436] "Caches populated" type="*v1.AdminPolicyBasedExternalRoute" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.204097 3645 factory.go:1977] *v1.AdminPolicyBasedExternalRoute informer cache synced successfully I0419 18:45:06.204255 3645 reflector.go:358] "Starting reflector" type="*v1.NetworkAttachmentDefinition" resyncPeriod="0s" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" I0419 18:45:06.204271 3645 reflector.go:404] "Listing and watching" type="*v1.NetworkAttachmentDefinition" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" I0419 18:45:06.206484 3645 factory.go:1977] *v1.AdminPolicyBasedExternalRoute informer cache synced successfully I0419 18:45:06.206632 3645 reflector.go:436] "Caches populated" type="*v1.NetworkAttachmentDefinition" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" I0419 18:45:06.209921 3645 factory.go:1977] *v1.NetworkAttachmentDefinition informer cache synced successfully I0419 18:45:06.210034 3645 reflector.go:358] "Starting reflector" type="*v1.UserDefinedNetwork" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.210055 3645 reflector.go:404] "Listing and watching" type="*v1.UserDefinedNetwork" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.210032 3645 reflector.go:358] "Starting reflector" type="*v1.ClusterUserDefinedNetwork" resyncPeriod="0s" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.210114 3645 reflector.go:404] "Listing and watching" type="*v1.ClusterUserDefinedNetwork" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.212266 3645 factory.go:1977] *v1.NetworkAttachmentDefinition informer cache synced successfully I0419 18:45:06.212616 3645 reflector.go:436] "Caches populated" type="*v1.UserDefinedNetwork" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.213043 3645 reflector.go:436] "Caches populated" type="*v1.ClusterUserDefinedNetwork" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" I0419 18:45:06.214963 3645 factory.go:1977] *v1.UserDefinedNetwork informer cache synced successfully I0419 18:45:06.214973 3645 factory.go:1977] *v1.ClusterUserDefinedNetwork informer cache synced successfully I0419 18:45:06.214980 3645 factory.go:680] Watch Factory start up complete, took: 150.358843ms I0419 18:45:06.215048 3645 default_node_network_controller.go:171] Enable node proxy healthz server on 0.0.0.0:10256 I0419 18:45:06.217345 3645 factory.go:1977] *v1.UserDefinedNetwork informer cache synced successfully I0419 18:45:06.217355 3645 factory.go:1977] *v1.ClusterUserDefinedNetwork informer cache synced successfully I0419 18:45:06.217362 3645 factory.go:680] Watch Factory start up complete, took: 149.604842ms I0419 18:45:06.217368 3645 controller_manager.go:406] Waiting up to 5m0s for a node to have "ip-10-0-2-180.ec2.internal" zone I0419 18:45:06.274583 3645 default_node_network_controller.go:710] Initializing the default node network controller I0419 18:45:06.362456 3645 udn_isolation.go:101] Starting UDN host isolation manager I0419 18:45:06.368082 3645 udn_isolation.go:112] Found kubelet cgroup path: system.slice/kubelet.service I0419 18:45:06.440035 3645 controller.go:133] Adding controller udn-host-isolation-manager event handlers I0419 18:45:06.440100 3645 shared_informer.go:349] "Waiting for caches to sync" controller="udn-host-isolation-manager" I0419 18:45:06.440118 3645 shared_informer.go:356] "Caches are synced" controller="udn-host-isolation-manager" I0419 18:45:06.473332 3645 controller.go:157] Starting controller udn-host-isolation-manager with 1 workers I0419 18:45:06.473398 3645 default_node_network_controller.go:809] Node ip-10-0-2-180.ec2.internal ready for ovn initialization with subnet 10.130.0.0/23 I0419 18:45:06.478900 3645 kube.go:131] Setting annotations map[k8s.ovn.org/node-encap-ips:["10.0.2.180"] k8s.ovn.org/zone-name:ip-10-0-2-180.ec2.internal] on node ip-10-0-2-180.ec2.internal I0419 18:45:06.500175 3645 gateway_init.go:234] Initializing Gateway Functionality for Gateway PreStart I0419 18:45:06.506576 3645 helper_linux.go:93] Provided gateway interface "br-ex", found as index: 5 I0419 18:45:06.506709 3645 helper_linux.go:118] Found default gateway interface br-ex 10.0.2.1 I0419 18:45:06.506827 3645 gateway_init.go:263] Preparing Gateway I0419 18:45:06.506839 3645 gateway_shared_intf.go:1687] Creating new gateway I0419 18:45:06.621332 3645 iptables.go:108] Creating table: filter chain: FORWARD I0419 18:45:06.628219 3645 iptables.go:108] Creating table: filter chain: OUTPUT I0419 18:45:06.634946 3645 gateway_shared_intf.go:1816] Gateway Creation Complete I0419 18:45:06.634964 3645 kube.go:131] Setting annotations map[k8s.ovn.org/gateway-mtu-support: k8s.ovn.org/l3-gateway-config:{"default":{"mode":"shared","bridge-id":"br-ex","interface-id":"br-ex_ip-10-0-2-180.ec2.internal","mac-address":"0e:31:c5:da:9b:db","ip-addresses":["10.0.2.180/24"],"ip-address":"10.0.2.180/24","next-hops":["10.0.2.1"],"next-hop":"10.0.2.1","node-port-enable":"true","vlan-id":"0"}} k8s.ovn.org/node-chassis-id:b4127fe2-9fdb-4ee3-bc09-e67f72a2922a k8s.ovn.org/node-encap-ips:["10.0.2.180"] k8s.ovn.org/node-primary-ifaddr:{"ipv4":"10.0.2.180/24"} k8s.ovn.org/zone-name:ip-10-0-2-180.ec2.internal] on node ip-10-0-2-180.ec2.internal I0419 18:45:06.654168 3645 default_node_network_controller.go:519] Node connection status = connected I0419 18:45:06.718451 3645 controller_manager.go:429] Waiting for node in zone sync took: 501.068927ms I0419 18:45:06.723914 3645 controller_manager.go:611] Switching to transit router for layer2 networks I0419 18:45:06.723930 3645 kube.go:131] Setting annotations map[k8s.ovn.org/layer2-topology-version:2.0] on node ip-10-0-2-180.ec2.internal I0419 18:45:06.741927 3645 metrics.go:164] OVN Kube log file not specified in config, therefore not starting the log file metric monitor I0419 18:45:06.749184 3645 controller_manager.go:290] SCTP support detected in OVN I0419 18:45:06.749792 3645 services_controller.go:69] Creating services controller for network=default I0419 18:45:06.749908 3645 udn_enabled_svc.go:80] Waiting for service informer to sync I0419 18:45:06.750050 3645 controller.go:133] Adding controller [zone-nad-controller NAD controller] event handlers I0419 18:45:06.750099 3645 shared_informer.go:349] "Waiting for caches to sync" controller="[zone-nad-controller NAD controller]" I0419 18:45:06.750113 3645 shared_informer.go:356] "Caches are synced" controller="[zone-nad-controller NAD controller]" I0419 18:45:06.750122 3645 nad_controller.go:541] [zone-nad-controller NAD controller]: 1 NADs are missing the network ID annotation, fetching from nodes I0419 18:45:06.750353 3645 controller.go:157] Starting controller [zone-nad-controller NAD controller] with 1 workers I0419 18:45:06.750440 3645 udn_enabled_svc.go:95] Performing full resync I0419 18:45:06.750454 3645 nad_controller.go:580] [zone-nad-controller NAD controller]: finished syncing NAD openshift-ovn-kubernetes/default, took 65.824µs I0419 18:45:06.750495 3645 network_controller.go:373] [zone-nad-controller network controller]: syncing all networks I0419 18:45:06.750530 3645 network_controller.go:402] [zone-nad-controller network controller]: finished syncing network default, took 26.138µs I0419 18:45:06.750537 3645 network_controller.go:384] [zone-nad-controller network controller]: finished syncing all networks. Time taken: 42.566µs I0419 18:45:06.750543 3645 controller.go:157] Starting controller [zone-nad-controller network controller] with 1 workers I0419 18:45:06.750553 3645 nad_controller.go:379] [zone-nad-controller NAD controller]: started I0419 18:45:06.750562 3645 default_network_controller.go:325] Starting the default network controller I0419 18:45:06.750574 3645 network_controller.go:402] [zone-nad-controller network controller]: finished syncing network default, took 13.259µs I0419 18:45:06.750699 3645 udn_enabled_svc.go:99] Waiting for handler to sync I0419 18:45:06.750708 3645 udn_enabled_svc.go:104] Starting worker I0419 18:45:06.759838 3645 default_network_controller.go:395] Cleaning External Gateway ECMP routes I0419 18:45:06.759902 3645 repair.go:33] Syncing exgw routes took 50.427µs I0419 18:45:06.759923 3645 default_network_controller.go:414] Starting all the Watchers... I0419 18:45:06.760062 3645 namespace.go:98] [openshift-cloud-network-config-controller] adding namespace I0419 18:45:06.760063 3645 namespace.go:98] [openshift-deployment-validation-operator] adding namespace I0419 18:45:06.760068 3645 namespace.go:98] [open-cluster-management-agent-addon] adding namespace I0419 18:45:06.760061 3645 namespace.go:98] [openshift-authentication-operator] adding namespace I0419 18:45:06.760080 3645 namespace.go:98] [openshift-cluster-samples-operator] adding namespace I0419 18:45:06.760079 3645 namespace.go:98] [openshift-backplane] adding namespace I0419 18:45:06.760095 3645 namespace.go:98] [openshift-kube-apiserver-operator] adding namespace I0419 18:45:06.760093 3645 namespace.go:98] [openshift-cluster-storage-operator] adding namespace I0419 18:45:06.760103 3645 namespace.go:98] [openshift-node] adding namespace I0419 18:45:06.760103 3645 namespace.go:98] [openshift-apiserver-operator] adding namespace I0419 18:45:06.760111 3645 namespace.go:98] [default] adding namespace I0419 18:45:06.760111 3645 namespace.go:98] [openshift-monitoring] adding namespace I0419 18:45:06.760120 3645 namespace.go:98] [openshift-operators-redhat] adding namespace I0419 18:45:06.760123 3645 namespace.go:98] [openshift-kube-apiserver] adding namespace I0419 18:45:06.760158 3645 namespace.go:98] [openshift] adding namespace I0419 18:45:06.760616 3645 namespace.go:102] [openshift-cloud-network-config-controller] adding namespace took 541.712µs I0419 18:45:06.760633 3645 namespace.go:98] [openshift-infra] adding namespace I0419 18:45:06.761075 3645 namespace.go:102] [openshift-infra] adding namespace took 436.19µs I0419 18:45:06.761092 3645 namespace.go:98] [openshift-host-network] adding namespace I0419 18:45:06.761685 3645 namespace.go:102] [openshift-host-network] adding namespace took 588.617µs I0419 18:45:06.761697 3645 namespace.go:98] [openshift-backplane-srep] adding namespace I0419 18:45:06.762111 3645 namespace.go:102] [openshift-deployment-validation-operator] adding namespace took 2.030692ms I0419 18:45:06.762131 3645 namespace.go:98] [openshift-machine-api] adding namespace I0419 18:45:06.762654 3645 namespace.go:102] [open-cluster-management-agent-addon] adding namespace took 2.570707ms I0419 18:45:06.762670 3645 namespace.go:98] [openshift-kube-storage-version-migrator-operator] adding namespace I0419 18:45:06.763127 3645 namespace.go:102] [openshift-authentication-operator] adding namespace took 3.041242ms I0419 18:45:06.763145 3645 namespace.go:98] [openshift-user-workload-monitoring] adding namespace I0419 18:45:06.763529 3645 namespace.go:102] [openshift-cluster-samples-operator] adding namespace took 3.44129ms I0419 18:45:06.763542 3645 namespace.go:98] [openshift-console] adding namespace I0419 18:45:06.764021 3645 namespace.go:102] [openshift-backplane] adding namespace took 3.929936ms I0419 18:45:06.764035 3645 namespace.go:98] [openshift-config-operator] adding namespace I0419 18:45:06.764444 3645 namespace.go:102] [openshift-kube-apiserver-operator] adding namespace took 4.344577ms I0419 18:45:06.764455 3645 namespace.go:98] [openshift-backplane-tam] adding namespace I0419 18:45:06.764920 3645 namespace.go:102] [openshift-node] adding namespace took 4.810384ms I0419 18:45:06.764940 3645 namespace.go:98] [openshift-route-controller-manager] adding namespace I0419 18:45:06.765308 3645 namespace.go:102] [openshift-cluster-storage-operator] adding namespace took 5.201613ms I0419 18:45:06.765321 3645 namespace.go:98] [openshift-package-operator] adding namespace I0419 18:45:06.765820 3645 namespace.go:102] [openshift-apiserver-operator] adding namespace took 5.70779ms I0419 18:45:06.765834 3645 namespace.go:98] [kube-node-lease] adding namespace I0419 18:45:06.766251 3645 namespace.go:102] [openshift-monitoring] adding namespace took 6.130231ms I0419 18:45:06.766261 3645 namespace.go:98] [openshift-console-user-settings] adding namespace I0419 18:45:06.767033 3645 namespace.go:102] [default] adding namespace took 6.910285ms I0419 18:45:06.767052 3645 namespace.go:98] [openshift-marketplace] adding namespace I0419 18:45:06.767537 3645 namespace.go:102] [openshift-operators-redhat] adding namespace took 7.411957ms I0419 18:45:06.767550 3645 namespace.go:98] [openshift-kube-controller-manager] adding namespace I0419 18:45:06.768000 3645 namespace.go:102] [openshift-kube-apiserver] adding namespace took 7.867491ms I0419 18:45:06.768019 3645 namespace.go:98] [openshift-network-console] adding namespace I0419 18:45:06.768419 3645 namespace.go:102] [openshift] adding namespace took 8.251332ms I0419 18:45:06.768431 3645 namespace.go:98] [openshift-customer-monitoring] adding namespace I0419 18:45:06.768822 3645 namespace.go:102] [openshift-backplane-srep] adding namespace took 7.120163ms I0419 18:45:06.768836 3645 namespace.go:98] [openshift-dns-operator] adding namespace I0419 18:45:06.769221 3645 namespace.go:102] [openshift-machine-api] adding namespace took 7.082453ms I0419 18:45:06.769236 3645 namespace.go:98] [openshift-cloud-credential-operator] adding namespace I0419 18:45:06.769630 3645 namespace.go:102] [openshift-kube-storage-version-migrator-operator] adding namespace took 6.954452ms I0419 18:45:06.769643 3645 namespace.go:98] [kube-system] adding namespace I0419 18:45:06.770035 3645 namespace.go:102] [openshift-user-workload-monitoring] adding namespace took 6.885037ms I0419 18:45:06.770048 3645 namespace.go:98] [openshift-cluster-csi-drivers] adding namespace I0419 18:45:06.770388 3645 namespace.go:102] [openshift-console] adding namespace took 6.841252ms I0419 18:45:06.770417 3645 namespace.go:98] [openshift-network-node-identity] adding namespace I0419 18:45:06.770869 3645 namespace.go:102] [openshift-config-operator] adding namespace took 6.829239ms I0419 18:45:06.770882 3645 namespace.go:98] [openshift-operator-lifecycle-manager] adding namespace I0419 18:45:06.771211 3645 namespace.go:102] [openshift-backplane-tam] adding namespace took 6.750622ms I0419 18:45:06.771224 3645 namespace.go:98] [openshift-backplane-cee] adding namespace I0419 18:45:06.771678 3645 namespace.go:102] [openshift-route-controller-manager] adding namespace took 6.732357ms I0419 18:45:06.771690 3645 namespace.go:98] [openshift-backplane-mcs-tier-two] adding namespace I0419 18:45:06.772053 3645 namespace.go:102] [openshift-package-operator] adding namespace took 6.726208ms I0419 18:45:06.772069 3645 namespace.go:98] [openshift-dns] adding namespace I0419 18:45:06.772511 3645 namespace.go:102] [kube-node-lease] adding namespace took 6.672252ms I0419 18:45:06.772525 3645 namespace.go:98] [openshift-service-ca-operator] adding namespace I0419 18:45:06.772880 3645 namespace.go:102] [openshift-console-user-settings] adding namespace took 6.613979ms I0419 18:45:06.772894 3645 namespace.go:98] [openshift-ingress-operator] adding namespace I0419 18:45:06.773335 3645 namespace.go:102] [openshift-marketplace] adding namespace took 6.275819ms I0419 18:45:06.773349 3645 namespace.go:98] [openshift-backplane-csm] adding namespace I0419 18:45:06.773741 3645 namespace.go:102] [openshift-kube-controller-manager] adding namespace took 6.186166ms I0419 18:45:06.773754 3645 namespace.go:98] [openshift-controller-manager-operator] adding namespace I0419 18:45:06.774124 3645 namespace.go:102] [openshift-network-console] adding namespace took 6.09892ms I0419 18:45:06.774135 3645 namespace.go:98] [kube-public] adding namespace I0419 18:45:06.774531 3645 namespace.go:102] [openshift-customer-monitoring] adding namespace took 6.094783ms I0419 18:45:06.774543 3645 namespace.go:98] [openshift-operators] adding namespace I0419 18:45:06.774934 3645 namespace.go:102] [openshift-dns-operator] adding namespace took 6.093214ms I0419 18:45:06.775322 3645 namespace.go:102] [openshift-cloud-credential-operator] adding namespace took 6.081409ms I0419 18:45:06.775333 3645 namespace.go:98] [openshift-network-diagnostics] adding namespace I0419 18:45:06.775728 3645 namespace.go:102] [kube-system] adding namespace took 6.079994ms I0419 18:45:06.775741 3645 namespace.go:98] [openshift-machine-config-operator] adding namespace I0419 18:45:06.776081 3645 namespace.go:102] [openshift-cluster-csi-drivers] adding namespace took 6.027362ms I0419 18:45:06.776095 3645 namespace.go:98] [openshift-config-managed] adding namespace I0419 18:45:06.776481 3645 namespace.go:102] [openshift-network-node-identity] adding namespace took 6.057451ms I0419 18:45:06.776493 3645 namespace.go:98] [openshift-controller-manager] adding namespace I0419 18:45:06.776969 3645 namespace.go:102] [openshift-operator-lifecycle-manager] adding namespace took 6.080656ms I0419 18:45:06.776988 3645 namespace.go:98] [openshift-insights] adding namespace I0419 18:45:06.777376 3645 namespace.go:102] [openshift-backplane-cee] adding namespace took 6.146209ms I0419 18:45:06.777393 3645 namespace.go:98] [openshift-cloud-controller-manager] adding namespace I0419 18:45:06.777825 3645 namespace.go:102] [openshift-backplane-mcs-tier-two] adding namespace took 6.128875ms I0419 18:45:06.777843 3645 namespace.go:98] [openshift-ovn-kubernetes] adding namespace I0419 18:45:06.778265 3645 namespace.go:102] [openshift-dns] adding namespace took 6.191705ms I0419 18:45:06.778276 3645 namespace.go:98] [open-cluster-management-2ppd21iub53sh96ramq6ul1mg6plu4eu] adding namespace I0419 18:45:06.778702 3645 namespace.go:102] [openshift-service-ca-operator] adding namespace took 6.171332ms I0419 18:45:06.778720 3645 namespace.go:98] [openshift-kube-controller-manager-operator] adding namespace I0419 18:45:06.779116 3645 namespace.go:102] [openshift-ingress-operator] adding namespace took 6.215911ms I0419 18:45:06.779131 3645 namespace.go:98] [openshift-backplane-mobb] adding namespace I0419 18:45:06.779545 3645 namespace.go:102] [openshift-backplane-csm] adding namespace took 6.189751ms I0419 18:45:06.779563 3645 namespace.go:98] [openshift-kube-scheduler-operator] adding namespace I0419 18:45:06.779962 3645 namespace.go:102] [openshift-controller-manager-operator] adding namespace took 6.20094ms I0419 18:45:06.779979 3645 namespace.go:98] [openshift-cluster-node-tuning-operator] adding namespace I0419 18:45:06.780372 3645 namespace.go:102] [kube-public] adding namespace took 6.231098ms I0419 18:45:06.780389 3645 namespace.go:98] [openshift-ingress-canary] adding namespace W0419 18:45:06.780785 3645 base_network_controller_namespace.go:410] Failed to get IPs for pod openshift-network-diagnostics/network-check-target-p9ldg: pod openshift-network-diagnostics/network-check-target-p9ldg: no pod IPs found I0419 18:45:06.780838 3645 namespace.go:102] [openshift-operators] adding namespace took 6.289714ms I0419 18:45:06.780854 3645 namespace.go:98] [openshift-cluster-machine-approver] adding namespace I0419 18:45:06.781275 3645 namespace.go:102] [openshift-network-diagnostics] adding namespace took 5.936804ms I0419 18:45:06.781285 3645 namespace.go:98] [openshift-backplane-cse] adding namespace I0419 18:45:06.781693 3645 namespace.go:102] [openshift-machine-config-operator] adding namespace took 5.946447ms I0419 18:45:06.781705 3645 namespace.go:98] [openshift-backplane-srep-ro] adding namespace I0419 18:45:06.782051 3645 namespace.go:102] [openshift-config-managed] adding namespace took 5.950237ms I0419 18:45:06.782065 3645 namespace.go:98] [openshift-logging] adding namespace I0419 18:45:06.782704 3645 namespace.go:102] [openshift-controller-manager] adding namespace took 6.20528ms I0419 18:45:06.782716 3645 namespace.go:98] [openshift-config] adding namespace I0419 18:45:06.783141 3645 namespace.go:102] [openshift-insights] adding namespace took 6.145113ms I0419 18:45:06.783160 3645 namespace.go:98] [openshift-console-operator] adding namespace I0419 18:45:06.783574 3645 namespace.go:102] [openshift-cloud-controller-manager] adding namespace took 6.174288ms I0419 18:45:06.783590 3645 namespace.go:98] [openshift-ingress] adding namespace I0419 18:45:06.784002 3645 namespace.go:102] [openshift-ovn-kubernetes] adding namespace took 6.152549ms I0419 18:45:06.784019 3645 namespace.go:98] [openshift-etcd] adding namespace I0419 18:45:06.784457 3645 namespace.go:102] [open-cluster-management-2ppd21iub53sh96ramq6ul1mg6plu4eu] adding namespace took 6.175474ms I0419 18:45:06.784474 3645 namespace.go:98] [openshift-must-gather-operator] adding namespace I0419 18:45:06.784879 3645 namespace.go:102] [openshift-kube-controller-manager-operator] adding namespace took 6.152378ms I0419 18:45:06.784891 3645 namespace.go:98] [openshift-apiserver] adding namespace I0419 18:45:06.785318 3645 namespace.go:102] [openshift-backplane-mobb] adding namespace took 6.181387ms I0419 18:45:06.785330 3645 namespace.go:98] [openshift-cluster-version] adding namespace I0419 18:45:06.785736 3645 namespace.go:102] [openshift-kube-scheduler-operator] adding namespace took 6.16666ms I0419 18:45:06.785748 3645 namespace.go:98] [openshift-backplane-lpsre] adding namespace I0419 18:45:06.786573 3645 namespace.go:102] [openshift-cluster-node-tuning-operator] adding namespace took 6.589527ms I0419 18:45:06.786584 3645 namespace.go:98] [openshift-image-registry] adding namespace I0419 18:45:06.786993 3645 namespace.go:102] [openshift-ingress-canary] adding namespace took 6.597999ms I0419 18:45:06.787007 3645 namespace.go:98] [openshift-multus] adding namespace I0419 18:45:06.787389 3645 namespace.go:102] [openshift-cluster-machine-approver] adding namespace took 6.529127ms I0419 18:45:06.787420 3645 namespace.go:98] [openshift-backplane-managed-scripts] adding namespace I0419 18:45:06.787804 3645 namespace.go:102] [openshift-backplane-cse] adding namespace took 6.514208ms I0419 18:45:06.788200 3645 namespace.go:102] [openshift-backplane-srep-ro] adding namespace took 6.490069ms I0419 18:45:06.788612 3645 namespace.go:102] [openshift-logging] adding namespace took 6.541169ms I0419 18:45:06.789031 3645 namespace.go:102] [openshift-config] adding namespace took 6.30901ms I0419 18:45:06.789430 3645 namespace.go:102] [openshift-console-operator] adding namespace took 6.262853ms I0419 18:45:06.789836 3645 namespace.go:102] [openshift-ingress] adding namespace took 6.240798ms I0419 18:45:06.789847 3645 namespace.go:98] [openshift-authentication] adding namespace I0419 18:45:06.790189 3645 namespace.go:102] [openshift-etcd] adding namespace took 6.163421ms I0419 18:45:06.790200 3645 namespace.go:98] [dedicated-admin] adding namespace I0419 18:45:06.790697 3645 namespace.go:102] [openshift-must-gather-operator] adding namespace took 6.217378ms I0419 18:45:06.790716 3645 namespace.go:98] [openshift-kube-scheduler] adding namespace I0419 18:45:06.791130 3645 namespace.go:102] [openshift-apiserver] adding namespace took 6.234618ms I0419 18:45:06.791142 3645 namespace.go:98] [openshift-network-operator] adding namespace I0419 18:45:06.791630 3645 namespace.go:102] [openshift-cluster-version] adding namespace took 6.295034ms I0419 18:45:06.792035 3645 namespace.go:102] [openshift-backplane-lpsre] adding namespace took 6.282274ms W0419 18:45:06.792484 3645 base_network_controller_namespace.go:410] Failed to get IPs for pod openshift-multus/network-metrics-daemon-l99ph: pod openshift-multus/network-metrics-daemon-l99ph: no pod IPs found I0419 18:45:06.792531 3645 namespace.go:102] [openshift-image-registry] adding namespace took 5.942163ms I0419 18:45:06.792958 3645 namespace.go:102] [openshift-multus] adding namespace took 5.947769ms I0419 18:45:06.793387 3645 namespace.go:102] [openshift-backplane-managed-scripts] adding namespace took 5.961558ms I0419 18:45:06.793741 3645 namespace.go:102] [openshift-authentication] adding namespace took 3.89039ms I0419 18:45:06.794157 3645 namespace.go:102] [dedicated-admin] adding namespace took 3.952446ms I0419 18:45:06.794501 3645 namespace.go:102] [openshift-kube-scheduler] adding namespace took 3.780863ms I0419 18:45:06.794912 3645 namespace.go:102] [openshift-network-operator] adding namespace took 3.766953ms I0419 18:45:06.795102 3645 master.go:548] Adding or Updating local node "ip-10-0-2-180.ec2.internal" for network "default" I0419 18:45:06.795857 3645 zone_ic_handler.go:195] Time taken to create transit switch: 252.542µs I0419 18:45:06.795871 3645 zone_ic_handler.go:277] Creating interconnect resources for remote zone node ip-10-0-0-39.ec2.internal for the network default E0419 18:45:06.796391 3645 obj_retry.go:575] Failed to create *v1.Node ip-10-0-1-170.ec2.internal, error: adding or updating remote node IC resources ip-10-0-1-170.ec2.internal failed, err - ensuring transit switch for remote zone node ip-10-0-1-170.ec2.internal for the network default failed : err - failed to create/update transit switch transit_switch: error in transact with ops [{Op:wait Table:Logical_Switch Row:map[] Rows:[map[name:transit_switch]] Columns:[name] Mutations:[] Timeout:0xc007ae9b18 Where:[where column _uuid == {ee5b8718-33f9-4c64-8fe8-e59283459f9b}] Until:!= Durable: Comment: Lock: UUID: UUIDName:} {Op:insert Table:Logical_Switch Row:map[name:transit_switch other_config:{GoMap:map[interconn-ts:transit_switch mcast_flood_unregistered:true mcast_querier:false mcast_snoop:true requested-tnl-key:16711683]}] Rows:[] Columns:[] Mutations:[] Timeout: Where:[] Until: Durable: Comment: Lock: UUID: UUIDName:u3971519239}] results [{Count:0 Error:timed out Details:"where" clause test failed UUID:{GoUUID:} Rows:[]} {Count:0 Error: Details: UUID:{GoUUID:} Rows:[]}] and errors [timed out: "where" clause test failed]: 1 ovsdb operations failed I0419 18:45:06.796511 3645 event.go:377] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"ip-10-0-1-170.ec2.internal", UID:"16b48080-7011-477d-bbe6-39c0adb9dc1a", APIVersion:"v1", ResourceVersion:"9181", FieldPath:""}): type: 'Warning' reason: 'ErrorAddingResource' adding or updating remote node IC resources ip-10-0-1-170.ec2.internal failed, err - ensuring transit switch for remote zone node ip-10-0-1-170.ec2.internal for the network default failed : err - failed to create/update transit switch transit_switch: error in transact with ops [{Op:wait Table:Logical_Switch Row:map[] Rows:[map[name:transit_switch]] Columns:[name] Mutations:[] Timeout:0xc007ae9b18 Where:[where column _uuid == {ee5b8718-33f9-4c64-8fe8-e59283459f9b}] Until:!= Durable: Comment: Lock: UUID: UUIDName:} {Op:insert Table:Logical_Switch Row:map[name:transit_switch other_config:{GoMap:map[interconn-ts:transit_switch mcast_flood_unregistered:true mcast_querier:false mcast_snoop:true requested-tnl-key:16711683]}] Rows:[] Columns:[] Mutations:[] Timeout: Where:[] Until: Durable: Comment: Lock: UUID: UUIDName:u3971519239}] results [{Count:0 Error:timed out Details:"where" clause test failed UUID:{GoUUID:} Rows:[]} {Count:0 Error: Details: UUID:{GoUUID:} Rows:[]}] and errors [timed out: "where" clause test failed]: 1 ovsdb operations failed I0419 18:45:06.796994 3645 zone_ic_handler.go:282] Creating Interconnect resources for node "ip-10-0-0-39.ec2.internal" on network "default" took: 1.397615ms W0419 18:45:06.803352 3645 gateway.go:207] Unable to fetch podIPs for pod openshift-network-diagnostics/network-check-target-p9ldg: pod openshift-network-diagnostics/network-check-target-p9ldg: no pod IPs found W0419 18:45:06.803372 3645 gateway.go:207] Unable to fetch podIPs for pod openshift-multus/network-metrics-daemon-l99ph: pod openshift-multus/network-metrics-daemon-l99ph: no pod IPs found I0419 18:45:06.804264 3645 zone_ic_handler.go:203] Creating interconnect resources for local zone node ip-10-0-2-180.ec2.internal for the network default I0419 18:45:06.804426 3645 zone_ic_handler.go:195] Time taken to create transit switch: 151.545µs I0419 18:45:06.805285 3645 services_controller.go:192] Starting controller ovn-lb-controller for network=default I0419 18:45:06.805309 3645 services_controller.go:200] Waiting for node tracker handler to sync for network=default I0419 18:45:06.805322 3645 shared_informer.go:349] "Waiting for caches to sync" controller="node-tracker-controller" I0419 18:45:06.805329 3645 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-0-39.ec2.internal in network "default" I0419 18:45:06.805754 3645 node_tracker.go:169] Node ip-10-0-0-39.ec2.internal switch + router changed, syncing services in network "default" I0419 18:45:06.805772 3645 services_controller.go:574] Full service sync requested for network=default I0419 18:45:06.805781 3645 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-1-170.ec2.internal in network "default" I0419 18:45:06.805807 3645 node_tracker.go:169] Node ip-10-0-1-170.ec2.internal switch + router changed, syncing services in network "default" I0419 18:45:06.805815 3645 services_controller.go:574] Full service sync requested for network=default I0419 18:45:06.805820 3645 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-2-180.ec2.internal in network "default" I0419 18:45:06.805841 3645 node_tracker.go:169] Node ip-10-0-2-180.ec2.internal switch + router changed, syncing services in network "default" I0419 18:45:06.805848 3645 services_controller.go:574] Full service sync requested for network=default I0419 18:45:06.905386 3645 shared_informer.go:356] "Caches are synced" controller="node-tracker-controller" I0419 18:45:06.905429 3645 services_controller.go:208] Setting up event handlers for services for network=default I0419 18:45:06.905526 3645 services_controller.go:218] Setting up event handlers for endpoint slices for network=default I0419 18:45:06.905573 3645 services_controller.go:233] Waiting for service and endpoint handlers to sync for network=default I0419 18:45:06.905588 3645 shared_informer.go:349] "Waiting for caches to sync" controller="ovn-lb-controller" I0419 18:45:07.006167 3645 shared_informer.go:356] "Caches are synced" controller="ovn-lb-controller" I0419 18:45:07.006189 3645 repair.go:60] Starting repairing loop for services I0419 18:45:07.006284 3645 repair.go:131] Deleted 0 stale service LBs I0419 18:45:07.006312 3645 repair.go:137] Deleted 0 stale Chassis Template Vars I0419 18:45:07.006338 3645 repair.go:62] Finished repairing loop for services: 151.804µs I0419 18:45:07.006372 3645 services_controller.go:363] Controller cache of 0 load balancers initialized for 0 services for network=default I0419 18:45:07.006386 3645 services_controller.go:254] Starting workers for network=default I0419 18:45:07.006744 3645 base_network_controller_pods.go:487] [default/openshift-multus/network-metrics-daemon-l99ph] creating logical port openshift-multus_network-metrics-daemon-l99ph for pod on switch ip-10-0-2-180.ec2.internal I0419 18:45:07.006831 3645 obj_retry.go:498] Detected object openshift-deployment-validation-operator/ca5b21c3869491e79768d2cfaf75f947507507b9e3b372dfa4bf565390zxp2c of type *v1.Pod in terminal state (e.g. completed) during add event: will remove it I0419 18:45:07.006872 3645 kube.go:256] Updating pod openshift-multus/network-metrics-daemon-l99ph I0419 18:45:07.006887 3645 base_network_controller_pods.go:487] [default/openshift-network-diagnostics/network-check-target-p9ldg] creating logical port openshift-network-diagnostics_network-check-target-p9ldg for pod on switch ip-10-0-2-180.ec2.internal I0419 18:45:07.006965 3645 kube.go:256] Updating pod openshift-network-diagnostics/network-check-target-p9ldg I0419 18:45:07.008094 3645 obj_retry.go:498] Detected object kube-system/ecr-credential-refresh-initial-ppdss of type *v1.Pod in terminal state (e.g. completed) during add event: will remove it I0419 18:45:07.027367 3645 pod.go:62] [openshift-network-diagnostics/network-check-target-p9ldg] pod update took 20.429022ms I0419 18:45:07.027392 3645 base_network_controller_pods.go:951] [default/openshift-network-diagnostics/network-check-target-p9ldg] addLogicalPort annotation time took 20.455775ms I0419 18:45:07.028545 3645 pods.go:271] [openshift-network-diagnostics/network-check-target-p9ldg] addLogicalPort took 21.664098ms, libovsdb time 875.661µs I0419 18:45:07.028648 3645 pod.go:62] [openshift-multus/network-metrics-daemon-l99ph] pod update took 21.830741ms I0419 18:45:07.028661 3645 base_network_controller_pods.go:951] [default/openshift-multus/network-metrics-daemon-l99ph] addLogicalPort annotation time took 21.849881ms I0419 18:45:07.029461 3645 pods.go:271] [openshift-multus/network-metrics-daemon-l99ph] addLogicalPort took 22.726145ms, libovsdb time 594.389µs I0419 18:45:07.029970 3645 admin_network_policy_controller.go:227] Starting controller default-network-controller I0419 18:45:07.029991 3645 shared_informer.go:349] "Waiting for caches to sync" controller="default-network-controller" I0419 18:45:07.030003 3645 shared_informer.go:356] "Caches are synced" controller="default-network-controller" I0419 18:45:07.030008 3645 admin_network_policy_controller.go:237] Repairing Admin Network Policies I0419 18:45:07.030189 3645 repair.go:29] Repairing admin network policies took 176.713µs I0419 18:45:07.030295 3645 repair.go:92] Repairing baseline admin network policies took 97.863µs I0419 18:45:07.030841 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-egress-to-api-server for network default I0419 18:45:07.030858 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-egress-to-api-server added to peer address sets [] I0419 18:45:07.032691 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-egress-to-api-server resources completed, update namespace loglevel I0419 18:45:07.032727 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-version/default-deny for network default I0419 18:45:07.032735 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-version/default-deny added to peer address sets [] I0419 18:45:07.034032 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-version/default-deny resources completed, update namespace loglevel I0419 18:45:07.034051 3645 base_network_controller_policy.go:1179] Adding network policy openshift-marketplace/marketplace-operator for network default I0419 18:45:07.034062 3645 base_network_controller_policy.go:1054] Policy openshift-marketplace/marketplace-operator added to peer address sets [] I0419 18:45:07.036110 3645 base_network_controller_policy.go:1232] Create network policy openshift-marketplace/marketplace-operator resources completed, update namespace loglevel I0419 18:45:07.036130 3645 base_network_controller_policy.go:1179] Adding network policy openshift-operators/default-allow-all for network default I0419 18:45:07.036138 3645 base_network_controller_policy.go:1054] Policy openshift-operators/default-allow-all added to peer address sets [] I0419 18:45:07.037970 3645 base_network_controller_policy.go:1232] Create network policy openshift-operators/default-allow-all resources completed, update namespace loglevel I0419 18:45:07.037989 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/allow-all-egress for network default I0419 18:45:07.037998 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/allow-all-egress added to peer address sets [] I0419 18:45:07.039576 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/allow-all-egress resources completed, update namespace loglevel I0419 18:45:07.039594 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/default-deny for network default I0419 18:45:07.039601 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/default-deny added to peer address sets [] I0419 18:45:07.039888 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/default-deny resources completed, update namespace loglevel I0419 18:45:07.039906 3645 base_network_controller_policy.go:1179] Adding network policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-grpc-server for network default I0419 18:45:07.039915 3645 base_network_controller_policy.go:1054] Policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-grpc-server added to peer address sets [] I0419 18:45:07.041350 3645 base_network_controller_policy.go:1232] Create network policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-grpc-server resources completed, update namespace loglevel I0419 18:45:07.041367 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/allow-ingress-pprof for network default I0419 18:45:07.041375 3645 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/allow-ingress-pprof added to peer address sets [] I0419 18:45:07.043145 3645 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/allow-ingress-pprof resources completed, update namespace loglevel I0419 18:45:07.043163 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/allow-egress-to-api-server for network default I0419 18:45:07.043172 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/allow-egress-to-api-server added to peer address sets [] I0419 18:45:07.043617 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/allow-egress-to-api-server resources completed, update namespace loglevel I0419 18:45:07.043636 3645 base_network_controller_policy.go:1179] Adding network policy openshift-deployment-validation-operator/allow-from-openshift-insights for network default I0419 18:45:07.043645 3645 base_network_controller_policy.go:1054] Policy openshift-deployment-validation-operator/allow-from-openshift-insights added to peer address sets [] I0419 18:45:07.044559 3645 obj_retry.go:498] Detected object openshift-deployment-validation-operator/ca5b21c3869491e79768d2cfaf75f947507507b9e3b372dfa4bf565390zxp2c of type *factory.localPodSelector in terminal state (e.g. completed) during add event: will remove it I0419 18:45:07.044572 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-deployment-validation-operator/allow-from-openshift-insights to delete 1 local pods... I0419 18:45:07.044591 3645 base_network_controller_policy.go:1232] Create network policy openshift-deployment-validation-operator/allow-from-openshift-insights resources completed, update namespace loglevel I0419 18:45:07.044604 3645 base_network_controller_policy.go:1179] Adding network policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-unpack-bundles for network default I0419 18:45:07.044613 3645 base_network_controller_policy.go:1054] Policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-unpack-bundles added to peer address sets [] I0419 18:45:07.045462 3645 obj_retry.go:498] Detected object openshift-deployment-validation-operator/ca5b21c3869491e79768d2cfaf75f947507507b9e3b372dfa4bf565390zxp2c of type *factory.localPodSelector in terminal state (e.g. completed) during add event: will remove it I0419 18:45:07.045476 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-deployment-validation-operator/deployment-validation-operator-catalog-unpack-bundles to delete 1 local pods... I0419 18:45:07.045491 3645 base_network_controller_policy.go:1232] Create network policy openshift-deployment-validation-operator/deployment-validation-operator-catalog-unpack-bundles resources completed, update namespace loglevel I0419 18:45:07.045504 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/allow-ingress-webhook for network default I0419 18:45:07.045513 3645 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/allow-ingress-webhook added to peer address sets [] I0419 18:45:07.045974 3645 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/allow-ingress-webhook resources completed, update namespace loglevel I0419 18:45:07.045991 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/allow-to-dns for network default I0419 18:45:07.046710 3645 pod_selector_address_set.go:213] Created shared address set for pod selector LS{ML:{kubernetes.io/metadata.name: openshift-dns,},}_LS{ML:{dns.operator.openshift.io/daemonset-dns: default,},} I0419 18:45:07.046723 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/allow-to-dns added to peer address sets [LS{ML:{kubernetes.io/metadata.name: openshift-dns,},}_LS{ML:{dns.operator.openshift.io/daemonset-dns: default,},}] I0419 18:45:07.047432 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/allow-to-dns resources completed, update namespace loglevel I0419 18:45:07.047465 3645 base_network_controller_policy.go:1179] Adding network policy openshift-deployment-validation-operator/allow-from-openshift-olm for network default I0419 18:45:07.047482 3645 base_network_controller_policy.go:1054] Policy openshift-deployment-validation-operator/allow-from-openshift-olm added to peer address sets [] I0419 18:45:07.048011 3645 base_network_controller_policy.go:1232] Create network policy openshift-deployment-validation-operator/allow-from-openshift-olm resources completed, update namespace loglevel I0419 18:45:07.048028 3645 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/packageserver for network default I0419 18:45:07.048039 3645 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/packageserver added to peer address sets [] I0419 18:45:07.050164 3645 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/packageserver resources completed, update namespace loglevel I0419 18:45:07.050182 3645 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/olm-operator for network default I0419 18:45:07.050191 3645 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/olm-operator added to peer address sets [] I0419 18:45:07.051145 3645 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/olm-operator resources completed, update namespace loglevel I0419 18:45:07.051163 3645 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/package-server-manager for network default I0419 18:45:07.051172 3645 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/package-server-manager added to peer address sets [] I0419 18:45:07.052093 3645 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/package-server-manager resources completed, update namespace loglevel I0419 18:45:07.052115 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/default-deny for network default I0419 18:45:07.052124 3645 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/default-deny added to peer address sets [] I0419 18:45:07.052429 3645 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/default-deny resources completed, update namespace loglevel I0419 18:45:07.052445 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-ingress-to-metrics-range for network default I0419 18:45:07.052453 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-ingress-to-metrics-range added to peer address sets [] I0419 18:45:07.052991 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-ingress-to-metrics-range resources completed, update namespace loglevel I0419 18:45:07.053007 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-ingress-to-operator-metrics-range for network default I0419 18:45:07.053016 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-ingress-to-operator-metrics-range added to peer address sets [] I0419 18:45:07.053461 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-ingress-to-operator-metrics-range resources completed, update namespace loglevel I0419 18:45:07.053474 3645 base_network_controller_policy.go:1179] Adding network policy openshift-marketplace/unpack-bundles for network default I0419 18:45:07.053483 3645 base_network_controller_policy.go:1054] Policy openshift-marketplace/unpack-bundles added to peer address sets [] I0419 18:45:07.054361 3645 base_network_controller_policy.go:1232] Create network policy openshift-marketplace/unpack-bundles resources completed, update namespace loglevel I0419 18:45:07.054386 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-to-dns for network default I0419 18:45:07.054434 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-to-dns added to peer address sets [LS{ML:{kubernetes.io/metadata.name: openshift-dns,},}_LS{ML:{dns.operator.openshift.io/daemonset-dns: default,},}] I0419 18:45:07.055155 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-to-dns resources completed, update namespace loglevel I0419 18:45:07.055174 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-storage-operator/allow-ingress-to-operator-metrics for network default I0419 18:45:07.055182 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-storage-operator/allow-ingress-to-operator-metrics added to peer address sets [] I0419 18:45:07.055733 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-storage-operator/allow-ingress-to-operator-metrics resources completed, update namespace loglevel I0419 18:45:07.055751 3645 base_network_controller_policy.go:1179] Adding network policy openshift-marketplace/default-deny-all for network default I0419 18:45:07.055757 3645 base_network_controller_policy.go:1054] Policy openshift-marketplace/default-deny-all added to peer address sets [] I0419 18:45:07.056058 3645 base_network_controller_policy.go:1232] Create network policy openshift-marketplace/default-deny-all resources completed, update namespace loglevel I0419 18:45:07.056075 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/allow-egress for network default I0419 18:45:07.056083 3645 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/allow-egress added to peer address sets [] I0419 18:45:07.056812 3645 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/allow-egress resources completed, update namespace loglevel I0419 18:45:07.056830 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cloud-credential-operator/allow-ingress-metrics for network default I0419 18:45:07.056837 3645 base_network_controller_policy.go:1054] Policy openshift-cloud-credential-operator/allow-ingress-metrics added to peer address sets [] I0419 18:45:07.057345 3645 base_network_controller_policy.go:1232] Create network policy openshift-cloud-credential-operator/allow-ingress-metrics resources completed, update namespace loglevel I0419 18:45:07.057362 3645 base_network_controller_policy.go:1179] Adding network policy openshift-cluster-csi-drivers/allow-all-egress for network default I0419 18:45:07.057369 3645 base_network_controller_policy.go:1054] Policy openshift-cluster-csi-drivers/allow-all-egress added to peer address sets [] I0419 18:45:07.057937 3645 base_network_controller_policy.go:1232] Create network policy openshift-cluster-csi-drivers/allow-all-egress resources completed, update namespace loglevel I0419 18:45:07.057954 3645 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/catalog-operator for network default I0419 18:45:07.057963 3645 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/catalog-operator added to peer address sets [] I0419 18:45:07.059220 3645 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/catalog-operator resources completed, update namespace loglevel I0419 18:45:07.059239 3645 base_network_controller_policy.go:1179] Adding network policy openshift-operator-lifecycle-manager/default-deny-all-traffic for network default I0419 18:45:07.059248 3645 base_network_controller_policy.go:1054] Policy openshift-operator-lifecycle-manager/default-deny-all-traffic added to peer address sets [] I0419 18:45:07.059576 3645 base_network_controller_policy.go:1232] Create network policy openshift-operator-lifecycle-manager/default-deny-all-traffic resources completed, update namespace loglevel I0419 18:45:07.059793 3645 obj_retry.go:498] Detected object openshift-deployment-validation-operator/ca5b21c3869491e79768d2cfaf75f947507507b9e3b372dfa4bf565390zxp2c of type *factory.egressIPPod in terminal state (e.g. completed) during add event: will remove it I0419 18:45:07.059807 3645 obj_retry.go:498] Detected object kube-system/ecr-credential-refresh-initial-ppdss of type *factory.egressIPPod in terminal state (e.g. completed) during add event: will remove it I0419 18:45:07.066225 3645 egressfirewall.go:374] Starting EgressFirewall controller I0419 18:45:07.066259 3645 controller.go:133] Adding controller egress-firewall-controller event handlers I0419 18:45:07.066308 3645 controller.go:133] Adding controller egress-firewall-controller-node event handlers I0419 18:45:07.066337 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egress-firewall-controller-node" I0419 18:45:07.066350 3645 shared_informer.go:356] "Caches are synced" controller="egress-firewall-controller-node" I0419 18:45:07.066310 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egress-firewall-controller" I0419 18:45:07.066359 3645 shared_informer.go:356] "Caches are synced" controller="egress-firewall-controller" I0419 18:45:07.066656 3645 controller.go:157] Starting controller egress-firewall-controller with 1 workers I0419 18:45:07.066672 3645 controller.go:157] Starting controller egress-firewall-controller-node with 1 workers I0419 18:45:07.066678 3645 controller.go:157] Starting controller egress-firewall-controller-NAD with 1 workers I0419 18:45:07.066687 3645 egressqos.go:193] Setting up event handlers for EgressQoS I0419 18:45:07.066697 3645 egressfirewall.go:723] Syncing node "ip-10-0-0-39.ec2.internal" for egress firewall I0419 18:45:07.066711 3645 egressfirewall.go:723] Syncing node "ip-10-0-1-170.ec2.internal" for egress firewall I0419 18:45:07.066714 3645 egressfirewall.go:723] Syncing node "ip-10-0-2-180.ec2.internal" for egress firewall I0419 18:45:07.066878 3645 egressqos.go:245] Starting EgressQoS Controller I0419 18:45:07.066890 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressqosnodes" I0419 18:45:07.066898 3645 shared_informer.go:356] "Caches are synced" controller="egressqosnodes" I0419 18:45:07.066904 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressqospods" I0419 18:45:07.066908 3645 shared_informer.go:356] "Caches are synced" controller="egressqospods" I0419 18:45:07.066913 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressqos" I0419 18:45:07.066917 3645 shared_informer.go:356] "Caches are synced" controller="egressqos" I0419 18:45:07.066921 3645 egressqos.go:259] Repairing EgressQoSes I0419 18:45:07.066925 3645 egressqos.go:399] Starting repairing loop for egressqos I0419 18:45:07.067008 3645 egressqos.go:401] Finished repairing loop for egressqos: 82.507µs I0419 18:45:07.067041 3645 egressservice_zone.go:132] Setting up event handlers for Egress Services I0419 18:45:07.067124 3645 egressqos.go:1007] Processing sync for EgressQoS node ip-10-0-2-180.ec2.internal I0419 18:45:07.067165 3645 egressqos.go:1010] Finished syncing EgressQoS node ip-10-0-2-180.ec2.internal : 42.81µs I0419 18:45:07.067204 3645 egressservice_zone.go:210] Starting Egress Services Controller I0419 18:45:07.067217 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices" I0419 18:45:07.067225 3645 shared_informer.go:356] "Caches are synced" controller="egressservices" I0419 18:45:07.067233 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_services" I0419 18:45:07.067239 3645 shared_informer.go:356] "Caches are synced" controller="egressservices_services" I0419 18:45:07.067247 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_endpointslices" I0419 18:45:07.067254 3645 shared_informer.go:356] "Caches are synced" controller="egressservices_endpointslices" I0419 18:45:07.067261 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_nodes" I0419 18:45:07.067265 3645 shared_informer.go:356] "Caches are synced" controller="egressservices_nodes" I0419 18:45:07.067269 3645 egressservice_zone.go:228] Repairing Egress Services I0419 18:45:07.067566 3645 master_controller.go:88] Starting Admin Policy Based Route Controller I0419 18:45:07.067577 3645 external_controller.go:278] Starting Admin Policy Based Route Controller I0419 18:45:07.067640 3645 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-0-39.ec2.internal I0419 18:45:07.067658 3645 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-0-39.ec2.internal: 18.614µs I0419 18:45:07.067668 3645 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-1-170.ec2.internal I0419 18:45:07.067675 3645 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-1-170.ec2.internal: 7.728µs I0419 18:45:07.067679 3645 default_network_controller.go:585] Completing all the Watchers took 307.747165ms I0419 18:45:07.067683 3645 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-2-180.ec2.internal I0419 18:45:07.067688 3645 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-2-180.ec2.internal: 6.469µs I0419 18:45:07.067689 3645 default_network_controller.go:589] Starting unidling controllers I0419 18:45:07.067709 3645 unidle.go:46] Registering OVN SB ControllerEvent handler I0419 18:45:07.067720 3645 unidle.go:63] Populating Initial ContollerEvent events I0419 18:45:07.067733 3645 unidle.go:79] Setting up event handlers for services I0419 18:45:07.067827 3645 ovnkube.go:545] Waiting for OVN northbound database changes to be processed by ovn-controller I0419 18:45:07.155032 3645 default_node_network_controller.go:519] Node connection status = connected I0419 18:45:07.655063 3645 default_node_network_controller.go:519] Node connection status = connected I0419 18:45:08.155205 3645 default_node_network_controller.go:519] Node connection status = connected I0419 18:45:08.413336 3645 ovnkube.go:550] Finished waiting for OVN northbound database changes to be processed by ovn-controller I0419 18:45:08.655341 3645 default_node_network_controller.go:519] Node connection status = connected I0419 18:45:08.673855 3645 bridgeconfig.go:537] Gateway is ready I0419 18:45:08.673874 3645 gateway_shared_intf.go:1733] Creating Gateway Openflow Manager I0419 18:45:08.686012 3645 node_ip_handler_linux.go:247] Node primary address changed to 10.0.2.180. Updating OVN encap IP. I0419 18:45:08.691601 3645 node_ip_handler_linux.go:542] Will not update encap IP 10.0.2.180 - it is already configured I0419 18:45:08.691617 3645 node_ip_handler_linux.go:485] Node address changed to map[10.0.2.180/24:{} 172.20.0.1/32:{}]. Updating annotations. I0419 18:45:08.691826 3645 kube.go:131] Setting annotations map[k8s.ovn.org/host-cidrs:["10.0.2.180/24","172.20.0.1/32"] k8s.ovn.org/l3-gateway-config:{"default":{"mode":"shared","bridge-id":"br-ex","interface-id":"br-ex_ip-10-0-2-180.ec2.internal","mac-address":"0e:31:c5:da:9b:db","ip-addresses":["10.0.2.180/24"],"ip-address":"10.0.2.180/24","next-hops":["10.0.2.1"],"next-hop":"10.0.2.1","node-port-enable":"true","vlan-id":"0"}} k8s.ovn.org/node-chassis-id:b4127fe2-9fdb-4ee3-bc09-e67f72a2922a k8s.ovn.org/node-primary-ifaddr:{"ipv4":"10.0.2.180/24"}] on node ip-10-0-2-180.ec2.internal I0419 18:45:08.706922 3645 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-2-180.ec2.internal I0419 18:45:08.706939 3645 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-2-180.ec2.internal: 26.234µs I0419 18:45:08.707011 3645 node_tracker.go:208] Processing possible switch / router updates for node ip-10-0-2-180.ec2.internal in network "default" I0419 18:45:08.707013 3645 egressfirewall.go:723] Syncing node "ip-10-0-2-180.ec2.internal" for egress firewall I0419 18:45:08.707040 3645 master.go:548] Adding or Updating local node "ip-10-0-2-180.ec2.internal" for network "default" I0419 18:45:08.707056 3645 node_tracker.go:169] Node ip-10-0-2-180.ec2.internal switch + router changed, syncing services in network "default" I0419 18:45:08.707069 3645 services_controller.go:574] Full service sync requested for network=default I0419 18:45:08.707515 3645 gateway_shared_intf.go:2023] Setting OVN Masquerade route with source: 10.0.2.180 I0419 18:45:08.707630 3645 kube.go:131] Setting annotations map[k8s.ovn.org/node-masquerade-subnet:{"ipv4":"169.254.0.0/17","ipv6":"fd69::/112"}] on node ip-10-0-2-180.ec2.internal I0419 18:45:08.711897 3645 default_network_controller.go:1084] Egress IP detected IP address change for node ip-10-0-2-180.ec2.internal. Updating no re-route policies I0419 18:45:08.711956 3645 obj_retry.go:334] Retry object setup: *v1.Pod kube-system/kube-apiserver-proxy-ip-10-0-2-180.ec2.internal I0419 18:45:08.711965 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xvfsf I0419 18:45:08.711968 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-cluster-node-tuning-operator/tuned-qfjkd I0419 18:45:08.711978 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xvfsf I0419 18:45:08.711977 3645 obj_retry.go:334] Retry object setup: *v1.Pod kube-system/konnectivity-agent-kzscw I0419 18:45:08.711973 3645 obj_retry.go:398] Adding new object: *v1.Pod kube-system/kube-apiserver-proxy-ip-10-0-2-180.ec2.internal I0419 18:45:08.711983 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-multus/multus-r4j42 I0419 18:45:08.711991 3645 obj_retry.go:398] Adding new object: *v1.Pod kube-system/konnectivity-agent-kzscw I0419 18:45:08.711996 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-l99ph I0419 18:45:08.711958 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-dns/node-resolver-bxzvc I0419 18:45:08.712002 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-l99ph I0419 18:45:08.712001 3645 obj_retry.go:418] Retry successful for *v1.Pod kube-system/kube-apiserver-proxy-ip-10-0-2-180.ec2.internal after 0 failed attempt(s) I0419 18:45:08.712000 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-42ccl I0419 18:45:08.712008 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-c9xmp I0419 18:45:08.712013 3645 obj_retry.go:418] Retry successful for *v1.Pod kube-system/konnectivity-agent-kzscw after 0 failed attempt(s) I0419 18:45:08.712015 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-42ccl I0419 18:45:08.711982 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-cluster-node-tuning-operator/tuned-qfjkd I0419 18:45:08.712021 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-rcwhk I0419 18:45:08.712020 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-p9ldg I0419 18:45:08.712030 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-rcwhk I0419 18:45:08.712032 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-42ccl after 0 failed attempt(s) I0419 18:45:08.712033 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-cluster-node-tuning-operator/tuned-qfjkd after 0 failed attempt(s) I0419 18:45:08.712037 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-p9ldg I0419 18:45:08.712041 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-rcwhk after 0 failed attempt(s) I0419 18:45:08.712068 3645 base_network_controller_pods.go:487] [default/openshift-network-diagnostics/network-check-target-p9ldg] creating logical port openshift-network-diagnostics_network-check-target-p9ldg for pod on switch ip-10-0-2-180.ec2.internal I0419 18:45:08.711996 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-multus/multus-r4j42 I0419 18:45:08.712114 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-multus/multus-r4j42 after 0 failed attempt(s) I0419 18:45:08.711980 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-image-registry/node-ca-ff4db I0419 18:45:08.712127 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-image-registry/node-ca-ff4db I0419 18:45:08.712138 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-image-registry/node-ca-ff4db after 0 failed attempt(s) I0419 18:45:08.712005 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-dns/node-resolver-bxzvc I0419 18:45:08.712160 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-dns/node-resolver-bxzvc after 0 failed attempt(s) I0419 18:45:08.711988 3645 obj_retry.go:334] Retry object setup: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-2-180.ec2.internal I0419 18:45:08.712175 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-2-180.ec2.internal I0419 18:45:08.712188 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-2-180.ec2.internal after 0 failed attempt(s) I0419 18:45:08.712003 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xvfsf after 0 failed attempt(s) I0419 18:45:08.712021 3645 obj_retry.go:398] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-c9xmp I0419 18:45:08.712209 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-c9xmp after 0 failed attempt(s) I0419 18:45:08.712026 3645 base_network_controller_pods.go:487] [default/openshift-multus/network-metrics-daemon-l99ph] creating logical port openshift-multus_network-metrics-daemon-l99ph for pod on switch ip-10-0-2-180.ec2.internal I0419 18:45:08.712756 3645 pods.go:271] [openshift-network-diagnostics/network-check-target-p9ldg] addLogicalPort took 697.192µs, libovsdb time 301.722µs I0419 18:45:08.712769 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-target-p9ldg after 0 failed attempt(s) I0419 18:45:08.712952 3645 pods.go:271] [openshift-multus/network-metrics-daemon-l99ph] addLogicalPort took 932.368µs, libovsdb time 309.966µs I0419 18:45:08.712964 3645 obj_retry.go:418] Retry successful for *v1.Pod openshift-multus/network-metrics-daemon-l99ph after 0 failed attempt(s) I0419 18:45:08.722890 3645 gateway_shared_intf.go:1799] Creating Gateway Node Port Watcher I0419 18:45:08.735874 3645 iptables.go:108] Creating table: mangle chain: OUTPUT I0419 18:45:08.739785 3645 iptables.go:108] Creating table: nat chain: OUTPUT I0419 18:45:08.771044 3645 iptables.go:108] Creating table: nat chain: PREROUTING I0419 18:45:08.821992 3645 iptables.go:108] Creating table: filter chain: FORWARD I0419 18:45:08.840130 3645 gateway_shared_intf.go:2124] Ensuring IP Neighbor entry for: 169.254.0.1 W0419 18:45:08.840249 3645 gateway_shared_intf.go:2130] Failed to remove IP neighbor entry for ip 169.254.0.1, on iface br-ex: failed to delete neighbour entry 169.254.0.1 : no such file or directory I0419 18:45:08.840300 3645 gateway_shared_intf.go:2124] Ensuring IP Neighbor entry for: 169.254.0.4 W0419 18:45:08.840386 3645 gateway_shared_intf.go:2130] Failed to remove IP neighbor entry for ip 169.254.0.4, on iface br-ex: failed to delete neighbour entry 169.254.0.4 : no such file or directory I0419 18:45:08.840475 3645 gateway_init.go:325] Gateway and management port readiness took 2.190557837s I0419 18:45:08.840510 3645 controller.go:133] Adding controller [node-nad-controller NAD controller] event handlers I0419 18:45:08.840595 3645 shared_informer.go:349] "Waiting for caches to sync" controller="[node-nad-controller NAD controller]" I0419 18:45:08.840607 3645 shared_informer.go:356] "Caches are synced" controller="[node-nad-controller NAD controller]" I0419 18:45:08.840615 3645 nad_controller.go:541] [node-nad-controller NAD controller]: 1 NADs are missing the network ID annotation, fetching from nodes I0419 18:45:08.840673 3645 controller.go:157] Starting controller [node-nad-controller NAD controller] with 1 workers I0419 18:45:08.840765 3645 nad_controller.go:580] [node-nad-controller NAD controller]: finished syncing NAD openshift-ovn-kubernetes/default, took 63.007µs I0419 18:45:08.841343 3645 network_controller.go:373] [node-nad-controller network controller]: syncing all networks I0419 18:45:08.841368 3645 network_controller.go:402] [node-nad-controller network controller]: finished syncing network default, took 15.373µs I0419 18:45:08.841381 3645 network_controller.go:384] [node-nad-controller network controller]: finished syncing all networks. Time taken: 38.22µs I0419 18:45:08.841387 3645 controller.go:157] Starting controller [node-nad-controller network controller] with 1 workers I0419 18:45:08.841398 3645 nad_controller.go:379] [node-nad-controller NAD controller]: started I0419 18:45:08.841425 3645 default_node_network_controller.go:899] Starting the default node network controller I0419 18:45:08.841430 3645 gateway_init.go:335] Initializing Gateway Functionality for gateway Start I0419 18:45:08.841501 3645 network_controller.go:402] [node-nad-controller network controller]: finished syncing network default, took 12.409µs I0419 18:45:08.841593 3645 default_node_network_controller.go:1460] MTU (9001) of network interface br-ex is big enough to deal with Geneve header overhead (sum 8959). I0419 18:45:08.841602 3645 default_node_network_controller.go:935] Waiting for gateway and management port readiness... I0419 18:45:08.841626 3645 obj_retry_gateway.go:33] [newRetryFrameworkNodeWithParameters] g.watchFactory=&{0xc000c3c7c0 0xc0002d32d0 0xc0002d33b0 0xc0002d3420 0xc0002d3490 0xc0002d3500 0xc0002d35e0 0xc000d18190 0xc0002d3650 0xc0002d36c0 0xc000b2a280 0xc00651b340 0xc0002d3730 map[0x2ff9080:0xc0024207e0 0x2ff95c0:0xc002420960 0x2ffa2e0:0xc001795b60 0x2ffa580:0xc002420780 0x2ffad60:0xc002420840 0x2ffb000:0xc0024208a0 0x2ffb2a0:0xc002420900 0x2ffba80:0xc002420b40 0x2ffbd20:0xc002420ba0 0x3026980:0xc0014828a0 0x3026d20:0xc001482900 0x3027800:0xc000c47740 0x3029160:0xc001cde000 0x30298a0:0xc001795b00 0x302a720:0xc000cb2480] 0xc000168930 0} I0419 18:45:08.841683 3645 gateway.go:151] Starting gateway service sync I0419 18:45:08.842065 3645 gateway_iptables.go:474] Recreating iptables rules for table: nat, chain: OVN-KUBE-ITP I0419 18:45:08.844459 3645 gateway_iptables.go:474] Recreating iptables rules for table: nat, chain: OVN-KUBE-NODEPORT I0419 18:45:08.868066 3645 gateway_iptables.go:474] Recreating iptables rules for table: nat, chain: OVN-KUBE-EXTERNALIP I0419 18:45:08.869430 3645 gateway_iptables.go:474] Recreating iptables rules for table: nat, chain: OVN-KUBE-ETP I0419 18:45:08.870604 3645 gateway_iptables.go:474] Recreating iptables rules for table: mangle, chain: OVN-KUBE-ITP I0419 18:45:08.906306 3645 gateway.go:168] Gateway service sync done. Time taken: 64.608535ms I0419 18:45:08.906901 3645 healthcheck.go:145] Opening healthcheck "openshift-ingress/router-default" on port 30381 I0419 18:45:08.907089 3645 obj_retry_gateway.go:33] [newRetryFrameworkNodeWithParameters] g.watchFactory=&{0xc000c3c7c0 0xc0002d32d0 0xc0002d33b0 0xc0002d3420 0xc0002d3490 0xc0002d3500 0xc0002d35e0 0xc000d18190 0xc0002d3650 0xc0002d36c0 0xc000b2a280 0xc00651b340 0xc0002d3730 map[0x2ff9080:0xc0024207e0 0x2ff95c0:0xc002420960 0x2ffa2e0:0xc001795b60 0x2ffa580:0xc002420780 0x2ffad60:0xc002420840 0x2ffb000:0xc0024208a0 0x2ffb2a0:0xc002420900 0x2ffba80:0xc002420b40 0x2ffbd20:0xc002420ba0 0x3026980:0xc0014828a0 0x3026d20:0xc001482900 0x3027800:0xc000c47740 0x3029160:0xc001cde000 0x30298a0:0xc001795b00 0x302a720:0xc000cb2480] 0xc000168930 0} I0419 18:45:08.907519 3645 obj_retry_gateway.go:33] [newRetryFrameworkNodeWithParameters] g.watchFactory=&{0xc000c3c7c0 0xc0002d32d0 0xc0002d33b0 0xc0002d3420 0xc0002d3490 0xc0002d3500 0xc0002d35e0 0xc000d18190 0xc0002d3650 0xc0002d36c0 0xc000b2a280 0xc00651b340 0xc0002d3730 map[0x2ff9080:0xc0024207e0 0x2ff95c0:0xc002420960 0x2ffa2e0:0xc001795b60 0x2ffa580:0xc002420780 0x2ffad60:0xc002420840 0x2ffb000:0xc0024208a0 0x2ffb2a0:0xc002420900 0x2ffba80:0xc002420b40 0x2ffbd20:0xc002420ba0 0x3026980:0xc0014828a0 0x3026d20:0xc001482900 0x3027800:0xc000c47740 0x3029160:0xc001cde000 0x30298a0:0xc001795b00 0x302a720:0xc000cb2480] 0xc000168930 0} I0419 18:45:08.907564 3645 gateway.go:532] Reconciling gateway with updates I0419 18:45:08.913335 3645 gateway.go:339] Spawning Conntrack Rule Check Thread I0419 18:45:08.913433 3645 default_node_network_controller.go:944] Gateway and management port readiness took 71.821026ms I0419 18:45:08.913428 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-diagnostics/network-check-source I0419 18:45:08.913450 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-version/cluster-version-operator I0419 18:45:08.913459 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-diagnostics/network-check-source I0419 18:45:08.913464 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-version/cluster-version-operator I0419 18:45:08.913466 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-image-registry/image-registry I0419 18:45:08.913473 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio I0419 18:45:08.913456 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-server I0419 18:45:08.913485 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-server I0419 18:45:08.913486 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio I0419 18:45:08.913489 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node I0419 18:45:08.913498 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node I0419 18:45:08.913505 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node after 0 failed attempt(s) I0419 18:45:08.913501 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator I0419 18:45:08.913466 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-diagnostics/network-check-source after 0 failed attempt(s) I0419 18:45:08.913508 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-authentication-operator/metrics I0419 18:45:08.913528 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics I0419 18:45:08.913530 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator I0419 18:45:08.913528 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-dns-operator/metrics I0419 18:45:08.913535 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-operator/metrics I0419 18:45:08.913537 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress-canary/ingress-canary I0419 18:45:08.913546 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-dns-operator/metrics I0419 18:45:08.913544 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics I0419 18:45:08.913552 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress-canary/ingress-canary I0419 18:45:08.913559 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress/router-default I0419 18:45:08.913561 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver I0419 18:45:08.913565 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress/router-default I0419 18:45:08.913563 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon I0419 18:45:08.913572 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver I0419 18:45:08.913576 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics I0419 18:45:08.913577 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics I0419 18:45:08.913581 3645 port_claim.go:71] Svc openshift-ingress/router-default: port 32321 is already open, no action needed I0419 18:45:08.913561 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics I0419 18:45:08.913587 3645 port_claim.go:71] Svc openshift-ingress/router-default: port 32654 is already open, no action needed I0419 18:45:08.913591 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics I0419 18:45:08.913592 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-console-operator/metrics I0419 18:45:08.913594 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-diagnostics/network-check-target I0419 18:45:08.913594 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog I0419 18:45:08.913600 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller I0419 18:45:08.913603 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-insights/metrics I0419 18:45:08.913606 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-diagnostics/network-check-target I0419 18:45:08.913609 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-console-operator/metrics I0419 18:45:08.913612 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-insights/metrics I0419 18:45:08.913608 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog I0419 18:45:08.913617 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress/router-internal-default I0419 18:45:08.913622 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-samples-operator/metrics I0419 18:45:08.913624 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-console/networking-console-plugin I0419 18:45:08.913636 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-console/networking-console-plugin I0419 18:45:08.913636 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift-apiserver I0419 18:45:08.913647 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/packageserver I0419 18:45:08.913651 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress/router-internal-default I0419 18:45:08.913653 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics after 0 failed attempt(s) I0419 18:45:08.913656 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio after 0 failed attempt(s) I0419 18:45:08.913658 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-service-ca-operator/metrics I0419 18:45:08.913664 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-apiserver/check-endpoints I0419 18:45:08.913560 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-server after 0 failed attempt(s) I0419 18:45:08.913665 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress-operator/metrics I0419 18:45:08.913672 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-service-ca-operator/metrics I0419 18:45:08.913677 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-apiserver/check-endpoints I0419 18:45:08.913680 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics I0419 18:45:08.913686 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-config-operator/metrics I0419 18:45:08.913687 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress-operator/metrics I0419 18:45:08.913691 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics I0419 18:45:08.913695 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-config-operator/metrics I0419 18:45:08.913549 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service I0419 18:45:08.913637 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-samples-operator/metrics I0419 18:45:08.913709 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service I0419 18:45:08.913510 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-apiserver-operator/metrics I0419 18:45:08.913586 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/kubernetes I0419 18:45:08.913540 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator I0419 18:45:08.913478 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-image-registry/image-registry I0419 18:45:08.913725 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/kubernetes I0419 18:45:08.913731 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator after 0 failed attempt(s) I0419 18:45:08.913657 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/packageserver I0419 18:45:08.913647 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller I0419 18:45:08.913652 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift-apiserver I0419 18:45:08.913711 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-samples-operator/metrics after 0 failed attempt(s) I0419 18:45:08.913580 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver after 0 failed attempt(s) I0419 18:45:08.913439 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics I0419 18:45:08.913849 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics I0419 18:45:08.913660 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-console-operator/metrics after 0 failed attempt(s) I0419 18:45:08.913520 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator I0419 18:45:08.913870 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator after 0 failed attempt(s) I0419 18:45:08.913476 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics I0419 18:45:08.913879 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics I0419 18:45:08.913921 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift I0419 18:45:08.913594 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-dns-operator/metrics after 0 failed attempt(s) I0419 18:45:08.913934 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift I0419 18:45:08.913542 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-operator/metrics I0419 18:45:08.913940 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift after 0 failed attempt(s) I0419 18:45:08.913947 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-operator/metrics after 0 failed attempt(s) I0419 18:45:08.913542 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-version/cluster-version-operator after 0 failed attempt(s) I0419 18:45:08.913550 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift-oauth-apiserver I0419 18:45:08.913961 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift-oauth-apiserver I0419 18:45:08.913537 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-authentication-operator/metrics I0419 18:45:08.913578 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon I0419 18:45:08.913588 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics I0419 18:45:08.913584 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics I0419 18:45:08.913610 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-multus/network-metrics-service I0419 18:45:08.914035 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-multus/network-metrics-service I0419 18:45:08.914039 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-multus/network-metrics-service after 0 failed attempt(s) I0419 18:45:08.913667 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics I0419 18:45:08.914046 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics I0419 18:45:08.913614 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress-canary/ingress-canary after 0 failed attempt(s) I0419 18:45:08.913669 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-diagnostics/network-check-target after 0 failed attempt(s) I0419 18:45:08.913722 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-apiserver-operator/metrics I0419 18:45:08.913664 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-insights/metrics after 0 failed attempt(s) I0419 18:45:08.913523 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-dns/dns-default I0419 18:45:08.914170 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-dns/dns-default I0419 18:45:08.914764 3645 node_ip_handler_linux.go:160] Node IP manager is running W0419 18:45:08.952835 3645 port_linux.go:245] Missing or unable to find route entry for subnet 10.128.0.0/14 via gateway 10.130.0.1 on link ovn-k8s-mp0 with MTU: 0 W0419 18:45:08.953005 3645 port_linux.go:245] Missing or unable to find route entry for subnet 169.254.0.3/32 via gateway 10.130.0.1 on link ovn-k8s-mp0 with MTU: 0 W0419 18:45:08.953145 3645 port_linux.go:261] Missing arp entry for MAC/IP binding (0a:58:0a:82:00:01/10.130.0.1) on link ovn-k8s-mp0 I0419 18:45:08.961627 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress/router-default after 0 failed attempt(s) I0419 18:45:08.961641 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog after 0 failed attempt(s) I0419 18:45:08.961631 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics after 0 failed attempt(s) I0419 18:45:08.961655 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress/router-internal-default after 0 failed attempt(s) I0419 18:45:08.961663 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-service-ca-operator/metrics after 0 failed attempt(s) I0419 18:45:08.961672 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics after 0 failed attempt(s) I0419 18:45:08.961684 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-config-operator/metrics after 0 failed attempt(s) I0419 18:45:08.961695 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress-operator/metrics after 0 failed attempt(s) I0419 18:45:08.961726 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/packageserver after 0 failed attempt(s) I0419 18:45:08.961740 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service after 0 failed attempt(s) I0419 18:45:08.961751 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-apiserver/check-endpoints after 0 failed attempt(s) I0419 18:45:08.961759 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift-apiserver after 0 failed attempt(s) I0419 18:45:08.961768 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics after 0 failed attempt(s) I0419 18:45:08.961779 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift-oauth-apiserver after 0 failed attempt(s) I0419 18:45:08.961787 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-authentication-operator/metrics after 0 failed attempt(s) I0419 18:45:08.961794 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon after 0 failed attempt(s) I0419 18:45:08.961801 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics after 0 failed attempt(s) I0419 18:45:08.961809 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-apiserver-operator/metrics after 0 failed attempt(s) I0419 18:45:09.003229 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-dns/dns-default after 0 failed attempt(s) I0419 18:45:09.003255 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-console/networking-console-plugin after 0 failed attempt(s) I0419 18:45:09.003261 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics after 0 failed attempt(s) I0419 18:45:09.003269 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics after 0 failed attempt(s) I0419 18:45:09.003279 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics after 0 failed attempt(s) I0419 18:45:09.003290 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-image-registry/image-registry after 0 failed attempt(s) I0419 18:45:09.003298 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller after 0 failed attempt(s) I0419 18:45:09.043123 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/kubernetes after 0 failed attempt(s) I0419 18:45:09.550616 3645 namespace.go:142] [openshift-deployment-validation-operator] updating namespace I0419 18:45:09.563731 3645 namespace.go:142] [openshift-deployment-validation-operator] updating namespace I0419 18:45:10.716296 3645 default_node_network_controller.go:1380] Starting node controller node sync W0419 18:45:10.716293 3645 egressip_healthcheck.go:76] Health checking using insecure connection I0419 18:45:10.716428 3645 egressip_healthcheck.go:109] Starting Egress IP Health Server on 10.130.0.2:9107 I0419 18:45:10.728604 3645 default_node_network_controller.go:1422] Node controller node sync done. Time taken: 12.283436ms I0419 18:45:10.728750 3645 default_node_network_controller.go:1293] Adding remote node "ip-10-0-1-170.ec2.internal", IP: 10.0.1.170 to PMTUD blocking rules I0419 18:45:10.728750 3645 default_node_network_controller.go:1293] Adding remote node "ip-10-0-0-39.ec2.internal", IP: 10.0.0.39 to PMTUD blocking rules I0419 18:45:10.728766 3645 default_node_network_controller.go:1293] Adding remote node "ip-10-0-1-170.ec2.internal", IP: 172.20.0.1 to PMTUD blocking rules I0419 18:45:10.728767 3645 default_node_network_controller.go:1293] Adding remote node "ip-10-0-0-39.ec2.internal", IP: 172.20.0.1 to PMTUD blocking rules I0419 18:45:10.741126 3645 healthcheck_node.go:124] "Starting node proxy healthz server" address="0.0.0.0:10256" I0419 18:45:10.741431 3645 egressservice_node.go:93] Setting up event handlers for Egress Services I0419 18:45:10.741577 3645 egressservice_node.go:183] Starting Egress Services Controller I0419 18:45:10.741597 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices" I0419 18:45:10.741608 3645 shared_informer.go:356] "Caches are synced" controller="egressservices" I0419 18:45:10.741616 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_services" I0419 18:45:10.741623 3645 shared_informer.go:356] "Caches are synced" controller="egressservices_services" I0419 18:45:10.741632 3645 shared_informer.go:349] "Waiting for caches to sync" controller="egressservices_endpointslices" I0419 18:45:10.741639 3645 shared_informer.go:356] "Caches are synced" controller="egressservices_endpointslices" I0419 18:45:10.741644 3645 egressservice_node.go:197] Repairing Egress Services I0419 18:45:10.751733 3645 node_controller.go:43] Starting Admin Policy Based Route Node Controller I0419 18:45:10.751750 3645 external_controller.go:278] Starting Admin Policy Based Route Controller I0419 18:45:10.751972 3645 default_node_network_controller.go:1077] Egress IP for secondary host network is disabled I0419 18:45:10.751990 3645 link_network_manager.go:119] Link manager is running I0419 18:45:10.752004 3645 default_node_network_controller.go:1098] Default node network controller initialized and ready. I0419 18:45:10.752156 3645 ovspinning_linux.go:50] OVS CPU affinity pinning disabled I0419 18:45:10.752332 3645 vrf_manager.go:128] VRF manager is running I0419 18:45:10.752433 3645 node_controller_manager.go:437] Removing flows to drop GARP I0419 18:45:10.752442 3645 gateway.go:532] Reconciling gateway with updates I0419 18:45:10.770428 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-apiserver-operator/metrics I0419 18:45:10.770448 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator I0419 18:45:10.770458 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-apiserver-operator/metrics I0419 18:45:10.770455 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift-oauth-apiserver I0419 18:45:10.770434 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics I0419 18:45:10.770471 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator I0419 18:45:10.770476 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift-oauth-apiserver I0419 18:45:10.770478 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics I0419 18:45:10.770480 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-config-operator/metrics I0419 18:45:10.770484 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics I0419 18:45:10.770484 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics I0419 18:45:10.770486 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics I0419 18:45:10.770495 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics I0419 18:45:10.770494 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator I0419 18:45:10.770504 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-monitoring/cluster-monitoring-operator after 0 failed attempt(s) I0419 18:45:10.770505 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio I0419 18:45:10.770505 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon I0419 18:45:10.770515 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-samples-operator/metrics I0419 18:45:10.770517 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio I0419 18:45:10.770520 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon I0419 18:45:10.770523 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-samples-operator/metrics I0419 18:45:10.770493 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-config-operator/metrics I0419 18:45:10.770532 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-samples-operator/metrics after 0 failed attempt(s) I0419 18:45:10.770534 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-dns-operator/metrics I0419 18:45:10.770433 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics I0419 18:45:10.770544 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics I0419 18:45:10.770468 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator I0419 18:45:10.770548 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-storage-operator/csi-snapshot-controller-operator-metrics after 0 failed attempt(s) I0419 18:45:10.770552 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics I0419 18:45:10.770554 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics I0419 18:45:10.770557 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-node-tuning-operator/node-tuning-operator after 0 failed attempt(s) I0419 18:45:10.770558 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-apiserver/check-endpoints I0419 18:45:10.770543 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-dns-operator/metrics I0419 18:45:10.770544 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift-apiserver I0419 18:45:10.770566 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-authentication-operator/metrics I0419 18:45:10.770575 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress-operator/metrics I0419 18:45:10.770575 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-server I0419 18:45:10.770581 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-multus/network-metrics-service I0419 18:45:10.770585 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress/router-internal-default I0419 18:45:10.770589 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-server I0419 18:45:10.770589 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics I0419 18:45:10.770592 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-multus/network-metrics-service I0419 18:45:10.770598 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics I0419 18:45:10.770599 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-operator/metrics I0419 18:45:10.770603 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-metrics after 0 failed attempt(s) I0419 18:45:10.770602 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver I0419 18:45:10.770604 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-diagnostics/network-check-source I0419 18:45:10.770611 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-operator/metrics I0419 18:45:10.770616 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-diagnostics/network-check-source I0419 18:45:10.770611 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/openshift I0419 18:45:10.770616 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller I0419 18:45:10.770622 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-operator/metrics after 0 failed attempt(s) I0419 18:45:10.770620 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-dns-operator/metrics after 0 failed attempt(s) I0419 18:45:10.770627 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift I0419 18:45:10.770624 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-diagnostics/network-check-source after 0 failed attempt(s) I0419 18:45:10.770630 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller I0419 18:45:10.770639 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift after 0 failed attempt(s) I0419 18:45:10.770642 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-server after 0 failed attempt(s) I0419 18:45:10.770571 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-apiserver/check-endpoints I0419 18:45:10.770665 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-kube-controller-manager-operator/metrics after 0 failed attempt(s) I0419 18:45:10.770669 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-controller after 0 failed attempt(s) I0419 18:45:10.770465 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/kubernetes I0419 18:45:10.770685 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/kubernetes I0419 18:45:10.770674 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-apiserver/check-endpoints after 0 failed attempt(s) I0419 18:45:10.770531 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics I0419 18:45:10.770704 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics I0419 18:45:10.770542 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-version/cluster-version-operator I0419 18:45:10.770716 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-version/cluster-version-operator I0419 18:45:10.770453 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node I0419 18:45:10.770749 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node I0419 18:45:10.770754 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ovn-kubernetes/ovn-kubernetes-node after 0 failed attempt(s) I0419 18:45:10.770556 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service I0419 18:45:10.770762 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service I0419 18:45:10.770564 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/machine-config-daemon after 0 failed attempt(s) I0419 18:45:10.770566 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics I0419 18:45:10.770781 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics I0419 18:45:10.770568 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-storage-operator/cluster-storage-operator-metrics after 0 failed attempt(s) I0419 18:45:10.770571 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-dns/dns-default I0419 18:45:10.770832 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-dns/dns-default I0419 18:45:10.770570 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift-oauth-apiserver after 0 failed attempt(s) I0419 18:45:10.770471 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-console-operator/metrics I0419 18:45:10.770877 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-console-operator/metrics I0419 18:45:10.770541 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-apiserver-operator/metrics after 0 failed attempt(s) I0419 18:45:10.770579 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/openshift-apiserver I0419 18:45:10.770581 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-authentication-operator/metrics I0419 18:45:10.770580 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cloud-credential-operator/cco-metrics after 0 failed attempt(s) I0419 18:45:10.770559 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-diagnostics/network-check-target I0419 18:45:10.770991 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-diagnostics/network-check-target I0419 18:45:10.770583 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress-operator/metrics I0419 18:45:10.770555 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway default/packageserver I0419 18:45:10.771064 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway default/packageserver I0419 18:45:10.770585 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-config-operator/metrics after 0 failed attempt(s) I0419 18:45:10.770571 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-image-registry/image-registry I0419 18:45:10.771109 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-image-registry/image-registry I0419 18:45:10.770591 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress/router-default I0419 18:45:10.771147 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress/router-default I0419 18:45:10.771164 3645 port_claim.go:71] Svc openshift-ingress/router-default: port 32321 is already open, no action needed I0419 18:45:10.771174 3645 port_claim.go:71] Svc openshift-ingress/router-default: port 32654 is already open, no action needed I0419 18:45:10.770591 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-insights/metrics I0419 18:45:10.771223 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-insights/metrics I0419 18:45:10.770593 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress/router-internal-default I0419 18:45:10.770594 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog I0419 18:45:10.771292 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog I0419 18:45:10.770601 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-network-console/networking-console-plugin I0419 18:45:10.771333 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-network-console/networking-console-plugin I0419 18:45:10.770602 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-operator-lifecycle-manager/catalog-operator-metrics after 0 failed attempt(s) I0419 18:45:10.770601 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-service-ca-operator/metrics I0419 18:45:10.771381 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-service-ca-operator/metrics I0419 18:45:10.770609 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-machine-config-operator/kube-rbac-proxy-crio after 0 failed attempt(s) I0419 18:45:10.770605 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-multus/network-metrics-service after 0 failed attempt(s) I0419 18:45:10.770613 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver I0419 18:45:10.771461 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-machine-approver/machine-approver after 0 failed attempt(s) I0419 18:45:10.770470 3645 obj_retry.go:334] Retry object setup: *factory.serviceForGateway openshift-ingress-canary/ingress-canary I0419 18:45:10.771476 3645 obj_retry.go:398] Adding new object: *factory.serviceForGateway openshift-ingress-canary/ingress-canary I0419 18:45:10.796253 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/kubernetes after 0 failed attempt(s) I0419 18:45:10.796279 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-operator-lifecycle-manager/olm-operator-metrics after 0 failed attempt(s) I0419 18:45:10.796286 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-version/cluster-version-operator after 0 failed attempt(s) I0419 18:45:10.796292 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-cluster-node-tuning-operator/performance-addon-operator-service after 0 failed attempt(s) I0419 18:45:10.796298 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-kube-storage-version-migrator-operator/metrics after 0 failed attempt(s) I0419 18:45:10.837288 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-dns/dns-default after 0 failed attempt(s) I0419 18:45:10.837315 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-console-operator/metrics after 0 failed attempt(s) I0419 18:45:10.837321 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/openshift-apiserver after 0 failed attempt(s) I0419 18:45:10.837329 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-authentication-operator/metrics after 0 failed attempt(s) I0419 18:45:10.837338 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-diagnostics/network-check-target after 0 failed attempt(s) I0419 18:45:10.837357 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress-operator/metrics after 0 failed attempt(s) I0419 18:45:10.837375 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway default/packageserver after 0 failed attempt(s) I0419 18:45:10.837382 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-image-registry/image-registry after 0 failed attempt(s) I0419 18:45:10.878432 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress/router-default after 0 failed attempt(s) I0419 18:45:10.878443 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-insights/metrics after 0 failed attempt(s) I0419 18:45:10.878464 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress/router-internal-default after 0 failed attempt(s) I0419 18:45:10.878471 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-deployment-validation-operator/deployment-validation-operator-catalog after 0 failed attempt(s) I0419 18:45:10.878477 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-network-console/networking-console-plugin after 0 failed attempt(s) I0419 18:45:10.878483 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-service-ca-operator/metrics after 0 failed attempt(s) I0419 18:45:10.878489 3645 obj_retry.go:418] Retry successful for *factory.serviceForGateway openshift-ingress-canary/ingress-canary after 0 failed attempt(s) I0419 18:45:12.286046 3645 egressservice_zone_node.go:111] Processing sync for Egress Service node ip-10-0-2-180.ec2.internal I0419 18:45:12.286076 3645 egressservice_zone_node.go:114] Finished syncing Egress Service node ip-10-0-2-180.ec2.internal: 36.087µs I0419 18:45:12.314358 3645 base_network_controller_pods.go:487] [default/openshift-image-registry/image-registry-7d465f4d6f-fhj4h] creating logical port openshift-image-registry_image-registry-7d465f4d6f-fhj4h for pod on switch ip-10-0-2-180.ec2.internal I0419 18:45:12.314517 3645 kube.go:256] Updating pod openshift-image-registry/image-registry-7d465f4d6f-fhj4h I0419 18:45:12.336048 3645 pod.go:62] [openshift-image-registry/image-registry-7d465f4d6f-fhj4h] pod update took 21.574837ms I0419 18:45:12.336076 3645 base_network_controller_pods.go:951] [default/openshift-image-registry/image-registry-7d465f4d6f-fhj4h] addLogicalPort annotation time took 21.604877ms I0419 18:45:12.337469 3645 pods.go:271] [openshift-image-registry/image-registry-7d465f4d6f-fhj4h] addLogicalPort took 23.12805ms, libovsdb time 1.088767ms I0419 18:45:12.345780 3645 base_network_controller_pods.go:487] [default/openshift-ingress-canary/ingress-canary-h276j] creating logical port openshift-ingress-canary_ingress-canary-h276j for pod on switch ip-10-0-2-180.ec2.internal I0419 18:45:12.345874 3645 kube.go:256] Updating pod openshift-ingress-canary/ingress-canary-h276j I0419 18:45:12.354389 3645 base_network_controller_pods.go:487] [default/openshift-dns/dns-default-d5bcz] creating logical port openshift-dns_dns-default-d5bcz for pod on switch ip-10-0-2-180.ec2.internal I0419 18:45:12.354522 3645 kube.go:256] Updating pod openshift-dns/dns-default-d5bcz I0419 18:45:12.362280 3645 pod.go:62] [openshift-ingress-canary/ingress-canary-h276j] pod update took 16.432761ms I0419 18:45:12.362309 3645 base_network_controller_pods.go:951] [default/openshift-ingress-canary/ingress-canary-h276j] addLogicalPort annotation time took 16.464307ms I0419 18:45:12.363394 3645 pods.go:271] [openshift-ingress-canary/ingress-canary-h276j] addLogicalPort took 17.62645ms, libovsdb time 790.923µs I0419 18:45:12.374163 3645 pod.go:62] [openshift-dns/dns-default-d5bcz] pod update took 19.661004ms I0419 18:45:12.374259 3645 base_network_controller_pods.go:951] [default/openshift-dns/dns-default-d5bcz] addLogicalPort annotation time took 19.761279ms I0419 18:45:12.376519 3645 pods.go:271] [openshift-dns/dns-default-d5bcz] addLogicalPort took 22.143526ms, libovsdb time 1.628719ms I0419 18:45:13.703141 3645 cni.go:452] [openshift-ingress-canary/ingress-canary-h276j f596effbb781be7a0b043b611f7fc6b89abe45174cabf1f104c044f08805ae1c network default NAD default NAD key ] ADD starting CNI request [openshift-ingress-canary/ingress-canary-h276j f596effbb781be7a0b043b611f7fc6b89abe45174cabf1f104c044f08805ae1c network default NAD default NAD key ] I0419 18:45:13.742590 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-ingress-canary, podName: ingress-canary-h276j, hostIfaceName: f596effbb781be7, network: default, NAD default, SandboxID: "f596effbb781be7a0b043b611f7fc6b89abe45174cabf1f104c044f08805ae1c", PCI device ID: , UID: "8b13b4c0-8339-4946-a37c-986106cefb28", MAC: 0a:58:0a:82:00:06, IPs: [10.130.0.6/23] I0419 18:45:13.791749 3645 cni.go:473] [openshift-ingress-canary/ingress-canary-h276j f596effbb781be7a0b043b611f7fc6b89abe45174cabf1f104c044f08805ae1c network default NAD default NAD key default] ADD finished CNI request [openshift-ingress-canary/ingress-canary-h276j f596effbb781be7a0b043b611f7fc6b89abe45174cabf1f104c044f08805ae1c network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"e6:bf:79:c6:73:c2\",\"name\":\"f596effbb781be7\"},{\"mac\":\"0a:58:0a:82:00:06\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/c4725370-449c-4e41-b3ba-5aa5b461668c\"}],\"ips\":[{\"address\":\"10.130.0.6/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:45:16.140284 3645 default_network_controller.go:1025] Node "ip-10-0-1-170.ec2.internal" in remote zone "ip-10-0-1-170.ec2.internal", network "default", needs interconnect zone sync up. Zone cluster changed: false I0419 18:45:16.141115 3645 zone_ic_handler.go:195] Time taken to create transit switch: 282.675µs I0419 18:45:16.141130 3645 zone_ic_handler.go:277] Creating interconnect resources for remote zone node ip-10-0-1-170.ec2.internal for the network default I0419 18:45:16.142190 3645 zone_ic_handler.go:282] Creating Interconnect resources for node "ip-10-0-1-170.ec2.internal" on network "default" took: 1.371901ms I0419 18:45:16.376188 3645 cni.go:452] [openshift-network-diagnostics/network-check-target-p9ldg 62af0bb631eeda21755a149436018210b83d0b0378dded7cf0853c0c461c6181 network default NAD default NAD key ] ADD starting CNI request [openshift-network-diagnostics/network-check-target-p9ldg 62af0bb631eeda21755a149436018210b83d0b0378dded7cf0853c0c461c6181 network default NAD default NAD key ] I0419 18:45:16.386284 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-network-diagnostics, podName: network-check-target-p9ldg, hostIfaceName: 62af0bb631eeda2, network: default, NAD default, SandboxID: "62af0bb631eeda21755a149436018210b83d0b0378dded7cf0853c0c461c6181", PCI device ID: , UID: "90b1bb67-2a12-4a77-b1b1-2c27357c5aea", MAC: 0a:58:0a:82:00:04, IPs: [10.130.0.4/23] I0419 18:45:16.434172 3645 cni.go:473] [openshift-network-diagnostics/network-check-target-p9ldg 62af0bb631eeda21755a149436018210b83d0b0378dded7cf0853c0c461c6181 network default NAD default NAD key default] ADD finished CNI request [openshift-network-diagnostics/network-check-target-p9ldg 62af0bb631eeda21755a149436018210b83d0b0378dded7cf0853c0c461c6181 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"6e:b7:11:1b:52:7c\",\"name\":\"62af0bb631eeda2\"},{\"mac\":\"0a:58:0a:82:00:04\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/55904b09-dbaf-4c95-ba42-257a3f3896b9\"}],\"ips\":[{\"address\":\"10.130.0.4/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:45:51.582826 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8080: empty IP address endpoints for service openshift-network-diagnostics/network-check-target I0419 18:46:38.815257 3645 base_network_controller_pods.go:487] [default/openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8] creating logical port openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-74478b59c6-9ktd8 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:46:38.815383 3645 kube.go:256] Updating pod openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8 I0419 18:46:38.815436 3645 base_network_controller_pods.go:487] [default/openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d] creating logical port openshift-cluster-storage-operator_volume-data-source-validator-6944588dcc-76d9d for pod on switch ip-10-0-2-180.ec2.internal I0419 18:46:38.815537 3645 kube.go:256] Updating pod openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d I0419 18:46:38.844328 3645 pod.go:62] [openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d] pod update took 28.806217ms I0419 18:46:38.844332 3645 pod.go:62] [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8] pod update took 29.010287ms I0419 18:46:38.844351 3645 base_network_controller_pods.go:951] [default/openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d] addLogicalPort annotation time took 28.833961ms I0419 18:46:38.844356 3645 base_network_controller_pods.go:951] [default/openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8] addLogicalPort annotation time took 29.039744ms I0419 18:46:38.845599 3645 pods.go:271] [openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d] addLogicalPort took 30.205796ms, libovsdb time 908.563µs I0419 18:46:38.845907 3645 pods.go:271] [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8] addLogicalPort took 30.667791ms, libovsdb time 1.194844ms I0419 18:46:38.914068 3645 base_network_controller_pods.go:487] [default/openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv] creating logical port openshift-network-diagnostics_network-check-source-84fd6d9d54-5glnv for pod on switch ip-10-0-2-180.ec2.internal I0419 18:46:38.914178 3645 kube.go:256] Updating pod openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv I0419 18:46:38.950844 3645 pod.go:62] [openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv] pod update took 36.685282ms I0419 18:46:38.950865 3645 base_network_controller_pods.go:951] [default/openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv] addLogicalPort annotation time took 36.71184ms I0419 18:46:38.952021 3645 pods.go:271] [openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv] addLogicalPort took 37.967996ms, libovsdb time 907.93µs I0419 18:46:39.045118 3645 base_network_controller_pods.go:487] [default/openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4] creating logical port openshift-service-ca-operator_service-ca-operator-bc9564db4-tp9p4 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:46:39.045253 3645 kube.go:256] Updating pod openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4 I0419 18:46:39.045313 3645 base_network_controller_pods.go:487] [default/openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8] creating logical port openshift-network-console_networking-console-plugin-7f449dfbfd-lshq8 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:46:39.045382 3645 kube.go:256] Updating pod openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8 I0419 18:46:39.080783 3645 base_network_controller_pods.go:487] [default/openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k] creating logical port openshift-monitoring_cluster-monitoring-operator-557b74df6-rpq6k for pod on switch ip-10-0-2-180.ec2.internal I0419 18:46:39.080887 3645 kube.go:256] Updating pod openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k I0419 18:46:39.098811 3645 pod.go:62] [openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8] pod update took 53.446184ms I0419 18:46:39.098837 3645 base_network_controller_pods.go:951] [default/openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8] addLogicalPort annotation time took 53.476659ms I0419 18:46:39.099930 3645 pods.go:271] [openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8] addLogicalPort took 54.627983ms, libovsdb time 738.207µs I0419 18:46:39.112603 3645 pod.go:62] [openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4] pod update took 67.37327ms I0419 18:46:39.112622 3645 base_network_controller_pods.go:951] [default/openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4] addLogicalPort annotation time took 67.397307ms I0419 18:46:39.113593 3645 pods.go:271] [openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4] addLogicalPort took 68.494903ms, libovsdb time 723.152µs I0419 18:46:39.126075 3645 pod.go:62] [openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k] pod update took 45.216469ms I0419 18:46:39.126094 3645 base_network_controller_pods.go:951] [default/openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k] addLogicalPort annotation time took 45.239986ms I0419 18:46:39.126953 3645 pods.go:271] [openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k] addLogicalPort took 46.189051ms, libovsdb time 620.567µs I0419 18:46:39.472186 3645 cni.go:452] [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8 5003289b11a5c0b8f06649d39b8d5ee8a28e5df6dab2cd6e957c3d9f395e1577 network default NAD default NAD key ] ADD starting CNI request [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8 5003289b11a5c0b8f06649d39b8d5ee8a28e5df6dab2cd6e957c3d9f395e1577 network default NAD default NAD key ] I0419 18:46:39.478224 3645 cni.go:452] [openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d 5b608b51d782f49e3018e53c79d93ec0b3caa52f7bb6503a554c4a9a92a699b3 network default NAD default NAD key ] ADD starting CNI request [openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d 5b608b51d782f49e3018e53c79d93ec0b3caa52f7bb6503a554c4a9a92a699b3 network default NAD default NAD key ] I0419 18:46:39.483224 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-kube-storage-version-migrator-operator, podName: kube-storage-version-migrator-operator-74478b59c6-9ktd8, hostIfaceName: 5003289b11a5c0b, network: default, NAD default, SandboxID: "5003289b11a5c0b8f06649d39b8d5ee8a28e5df6dab2cd6e957c3d9f395e1577", PCI device ID: , UID: "79d41320-0eb5-4de5-8272-0567178d0c3a", MAC: 0a:58:0a:82:00:08, IPs: [10.130.0.8/23] I0419 18:46:39.488392 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-cluster-storage-operator, podName: volume-data-source-validator-6944588dcc-76d9d, hostIfaceName: 5b608b51d782f49, network: default, NAD default, SandboxID: "5b608b51d782f49e3018e53c79d93ec0b3caa52f7bb6503a554c4a9a92a699b3", PCI device ID: , UID: "247fe720-e983-4be4-92d7-c0a7c8104be4", MAC: 0a:58:0a:82:00:09, IPs: [10.130.0.9/23] I0419 18:46:39.531364 3645 cni.go:473] [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8 5003289b11a5c0b8f06649d39b8d5ee8a28e5df6dab2cd6e957c3d9f395e1577 network default NAD default NAD key default] ADD finished CNI request [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-74478b59c6-9ktd8 5003289b11a5c0b8f06649d39b8d5ee8a28e5df6dab2cd6e957c3d9f395e1577 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"7e:d7:42:11:f9:4d\",\"name\":\"5003289b11a5c0b\"},{\"mac\":\"0a:58:0a:82:00:08\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/d6c68365-8037-4c45-941b-086064945cc0\"}],\"ips\":[{\"address\":\"10.130.0.8/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:46:39.548303 3645 cni.go:473] [openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d 5b608b51d782f49e3018e53c79d93ec0b3caa52f7bb6503a554c4a9a92a699b3 network default NAD default NAD key default] ADD finished CNI request [openshift-cluster-storage-operator/volume-data-source-validator-6944588dcc-76d9d 5b608b51d782f49e3018e53c79d93ec0b3caa52f7bb6503a554c4a9a92a699b3 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"6a:12:ae:ed:8f:5b\",\"name\":\"5b608b51d782f49\"},{\"mac\":\"0a:58:0a:82:00:09\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/00e3d208-30ec-4a6d-89a5-1c0c7e0aa34c\"}],\"ips\":[{\"address\":\"10.130.0.9/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:46:39.572126 3645 cni.go:452] [openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv 751937cccf203c3f34fe94572f1056a925ea0a011b634712ad8313c3b83411e1 network default NAD default NAD key ] ADD starting CNI request [openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv 751937cccf203c3f34fe94572f1056a925ea0a011b634712ad8313c3b83411e1 network default NAD default NAD key ] I0419 18:46:39.584518 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-network-diagnostics, podName: network-check-source-84fd6d9d54-5glnv, hostIfaceName: 751937cccf203c3, network: default, NAD default, SandboxID: "751937cccf203c3f34fe94572f1056a925ea0a011b634712ad8313c3b83411e1", PCI device ID: , UID: "5614b656-95b0-4b8b-9877-33d9ee4802df", MAC: 0a:58:0a:82:00:0a, IPs: [10.130.0.10/23] I0419 18:46:39.632252 3645 cni.go:473] [openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv 751937cccf203c3f34fe94572f1056a925ea0a011b634712ad8313c3b83411e1 network default NAD default NAD key default] ADD finished CNI request [openshift-network-diagnostics/network-check-source-84fd6d9d54-5glnv 751937cccf203c3f34fe94572f1056a925ea0a011b634712ad8313c3b83411e1 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"2a:1e:bf:ce:eb:21\",\"name\":\"751937cccf203c3\"},{\"mac\":\"0a:58:0a:82:00:0a\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/1d05af3d-7f56-43fe-9fc9-5e7b92244f2f\"}],\"ips\":[{\"address\":\"10.130.0.10/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:46:41.038020 3645 cni.go:452] [openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4 929d217f13def34e8da4a9196a1a9ddef7e820d6c545904f1a23b59247d48765 network default NAD default NAD key ] ADD starting CNI request [openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4 929d217f13def34e8da4a9196a1a9ddef7e820d6c545904f1a23b59247d48765 network default NAD default NAD key ] I0419 18:46:41.048161 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-service-ca-operator, podName: service-ca-operator-bc9564db4-tp9p4, hostIfaceName: 929d217f13def34, network: default, NAD default, SandboxID: "929d217f13def34e8da4a9196a1a9ddef7e820d6c545904f1a23b59247d48765", PCI device ID: , UID: "8b8f06a5-1de8-4e27-a030-77838786925e", MAC: 0a:58:0a:82:00:0b, IPs: [10.130.0.11/23] I0419 18:46:41.096286 3645 cni.go:473] [openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4 929d217f13def34e8da4a9196a1a9ddef7e820d6c545904f1a23b59247d48765 network default NAD default NAD key default] ADD finished CNI request [openshift-service-ca-operator/service-ca-operator-bc9564db4-tp9p4 929d217f13def34e8da4a9196a1a9ddef7e820d6c545904f1a23b59247d48765 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"66:f7:58:a9:5c:bc\",\"name\":\"929d217f13def34\"},{\"mac\":\"0a:58:0a:82:00:0b\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/55f26f89-1242-4d0b-8fb1-eb626030cfc2\"}],\"ips\":[{\"address\":\"10.130.0.11/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:46:42.637162 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console-operator/metrics I0419 18:46:43.933567 3645 namespace.go:98] [openshift-kube-storage-version-migrator] adding namespace I0419 18:46:43.934447 3645 namespace.go:102] [openshift-kube-storage-version-migrator] adding namespace took 858.14µs I0419 18:46:43.959657 3645 namespace.go:142] [openshift-kube-storage-version-migrator] updating namespace I0419 18:46:45.917355 3645 namespace.go:98] [openshift-service-ca] adding namespace I0419 18:46:45.918367 3645 namespace.go:102] [openshift-service-ca] adding namespace took 985.456µs I0419 18:46:45.941712 3645 namespace.go:142] [openshift-service-ca] updating namespace I0419 18:46:47.601483 3645 base_network_controller_pods.go:487] [default/openshift-service-ca/service-ca-d9cb48497-47597] creating logical port openshift-service-ca_service-ca-d9cb48497-47597 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:46:47.601574 3645 kube.go:256] Updating pod openshift-service-ca/service-ca-d9cb48497-47597 I0419 18:46:47.621349 3645 pod.go:62] [openshift-service-ca/service-ca-d9cb48497-47597] pod update took 19.796634ms I0419 18:46:47.621374 3645 base_network_controller_pods.go:951] [default/openshift-service-ca/service-ca-d9cb48497-47597] addLogicalPort annotation time took 19.826663ms I0419 18:46:47.622625 3645 pods.go:271] [openshift-service-ca/service-ca-d9cb48497-47597] addLogicalPort took 21.160341ms, libovsdb time 876.476µs I0419 18:46:47.957082 3645 cni.go:452] [openshift-service-ca/service-ca-d9cb48497-47597 8d38b5208e153780efbf1081b8866587aae9ee1b88eff475e7c76c5745d3525c network default NAD default NAD key ] ADD starting CNI request [openshift-service-ca/service-ca-d9cb48497-47597 8d38b5208e153780efbf1081b8866587aae9ee1b88eff475e7c76c5745d3525c network default NAD default NAD key ] I0419 18:46:47.966814 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-service-ca, podName: service-ca-d9cb48497-47597, hostIfaceName: 8d38b5208e15378, network: default, NAD default, SandboxID: "8d38b5208e153780efbf1081b8866587aae9ee1b88eff475e7c76c5745d3525c", PCI device ID: , UID: "b73bb7c1-0b42-47d6-86f8-767308708dbe", MAC: 0a:58:0a:82:00:0e, IPs: [10.130.0.14/23] I0419 18:46:48.015547 3645 cni.go:473] [openshift-service-ca/service-ca-d9cb48497-47597 8d38b5208e153780efbf1081b8866587aae9ee1b88eff475e7c76c5745d3525c network default NAD default NAD key default] ADD finished CNI request [openshift-service-ca/service-ca-d9cb48497-47597 8d38b5208e153780efbf1081b8866587aae9ee1b88eff475e7c76c5745d3525c network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"d2:b6:9c:37:3d:21\",\"name\":\"8d38b5208e15378\"},{\"mac\":\"0a:58:0a:82:00:0e\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/7eeae358-7555-4ea1-9e42-ce6f6ff7c683\"}],\"ips\":[{\"address\":\"10.130.0.14/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:46:52.439545 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console-operator/metrics I0419 18:46:53.301194 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-78774c6778-p5x98 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:46:53.301845 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-78774c6778-p5x98 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:46:53.320475 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-78774c6778-p5x98 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:46:53.320496 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-78774c6778-p5x98 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:46:53.329116 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-image-registry/image-registry-78774c6778-p5x98 I0419 18:46:53.329135 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-image-registry/image-registry-78774c6778-p5x98 I0419 18:46:53.803698 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-7d465f4d6f-fhj4h of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:46:53.803732 3645 pods.go:176] Deleting pod: openshift-image-registry/image-registry-7d465f4d6f-fhj4h I0419 18:46:53.808861 3645 pods.go:236] Attempting to release IPs for pod: openshift-image-registry/image-registry-7d465f4d6f-fhj4h, ips: 10.130.0.5 I0419 18:46:53.808994 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-7d465f4d6f-fhj4h of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:46:53.813952 3645 base_network_controller_pods.go:487] [default/openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx] creating logical port openshift-image-registry_image-registry-7f8cfbbcf5-dgtbx for pod on switch ip-10-0-2-180.ec2.internal I0419 18:46:53.814311 3645 kube.go:256] Updating pod openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx I0419 18:46:53.823268 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-7d465f4d6f-fhj4h of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:46:53.823301 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-7d465f4d6f-fhj4h of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:46:53.834508 3645 pod.go:62] [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx] pod update took 20.36171ms I0419 18:46:53.834530 3645 base_network_controller_pods.go:951] [default/openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx] addLogicalPort annotation time took 20.388284ms I0419 18:46:53.835639 3645 pods.go:271] [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx] addLogicalPort took 21.703416ms, libovsdb time 847.443µs I0419 18:46:53.839757 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-image-registry/image-registry-7d465f4d6f-fhj4h I0419 18:46:53.839782 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-image-registry/image-registry-7d465f4d6f-fhj4h I0419 18:46:54.174958 3645 cni.go:452] [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d network default NAD default NAD key ] ADD starting CNI request [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d network default NAD default NAD key ] I0419 18:46:54.184547 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-image-registry, podName: image-registry-7f8cfbbcf5-dgtbx, hostIfaceName: d74ba2e7e7e473c, network: default, NAD default, SandboxID: "d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d", PCI device ID: , UID: "063f4656-ae7a-4996-b79a-6f254c0a552c", MAC: 0a:58:0a:82:00:0f, IPs: [10.130.0.15/23] I0419 18:46:54.233845 3645 cni.go:473] [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d network default NAD default NAD key default] ADD finished CNI request [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"8a:7a:81:6e:c8:68\",\"name\":\"d74ba2e7e7e473c\"},{\"mac\":\"0a:58:0a:82:00:0f\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/4da4ecfe-7ad6-4a80-b012-76fada677ea2\"}],\"ips\":[{\"address\":\"10.130.0.15/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:47:11.233535 3645 cni.go:452] [openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k abf629e5e71b01573081c24d15a020da568420e28f24d1a9fe6d9bf60d5d35d0 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k abf629e5e71b01573081c24d15a020da568420e28f24d1a9fe6d9bf60d5d35d0 network default NAD default NAD key ] I0419 18:47:11.242825 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: cluster-monitoring-operator-557b74df6-rpq6k, hostIfaceName: abf629e5e71b015, network: default, NAD default, SandboxID: "abf629e5e71b01573081c24d15a020da568420e28f24d1a9fe6d9bf60d5d35d0", PCI device ID: , UID: "f77af483-9260-4e72-9f9d-dde2658e736b", MAC: 0a:58:0a:82:00:0d, IPs: [10.130.0.13/23] I0419 18:47:11.294202 3645 cni.go:473] [openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k abf629e5e71b01573081c24d15a020da568420e28f24d1a9fe6d9bf60d5d35d0 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/cluster-monitoring-operator-557b74df6-rpq6k abf629e5e71b01573081c24d15a020da568420e28f24d1a9fe6d9bf60d5d35d0 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"fa:12:21:30:f5:94\",\"name\":\"abf629e5e71b015\"},{\"mac\":\"0a:58:0a:82:00:0d\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/4e27219c-009e-4dd1-a06e-f2b436fb1b32\"}],\"ips\":[{\"address\":\"10.130.0.13/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:47:12.093770 3645 cni.go:452] [openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8 761f826aa3df563aa329a70e789f1f3844113d6a18f80c2234c599f20c6d675d network default NAD default NAD key ] ADD starting CNI request [openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8 761f826aa3df563aa329a70e789f1f3844113d6a18f80c2234c599f20c6d675d network default NAD default NAD key ] I0419 18:47:12.104448 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-network-console, podName: networking-console-plugin-7f449dfbfd-lshq8, hostIfaceName: 761f826aa3df563, network: default, NAD default, SandboxID: "761f826aa3df563aa329a70e789f1f3844113d6a18f80c2234c599f20c6d675d", PCI device ID: , UID: "ba79b833-05df-40f9-b6e5-fc7f80879d5b", MAC: 0a:58:0a:82:00:0c, IPs: [10.130.0.12/23] I0419 18:47:12.155522 3645 cni.go:473] [openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8 761f826aa3df563aa329a70e789f1f3844113d6a18f80c2234c599f20c6d675d network default NAD default NAD key default] ADD finished CNI request [openshift-network-console/networking-console-plugin-7f449dfbfd-lshq8 761f826aa3df563aa329a70e789f1f3844113d6a18f80c2234c599f20c6d675d network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"0a:13:9e:26:86:15\",\"name\":\"761f826aa3df563\"},{\"mac\":\"0a:58:0a:82:00:0c\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/8b6ee658-c6c6-4278-82a3-4a49e4504977\"}],\"ips\":[{\"address\":\"10.130.0.12/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:47:14.570335 3645 util.go:870] Failed to build global endpoints for port TCP/5000-tcp: empty IP address endpoints for service openshift-image-registry/image-registry I0419 18:47:14.702597 3645 base_network_controller_pods.go:487] [default/openshift-insights/insights-runtime-extractor-78xtc] creating logical port openshift-insights_insights-runtime-extractor-78xtc for pod on switch ip-10-0-2-180.ec2.internal I0419 18:47:14.702688 3645 kube.go:256] Updating pod openshift-insights/insights-runtime-extractor-78xtc I0419 18:47:14.728515 3645 pod.go:62] [openshift-insights/insights-runtime-extractor-78xtc] pod update took 25.85753ms I0419 18:47:14.728538 3645 base_network_controller_pods.go:951] [default/openshift-insights/insights-runtime-extractor-78xtc] addLogicalPort annotation time took 25.882992ms I0419 18:47:14.729674 3645 pods.go:271] [openshift-insights/insights-runtime-extractor-78xtc] addLogicalPort took 27.090832ms, libovsdb time 858.097µs I0419 18:47:14.739575 3645 base_network_controller_pods.go:487] [default/openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97] creating logical port openshift-monitoring_prometheus-operator-admission-webhook-59b79769fb-llh97 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:47:14.739678 3645 kube.go:256] Updating pod openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97 I0419 18:47:14.757275 3645 pod.go:62] [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97] pod update took 17.630939ms I0419 18:47:14.757296 3645 base_network_controller_pods.go:951] [default/openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97] addLogicalPort annotation time took 17.658315ms I0419 18:47:14.758448 3645 pods.go:271] [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97] addLogicalPort took 18.880943ms, libovsdb time 844.932µs W0419 18:47:14.784553 3645 util.go:870] Failed to build global endpoints for port TCP/5000-tcp: empty IP address endpoints for service openshift-image-registry/image-registry I0419 18:47:15.060472 3645 cni.go:452] [openshift-insights/insights-runtime-extractor-78xtc 2dc1a74276ccab2c43bcffccf14a96d6edc5b82a17496947dd22df0265a89f53 network default NAD default NAD key ] ADD starting CNI request [openshift-insights/insights-runtime-extractor-78xtc 2dc1a74276ccab2c43bcffccf14a96d6edc5b82a17496947dd22df0265a89f53 network default NAD default NAD key ] I0419 18:47:15.069685 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-insights, podName: insights-runtime-extractor-78xtc, hostIfaceName: 2dc1a74276ccab2, network: default, NAD default, SandboxID: "2dc1a74276ccab2c43bcffccf14a96d6edc5b82a17496947dd22df0265a89f53", PCI device ID: , UID: "f6f3b323-a791-4c13-873b-92d67e08b07b", MAC: 0a:58:0a:82:00:10, IPs: [10.130.0.16/23] I0419 18:47:15.099139 3645 cni.go:452] [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97 a59e8a7ac27315f58f52902481a3419d6286caef5f2954aa64b4e5862b5c5340 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97 a59e8a7ac27315f58f52902481a3419d6286caef5f2954aa64b4e5862b5c5340 network default NAD default NAD key ] I0419 18:47:15.114172 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: prometheus-operator-admission-webhook-59b79769fb-llh97, hostIfaceName: a59e8a7ac27315f, network: default, NAD default, SandboxID: "a59e8a7ac27315f58f52902481a3419d6286caef5f2954aa64b4e5862b5c5340", PCI device ID: , UID: "7c68a11c-6fd3-4bb7-865f-76e7497c03f6", MAC: 0a:58:0a:82:00:11, IPs: [10.130.0.17/23] I0419 18:47:15.122530 3645 cni.go:473] [openshift-insights/insights-runtime-extractor-78xtc 2dc1a74276ccab2c43bcffccf14a96d6edc5b82a17496947dd22df0265a89f53 network default NAD default NAD key default] ADD finished CNI request [openshift-insights/insights-runtime-extractor-78xtc 2dc1a74276ccab2c43bcffccf14a96d6edc5b82a17496947dd22df0265a89f53 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"d6:db:0a:04:ce:2c\",\"name\":\"2dc1a74276ccab2\"},{\"mac\":\"0a:58:0a:82:00:10\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/698ceb8c-1484-4f71-a3ac-c5290c5ea042\"}],\"ips\":[{\"address\":\"10.130.0.16/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:47:15.162579 3645 cni.go:473] [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97 a59e8a7ac27315f58f52902481a3419d6286caef5f2954aa64b4e5862b5c5340 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/prometheus-operator-admission-webhook-59b79769fb-llh97 a59e8a7ac27315f58f52902481a3419d6286caef5f2954aa64b4e5862b5c5340 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"16:e1:c7:c6:9a:9d\",\"name\":\"a59e8a7ac27315f\"},{\"mac\":\"0a:58:0a:82:00:11\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/82201600-61f0-45c4-a296-00a3ceddaeea\"}],\"ips\":[{\"address\":\"10.130.0.17/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:47:15.742762 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-78774c6778-4vjzr of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:47:15.743372 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-78774c6778-4vjzr of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:47:15.765621 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-78774c6778-4vjzr of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:47:15.765646 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-78774c6778-4vjzr of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:47:15.776481 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-image-registry/image-registry-78774c6778-4vjzr I0419 18:47:15.776508 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-image-registry/image-registry-78774c6778-4vjzr W0419 18:47:16.749498 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-monitoring/prometheus-operator-admission-webhook W0419 18:47:16.778504 3645 util.go:870] Failed to build global endpoints for port TCP/5000-tcp: empty IP address endpoints for service openshift-image-registry/image-registry W0419 18:47:16.911492 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-monitoring/prometheus-operator-admission-webhook I0419 18:47:17.707490 3645 base_network_controller_pods.go:487] [default/openshift-monitoring/prometheus-operator-76b589d668-z5xb9] creating logical port openshift-monitoring_prometheus-operator-76b589d668-z5xb9 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:47:17.707613 3645 kube.go:256] Updating pod openshift-monitoring/prometheus-operator-76b589d668-z5xb9 I0419 18:47:17.723307 3645 pod.go:62] [openshift-monitoring/prometheus-operator-76b589d668-z5xb9] pod update took 15.718256ms I0419 18:47:17.723338 3645 base_network_controller_pods.go:951] [default/openshift-monitoring/prometheus-operator-76b589d668-z5xb9] addLogicalPort annotation time took 15.751787ms I0419 18:47:17.724653 3645 pods.go:271] [openshift-monitoring/prometheus-operator-76b589d668-z5xb9] addLogicalPort took 17.179191ms, libovsdb time 910.677µs I0419 18:47:18.066627 3645 cni.go:452] [openshift-monitoring/prometheus-operator-76b589d668-z5xb9 8a8ee46b107d8e314b29796a69e0df927241e17428e6daa243150a8f3e15639f network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/prometheus-operator-76b589d668-z5xb9 8a8ee46b107d8e314b29796a69e0df927241e17428e6daa243150a8f3e15639f network default NAD default NAD key ] I0419 18:47:18.076316 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: prometheus-operator-76b589d668-z5xb9, hostIfaceName: 8a8ee46b107d8e3, network: default, NAD default, SandboxID: "8a8ee46b107d8e314b29796a69e0df927241e17428e6daa243150a8f3e15639f", PCI device ID: , UID: "ba0f3e52-5086-4f1c-b7dc-c5108b602356", MAC: 0a:58:0a:82:00:12, IPs: [10.130.0.18/23] I0419 18:47:18.126227 3645 cni.go:473] [openshift-monitoring/prometheus-operator-76b589d668-z5xb9 8a8ee46b107d8e314b29796a69e0df927241e17428e6daa243150a8f3e15639f network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/prometheus-operator-76b589d668-z5xb9 8a8ee46b107d8e314b29796a69e0df927241e17428e6daa243150a8f3e15639f network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"86:78:86:06:f1:56\",\"name\":\"8a8ee46b107d8e3\"},{\"mac\":\"0a:58:0a:82:00:12\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/f13356e6-61a5-4ece-b3c1-c7a18c5d524b\"}],\"ips\":[{\"address\":\"10.130.0.18/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:47:22.264098 3645 base_network_controller_pods.go:487] [default/openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc] creating logical port openshift-monitoring_openshift-state-metrics-b8587444b-9m5xc for pod on switch ip-10-0-2-180.ec2.internal I0419 18:47:22.264220 3645 kube.go:256] Updating pod openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc I0419 18:47:22.286499 3645 pod.go:62] [openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc] pod update took 22.311583ms I0419 18:47:22.286533 3645 base_network_controller_pods.go:951] [default/openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc] addLogicalPort annotation time took 22.347545ms I0419 18:47:22.287726 3645 pods.go:271] [openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc] addLogicalPort took 23.648496ms, libovsdb time 814.753µs I0419 18:47:23.228482 3645 cni.go:452] [openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc fb90b718faa78b606edf5cd41f9c9c590dc4988c756e6dafd63434c8e7b782cb network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc fb90b718faa78b606edf5cd41f9c9c590dc4988c756e6dafd63434c8e7b782cb network default NAD default NAD key ] I0419 18:47:23.238968 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: openshift-state-metrics-b8587444b-9m5xc, hostIfaceName: fb90b718faa78b6, network: default, NAD default, SandboxID: "fb90b718faa78b606edf5cd41f9c9c590dc4988c756e6dafd63434c8e7b782cb", PCI device ID: , UID: "63a2ace0-2d22-44e6-ad2d-c3d30f048aa3", MAC: 0a:58:0a:82:00:13, IPs: [10.130.0.19/23] I0419 18:47:23.296875 3645 cni.go:473] [openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc fb90b718faa78b606edf5cd41f9c9c590dc4988c756e6dafd63434c8e7b782cb network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/openshift-state-metrics-b8587444b-9m5xc fb90b718faa78b606edf5cd41f9c9c590dc4988c756e6dafd63434c8e7b782cb network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"2a:93:d7:a6:56:b5\",\"name\":\"fb90b718faa78b6\"},{\"mac\":\"0a:58:0a:82:00:13\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/55c8fa66-d5be-43c8-8a63-95ca87c67215\"}],\"ips\":[{\"address\":\"10.130.0.19/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:47:25.430075 3645 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0419 18:47:25.430096 3645 util.go:870] Failed to build global endpoints for port TCP/tenancy: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0419 18:47:25.430102 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/alertmanager-main I0419 18:47:25.552220 3645 base_network_controller_pods.go:487] [default/openshift-monitoring/thanos-querier-6d68645545-rlp84] creating logical port openshift-monitoring_thanos-querier-6d68645545-rlp84 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:47:25.552316 3645 kube.go:256] Updating pod openshift-monitoring/thanos-querier-6d68645545-rlp84 I0419 18:47:25.574592 3645 pod.go:62] [openshift-monitoring/thanos-querier-6d68645545-rlp84] pod update took 22.305226ms I0419 18:47:25.574619 3645 base_network_controller_pods.go:951] [default/openshift-monitoring/thanos-querier-6d68645545-rlp84] addLogicalPort annotation time took 22.337323ms I0419 18:47:25.575823 3645 pods.go:271] [openshift-monitoring/thanos-querier-6d68645545-rlp84] addLogicalPort took 23.616832ms, libovsdb time 854.926µs W0419 18:47:25.841215 3645 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0419 18:47:25.841236 3645 util.go:870] Failed to build global endpoints for port TCP/tenancy: empty IP address endpoints for service openshift-monitoring/alertmanager-main W0419 18:47:25.841241 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/alertmanager-main I0419 18:47:25.914080 3645 cni.go:452] [openshift-monitoring/thanos-querier-6d68645545-rlp84 7cccc9b0f6757743c541c3bce2132ebaa3be0dad6080aec03a4cbe2a1c8b7e4e network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/thanos-querier-6d68645545-rlp84 7cccc9b0f6757743c541c3bce2132ebaa3be0dad6080aec03a4cbe2a1c8b7e4e network default NAD default NAD key ] I0419 18:47:25.924564 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: thanos-querier-6d68645545-rlp84, hostIfaceName: 7cccc9b0f675774, network: default, NAD default, SandboxID: "7cccc9b0f6757743c541c3bce2132ebaa3be0dad6080aec03a4cbe2a1c8b7e4e", PCI device ID: , UID: "5efc9eef-19d0-4625-892f-80c0d3366e7a", MAC: 0a:58:0a:82:00:14, IPs: [10.130.0.20/23] I0419 18:47:25.973228 3645 cni.go:473] [openshift-monitoring/thanos-querier-6d68645545-rlp84 7cccc9b0f6757743c541c3bce2132ebaa3be0dad6080aec03a4cbe2a1c8b7e4e network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/thanos-querier-6d68645545-rlp84 7cccc9b0f6757743c541c3bce2132ebaa3be0dad6080aec03a4cbe2a1c8b7e4e network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"da:69:cf:74:8a:6d\",\"name\":\"7cccc9b0f675774\"},{\"mac\":\"0a:58:0a:82:00:14\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/e25c4df2-3a96-4cfe-983d-e4f4f3362de7\"}],\"ips\":[{\"address\":\"10.130.0.20/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:47:26.832640 3645 base_network_controller_pods.go:487] [default/openshift-monitoring/metrics-server-59f79555dc-j8m8g] creating logical port openshift-monitoring_metrics-server-59f79555dc-j8m8g for pod on switch ip-10-0-2-180.ec2.internal I0419 18:47:26.832730 3645 kube.go:256] Updating pod openshift-monitoring/metrics-server-59f79555dc-j8m8g I0419 18:47:26.853613 3645 pod.go:62] [openshift-monitoring/metrics-server-59f79555dc-j8m8g] pod update took 20.90276ms I0419 18:47:26.853641 3645 base_network_controller_pods.go:951] [default/openshift-monitoring/metrics-server-59f79555dc-j8m8g] addLogicalPort annotation time took 20.937202ms I0419 18:47:26.854889 3645 pods.go:271] [openshift-monitoring/metrics-server-59f79555dc-j8m8g] addLogicalPort took 22.259328ms, libovsdb time 870.192µs I0419 18:47:27.197265 3645 cni.go:452] [openshift-monitoring/metrics-server-59f79555dc-j8m8g 7b12bcd5987e4595e1d666b5331147737f02c332b6ead5452eb7de0692ff4e7b network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/metrics-server-59f79555dc-j8m8g 7b12bcd5987e4595e1d666b5331147737f02c332b6ead5452eb7de0692ff4e7b network default NAD default NAD key ] I0419 18:47:27.206627 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: metrics-server-59f79555dc-j8m8g, hostIfaceName: 7b12bcd5987e459, network: default, NAD default, SandboxID: "7b12bcd5987e4595e1d666b5331147737f02c332b6ead5452eb7de0692ff4e7b", PCI device ID: , UID: "b94c8bf9-fd05-4cc3-92ca-8ced0c952b46", MAC: 0a:58:0a:82:00:15, IPs: [10.130.0.21/23] I0419 18:47:27.258749 3645 cni.go:473] [openshift-monitoring/metrics-server-59f79555dc-j8m8g 7b12bcd5987e4595e1d666b5331147737f02c332b6ead5452eb7de0692ff4e7b network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/metrics-server-59f79555dc-j8m8g 7b12bcd5987e4595e1d666b5331147737f02c332b6ead5452eb7de0692ff4e7b network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"86:b7:11:4f:c0:2b\",\"name\":\"7b12bcd5987e459\"},{\"mac\":\"0a:58:0a:82:00:15\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/59f5bd7e-12e1-4338-81d3-4c5076972a74\"}],\"ips\":[{\"address\":\"10.130.0.21/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:47:28.440683 3645 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:28.440707 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:28.440713 3645 util.go:870] Failed to build global endpoints for port TCP/tenancy: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:28.440719 3645 util.go:870] Failed to build global endpoints for port TCP/tenancy-rules: empty IP address endpoints for service openshift-monitoring/thanos-querier I0419 18:47:28.690807 3645 base_network_controller_pods.go:487] [default/openshift-monitoring/prometheus-k8s-1] creating logical port openshift-monitoring_prometheus-k8s-1 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:47:28.691003 3645 kube.go:256] Updating pod openshift-monitoring/prometheus-k8s-1 I0419 18:47:28.692228 3645 cni.go:452] [openshift-dns/dns-default-d5bcz 949a24409fb1399d6b4ac81b648600519129b3eb92f05f838a0de756b560634c network default NAD default NAD key ] ADD starting CNI request [openshift-dns/dns-default-d5bcz 949a24409fb1399d6b4ac81b648600519129b3eb92f05f838a0de756b560634c network default NAD default NAD key ] I0419 18:47:28.705537 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-dns, podName: dns-default-d5bcz, hostIfaceName: 949a24409fb1399, network: default, NAD default, SandboxID: "949a24409fb1399d6b4ac81b648600519129b3eb92f05f838a0de756b560634c", PCI device ID: , UID: "32114b48-d457-45eb-8f9e-6680120c8677", MAC: 0a:58:0a:82:00:07, IPs: [10.130.0.7/23] I0419 18:47:28.715737 3645 pod.go:62] [openshift-monitoring/prometheus-k8s-1] pod update took 24.784089ms I0419 18:47:28.715766 3645 base_network_controller_pods.go:951] [default/openshift-monitoring/prometheus-k8s-1] addLogicalPort annotation time took 24.822135ms I0419 18:47:28.717650 3645 pods.go:271] [openshift-monitoring/prometheus-k8s-1] addLogicalPort took 26.86124ms, libovsdb time 1.426354ms I0419 18:47:28.766417 3645 cni.go:473] [openshift-dns/dns-default-d5bcz 949a24409fb1399d6b4ac81b648600519129b3eb92f05f838a0de756b560634c network default NAD default NAD key default] ADD finished CNI request [openshift-dns/dns-default-d5bcz 949a24409fb1399d6b4ac81b648600519129b3eb92f05f838a0de756b560634c network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ba:cb:83:b6:80:f4\",\"name\":\"949a24409fb1399\"},{\"mac\":\"0a:58:0a:82:00:07\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/9c02ec63-6068-423f-af3e-4a22ff5f9fe8\"}],\"ips\":[{\"address\":\"10.130.0.7/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:47:28.857012 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-monitoring/metrics-server W0419 18:47:28.956204 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-monitoring/metrics-server I0419 18:47:29.079788 3645 cni.go:452] [openshift-monitoring/prometheus-k8s-1 6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/prometheus-k8s-1 6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38 network default NAD default NAD key ] I0419 18:47:29.090762 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: prometheus-k8s-1, hostIfaceName: 6803d8e34c4fb1c, network: default, NAD default, SandboxID: "6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38", PCI device ID: , UID: "f5e1a0b6-0dce-4662-8ce9-29f4c347b071", MAC: 0a:58:0a:82:00:16, IPs: [10.130.0.22/23] I0419 18:47:29.149301 3645 cni.go:473] [openshift-monitoring/prometheus-k8s-1 6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/prometheus-k8s-1 6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"7e:07:4f:46:d3:5f\",\"name\":\"6803d8e34c4fb1c\"},{\"mac\":\"0a:58:0a:82:00:16\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/09bb5ecd-8e60-43cb-9998-ef6acaeed137\"}],\"ips\":[{\"address\":\"10.130.0.22/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:47:29.828385 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-monitoring/monitoring-plugin W0419 18:47:29.920869 3645 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:47:29.920894 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:47:29.985516 3645 util.go:870] Failed to build global endpoints for port TCP/tenancy: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:29.985534 3645 util.go:870] Failed to build global endpoints for port TCP/tenancy-rules: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:29.985539 3645 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:29.985543 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:30.997884 3645 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:47:30.997905 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:47:31.029492 3645 util.go:870] Failed to build global endpoints for port TCP/dns-tcp: empty IP address endpoints for service openshift-dns/dns-default W0419 18:47:31.029517 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-dns/dns-default W0419 18:47:31.029524 3645 util.go:870] Failed to build global endpoints for port UDP/dns: empty IP address endpoints for service openshift-dns/dns-default W0419 18:47:34.458623 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:34.458650 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9094: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:34.458658 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:34.458670 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9093: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:34.458692 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:34.458699 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9094: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:34.458706 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:34.458712 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9093: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:36.994072 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:36.994095 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9094: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:36.994102 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/thanos-querier W0419 18:47:36.994108 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9093: empty IP address endpoints for service openshift-monitoring/thanos-querier I0419 18:47:40.065017 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-7f8cfbbcf5-wtbtq of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:47:40.065630 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-7f8cfbbcf5-wtbtq of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it W0419 18:47:40.073558 3645 util.go:870] Failed to build global endpoints for port TCP/5000-tcp: empty IP address endpoints for service openshift-image-registry/image-registry I0419 18:47:40.485344 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-7f8cfbbcf5-wtbtq of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:47:40.485366 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-7f8cfbbcf5-wtbtq of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:47:40.494277 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-image-registry/image-registry-7f8cfbbcf5-wtbtq I0419 18:47:40.494306 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-image-registry/image-registry-7f8cfbbcf5-wtbtq W0419 18:47:40.983633 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 5353: empty IP address endpoints for service openshift-dns/dns-default W0419 18:47:40.983657 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9154: empty IP address endpoints for service openshift-dns/dns-default W0419 18:47:40.983666 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 5353: empty IP address endpoints for service openshift-dns/dns-default W0419 18:47:41.498194 3645 util.go:870] Failed to build global endpoints for port TCP/5000-tcp: empty IP address endpoints for service openshift-image-registry/image-registry W0419 18:47:47.767422 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 5000: empty IP address endpoints for service openshift-image-registry/image-registry W0419 18:47:47.767452 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 5000: empty IP address endpoints for service openshift-image-registry/image-registry W0419 18:47:49.842067 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 10250: empty IP address endpoints for service openshift-monitoring/metrics-server W0419 18:47:49.842099 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 10250: empty IP address endpoints for service openshift-monitoring/metrics-server W0419 18:47:49.980660 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 10250: empty IP address endpoints for service openshift-monitoring/metrics-server W0419 18:47:52.538575 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service openshift-console/downloads W0419 18:47:53.950837 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-insights/metrics I0419 18:47:54.392380 3645 base_network_controller_pods.go:487] [default/openshift-console/console-bff75d556-qjghg] creating logical port openshift-console_console-bff75d556-qjghg for pod on switch ip-10-0-2-180.ec2.internal I0419 18:47:54.392533 3645 kube.go:256] Updating pod openshift-console/console-bff75d556-qjghg I0419 18:47:54.409351 3645 pod.go:62] [openshift-console/console-bff75d556-qjghg] pod update took 16.852663ms I0419 18:47:54.409378 3645 base_network_controller_pods.go:951] [default/openshift-console/console-bff75d556-qjghg] addLogicalPort annotation time took 16.883952ms I0419 18:47:54.411193 3645 pods.go:271] [openshift-console/console-bff75d556-qjghg] addLogicalPort took 18.827304ms, libovsdb time 1.401728ms I0419 18:47:54.761510 3645 cni.go:452] [openshift-console/console-bff75d556-qjghg 916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea network default NAD default NAD key ] ADD starting CNI request [openshift-console/console-bff75d556-qjghg 916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea network default NAD default NAD key ] I0419 18:47:54.770629 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-console, podName: console-bff75d556-qjghg, hostIfaceName: 916dd044117aca0, network: default, NAD default, SandboxID: "916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea", PCI device ID: , UID: "f14b254d-6270-43b8-8b7e-4069077dc823", MAC: 0a:58:0a:82:00:17, IPs: [10.130.0.23/23] I0419 18:47:54.820556 3645 cni.go:473] [openshift-console/console-bff75d556-qjghg 916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea network default NAD default NAD key default] ADD finished CNI request [openshift-console/console-bff75d556-qjghg 916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"66:5b:76:4a:d5:c4\",\"name\":\"916dd044117aca0\"},{\"mac\":\"0a:58:0a:82:00:17\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/7f7adfe8-52da-475d-a6b8-16d253fea1e5\"}],\"ips\":[{\"address\":\"10.130.0.23/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:47:55.560061 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console/console W0419 18:47:55.983626 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-console/console W0419 18:48:02.509201 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 5000: empty IP address endpoints for service openshift-image-registry/image-registry W0419 18:48:02.509223 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 5000: empty IP address endpoints for service openshift-image-registry/image-registry W0419 18:48:03.106163 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-kube-storage-version-migrator-operator/metrics W0419 18:48:03.106224 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-kube-storage-version-migrator-operator/metrics W0419 18:48:03.106270 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-kube-storage-version-migrator-operator/metrics W0419 18:48:04.119871 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-kube-storage-version-migrator-operator/metrics W0419 18:48:05.134564 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console W0419 18:48:15.156742 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-service-ca-operator/metrics W0419 18:48:15.156782 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-service-ca-operator/metrics W0419 18:48:15.156822 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-service-ca-operator/metrics W0419 18:48:16.166478 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-service-ca-operator/metrics I0419 18:48:20.844469 3645 obj_retry.go:498] Detected object openshift-console/console-6d4854c664-sj4h2 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:20.845171 3645 obj_retry.go:498] Detected object openshift-console/console-6d4854c664-sj4h2 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:21.708076 3645 obj_retry.go:492] Detected object openshift-console/console-6d4854c664-sj4h2 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:21.708099 3645 obj_retry.go:492] Detected object openshift-console/console-6d4854c664-sj4h2 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:21.725206 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-6d4854c664-sj4h2 I0419 18:48:21.725234 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-6d4854c664-sj4h2 I0419 18:48:25.232653 3645 cni.go:452] [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d network default NAD default NAD key ] DEL starting CNI request [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d network default NAD default NAD key ] I0419 18:48:25.319603 3645 cni.go:473] [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d network default NAD default NAD key default] DEL finished CNI request [openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx d74ba2e7e7e473c51152de0bc505a9de50209d5df1ce86b2cd4deef07d0e6c6d network default NAD default NAD key default], result "{}", err I0419 18:48:25.376029 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:25.376065 3645 pods.go:176] Deleting pod: openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx I0419 18:48:25.378001 3645 pods.go:236] Attempting to release IPs for pod: openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx, ips: 10.130.0.15 I0419 18:48:25.378042 3645 obj_retry.go:498] Detected object openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:26.199618 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:26.199645 3645 obj_retry.go:492] Detected object openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:26.206955 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx I0419 18:48:26.206982 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-image-registry/image-registry-7f8cfbbcf5-dgtbx I0419 18:48:28.369158 3645 obj_retry.go:498] Detected object openshift-ingress/router-default-5f4bcd7fcc-9x8qd of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:28.369725 3645 obj_retry.go:498] Detected object openshift-ingress/router-default-5f4bcd7fcc-9x8qd of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:29.099020 3645 obj_retry.go:492] Detected object openshift-ingress/router-default-5f4bcd7fcc-9x8qd of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:29.099048 3645 obj_retry.go:492] Detected object openshift-ingress/router-default-5f4bcd7fcc-9x8qd of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:29.105932 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-ingress/router-default-5f4bcd7fcc-9x8qd I0419 18:48:29.105954 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-ingress/router-default-5f4bcd7fcc-9x8qd W0419 18:48:29.160948 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:48:29.160966 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:48:29.160976 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:48:29.160981 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:48:29.256161 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:48:29.256181 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/prometheus-k8s I0419 18:48:30.514375 3645 obj_retry.go:498] Detected object openshift-console/console-6d4854c664-czhrp of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:30.514938 3645 obj_retry.go:498] Detected object openshift-console/console-6d4854c664-czhrp of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:31.107932 3645 obj_retry.go:492] Detected object openshift-console/console-6d4854c664-czhrp of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:31.107959 3645 obj_retry.go:492] Detected object openshift-console/console-6d4854c664-czhrp of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:31.114233 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-6d4854c664-czhrp I0419 18:48:31.114257 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-6d4854c664-czhrp I0419 18:48:44.250574 3645 obj_retry.go:498] Detected object openshift-monitoring/alertmanager-main-1 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:44.251119 3645 obj_retry.go:498] Detected object openshift-monitoring/alertmanager-main-1 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:44.752678 3645 obj_retry.go:492] Detected object openshift-monitoring/alertmanager-main-1 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:44.752707 3645 obj_retry.go:492] Detected object openshift-monitoring/alertmanager-main-1 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:44.759702 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-monitoring/alertmanager-main-1 I0419 18:48:44.759723 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-monitoring/alertmanager-main-1 I0419 18:48:48.735880 3645 cni.go:452] [openshift-monitoring/prometheus-k8s-1 6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38 network default NAD default NAD key ] DEL starting CNI request [openshift-monitoring/prometheus-k8s-1 6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38 network default NAD default NAD key ] I0419 18:48:48.813940 3645 cni.go:473] [openshift-monitoring/prometheus-k8s-1 6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38 network default NAD default NAD key default] DEL finished CNI request [openshift-monitoring/prometheus-k8s-1 6803d8e34c4fb1cec0bcaad3f99b2287b5e8bb2cec970443950ef88fc1fb8d38 network default NAD default NAD key default], result "{}", err I0419 18:48:48.879695 3645 obj_retry.go:498] Detected object openshift-monitoring/prometheus-k8s-1 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:48.879734 3645 pods.go:176] Deleting pod: openshift-monitoring/prometheus-k8s-1 I0419 18:48:48.881090 3645 pods.go:236] Attempting to release IPs for pod: openshift-monitoring/prometheus-k8s-1, ips: 10.130.0.22 I0419 18:48:48.881120 3645 obj_retry.go:498] Detected object openshift-monitoring/prometheus-k8s-1 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:48:49.312299 3645 obj_retry.go:492] Detected object openshift-monitoring/prometheus-k8s-1 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:49.312321 3645 obj_retry.go:492] Detected object openshift-monitoring/prometheus-k8s-1 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:48:49.320680 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-monitoring/prometheus-k8s-1 I0419 18:48:49.320705 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-monitoring/prometheus-k8s-1 I0419 18:48:49.357597 3645 base_network_controller_pods.go:487] [default/openshift-monitoring/prometheus-k8s-1] creating logical port openshift-monitoring_prometheus-k8s-1 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:48:49.357747 3645 kube.go:256] Updating pod openshift-monitoring/prometheus-k8s-1 I0419 18:48:49.378509 3645 pod.go:62] [openshift-monitoring/prometheus-k8s-1] pod update took 20.843065ms I0419 18:48:49.378531 3645 base_network_controller_pods.go:951] [default/openshift-monitoring/prometheus-k8s-1] addLogicalPort annotation time took 20.871366ms I0419 18:48:49.379734 3645 pods.go:271] [openshift-monitoring/prometheus-k8s-1] addLogicalPort took 22.158248ms, libovsdb time 877.821µs I0419 18:48:49.720251 3645 cni.go:452] [openshift-monitoring/prometheus-k8s-1 b40aa4a8acccca33b1b94666d3ed69d691ff9398f307ae1bef82ea6476792d41 network default NAD default NAD key ] ADD starting CNI request [openshift-monitoring/prometheus-k8s-1 b40aa4a8acccca33b1b94666d3ed69d691ff9398f307ae1bef82ea6476792d41 network default NAD default NAD key ] I0419 18:48:49.730485 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-monitoring, podName: prometheus-k8s-1, hostIfaceName: b40aa4a8acccca3, network: default, NAD default, SandboxID: "b40aa4a8acccca33b1b94666d3ed69d691ff9398f307ae1bef82ea6476792d41", PCI device ID: , UID: "c4bdfe46-f72a-4eaf-a464-c12cf12f6ce3", MAC: 0a:58:0a:82:00:18, IPs: [10.130.0.24/23] I0419 18:48:49.782616 3645 cni.go:473] [openshift-monitoring/prometheus-k8s-1 b40aa4a8acccca33b1b94666d3ed69d691ff9398f307ae1bef82ea6476792d41 network default NAD default NAD key default] ADD finished CNI request [openshift-monitoring/prometheus-k8s-1 b40aa4a8acccca33b1b94666d3ed69d691ff9398f307ae1bef82ea6476792d41 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"7a:bd:42:12:0b:e0\",\"name\":\"b40aa4a8acccca3\"},{\"mac\":\"0a:58:0a:82:00:18\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/59b47099-3afa-4060-902c-4fb7598c12e0\"}],\"ips\":[{\"address\":\"10.130.0.24/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:48:54.363611 3645 cni.go:452] [openshift-multus/network-metrics-daemon-l99ph 013385d624def2f0044861675c75eae4e4baa67c22c37fcb347caa69041b28ae network default NAD default NAD key ] ADD starting CNI request [openshift-multus/network-metrics-daemon-l99ph 013385d624def2f0044861675c75eae4e4baa67c22c37fcb347caa69041b28ae network default NAD default NAD key ] I0419 18:48:54.373418 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-multus, podName: network-metrics-daemon-l99ph, hostIfaceName: 013385d624def2f, network: default, NAD default, SandboxID: "013385d624def2f0044861675c75eae4e4baa67c22c37fcb347caa69041b28ae", PCI device ID: , UID: "de0c68e2-469a-4ff1-9b56-b5586baac645", MAC: 0a:58:0a:82:00:03, IPs: [10.130.0.3/23] I0419 18:48:54.426949 3645 cni.go:473] [openshift-multus/network-metrics-daemon-l99ph 013385d624def2f0044861675c75eae4e4baa67c22c37fcb347caa69041b28ae network default NAD default NAD key default] ADD finished CNI request [openshift-multus/network-metrics-daemon-l99ph 013385d624def2f0044861675c75eae4e4baa67c22c37fcb347caa69041b28ae network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"c6:35:a4:ef:84:34\",\"name\":\"013385d624def2f\"},{\"mac\":\"0a:58:0a:82:00:03\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/7b017d87-26bc-47b6-9a70-4d1bc97470f4\"}],\"ips\":[{\"address\":\"10.130.0.3/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:49:06.527872 3645 obj_retry.go:498] Detected object openshift-monitoring/alertmanager-main-0 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:49:06.528467 3645 obj_retry.go:498] Detected object openshift-monitoring/alertmanager-main-0 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:49:07.271819 3645 obj_retry.go:492] Detected object openshift-monitoring/alertmanager-main-0 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:49:07.271844 3645 obj_retry.go:492] Detected object openshift-monitoring/alertmanager-main-0 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:49:07.282083 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-monitoring/alertmanager-main-0 I0419 18:49:07.282110 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-monitoring/alertmanager-main-0 I0419 18:49:16.316487 3645 obj_retry.go:498] Detected object openshift-console/console-bff75d556-n5mrk of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:49:16.317053 3645 obj_retry.go:498] Detected object openshift-console/console-bff75d556-n5mrk of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:49:16.915818 3645 obj_retry.go:492] Detected object openshift-console/console-bff75d556-n5mrk of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:49:16.915840 3645 obj_retry.go:492] Detected object openshift-console/console-bff75d556-n5mrk of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:49:16.922763 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-bff75d556-n5mrk I0419 18:49:16.922787 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-bff75d556-n5mrk I0419 18:49:27.491525 3645 cni.go:452] [openshift-console/console-bff75d556-qjghg 916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea network default NAD default NAD key ] DEL starting CNI request [openshift-console/console-bff75d556-qjghg 916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea network default NAD default NAD key ] I0419 18:49:27.580093 3645 cni.go:473] [openshift-console/console-bff75d556-qjghg 916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea network default NAD default NAD key default] DEL finished CNI request [openshift-console/console-bff75d556-qjghg 916dd044117aca0a731150cb26c076ff076f1b1abed14c209ed194a85d1666ea network default NAD default NAD key default], result "{}", err I0419 18:49:27.628255 3645 obj_retry.go:498] Detected object openshift-console/console-bff75d556-qjghg of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:49:27.628288 3645 pods.go:176] Deleting pod: openshift-console/console-bff75d556-qjghg I0419 18:49:27.629562 3645 pods.go:236] Attempting to release IPs for pod: openshift-console/console-bff75d556-qjghg, ips: 10.130.0.23 I0419 18:49:27.629590 3645 obj_retry.go:498] Detected object openshift-console/console-bff75d556-qjghg of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:49:28.486527 3645 obj_retry.go:492] Detected object openshift-console/console-bff75d556-qjghg of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:49:28.486569 3645 obj_retry.go:492] Detected object openshift-console/console-bff75d556-qjghg of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:49:28.498489 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-bff75d556-qjghg I0419 18:49:28.498515 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-bff75d556-qjghg W0419 18:49:50.610966 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9091: empty IP address endpoints for service openshift-monitoring/prometheus-k8s W0419 18:49:50.610995 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9092: empty IP address endpoints for service openshift-monitoring/prometheus-k8s I0419 18:49:51.918911 3645 obj_retry.go:498] Detected object openshift-monitoring/prometheus-k8s-0 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:49:51.919516 3645 obj_retry.go:498] Detected object openshift-monitoring/prometheus-k8s-0 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:49:52.473366 3645 obj_retry.go:492] Detected object openshift-monitoring/prometheus-k8s-0 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:49:52.473428 3645 obj_retry.go:492] Detected object openshift-monitoring/prometheus-k8s-0 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:49:52.484119 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-monitoring/prometheus-k8s-0 I0419 18:49:52.484151 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-monitoring/prometheus-k8s-0 I0419 18:50:00.232574 3645 base_network_controller_pods.go:487] [default/openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2] creating logical port openshift-backplane_osd-delete-backplane-serviceaccounts-29610410-8gpt2 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:50:00.232667 3645 kube.go:256] Updating pod openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 I0419 18:50:00.246155 3645 pod.go:62] [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2] pod update took 13.514857ms I0419 18:50:00.246183 3645 base_network_controller_pods.go:951] [default/openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2] addLogicalPort annotation time took 13.547241ms I0419 18:50:00.247415 3645 pods.go:271] [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2] addLogicalPort took 14.843994ms, libovsdb time 862.079µs I0419 18:50:00.612598 3645 cni.go:452] [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684 network default NAD default NAD key ] ADD starting CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684 network default NAD default NAD key ] I0419 18:50:00.621735 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-backplane, podName: osd-delete-backplane-serviceaccounts-29610410-8gpt2, hostIfaceName: 5ff52970eb55f10, network: default, NAD default, SandboxID: "5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684", PCI device ID: , UID: "846c08f9-32d3-4ea1-a8eb-a8eaba7f2c99", MAC: 0a:58:0a:82:00:19, IPs: [10.130.0.25/23] I0419 18:50:00.672967 3645 cni.go:473] [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684 network default NAD default NAD key default] ADD finished CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"d6:36:6e:57:74:61\",\"name\":\"5ff52970eb55f10\"},{\"mac\":\"0a:58:0a:82:00:19\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/311b9a1d-50cc-43b0-ba1d-89e7e3175420\"}],\"ips\":[{\"address\":\"10.130.0.25/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:50:01.080132 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610400-tvb5j of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:50:01.080734 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610400-tvb5j of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:50:01.102147 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610400-tvb5j of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:50:01.102174 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610400-tvb5j of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:50:01.113783 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-backplane/osd-delete-backplane-serviceaccounts-29610400-tvb5j I0419 18:50:01.113806 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-backplane/osd-delete-backplane-serviceaccounts-29610400-tvb5j I0419 18:50:04.626633 3645 cni.go:452] [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684 network default NAD default NAD key ] DEL starting CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684 network default NAD default NAD key ] I0419 18:50:04.715855 3645 cni.go:473] [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684 network default NAD default NAD key default] DEL finished CNI request [openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 5ff52970eb55f10ff539c89ab65e3aa564b39c951eb4fb5ca546a4567b43a684 network default NAD default NAD key default], result "{}", err I0419 18:50:04.764442 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:50:04.764487 3645 pods.go:176] Deleting pod: openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 I0419 18:50:04.765798 3645 pods.go:236] Attempting to release IPs for pod: openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2, ips: 10.130.0.25 I0419 18:50:04.765849 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:50:05.639631 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:50:05.639665 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed W0419 18:50:15.141766 3645 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-deployment-validation-operator/deployment-validation-operator-metrics I0419 18:50:23.750344 3645 namespace.go:142] [openshift-kube-storage-version-migrator] updating namespace I0419 18:50:26.178774 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-deployment-validation-operator/allow-from-openshift-insights to delete 1 local pods... W0419 18:50:26.178860 3645 base_network_controller_namespace.go:458] Unable to remove remote zone pod's openshift-deployment-validation-operator/deployment-validation-operator-847f6f7b8-n8pjq IP address from the namespace address-set, err: pod openshift-deployment-validation-operator/deployment-validation-operator-847f6f7b8-n8pjq: no pod IPs found I0419 18:50:27.187019 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=7 I0419 18:50:34.209323 3645 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=7 I0419 18:50:54.215819 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=5 I0419 18:51:09.177223 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=8 I0419 18:51:11.193337 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=8 I0419 18:51:14.216023 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=6 I0419 18:51:18.088683 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=123 I0419 18:51:21.002501 3645 base_network_controller_pods.go:487] [default/openshift-console/console-5c758fc7dd-9bdrg] creating logical port openshift-console_console-5c758fc7dd-9bdrg for pod on switch ip-10-0-2-180.ec2.internal I0419 18:51:21.002621 3645 kube.go:256] Updating pod openshift-console/console-5c758fc7dd-9bdrg I0419 18:51:21.030499 3645 pod.go:62] [openshift-console/console-5c758fc7dd-9bdrg] pod update took 27.910764ms I0419 18:51:21.030526 3645 base_network_controller_pods.go:951] [default/openshift-console/console-5c758fc7dd-9bdrg] addLogicalPort annotation time took 27.942385ms I0419 18:51:21.031710 3645 pods.go:271] [openshift-console/console-5c758fc7dd-9bdrg] addLogicalPort took 29.230389ms, libovsdb time 836.933µs I0419 18:51:21.359805 3645 cni.go:452] [openshift-console/console-5c758fc7dd-9bdrg c5d21c2576ead520387b9d2ac033700ba4357c3b5f6e458a463c917b87f538c5 network default NAD default NAD key ] ADD starting CNI request [openshift-console/console-5c758fc7dd-9bdrg c5d21c2576ead520387b9d2ac033700ba4357c3b5f6e458a463c917b87f538c5 network default NAD default NAD key ] I0419 18:51:21.369583 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-console, podName: console-5c758fc7dd-9bdrg, hostIfaceName: c5d21c2576ead52, network: default, NAD default, SandboxID: "c5d21c2576ead520387b9d2ac033700ba4357c3b5f6e458a463c917b87f538c5", PCI device ID: , UID: "6dd4c951-427b-4bb6-b91d-6395bb7c69ff", MAC: 0a:58:0a:82:00:1a, IPs: [10.130.0.26/23] I0419 18:51:21.420493 3645 cni.go:473] [openshift-console/console-5c758fc7dd-9bdrg c5d21c2576ead520387b9d2ac033700ba4357c3b5f6e458a463c917b87f538c5 network default NAD default NAD key default] ADD finished CNI request [openshift-console/console-5c758fc7dd-9bdrg c5d21c2576ead520387b9d2ac033700ba4357c3b5f6e458a463c917b87f538c5 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"46:68:73:be:89:18\",\"name\":\"c5d21c2576ead52\"},{\"mac\":\"0a:58:0a:82:00:1a\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/e137a94c-1161-452c-a300-6f876307afdb\"}],\"ips\":[{\"address\":\"10.130.0.26/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:51:31.977859 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-console/console I0419 18:51:36.181797 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=7 I0419 18:51:46.313583 3645 obj_retry.go:498] Detected object openshift-console/console-85f7f99c6-fxpqn of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:51:46.314199 3645 obj_retry.go:498] Detected object openshift-console/console-85f7f99c6-fxpqn of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:51:46.457705 3645 obj_retry.go:492] Detected object openshift-console/console-85f7f99c6-fxpqn of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:51:46.457758 3645 obj_retry.go:492] Detected object openshift-console/console-85f7f99c6-fxpqn of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:51:46.464459 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-85f7f99c6-fxpqn I0419 18:51:46.464491 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-85f7f99c6-fxpqn I0419 18:51:57.312201 3645 obj_retry.go:498] Detected object openshift-console/console-85f7f99c6-lvgbn of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:51:57.312757 3645 obj_retry.go:498] Detected object openshift-console/console-85f7f99c6-lvgbn of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:51:57.972907 3645 obj_retry.go:492] Detected object openshift-console/console-85f7f99c6-lvgbn of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:51:57.972934 3645 obj_retry.go:492] Detected object openshift-console/console-85f7f99c6-lvgbn of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:51:57.992270 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-console/console-85f7f99c6-lvgbn I0419 18:51:57.992298 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-console/console-85f7f99c6-lvgbn I0419 18:52:24.203773 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=8 I0419 18:52:41.176890 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=8 I0419 18:53:30.093749 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=16 I0419 18:53:40.102131 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=573 I0419 18:54:13.198697 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=11 I0419 18:54:14.088345 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=10 I0419 18:54:32.087808 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=61 I0419 18:54:35.088903 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=72 I0419 18:56:09.129748 3645 namespace.go:98] [tsf] adding namespace I0419 18:56:09.130674 3645 namespace.go:102] [tsf] adding namespace took 900.61µs I0419 18:56:09.151194 3645 namespace.go:142] [tsf] updating namespace I0419 18:56:09.161622 3645 namespace.go:142] [tsf] updating namespace I0419 18:56:12.873255 3645 namespace.go:98] [rhtpa-operator] adding namespace I0419 18:56:12.874118 3645 namespace.go:102] [rhtpa-operator] adding namespace took 836.767µs I0419 18:56:12.880833 3645 namespace.go:98] [konflux-ui] adding namespace I0419 18:56:12.880836 3645 namespace.go:98] [cert-manager-operator] adding namespace I0419 18:56:12.881606 3645 namespace.go:102] [konflux-ui] adding namespace took 755.377µs I0419 18:56:12.882198 3645 namespace.go:102] [cert-manager-operator] adding namespace took 1.340497ms I0419 18:56:12.886489 3645 namespace.go:98] [konflux-operator] adding namespace I0419 18:56:12.886626 3645 namespace.go:98] [tsf-keycloak] adding namespace I0419 18:56:12.886665 3645 namespace.go:98] [tsf-tas] adding namespace I0419 18:56:12.887177 3645 namespace.go:102] [konflux-operator] adding namespace took 670.11µs I0419 18:56:12.887203 3645 namespace.go:98] [tsf-tpa] adding namespace I0419 18:56:12.887896 3645 namespace.go:102] [tsf-tpa] adding namespace took 684.446µs I0419 18:56:12.888525 3645 namespace.go:102] [tsf-keycloak] adding namespace took 1.840353ms I0419 18:56:12.888980 3645 namespace.go:102] [tsf-tas] adding namespace took 2.287037ms I0419 18:56:12.890295 3645 namespace.go:98] [rhbk-operator] adding namespace I0419 18:56:12.890992 3645 namespace.go:102] [rhbk-operator] adding namespace took 684.722µs I0419 18:56:12.913136 3645 namespace.go:142] [rhtpa-operator] updating namespace I0419 18:56:12.926625 3645 namespace.go:142] [rhtpa-operator] updating namespace I0419 18:56:12.950663 3645 namespace.go:142] [konflux-ui] updating namespace I0419 18:56:12.950696 3645 namespace.go:142] [konflux-operator] updating namespace I0419 18:56:12.954930 3645 namespace.go:142] [cert-manager-operator] updating namespace I0419 18:56:12.959952 3645 namespace.go:142] [tsf-tpa] updating namespace I0419 18:56:12.971241 3645 namespace.go:142] [tsf-keycloak] updating namespace I0419 18:56:12.971278 3645 namespace.go:142] [konflux-ui] updating namespace I0419 18:56:12.972013 3645 namespace.go:142] [tsf-tas] updating namespace I0419 18:56:12.979912 3645 namespace.go:142] [rhtpa-operator] updating namespace I0419 18:56:12.985155 3645 namespace.go:142] [rhbk-operator] updating namespace I0419 18:56:13.047844 3645 namespace.go:142] [konflux-operator] updating namespace I0419 18:56:13.061455 3645 namespace.go:142] [konflux-operator] updating namespace I0419 18:56:13.089128 3645 namespace.go:142] [tsf-tpa] updating namespace I0419 18:56:13.095043 3645 namespace.go:142] [konflux-ui] updating namespace I0419 18:56:13.100790 3645 namespace.go:142] [tsf-tpa] updating namespace I0419 18:56:13.112638 3645 namespace.go:142] [tsf-keycloak] updating namespace I0419 18:56:13.123461 3645 namespace.go:142] [tsf-keycloak] updating namespace I0419 18:56:13.154261 3645 namespace.go:142] [tsf-tas] updating namespace I0419 18:56:13.167087 3645 namespace.go:142] [tsf-tas] updating namespace I0419 18:56:13.175476 3645 namespace.go:142] [rhbk-operator] updating namespace I0419 18:56:13.202570 3645 namespace.go:142] [rhbk-operator] updating namespace I0419 18:56:13.264919 3645 namespace.go:142] [cert-manager-operator] updating namespace I0419 18:56:13.310092 3645 namespace.go:142] [cert-manager-operator] updating namespace I0419 18:56:14.247364 3645 namespace.go:142] [tsf-tpa] updating namespace I0419 18:56:14.255938 3645 namespace.go:142] [tsf-keycloak] updating namespace I0419 18:56:20.133215 3645 obj_retry.go:498] Detected object openshift-marketplace/53634cc716e3784f8e237b128ae14253d65dd14b0e9dcd086f5ff88ece59ftq of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:20.133759 3645 obj_retry.go:498] Detected object openshift-marketplace/53634cc716e3784f8e237b128ae14253d65dd14b0e9dcd086f5ff88ece59ftq of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:20.133785 3645 obj_retry.go:498] Detected object openshift-marketplace/53634cc716e3784f8e237b128ae14253d65dd14b0e9dcd086f5ff88ece59ftq of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:20.133794 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0419 18:56:20.133806 3645 obj_retry.go:498] Detected object openshift-marketplace/53634cc716e3784f8e237b128ae14253d65dd14b0e9dcd086f5ff88ece59ftq of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:20.133810 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0419 18:56:21.131963 3645 obj_retry.go:498] Detected object openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5zcr4f of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:21.132561 3645 obj_retry.go:498] Detected object openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5zcr4f of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:21.132585 3645 obj_retry.go:498] Detected object openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5zcr4f of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:21.132594 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0419 18:56:21.132610 3645 obj_retry.go:498] Detected object openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5zcr4f of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:21.132617 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0419 18:56:21.160304 3645 obj_retry.go:492] Detected object openshift-marketplace/53634cc716e3784f8e237b128ae14253d65dd14b0e9dcd086f5ff88ece59ftq of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:21.160329 3645 obj_retry.go:492] Detected object openshift-marketplace/53634cc716e3784f8e237b128ae14253d65dd14b0e9dcd086f5ff88ece59ftq of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:21.160342 3645 obj_retry.go:492] Detected object openshift-marketplace/53634cc716e3784f8e237b128ae14253d65dd14b0e9dcd086f5ff88ece59ftq of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:21.160355 3645 obj_retry.go:492] Detected object openshift-marketplace/53634cc716e3784f8e237b128ae14253d65dd14b0e9dcd086f5ff88ece59ftq of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:22.165121 3645 obj_retry.go:492] Detected object openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5zcr4f of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:22.165141 3645 obj_retry.go:492] Detected object openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5zcr4f of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:22.165153 3645 obj_retry.go:492] Detected object openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5zcr4f of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:22.165171 3645 obj_retry.go:492] Detected object openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5zcr4f of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:24.161849 3645 obj_retry.go:498] Detected object openshift-marketplace/f79378b9836da0a180436b794e6453783590867ba696f35668b1ac4568k2kr2 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:24.162367 3645 obj_retry.go:498] Detected object openshift-marketplace/f79378b9836da0a180436b794e6453783590867ba696f35668b1ac4568k2kr2 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:24.162392 3645 obj_retry.go:498] Detected object openshift-marketplace/f79378b9836da0a180436b794e6453783590867ba696f35668b1ac4568k2kr2 of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:24.162422 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0419 18:56:24.162437 3645 obj_retry.go:498] Detected object openshift-marketplace/f79378b9836da0a180436b794e6453783590867ba696f35668b1ac4568k2kr2 of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:24.162441 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0419 18:56:25.169385 3645 obj_retry.go:498] Detected object openshift-marketplace/a54ee9eda2828a12414ea519ebe1984caa1e89a473cf1b21256364ccf2rzg5w of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:25.169974 3645 obj_retry.go:498] Detected object openshift-marketplace/a54ee9eda2828a12414ea519ebe1984caa1e89a473cf1b21256364ccf2rzg5w of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:25.170003 3645 obj_retry.go:498] Detected object openshift-marketplace/a54ee9eda2828a12414ea519ebe1984caa1e89a473cf1b21256364ccf2rzg5w of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:25.170011 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0419 18:56:25.170022 3645 obj_retry.go:498] Detected object openshift-marketplace/a54ee9eda2828a12414ea519ebe1984caa1e89a473cf1b21256364ccf2rzg5w of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:25.170027 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0419 18:56:25.189317 3645 obj_retry.go:492] Detected object openshift-marketplace/f79378b9836da0a180436b794e6453783590867ba696f35668b1ac4568k2kr2 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:25.189342 3645 obj_retry.go:492] Detected object openshift-marketplace/f79378b9836da0a180436b794e6453783590867ba696f35668b1ac4568k2kr2 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:25.189358 3645 obj_retry.go:492] Detected object openshift-marketplace/f79378b9836da0a180436b794e6453783590867ba696f35668b1ac4568k2kr2 of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:25.189370 3645 obj_retry.go:492] Detected object openshift-marketplace/f79378b9836da0a180436b794e6453783590867ba696f35668b1ac4568k2kr2 of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:26.259229 3645 obj_retry.go:492] Detected object openshift-marketplace/a54ee9eda2828a12414ea519ebe1984caa1e89a473cf1b21256364ccf2rzg5w of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:26.259253 3645 obj_retry.go:492] Detected object openshift-marketplace/a54ee9eda2828a12414ea519ebe1984caa1e89a473cf1b21256364ccf2rzg5w of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:26.259269 3645 obj_retry.go:492] Detected object openshift-marketplace/a54ee9eda2828a12414ea519ebe1984caa1e89a473cf1b21256364ccf2rzg5w of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:26.259281 3645 obj_retry.go:492] Detected object openshift-marketplace/a54ee9eda2828a12414ea519ebe1984caa1e89a473cf1b21256364ccf2rzg5w of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:27.190258 3645 obj_retry.go:498] Detected object openshift-marketplace/18ab2eaa607a00575555d3d5b97fe3206aa96262fc8b848ad325c51ad7crscj of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:27.190795 3645 obj_retry.go:498] Detected object openshift-marketplace/18ab2eaa607a00575555d3d5b97fe3206aa96262fc8b848ad325c51ad7crscj of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:27.190822 3645 obj_retry.go:498] Detected object openshift-marketplace/18ab2eaa607a00575555d3d5b97fe3206aa96262fc8b848ad325c51ad7crscj of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:27.190834 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0419 18:56:27.190849 3645 obj_retry.go:498] Detected object openshift-marketplace/18ab2eaa607a00575555d3d5b97fe3206aa96262fc8b848ad325c51ad7crscj of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:27.190857 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0419 18:56:28.192495 3645 obj_retry.go:498] Detected object openshift-marketplace/30e1fee4a656579f055e90bd50374338b32290af440ed32e709fbffaafgl2q9 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:28.193025 3645 obj_retry.go:498] Detected object openshift-marketplace/30e1fee4a656579f055e90bd50374338b32290af440ed32e709fbffaafgl2q9 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:28.193056 3645 obj_retry.go:498] Detected object openshift-marketplace/30e1fee4a656579f055e90bd50374338b32290af440ed32e709fbffaafgl2q9 of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:28.193064 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/unpack-bundles to delete 1 local pods... I0419 18:56:28.193079 3645 obj_retry.go:498] Detected object openshift-marketplace/30e1fee4a656579f055e90bd50374338b32290af440ed32e709fbffaafgl2q9 of type *factory.localPodSelector in terminal state (e.g. completed) during update event: will remove it I0419 18:56:28.193084 3645 base_network_controller_policy.go:642] Processing NetworkPolicy openshift-marketplace/default-deny-all to delete 1 local pods... I0419 18:56:28.219821 3645 obj_retry.go:492] Detected object openshift-marketplace/18ab2eaa607a00575555d3d5b97fe3206aa96262fc8b848ad325c51ad7crscj of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:28.219846 3645 obj_retry.go:492] Detected object openshift-marketplace/18ab2eaa607a00575555d3d5b97fe3206aa96262fc8b848ad325c51ad7crscj of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:28.219866 3645 obj_retry.go:492] Detected object openshift-marketplace/18ab2eaa607a00575555d3d5b97fe3206aa96262fc8b848ad325c51ad7crscj of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:28.219881 3645 obj_retry.go:492] Detected object openshift-marketplace/18ab2eaa607a00575555d3d5b97fe3206aa96262fc8b848ad325c51ad7crscj of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:29.228870 3645 obj_retry.go:492] Detected object openshift-marketplace/30e1fee4a656579f055e90bd50374338b32290af440ed32e709fbffaafgl2q9 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:29.228896 3645 obj_retry.go:492] Detected object openshift-marketplace/30e1fee4a656579f055e90bd50374338b32290af440ed32e709fbffaafgl2q9 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:29.228909 3645 obj_retry.go:492] Detected object openshift-marketplace/30e1fee4a656579f055e90bd50374338b32290af440ed32e709fbffaafgl2q9 of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:29.228926 3645 obj_retry.go:492] Detected object openshift-marketplace/30e1fee4a656579f055e90bd50374338b32290af440ed32e709fbffaafgl2q9 of type *factory.localPodSelector in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:31.145119 3645 base_network_controller_pods.go:487] [default/tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr] creating logical port tsf-tpa_rhtpa-operator-controller-manager-576c85d8f-rnjrr for pod on switch ip-10-0-2-180.ec2.internal I0419 18:56:31.145243 3645 kube.go:256] Updating pod tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr I0419 18:56:31.172202 3645 pod.go:62] [tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr] pod update took 26.983456ms I0419 18:56:31.172230 3645 base_network_controller_pods.go:951] [default/tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr] addLogicalPort annotation time took 27.018079ms I0419 18:56:31.173423 3645 pods.go:271] [tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr] addLogicalPort took 28.321047ms, libovsdb time 826.63µs I0419 18:56:31.503430 3645 cni.go:452] [tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr b369cc5967536da52babdcf16de8f85fca97a7c34c9bd6d47cb39288e2783e5b network default NAD default NAD key ] ADD starting CNI request [tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr b369cc5967536da52babdcf16de8f85fca97a7c34c9bd6d47cb39288e2783e5b network default NAD default NAD key ] I0419 18:56:31.512740 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tpa, podName: rhtpa-operator-controller-manager-576c85d8f-rnjrr, hostIfaceName: b369cc5967536da, network: default, NAD default, SandboxID: "b369cc5967536da52babdcf16de8f85fca97a7c34c9bd6d47cb39288e2783e5b", PCI device ID: , UID: "706d1a44-3094-40e2-8e41-0d2af90db7fe", MAC: 0a:58:0a:82:00:1b, IPs: [10.130.0.27/23] I0419 18:56:31.562593 3645 cni.go:473] [tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr b369cc5967536da52babdcf16de8f85fca97a7c34c9bd6d47cb39288e2783e5b network default NAD default NAD key default] ADD finished CNI request [tsf-tpa/rhtpa-operator-controller-manager-576c85d8f-rnjrr b369cc5967536da52babdcf16de8f85fca97a7c34c9bd6d47cb39288e2783e5b network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"2e:d9:88:e7:72:42\",\"name\":\"b369cc5967536da\"},{\"mac\":\"0a:58:0a:82:00:1b\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/fac5c92a-ad74-43ab-ba62-a0bbb7407e07\"}],\"ips\":[{\"address\":\"10.130.0.27/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:56:33.449182 3645 namespace.go:98] [cert-manager] adding namespace I0419 18:56:33.450015 3645 namespace.go:102] [cert-manager] adding namespace took 810.313µs I0419 18:56:33.519954 3645 namespace.go:142] [cert-manager] updating namespace I0419 18:56:33.564231 3645 namespace.go:142] [cert-manager] updating namespace W0419 18:56:35.199993 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service W0419 18:56:38.139147 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service W0419 18:56:38.146619 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service I0419 18:56:38.362659 3645 base_network_controller_pods.go:487] [default/cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj] creating logical port cert-manager_cert-manager-webhook-fd44f5cbb-vwbbj for pod on switch ip-10-0-2-180.ec2.internal I0419 18:56:38.362755 3645 kube.go:256] Updating pod cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj I0419 18:56:38.376235 3645 pod.go:62] [cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj] pod update took 13.504276ms I0419 18:56:38.376256 3645 base_network_controller_pods.go:951] [default/cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj] addLogicalPort annotation time took 13.530866ms I0419 18:56:38.377504 3645 pods.go:271] [cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj] addLogicalPort took 14.865182ms, libovsdb time 854.064µs I0419 18:56:38.738310 3645 cni.go:452] [cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj 71aca5701a89efc3148691ef75dafc90edd3793e77bee7232b9302a1d9d8fd7f network default NAD default NAD key ] ADD starting CNI request [cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj 71aca5701a89efc3148691ef75dafc90edd3793e77bee7232b9302a1d9d8fd7f network default NAD default NAD key ] I0419 18:56:38.747578 3645 helper_linux.go:556] ConfigureOVS: namespace: cert-manager, podName: cert-manager-webhook-fd44f5cbb-vwbbj, hostIfaceName: 71aca5701a89efc, network: default, NAD default, SandboxID: "71aca5701a89efc3148691ef75dafc90edd3793e77bee7232b9302a1d9d8fd7f", PCI device ID: , UID: "ca76d547-4e02-483e-9352-5423ee83f75f", MAC: 0a:58:0a:82:00:1c, IPs: [10.130.0.28/23] I0419 18:56:38.800155 3645 cni.go:473] [cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj 71aca5701a89efc3148691ef75dafc90edd3793e77bee7232b9302a1d9d8fd7f network default NAD default NAD key default] ADD finished CNI request [cert-manager/cert-manager-webhook-fd44f5cbb-vwbbj 71aca5701a89efc3148691ef75dafc90edd3793e77bee7232b9302a1d9d8fd7f network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"66:af:cb:19:73:d8\",\"name\":\"71aca5701a89efc\"},{\"mac\":\"0a:58:0a:82:00:1c\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/a018f254-4ba4-4cca-a0e8-ccabbbbc36a6\"}],\"ips\":[{\"address\":\"10.130.0.28/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:56:39.179433 3645 base_network_controller_pods.go:487] [default/cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv] creating logical port cert-manager_cert-manager-cainjector-5b4bf68748-4fmxv for pod on switch ip-10-0-2-180.ec2.internal I0419 18:56:39.179528 3645 kube.go:256] Updating pod cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv I0419 18:56:39.194141 3645 pod.go:62] [cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv] pod update took 14.640001ms I0419 18:56:39.194166 3645 base_network_controller_pods.go:951] [default/cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv] addLogicalPort annotation time took 14.667301ms I0419 18:56:39.195429 3645 pods.go:271] [cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv] addLogicalPort took 16.030305ms, libovsdb time 819.242µs I0419 18:56:39.543290 3645 cni.go:452] [cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv afa37716d5928b8bb2906a0640839abec3e40c8040c474a687ab2c4c3de25d95 network default NAD default NAD key ] ADD starting CNI request [cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv afa37716d5928b8bb2906a0640839abec3e40c8040c474a687ab2c4c3de25d95 network default NAD default NAD key ] I0419 18:56:39.553979 3645 helper_linux.go:556] ConfigureOVS: namespace: cert-manager, podName: cert-manager-cainjector-5b4bf68748-4fmxv, hostIfaceName: afa37716d5928b8, network: default, NAD default, SandboxID: "afa37716d5928b8bb2906a0640839abec3e40c8040c474a687ab2c4c3de25d95", PCI device ID: , UID: "c92c8f21-10b3-4874-8c20-3cb6f023a76a", MAC: 0a:58:0a:82:00:1d, IPs: [10.130.0.29/23] I0419 18:56:39.605486 3645 cni.go:473] [cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv afa37716d5928b8bb2906a0640839abec3e40c8040c474a687ab2c4c3de25d95 network default NAD default NAD key default] ADD finished CNI request [cert-manager/cert-manager-cainjector-5b4bf68748-4fmxv afa37716d5928b8bb2906a0640839abec3e40c8040c474a687ab2c4c3de25d95 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"06:ed:aa:fd:d5:bd\",\"name\":\"afa37716d5928b8\"},{\"mac\":\"0a:58:0a:82:00:1d\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/e4bd18eb-6fb9-4c01-93ba-314bbe2591e6\"}],\"ips\":[{\"address\":\"10.130.0.29/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:56:43.232189 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service cert-manager/cert-manager-webhook W0419 18:56:43.232210 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service cert-manager/cert-manager-webhook I0419 18:56:44.401665 3645 namespace.go:142] [openshift-operators] updating namespace I0419 18:56:44.418092 3645 namespace.go:142] [openshift-operators] updating namespace I0419 18:56:45.172122 3645 base_network_controller_pods.go:487] [default/rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d] creating logical port rhbk-operator_rhbk-operator-fb86cd8d9-jkw9d for pod on switch ip-10-0-2-180.ec2.internal I0419 18:56:45.172233 3645 kube.go:256] Updating pod rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d I0419 18:56:45.193625 3645 pod.go:62] [rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d] pod update took 21.418861ms I0419 18:56:45.193649 3645 base_network_controller_pods.go:951] [default/rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d] addLogicalPort annotation time took 21.446612ms I0419 18:56:45.194791 3645 pods.go:271] [rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d] addLogicalPort took 22.689163ms, libovsdb time 795.191µs I0419 18:56:45.535943 3645 cni.go:452] [rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d 8a2ae8561159fa519d82ecf11b5c22c4823b9b59c964cc058d020d0197a0fb67 network default NAD default NAD key ] ADD starting CNI request [rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d 8a2ae8561159fa519d82ecf11b5c22c4823b9b59c964cc058d020d0197a0fb67 network default NAD default NAD key ] I0419 18:56:45.546380 3645 helper_linux.go:556] ConfigureOVS: namespace: rhbk-operator, podName: rhbk-operator-fb86cd8d9-jkw9d, hostIfaceName: 8a2ae8561159fa5, network: default, NAD default, SandboxID: "8a2ae8561159fa519d82ecf11b5c22c4823b9b59c964cc058d020d0197a0fb67", PCI device ID: , UID: "0ac8c801-aef9-4445-824e-8cc0a5de2a3c", MAC: 0a:58:0a:82:00:1e, IPs: [10.130.0.30/23] I0419 18:56:45.598520 3645 cni.go:473] [rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d 8a2ae8561159fa519d82ecf11b5c22c4823b9b59c964cc058d020d0197a0fb67 network default NAD default NAD key default] ADD finished CNI request [rhbk-operator/rhbk-operator-fb86cd8d9-jkw9d 8a2ae8561159fa519d82ecf11b5c22c4823b9b59c964cc058d020d0197a0fb67 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"de:0a:3e:fa:5e:44\",\"name\":\"8a2ae8561159fa5\"},{\"mac\":\"0a:58:0a:82:00:1e\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/123845c5-9125-4a52-ae93-ba9f24b57d6a\"}],\"ips\":[{\"address\":\"10.130.0.30/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:56:46.220446 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service tsf-tpa/rhtpa-operator-controller-manager-metrics-service W0419 18:56:47.198655 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service konflux-operator/konflux-operator-controller-manager-metrics-service W0419 18:56:48.575856 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service konflux-operator/konflux-operator-controller-manager-metrics-service W0419 18:56:48.600363 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service konflux-operator/konflux-operator-controller-manager-metrics-service W0419 18:56:48.637210 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service konflux-operator/konflux-operator-controller-manager-metrics-service W0419 18:56:48.655449 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service konflux-operator/konflux-operator-controller-manager-metrics-service W0419 18:56:49.272640 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 10250: empty IP address endpoints for service cert-manager/cert-manager-webhook W0419 18:56:49.272660 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9402: empty IP address endpoints for service cert-manager/cert-manager-webhook I0419 18:56:51.962740 3645 namespace.go:142] [openshift-operators] updating namespace I0419 18:56:52.312063 3645 base_network_controller_pods.go:487] [default/openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq] creating logical port openshift-operators_openshift-pipelines-operator-6684678d86-l8ptq for pod on switch ip-10-0-2-180.ec2.internal I0419 18:56:52.312176 3645 kube.go:256] Updating pod openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq I0419 18:56:52.346300 3645 pod.go:62] [openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq] pod update took 34.148944ms I0419 18:56:52.346322 3645 base_network_controller_pods.go:951] [default/openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq] addLogicalPort annotation time took 34.176042ms I0419 18:56:52.347621 3645 pods.go:271] [openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq] addLogicalPort took 35.574372ms, libovsdb time 948.049µs I0419 18:56:52.674086 3645 cni.go:452] [openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq 3809dc62e9c352309f65b2b50611fcb2675028ac333d39b108dc8b8dd8d0e916 network default NAD default NAD key ] ADD starting CNI request [openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq 3809dc62e9c352309f65b2b50611fcb2675028ac333d39b108dc8b8dd8d0e916 network default NAD default NAD key ] I0419 18:56:52.684722 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-operators, podName: openshift-pipelines-operator-6684678d86-l8ptq, hostIfaceName: 3809dc62e9c3523, network: default, NAD default, SandboxID: "3809dc62e9c352309f65b2b50611fcb2675028ac333d39b108dc8b8dd8d0e916", PCI device ID: , UID: "2ebca803-60c1-4455-8a9e-7c22790d4802", MAC: 0a:58:0a:82:00:1f, IPs: [10.130.0.31/23] I0419 18:56:52.738031 3645 cni.go:473] [openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq 3809dc62e9c352309f65b2b50611fcb2675028ac333d39b108dc8b8dd8d0e916 network default NAD default NAD key default] ADD finished CNI request [openshift-operators/openshift-pipelines-operator-6684678d86-l8ptq 3809dc62e9c352309f65b2b50611fcb2675028ac333d39b108dc8b8dd8d0e916 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"42:a1:b9:fa:e9:2a\",\"name\":\"3809dc62e9c3523\"},{\"mac\":\"0a:58:0a:82:00:1f\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/3086aa42-d0dd-4ed1-b84b-750bdabc91cd\"}],\"ips\":[{\"address\":\"10.130.0.31/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:56:54.312217 3645 base_network_controller_pods.go:487] [default/cert-manager/cert-manager-59ccb4bb6b-qcknv] creating logical port cert-manager_cert-manager-59ccb4bb6b-qcknv for pod on switch ip-10-0-2-180.ec2.internal I0419 18:56:54.312356 3645 kube.go:256] Updating pod cert-manager/cert-manager-59ccb4bb6b-qcknv I0419 18:56:54.353136 3645 pod.go:62] [cert-manager/cert-manager-59ccb4bb6b-qcknv] pod update took 40.816042ms I0419 18:56:54.353164 3645 base_network_controller_pods.go:951] [default/cert-manager/cert-manager-59ccb4bb6b-qcknv] addLogicalPort annotation time took 40.851787ms I0419 18:56:54.354580 3645 pods.go:271] [cert-manager/cert-manager-59ccb4bb6b-qcknv] addLogicalPort took 42.381184ms, libovsdb time 953.894µs I0419 18:56:55.625373 3645 cni.go:452] [cert-manager/cert-manager-59ccb4bb6b-qcknv 5c97dfbce8bb05204ef9473f7a094a9a2bc73eab4b8be0d99410f709be1f5a90 network default NAD default NAD key ] ADD starting CNI request [cert-manager/cert-manager-59ccb4bb6b-qcknv 5c97dfbce8bb05204ef9473f7a094a9a2bc73eab4b8be0d99410f709be1f5a90 network default NAD default NAD key ] I0419 18:56:55.636141 3645 helper_linux.go:556] ConfigureOVS: namespace: cert-manager, podName: cert-manager-59ccb4bb6b-qcknv, hostIfaceName: 5c97dfbce8bb052, network: default, NAD default, SandboxID: "5c97dfbce8bb05204ef9473f7a094a9a2bc73eab4b8be0d99410f709be1f5a90", PCI device ID: , UID: "902a2c9d-8cff-47a3-89ad-d1b290afb8b9", MAC: 0a:58:0a:82:00:20, IPs: [10.130.0.32/23] I0419 18:56:55.695236 3645 cni.go:473] [cert-manager/cert-manager-59ccb4bb6b-qcknv 5c97dfbce8bb05204ef9473f7a094a9a2bc73eab4b8be0d99410f709be1f5a90 network default NAD default NAD key default] ADD finished CNI request [cert-manager/cert-manager-59ccb4bb6b-qcknv 5c97dfbce8bb05204ef9473f7a094a9a2bc73eab4b8be0d99410f709be1f5a90 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"86:c4:d7:08:a3:7f\",\"name\":\"5c97dfbce8bb052\"},{\"mac\":\"0a:58:0a:82:00:20\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/faff91b1-20e5-4610-bd24-23bce23bade8\"}],\"ips\":[{\"address\":\"10.130.0.32/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:56:59.753244 3645 obj_retry.go:498] Detected object tsf/test-tsf-subscriptions of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:59.753791 3645 obj_retry.go:498] Detected object tsf/test-tsf-subscriptions of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:56:59.793155 3645 obj_retry.go:492] Detected object tsf/test-tsf-subscriptions of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:59.793178 3645 obj_retry.go:492] Detected object tsf/test-tsf-subscriptions of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:56:59.805202 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf/test-tsf-subscriptions I0419 18:56:59.805236 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf/test-tsf-subscriptions I0419 18:57:00.711243 3645 namespace.go:98] [trusted-artifact-signer] adding namespace I0419 18:57:00.712176 3645 namespace.go:102] [trusted-artifact-signer] adding namespace took 909.753µs I0419 18:57:00.744185 3645 namespace.go:142] [trusted-artifact-signer] updating namespace I0419 18:57:00.778311 3645 namespace.go:142] [trusted-artifact-signer] updating namespace W0419 18:57:01.331032 3645 util.go:870] Failed to build global endpoints for port TCP/metrics: empty IP address endpoints for service openshift-operators/rhtas-operator-controller-manager-metrics-service W0419 18:57:16.676088 3645 util.go:870] Failed to build global endpoints for port TCP/data: empty IP address endpoints for service tsf-keycloak/keycloak-pgsql W0419 18:57:16.699027 3645 util.go:870] Failed to build global endpoints for port TCP/data: empty IP address endpoints for service tsf-tpa/tpa-pgsql I0419 18:57:16.728965 3645 namespace.go:98] [openshift-pipelines] adding namespace I0419 18:57:16.729875 3645 namespace.go:102] [openshift-pipelines] adding namespace took 885.844µs I0419 18:57:16.750130 3645 namespace.go:142] [openshift-pipelines] updating namespace I0419 18:57:17.520121 3645 namespace.go:142] [cert-manager] updating namespace I0419 18:57:17.714635 3645 namespace.go:142] [cert-manager-operator] updating namespace I0419 18:57:18.113759 3645 namespace.go:142] [konflux-operator] updating namespace I0419 18:57:18.313637 3645 namespace.go:142] [konflux-ui] updating namespace I0419 18:57:18.517182 3645 namespace.go:142] [open-cluster-management-2ppd21iub53sh96ramq6ul1mg6plu4eu] updating namespace I0419 18:57:18.919490 3645 namespace.go:142] [rhbk-operator] updating namespace I0419 18:57:19.118549 3645 namespace.go:142] [rhtpa-operator] updating namespace I0419 18:57:19.319679 3645 namespace.go:142] [trusted-artifact-signer] updating namespace I0419 18:57:19.519383 3645 namespace.go:142] [tsf] updating namespace I0419 18:57:19.722555 3645 namespace.go:142] [tsf-keycloak] updating namespace I0419 18:57:19.923150 3645 namespace.go:142] [tsf-tas] updating namespace I0419 18:57:20.124031 3645 namespace.go:142] [tsf-tpa] updating namespace I0419 18:57:20.258776 3645 namespace.go:142] [cert-manager] updating namespace I0419 18:57:20.269587 3645 namespace.go:142] [cert-manager-operator] updating namespace I0419 18:57:20.280698 3645 namespace.go:142] [default] updating namespace I0419 18:57:20.289907 3645 namespace.go:142] [konflux-operator] updating namespace I0419 18:57:20.300122 3645 namespace.go:142] [konflux-ui] updating namespace I0419 18:57:20.310176 3645 namespace.go:142] [open-cluster-management-2ppd21iub53sh96ramq6ul1mg6plu4eu] updating namespace I0419 18:57:20.318169 3645 namespace.go:142] [openshift] updating namespace I0419 18:57:20.326588 3645 namespace.go:142] [rhbk-operator] updating namespace I0419 18:57:20.337915 3645 namespace.go:142] [rhtpa-operator] updating namespace I0419 18:57:20.346608 3645 namespace.go:142] [trusted-artifact-signer] updating namespace I0419 18:57:20.356997 3645 namespace.go:142] [tsf] updating namespace I0419 18:57:20.367955 3645 namespace.go:142] [tsf-keycloak] updating namespace I0419 18:57:20.390824 3645 namespace.go:142] [tsf-tas] updating namespace I0419 18:57:20.406932 3645 namespace.go:142] [tsf-tpa] updating namespace I0419 18:57:20.508573 3645 namespace.go:142] [cert-manager] updating namespace I0419 18:57:20.615680 3645 namespace.go:142] [cert-manager-operator] updating namespace I0419 18:57:20.707335 3645 namespace.go:142] [default] updating namespace I0419 18:57:20.808994 3645 namespace.go:142] [konflux-operator] updating namespace I0419 18:57:20.909583 3645 namespace.go:142] [konflux-ui] updating namespace I0419 18:57:21.010703 3645 namespace.go:142] [open-cluster-management-2ppd21iub53sh96ramq6ul1mg6plu4eu] updating namespace I0419 18:57:21.109586 3645 namespace.go:142] [openshift] updating namespace I0419 18:57:21.212067 3645 namespace.go:142] [rhbk-operator] updating namespace I0419 18:57:21.309020 3645 namespace.go:142] [rhtpa-operator] updating namespace I0419 18:57:21.408089 3645 namespace.go:142] [trusted-artifact-signer] updating namespace I0419 18:57:21.511641 3645 namespace.go:142] [tsf] updating namespace I0419 18:57:21.609834 3645 namespace.go:142] [tsf-keycloak] updating namespace I0419 18:57:21.706748 3645 namespace.go:142] [tsf-tas] updating namespace I0419 18:57:21.809750 3645 namespace.go:142] [tsf-tpa] updating namespace I0419 18:57:22.077430 3645 namespace.go:142] [openshift-pipelines] updating namespace W0419 18:57:35.530357 3645 util.go:870] Failed to build global endpoints for port TCP/probes: empty IP address endpoints for service openshift-pipelines/tekton-events-controller W0419 18:57:35.530376 3645 util.go:870] Failed to build global endpoints for port TCP/http-profiling: empty IP address endpoints for service openshift-pipelines/tekton-events-controller W0419 18:57:35.530385 3645 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/tekton-events-controller W0419 18:57:35.591453 3645 util.go:870] Failed to build global endpoints for port TCP/probes: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-controller W0419 18:57:35.591474 3645 util.go:870] Failed to build global endpoints for port TCP/http-profiling: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-controller W0419 18:57:35.591479 3645 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-controller W0419 18:57:35.791916 3645 util.go:870] Failed to build global endpoints for port TCP/https-webhook: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 18:57:35.791935 3645 util.go:870] Failed to build global endpoints for port TCP/http-profiling: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 18:57:35.791939 3645 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 18:57:35.791944 3645 util.go:870] Failed to build global endpoints for port TCP/probes: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook I0419 18:57:42.180008 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=7 I0419 18:57:48.963511 3645 obj_retry.go:498] Detected object tsf/test-tsf-infrastructure of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:57:48.964110 3645 obj_retry.go:498] Detected object tsf/test-tsf-infrastructure of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:57:48.994722 3645 obj_retry.go:492] Detected object tsf/test-tsf-infrastructure of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:57:48.994751 3645 obj_retry.go:492] Detected object tsf/test-tsf-infrastructure of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:57:49.001753 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf/test-tsf-infrastructure I0419 18:57:49.001775 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf/test-tsf-infrastructure I0419 18:57:49.091230 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=60 I0419 18:57:50.981053 3645 base_network_controller_policy.go:1179] Adding network policy tsf-keycloak/keycloak-network-policy for network default I0419 18:57:50.981709 3645 pod_selector_address_set.go:213] Created shared address set for pod selector tsf-keycloak_LS{ML:{app: keycloak,app.kubernetes.io/instance: keycloak,app.kubernetes.io/managed-by: keycloak-operator,},} I0419 18:57:50.981730 3645 base_network_controller_policy.go:1054] Policy tsf-keycloak/keycloak-network-policy added to peer address sets [tsf-keycloak_LS{ML:{app: keycloak,app.kubernetes.io/instance: keycloak,app.kubernetes.io/managed-by: keycloak-operator,},}] I0419 18:57:50.984085 3645 base_network_controller_policy.go:1232] Create network policy tsf-keycloak/keycloak-network-policy resources completed, update namespace loglevel W0419 18:57:56.898218 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-keycloak/keycloak W0419 18:57:56.898239 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-keycloak/keycloak W0419 18:57:56.898221 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service tsf-keycloak/keycloak-service W0419 18:57:56.898293 3645 util.go:870] Failed to build global endpoints for port TCP/management: empty IP address endpoints for service tsf-keycloak/keycloak-service W0419 18:57:57.576581 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service openshift-pipelines/tekton-triggers-core-interceptors I0419 18:58:10.563143 3645 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll] creating logical port openshift-pipelines_tekton-chains-controller-5fb999ffff-khbll for pod on switch ip-10-0-2-180.ec2.internal I0419 18:58:10.563245 3645 kube.go:256] Updating pod openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll I0419 18:58:10.579557 3645 pod.go:62] [openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll] pod update took 16.334873ms I0419 18:58:10.579580 3645 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll] addLogicalPort annotation time took 16.364936ms I0419 18:58:10.580899 3645 pods.go:271] [openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll] addLogicalPort took 17.771714ms, libovsdb time 943.859µs I0419 18:58:10.923805 3645 cni.go:452] [openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll a7ebf1479e928a5dd8a7d65b0be2514859ccbc3e0c8c9fbb70e93048385d18dc network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll a7ebf1479e928a5dd8a7d65b0be2514859ccbc3e0c8c9fbb70e93048385d18dc network default NAD default NAD key ] I0419 18:58:10.934108 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-chains-controller-5fb999ffff-khbll, hostIfaceName: a7ebf1479e928a5, network: default, NAD default, SandboxID: "a7ebf1479e928a5dd8a7d65b0be2514859ccbc3e0c8c9fbb70e93048385d18dc", PCI device ID: , UID: "fa7bf4a8-52dc-486e-9b8c-4ada56d9c704", MAC: 0a:58:0a:82:00:21, IPs: [10.130.0.33/23] I0419 18:58:10.986992 3645 cni.go:473] [openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll a7ebf1479e928a5dd8a7d65b0be2514859ccbc3e0c8c9fbb70e93048385d18dc network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-chains-controller-5fb999ffff-khbll a7ebf1479e928a5dd8a7d65b0be2514859ccbc3e0c8c9fbb70e93048385d18dc network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"b6:c3:d6:91:00:28\",\"name\":\"a7ebf1479e928a5\"},{\"mac\":\"0a:58:0a:82:00:21\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/6d2d4631-a739-4e28-aa8b-f9fb4227c14f\"}],\"ips\":[{\"address\":\"10.130.0.33/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:58:15.691300 3645 iptables.go:108] Creating table: nat chain: OVN-KUBE-NODEPORT I0419 18:58:18.811538 3645 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-results-postgres-0] creating logical port openshift-pipelines_tekton-results-postgres-0 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:58:18.811638 3645 kube.go:256] Updating pod openshift-pipelines/tekton-results-postgres-0 I0419 18:58:18.824346 3645 pod.go:62] [openshift-pipelines/tekton-results-postgres-0] pod update took 12.730159ms I0419 18:58:18.824373 3645 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-results-postgres-0] addLogicalPort annotation time took 12.763496ms I0419 18:58:18.825793 3645 pods.go:271] [openshift-pipelines/tekton-results-postgres-0] addLogicalPort took 14.273101ms, libovsdb time 1.005957ms I0419 18:58:23.070111 3645 cni.go:452] [openshift-pipelines/tekton-results-postgres-0 6ab517bf13e9006d2522bae7a7f3b6b1a6e61dfdb67d0d88e30161cbc3bac3cd network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-results-postgres-0 6ab517bf13e9006d2522bae7a7f3b6b1a6e61dfdb67d0d88e30161cbc3bac3cd network default NAD default NAD key ] I0419 18:58:23.080040 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-results-postgres-0, hostIfaceName: 6ab517bf13e9006, network: default, NAD default, SandboxID: "6ab517bf13e9006d2522bae7a7f3b6b1a6e61dfdb67d0d88e30161cbc3bac3cd", PCI device ID: , UID: "a924d11a-61f0-45b7-9efb-e3b312dc0b52", MAC: 0a:58:0a:82:00:22, IPs: [10.130.0.34/23] I0419 18:58:23.132653 3645 cni.go:473] [openshift-pipelines/tekton-results-postgres-0 6ab517bf13e9006d2522bae7a7f3b6b1a6e61dfdb67d0d88e30161cbc3bac3cd network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-results-postgres-0 6ab517bf13e9006d2522bae7a7f3b6b1a6e61dfdb67d0d88e30161cbc3bac3cd network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"26:6a:c4:8a:8e:69\",\"name\":\"6ab517bf13e9006\"},{\"mac\":\"0a:58:0a:82:00:22\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/d12dbf8f-cd46-4b7e-b145-cb5e737009da\"}],\"ips\":[{\"address\":\"10.130.0.34/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:58:23.176683 3645 obj_retry.go:498] Detected object tsf/test-tsf-iam of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:58:23.177210 3645 obj_retry.go:498] Detected object tsf/test-tsf-iam of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:58:23.218792 3645 obj_retry.go:492] Detected object tsf/test-tsf-iam of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:58:23.218815 3645 obj_retry.go:492] Detected object tsf/test-tsf-iam of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:58:23.225607 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf/test-tsf-iam I0419 18:58:23.225627 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf/test-tsf-iam I0419 18:58:23.935918 3645 namespace.go:142] [openshift-pipelines] updating namespace I0419 18:58:24.957201 3645 base_network_controller_pods.go:487] [default/tsf-tas/segment-backup-installation-nklcg-2m7k6] creating logical port tsf-tas_segment-backup-installation-nklcg-2m7k6 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:58:24.957303 3645 kube.go:256] Updating pod tsf-tas/segment-backup-installation-nklcg-2m7k6 I0419 18:58:24.969935 3645 pod.go:62] [tsf-tas/segment-backup-installation-nklcg-2m7k6] pod update took 12.657515ms I0419 18:58:24.970035 3645 base_network_controller_pods.go:951] [default/tsf-tas/segment-backup-installation-nklcg-2m7k6] addLogicalPort annotation time took 12.758034ms I0419 18:58:24.971666 3645 pods.go:271] [tsf-tas/segment-backup-installation-nklcg-2m7k6] addLogicalPort took 14.481968ms, libovsdb time 1.041719ms I0419 18:58:25.844863 3645 base_network_controller_pods.go:487] [default/tsf-tas/ctlog-createtree-job-7h8nc-qn6gd] creating logical port tsf-tas_ctlog-createtree-job-7h8nc-qn6gd for pod on switch ip-10-0-2-180.ec2.internal I0419 18:58:25.844986 3645 kube.go:256] Updating pod tsf-tas/ctlog-createtree-job-7h8nc-qn6gd I0419 18:58:25.858190 3645 pod.go:62] [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd] pod update took 13.24035ms I0419 18:58:25.858223 3645 base_network_controller_pods.go:951] [default/tsf-tas/ctlog-createtree-job-7h8nc-qn6gd] addLogicalPort annotation time took 13.279487ms I0419 18:58:25.859788 3645 pods.go:271] [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd] addLogicalPort took 14.9442ms, libovsdb time 976.063µs I0419 18:58:26.243191 3645 base_network_controller_pods.go:487] [default/tsf-tas/trillian-logserver-574f7d45bb-8vcrg] creating logical port tsf-tas_trillian-logserver-574f7d45bb-8vcrg for pod on switch ip-10-0-2-180.ec2.internal I0419 18:58:26.243324 3645 kube.go:256] Updating pod tsf-tas/trillian-logserver-574f7d45bb-8vcrg I0419 18:58:26.262202 3645 pod.go:62] [tsf-tas/trillian-logserver-574f7d45bb-8vcrg] pod update took 18.938043ms I0419 18:58:26.262227 3645 base_network_controller_pods.go:951] [default/tsf-tas/trillian-logserver-574f7d45bb-8vcrg] addLogicalPort annotation time took 18.968912ms I0419 18:58:26.263715 3645 pods.go:271] [tsf-tas/trillian-logserver-574f7d45bb-8vcrg] addLogicalPort took 20.544296ms, libovsdb time 896.658µs I0419 18:58:26.428152 3645 base_network_controller_pods.go:487] [default/tsf-tas/trillian-logsigner-6cd6c547fd-lxllp] creating logical port tsf-tas_trillian-logsigner-6cd6c547fd-lxllp for pod on switch ip-10-0-2-180.ec2.internal I0419 18:58:26.428263 3645 kube.go:256] Updating pod tsf-tas/trillian-logsigner-6cd6c547fd-lxllp I0419 18:58:26.446475 3645 pod.go:62] [tsf-tas/trillian-logsigner-6cd6c547fd-lxllp] pod update took 18.240899ms I0419 18:58:26.446505 3645 base_network_controller_pods.go:951] [default/tsf-tas/trillian-logsigner-6cd6c547fd-lxllp] addLogicalPort annotation time took 18.275979ms I0419 18:58:26.447761 3645 pods.go:271] [tsf-tas/trillian-logsigner-6cd6c547fd-lxllp] addLogicalPort took 19.634632ms, libovsdb time 844.493µs I0419 18:58:27.949116 3645 cni.go:452] [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd 193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f network default NAD default NAD key ] ADD starting CNI request [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd 193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f network default NAD default NAD key ] I0419 18:58:27.956002 3645 cni.go:452] [tsf-tas/segment-backup-installation-nklcg-2m7k6 cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758 network default NAD default NAD key ] ADD starting CNI request [tsf-tas/segment-backup-installation-nklcg-2m7k6 cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758 network default NAD default NAD key ] I0419 18:58:27.960576 3645 cni.go:452] [tsf-tas/trillian-logsigner-6cd6c547fd-lxllp 36abbfa783707f3ced16fb2242a380c8a36d0345a89cced812e675e5de8b30ce network default NAD default NAD key ] ADD starting CNI request [tsf-tas/trillian-logsigner-6cd6c547fd-lxllp 36abbfa783707f3ced16fb2242a380c8a36d0345a89cced812e675e5de8b30ce network default NAD default NAD key ] I0419 18:58:27.960778 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: ctlog-createtree-job-7h8nc-qn6gd, hostIfaceName: 193a811ee5aee75, network: default, NAD default, SandboxID: "193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f", PCI device ID: , UID: "7e853277-57bc-40a7-a0ec-411dff7ee151", MAC: 0a:58:0a:82:00:24, IPs: [10.130.0.36/23] I0419 18:58:27.961753 3645 cni.go:452] [tsf-tas/trillian-logserver-574f7d45bb-8vcrg 100a2b35bb54cd0d5ef4d3fb6daedfb9a7315d3a878e5dc041b73ae8203a2e30 network default NAD default NAD key ] ADD starting CNI request [tsf-tas/trillian-logserver-574f7d45bb-8vcrg 100a2b35bb54cd0d5ef4d3fb6daedfb9a7315d3a878e5dc041b73ae8203a2e30 network default NAD default NAD key ] I0419 18:58:27.966790 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: segment-backup-installation-nklcg-2m7k6, hostIfaceName: cf59e5f0cc22071, network: default, NAD default, SandboxID: "cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758", PCI device ID: , UID: "00cd205f-ae9a-4f0a-93a3-eb723dfa3060", MAC: 0a:58:0a:82:00:23, IPs: [10.130.0.35/23] I0419 18:58:27.976447 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: trillian-logserver-574f7d45bb-8vcrg, hostIfaceName: 100a2b35bb54cd0, network: default, NAD default, SandboxID: "100a2b35bb54cd0d5ef4d3fb6daedfb9a7315d3a878e5dc041b73ae8203a2e30", PCI device ID: , UID: "6f05bf15-fc06-4857-b1d1-7b7381ba46e6", MAC: 0a:58:0a:82:00:25, IPs: [10.130.0.37/23] I0419 18:58:27.977994 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: trillian-logsigner-6cd6c547fd-lxllp, hostIfaceName: 36abbfa783707f3, network: default, NAD default, SandboxID: "36abbfa783707f3ced16fb2242a380c8a36d0345a89cced812e675e5de8b30ce", PCI device ID: , UID: "0b031b08-a3e7-4175-9ad2-42180f97fd6e", MAC: 0a:58:0a:82:00:26, IPs: [10.130.0.38/23] I0419 18:58:28.034984 3645 cni.go:473] [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd 193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f network default NAD default NAD key default] ADD finished CNI request [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd 193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"9a:b8:65:4c:2a:e0\",\"name\":\"193a811ee5aee75\"},{\"mac\":\"0a:58:0a:82:00:24\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/75b2f5fc-8247-4c70-baa7-dc028b91cd30\"}],\"ips\":[{\"address\":\"10.130.0.36/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:58:28.264060 3645 cni.go:473] [tsf-tas/trillian-logserver-574f7d45bb-8vcrg 100a2b35bb54cd0d5ef4d3fb6daedfb9a7315d3a878e5dc041b73ae8203a2e30 network default NAD default NAD key default] ADD finished CNI request [tsf-tas/trillian-logserver-574f7d45bb-8vcrg 100a2b35bb54cd0d5ef4d3fb6daedfb9a7315d3a878e5dc041b73ae8203a2e30 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"72:d2:23:4d:9d:83\",\"name\":\"100a2b35bb54cd0\"},{\"mac\":\"0a:58:0a:82:00:25\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/74538914-39ad-42db-9261-5de53f3e38bb\"}],\"ips\":[{\"address\":\"10.130.0.37/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:58:28.264153 3645 cni.go:473] [tsf-tas/segment-backup-installation-nklcg-2m7k6 cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758 network default NAD default NAD key default] ADD finished CNI request [tsf-tas/segment-backup-installation-nklcg-2m7k6 cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"56:27:09:77:2d:1c\",\"name\":\"cf59e5f0cc22071\"},{\"mac\":\"0a:58:0a:82:00:23\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/1bea8b07-c602-4ed1-a5bb-de6eb3eb8d25\"}],\"ips\":[{\"address\":\"10.130.0.35/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:58:28.264328 3645 cni.go:473] [tsf-tas/trillian-logsigner-6cd6c547fd-lxllp 36abbfa783707f3ced16fb2242a380c8a36d0345a89cced812e675e5de8b30ce network default NAD default NAD key default] ADD finished CNI request [tsf-tas/trillian-logsigner-6cd6c547fd-lxllp 36abbfa783707f3ced16fb2242a380c8a36d0345a89cced812e675e5de8b30ce network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"66:ed:8c:b6:1a:63\",\"name\":\"36abbfa783707f3\"},{\"mac\":\"0a:58:0a:82:00:26\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/0a6d4e65-f22c-4c13-99f0-d58b517d3429\"}],\"ips\":[{\"address\":\"10.130.0.38/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:58:28.718455 3645 util.go:870] Failed to build global endpoints for port TCP/grpc: empty IP address endpoints for service tsf-tas/fulcio-server W0419 18:58:28.718479 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/fulcio-server I0419 18:58:28.759399 3645 iptables.go:108] Creating table: nat chain: OVN-KUBE-NODEPORT I0419 18:58:30.205587 3645 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4] creating logical port openshift-pipelines_tekton-results-retention-policy-agent-55cbb85cb8-sjvf4 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:58:30.205709 3645 kube.go:256] Updating pod openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4 I0419 18:58:30.223004 3645 pod.go:62] [openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4] pod update took 17.326353ms I0419 18:58:30.223036 3645 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4] addLogicalPort annotation time took 17.362677ms I0419 18:58:30.224839 3645 pods.go:271] [openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4] addLogicalPort took 19.270916ms, libovsdb time 1.219425ms I0419 18:58:30.295625 3645 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv] creating logical port openshift-pipelines_tekton-results-watcher-6f946f97cf-lwkfv for pod on switch ip-10-0-2-180.ec2.internal I0419 18:58:30.295838 3645 kube.go:256] Updating pod openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv I0419 18:58:30.311209 3645 pod.go:62] [openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv] pod update took 15.408327ms I0419 18:58:30.311242 3645 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv] addLogicalPort annotation time took 15.445668ms I0419 18:58:30.313845 3645 pods.go:271] [openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv] addLogicalPort took 18.262328ms, libovsdb time 2.012283ms W0419 18:58:34.755240 3645 util.go:870] Failed to build global endpoints for port TCP/prometheus: empty IP address endpoints for service openshift-pipelines/tekton-results-api-service W0419 18:58:34.755265 3645 util.go:870] Failed to build global endpoints for port TCP/profiling: empty IP address endpoints for service openshift-pipelines/tekton-results-api-service W0419 18:58:34.755272 3645 util.go:870] Failed to build global endpoints for port TCP/server: empty IP address endpoints for service openshift-pipelines/tekton-results-api-service W0419 18:58:36.775684 3645 util.go:870] Failed to build global endpoints for port TCP/trillian-mysql: empty IP address endpoints for service tsf-tas/trillian-mysql I0419 18:58:39.196564 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=8 I0419 18:58:39.263303 3645 obj_retry.go:498] Detected object tsf-keycloak/tsf-iam-4xnzw of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:58:39.263902 3645 obj_retry.go:498] Detected object tsf-keycloak/tsf-iam-4xnzw of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:58:40.163175 3645 obj_retry.go:492] Detected object tsf-keycloak/tsf-iam-4xnzw of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:58:40.163208 3645 obj_retry.go:492] Detected object tsf-keycloak/tsf-iam-4xnzw of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:58:41.319912 3645 cni.go:452] [openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv d87db8c5dcc69fdbb3a98a5a98f40aa74c2c3287f48752e0c2484fabff5a48b6 network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv d87db8c5dcc69fdbb3a98a5a98f40aa74c2c3287f48752e0c2484fabff5a48b6 network default NAD default NAD key ] I0419 18:58:41.331152 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-results-watcher-6f946f97cf-lwkfv, hostIfaceName: d87db8c5dcc69fd, network: default, NAD default, SandboxID: "d87db8c5dcc69fdbb3a98a5a98f40aa74c2c3287f48752e0c2484fabff5a48b6", PCI device ID: , UID: "54406f0e-6e95-40f1-85f9-c2694102c7e0", MAC: 0a:58:0a:82:00:28, IPs: [10.130.0.40/23] I0419 18:58:41.392525 3645 cni.go:473] [openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv d87db8c5dcc69fdbb3a98a5a98f40aa74c2c3287f48752e0c2484fabff5a48b6 network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-results-watcher-6f946f97cf-lwkfv d87db8c5dcc69fdbb3a98a5a98f40aa74c2c3287f48752e0c2484fabff5a48b6 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"5e:ff:e7:12:ac:bf\",\"name\":\"d87db8c5dcc69fd\"},{\"mac\":\"0a:58:0a:82:00:28\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/d2042ad9-7201-4786-a8c6-de33c723eb2e\"}],\"ips\":[{\"address\":\"10.130.0.40/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:58:44.046555 3645 cni.go:452] [openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4 a9342a2fd9d3b77527e6cc51ed34c8211bf65c8595d5f884e232f41c30fcf84d network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4 a9342a2fd9d3b77527e6cc51ed34c8211bf65c8595d5f884e232f41c30fcf84d network default NAD default NAD key ] I0419 18:58:44.074776 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-results-retention-policy-agent-55cbb85cb8-sjvf4, hostIfaceName: a9342a2fd9d3b77, network: default, NAD default, SandboxID: "a9342a2fd9d3b77527e6cc51ed34c8211bf65c8595d5f884e232f41c30fcf84d", PCI device ID: , UID: "9e4ed30d-cbb3-49ac-8bb5-9a58b47a9e12", MAC: 0a:58:0a:82:00:27, IPs: [10.130.0.39/23] I0419 18:58:44.133472 3645 cni.go:473] [openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4 a9342a2fd9d3b77527e6cc51ed34c8211bf65c8595d5f884e232f41c30fcf84d network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-results-retention-policy-agent-55cbb85cb8-sjvf4 a9342a2fd9d3b77527e6cc51ed34c8211bf65c8595d5f884e232f41c30fcf84d network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"3a:07:4d:3c:b2:99\",\"name\":\"a9342a2fd9d3b77\"},{\"mac\":\"0a:58:0a:82:00:27\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/7324df06-a817-4dda-82a3-738f6635c220\"}],\"ips\":[{\"address\":\"10.130.0.39/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:58:45.219059 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=8 W0419 18:58:45.812717 3645 util.go:870] Failed to build global endpoints for port TCP/grpc: empty IP address endpoints for service tsf-tas/trillian-logsigner W0419 18:58:46.825669 3645 util.go:870] Failed to build global endpoints for port TCP/grpc: empty IP address endpoints for service tsf-tas/trillian-logserver I0419 18:58:49.847609 3645 cni.go:452] [tsf-tas/segment-backup-installation-nklcg-2m7k6 cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758 network default NAD default NAD key ] DEL starting CNI request [tsf-tas/segment-backup-installation-nklcg-2m7k6 cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758 network default NAD default NAD key ] I0419 18:58:49.946373 3645 cni.go:473] [tsf-tas/segment-backup-installation-nklcg-2m7k6 cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758 network default NAD default NAD key default] DEL finished CNI request [tsf-tas/segment-backup-installation-nklcg-2m7k6 cf59e5f0cc22071530918870104fbbe8719b3d191a2de7c77800163803456758 network default NAD default NAD key default], result "{}", err I0419 18:58:49.999628 3645 obj_retry.go:498] Detected object tsf-tas/segment-backup-installation-nklcg-2m7k6 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:58:49.999737 3645 pods.go:176] Deleting pod: tsf-tas/segment-backup-installation-nklcg-2m7k6 I0419 18:58:50.001202 3645 pods.go:236] Attempting to release IPs for pod: tsf-tas/segment-backup-installation-nklcg-2m7k6, ips: 10.130.0.35 I0419 18:58:50.001232 3645 obj_retry.go:498] Detected object tsf-tas/segment-backup-installation-nklcg-2m7k6 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:58:50.887647 3645 obj_retry.go:492] Detected object tsf-tas/segment-backup-installation-nklcg-2m7k6 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:58:50.887672 3645 obj_retry.go:492] Detected object tsf-tas/segment-backup-installation-nklcg-2m7k6 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed W0419 18:59:04.888676 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8091: empty IP address endpoints for service tsf-tas/trillian-logserver W0419 18:59:04.915662 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8091: empty IP address endpoints for service tsf-tas/trillian-logsigner I0419 18:59:13.505455 3645 obj_retry.go:498] Detected object tsf-tas/test-tsf-tas of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:13.506016 3645 obj_retry.go:498] Detected object tsf-tas/test-tsf-tas of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:13.526090 3645 obj_retry.go:492] Detected object tsf-tas/test-tsf-tas of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:13.526109 3645 obj_retry.go:492] Detected object tsf-tas/test-tsf-tas of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:13.540702 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tas/test-tsf-tas I0419 18:59:13.540726 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tas/test-tsf-tas W0419 18:59:14.305715 3645 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-controller W0419 18:59:14.305738 3645 util.go:870] Failed to build global endpoints for port TCP/http-listener: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-controller W0419 18:59:14.356109 3645 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-watcher I0419 18:59:15.990360 3645 cni.go:452] [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd 193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f network default NAD default NAD key ] DEL starting CNI request [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd 193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f network default NAD default NAD key ] I0419 18:59:16.078753 3645 cni.go:473] [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd 193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f network default NAD default NAD key default] DEL finished CNI request [tsf-tas/ctlog-createtree-job-7h8nc-qn6gd 193a811ee5aee75742ebfbd51e53917543edfa4d9c8f6fb180c67b82f67e7b3f network default NAD default NAD key default], result "{}", err I0419 18:59:16.228068 3645 obj_retry.go:498] Detected object tsf-tas/ctlog-createtree-job-7h8nc-qn6gd of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:16.228108 3645 pods.go:176] Deleting pod: tsf-tas/ctlog-createtree-job-7h8nc-qn6gd I0419 18:59:16.229538 3645 pods.go:236] Attempting to release IPs for pod: tsf-tas/ctlog-createtree-job-7h8nc-qn6gd, ips: 10.130.0.36 I0419 18:59:16.229565 3645 obj_retry.go:498] Detected object tsf-tas/ctlog-createtree-job-7h8nc-qn6gd of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:17.003829 3645 obj_retry.go:492] Detected object tsf-tas/ctlog-createtree-job-7h8nc-qn6gd of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:17.003852 3645 obj_retry.go:492] Detected object tsf-tas/ctlog-createtree-job-7h8nc-qn6gd of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:19.160616 3645 obj_retry.go:498] Detected object tsf-tas/rekor-createtree-job-l2r9c-wh8q5 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:19.161171 3645 obj_retry.go:498] Detected object tsf-tas/rekor-createtree-job-l2r9c-wh8q5 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:20.035618 3645 obj_retry.go:492] Detected object tsf-tas/rekor-createtree-job-l2r9c-wh8q5 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:20.035643 3645 obj_retry.go:492] Detected object tsf-tas/rekor-createtree-job-l2r9c-wh8q5 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:21.140677 3645 base_network_controller_pods.go:487] [default/tsf-tas/ctlog-bb98b7cf6-x6cvz] creating logical port tsf-tas_ctlog-bb98b7cf6-x6cvz for pod on switch ip-10-0-2-180.ec2.internal I0419 18:59:21.140846 3645 kube.go:256] Updating pod tsf-tas/ctlog-bb98b7cf6-x6cvz I0419 18:59:21.166857 3645 pod.go:62] [tsf-tas/ctlog-bb98b7cf6-x6cvz] pod update took 26.031111ms I0419 18:59:21.166882 3645 base_network_controller_pods.go:951] [default/tsf-tas/ctlog-bb98b7cf6-x6cvz] addLogicalPort annotation time took 26.063405ms I0419 18:59:21.168181 3645 pods.go:271] [tsf-tas/ctlog-bb98b7cf6-x6cvz] addLogicalPort took 27.522468ms, libovsdb time 836.187µs I0419 18:59:21.796157 3645 cni.go:452] [tsf-tas/ctlog-bb98b7cf6-x6cvz b35143b60d5fc8a734edb4d6457e28fd467490cf90b8bbfb225a635b4543ad1b network default NAD default NAD key ] ADD starting CNI request [tsf-tas/ctlog-bb98b7cf6-x6cvz b35143b60d5fc8a734edb4d6457e28fd467490cf90b8bbfb225a635b4543ad1b network default NAD default NAD key ] I0419 18:59:21.806289 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tas, podName: ctlog-bb98b7cf6-x6cvz, hostIfaceName: b35143b60d5fc8a, network: default, NAD default, SandboxID: "b35143b60d5fc8a734edb4d6457e28fd467490cf90b8bbfb225a635b4543ad1b", PCI device ID: , UID: "ec67498d-88e7-45a1-9a88-ebb46bec0c5c", MAC: 0a:58:0a:82:00:29, IPs: [10.130.0.41/23] I0419 18:59:21.860589 3645 cni.go:473] [tsf-tas/ctlog-bb98b7cf6-x6cvz b35143b60d5fc8a734edb4d6457e28fd467490cf90b8bbfb225a635b4543ad1b network default NAD default NAD key default] ADD finished CNI request [tsf-tas/ctlog-bb98b7cf6-x6cvz b35143b60d5fc8a734edb4d6457e28fd467490cf90b8bbfb225a635b4543ad1b network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"c6:3c:df:a3:7d:80\",\"name\":\"b35143b60d5fc8a\"},{\"mac\":\"0a:58:0a:82:00:29\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/17ee83bf-5a5d-43cf-abda-31f2983c2cab\"}],\"ips\":[{\"address\":\"10.130.0.41/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:59:22.105824 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=397 I0419 18:59:22.473641 3645 obj_retry.go:498] Detected object tsf/patch-tekton-config-wl884 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:22.474167 3645 obj_retry.go:498] Detected object tsf/patch-tekton-config-wl884 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:23.379810 3645 obj_retry.go:492] Detected object tsf/patch-tekton-config-wl884 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:23.379835 3645 obj_retry.go:492] Detected object tsf/patch-tekton-config-wl884 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed W0419 18:59:23.460887 3645 util.go:870] Failed to build global endpoints for port TCP/http-metrics: empty IP address endpoints for service openshift-pipelines/pipelines-as-code-watcher I0419 18:59:23.498603 3645 base_network_controller_pods.go:487] [default/tsf/tsf-tekton-configuration-c29zz] creating logical port tsf_tsf-tekton-configuration-c29zz for pod on switch ip-10-0-2-180.ec2.internal I0419 18:59:23.498732 3645 kube.go:256] Updating pod tsf/tsf-tekton-configuration-c29zz I0419 18:59:23.514013 3645 pod.go:62] [tsf/tsf-tekton-configuration-c29zz] pod update took 15.306557ms I0419 18:59:23.514038 3645 base_network_controller_pods.go:951] [default/tsf/tsf-tekton-configuration-c29zz] addLogicalPort annotation time took 15.332759ms I0419 18:59:23.515241 3645 pods.go:271] [tsf/tsf-tekton-configuration-c29zz] addLogicalPort took 16.653514ms, libovsdb time 803.382µs I0419 18:59:23.927107 3645 cni.go:452] [tsf/tsf-tekton-configuration-c29zz c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2 network default NAD default NAD key ] ADD starting CNI request [tsf/tsf-tekton-configuration-c29zz c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2 network default NAD default NAD key ] I0419 18:59:23.937356 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf, podName: tsf-tekton-configuration-c29zz, hostIfaceName: c5e98407d5f126e, network: default, NAD default, SandboxID: "c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2", PCI device ID: , UID: "fdb680d4-b075-4a09-96f5-13e1ddb8e310", MAC: 0a:58:0a:82:00:2a, IPs: [10.130.0.42/23] I0419 18:59:23.995442 3645 cni.go:473] [tsf/tsf-tekton-configuration-c29zz c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2 network default NAD default NAD key default] ADD finished CNI request [tsf/tsf-tekton-configuration-c29zz c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"72:4b:6a:8a:cd:32\",\"name\":\"c5e98407d5f126e\"},{\"mac\":\"0a:58:0a:82:00:2a\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/905ac4e4-66a2-4d72-85fc-77de088dac66\"}],\"ips\":[{\"address\":\"10.130.0.42/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:59:25.062241 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/ctlog I0419 18:59:25.091898 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=8 W0419 18:59:27.075676 3645 util.go:870] Failed to build global endpoints for port TCP/resp: empty IP address endpoints for service tsf-tas/rekor-redis W0419 18:59:27.097756 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/rekor-search-ui I0419 18:59:27.202193 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=5 I0419 18:59:29.072371 3645 cni.go:452] [tsf/tsf-tekton-configuration-c29zz c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2 network default NAD default NAD key ] DEL starting CNI request [tsf/tsf-tekton-configuration-c29zz c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2 network default NAD default NAD key ] I0419 18:59:29.154476 3645 cni.go:473] [tsf/tsf-tekton-configuration-c29zz c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2 network default NAD default NAD key default] DEL finished CNI request [tsf/tsf-tekton-configuration-c29zz c5e98407d5f126e0e1da860905ec34b643d41df229d42be603c6a59d1f1a2ed2 network default NAD default NAD key default], result "{}", err I0419 18:59:29.202803 3645 obj_retry.go:498] Detected object tsf/tsf-tekton-configuration-c29zz of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:29.202840 3645 pods.go:176] Deleting pod: tsf/tsf-tekton-configuration-c29zz I0419 18:59:29.204104 3645 pods.go:236] Attempting to release IPs for pod: tsf/tsf-tekton-configuration-c29zz, ips: 10.130.0.42 I0419 18:59:29.204131 3645 obj_retry.go:498] Detected object tsf/tsf-tekton-configuration-c29zz of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:30.229762 3645 obj_retry.go:492] Detected object tsf/tsf-tekton-configuration-c29zz of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:30.229790 3645 obj_retry.go:492] Detected object tsf/tsf-tekton-configuration-c29zz of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:30.267909 3645 obj_retry.go:492] Detected object tsf/patch-tekton-config-wl884 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:30.267928 3645 obj_retry.go:492] Detected object tsf/patch-tekton-config-wl884 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:30.271597 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf/patch-tekton-config-wl884 I0419 18:59:30.271620 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf/patch-tekton-config-wl884 W0419 18:59:33.442030 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/rekor-server I0419 18:59:35.651909 3645 obj_retry.go:498] Detected object tsf-tpa/pre-install-tsf-tpa of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:35.652426 3645 obj_retry.go:498] Detected object tsf-tpa/pre-install-tsf-tpa of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:35.672970 3645 obj_retry.go:492] Detected object tsf-tpa/pre-install-tsf-tpa of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:35.672987 3645 obj_retry.go:492] Detected object tsf-tpa/pre-install-tsf-tpa of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:35.681641 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/pre-install-tsf-tpa I0419 18:59:35.681666 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/pre-install-tsf-tpa W0419 18:59:36.055425 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 6962: empty IP address endpoints for service tsf-tas/ctlog I0419 18:59:37.102508 3645 base_network_controller_pods.go:487] [default/tsf-tpa/create-db-9pgx5] creating logical port tsf-tpa_create-db-9pgx5 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:59:37.102595 3645 kube.go:256] Updating pod tsf-tpa/create-db-9pgx5 I0419 18:59:37.116595 3645 pod.go:62] [tsf-tpa/create-db-9pgx5] pod update took 14.021521ms I0419 18:59:37.116623 3645 base_network_controller_pods.go:951] [default/tsf-tpa/create-db-9pgx5] addLogicalPort annotation time took 14.050678ms I0419 18:59:37.118093 3645 pods.go:271] [tsf-tpa/create-db-9pgx5] addLogicalPort took 15.60513ms, libovsdb time 1.001087ms I0419 18:59:37.463525 3645 cni.go:452] [tsf-tpa/create-db-9pgx5 b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4 network default NAD default NAD key ] ADD starting CNI request [tsf-tpa/create-db-9pgx5 b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4 network default NAD default NAD key ] I0419 18:59:37.473044 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tpa, podName: create-db-9pgx5, hostIfaceName: b442203349928cf, network: default, NAD default, SandboxID: "b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4", PCI device ID: , UID: "743378c6-05bd-42bf-8cb5-066faba288b2", MAC: 0a:58:0a:82:00:2b, IPs: [10.130.0.43/23] I0419 18:59:37.529009 3645 cni.go:473] [tsf-tpa/create-db-9pgx5 b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4 network default NAD default NAD key default] ADD finished CNI request [tsf-tpa/create-db-9pgx5 b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"be:67:63:92:70:c4\",\"name\":\"b442203349928cf\"},{\"mac\":\"0a:58:0a:82:00:2b\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/03560ba0-a091-4366-9163-00556797ca30\"}],\"ips\":[{\"address\":\"10.130.0.43/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:59:42.145626 3645 cni.go:452] [tsf-tpa/create-db-9pgx5 b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4 network default NAD default NAD key ] DEL starting CNI request [tsf-tpa/create-db-9pgx5 b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4 network default NAD default NAD key ] I0419 18:59:42.180520 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=7 I0419 18:59:42.229020 3645 cni.go:473] [tsf-tpa/create-db-9pgx5 b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4 network default NAD default NAD key default] DEL finished CNI request [tsf-tpa/create-db-9pgx5 b442203349928cfdfecda2d68ca4dd7739e3b03a7483771df1d8569cbab3f8e4 network default NAD default NAD key default], result "{}", err I0419 18:59:42.279682 3645 obj_retry.go:498] Detected object tsf-tpa/create-db-9pgx5 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:42.279718 3645 pods.go:176] Deleting pod: tsf-tpa/create-db-9pgx5 I0419 18:59:42.281043 3645 pods.go:236] Attempting to release IPs for pod: tsf-tpa/create-db-9pgx5, ips: 10.130.0.43 I0419 18:59:42.281067 3645 obj_retry.go:498] Detected object tsf-tpa/create-db-9pgx5 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:43.306316 3645 obj_retry.go:492] Detected object tsf-tpa/create-db-9pgx5 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:43.306341 3645 obj_retry.go:492] Detected object tsf-tpa/create-db-9pgx5 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:43.421193 3645 base_network_controller_pods.go:487] [default/tsf-tpa/migrate-db-5c7zr] creating logical port tsf-tpa_migrate-db-5c7zr for pod on switch ip-10-0-2-180.ec2.internal I0419 18:59:43.421286 3645 kube.go:256] Updating pod tsf-tpa/migrate-db-5c7zr I0419 18:59:43.433656 3645 pod.go:62] [tsf-tpa/migrate-db-5c7zr] pod update took 12.38986ms I0419 18:59:43.433688 3645 base_network_controller_pods.go:951] [default/tsf-tpa/migrate-db-5c7zr] addLogicalPort annotation time took 12.427444ms I0419 18:59:43.434911 3645 pods.go:271] [tsf-tpa/migrate-db-5c7zr] addLogicalPort took 13.733827ms, libovsdb time 809.864µs I0419 18:59:43.788188 3645 cni.go:452] [tsf-tpa/migrate-db-5c7zr c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6 network default NAD default NAD key ] ADD starting CNI request [tsf-tpa/migrate-db-5c7zr c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6 network default NAD default NAD key ] I0419 18:59:43.797843 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tpa, podName: migrate-db-5c7zr, hostIfaceName: c12fed5309e7e3a, network: default, NAD default, SandboxID: "c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6", PCI device ID: , UID: "21bf2db7-a516-4f11-ba10-b8dd3e1c45e8", MAC: 0a:58:0a:82:00:2c, IPs: [10.130.0.44/23] I0419 18:59:43.854296 3645 cni.go:473] [tsf-tpa/migrate-db-5c7zr c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6 network default NAD default NAD key default] ADD finished CNI request [tsf-tpa/migrate-db-5c7zr c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"b2:48:9f:87:a6:a5\",\"name\":\"c12fed5309e7e3a\"},{\"mac\":\"0a:58:0a:82:00:2c\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/d1998870-6aac-4fdf-8739-5a22e9215191\"}],\"ips\":[{\"address\":\"10.130.0.44/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:59:46.173160 3645 cni.go:452] [tsf-tpa/migrate-db-5c7zr c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6 network default NAD default NAD key ] DEL starting CNI request [tsf-tpa/migrate-db-5c7zr c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6 network default NAD default NAD key ] I0419 18:59:46.265317 3645 cni.go:473] [tsf-tpa/migrate-db-5c7zr c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6 network default NAD default NAD key default] DEL finished CNI request [tsf-tpa/migrate-db-5c7zr c12fed5309e7e3a17dd496b2e23a2b19203225cc3649025cc13f6046ba8bccf6 network default NAD default NAD key default], result "{}", err I0419 18:59:46.314934 3645 obj_retry.go:498] Detected object tsf-tpa/migrate-db-5c7zr of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:46.314974 3645 pods.go:176] Deleting pod: tsf-tpa/migrate-db-5c7zr I0419 18:59:46.316302 3645 pods.go:236] Attempting to release IPs for pod: tsf-tpa/migrate-db-5c7zr, ips: 10.130.0.44 I0419 18:59:46.316330 3645 obj_retry.go:498] Detected object tsf-tpa/migrate-db-5c7zr of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:46.451190 3645 obj_retry.go:492] Detected object tsf-tpa/migrate-db-5c7zr of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:46.451214 3645 obj_retry.go:492] Detected object tsf-tpa/migrate-db-5c7zr of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:49.590821 3645 base_network_controller_pods.go:487] [default/tsf-tpa/server-7847885b69-sg4h2] creating logical port tsf-tpa_server-7847885b69-sg4h2 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:59:49.590957 3645 kube.go:256] Updating pod tsf-tpa/server-7847885b69-sg4h2 I0419 18:59:49.607129 3645 pod.go:62] [tsf-tpa/server-7847885b69-sg4h2] pod update took 16.198595ms I0419 18:59:49.607153 3645 base_network_controller_pods.go:951] [default/tsf-tpa/server-7847885b69-sg4h2] addLogicalPort annotation time took 16.229009ms I0419 18:59:49.608532 3645 pods.go:271] [tsf-tpa/server-7847885b69-sg4h2] addLogicalPort took 17.731969ms, libovsdb time 967.417µs I0419 18:59:51.031738 3645 base_network_controller_pods.go:487] [default/tsf-tpa/importer-5b5b6dd8fb-tbgl4] creating logical port tsf-tpa_importer-5b5b6dd8fb-tbgl4 for pod on switch ip-10-0-2-180.ec2.internal I0419 18:59:51.031844 3645 kube.go:256] Updating pod tsf-tpa/importer-5b5b6dd8fb-tbgl4 I0419 18:59:51.046684 3645 pod.go:62] [tsf-tpa/importer-5b5b6dd8fb-tbgl4] pod update took 14.860959ms I0419 18:59:51.046714 3645 base_network_controller_pods.go:951] [default/tsf-tpa/importer-5b5b6dd8fb-tbgl4] addLogicalPort annotation time took 14.8981ms I0419 18:59:51.048073 3645 pods.go:271] [tsf-tpa/importer-5b5b6dd8fb-tbgl4] addLogicalPort took 16.356673ms, libovsdb time 913.803µs I0419 18:59:51.670596 3645 obj_retry.go:498] Detected object tsf-tpa/create-importers-rvng5 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:51.671249 3645 obj_retry.go:498] Detected object tsf-tpa/create-importers-rvng5 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:52.697394 3645 obj_retry.go:492] Detected object tsf-tpa/create-importers-rvng5 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:52.697438 3645 obj_retry.go:492] Detected object tsf-tpa/create-importers-rvng5 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:54.751877 3645 cni.go:452] [tsf-tpa/server-7847885b69-sg4h2 801194cbf04312292ec742150354014a1f9797c044952fabbc01a2a5d9d018e0 network default NAD default NAD key ] ADD starting CNI request [tsf-tpa/server-7847885b69-sg4h2 801194cbf04312292ec742150354014a1f9797c044952fabbc01a2a5d9d018e0 network default NAD default NAD key ] I0419 18:59:54.761473 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tpa, podName: server-7847885b69-sg4h2, hostIfaceName: 801194cbf043122, network: default, NAD default, SandboxID: "801194cbf04312292ec742150354014a1f9797c044952fabbc01a2a5d9d018e0", PCI device ID: , UID: "634bdefd-4829-4996-b59a-fe34e04990d9", MAC: 0a:58:0a:82:00:2d, IPs: [10.130.0.45/23] I0419 18:59:54.816283 3645 cni.go:473] [tsf-tpa/server-7847885b69-sg4h2 801194cbf04312292ec742150354014a1f9797c044952fabbc01a2a5d9d018e0 network default NAD default NAD key default] ADD finished CNI request [tsf-tpa/server-7847885b69-sg4h2 801194cbf04312292ec742150354014a1f9797c044952fabbc01a2a5d9d018e0 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"be:cf:c4:90:92:f8\",\"name\":\"801194cbf043122\"},{\"mac\":\"0a:58:0a:82:00:2d\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/dbca6232-a302-4c21-a39b-0375e1e2c20a\"}],\"ips\":[{\"address\":\"10.130.0.45/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err W0419 18:59:55.225566 3645 util.go:870] Failed to build global endpoints for port TCP/endpoint: empty IP address endpoints for service tsf-tpa/server I0419 18:59:56.097256 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=96 I0419 18:59:56.195271 3645 cni.go:452] [tsf-tpa/importer-5b5b6dd8fb-tbgl4 c704bd24f0e85b47d625f8e04e853489cf074ef4ec76d40aa2125f6b55b67649 network default NAD default NAD key ] ADD starting CNI request [tsf-tpa/importer-5b5b6dd8fb-tbgl4 c704bd24f0e85b47d625f8e04e853489cf074ef4ec76d40aa2125f6b55b67649 network default NAD default NAD key ] I0419 18:59:56.205394 3645 helper_linux.go:556] ConfigureOVS: namespace: tsf-tpa, podName: importer-5b5b6dd8fb-tbgl4, hostIfaceName: c704bd24f0e85b4, network: default, NAD default, SandboxID: "c704bd24f0e85b47d625f8e04e853489cf074ef4ec76d40aa2125f6b55b67649", PCI device ID: , UID: "46763a00-7192-4caf-b3ff-85070cea8570", MAC: 0a:58:0a:82:00:2e, IPs: [10.130.0.46/23] W0419 18:59:56.241637 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8080: empty IP address endpoints for service tsf-tpa/server I0419 18:59:56.260485 3645 cni.go:473] [tsf-tpa/importer-5b5b6dd8fb-tbgl4 c704bd24f0e85b47d625f8e04e853489cf074ef4ec76d40aa2125f6b55b67649 network default NAD default NAD key default] ADD finished CNI request [tsf-tpa/importer-5b5b6dd8fb-tbgl4 c704bd24f0e85b47d625f8e04e853489cf074ef4ec76d40aa2125f6b55b67649 network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"42:57:a5:08:c5:d7\",\"name\":\"c704bd24f0e85b4\"},{\"mac\":\"0a:58:0a:82:00:2e\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/90ae3b6e-c558-48e7-8797-b37abbb79413\"}],\"ips\":[{\"address\":\"10.130.0.46/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 18:59:58.383993 3645 obj_retry.go:498] Detected object tsf-tas/tuf-repository-init-jt9fv-64ggn of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:58.384644 3645 obj_retry.go:498] Detected object tsf-tas/tuf-repository-init-jt9fv-64ggn of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:59.274465 3645 obj_retry.go:492] Detected object tsf-tas/tuf-repository-init-jt9fv-64ggn of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:59.274497 3645 obj_retry.go:492] Detected object tsf-tas/tuf-repository-init-jt9fv-64ggn of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:59.699726 3645 obj_retry.go:498] Detected object tsf-tpa/test-tsf-tpa of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:59.700271 3645 obj_retry.go:498] Detected object tsf-tpa/test-tsf-tpa of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 18:59:59.718932 3645 obj_retry.go:492] Detected object tsf-tpa/test-tsf-tpa of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:59.718952 3645 obj_retry.go:492] Detected object tsf-tpa/test-tsf-tpa of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 18:59:59.726753 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/test-tsf-tpa I0419 18:59:59.726784 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/test-tsf-tpa I0419 19:00:00.440822 3645 namespace.go:98] [image-controller] adding namespace I0419 19:00:00.442154 3645 namespace.go:102] [image-controller] adding namespace took 1.306941ms I0419 19:00:00.443271 3645 namespace.go:98] [default-tenant] adding namespace I0419 19:00:00.443469 3645 namespace.go:98] [integration-service] adding namespace I0419 19:00:00.443504 3645 namespace.go:98] [build-service] adding namespace I0419 19:00:00.444129 3645 namespace.go:102] [default-tenant] adding namespace took 841.556µs I0419 19:00:00.444825 3645 namespace.go:102] [integration-service] adding namespace took 1.340921ms I0419 19:00:00.445457 3645 namespace.go:102] [build-service] adding namespace took 1.942432ms I0419 19:00:00.485137 3645 namespace.go:142] [image-controller] updating namespace I0419 19:00:00.505074 3645 namespace.go:142] [image-controller] updating namespace I0419 19:00:00.520483 3645 namespace.go:142] [default-tenant] updating namespace I0419 19:00:00.548743 3645 namespace.go:142] [default-tenant] updating namespace I0419 19:00:00.594183 3645 namespace.go:142] [integration-service] updating namespace I0419 19:00:00.611471 3645 namespace.go:142] [integration-service] updating namespace I0419 19:00:00.619723 3645 namespace.go:142] [build-service] updating namespace I0419 19:00:00.672163 3645 namespace.go:142] [build-service] updating namespace I0419 19:00:00.748141 3645 namespace.go:142] [build-service] updating namespace I0419 19:00:00.779238 3645 namespace.go:142] [integration-service] updating namespace I0419 19:00:00.814035 3645 namespace.go:98] [release-service] adding namespace I0419 19:00:00.814960 3645 namespace.go:102] [release-service] adding namespace took 901.471µs I0419 19:00:00.852192 3645 namespace.go:142] [release-service] updating namespace I0419 19:00:00.852186 3645 namespace.go:142] [konflux-ui] updating namespace I0419 19:00:00.877035 3645 namespace.go:142] [release-service] updating namespace I0419 19:00:00.959111 3645 namespace.go:98] [konflux-info] adding namespace I0419 19:00:00.959937 3645 namespace.go:102] [konflux-info] adding namespace took 807.164µs I0419 19:00:01.040139 3645 namespace.go:98] [namespace-lister] adding namespace I0419 19:00:01.040965 3645 namespace.go:102] [namespace-lister] adding namespace took 800.637µs I0419 19:00:01.081072 3645 namespace.go:142] [konflux-info] updating namespace I0419 19:00:01.140750 3645 namespace.go:142] [konflux-info] updating namespace I0419 19:00:01.178260 3645 namespace.go:98] [enterprise-contract-service] adding namespace I0419 19:00:01.179209 3645 namespace.go:102] [enterprise-contract-service] adding namespace took 930.155µs I0419 19:00:01.183138 3645 namespace.go:142] [namespace-lister] updating namespace I0419 19:00:01.229136 3645 namespace.go:142] [image-controller] updating namespace I0419 19:00:01.258991 3645 namespace.go:142] [namespace-lister] updating namespace I0419 19:00:01.301111 3645 namespace.go:142] [enterprise-contract-service] updating namespace I0419 19:00:01.385165 3645 namespace.go:142] [enterprise-contract-service] updating namespace I0419 19:00:01.469628 3645 namespace.go:142] [default-tenant] updating namespace I0419 19:00:02.004733 3645 namespace.go:98] [segment-bridge] adding namespace I0419 19:00:02.005533 3645 namespace.go:102] [segment-bridge] adding namespace took 782.38µs I0419 19:00:02.371108 3645 namespace.go:142] [segment-bridge] updating namespace I0419 19:00:02.450870 3645 base_network_controller_policy.go:1179] Adding network policy namespace-lister/namespace-lister-allow-from-konfluxui for network default I0419 19:00:02.451529 3645 pod_selector_address_set.go:213] Created shared address set for pod selector namespace-lister_LS{ML:{app: proxy,},} I0419 19:00:02.451550 3645 base_network_controller_policy.go:1054] Policy namespace-lister/namespace-lister-allow-from-konfluxui added to peer address sets [namespace-lister_LS{ML:{app: proxy,},}] I0419 19:00:02.454205 3645 base_network_controller_policy.go:1232] Create network policy namespace-lister/namespace-lister-allow-from-konfluxui resources completed, update namespace loglevel I0419 19:00:02.510782 3645 namespace.go:142] [segment-bridge] updating namespace I0419 19:00:02.514127 3645 base_network_controller_policy.go:1179] Adding network policy namespace-lister/namespace-lister-allow-to-apiserver for network default I0419 19:00:02.514150 3645 base_network_controller_policy.go:1054] Policy namespace-lister/namespace-lister-allow-to-apiserver added to peer address sets [] I0419 19:00:02.515044 3645 base_network_controller_policy.go:1232] Create network policy namespace-lister/namespace-lister-allow-to-apiserver resources completed, update namespace loglevel I0419 19:00:03.590136 3645 namespace.go:142] [build-service] updating namespace I0419 19:00:04.114441 3645 namespace.go:142] [default-tenant] updating namespace I0419 19:00:04.507860 3645 namespace.go:142] [enterprise-contract-service] updating namespace I0419 19:00:04.807018 3645 namespace.go:142] [image-controller] updating namespace I0419 19:00:05.023223 3645 namespace.go:142] [integration-service] updating namespace I0419 19:00:05.332617 3645 namespace.go:142] [konflux-info] updating namespace I0419 19:00:05.532721 3645 namespace.go:142] [namespace-lister] updating namespace I0419 19:00:05.741212 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610420-fwt8k of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:00:05.741856 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610420-fwt8k of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:00:05.762866 3645 namespace.go:142] [release-service] updating namespace I0419 19:00:05.904492 3645 namespace.go:142] [build-service] updating namespace I0419 19:00:05.938684 3645 namespace.go:142] [default-tenant] updating namespace I0419 19:00:05.973285 3645 namespace.go:142] [enterprise-contract-service] updating namespace I0419 19:00:06.011834 3645 namespace.go:142] [image-controller] updating namespace I0419 19:00:06.036114 3645 namespace.go:142] [integration-service] updating namespace I0419 19:00:06.071708 3645 namespace.go:142] [konflux-info] updating namespace I0419 19:00:06.100378 3645 namespace.go:142] [namespace-lister] updating namespace I0419 19:00:06.122730 3645 namespace.go:142] [release-service] updating namespace I0419 19:00:06.263916 3645 namespace.go:142] [build-service] updating namespace W0419 19:00:06.307753 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service namespace-lister/namespace-lister I0419 19:00:06.423254 3645 namespace.go:142] [default-tenant] updating namespace I0419 19:00:06.599373 3645 namespace.go:142] [enterprise-contract-service] updating namespace I0419 19:00:06.755317 3645 namespace.go:142] [image-controller] updating namespace I0419 19:00:06.812641 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610420-fwt8k of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:00:06.812673 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610420-fwt8k of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:00:06.913208 3645 namespace.go:142] [integration-service] updating namespace I0419 19:00:07.063130 3645 namespace.go:142] [konflux-info] updating namespace I0419 19:00:07.171417 3645 namespace.go:142] [namespace-lister] updating namespace I0419 19:00:07.218750 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=9 I0419 19:00:07.271816 3645 namespace.go:142] [release-service] updating namespace W0419 19:00:08.315294 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service tsf-tas/tuf I0419 19:00:08.542005 3645 namespace.go:142] [segment-bridge] updating namespace I0419 19:00:08.673643 3645 namespace.go:142] [segment-bridge] updating namespace I0419 19:00:08.770583 3645 namespace.go:142] [segment-bridge] updating namespace W0419 19:00:09.619727 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service image-controller/image-controller-controller-manager-metrics-service W0419 19:00:09.650051 3645 util.go:870] Failed to build global endpoints for port TCP/https: empty IP address endpoints for service build-service/build-service-controller-manager-metrics-service I0419 19:00:10.212878 3645 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=10 W0419 19:00:13.344562 3645 util.go:870] Failed to build global endpoints for port TCP/: empty IP address endpoints for service integration-service/integration-service-webhook-service W0419 19:00:13.345298 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service integration-service/integration-service-controller-manager-metrics-service W0419 19:00:15.364598 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service release-service/release-service-controller-manager-metrics-service W0419 19:00:15.364605 3645 util.go:870] Failed to build global endpoints for port TCP/: empty IP address endpoints for service release-service/release-service-webhook-service W0419 19:00:15.660048 3645 util.go:870] Failed to build global endpoints for port TCP/dex: empty IP address endpoints for service konflux-ui/dex W0419 19:00:15.687639 3645 util.go:870] Failed to build global endpoints for port TCP/: empty IP address endpoints for service integration-service/integration-service-webhook-service W0419 19:00:15.688312 3645 util.go:870] Failed to build global endpoints for port TCP/http: empty IP address endpoints for service integration-service/integration-service-controller-manager-metrics-service W0419 19:00:19.409958 3645 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service konflux-ui/proxy W0419 19:00:19.409977 3645 util.go:870] Failed to build global endpoints for port TCP/web-tls: empty IP address endpoints for service konflux-ui/proxy W0419 19:00:19.699371 3645 util.go:870] Failed to build global endpoints for port TCP/web: empty IP address endpoints for service konflux-ui/proxy W0419 19:00:19.699389 3645 util.go:870] Failed to build global endpoints for port TCP/web-tls: empty IP address endpoints for service konflux-ui/proxy I0419 19:00:22.189838 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=10 I0419 19:00:25.367544 3645 obj_retry.go:498] Detected object integration-service/integration-service-controller-manager-7b666bf6d8-l4pjp of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:00:25.368120 3645 obj_retry.go:498] Detected object integration-service/integration-service-controller-manager-7b666bf6d8-l4pjp of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:00:25.742596 3645 obj_retry.go:492] Detected object integration-service/integration-service-controller-manager-7b666bf6d8-l4pjp of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:00:25.742620 3645 obj_retry.go:492] Detected object integration-service/integration-service-controller-manager-7b666bf6d8-l4pjp of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:00:25.749533 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod integration-service/integration-service-controller-manager-7b666bf6d8-l4pjp I0419 19:00:25.749549 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod integration-service/integration-service-controller-manager-7b666bf6d8-l4pjp I0419 19:01:02.206591 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=9 I0419 19:01:15.184784 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=11 I0419 19:01:31.377907 3645 obj_retry.go:498] Detected object konflux-ui/proxy-7684b9bb65-84vwh of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:01:31.378498 3645 obj_retry.go:498] Detected object konflux-ui/proxy-7684b9bb65-84vwh of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:01:31.811391 3645 obj_retry.go:492] Detected object konflux-ui/proxy-7684b9bb65-84vwh of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:01:31.811446 3645 obj_retry.go:492] Detected object konflux-ui/proxy-7684b9bb65-84vwh of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:01:31.832667 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod konflux-ui/proxy-7684b9bb65-84vwh I0419 19:01:31.832697 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod konflux-ui/proxy-7684b9bb65-84vwh I0419 19:01:37.092828 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=78 I0419 19:01:47.270898 3645 obj_retry.go:498] Detected object konflux-ui/test-tsf-konflux of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:01:47.271468 3645 obj_retry.go:498] Detected object konflux-ui/test-tsf-konflux of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:01:47.318255 3645 obj_retry.go:492] Detected object konflux-ui/test-tsf-konflux of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:01:47.318277 3645 obj_retry.go:492] Detected object konflux-ui/test-tsf-konflux of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:01:47.322146 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod konflux-ui/test-tsf-konflux I0419 19:01:47.322169 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod konflux-ui/test-tsf-konflux I0419 19:02:15.693733 3645 namespace.go:98] [default-managed-tenant-ea5h9] adding namespace I0419 19:02:15.694642 3645 namespace.go:102] [default-managed-tenant-ea5h9] adding namespace took 882.543µs I0419 19:02:15.724890 3645 namespace.go:142] [default-managed-tenant-ea5h9] updating namespace I0419 19:02:15.741230 3645 namespace.go:142] [default-managed-tenant-ea5h9] updating namespace I0419 19:02:16.706291 3645 namespace.go:142] [default-managed-tenant-ea5h9] updating namespace I0419 19:02:16.838032 3645 namespace.go:142] [default-managed-tenant-ea5h9] updating namespace I0419 19:02:16.939115 3645 namespace.go:142] [default-managed-tenant-ea5h9] updating namespace I0419 19:02:52.091052 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=76 I0419 19:04:04.222927 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=6 I0419 19:04:25.109561 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=229 I0419 19:04:38.200812 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=6 I0419 19:05:04.101007 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=72 I0419 19:05:30.094780 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=111 I0419 19:05:36.221916 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=6 I0419 19:06:55.184458 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=9 I0419 19:07:20.210146 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=7 I0419 19:07:24.182850 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=11 I0419 19:07:50.095205 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=11 I0419 19:07:51.205937 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=10 I0419 19:08:32.215769 3645 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=9 I0419 19:08:52.187330 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=9 I0419 19:09:43.060294 3645 obj_retry.go:492] Detected object tsf-tpa/create-db-9pgx5 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:09:43.060337 3645 obj_retry.go:492] Detected object tsf-tpa/create-db-9pgx5 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:09:43.065629 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/create-db-9pgx5 I0419 19:09:43.065650 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/create-db-9pgx5 I0419 19:09:46.044600 3645 obj_retry.go:492] Detected object tsf-tpa/migrate-db-5c7zr of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:09:46.044646 3645 obj_retry.go:492] Detected object tsf-tpa/migrate-db-5c7zr of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:09:46.050523 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/migrate-db-5c7zr I0419 19:09:46.050555 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/migrate-db-5c7zr I0419 19:09:52.034978 3645 obj_retry.go:492] Detected object tsf-tpa/create-importers-rvng5 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:09:52.035029 3645 obj_retry.go:492] Detected object tsf-tpa/create-importers-rvng5 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:09:52.040765 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod tsf-tpa/create-importers-rvng5 I0419 19:09:52.040795 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod tsf-tpa/create-importers-rvng5 I0419 19:09:58.203733 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=7 I0419 19:10:03.720532 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610430-gnmzk of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:10:03.721166 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610430-gnmzk of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:10:04.585315 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610430-gnmzk of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:10:04.585362 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610430-gnmzk of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:10:14.193682 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=11 I0419 19:10:25.096053 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=10 I0419 19:10:51.093750 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=37 I0419 19:11:13.112298 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=22 I0419 19:11:20.225721 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=8 I0419 19:11:34.103683 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=7 I0419 19:12:13.225118 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=7 I0419 19:12:52.213510 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=6 I0419 19:12:52.401235 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-pull-request-zzsrd-init-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:12:52.401858 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-pull-request-zzsrd-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:12:52.714054 3645 obj_retry.go:492] Detected object default-tenant/tsf-demo-comp-on-pull-request-zzsrd-init-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:12:52.714102 3645 obj_retry.go:492] Detected object default-tenant/tsf-demo-comp-on-pull-request-zzsrd-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:12:52.721528 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod default-tenant/tsf-demo-comp-on-pull-request-zzsrd-init-pod I0419 19:12:52.721554 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod default-tenant/tsf-demo-comp-on-pull-request-zzsrd-init-pod I0419 19:13:04.007275 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-init-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:13:04.007902 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:13:19.772282 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-clone-repository-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:13:19.772848 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-clone-repository-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:13:32.188279 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=8 I0419 19:14:04.138616 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-prefetch-dependencies-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:14:04.139217 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-prefetch-dependencies-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:14:26.218855 3645 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=6 I0419 19:14:42.097924 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.EndpointSlice" totalItems=13 I0419 19:15:34.098327 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.NetworkPolicy" totalItems=8 I0419 19:16:35.416852 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-build-container-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:16:35.417467 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-build-container-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:16:43.185990 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.AdminNetworkPolicy" totalItems=10 I0419 19:17:05.084195 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-build-image-index-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:17:05.084769 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-build-image-index-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:17:27.918691 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-pull-request-fp54g-init-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:17:27.919266 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-pull-request-fp54g-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:17:38.209547 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressService" totalItems=10 I0419 19:17:56.228609 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.UserDefinedNetwork" totalItems=6 I0419 19:17:59.214065 3645 base_network_controller_pods.go:487] [default/openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c] creating logical port openshift-pipelines_tekton-pipelines-webhook-584c448755-phh2c for pod on switch ip-10-0-2-180.ec2.internal I0419 19:17:59.214205 3645 kube.go:256] Updating pod openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c I0419 19:17:59.230772 3645 pod.go:62] [openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c] pod update took 16.598828ms I0419 19:17:59.230798 3645 base_network_controller_pods.go:951] [default/openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c] addLogicalPort annotation time took 16.630379ms I0419 19:17:59.232090 3645 pods.go:271] [openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c] addLogicalPort took 18.04056ms, libovsdb time 906.938µs I0419 19:17:59.584057 3645 cni.go:452] [openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c 52b7c2dbf33e0a178b3bcbbfa987224223ca83b1180e0e74b19f17738d99848f network default NAD default NAD key ] ADD starting CNI request [openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c 52b7c2dbf33e0a178b3bcbbfa987224223ca83b1180e0e74b19f17738d99848f network default NAD default NAD key ] I0419 19:17:59.593658 3645 helper_linux.go:556] ConfigureOVS: namespace: openshift-pipelines, podName: tekton-pipelines-webhook-584c448755-phh2c, hostIfaceName: 52b7c2dbf33e0a1, network: default, NAD default, SandboxID: "52b7c2dbf33e0a178b3bcbbfa987224223ca83b1180e0e74b19f17738d99848f", PCI device ID: , UID: "79bdacdd-081a-47e6-a8dc-60e46b1089bd", MAC: 0a:58:0a:82:00:2f, IPs: [10.130.0.47/23] I0419 19:17:59.652532 3645 cni.go:473] [openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c 52b7c2dbf33e0a178b3bcbbfa987224223ca83b1180e0e74b19f17738d99848f network default NAD default NAD key default] ADD finished CNI request [openshift-pipelines/tekton-pipelines-webhook-584c448755-phh2c 52b7c2dbf33e0a178b3bcbbfa987224223ca83b1180e0e74b19f17738d99848f network default NAD default NAD key default], result "{\"interfaces\":[{\"mac\":\"ba:40:ac:9f:c0:12\",\"name\":\"52b7c2dbf33e0a1\"},{\"mac\":\"0a:58:0a:82:00:2f\",\"name\":\"eth0\",\"sandbox\":\"/var/run/netns/c0ebab08-824f-492a-87eb-58d6efe631f8\"}],\"ips\":[{\"address\":\"10.130.0.47/23\",\"gateway\":\"10.130.0.1\",\"interface\":1}]}", err I0419 19:18:03.331116 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-pull-request-fp54g-clone-repository-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:03.331706 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-pull-request-fp54g-clone-repository-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:03.979584 3645 obj_retry.go:492] Detected object default-tenant/tsf-demo-comp-on-pull-request-fp54g-clone-repository-pod of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:18:03.979614 3645 obj_retry.go:492] Detected object default-tenant/tsf-demo-comp-on-pull-request-fp54g-clone-repository-pod of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:18:03.989960 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod default-tenant/tsf-demo-comp-on-pull-request-fp54g-clone-repository-pod I0419 19:18:03.989985 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod default-tenant/tsf-demo-comp-on-pull-request-fp54g-clone-repository-pod I0419 19:18:05.202681 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-init-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:05.203726 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-init-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:09.242719 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-sast-unicode-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:09.243267 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-sast-unicode-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:10.196979 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressFirewall" totalItems=9 I0419 19:18:11.190659 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressIP" totalItems=11 W0419 19:18:11.933346 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8080: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:11.933368 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:11.933374 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8008: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:11.933379 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9090: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:11.933395 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:11.933419 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8008: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:11.933424 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9090: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:11.933429 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8080: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook I0419 19:18:12.270304 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-deprecated-base-image-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:12.270890 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-deprecated-base-image-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:13.326882 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-sast-shell-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:13.327518 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-sast-shell-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:13.365990 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-clone-repository-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:13.366586 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-clone-repository-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it W0419 19:18:14.048062 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8008: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:14.048082 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 9090: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:14.048088 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8080: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook W0419 19:18:14.048093 3645 util.go:959] Failed to build node endpoints for node ip-10-0-2-180.ec2.internal port 8443: empty IP address endpoints for service openshift-pipelines/tekton-pipelines-webhook I0419 19:18:14.713868 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-tpa-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:14.714428 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-tpa-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:28.731059 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-rpms-signature-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:18:28.731645 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-rpms-signature-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:19:03.183705 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-prefetch-dependencies-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:19:03.184297 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-prefetch-dependencies-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:19:09.097005 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" totalItems=55 I0419 19:19:39.115679 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Pod" totalItems=256 I0419 19:19:51.207978 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140" type="*v1.EgressQoS" totalItems=10 I0419 19:20:03.216452 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140" type="*v1.AdminPolicyBasedExternalRoute" totalItems=9 I0419 19:20:04.307546 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610440-wkqt4 of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:20:04.308141 3645 obj_retry.go:498] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610440-wkqt4 of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:20:05.175367 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610440-wkqt4 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:20:05.175391 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610440-wkqt4 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:20:05.213959 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 of type *v1.Pod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:20:05.213979 3645 obj_retry.go:492] Detected object openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 of type *factory.egressIPPod in terminal state (e.g. completed) will be ignored as it has already been processed I0419 19:20:05.221242 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *factory.egressIPPod openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 I0419 19:20:05.221271 3645 obj_retry.go:771] Ignoring delete event for resource in terminal state *v1.Pod openshift-backplane/osd-delete-backplane-serviceaccounts-29610410-8gpt2 I0419 19:20:09.099021 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" totalItems=10 I0419 19:20:29.192028 3645 reflector.go:979] "Watch close" reflector="sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141" type="*v1alpha1.BaselineAdminNetworkPolicy" totalItems=7 I0419 19:20:43.227971 3645 reflector.go:979] "Watch close" reflector="github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140" type="*v1.ClusterUserDefinedNetwork" totalItems=9 I0419 19:20:54.107524 3645 reflector.go:979] "Watch close" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Namespace" totalItems=10 I0419 19:21:04.790117 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-build-container-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:04.790752 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-build-container-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:09.814867 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-build-image-index-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:09.815455 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-build-image-index-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:16.948614 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-clair-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:16.949225 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-clair-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:17.946189 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-sast-unicode-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:17.946726 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-sast-unicode-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:22.999474 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-sast-shell-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:23.000052 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-sast-shell-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:24.023709 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-deprecated-base-image-check-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:24.024255 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-deprecated-base-image-check-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:30.924876 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-rpms-signature-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:30.925431 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-rpms-signature-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:38.006236 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-clamav-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:38.006836 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-s6tm2-clamav-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:21:40.222237 3645 reflector.go:979] "Watch close" reflector="github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117" type="*v1.NetworkAttachmentDefinition" totalItems=8 I0419 19:22:00.180548 3645 obj_retry.go:498] Detected object default-tenant/my-integration-test-mkpp-kbpdc-collect-keyless-params-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:22:00.181133 3645 obj_retry.go:498] Detected object default-tenant/my-integration-test-mkpp-kbpdc-collect-keyless-params-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:22:05.411550 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-tpa-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:22:05.412146 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-tpa-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:22:13.338266 3645 obj_retry.go:498] Detected object default-tenant/my-integration-test-mkpp-kbpdc-verify-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:22:13.338875 3645 obj_retry.go:498] Detected object default-tenant/my-integration-test-mkpp-kbpdc-verify-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it I0419 19:22:32.662510 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-clair-scan-pod of type *v1.Pod in terminal state (e.g. completed) during update event: will remove it I0419 19:22:32.663101 3645 obj_retry.go:498] Detected object default-tenant/tsf-demo-comp-on-push-xztm2-clair-scan-pod of type *factory.egressIPPod in terminal state (e.g. completed) during update event: will remove it