You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We have a high (>20%) packet loss and latency (>10ms) between VMs (debian 10 and 11) running on the same KVM hypervisor host (debian 11) on a bridge that connects only virtual interfaces with no physical interface involved. The packet loss takes only place on the red connectivity. All other links that involve physical interfaces are not affected.
The physical interfaces A and B are attached to the bgp-VM via passthru. Interface C is a bridge (br0) and attached to a physical interface.
Interface "D" is only virtual (virbr1) with no physical interfaces attached.
ping from VM1 to bgp-VM:
ping 109.70.100.65
PING 109.70.100.65 (109.70.100.65): 56 data bytes
64 bytes from 109.70.100.65: icmp_seq=0 ttl=64 time=0.478 ms
64 bytes from 109.70.100.65: icmp_seq=2 ttl=64 time=19.331 ms
64 bytes from 109.70.100.65: icmp_seq=3 ttl=64 time=6.489 ms
[...]
64 bytes from 109.70.100.65: icmp_seq=8 ttl=64 time=19.009 ms
64 bytes from 109.70.100.65: icmp_seq=9 ttl=64 time=0.412 ms
64 bytes from 109.70.100.65: icmp_seq=10 ttl=64 time=2.670 ms
^C
--- 109.70.100.65 ping statistics ---
12 packets transmitted, 9 packets received, 25.0% packet loss
round-trip min/avg/max/stddev = 0.412/8.830/21.736/8.391 ms
The following graph is from the hypervisor and shows network traffic in pps and dropped packets per second as the network traffic ramps up.
Drops start at a network load of around 20k pps.
a few log entries from the KVM host (/var/log/messages), no relevant logs in /var/log/syslog:
kernel: [1025010.317588] br0: port 2(vnet9) entered blocking state
kernel: [1025010.317595] br0: port 2(vnet9) entered disabled state
kernel: [1025010.317784] device vnet9 entered promiscuous mode
kernel: [1025010.318392] br0: port 2(vnet9) entered blocking state
kernel: [1025010.318397] br0: port 2(vnet9) entered forwarding state
kernel: [1025010.532324] audit: type=1400 audit(1635007733.797:79): apparmor="STATUS" operation="profile_replace" profile="unconfined" name="libvirt-56ec6397-5928-483d-a9f1-5c61246ece9e" pid=2021336 comm="apparmor_parser"
kernel: [1025010.541096] virbr1: port 1(vnet10) entered blocking state
kernel: [1025010.541100] virbr1: port 1(vnet10) entered disabled state
kernel: [1025010.541279] device vnet10 entered promiscuous mode
kernel: [1025010.541778] virbr1: port 1(vnet10) entered blocking state
kernel: [1025010.541783] virbr1: port 1(vnet10) entered listening state
kernel: [1025010.762247] audit: type=1400 audit(1635007734.029:80): apparmor="STATUS" operation="profile_replace" info="same as current profile, skipping" profile="unconfined" name="libvirt-56ec6397-5928-483d-a9f1-5c61246ece9e" pid=2021348 comm="apparmor_parser"
kernel: [1025010.826756] audit: type=1400 audit(1635007734.093:81): apparmor="DENIED" operation="capable" profile="libvirtd" pid=973 comm="rpc-worker" capability=38 capname="perfmon"
kernel: [1025012.389140] vfio-pci 0000:06:00.0: Masking broken INTx support
kernel: [1025012.389230] vfio-pci 0000:06:00.0: vfio_ecap_init: hiding ecap 0x19@0x1d0
kernel: [1025012.517115] vfio-pci 0000:06:00.1: Masking broken INTx support
kernel: [1025012.560950] virbr1: port 1(vnet10) entered learning state
kernel: [1025014.546677] vfio-pci 0000:06:00.0: vfio_bar_restore: reset recovery - restoring BARs
kernel: [1025014.576772] virbr1: port 1(vnet10) entered forwarding state
kernel: [1025014.576784] virbr1: topology change detected, propagating
kernel: [1025014.780404] vfio-pci 0000:06:00.1: vfio_bar_restore: reset recovery - restoring BARs
kernel: [1027370.083357] vfio-pci 0000:06:00.0: vfio_bar_restore: reset recovery - restoring BARs
kernel: [1027370.116192] vfio-pci 0000:06:00.1: vfio_bar_restore: reset recovery - restoring BARs
appliedprivacy
changed the title
high packet loss between VMs on KVM host clamps
high packet loss on bridge between VMs on debian KVM hypervisor
Oct 23, 2021
We have a high (>20%) packet loss and latency (>10ms) between VMs (debian 10 and 11) running on the same KVM hypervisor host (debian 11) on a bridge that connects only virtual interfaces with no physical interface involved. The packet loss takes only place on the red connectivity. All other links that involve physical interfaces are not affected.
The physical interfaces A and B are attached to the bgp-VM via passthru. Interface C is a bridge (br0) and attached to a physical interface.
Interface "D" is only virtual (virbr1) with no physical interfaces attached.
ping from VM1 to bgp-VM:
Same issue when pinging VM1 <-> VM2.
domifstat shows it on RX (not TX):
KVM host:
The hypervisor has 32 CPU cores. Usage is at 75%.
Besides the qemu processes there are two processes on the KVM host that take up ~50% of a single CPU core named
vhost-<number>
.ping from VM2 to bgp-VM:
these metrics are also included in node_exporter output:
The following graph is from the hypervisor and shows network traffic in pps and dropped packets per second as the network traffic ramps up.
Drops start at a network load of around 20k pps.
a few log entries from the KVM host (/var/log/messages), no relevant logs in /var/log/syslog:
a few web search results on this topic
The text was updated successfully, but these errors were encountered: