@stormi
It seems to be good here!

@olivierlambert when changes are made (like IP change in the VM), it is now instantly reported by XO 
I will be able to go back to my work to integrate a BGP deamon to announce locals IPs to the network (I started last year to work on the xe-deamon but never had the time to finish it, and some months after I saw you blog post about moving to rust for it).
Thank you!
PS: i'm using these 2 scripts to list all interfaces drivers version accross our servers :
$ cat get_network_drivers_info.sh
#!/bin/bash
format="| %-13.13s | %-20.20s | %-20.20s | %-10.10s | %-7.7s | %-10.10s | %-30.30s | %-s \n"
printf "${format}" "date" "hostname" "OS" "interface" "driver" "version" "firmware" "yum"
printf "${format}" "----------------------------" "----------------------------" "----------------------------" "----------------------------" "----------------------------" "----------------------------" "----------------------------" "----------------------------"
if [ $# -gt 0 ]; then
servers=($(echo ${BASH_ARGV[*]}))
else
servers=($(cat host.json | jq -r '.[] | .address' | egrep -v "^192.168.124.9$"))
fi
for line in ${servers[@]}; do
scp get_network_drivers_info.sh.tpl ${line}:/tmp/get_network_drivers_info.sh > /dev/null 2>&1;
ssh -n ${line} bash /tmp/get_network_drivers_info.sh 2> /dev/null;
if [ $? -ne 0 ]; then
echo "${line} fail" >&2
fi
done
$ cat get_network_drivers_info.sh.tpl
#!/bin/bash
format="| %-13.13s | %-20.20s | %-20.20s | %-10.10s | %-7.7s | %-10.10s | %-30.30s | %-s \n"
d=$(date '+%Y%m%d-%H%M')
name=$(hostname)
cd /sys/class/net/
for interface in $(ls -l /sys/class/net/ | awk '/\/pci/ {print $9}'); do
version=$(ethtool -i ${interface} | awk '/^version:/ {$1=""; print}')
firmware=$(ethtool -i ${interface} | awk '/^firmware-version:/ {$1=""; print}')
driver=$(ethtool -i ${interface} | awk '/^driver:/ {$1=""; print}')
YUM=$(which yum)
if [ $? -eq 0 ]; then
packages=$(yum list installed | awk '/ixgbe/ {print $1"@"$2}' | tr '\n' ',')
else
packages="NA"
fi
os_version=$(lsb_release -d | awk '{$1=""} 1' | sed 's/XenServer/XS/; s/ (xenenterprise)//; s/release //')
printf "${format}" "${d}" "${name}" "${os_version}" "${interface}" "${driver}" "${version}" "${firmware}" "${packages}"
done
PS: host.json file is generated via : xo-cli --list-objects type=host
@stormi Hello, some week after, I can confirm that the problem is solved here by using intel-ixgbe.x86_64@5.5.2-2.1.xcpng8.1 or intel-ixgbe.x86_64@5.5.2-2.1.xcpng8.2
@stormi I have installed intel-ixgbe 5.5.2-2.1.xcpng8.2 on my server s0267. Let's wait a some days to check if the memleak is solved by this patch.
@stormi

The 2 servers have been reinstalled with an up to date 8.2. They host each 2 VMs that are doing the same thing (~100Mb/s of netdata stream).
The right one has the 5.9.4-1.xcpng8.2, the left one has 5.5.2-2.xcpng8.2.
The patch seem to be OK for me.
server 266 with alt-kernel: still no problem.

server 268 with 4.19.19-6.0.10.1.xcpng8.1: the problem has begun some days ago after some stable days.

server 272 with 4.19.19-6.0.11.1.0.1.patch53disabled.xcpng8.1:
)
server 273 with 4.19.19-6.0.11.1.0.1.patch62disabled.xcpng8.1:

It seems that 4.19.19-6.0.11.1.0.1.patch62disabled.xcpng8.1 is more stable than 4.19.19-6.0.11.1.0.1.patch53disabled.xcpng8.1. But it is a but early to be sure.
@stormi For the kernel-4.19.19-6.0.10.1.xcpng8.1 test, i'm not sure it solve the problem because I get a small memory increase. We have to wait a bit more 
@stormi I have installed the two kernels
272 ~]# yum list installed kernel | grep kernel
kernel.x86_64 4.19.19-6.0.11.1.0.1.patch53disabled.xcpng8.1
273 ~]# yum list installed kernel | grep kernel
kernel.x86_64 4.19.19-6.0.11.1.0.1.patch62disabled.xcpng8.1
I have removed the modification in /etc/modprobe.d/dist.conf on server 273.
We have to wait a little bit now 
@stormi @r1
Four days later, I get:
search extra built-in weak-updates override updates: problem still present@stormi I have a server with only search extra built-in weak-updates override updates. We will see if it is better.
@r1 installed and it works :
# dmesg | grep kmem
[ 6.181218] kmemleak: Kernel memory leak detector initialized
[ 6.181223] kmemleak: Automatic memory scanning thread started
I will check the leaks tomorrow.
Hello,
We have the same problem here on multiple servers.
We also have 10G interfaces. We use ext local SR.
(::bbxl0271) (2 running) [08:43 bbxl0271 ~]# lsmod | sort -k 2 -n -r
ipv6 548864 313 nf_nat_ipv6
sunrpc 413696 18 lockd,nfsv3,nfs_acl,nfs
ixgbe 380928 0
fscache 380928 1 nfs
nfs 307200 2 nfsv3
libata 274432 2 libahci,ahci
xhci_hcd 258048 1 xhci_pci
scsi_mod 253952 15 fcoe,scsi_dh_emc,sd_mod,dm_multipath,scsi_dh_alua,scsi_transport_fc,usb_storage,libfc,bnx2fc,uas,megaraid_sas,libata,sg,scsi_dh_rdac,scsi_dh_hp_sw
aesni_intel 200704 0
megaraid_sas 167936 4
nf_conntrack 163840 6 xt_conntrack,nf_nat,nf_nat_ipv6,nf_nat_ipv4,openvswitch,nf_conncount
bnx2fc 159744 0
dm_mod 151552 5 dm_multipath
openvswitch 147456 12
libfc 147456 3 fcoe,bnx2fc,libfcoe
hid 122880 2 usbhid,hid_generic
mei 114688 1 mei_me
lockd 110592 2 nfsv3,nfs
cnic 81920 1 bnx2fc
libfcoe 77824 2 fcoe,bnx2fc
usb_storage 73728 1 uas
scsi_transport_fc 69632 3 fcoe,libfc,bnx2fc
ipmi_si 65536 0
ipmi_msghandler 61440 2 ipmi_devintf,ipmi_si
usbhid 57344 0
sd_mod 53248 5
tun 49152 0
nfsv3 49152 1
x_tables 45056 6 xt_conntrack,iptable_filter,xt_multiport,xt_tcpudp,ipt_REJECT,ip_tables
mei_me 45056 0
sg 40960 0
libahci 40960 1 ahci
ahci 40960 0
8021q 40960 0
nf_nat 36864 3 nf_nat_ipv6,nf_nat_ipv4,openvswitch
fcoe 32768 0
dm_multipath 32768 0
uas 28672 0
lpc_ich 28672 0
ip_tables 28672 2 iptable_filter
i2c_i801 28672 0
cryptd 28672 3 crypto_simd,ghash_clmulni_intel,aesni_intel
uio 20480 1 cnic
scsi_dh_alua 20480 0
nf_defrag_ipv6 20480 2 nf_conntrack,openvswitch
mrp 20480 1 8021q
ipmi_devintf 20480 0
aes_x86_64 20480 1 aesni_intel
acpi_power_meter 20480 0
xt_tcpudp 16384 9
xt_multiport 16384 1
xt_conntrack 16384 5
xhci_pci 16384 0
stp 16384 1 garp
skx_edac 16384 0
scsi_dh_rdac 16384 0
scsi_dh_hp_sw 16384 0
scsi_dh_emc 16384 0
pcbc 16384 0
nsh 16384 1 openvswitch
nfs_acl 16384 1 nfsv3
nf_reject_ipv4 16384 1 ipt_REJECT
nf_nat_ipv6 16384 1 openvswitch
nf_nat_ipv4 16384 1 openvswitch
nf_defrag_ipv4 16384 1 nf_conntrack
nf_conncount 16384 1 openvswitch
llc 16384 2 stp,garp
libcrc32c 16384 3 nf_conntrack,nf_nat,openvswitch
ipt_REJECT 16384 3
iptable_filter 16384 1
intel_rapl_perf 16384 0
intel_powerclamp 16384 0
hid_generic 16384 0
grace 16384 1 lockd
glue_helper 16384 1 aesni_intel
ghash_clmulni_intel 16384 0
garp 16384 1 8021q
crypto_simd 16384 1 aesni_intel
crct10dif_pclmul 16384 0
crc_ccitt 16384 1 ipv6
crc32_pclmul 16384 0
I'll install the kmemleak kernel on one server today.