Update published: https://xcp-ng.org/blog/2024/09/27/september-2024-security-updates/
Thank you for the tests!
Update published: https://xcp-ng.org/blog/2024/09/27/september-2024-security-updates/
Thank you for the tests!
Two new XSAs were published on 30th of January.
xen-*
:yum clean metadata --enablerepo=xcp-ng-testing
yum update "xen-*" --enablerepo=xcp-ng-testing
reboot
The usual update rules apply: pool coordinator first, etc.
xen
: 4.13.5-9.38.2.xcpng8.2Normal use and anything else you want to test, if you are using PCI passthrough devices that's even better, but we also would be glad to have confirmation from others that their normal use case still works as intended.
Test window before official release of the updates
2 day because of security updates.
A new XSA was published on the 23rd of January, so we have a new security update to include it.
kernel
:yum clean metadata --enablerepo=xcp-ng-testing
yum update kernel --enablerepo=xcp-ng-testing
reboot
The usual update rules apply: pool coordinator first, etc.
kernel
: 4.19.19-7.0.23.1.xcpng8.2Normal use and anything else you want to test. The closer to your actual use of XCP-ng, the better.
Test window before official release of the updates
~2 days due to security updates.
Update published https://xcp-ng.org/blog/2024/07/18/july-2024-security-updates/
Thank you everyone for your tests!
Two new XSAs were published on 16th of July.
xen-*
:
xapi
, xsconsole
:
yum clean metadata --enablerepo=xcp-ng-testing
yum update "xen-*" "xapi-*" xsconsole --enablerepo=xcp-ng-testing
reboot
The usual update rules apply: pool coordinator first, etc.
xen
: xen-4.13.5-9.40.2.xcpng8.2xapi
: xapi-1.249.36-1.2.xcpng8.2xsconsole
: xsconsole-10.1.13-1.2.xcpng8.2Normal use and anything else you want to test.
~ 1 day because of security updates.
The update has been published, thanks for testing.
https://xcp-ng.org/blog/2024/02/02/february-2024-security-update/
The update has been published, thanks for the feedback and tests.
https://xcp-ng.org/blog/2024/01/26/january-2024-security-update/
Hello guys,
I'll be the one investigating this further, we're trying to compile a list of CPUs and their behavior. First, thank you for your reports and tests, that's already very helpful and gave us some insight already.
If some of you can help us cover more ground that would be awesome, so here is what would be an ideal for testing to get everyone on the same page:
yum install iperf
²¹: it seems some recent kernels do provide a slight boost, but in any case the performance is pretty low for such high grade CPUs.
²: iperf3 is singlethreaded, the -P
option will establish multiple connexions, but it will process all of them in a single thread, so if reaching a 100% cpu usage, it won't get much increase and won't help identifying the scaling on such a cpu. For example on a Ryzen 5 7600
processor, we do have about the same low perfomances, but using multiple thread will scale, which does not seem to be the case for EPYC Zen1 CPUs.
xentop
on host, and try to get an idea of the top values of each domain when the test is runningiperf -s
on VM1, and let it run (no -P X
this would stop after X connexion established)iperf -c <ip_VM1> -t 60
, note result for v2v 1 threadiperf -c <ip_VM1> -t 60 -P4
, note result for v2v 4 threadsiperf -c <ip_VM1> -t 60
, note result for h2v 1 threadiperf -c <ip_VM1> -t 60 -P4
, note result for h2v 4 threadsHere is an example of report template
xl info -n
especially the cpu_topology
section in a code block.³: I note the max I see while test is running in vm-client/vm-server/host order.
Mostly for information, here are a few tests I ran which did not seem to improve performances.
noibrs noibpb nopti nospectre_v2 spectre_v2_user=off spectre_v2=off nospectre_v1 l1tf=off nospec_store_bypass_disable no_stf_barrier mds=off mitigations=off
. Note this won't disable them at xen level as there are patches that enable the fixes for the related hardware with no flags to disable them.noxsave
in kernel boot parameters as there is a known issue on Zen CPU avoided boosting when a core is under heavy AVX load, still no changes.Not tested it myself but @nicols tried on the same machine giving him about 3Gbps as we all see, on VMWare, and it went to ~25Gbps single threaded and about 40Gbps with 4 threads, and with proxmox about 21.7Gbps (I assume single threaded) which are both a lot more along what I would expect this hardware to produce.
@JamesG did test windows and debian guests and got about the same results.
Althrough we do get a small boost by increasing threads (or connexions in case of iperf3), it still is far from what we can see on other setups with vmware or proxmox).
Althrough Olivier's pool with zen4 desktop cpu do scale a lot better than EPYCs when increasing the number of threads, it still is not providing us with expected results for such powerful cpus in single thread (we do not even reach vmware single thread performances with 4 threads).
Althrough @Ajmind-0 test show a difference between debian versions, results even on debian 11 are stil not on par with expected results.
Disabling AVX only provided an improvement on my home FX cpu, which are known to not have real "threads" and share computing unit between 2 threads of a core, so it does make sense. (this is not shown in the table)
It seems that memcpy in the glibc is not related to the issue, dd if=/dev/zero of=/dev/null
has decent performances on these machines (1.2-1.3GBytes/s), and it's worth keeping in mind that both kernel and xen have their own implementation, so it could play a small role in filling the ring buffer in iperf, but I feel like the libc memcpy() is not at play here.
I'll update this table with updated results, or maybe repost it in further post.
Throughputs are in Gbit/s, noted as G
for shorter table entries.
CPU usages are for (VMclient/VMserver/dom0) in percentage as shown in xentop
.
user | cpu | family | market | v2v 1T | v2v 4T | h2v 1T | h2v 4T | notes |
---|---|---|---|---|---|---|---|---|
vates | fx8320-e | piledriver | desktop | 5.64 G (120/150/220) | 7.5 G (180/230/330) | 9.5 G (0/110/160) | 13.6 G (0/300/350) | not a zen cpu, no boost |
vates | EPYC 7451 | Zen1 | server | 4.6 G (110/180/250) | 6.08 G (180/220/300) | 7.73 G (0/150/230) | 11.2 G (0/320/350) | no boost |
vates | Ryzen 5 7600 | Zen4 | desktop | 9.74 G (70/80/100) | 19.7 G (190/260/300) | 19.2G (0/110/140) | 33.9 G (0/310/350) | Olivier's pool, no boost |
nicols | EPYC 7443 | Zen3 | server | 3.38 G (?) | iperf3 | |||
nicols | EPYC 7443 | Zen3 | server | 2.78 G (?) | 4.44 G (?) | iperf2 | ||
nicols | EPYC 7502 | Zen2 | server | similar ^ | similar ^ | iperf2 | ||
JamesG | EPYC 7302p | Zen2 | server | 6.58 G (?) | iperf3 | |||
Ajmind-0 | EPYC 7313P | Zen3 | server | 7.6 G (?) | 10.3 G (?) | iperf3, debian11 | ||
Ajmind-0 | EPYC 7313P | Zen3 | server | 4.4 G (?) | 3.07G (?) | iperf3, debian12 | ||
vates | EPYC 9124 | Zen4 | server | 1.16 G (16/17/??⁴) | 1.35 G (20/25/??⁴) | N/A | N/A | !xcp-ng, Xen 4.18-rc + suse 15 |
vates | EPYC 9124 | Zen4 | server | 5.70 G (100/140/200) | 10.4 G (230/250/420) | 10.7 G (0/120/200) | 15.8 G (0/320/380) | no boost |
vates | Ryzen 9 5950x | Zen3 | desktop | 7.25 G (30/35/60) | 16.5 G (160/210/300) | 17.5 G (0/110/140) | 27.6 G (0/270/330) | no boost |
⁴: xentop on this host shows 3200% on dom0 all the time, profiling does not seem to show anything actually using CPU, but may be related to the extremely poor performance
last updated: 2023-11-29 16:46
All help is welcome! For those of you who already provided tests I integrated in the table, feel free to not rerun tests, it looks like following the exact protocol and provided more data won't make much of a difference and I don't want to waste your time!
Thanks again to all of you for your insight and your patience, it looks like this is going to be a deep rabbit hole, I'll do my best to get to the bottom of this as soon as possible.
So, after our investigations, we were able to pinpoint the issue.
It seem to happen on most RHEL derivative distributions when migrating from 8.7 to 8.8. As suggested, the bug is in the kernel.
Starting with 4.18.0-466.el8
the patch: x86/idt: Annotate alloc_intr_gate() with __init is integrated and will create the issue. It is missing x86/xen: Split HVM vector callback setup and interrupt gate allocation that should have been integrated as well.
The migration to 8.8 will move you to 4.18.0-477.*
versions that are also raising this issue, that's what you reported.
We found that the 4.18.0-488
that can be found in CentOS 8 Stream integrates the missing patch, and do indeed work when installed manually.
Your report helped us identify and reproduce the issues. That allowed us to provide a callstack to Xen devs. Then Roger Pau Monné found that it was this patch missing quickly, and we were able to find which versions of the kernel RPMs were integrating it and when the fix was integrated.
This means the issue was identified on RH side, and it is now a matter of having an updated kernel in derivative distributions like Rocky and Alma.
The question that I'm asking here is how does the Vates Team evaluate these vulnerabilities, Qualys, Greenbone, something else?
I'm not sure what you mean by evaluate vulnerability, especially the list about Qualys, Greenbone…
If you mean how do we track and process them, I cannot talk about XO side, but I can shed some light on XCP-ng side:
That's for the dom0 side, on the hypervisor side, we're part of the security list of the Xen Project, so we receive the XSAs and integrate them as fast as we can in following our release process, sometime integrating the patches ourselves, sometime going with the XenServer fixes. If we integrate them ourselves we most of the time remove our own integration and move to the one from XenServer as the people working on these fixes are mostly the ones working on the XSAs in the first place so they have a better knowledge and insigts than us.
I hope this answers this question.
Is the Vates team open to the community reporting these vulnerabilities openly or would a ticket be best?
On XCP-ng side, everything that are packages from open source would be reports of publicly disclosed CVEs, so you can openly report them. If people were to find new vulnerabilities it would depend, but should follow a classic private disclosure in the first place:
Sorry, you asked about the whole ecosystem, but I'm only able to answer from the XCP-ng side of things.
@nick.lloyd As mentionned by Olivier the documentation has a mail to contact us, which will create a ticket internally, unrelated to support, that will reach the security team directly.
The question that I'm asking here is how does the Vates Team evaluate these vulnerabilities, Qualys, Greenbone, something else?
I'm not sure what you mean by evaluate vulnerability, especially the list about Qualys, Greenbone…
If you mean how do we track and process them, I cannot talk about XO side, but I can shed some light on XCP-ng side:
That's for the dom0 side, on the hypervisor side, we're part of the security list of the Xen Project, so we receive the XSAs and integrate them as fast as we can in following our release process, sometime integrating the patches ourselves, sometime going with the XenServer fixes. If we integrate them ourselves we most of the time remove our own integration and move to the one from XenServer as the people working on these fixes are mostly the ones working on the XSAs in the first place so they have a better knowledge and insigts than us.
I hope this answers this question.
Is the Vates team open to the community reporting these vulnerabilities openly or would a ticket be best?
On XCP-ng side, everything that are packages from open source would be reports of publicly disclosed CVEs, so you can openly report them. If people were to find new vulnerabilities it would depend, but should follow a classic private disclosure in the first place:
Sorry, you asked about the whole ecosystem, but I'm only able to answer from the XCP-ng side of things.
Running tcpdump switches the interface to promiscuous to allow all traffic that reaches the NIC to be dumped. So I assume the issue you had on your switches allowed traffic to reach the host, that was forwarding it to the VMs, and wasn't dropped because tcpdump switched the VIF into promiscuous mode.
If it seems resolved, that's good, otherwise let us know if we need to investigate further on this
@carldotcliff if you are 100% positive you see traffic on the VM that should not reach them, it is worth opening a ticket as this is not an intended behavior. If you do, tell in the ticket that this was discussed in the forum with David (me), so our support team can assign it to me if they want to.
For the dropped packets, I do not see any on my home setup, which is a pretty "small" network, in our lab, we do have some on our hosts. On bigger network, that could be pretty much anything, broadcast or multicast reaching the host that the NIC is chosing to drop itself, some NIC will also drop some discovery protocol frames, it would be hard to identify unfortunately, but that would not worry me as long as it is not a high count and not impacting performances.
I think the promisc mode is due to the fact the interfaces end up in OVS bridges, without that, the traffic coming from the outside to the VMs MAC addresses would be dropped.
Once it reach the OVS bridge the interface is in, it is up to OVS to act as a switch and only forward packets to the MAC he knows on its ports so all the traffic should not be forwarded to all the VIFs.
I just tested on 8.2 and 8.3:
So to answer your question, yes it is normal the NICs are in promiscuous, but that should not lead to all traffic going to all the VMs.
@irtaza9 Xen Orchestra premium (and from sources) has an SDN Controller plugin, it allows to create private networks and relies on GRE or VXLAN to create private networks, so as long as there are IP connectivity this can do the trick.
There are 2 blog posts on the subject:
https://xen-orchestra.com/blog/xo-sdn-controller/
https://xen-orchestra.com/blog/devblog-3-extending-the-sdn-controller/
And the documentation:
https://docs.xen-orchestra.com/sdn_controller
There are 2 main issues:
Is that answering your question?
To be honnest, I'm unsure, generally the XSAs have pretty clear impact description, here it just states:
resulting in e.g. guest user processes
being able to read data they ought not have access to.
No detail here if that's only inside the guest or if it could maybe reach data outside its domain scope. So I would not be able to say, but generally it is pretty clear in XSAs when there is a risk of accessing other guests data, my assumption would be that this is only inside the guest domain.
Hello @NielsH, no, that XSA is on the guest side, the fixes will be in the kernel used by the guest, unless we missed something, there is currently nothing to be done on the host kernel side.
Thanks for letting us know, and I'm happy you have thing working nicely now.
I think to mark this as resolved you need to convert your original post as a question, and it can then be marked as resolved. I actually cannot do it myself, I think only a few people have the permission to do it for others at Vates.
Wow, that's not great indeed.
The console too small, is pretty odd, as this should be at least the standard 80x24 for any kind of VGA.
Have you tried, from the idrac to login on the third terminal (alt+F3) ? It would be good to see if you can login or if the update + failing disk really broke everything. On the 2nd terminal (alt+F2) you should have system messages too, which in the ideal case should be empty…