XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    XCP-ng 8.3 updates announcements and testing

    Scheduled Pinned Locked Moved News
    362 Posts 40 Posters 134.9k Views 55 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • stormiS Offline
      stormi Vates 🪐 XCP-ng Team @marcoi
      last edited by

      @marcoi said in XCP-ng 8.3 updates announcements and testing:

      also noticed a new issue- seems like changes i had in the /etc/xensource/usb-policy.conf file for usb was lost during the upgrade.

      I have some usb comm devices i use with a home assistant VM and they were gone post the upgrade.

      anyway to make those change last post upgrade? Maybe make then options in gui so a config file can always be reflective of gui settings?

      This one is known. I had opened an issue about it, but it didn't get much traction yet. We also have a related item in our backlog, but it's a matter of finding resources to handle it.

      https://github.com/xapi-project/xen-api/issues/4935

      stormi created this issue in xapi-project/xen-api

      open Packaging: xapi-core RPM updates overwrite custom user config in usb-policy.conf #4935

      olivierlambertO 1 Reply Last reply Reply Quote 1
      • S Offline
        shorian @stormi
        last edited by

        @stormi https://xcp-ng.org/forum/post/101015 V kind, thanks

        1 Reply Last reply Reply Quote 0
        • olivierlambertO Offline
          olivierlambert Vates 🪐 Co-Founder CEO @stormi
          last edited by

          @stormi Yes, I'm also affected in here, due to an USB device that's disabled by default.

          G 1 Reply Last reply Reply Quote 0
          • G Offline
            gb.123 @olivierlambert
            last edited by gb.123

            @olivierlambert

            I did post an alternative script here:
            https://xcp-ng.org/forum/topic/8620/usb-passthrough-override-script-to-ensure-usb-policy-conf-consistency

            But it was later removed due to request by @stormi .

            I'm currently using my script which basically backs up your settings and overrides the default one (after backing it up on first install) on every boot. I know its a crude way to handle but it was only meant to be temporary till you guys find a solution.

            I have reposted it.

            Please note : Issues caused by this script (if any) shall not be covered by the XCP-NG Support Team 
            
            stormiS 1 Reply Last reply Reply Quote 0
            • stormiS Offline
              stormi Vates 🪐 XCP-ng Team @gb.123
              last edited by

              @gb.123 Please mention in your post that any issue caused by this script will not be covered by official support. The concerns I voiced then still hold.

              G 1 Reply Last reply Reply Quote 0
              • G Offline
                gb.123 @stormi
                last edited by gb.123

                @stormi

                Great Idea!
                Post updated ! 🙂

                Update: I also added 'Automatic Backup' which backs up your original file in case something goes wrong.

                1 Reply Last reply Reply Quote 1
                • gduperreyG Offline
                  gduperrey Vates 🪐 XCP-ng Team
                  last edited by gduperrey

                  New security and maintenance update candidates for you to test!

                  Security vulnerabilities have been detected and fixed for xen and varstored. We also publish other non-urgent updates which we had in the pipe for the next update release.

                  Security updates:

                  • xen:

                    • XSA-477 / VSA-2026-001: A buffer overflow in the Xen shadow tracing code could allow a DomU virtual machine to crash Xen, or potentially escalate privileges.
                    • XSA-479 / VSA-2026-003: Some Xen optimizations to avoid clearing internal CPU buffers when not required could allow one guest to leak data of another guest. A mitigation can be applied without the fix by rebooting vulnerable Xen with "spec-ctrl=ibpb-entry=hvm,ibpb-entry=pv" on the Xen command line at the cost of decreased performances.
                  • varstored:

                    • XSA-478 / VSA-2026-002: Within varstored, there were insufficient compiler barriers, creating TOCTOU issues with data in the shared buffer. An attacker with kernel level access in a VM can escalate privilege via gaining code execution within varstored.

                  Maintenance updates:

                  • guest-templates-json:

                    • Update VM template labels
                    • Sync RHEL10 template with XenServer's
                  • intel-microcode:

                    • Update to publicly released microcode-20251111
                    • Updates for multiple functional issues
                  • kernel: Bug fixes in the NFS and NBD stacks for various deadlocks and other race conditions.

                  • qemu: Backport for CVE-2021-3929, fixing a DMA reentrancy flaw in NVMe emulation, that could lead to use-after-free from a malicious guest and potential arbitrary code execution.

                  • smartmontools: Update to minor release 7.5

                  • swtpm: Synchronize with release 0.7.3-12 from XenServer. No functional changes.

                  • xapi: Fix regression on dynamic memory management during live migration, causing VMs not to balloon down before the migration.

                  • xcp-ng-release: Prevent remote syslog from being overwritten by system updates.

                  XOSTOR
                  In addition to the changes in common packages, the following XOSTOR-specific packages received updates:

                  • drbd: Reduces the I/O load and time during resync.
                  • drbd-reactor: Misc improvements regarding drbd-reactor and events
                  • linstor:
                    • Resource delete: Fixed rare race condition where a delayed DRBD event causes "resource not found" ErrorReports
                    • Misc changes to robustify LINSTOR API calls and checks

                  If you are using Xostor, please refer to our documentation for the update method.

                  Test on XCP-ng 8.3

                  yum clean metadata --enablerepo=xcp-ng-testing,xcp-ng-candidates
                  yum update --enablerepo=xcp-ng-testing,xcp-ng-candidates
                  reboot
                  

                  The usual update rules apply: pool coordinator first, etc.

                  Versions:

                  • guest-templates-json: 2.0.15-1.1.xcpng8.3
                  • intel-microcode: 20251029-1.xcpng8.3
                  • kernel: 4.19.19-8.0.44.1.xcpng8.3
                  • qemu: 4.2.1-5.2.15.2.xcpng8.3
                  • smartmontools: 7.5-1.xcpng8.3
                  • swtpm: 0.7.3-12.xcpng8.3
                  • xapi: 25.33.1-2.3.xcpng8.3
                  • xcp-ng-release: 8.3.0-36
                  • xcp-python-libs: 3.0.10-1.1.xcpng8.3
                  • xen: 4.17.5-23.2.xcpng8.3
                  • varstored: 1.2.0-3.5.xcpng8.3

                  XOSTOR

                  • drbd: 9.33.0-1.el7_9
                  • drbd-reactor: 1.9.0-1
                  • kmod-drbd: 9.2.16-1.0.xcpng8.3
                  • linstor: 1.33.0~rc.2-1.el8
                  • linstor-client: 1.27.0-1.xcpng8.3
                  • python-linstor: 1.27.0-1.xcpng8.3
                  • xcp-ng-linstor: 1.2-4.xcpng8.3

                  What to test

                  Normal use and anything else you want to test.

                  Test window before official release of the updates

                  2 days max.

                  F A P 3 Replies Last reply Reply Quote 3
                  • F Offline
                    flakpyro @gduperrey
                    last edited by

                    Installed on my usual selection of hosts. (A mixture of AMD and Intel hosts, SuperMicro, Asus, and Minisforum). No issues after a reboot, PCI Passthru, backups, etc continue to work smoothly

                    1 Reply Last reply Reply Quote 4
                    • A Online
                      Andrew Top contributor @gduperrey
                      last edited by

                      @gduperrey Standard XCP 8.3 pools updated and running.

                      1 Reply Last reply Reply Quote 2
                      • gduperreyG Offline
                        gduperrey Vates 🪐 XCP-ng Team
                        last edited by

                        Thank you everyone for your tests and your feedback!

                        The updates are live now: https://xcp-ng.org/blog/2026/01/29/january-2026-security-and-maintenance-updates-for-xcp-ng-8-3-lts/

                        M 1 Reply Last reply Reply Quote 2
                        • D dcskinner referenced this topic
                        • M Offline
                          manilx @gduperrey
                          last edited by

                          @gduperrey updated 2 hosts @home. 5 @office.
                          Had to run yum clean metadata ; yum update on cli (cancelling to run RPU in XO) for updates to appear.

                          1 Reply Last reply Reply Quote 2
                          • P Offline
                            Pilow @gduperrey
                            last edited by Pilow

                            @gduperrey we had the XOA update alert, upgraded to XOA 6.1.0

                            but no sign of XCP hosts updates ?
                            da38c4a3-539e-483d-ac5b-93d60360d287-image.png

                            When patches are available, it usually pops up on its own, is there something to do on cli now ?

                            EDIT : my bad, we had a DNS resolution problem... I now see a bunch of updates available...

                            R 1 Reply Last reply Reply Quote 2
                            • R Offline
                              robertblissitt @Pilow
                              last edited by

                              Yesterday, from memory: Up-to-date XO (CE) said there were Pool updates available, but the three individual XCP-ng hosts showed nothing available. I went to https://xcp-ng.org/blog/tag/security/ and did not see any new patches published for January, and I feared that the previous updates from October had somehow not been fully installed. I put hosts into Maintenance Mode and rebooted them, and patches were seemingly installed as part of the reboot. I don't recall if I rebooted (and therefore patched) the Master first or not as you are supposed to do. This was a bit unsettling.

                              As of this morning, Central Time US, for our three-node XS 8.4 Pool also managed by the same XO (CE), I see patches available in XO both at the Pool level and at the Host level as expected. (Yesterday, I did not see any patches reported by XO for our XS 8.4 Pool.)

                              DanpD 1 Reply Last reply Reply Quote 0
                              • DanpD Online
                                Danp Pro Support Team @robertblissitt
                                last edited by

                                @robertblissitt You can check /var/log/yum.log on the XCP-ng hosts to see when the updates were actually applied, but there isn't anything in a standard installation of XO / XCP-ng that would trigger an "automated" update of missing patches.

                                R 1 Reply Last reply Reply Quote 2
                                • R Offline
                                  robertblissitt @Danp
                                  last edited by

                                  @Danp Thank you, and I could easily be misremembering how the patches got installed - I may have clicked a button (at the Host level?) to install them even though I could not see any to install. The other events I mention, however, I am more certain of.

                                  1 Reply Last reply Reply Quote 0
                                  • marcoiM Offline
                                    marcoi
                                    last edited by

                                    applied latest patches to my two host pool without issue.

                                    P 1 Reply Last reply Reply Quote 1
                                    • P Offline
                                      Pilow @marcoi
                                      last edited by

                                      currently having heavy issues with a production cluster of 3 hosts
                                      RPU launched, all VMs except one did evacuate the Master. we managed to shutdown this VM/restart it on another host
                                      we had 0b1cff9f-21d6-4912-99aa-f2223fc0f665-image.png
                                      Master patch & reboot proceeded
                                      Then RPU tried to evacuate a slave host and all VM are now locked we can't shutdown/hard shutdown them,
                                      we have a critical VM on this host that is still running, we tried to snapshot it in case of need of hard reboot of the host, but OPERATION NOT SUPPORTED DURING AN UPGRADE
                                      we manually install patches on the host without reboot and then snapshot proceeded
                                      I hope this VM is secured by this snapshot...

                                      ticket is open with pro support but quite stalled for now... no news since yesterday Ticket#7751752

                                      1 Reply Last reply Reply Quote 0
                                      • olivierlambertO Offline
                                        olivierlambert Vates 🪐 Co-Founder CEO
                                        last edited by

                                        That's a weird one 🤔 Ping @Team-Hypervisor-Kernel

                                        P 1 Reply Last reply Reply Quote 0
                                        • P Offline
                                          Pilow @olivierlambert
                                          last edited by Pilow

                                          @olivierlambert shoutout to @danp that did a takeover of the incident ticket

                                          he headed me the right way to resolution of the problem, my production pool is back up & running with its VMs.

                                          there was indeed a diff between what was seen by "xl list"/"xenops-cli list" and what was seen by XOA in the web ui.
                                          a couple "xl destroy pid" to destroy zombie VMs, and toolstack restarts later, all is now up.

                                          I don't know how the hell a simple RPU did get me in this situation though...

                                          R 1 Reply Last reply Reply Quote 0
                                          • olivierlambertO Offline
                                            olivierlambert Vates 🪐 Co-Founder CEO
                                            last edited by

                                            Oh wow. Indeed, that's strange. And big kudos to @danp then!!

                                            1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post