XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    XCP-ng 8.3 updates announcements and testing

    Scheduled Pinned Locked Moved News
    296 Posts 34 Posters 91.9k Views 50 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • M Offline
      manilx @gduperrey
      last edited by

      @gduperrey Updated 2 pools/2 servers @home, updated 3 pools/5 servers @ office with RPU.
      All fine.

      F 1 Reply Last reply Reply Quote 2
      • F Offline
        flakpyro @manilx
        last edited by flakpyro

        Updates 3 pools. All fine except for the last host on the last pool ran into a yum mirror error and failed. Manually running yum update and rebooting the host worked. Sadly i then had to move VMs back around since the RPU failed and the process of migrating VMs back did not kick off as a result.

        Error was

        One of the configured repositories failed (Unknown),
         and yum doesn't have enough cached data to continue. At this point the only
         safe thing yum can do is fail. There are a few ways to work "fix" this:
        
             1. Contact the upstream for the repository and get them to fix the problem.
        
             2. Reconfigure the baseurl/etc. for the repository, to point to a working
                upstream. This is most often useful if you are using a newer
                distribution release than is supported by the repository (and the
                packages for the previous distribution release still work).
        
             3. Run the command with the repository temporarily disabled
                    yum --disablerepo=<repoid> ...
        
             4. Disable the repository permanently, so yum won't use it by default. Yum
                will then just ignore the repository until you permanently enable it
                again or use --enablerepo for temporary usage:
        
                    yum-config-manager --disable <repoid>
                or
                    subscription-manager repos --disable=<repoid>
        
             5. Configure the failing repository to be skipped, if it is unavailable.
                Note that yum will try to contact the repo. when it runs most commands,
                so will have to try and fail each time (and thus. yum will be be much
                slower). If it is a very temporary problem though, this is often a nice
                compromise:
        
                    yum-config-manager --save --setopt=<repoid>.skip_if_unavailable=true
        
        xcp-ng-base: Check uncompressed DB failed
        
        
        stormiS 1 Reply Last reply Reply Quote 1
        • A Offline
          acebmxer
          last edited by

          Updated AMD Ryzen pool at home and update two Intel Dell r660 and r640 pools at work. No issues to report back.

          1 Reply Last reply Reply Quote 2
          • marcoiM Offline
            marcoi
            last edited by

            updated my two hosts without issues.

            1 Reply Last reply Reply Quote 2
            • stormiS Offline
              stormi Vates 🪐 XCP-ng Team
              last edited by stormi

              📣 IMPORTANT NOTICE!

              After publishing the updates, we discovered a very nasty bug when using the UEFI certificates that we distribute. Long story short, they're too big, and there's only limited space (57K), and combined to a preexisting bug in varstored, this will cause the VM to stop booting after Windows or any other OS attempts to append to the DBX (revocation database).

              We pulled the varstored update, but those who updated can be affected.

              There are conditions for the issue:

              • Existing VMs are not affected, unless you propagated the new certs to them
              • New VMs are affected only if you never installed UEFI certs to the pool yourself (through XOA or secureboot-certs install), or cleared them using secureboot-certs clear in order to use our default certificates.

              If you have the affected version of varstored (rpm -q varstored yields varstored-1.2.0-3.1.xcpng8.3) :

              • on every host, downgrade it with yum downgrade varstored-1.2.0-2.3.xcpng8.3. No reboot or toolstack restart required.
              • if you have affected UEFI VMs, that is VMs that meet the conditions above but are not broken yet, don't install updates, turn them off, and fix them by deleting their DBX database: https://docs.xcp-ng.org/guides/guest-UEFI-Secure-Boot/#remove-certificates-from-a-vm. This has to be done when the VM is off. Your OS will add its own DBX afterwards.
              • If you already have broken VMs (this warning reaching you too late), revert to a snapshot or backup. Other ways to fix them will require a patched varstored currently in the making.
              M 1 Reply Last reply Reply Quote 2
              • M Offline
                manilx @stormi
                last edited by

                @stormi Downgraded all hosts @home and @office. I have not done anything to running Windows Server VM's or touch anything regarding certs. So I guess I'm all good.

                1 Reply Last reply Reply Quote 1
                • stormiS Offline
                  stormi Vates 🪐 XCP-ng Team @flakpyro
                  last edited by

                  @flakpyro Thanks for letting us know. I suppose there was a mirror that was not ready yet, or had a transient issue, and unfortunately XOA's rolling pool update feature is not very resilient to that at the moment.

                  F 1 Reply Last reply Reply Quote 1
                  • F Offline
                    flakpyro @stormi
                    last edited by flakpyro

                    @stormi

                    I reverted the package however i initially followed the directions provided by vates in the release blog post and ran "secureboot-certs clear" then on each VM with Secure boot enabled i clicked "Copy the pools default UEFI Certificates to the VM".

                    After reverting the updates and running secureboot-certs install again i went back and clicked "Copy the pools default UEFI Certificates to the VM" again thinking it would put the old certs back.

                    It sounds like this may not be enough and i need to remove the dbx record from each of these VMs. Am i correct or was that enough to fix these VMs?

                    Per the docs:

                    varstore-rm <vm-uuid> d719b2cb-3d3a-4596-a3bc-dad00e67656f dbx
                    
                    Note that the GUID may be found by using varstore-ls <vm-uuid>.
                    

                    When i run the command i see

                    Exmaple:

                    varstore-ls f9166a11-3c3f-33f1-505c-542ce8e1764d
                    8be4df61-93ca-11d2-aa0d-00e098032b8c SecureBoot
                    8be4df61-93ca-11d2-aa0d-00e098032b8c DeployedMode
                    8be4df61-93ca-11d2-aa0d-00e098032b8c AuditMode
                    8be4df61-93ca-11d2-aa0d-00e098032b8c SetupMode
                    8be4df61-93ca-11d2-aa0d-00e098032b8c SignatureSupport
                    8be4df61-93ca-11d2-aa0d-00e098032b8c PK
                    8be4df61-93ca-11d2-aa0d-00e098032b8c KEK
                    d719b2cb-3d3a-4596-a3bc-dad00e67656f db
                    d719b2cb-3d3a-4596-a3bc-dad00e67656f dbx
                    605dab50-e046-4300-abb6-3dd810dd8b23 SbatLevel
                    fab7e9e1-39dd-4f2b-8408-e20e906cb6de HDDP
                    e20939be-32d4-41be-a150-897f85d49829 MemoryOverwriteRequestControl
                    bb983ccf-151d-40e1-a07b-4a17be168292 MemoryOverwriteRequestControlLock
                    9d1947eb-09bb-4780-a3cd-bea956e0e056 PPIBuffer
                    9d1947eb-09bb-4780-a3cd-bea956e0e056 Tcg2PhysicalPresenceFlagsLock
                    eb704011-1402-11d3-8e77-00a0c969723b MTC
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Boot0000
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Timeout
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Lang
                    8be4df61-93ca-11d2-aa0d-00e098032b8c PlatformLang
                    8be4df61-93ca-11d2-aa0d-00e098032b8c ConIn
                    8be4df61-93ca-11d2-aa0d-00e098032b8c ConOut
                    8be4df61-93ca-11d2-aa0d-00e098032b8c ErrOut
                    9d1947eb-09bb-4780-a3cd-bea956e0e056 Tcg2PhysicalPresenceFlags
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Key0000
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Key0001
                    5b446ed1-e30b-4faa-871a-3654eca36080 0050569B1890
                    937fe521-95ae-4d1a-8929-48bcd90ad31a 0050569B1890
                    9fb9a8a1-2f4a-43a6-889c-d0f7b6c47ad5 ClientId
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Boot0003
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Boot0004
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Boot0005
                    4c19049f-4137-4dd3-9c10-8b97a83ffdfa MemoryTypeInformation
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Boot0006
                    8be4df61-93ca-11d2-aa0d-00e098032b8c BootOrder
                    8c136d32-039a-4016-8bb4-9e985e62786f SecretKey
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Boot0001
                    8be4df61-93ca-11d2-aa0d-00e098032b8c Boot0002
                    

                    So the command would be:

                    varstore-rm f9166a11-3c3f-33f1-505c-542ce8e1764d d719b2cb-3d3a-4596-a3bc-dad00e67656f dbx correct? 
                    
                    Does "d719b2cb-3d3a-4596-a3bc-dad00e67656f " indicate the old certs have been re-installed?
                    
                    D 1 Reply Last reply Reply Quote 0
                    • D Offline
                      dinhngtu Vates 🪐 XCP-ng Team @flakpyro
                      last edited by dinhngtu

                      @flakpyro It's in the middle. But the "d719b2cb-3d3a-4596-a3bc-dad00e67656f dbx" part is always the same across all VMs.

                      After reverting the updates and running secureboot-certs install again

                      This will create a pool-level dbx which does not have the problem seen in varstored-1.2.0-3.1. So as long as you propagated the variables to all affected VMs without any errors, there's no need to delete them manually. But you can always delete them again if you're not sure.

                      F 1 Reply Last reply Reply Quote 2
                      • F Offline
                        flakpyro @dinhngtu
                        last edited by

                        @dinhngtu Ah okay, i was wondering if "d719b2cb-3d3a-4596-a3bc-dad00e67656f" indicated i was back on the safe certs since it is the same on all VMs since reverting and clicking "Copy the pools default UEFI Certificates to the VM"

                        So i need to run

                        varstore-rm f9166a11-3c3f-33f1-505c-542ce8e1764d d719b2cb-3d3a-4596-a3bc-dad00e67656f dbx
                        

                        while powered off to be safe?

                        F 1 Reply Last reply Reply Quote 0
                        • F Offline
                          flakpyro @flakpyro
                          last edited by

                          Awesome thanks for the response. I took a snapshot and tried rebooting a VM and it booted back up without issue simply by clicking the propagate button on each affecting VM after reverting and running "secureboot-certs install"

                          1 Reply Last reply Reply Quote 0
                          • A Offline
                            acebmxer
                            last edited by

                            @stormi would this issue cause this behavior? VM stuck at startup...

                            Screenshot 2025-10-24 091455.png

                            Screenshot 2025-10-24 091510.png

                            D 1 Reply Last reply Reply Quote 0
                            • D Offline
                              dinhngtu Vates 🪐 XCP-ng Team @acebmxer
                              last edited by dinhngtu

                              @acebmxer Looks like it. You can recover by installing a preliminary test update here:

                              wget https://koji.xcp-ng.org/repos/user/8/8.3/xcpng-users.repo -O /etc/yum.repos.d/xcpng-users.repo
                              yum update --enablerepo=xcp-ng-ndinh1 varstored varstored-tools
                              secureboot-certs clear
                              

                              Then propagate SB certs to make sure that subsequent dbx updates will run normally.

                              Sorry for the inconvenience.

                              G A 2 Replies Last reply Reply Quote 0
                              • G Offline
                                Greg_E @dinhngtu
                                last edited by

                                @dinhngtu

                                I was just getting ready to go update my pool... So is it safe now? And do these certs only affect Secure Boot VMs or UEFI VMs as well?

                                D 1 Reply Last reply Reply Quote 0
                                • A Offline
                                  acebmxer @dinhngtu
                                  last edited by acebmxer

                                  @dinhngtu Thank you for that reply.

                                  Just to clarify I have 4 host. 2 host i updated to latest update once they became available to public via pool updates. The other 2 host I had to push the second beta release to get command secureboot-certs install to work. Then I pushed the rest of the update again once they became public.

                                  So far im only seeing 1 vm having the issue on the later 2 host the had the beta updates installed. There that test PC and 1 windows server (Printsrver), xoproxy, and xoce.

                                  Should i run those commands on all 4 host or just the later 2 with the troubled vm?

                                  1 Reply Last reply Reply Quote 0
                                  • D Offline
                                    dinhngtu Vates 🪐 XCP-ng Team @Greg_E
                                    last edited by

                                    @Greg_E said in XCP-ng 8.3 updates announcements and testing:

                                    @dinhngtu

                                    I was just getting ready to go update my pool... So is it safe now? And do these certs only affect Secure Boot VMs or UEFI VMs as well?

                                    This problem will affect UEFI VMs. See the announcement above for more details.

                                    In short, if you haven't updated, you should be safe; the offending update has been withdrawn and no longer available from yum update (as long as you're not using the testing branch).

                                    @acebmxer said in XCP-ng 8.3 updates announcements and testing:

                                    @dinhngtu Thank you for that reply.

                                    Just to clarify I have 4 host. 2 host i updated to latest update once they became available to public via pool updates. The other 2 host I had to push the second beta release to get command secureboot-certs install to work. Then I pushed the rest of the update again once they became public.

                                    So far im only seeing 1 vm having the issue on the later 2 host the had the beta updates installed. There that test PC and 1 windows server (Printsrver), xoproxy, and xoce.

                                    Should i run those commands on all 4 host or just the later 2 with the troubled vm?

                                    I haven't tested using different versions of varstored within the same pool.

                                    A 1 Reply Last reply Reply Quote 0
                                    • A Offline
                                      acebmxer @dinhngtu
                                      last edited by acebmxer

                                      @dinhngtu

                                      All 4 host have the same version. They were just not updated the same way at the same time was the point i was making and that so far only 1 vm seems to be affected.

                                      But i guess to play it self i should downgrade all 4 host.

                                      Edit - ran the following commands on master host of pool with the vm with issue. VM now boots up.

                                      Sorry for all the questions.

                                      wget https://koji.xcp-ng.org/repos/user/8/8.3/xcpng-users.repo -O /etc/yum.repos.d/xcpng-users.repo
                                      yum update --enablerepo=xcp-ng-ndinh1 varstored varstored-tools
                                      secureboot-certs clear
                                      

                                      Should run that above commands on all the remaining host?

                                      Or should i run this command on any of the four host?

                                      yum downgrade varstored-1.2.0-2.3.xcpng8.3
                                      
                                      D 1 Reply Last reply Reply Quote 0
                                      • D Offline
                                        dinhngtu Vates 🪐 XCP-ng Team @acebmxer
                                        last edited by dinhngtu

                                        @acebmxer You can confirm if there are other affected VMs using this command:

                                        xe vm-list --minimal | xargs -d, -Ix sh -c "echo Scanning x 1>&2; varstore-ls x" > /dev/null
                                        

                                        If there are still affected VMs, you must fix them on varstored-1.2.0-3.2 by secureboot-certs clear + propagating before downgrading to 2.3 (if you wish).

                                        Note that downgrading to 2.3 with pool certs cleared means you'll need to set up guest Secure Boot again. So I recommend staying on 3.2 if it's possible for you.

                                        F 1 Reply Last reply Reply Quote 0
                                        • F Offline
                                          flakpyro @dinhngtu
                                          last edited by

                                          @dinhngtu The output of that command seems to result in the output of "Scanningn every UUID in my pool", most of which do not use SB. Im guessing that normal and i do not have any affected VMs any longer?

                                          1 Reply Last reply Reply Quote 0
                                          • A Offline
                                            acebmxer
                                            last edited by

                                            I opened a support ticket #7746470

                                            I get the following from different host...

                                            Pool1 Master host:
                                            [09:48 xcp-ng-vyadytkn ~]# xe vm-list --minimal | xargs -d, -Ix sh -c "echo Scanning x 1>&2; varstore-ls x" > /dev/null
                                            Scanning ab9e8cde-fad8-5089-29c8-c904a810857a
                                            Scanning d2025b59-6268-6bb9-28cd-e4812395563c
                                            Scanning 4567b8da-3a3c-00a4-cf6b-ccdae302a645
                                            Scanning 291f9d05-88c8-fc37-107a-368a55c649be
                                            Scanning 88d7ba09-e65a-dd82-8d45-ec3df3e3f9d4
                                            Scanning bcc182c6-dbb7-33c4-3e8b-c7db9b246e7b
                                            Scanning 5a2c18ca-e9ce-abfc-8bd7-2794f94c6c3a
                                            Scanning f4a4669a-2baf-47f2-b16a-08f60963e922
                                            Scanning d4352870-d73c-5cda-9470-ffd7cd01e251
                                            Scanning b0c21701-7309-dd9c-fa2a-bdc84a24565e
                                            Scanning 342b470e-fac9-6579-f82d-2517b8de6112
                                            Scanning 2c21c005-9230-f3f3-cdf6-2d39407b3456
                                            Scanning 102a5ac2-ab7f-68ca-9fa1-b549d64ec4a9
                                            Scanning c9dadc90-6fb8-eee7-d3e5-e1ae1024a77f
                                            Scanning 5703adef-d804-6b15-ba2f-7b3357a711bb
                                            Scanning a8858f86-ceab-2d5d-5d46-1b083c3fa399
                                            Scanning 82b191df-a63c-4357-b618-4467f714008a
                                            
                                            Pool 2 Master host (Problem VM)
                                            
                                            [10:02 localhost ~]# xe vm-list --minimal | xargs -d, -Ix sh -c "echo Scanning x 1>&2; varstore-ls x" > /dev/null
                                            Scanning ece6db75-533b-5359-1a99-28d5dc1a6912
                                            Scanning c81f1f1b-9f95-38ec-f96e-5ce8fea073e6
                                            unserialize_variables: Tolerating oversized data 60440 > 57344
                                            Scanning 4fb6dee8-c251-93e6-0c20-b1c1c67351ba
                                            Scanning 8a570993-6522-c837-0931-6fd1812dbd96
                                            Scanning b31749b8-bc1a-0e8a-a864-50a06391592e
                                            Scanning bf1b61d5-14ca-9f0e-7040-24c177693cea
                                            Scanning af6eeff7-5640-28a9-dbdb-87f094a04665
                                            Scanning 43ba924b-8bf7-4a50-b258-2e977f6f2f0c
                                            Scanning 5a625c92-3702-2ce6-1790-39497745e83c
                                            Scanning 9f36a8c5-c988-8546-b0f9-30ac9d0a5015
                                            Scanning a7f5b1bb-7fe2-d0b5-5187-f29e5d4d15d7
                                            
                                            
                                            D 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post