XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    huge number of api call "sr.getAllUnhealthyVdiChainsLength" in tasks

    Scheduled Pinned Locked Moved Xen Orchestra
    28 Posts 8 Posters 4.2k Views 8 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • F Offline
      flakpyro @manilx
      last edited by

      @manilx the .cbt logs are located on your SR (/run/sr-mount/UUID) i don't totally know if removing them is what fixed things for me but once backups started failing my goal was to "revert" everything back to a known good working state. Doing this, reverting XOA and restarting the tool stack on the host got things back to how they were.

      I'm very much looking forward to CBT backups though as i've been spoiled by Veeam / ESX CBT backups for a lot of years!

      1 Reply Last reply Reply Quote 0
      • ryan.gravelR Offline
        ryan.gravel @olivierlambert
        last edited by

        @olivierlambert I'm not sure that I am doing it correctly. XOA is up to date (Master, commit 6ee7a) and I went through and re scanned the storage repositories to get snapshots to coalesce. Backups work properly now.

        I have 50 or so instances of the following in XO tasks

        • API call: sr.getAllUnhealthyVdiChainsLength
        • API call: sr.getVdiChainsInfo
        • API call: sr.reclaimSpace
        • API call: proxy.getAll
        • API call: session.getUser
        • API call: pool.listMissingPatches

        I've rebooted the host and ran 'xe-toolstack-restart' but I'm getting the same results. I must be missing something.

        Thanks for your help.

        1 Reply Last reply Reply Quote 0
        • olivierlambertO Offline
          olivierlambert Vates 🪐 Co-Founder CEO
          last edited by

          Hi,

          XO tasks aren't related to XAPI tasks, it's normal to see those. They were hidden before, now it's helpful to see when we have so much calls 🙂

          1 Reply Last reply Reply Quote 1
          • olivierlambertO Offline
            olivierlambert Vates 🪐 Co-Founder CEO
            last edited by

            FYI, CBT was reverted, we'll continue to fix bugs on it in a dedicated branch for people to test (with a community and Vates joint test campaign).

            M 1 Reply Last reply Reply Quote 2
            • M Offline
              manilx @olivierlambert
              last edited by

              @olivierlambert I applaud this decision. Running stable for many many months suddenly to have all these issues, without really knowing what caused it and how to resolve has been a nightmare for me 😞

              I'm all in for those new features (!) but if updating to new build to get fixes results in those issues is not good.

              I've spent many hours lately, rebooted the hosts more times than the last 5 years to get backups/coalesce working again and still am not sure if everything is back to normal....

              So, yes, this is a good decision.

              1 Reply Last reply Reply Quote 1
              • olivierlambertO Offline
                olivierlambert Vates 🪐 Co-Founder CEO
                last edited by

                This thread will be used to centralize all backup issues related to CBT: https://xcp-ng.org/forum/topic/9268/cbt-the-thread-to-centralize-your-feedback

                F 1 Reply Last reply Reply Quote 1
                • F Offline
                  flakpyro @olivierlambert
                  last edited by

                  @olivierlambert

                  After i seen the CBT stuff was reverted on github, i updated to the latest commit on my home server, (253aa) i can report my backups are now working as they should and coalesce runs without issues leaving a clean health check dashboard again. :). Glad to see this has been held back on XOA as well as i was planning to stay on 5.95.1 otherwise! Looking forward to CBT eventually all the same!

                  M 1 Reply Last reply Reply Quote 0
                  • M Offline
                    manilx @flakpyro
                    last edited by

                    @flakpyro Just updated, run my delta backups and all is fine 🎂

                    M 1 Reply Last reply Reply Quote 0
                    • M Offline
                      manilx @manilx
                      last edited by

                      @manilx Was fine for a short while only: https://xcp-ng.org/forum/topic/9275/starting-getting-again-retry-the-vm-backup-due-to-an-error-error-vdi-must-be-free-or-attached-to-exactly-one-vm

                      1 Reply Last reply Reply Quote 0
                      • rvreugdeR Offline
                        rvreugde
                        last edited by rvreugde

                        Same problem here 300+ sessions and counting
                        07564582-c58a-490e-8921-5f5de0e98fa6-image.png
                        XAPI restart did not solve the issue...

                        1 Reply Last reply Reply Quote 1
                        • ryan.gravelR Offline
                          ryan.gravel
                          last edited by

                          Doesn't seem to be respecting [NOBAK] anymore for storage repos. Tried '[NOSNAP][NOBAK] StorageName' and it still grabs it.

                          1 Reply Last reply Reply Quote 0
                          • olivierlambertO Offline
                            olivierlambert Vates 🪐 Co-Founder CEO
                            last edited by

                            On the VDI name, right?

                            ryan.gravelR 1 Reply Last reply Reply Quote 0
                            • ryan.gravelR Offline
                              ryan.gravel @olivierlambert
                              last edited by ryan.gravel

                              @olivierlambert I believe so. It is the Disks tab of a VM. [NOBAK] was working but now the NAS is complaining about storage space. This VM has a disk on the NAS so it is backing it up twice. Added [NOSNAP] just as a test with the same result.
                              24516021-d80c-4418-a539-a89b7ca17797-image.png

                              It is running a full backup if that makes any difference.

                              julien-fJ 1 Reply Last reply Reply Quote 0
                              • olivierlambertO Offline
                                olivierlambert Vates 🪐 Co-Founder CEO
                                last edited by

                                Ping @julien-f

                                1 Reply Last reply Reply Quote 0
                                • julien-fJ Offline
                                  julien-f Vates 🪐 Co-Founder XO Team @ryan.gravel
                                  last edited by

                                  @ryan-gravel I cannot reproduce, both Full Backup and Incremental Backup correctly ignore VDIs with [NOBAK] in their name label.

                                  A 1 Reply Last reply Reply Quote 0
                                  • A Offline
                                    Andrew Top contributor @julien-f
                                    last edited by

                                    @julien-f It's happening to me too... my [NOBAK] disk is being backed up now (commit 0e4a3) using the normal Backup (running a full).

                                    ryan.gravelR A 2 Replies Last reply Reply Quote 1
                                    • ryan.gravelR Offline
                                      ryan.gravel @Andrew
                                      last edited by

                                      @Andrew @julien-f Ran a full backup with commit 96b76 and it is working properly now. Started thinking that I was the only one 🙂 Thanks a lot for everyone's help!

                                      1 Reply Last reply Reply Quote 0
                                      • A Offline
                                        Andrew Top contributor @Andrew
                                        last edited by

                                        With today's update (feat: release 5.96.0, master 96b7673) things are working well again. I have not seen any Orphan VDI/snapshots or stuck Control Domain VDIs (but it may take time to test). Thanks to the Vates team!

                                        julien-fJ 1 Reply Last reply Reply Quote 1
                                        • julien-fJ Offline
                                          julien-f Vates 🪐 Co-Founder XO Team @Andrew
                                          last edited by

                                          Hello, can you please try the buffered-task-events branch and let me know if that solves the issue?

                                          1 Reply Last reply Reply Quote 0
                                          • First post
                                            Last post