XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    CBT: the thread to centralize your feedback

    Scheduled Pinned Locked Moved Backup
    439 Posts 37 Posters 387.1k Views 29 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • R Offline
      rtjdamen @olivierlambert
      last edited by

      @olivierlambert, I first want to compliment the work that has been done. As a first release, it already seems very stable. I have shared some logs with support to investigate the data_destroy issue and some minor error messages that appeared. We managed to migrate all our backup jobs to CBT over the weekend. It was challenging to coalesce all the snapshots, but it has been completed. The difference in coalesce speed is significant, which is a great improvement for XOA backups. I will monitor the backups and observe how they evolve in the coming weeks.

      Please let us know if you need any additional input or if there are any updates regarding the data_destroy issue.

      1 Reply Last reply Reply Quote 1
      • olivierlambertO Offline
        olivierlambert Vates 🪐 Co-Founder CEO
        last edited by

        Thanks for the feedback @rtjdamen

        We will do a full sweep with the feedback in this thread and continue to monitor the situation. Also, we switched our own production on it, so if there's issue we will also be on the first seat 😛

        1 Reply Last reply Reply Quote 0
        • robytR Offline
          robyt
          last edited by robyt

          A little question: in my environment i've big problem with SR migration.
          i've opened a ticket and the answer is:
          There is currently a problem with NBD backup where the VDI are not correctly disconnected from the Dom0.
          The team is investigating but for now if you want to avoid more being accumulated, it is advised to disable NBD backups.

          I've disabled NBD and for small SR work fine, for big SR i can't try (three days estimated for migration..)

          This problem is resolved?
          Or i can't use CBT?

          1 Reply Last reply Reply Quote 0
          • olivierlambertO Offline
            olivierlambert Vates 🪐 Co-Founder CEO
            last edited by

            It depends on the SR you are using. If it's XOSTOR or not. On XOSTOR, NBD is causing some challenges. But if it's a regular SR, the issue should be fixed since we are enforcing removal of VDI attached to the control domain.

            robytR 1 Reply Last reply Reply Quote 0
            • robytR Offline
              robyt @olivierlambert
              last edited by

              @olivierlambert no xostor, sr is on fiber channel san (two san with some volumes, HDD or SSD)
              may i try the reactivation of both?

              1 Reply Last reply Reply Quote 0
              • olivierlambertO Offline
                olivierlambert Vates 🪐 Co-Founder CEO
                last edited by

                Without more context, I would say "yes" 🙂 CBT will reduce the amount to coalesce. Try to keep NBD "Number of NBD connection per disk" at 1 to be on the safe side.

                1 Reply Last reply Reply Quote 0
                • R Offline
                  rtjdamen
                  last edited by rtjdamen

                  @olivierlambert @florent regarding the error

                  error
                  {"code":"VDI_IN_USE","params":["OpaqueRef:fbd3bedd-ea60-4984-afca-9b2ec1b7b885","data_destroy"],"call":{"method":"VDI.data_destroy","params":["OpaqueRef:fbd3bedd-ea60-4984-afca-9b2ec1b7b885"]}}
                  vdiRef
                  "OpaqueRef:fbd3bedd-ea60-4984-afca-9b2ec1b7b885"
                  

                  As a test i did the command on this snapshot manually over ssh

                  xe vdi-data-destroy uuid=
                  

                  This seems to purge the snapshot data correct, so i believe this issue is temporary and maybe a retry itself during the backupjob does resolve it. We see it not so ofter but enough to investigate it more deeper.

                  Inside the Knowledgebase from xenserver i found this

                  VDI_IN_USE:
                  
                  The VDI snapshot is currently in use by another operation.
                  
                  Check that the VDI snapshot is not being accessed by another client or operation. Check that the VDI is not attached to a VM.
                  
                  If the VDI snapshot is connected to a VM snapshot by a VBD, you receive this error. Before you can run VDI.data_destroy on this VDI snapshot, you must remove the VM snapshot. Use VM.destroy to remove the VM snapshot.
                  

                  I believe the vdi.destroy is not yet finished complete before the data-destroy is issued, resulting in the vdi in use error.

                  1 Reply Last reply Reply Quote 0
                  • robytR Offline
                    robyt
                    last edited by

                    Hi, why this difference?
                    63eadf76-9d0f-4503-9c0d-db213a16be06-immagine.png
                    The backup joband the VM are the same

                    1 Reply Last reply Reply Quote 0
                    • M Offline
                      manilx
                      last edited by

                      While my backups (NBD, Delta) have been running now fine and garbage collection fnishes I just noticed that I have multiple snapshots referring to the backup on all VM's:
                      ScreenShot 2024-07-10 at 10.23.11.png
                      Deleted all but the last and will see how it goes.

                      olivierlambertO 1 Reply Last reply Reply Quote 0
                      • olivierlambertO Offline
                        olivierlambert Vates 🪐 Co-Founder CEO @manilx
                        last edited by

                        @manilx Have you enabled CBT+data removal? Otherwise you'll continue to see snapshots (but yeah, the old ones should be removed at some point)

                        M 2 Replies Last reply Reply Quote 0
                        • M Offline
                          manilx @olivierlambert
                          last edited by

                          @olivierlambert No, I haven't. Will do now.

                          M 1 Reply Last reply Reply Quote 0
                          • M Offline
                            manilx @manilx
                            last edited by

                            @manilx P.S. It wasn't clear to me that I had to turn this on. I just left everything as it was after updating XO..... (Was always using NBT before):

                            1 Reply Last reply Reply Quote 0
                            • M Offline
                              manilx @olivierlambert
                              last edited by manilx

                              @olivierlambert The next backup has run and now there a no snapshots at all (even the last one I left was removed). Is this how this is supposed to be? Totally confused.

                              I get this: ScreenShot 2024-07-10 at 12.59.42.png.
                              OK. But if I want the "old" way to have the snapshots, they accumulate as we have seen...... ??

                              1 Reply Last reply Reply Quote 0
                              • olivierlambertO Offline
                                olivierlambert Vates 🪐 Co-Founder CEO
                                last edited by

                                1. That's thanks to CBT
                                2. The old way can be used again, just disable the data removal toggle. And yes, it shouldn't accumulate more than 1 snap per schedule.
                                M 1 Reply Last reply Reply Quote 0
                                • M Offline
                                  manilx @olivierlambert
                                  last edited by

                                  @olivierlambert Understood. Then it's up to @florent to fix this snapshot accumulation.
                                  Thx

                                  1 Reply Last reply Reply Quote 0
                                  • olivierlambertO Offline
                                    olivierlambert Vates 🪐 Co-Founder CEO
                                    last edited by

                                    If we can reproduce the issue, because for now I do not have the problem here. Double check you are on the latest commit (or XOA latest fully up to date)

                                    M 2 Replies Last reply Reply Quote 0
                                    • M Offline
                                      manilx @olivierlambert
                                      last edited by

                                      @olivierlambert Positive. Will disable it again and see if it happens again.

                                      1 Reply Last reply Reply Quote 1
                                      • D Offline
                                        DustinB
                                        last edited by

                                        Seeing a nice decrease in space used.

                                        a22ff9a8-4e35-49d4-bb13-d17a1037fc99-image.png

                                        I'm keeping the snapshot around in case I need it, thought I doubt that I would.

                                        1 Reply Last reply Reply Quote 0
                                        • M Offline
                                          manilx @olivierlambert
                                          last edited by

                                          @olivierlambert Disabled the data removal option again.
                                          1st backup run: snapshot created
                                          2nd one: still just the one snapshot.

                                          Seems like enabling, backing up, disabling the option cleared this up.

                                          Will monitor.

                                          M 1 Reply Last reply Reply Quote 1
                                          • R Offline
                                            rtjdamen
                                            last edited by

                                            We have this error "stream has ended with not enough data (actual: 446, expected: 512)" on multiple vms in the last few days anyone seeing this issue?

                                            robytR 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post