XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Job canceled to protect the VDI chain

    Scheduled Pinned Locked Moved Backup
    15 Posts 3 Posters 87 Views 3 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • M Offline
      McHenry @olivierlambert
      last edited by

      @olivierlambert

      I spoke too soon. The backups started working however the problem has returned.
      221c9885-e718-4a43-b948-db80820666e3-image.png

      I do see 44 items waiting to coalesce. This is new as these would coalesce faster previously without causing this issue.
      06fdf997-95bc-4670-ae48-6c6dd31bbe33-image.png

      Is there a reason the coalesce is taking longer now or is there a way I can add resources to speed up the process?

      M 1 Reply Last reply Reply Quote 0
      • M Offline
        McHenry @McHenry
        last edited by

        I have the following entry in the logs, over and over. Not sure if this is a problem:

        Oct 29 15:25:08 HST106 SMGC: [1009624] Found 1 orphaned vdis
        Oct 29 15:25:08 HST106 SM: [1009624] lock: tried lock /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/sr, acquired: True (exists: True)
        Oct 29 15:25:08 HST106 SMGC: [1009624] Found 1 VDIs for deletion:
        Oct 29 15:25:08 HST106 SMGC: [1009624]   *d4a17b38(100.000G/21.652G?)
        Oct 29 15:25:08 HST106 SMGC: [1009624] Deleting unlinked VDI *d4a17b38(100.000G/21.652G?)
        Oct 29 15:25:08 HST106 SMGC: [1009624] Checking with slave: ('OpaqueRef:16797af5-c5d1-08d5-0e26-e17149c2807b', 'nfs-on-slave', 'check', {'path': '/var/run/sr-mount/be743b1c-7803-1943-0a70-baf5fcbfeaaf/d4a17b38-5a3c-438a-b394-fcbb64784499.vhd'})
        Oct 29 15:25:08 HST106 SM: [1009624] lock: released /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/sr
        Oct 29 15:25:08 HST106 SM: [1009624] lock: released /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/running
        Oct 29 15:25:08 HST106 SMGC: [1009624] GC process exiting, no work left
        Oct 29 15:25:08 HST106 SM: [1009624] lock: released /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/gc_active
        Oct 29 15:25:08 HST106 SMGC: [1009624] In cleanup
        Oct 29 15:25:08 HST106 SMGC: [1009624] SR be74 ('Shared NAS002') (166 VDIs in 27 VHD trees): no changes
        Oct 29 15:25:08 HST106 SM: [1009624] lock: closed /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/running
        Oct 29 15:25:08 HST106 SM: [1009624] lock: closed /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/gc_active
        Oct 29 15:25:08 HST106 SM: [1009624] lock: closed /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/sr
        
        1 Reply Last reply Reply Quote 0
        • olivierlambertO Offline
          olivierlambert Vates 🪐 Co-Founder CEO
          last edited by

          How long roughly before coalesce is done? Coalesce is a storage task directly done by XCP-ng, XO is just witnessing it.

          It's normal to have disks to coalesce after snapshot removal.

          P 1 Reply Last reply Reply Quote 0
          • P Offline
            Pilow @olivierlambert
            last edited by

            @olivierlambert if its at XCP level, is there a ratio where storage access performance should be especially monitored ?

            i guess that 500 VMs desnapshoted on 1 host is managed differently than 500VMs on 50 hosts.

            but if all hosts are on same shared storage... performance constraint ?

            1 Reply Last reply Reply Quote 0
            • olivierlambertO Offline
              olivierlambert Vates 🪐 Co-Founder CEO
              last edited by

              Coalescing is indeed generating some IOPS, latency is in general the most visible impact on it. However, it's managed per storage, regardless the number of VDIs, they are coalesced one by one.

              P 1 Reply Last reply Reply Quote 1
              • P Offline
                Pilow @olivierlambert
                last edited by

                @olivierlambert sor for high number of VMs, there is a point in time where you should be aware of not doing DR too frequently to let coalesce proceed.
                50VMs in a DR job, every hour, if coalesce takes 2min by VM, it didnt finish when the next DR is starting ?

                i'm looking for the edge case

                1 Reply Last reply Reply Quote 0
                • olivierlambertO Offline
                  olivierlambert Vates 🪐 Co-Founder CEO
                  last edited by

                  This is entirely dependent of your setup, there's no universal rule. But thanks to XO skipping the job, you can adapt your configuration to reduce the number of times it happens.

                  M 1 Reply Last reply Reply Quote 0
                  • M Offline
                    McHenry @olivierlambert
                    last edited by McHenry

                    @olivierlambert

                    As per yesterday, the backups are still being "Skipped". Checking the logs I see the following message being repeated:

                    Oct 30 08:30:09 HST106 SMGC: [1866514] Found 1 orphaned vdis
                    Oct 30 08:30:09 HST106 SM: [1866514] lock: tried lock /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/sr, acquired: True (exists: True)
                    Oct 30 08:30:09 HST106 SMGC: [1866514] Found 1 VDIs for deletion:
                    Oct 30 08:30:09 HST106 SMGC: [1866514]   *d4a17b38(100.000G/21.652G?)
                    Oct 30 08:30:09 HST106 SMGC: [1866514] Deleting unlinked VDI *d4a17b38(100.000G/21.652G?)
                    Oct 30 08:30:09 HST106 SMGC: [1866514] Checking with slave: ('OpaqueRef:16797af5-c5d1-08d5-0e26-e17149c2807b', 'nfs-on-slave', 'check', {'path': '/var/run/sr-mount/be743b1c-7803-1943-0a70-baf5fcbfeaaf/d4a17b38-5a3c-438a-b394-fcbb64784499.vhd'})
                    Oct 30 08:30:09 HST106 SM: [1866514] lock: released /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/sr
                    Oct 30 08:30:09 HST106 SM: [1866514] lock: released /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/running
                    Oct 30 08:30:09 HST106 SMGC: [1866514] GC process exiting, no work left
                    Oct 30 08:30:09 HST106 SM: [1866514] lock: released /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/gc_active
                    Oct 30 08:30:09 HST106 SMGC: [1866514] In cleanup
                    Oct 30 08:30:09 HST106 SMGC: [1866514] SR be74 ('Shared NAS002') (166 VDIs in 27 VHD trees): no changes
                    Oct 30 08:30:09 HST106 SM: [1866514] lock: closed /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/running
                    Oct 30 08:30:09 HST106 SM: [1866514] lock: closed /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/gc_active
                    Oct 30 08:30:09 HST106 SM: [1866514] lock: closed /var/lock/sm/be743b1c-7803-1943-0a70-baf5fcbfeaaf/sr
                    

                    It appears the unlinked VDI is never deleted. Could this be blocking and should this be deleted manually?

                    Deleting unlinked VDI *d4a17b38(100.000G/21.652G?)
                    

                    In regards to the following line, I can identify the VM UUID however is the 2nd UUID a snapshot? (d4a17b38-5a3c-438a-b394-fcbb64784499.vhd)

                    Oct 30 08:30:09 HST106 SMGC: [1866514] Checking with slave: ('OpaqueRef:16797af5-c5d1-08d5-0e26-e17149c2807b', 'nfs-on-slave', 'check', {'path': '/var/run/sr-mount/be743b1c-7803-1943-0a70-baf5fcbfeaaf/d4a17b38-5a3c-438a-b394-fcbb64784499.vhd'})
                    
                    M 1 Reply Last reply Reply Quote 0
                    • M Offline
                      McHenry @McHenry
                      last edited by

                      Appears to be the same as:
                      https://xcp-ng.org/forum/topic/1751/smgc-stuck-with-xcp-ng-8-0?_=1761802212787

                      It appears this snapshot is locked by a slave host that is currently offline.

                      Oct 30 08:30:09 HST106 SMGC: [1866514] Checking with slave: ('OpaqueRef:16797af5-c5d1-08d5-0e26-e17149c2807b', 'nfs-on-slave', 'check'
                      

                      When using shared storage how does a snapshot become locked by a host?

                      Is the scenario where a slave host is offline how can this lock be cleared?

                      1 Reply Last reply Reply Quote 0
                      • olivierlambertO Offline
                        olivierlambert Vates 🪐 Co-Founder CEO
                        last edited by

                        Can you start this host? If it's online, XAPI can't tell if it's really off or maybe disconnected from the management but still connected to the shared storage and creating a potential data corruption.

                        1 Reply Last reply Reply Quote 0
                        • First post
                          Last post