XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Deleting snapshots and base copies

    Scheduled Pinned Locked Moved Xen Orchestra
    18 Posts 5 Posters 4.0k Views 3 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • S Offline
      sumansaha @Darkbeldin
      last edited by

      @Darkbeldin currently no back up job in there.

      1 Reply Last reply Reply Quote 0
      • olivierlambertO Offline
        olivierlambert Vates 🪐 Co-Founder CEO
        last edited by

        @sumansaha check the "Advanced" view of your SR, you'll see if there's VDI to coalesce and how much/which depth.

        Then, you should take a look at the SMlog to see if it's moving forward or not.

        S 1 Reply Last reply Reply Quote 0
        • S Offline
          sumansaha @olivierlambert
          last edited by

          @olivierlambert Screenshot 2022-12-20 at 11.57.32 PM.png

          S 1 Reply Last reply Reply Quote 0
          • S Offline
            sumansaha @sumansaha
            last edited by

            @sumansaha

            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] ['/usr/bin/vhd-util', 'scan', '-f', '-m', '/var/run/sr-mount/c2de6039-fa02-837a-8fcb-87e47392c7fe/.vhd']
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] pread SUCCESS
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] ['ls', '/var/run/sr-mount/c2de6039-fa02-837a-8fcb-87e47392c7fe', '-1', '--color=never']
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] pread SUCCESS
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/running
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] lock: tried lock /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/running, acquired: True (exists: True)
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] lock: released /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/running
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] Kicking GC
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27693] === SR c2de6039-fa02-837a-8fcb-87e47392c7fe: gc ===
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27712] Will finish as PID [27713]
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27693] New PID [27712]
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/running
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/gc_active
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] lock: released /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] Found 0 cache files
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: tried lock /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/gc_active, acquired: True (exists: True)
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: tried lock /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr, acquired: True (exists: True)
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] ['/usr/bin/vhd-util', 'scan', '-f', '-m', '/var/run/sr-mount/c2de6039-fa02-837a-8fcb-87e47392c7fe/
            .vhd']
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] pread SUCCESS
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] SR c2de ('nasnew') (0 VDIs in 0 VHD trees): no changes
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: released /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] No work, exiting
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] GC process exiting, no work left
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: released /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/gc_active
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] In cleanup
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] SR c2de ('nasnew') (0 VDIs in 0 VHD trees): no changes
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27737] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27737] sr_update {'sr_uuid': 'c2de6039-fa02-837a-8fcb-87e47392c7fe', 'subtask_of': 'DummyRef:|073fd04b-d097-4bd7-8dba-e76a353279bd|SR.stat', 'args': [], 'host_ref': 'OpaqueRef:e5dea4f6-197b-47dd-a725-6c57ddb6d70b', 'session_ref': 'OpaqueRef:b545cd23-51ca-4e66-8d0f-c77d43364dd0', 'device_config': {'server': '192.168.196.121', 'SRmaster': 'true', 'serverpath': '/mnt/nas1/portawebback2', 'options': 'hard'}, 'command': 'sr_update', 'sr_ref': 'OpaqueRef:a9e94842-c873-408c-b128-78e189539b6a'}

            Not sure whether coalesce is going fine through .

            S 1 Reply Last reply Reply Quote 0
            • S Offline
              sumansaha @sumansaha
              last edited by

              @sumansaha /usr/bin/vhd-util coalesce --debug -n /dev/VG_XenStorage-da208944-11c3-c286-b097-2dbf5eb37103/VHD-8297ea93-a76a-446a-a37c-471efe1b2847

              This process is running behind.

              S 1 Reply Last reply Reply Quote 0
              • S Offline
                sumansaha @sumansaha
                last edited by

                @sumansaha

                Actually no luck in removing the base files after removing the orphand disks.

                1 Reply Last reply Reply Quote 0
                • olivierlambertO Offline
                  olivierlambert Vates 🪐 Co-Founder CEO
                  last edited by

                  So you have a coalesce process. It could take some time, leave it as is and NEVER remove a base copy manually.

                  tjkreidlT S 2 Replies Last reply Reply Quote 0
                  • tjkreidlT Offline
                    tjkreidl Ambassador @olivierlambert
                    last edited by

                    Note that the coalesce process can take up to 24 hours. If there are issues, this article might be helpful: https://support.citrix.com/article/CTX201296/understanding-garbage-collection-and-coalesce-process-troubleshooting

                    1 Reply Last reply Reply Quote 0
                    • S Offline
                      sumansaha @olivierlambert
                      last edited by

                      @olivierlambert That works. XCP-NG Rocks.

                      S 1 Reply Last reply Reply Quote 1
                      • S Offline
                        sumansaha @sumansaha
                        last edited by sumansaha

                        @sumansaha In summary , I've deleted Orphan VDIs from Dashboard->health . But in deletion, I've maintained the sequence of their age.It has taken 6 hours, though that depends on the VM size.

                        1 Reply Last reply Reply Quote 1
                        • K kamil-v4 referenced this topic on
                        • First post
                          Last post