XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Deleting snapshots and base copies

    Scheduled Pinned Locked Moved Xen Orchestra
    18 Posts 5 Posters 6.5k Views 3 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • S Offline
      sumansaha @Darkbeldin
      last edited by

      @Darkbeldin currently no back up job in there.

      1 Reply Last reply Reply Quote 0
      • olivierlambertO Offline
        olivierlambert Vates 🪐 Co-Founder CEO
        last edited by

        @sumansaha check the "Advanced" view of your SR, you'll see if there's VDI to coalesce and how much/which depth.

        Then, you should take a look at the SMlog to see if it's moving forward or not.

        S 1 Reply Last reply Reply Quote 0
        • S Offline
          sumansaha @olivierlambert
          last edited by

          @olivierlambert Screenshot 2022-12-20 at 11.57.32 PM.png

          S 1 Reply Last reply Reply Quote 0
          • S Offline
            sumansaha @sumansaha
            last edited by

            @sumansaha

            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] ['/usr/bin/vhd-util', 'scan', '-f', '-m', '/var/run/sr-mount/c2de6039-fa02-837a-8fcb-87e47392c7fe/.vhd']
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] pread SUCCESS
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] ['ls', '/var/run/sr-mount/c2de6039-fa02-837a-8fcb-87e47392c7fe', '-1', '--color=never']
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] pread SUCCESS
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/running
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] lock: tried lock /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/running, acquired: True (exists: True)
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] lock: released /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/running
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] Kicking GC
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27693] === SR c2de6039-fa02-837a-8fcb-87e47392c7fe: gc ===
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27712] Will finish as PID [27713]
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27693] New PID [27712]
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/running
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/gc_active
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27693] lock: released /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] Found 0 cache files
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: tried lock /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/gc_active, acquired: True (exists: True)
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: tried lock /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr, acquired: True (exists: True)
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] ['/usr/bin/vhd-util', 'scan', '-f', '-m', '/var/run/sr-mount/c2de6039-fa02-837a-8fcb-87e47392c7fe/
            .vhd']
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] pread SUCCESS
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] SR c2de ('nasnew') (0 VDIs in 0 VHD trees): no changes
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: released /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] No work, exiting
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] GC process exiting, no work left
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27713] lock: released /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/gc_active
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] In cleanup
            Dec 21 00:03:34 xcp-ng-slqtflub SMGC: [27713] SR c2de ('nasnew') (0 VDIs in 0 VHD trees): no changes
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27737] lock: opening lock file /var/lock/sm/c2de6039-fa02-837a-8fcb-87e47392c7fe/sr
            Dec 21 00:03:34 xcp-ng-slqtflub SM: [27737] sr_update {'sr_uuid': 'c2de6039-fa02-837a-8fcb-87e47392c7fe', 'subtask_of': 'DummyRef:|073fd04b-d097-4bd7-8dba-e76a353279bd|SR.stat', 'args': [], 'host_ref': 'OpaqueRef:e5dea4f6-197b-47dd-a725-6c57ddb6d70b', 'session_ref': 'OpaqueRef:b545cd23-51ca-4e66-8d0f-c77d43364dd0', 'device_config': {'server': '192.168.196.121', 'SRmaster': 'true', 'serverpath': '/mnt/nas1/portawebback2', 'options': 'hard'}, 'command': 'sr_update', 'sr_ref': 'OpaqueRef:a9e94842-c873-408c-b128-78e189539b6a'}

            Not sure whether coalesce is going fine through .

            S 1 Reply Last reply Reply Quote 0
            • S Offline
              sumansaha @sumansaha
              last edited by

              @sumansaha /usr/bin/vhd-util coalesce --debug -n /dev/VG_XenStorage-da208944-11c3-c286-b097-2dbf5eb37103/VHD-8297ea93-a76a-446a-a37c-471efe1b2847

              This process is running behind.

              S 1 Reply Last reply Reply Quote 0
              • S Offline
                sumansaha @sumansaha
                last edited by

                @sumansaha

                Actually no luck in removing the base files after removing the orphand disks.

                1 Reply Last reply Reply Quote 0
                • olivierlambertO Offline
                  olivierlambert Vates 🪐 Co-Founder CEO
                  last edited by

                  So you have a coalesce process. It could take some time, leave it as is and NEVER remove a base copy manually.

                  tjkreidlT S 2 Replies Last reply Reply Quote 0
                  • tjkreidlT Offline
                    tjkreidl Ambassador @olivierlambert
                    last edited by

                    Note that the coalesce process can take up to 24 hours. If there are issues, this article might be helpful: https://support.citrix.com/article/CTX201296/understanding-garbage-collection-and-coalesce-process-troubleshooting

                    1 Reply Last reply Reply Quote 0
                    • S Offline
                      sumansaha @olivierlambert
                      last edited by

                      @olivierlambert That works. XCP-NG Rocks.

                      S 1 Reply Last reply Reply Quote 1
                      • S Offline
                        sumansaha @sumansaha
                        last edited by sumansaha

                        @sumansaha In summary , I've deleted Orphan VDIs from Dashboard->health . But in deletion, I've maintained the sequence of their age.It has taken 6 hours, though that depends on the VM size.

                        1 Reply Last reply Reply Quote 1
                        • K kamil-v4 referenced this topic on

                        Hello! It looks like you're interested in this conversation, but you don't have an account yet.

                        Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.

                        With your input, this post could be even better 💗

                        Register Login
                        • First post
                          Last post