XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Migrating an offline VM disk between two local SRs is slow

    Scheduled Pinned Locked Moved Xen Orchestra
    22 Posts 6 Posters 615 Views 5 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • olivierlambertO Offline
      olivierlambert Vates 🪐 Co-Founder CEO
      last edited by olivierlambert

      Mostly CPU bound (single disk migration isn't multithreaded). Higher your CPU frequency, faster the migration.

      P D 2 Replies Last reply Reply Quote 0
      • P Offline
        pkgw @olivierlambert
        last edited by pkgw

        @olivierlambert Thanks, that's good to know. I appreciate your taking the time to discuss. I don't suppose there are any settings we can fiddle with that would speed up the single-disk scenario? Or some workaround approach that might get closer to the hardware's native speed? (In one experiment, someone did something that caused the system to transfer the OS disk with the log message "Cloning VDI" rather than "Creating a blank remote VDI", and the effective throughput was higher by a factor of 20 ...)

        1 Reply Last reply Reply Quote 0
        • olivierlambertO Offline
          olivierlambert Vates 🪐 Co-Founder CEO
          last edited by olivierlambert

          Can you describe exactly the steps that were done so we can double check/compare and understand the why?

          edit: also, are you comparing a live migration vs an offline copy? It's very different, since in live you have to replicate the blocks while the VM on top is running.

          P 1 Reply Last reply Reply Quote 0
          • P Offline
            pkgw @olivierlambert
            last edited by

            @olivierlambert This is all offline. Unfortunately I can't describe exactly what was done, since someone else was doing the work and they were trying a bunch of different things all in a row. I suspect that the apparently fast migration is a red herring (maybe a previous attempt left a copy of the disk on the destination SR, and the system noticed that and avoided the actual I/O?) but if there turned out to be a magical fast path, I wouldn't complain!

            1 Reply Last reply Reply Quote 0
            • olivierlambertO Offline
              olivierlambert Vates 🪐 Co-Founder CEO
              last edited by

              You can also try warm migration, which can go a lot faster.

              ForzaF 1 Reply Last reply Reply Quote 1
              • ForzaF Offline
                Forza @olivierlambert
                last edited by Forza

                Using XOA "Disaster Recovery" backup method can be a lot faster than normal offline migration.

                One time I did it, it took approx 10 minutes instead of 2 hours...

                M 1 Reply Last reply Reply Quote 0
                • M Offline
                  magicker @Forza
                  last edited by

                  I think I am seeing a similar issue. Raid1 NVME copy to raid 10 4x2tb HDD on same host

                  a 300gb transfer is estimated at 7 hours. (11% done in 50 mins)

                  the vm is live.

                  according to the stats almost nothing is happening on this server or the 2 storage

                  1 Reply Last reply Reply Quote 1
                  • D Offline
                    Davidj 0 @olivierlambert
                    last edited by

                    @olivierlambert
                    Is the CPU on the sending host or the receiving host the limiting factor for single disk migrations?

                    M 1 Reply Last reply Reply Quote 0
                    • olivierlambertO Offline
                      olivierlambert Vates 🪐 Co-Founder CEO
                      last edited by

                      I can't really tell, gut feeling is the sending host, but I have no numbers to confirm.

                      1 Reply Last reply Reply Quote 0
                      • M Offline
                        magicker @Davidj 0
                        last edited by magicker

                        @Davidj-0 in my case there CPU activity is minimal. I think something is wrong with the software raid 10 setup. On an identical setup warm migration between to the raid 10 array between hosts is showing horrible iowait similar to the sr to sr transfer on the other host

                        bb40264a-7678-4906-aca5-788bc217c7f4-image.png

                        1 Reply Last reply Reply Quote 0
                        • olivierlambertO Offline
                          olivierlambert Vates 🪐 Co-Founder CEO
                          last edited by

                          Maybe the IO scheduler is not the right one?

                          1 Reply Last reply Reply Quote 0
                          • First post
                            Last post