XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Migration between pools slow down or interrupted

    Scheduled Pinned Locked Moved Xen Orchestra
    21 Posts 2 Posters 3.0k Views 2 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Tristis OrisT Offline
      Tristis Oris Top contributor @olivierlambert
      last edited by

      @olivierlambert migration.

      1 Reply Last reply Reply Quote 0
      • olivierlambertO Offline
        olivierlambert Vates 🪐 Co-Founder CEO
        last edited by

        So why the log your provided is talking about the vm.copy operation? Please provide a screenshot of the exact button your are clicking in XO.

        Tristis OrisT 1 Reply Last reply Reply Quote 0
        • Tristis OrisT Offline
          Tristis Oris Top contributor @olivierlambert
          last edited by

          @olivierlambert ah, that maybe another log. Too many tests for today.
          But copy also didn't work very good.

          migration at this one.

          vm.migrate
          {
            "vm": "65ab6d31-60ef-7352-bde9-e509a5727660",
            "mapVifsNetworks": {
              "36099318-9cdb-751f-d127-8447a00c66cb": "614b31ae-923d-f049-af90-6a1989d34d28"
            },
            "migrationNetwork": "490f3252-51b1-679f-7b31-dc6fb13ab433",
            "sr": "c1715cfc-59af-ea9d-f1f9-95fc0436c567",
            "targetHost": "0dd76f8c-f33f-4171-b9e9-a0eb9eb2970a"
          }
          {
            "code": 21,
            "data": {
              "objectId": "65ab6d31-60ef-7352-bde9-e509a5727660",
              "code": "SR_BACKEND_FAILURE_47"
            },
            "message": "operation failed",
            "name": "XoError",
            "stack": "XoError: operation failed
              at operationFailed (/opt/xo/xo-builds/xen-orchestra-202212191715/packages/xo-common/api-errors.js:26:11)
              at file:///opt/xo/xo-builds/xen-orchestra-202212191715/packages/xo-server/src/api/vm.mjs:559:15
              at runMicrotasks (<anonymous>)
              at runNextTicks (node:internal/process/task_queues:61:5)
              at processImmediate (node:internal/timers:437:9)
              at process.callbackTrampoline (node:internal/async_hooks:130:17)
              at Xo.migrate (file:///opt/xo/xo-builds/xen-orchestra-202212191715/packages/xo-server/src/api/vm.mjs:545:3)
              at Api.#callApiMethod (file:///opt/xo/xo-builds/xen-orchestra-202212191715/packages/xo-server/src/xo-mixins/api.mjs:394:20)"
          } 
          
          1 Reply Last reply Reply Quote 0
          • olivierlambertO Offline
            olivierlambert Vates 🪐 Co-Founder CEO
            last edited by

            So, as you can read in the log: SR_BACKEND_FAILURE_47. All CAPS messages are coming directly from the host. So it's NOT a Xen Orchestra issue, but something bad on your host.

            Tristis OrisT 1 Reply Last reply Reply Quote 0
            • Tristis OrisT Offline
              Tristis Oris Top contributor @olivierlambert
              last edited by

              @olivierlambert that must like this
              xe vm-migrate uuid=65ab6d31-60ef-7352-bde9-e509a5727660 remote-master=1.2.3.4 remote-username=root remote-password=password host-uuid=1396f539-9f94-42db-a968-289163e0e26e vdi=? VM's disk vdi?

              Performing a Storage XenMotion migration. Your VM's VDIs will be migrated with the VM.
              Will migrate to remote host: hostname, using remote network: eth2 Mng. Here is the VDI mapping:
              VDI 9fcba1f0-bd8a-4f60-8298-519a22793cc8 -> SR 7d01ed49-f06a-31f0-42a3-c7fc8527e65c
              This VIF was not mapped to a destination Network in VM.migrate_send operation
              vif: 36099318-9cdb-751f-d127-8447a00c66cb
              

              what else i need to specify?

              1 Reply Last reply Reply Quote 0
              • olivierlambertO Offline
                olivierlambert Vates 🪐 Co-Founder CEO
                last edited by olivierlambert

                As the answer given by the command, you need to provide a map for your VIFs and your VDIs.

                A map is telling which disk is going to which SR on destination, and which virtual interface is going to which network on destination

                Tristis OrisT 1 Reply Last reply Reply Quote 0
                • Tristis OrisT Offline
                  Tristis Oris Top contributor @olivierlambert
                  last edited by

                  @olivierlambert finally got it.
                  vdi:vm vif id=new pool vif id, not new host.

                  same speed as with XO.
                  c07d9171-cbcf-4388-b065-cdb301f67916-image.png

                  1 Reply Last reply Reply Quote 0
                  • olivierlambertO Offline
                    olivierlambert Vates 🪐 Co-Founder CEO
                    last edited by

                    As I said, then 🙂 Your issue isn't XO related at all.

                    Tristis OrisT 1 Reply Last reply Reply Quote 0
                    • Tristis OrisT Offline
                      Tristis Oris Top contributor @olivierlambert
                      last edited by Tristis Oris

                      @olivierlambert yay!
                      i'm rollback to 6b25a211518dc85700e9a7b81c1f8259f9475826 which is working for me from XO, then start migration from cli:
                      c899d6f8-ed7f-4ba5-a350-e6f3e283dd38-image.png

                      so, that HW or xen problem. could it be related with 8.2.1 clean installation? https://xcp-ng.org/forum/topic/6669/vdi_io_error-continuous-replication-on-clean-install?_=1672425861045

                      but migration to\from this pool works after install 20 days ago.

                      1 Reply Last reply Reply Quote 0
                      • olivierlambertO Offline
                        olivierlambert Vates 🪐 Co-Founder CEO
                        last edited by

                        1. XO version you are using doesn't matter in your case.
                        2. It's hard to answer, I'm not inside your infrastructure. So many moving pieces could affect the migration speed. Network, I/O load, VM size etc.
                        Tristis OrisT 1 Reply Last reply Reply Quote 0
                        • Tristis OrisT Offline
                          Tristis Oris Top contributor @olivierlambert
                          last edited by Tristis Oris

                          @olivierlambert and last test, migration to 3rd pool - same speed.
                          so i have no idea what can be a cause.

                          upd
                          pool2<>pool3 migration fastas should be.
                          3rd was disabled, i just turn it on for this test.
                          so that only pool1 problem.

                          1 Reply Last reply Reply Quote 0
                          • olivierlambertO Offline
                            olivierlambert Vates 🪐 Co-Founder CEO
                            last edited by

                            You can read https://xcp-ng.org/docs/troubleshooting.html and start digging in the logs 🙂

                            Tristis OrisT 1 Reply Last reply Reply Quote 0
                            • Tristis OrisT Offline
                              Tristis Oris Top contributor @olivierlambert
                              last edited by

                              @olivierlambert still trying to find reason of problem. This question more about informative of logs.
                              something happens and now i can't migrate vms outside of pool
                              https://pastebin.com/afHQUiFB

                              ok it say The VDI is not available [opterr=Interrupted operation detected on this VDI, scan SR first to trigger auto-repair]
                              but on scan i got same log
                              https://pastebin.com/WKA1juqz

                              also problem with backups. That another physical storage, nfs connection but same error again.
                              99d5182f-ca0c-4f95-900c-efe0671d64d5-image.png

                              Also snapshot at few vms became orphaned and i can't delete them. again same log here.
                              https://pastebin.com/XPm1Y9Df

                              storage have no error, raid is healthy, i can rescan 2nd lun but have same problems with migration.
                              again nothing interesting at SMlog

                              Jan  9 20:05:07 : [3865] lock: released /var/lock/sm/lvm-22fc6b8e-19f0-880b-e604-68431a93f91c/9fcba1f0-bd8a-4f60-8298-519a22793cc8
                              Jan  9 20:05:07 : [3865] ***** LVHD over iSCSI: EXCEPTION <class 'util.CommandException'>, Invalid argument
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/SRCommand.py", line 376, in run
                              Jan  9 20:05:07 : [3865]     sr = driver(cmd, cmd.sr_uuid)
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/SR.py", line 156, in init
                              Jan  9 20:05:07 : [3865]     self.load(sr_uuid)
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/LVMoISCSISR", line 329, in load
                              Jan  9 20:05:07 : [3865]     LVHDSR.LVHDSR.load(self, sr_uuid)
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/LVHDSR.py", line 197, in load
                              Jan  9 20:05:07 : [3865]     self._undoAllJournals()
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/LVHDSR.py", line 1155, in _undoAllJournals
                              Jan  9 20:05:07 : [3865]     self._undoAllInflateJournals()
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/LVHDSR.py", line 1179, in _undoAllInflateJournals
                              Jan  9 20:05:07 : [3865]     lvhdutil.deflate(self.lvmCache, vdi.lvname, int(val))
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/lvhdutil.py", line 197, in deflate
                              Jan  9 20:05:07 : [3865]     vhdutil.setSizePhys(path, newSize)
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/vhdutil.py", line 262, in setSizePhys
                              Jan  9 20:05:07 : [3865]     ioretry(cmd)
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/vhdutil.py", line 102, in ioretry
                              Jan  9 20:05:07 : [3865]     errlist = [errno.EIO, errno.EAGAIN])
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/util.py", line 330, in ioretry
                              Jan  9 20:05:07 : [3865]     return f()
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/vhdutil.py", line 101, in <lambda>
                              Jan  9 20:05:07 : [3865]     return util.ioretry(lambda: util.pread2(cmd),
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/util.py", line 227, in pread2
                              Jan  9 20:05:07 : [3865]     return pread(cmdlist, quiet = quiet)
                              Jan  9 20:05:07 : [3865]   File "/opt/xensource/sm/util.py", line 190, in pread
                              Jan  9 20:05:07 : [3865]     raise CommandException(rc, str(cmdlist), stderr.strip())
                              

                              is it possible to get more information from XO about whats really wrong?

                              1 Reply Last reply Reply Quote 0
                              • olivierlambertO Offline
                                olivierlambert Vates 🪐 Co-Founder CEO
                                last edited by

                                No, the issue is deep within your host, and not in XO. You need to tail the SMlog while doing a scan, check if you have ISO mounted in your VM and all of that. This can take time, but if you do it carefully, you might your problem 🙂

                                Tristis OrisT 1 Reply Last reply Reply Quote 0
                                • Tristis OrisT Offline
                                  Tristis Oris Top contributor @olivierlambert
                                  last edited by

                                  @olivierlambert solved. that was somewhere here.

                                  Jan 10 16:32:51 SM: [6019] lock: released /var/lock/sm/.nil/lvm
                                  Jan 10 16:32:51 SM: [6019] ['/sbin/dmsetup', 'status', 'VG_XenStorage--22fc6b8e--19f0--880b--e604--68431a93f91c-VHD--9fcba1f0--bd8a--4f60--8298--519a22793cc8']
                                  Jan 10 16:32:51 SM: [6019]   pread SUCCESS
                                  Jan 10 16:32:51 SM: [6019] lock: released /var/lock/sm/lvm-22fc6b8e-19f0-880b-e604-68431a93f91c/9fcba1f0-bd8a-4f60-8298-519a22793cc8
                                  Jan 10 16:32:51 SM: [6019] ***** LVHD over iSCSI: EXCEPTION <class 'util.CommandException'>, Invalid argument
                                  

                                  one of VHD became corrupted so no backups, storage can't be mounted, speed issues, etc. Need to remove it manually and it works again.

                                  in my opinion xen should ignore such error and work with SR same way like with hdd bad blocks. But it works how it works.

                                  1 Reply Last reply Reply Quote 0
                                  • olivierlambertO Offline
                                    olivierlambert Vates 🪐 Co-Founder CEO
                                    last edited by

                                    Good catch 🙂

                                    1 Reply Last reply Reply Quote 0
                                    • First post
                                      Last post