XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,

    Scheduled Pinned Locked Moved Management
    16 Posts 3 Posters 650 Views 2 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • tjkreidlT Offline
      tjkreidl Ambassador @vahric 0
      last edited by

      @vahric-0 VM migration takes a lot of resources; be sure to give dom0 plenty of VCPUs and RAM. You can run top or xentop from the CLI to see the impact during migration and watch for CPU and memory saturation as a sign that dom0 does not have adequate resources.

      1 Reply Last reply Reply Quote 0
      • tjkreidlT Offline
        tjkreidl Ambassador @nikade
        last edited by

        @nikade In XenServer at least, I thought the limit was three VMs being able to be migrated in parallel, according to this:
        https://docs.xenserver.com/en-us/xencenter/current-release/vms-relocate.html

        nikadeN 1 Reply Last reply Reply Quote 0
        • nikadeN Offline
          nikade Top contributor @tjkreidl
          last edited by

          @tjkreidl said in Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,:

          @nikade In XenServer at least, I thought the limit was three VMs being able to be migrated in parallel, according to this:
          https://docs.xenserver.com/en-us/xencenter/current-release/vms-relocate.html

          Yeah, that seems to be correct. I would asume XCP has the same baseline.

          tjkreidlT 1 Reply Last reply Reply Quote 0
          • tjkreidlT Offline
            tjkreidl Ambassador @nikade
            last edited by

            @nikade And if you queue up more than three migration instances, my experience has been that then they are processed such that no more than three run concurrently.

            nikadeN 1 Reply Last reply Reply Quote 0
            • nikadeN Offline
              nikade Top contributor @tjkreidl
              last edited by

              @tjkreidl said in Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,:

              @nikade And if you queue up more than three migration instances, my experience has been that then they are processed such that no more than three run concurrently.

              Yeah, thats good to hear.
              Any idea if there's a max number of migrations you can queue?

              tjkreidlT 1 Reply Last reply Reply Quote 0
              • tjkreidlT Offline
                tjkreidl Ambassador @nikade
                last edited by tjkreidl

                @nikade No, but I've done a lot -- probably one or two dozen -- when doing updates to help speed up the evacuation of hosts. You can check the queue with
                "xe task-list" to see what's being processed or queued.

                nikadeN 1 Reply Last reply Reply Quote 0
                • nikadeN Offline
                  nikade Top contributor @tjkreidl
                  last edited by

                  @tjkreidl said in Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,:

                  @nikade No, but I've done a lot -- probably one or two dozen -- when doing updates to help speed up the evacuation of hosts. You can check the queue with
                  "xe task-list" to see what's being processed or queued.

                  Cool, I rarely do migrate them manually so I wouldn't know.

                  V 1 Reply Last reply Reply Quote 0
                  • V Offline
                    vahric 0 @nikade
                    last edited by

                    @nikade @tjkreidl
                    increasing memory of Dom0 did not effect ...
                    increasing vCPU of Dom0 did not effect (actually not all vcpu already used for it but i just want to try )
                    I run stress*ng for load vms memory but did not effect
                    No pinning or numa config need because single cpu and shared L3 cache for all cores
                    Also MTU size is not effecting its working same with 1500 and 9000 MTU
                    I saw and change tcp_limit_output_bytes but did not help me

                    Only what effect is changing the hardware
                    My Intel servers are Intel(R) Xeon(R) CPU E5-2620 v3 @ 2.40GHz --> 0.9 Gbit/s per migration
                    My AMD servers are AMD EPYC 7502P 32-Core Processor --> 1.76 Gbit/s per migration

                    Do you have any advise ?

                    nikadeN 1 Reply Last reply Reply Quote 0
                    • nikadeN Offline
                      nikade Top contributor @vahric 0
                      last edited by

                      @vahric-0 Are you using shared storage? Your numbers looks like you are migrating storage as well, im seeing way higher numbers within our pools and they also have the same CPU (Dual socket Intel(R) Xeon(R) CPU E5-2620 v3 @ 2.40GHz).

                      V 1 Reply Last reply Reply Quote 0
                      • V Offline
                        vahric 0 @nikade
                        last edited by

                        @nikade No storage migration , they are all sit in shared storage

                        nikadeN 1 Reply Last reply Reply Quote 0
                        • nikadeN Offline
                          nikade Top contributor @vahric 0
                          last edited by

                          @vahric-0 said in Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,:

                          @nikade No storage migration , they are all sit in shared storage

                          Thats strange, are all the XCP hosts within the same network as well? Not different VLAN's or subnets?

                          V 1 Reply Last reply Reply Quote 0
                          • V Offline
                            vahric 0 @nikade
                            last edited by

                            @nikade yes same, they are running on same broadcast domain

                            nikadeN 1 Reply Last reply Reply Quote 0
                            • nikadeN Offline
                              nikade Top contributor @vahric 0
                              last edited by

                              @vahric-0 Then I am out of suggestions, sorry.

                              1 Reply Last reply Reply Quote 0
                              • First post
                                Last post