XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,

    Scheduled Pinned Locked Moved Management
    16 Posts 3 Posters 650 Views 2 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • V Offline
      vahric 0
      last edited by

      Hello Everyone,

      After enable HA we saw that if you have two VMs one HA disabled other is reset mode we got "HOST_NOT_ENOUGH_FREE_MEMORY" error, after set disabled one reset then no error appear any more.

      We don't understand why step by step we should enable from each VM HA mode reset , i mean if we configured HA all VM should be protected isn't it ? it should be default behaviour i guess 🙂

      When try to soft reboot host, could't find where i have to adjust parallel migration number for VMs, its 1 by 1 migrating ....

      Looks like balance mode do not care about until CPU and Mem thresholds kicked , right ?

      Thanks
      VM

      tjkreidlT 1 Reply Last reply Reply Quote 0
      • nikadeN Offline
        nikade Top contributor
        last edited by

        What does the memory utilization looks on all the hosts in the pool?
        How much max RAM did you allocate the VM's that has the restart option enabled?

        About the parallel migration I think you are right, it will migrate the VM's one by one to another host in the pool. Haven't found any setting to change this.

        V tjkreidlT 2 Replies Last reply Reply Quote 0
        • V Offline
          vahric 0 @nikade
          last edited by vahric 0

          @nikade I guess limiter is Dom0 ,

          f157ac8a-1a13-4761-8da3-2e1c25757367-image.png

          I will try it with new hardware ....

          1 Reply Last reply Reply Quote 0
          • tjkreidlT Offline
            tjkreidl Ambassador @vahric 0
            last edited by

            @vahric-0 VM migration takes a lot of resources; be sure to give dom0 plenty of VCPUs and RAM. You can run top or xentop from the CLI to see the impact during migration and watch for CPU and memory saturation as a sign that dom0 does not have adequate resources.

            1 Reply Last reply Reply Quote 0
            • tjkreidlT Offline
              tjkreidl Ambassador @nikade
              last edited by

              @nikade In XenServer at least, I thought the limit was three VMs being able to be migrated in parallel, according to this:
              https://docs.xenserver.com/en-us/xencenter/current-release/vms-relocate.html

              nikadeN 1 Reply Last reply Reply Quote 0
              • nikadeN Offline
                nikade Top contributor @tjkreidl
                last edited by

                @tjkreidl said in Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,:

                @nikade In XenServer at least, I thought the limit was three VMs being able to be migrated in parallel, according to this:
                https://docs.xenserver.com/en-us/xencenter/current-release/vms-relocate.html

                Yeah, that seems to be correct. I would asume XCP has the same baseline.

                tjkreidlT 1 Reply Last reply Reply Quote 0
                • tjkreidlT Offline
                  tjkreidl Ambassador @nikade
                  last edited by

                  @nikade And if you queue up more than three migration instances, my experience has been that then they are processed such that no more than three run concurrently.

                  nikadeN 1 Reply Last reply Reply Quote 0
                  • nikadeN Offline
                    nikade Top contributor @tjkreidl
                    last edited by

                    @tjkreidl said in Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,:

                    @nikade And if you queue up more than three migration instances, my experience has been that then they are processed such that no more than three run concurrently.

                    Yeah, thats good to hear.
                    Any idea if there's a max number of migrations you can queue?

                    tjkreidlT 1 Reply Last reply Reply Quote 0
                    • tjkreidlT Offline
                      tjkreidl Ambassador @nikade
                      last edited by tjkreidl

                      @nikade No, but I've done a lot -- probably one or two dozen -- when doing updates to help speed up the evacuation of hosts. You can check the queue with
                      "xe task-list" to see what's being processed or queued.

                      nikadeN 1 Reply Last reply Reply Quote 0
                      • nikadeN Offline
                        nikade Top contributor @tjkreidl
                        last edited by

                        @tjkreidl said in Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,:

                        @nikade No, but I've done a lot -- probably one or two dozen -- when doing updates to help speed up the evacuation of hosts. You can check the queue with
                        "xe task-list" to see what's being processed or queued.

                        Cool, I rarely do migrate them manually so I wouldn't know.

                        V 1 Reply Last reply Reply Quote 0
                        • V Offline
                          vahric 0 @nikade
                          last edited by

                          @nikade @tjkreidl
                          increasing memory of Dom0 did not effect ...
                          increasing vCPU of Dom0 did not effect (actually not all vcpu already used for it but i just want to try )
                          I run stress*ng for load vms memory but did not effect
                          No pinning or numa config need because single cpu and shared L3 cache for all cores
                          Also MTU size is not effecting its working same with 1500 and 9000 MTU
                          I saw and change tcp_limit_output_bytes but did not help me

                          Only what effect is changing the hardware
                          My Intel servers are Intel(R) Xeon(R) CPU E5-2620 v3 @ 2.40GHz --> 0.9 Gbit/s per migration
                          My AMD servers are AMD EPYC 7502P 32-Core Processor --> 1.76 Gbit/s per migration

                          Do you have any advise ?

                          nikadeN 1 Reply Last reply Reply Quote 0
                          • nikadeN Offline
                            nikade Top contributor @vahric 0
                            last edited by

                            @vahric-0 Are you using shared storage? Your numbers looks like you are migrating storage as well, im seeing way higher numbers within our pools and they also have the same CPU (Dual socket Intel(R) Xeon(R) CPU E5-2620 v3 @ 2.40GHz).

                            V 1 Reply Last reply Reply Quote 0
                            • V Offline
                              vahric 0 @nikade
                              last edited by

                              @nikade No storage migration , they are all sit in shared storage

                              nikadeN 1 Reply Last reply Reply Quote 0
                              • nikadeN Offline
                                nikade Top contributor @vahric 0
                                last edited by

                                @vahric-0 said in Some HA Questions Memory Error, Parallel Migrate, HA for all VMs,:

                                @nikade No storage migration , they are all sit in shared storage

                                Thats strange, are all the XCP hosts within the same network as well? Not different VLAN's or subnets?

                                V 1 Reply Last reply Reply Quote 0
                                • V Offline
                                  vahric 0 @nikade
                                  last edited by

                                  @nikade yes same, they are running on same broadcast domain

                                  nikadeN 1 Reply Last reply Reply Quote 0
                                  • nikadeN Offline
                                    nikade Top contributor @vahric 0
                                    last edited by

                                    @vahric-0 Then I am out of suggestions, sorry.

                                    1 Reply Last reply Reply Quote 0
                                    • First post
                                      Last post