XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Hosts fenced, new master, rebooted slaves not reconnecting to pool

    Scheduled Pinned Locked Moved XCP-ng
    1 Posts 1 Posters 10 Views 1 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • nikadeN Offline
      nikade Top contributor
      last edited by nikade

      I'm doing a bit of labbing this friday evening and I ran into a scenario I haven't encountered before and I just wanted to see if this is a bug or if im just being unlucky.

      7st virtual xcp-ng 8.3 vm's running on a physical xcp-ng 8.3, all vm's have nested virtualization enabled.
      I fired up some VM's in the pool to try the load balancer out, by running some "benchmark"-scripts within the vm's to create some load.
      After a while 3/7 hosts failed, because of not enough ram (only 4gb per VM) which isn't really that strange, but after they failed they're not able to "connect" to the pool again:

      e0b56a17-937f-488a-a11c-63a46f6b7491-bild.png

      I then went to the sto-xcp7 vm and checked the pool.conf, only to see that it actually listed sto-xcp8 (which is the master after the fencing):

      [17:48 sto-xcp7 ~]# cat /etc/xensource/pool.conf
      slave:10.200.0.98[17:49 sto-xcp7 ~]#

      I can also go the host in XO and see that it's "halted" but yet, it displays the console:

      479624f6-fe9c-4205-b303-8f1f1c772907-bild.png

      Just for a sanity check, I checked xcp-ng center as well, and it agrees with XO, that the hosts are offline:

      933c545f-65f7-4b48-a2a1-eb8681c9ff29-bild.png

      Is this a bug or what's actually going on? I tried rebooting the failed hosts, without any luck. Any pointers on where to look?

      1 Reply Last reply Reply Quote 0
      • First post
        Last post