XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login
    1. Home
    2. Popular
    Log in to post
    • All Time
    • Day
    • Week
    • Month
    • All Topics
    • New Topics
    • Watched Topics
    • Unreplied Topics

    • All categories
    • AlexanderKA

      Nested Virtualization in xcp-ng

      Watching Ignoring Scheduled Pinned Locked Moved Compute
      13
      1 Votes
      13 Posts
      313 Views
      Y
      @abudef Some quotes from the documentation to clarify the situation: https://docs.xcp-ng.org/compute/#-nested-virtualization
    • okynnorO

      xcp-ng patches install fail

      Watching Ignoring Scheduled Pinned Locked Moved XCP-ng
      3
      0 Votes
      3 Posts
      82 Views
      P
      https://xcp-ng.org/forum/topic/11951/just-fyi-current-update-seams-to-break-nut-dependancies
    • DustyArmstrongD

      AMD 'Barcelo' passthrough issues - any success stories?

      Watching Ignoring Scheduled Pinned Locked Moved Hardware
      12
      1
      0 Votes
      12 Posts
      492 Views
      T
      @DustyArmstrong Thanks for responding to the GitHub issue. It’s great that more people want this working; it’s difficult to gain traction otherwise. Regarding your list, it’s correct. A reboot should be on the second place. You need to reboot only to detach your PCI device (video card) from its driver and assign it to the pciback driver instead on the next boot. This effectively creates a reservation for the device and allows you to dynamically assign it to VMs. Once your card is free from other kernel drivers, the rest doesn’t require a reboot.
    • kruessK

      XenServer 7.1.2 to XCP-ng 8.3: INTERNAL_ERROR(Not_found)

      Watching Ignoring Scheduled Pinned Locked Moved Migrate to XCP-ng
      2
      0 Votes
      2 Posts
      15 Views
      DanpD
      @kruess I can see numerous instances of VM_HOST_INCOMPATIBLE_VERSION in your xensource log. It is difficult to know the exact cause ithout more details. Was the pool master restarted after the patches were applied? Are you trying to start the VM on the master? Are you upgrading directly from XS 7.1.2 to XCP 8.3?
    • A

      Replication is leaving VDIs attached to Control Domain, again

      Watching Ignoring Scheduled Pinned Locked Moved Backup
      9
      0 Votes
      9 Posts
      345 Views
      A
      @florent I rebuilt my XCP hosting environment (everything is faster and bigger stuffed into one rack).... and this issue is now worse. The main changes to this new setup are 2x40Gb networking, faster NFS NVMe NAS, faster pool servers, more memory, much faster CR destination machine with ZFS. Running XCP 8.3 (March 2026 updates) and XO (master a2e33). Replication is leaving many attached to control domain every day with NBD connection set to 2. Changing it to 1 seems to resolve the issue (no more stuck to control domain).
    • A

      Migrations after updates

      Watching Ignoring Scheduled Pinned Locked Moved Xen Orchestra
      13
      2
      0 Votes
      13 Posts
      370 Views
      Bastien NolletB
      Hi @acebmxer, I've made some tests with a small infrastructure, which helped me understand the behaviour you encounter. With the performance plan, the load balancer can trigger migrations in the following cases: to better satisfy affinity or anti-affinity constraints if a host has a memory or CPU usage exceeds a threshold (85% of the CPU critical threshold, of 1.2 times the free memory critical threshold) with vCPU balancing behaviour, if the vCPU/CPU ratio differs too much from one host to another AND at least one host has more vCPUs than CPUs with preventive behaviour, if CPU usage differs too much from one host to another AND at least one host has more than 25% CPU usage After a host restart, your VMs will be unevenly distributed, but this will not trigger a migration if there are no anti-affinity constraints to satisfy, if no memory or CPU usage thresholds are exceeded, and if no host has more CPUs than vCPUs. If you want migrations to happen after a host restart, you should probably try using the "preventive" behaviour, which can trigger migrations even if thresholds are not reached. However it's based on CPU usage, so if your VMs use a lot of memory but don't use much CPU, this might not be ideal as well. We've received very few feedback about the "preventive" behaviour, so we'd be happy to have yours. As we said before, lowering the critical thresholds might also be a solution, but I think it will make the load balancer less effective if you encounter heavy load a some point.
    • A

      "NOT_SUPPORTED_DURING_UPGRADE()" error after yesterday's update

      Watching Ignoring Scheduled Pinned Locked Moved Backup
      24
      0 Votes
      24 Posts
      1k Views
      M
      @MajorP93 said: -disable HA on pool level -disable load balancer plugin -upgrade master -upgrade all other nodes -restart toolstack on master -restart toolstack on all other nodes -live migrate all VMs running on master to other node(s) -reboot master -reboot next node (live migrate all VMs running on that particular node away before doing so) -repeat until all nodes have been rebooted (one node at a time) -re-enable HA on pool level -re-enable load balancer plugin Never had any issues with that. No downtime for none of the VMs. update time again. and same issue I followed these steps -upgrade master -upgrade all other nodes -restart toolstack on master -restart toolstack on all other nodes -live migrate all VMs running on master to other node(s) -reboot master now cant migrate anything else. live migration : NOT_SUPPORTED_DURING_UPGRADE warm migration: fails and VM shuts down immediately and needs to be forced back to life CR backup to another server : NOT_SUPPORTED_DURING_UPGRADE