Categories

  • All news regarding Xen and XCP-ng ecosystem

    143 Topics
    4k Posts
    A
    @stormi I'm also getting error on some VMs while trying to export a disk and also trying to even start some VMs from NFS (that were fine before). xo-server[565]: 2026-05-13T02:53:15.746Z xo:api WARN admin | vm.start(...) [2s] =!> XapiError: INTERNAL_ERROR(xenopsd internal error: Storage_error ([S(Illegal_transition);[[S(Activated);S(RO)];[S(Activated);S(RW)]]])) xo-server[565]: 2026-05-13T02:53:40.652Z xo:api WARN admin | vm.start(...) [3s] =!> XapiError: SR_BACKEND_FAILURE_46(, The VDI is not available [opterr=VDI 399734eb-5965-4799-ac36-f6dd774db867 not detached cleanly], )
  • Everything related to the virtualization platform

    1k Topics
    15k Posts
    T
    Hi @maximsachs, Sorry for the delay. From what you describe, the driver seems to be probed in all cases but I have a doubt regarding the driver from xcp-ng 8.2. I rebuilt it specifically for 8.3. To completely eliminate a driver issue, can you try this RPM ? From the host, this can be done by running the following commands: $ wget https://nextcloud.vates.tech/public.php/dav/files/R33Dwpt5gjy6CCr/broadcom-bnxt-en-1.10.0_216.0.119.1-1.0.82srcs.0.xcpng8.3.x86_64.rpm $ yum update ./broadcom-bnxt-en-1.10.0_216.0.119.1-1.0.82srcs.0.xcpng8.3.x86_64.rpm Also, since this xcpng-8.2 release seems to have support for device IDs that have been removed from the 8.3 one, can you give the output of the following shell commands: $ lspci -nn -s 0001:86:00.0 and $ lspci -nn -s 0001:86:00.1 Regards, Thierry
  • 3k Topics
    28k Posts
    J
    WHAT: Automatically assigns CPU weights and I/O priorities based on assigned VM tag (i.e. replicating what vcenter did via resource pools etc.). HOW: Run via cron for regular enforcement. WHY: Automatically assign performance metrics on all pool VMs (as well as preventing configuration drift if settings are accidentally changed). TAGS: The Performance Tiering Concept: 4-tier system with a naming convention that sorts logically in XO: TAG CPU WEIGHT I/O PRIORITY USE CASE 0-core 2048 7 (Highest) Domain Controllers, DNS, DHCP, Core DBs 1-high 1024 7 Critical App Servers 2-normal 256 4 Standard Workloads 3-low 128 1 Dev/Test, Noisy Neighbors Why the "0-" prefix? It forces core VMs to the top of the VM list in XO for easy visibility and management. Important: CPU weights only matter during contention. When the host is under-utilized, all VMs get the performance they need regardless of weight. These are an insurance policy. Script: set-performace.sh bash #!/bin/bash # ============================================ # XCP-ng set-performace.sh script # Tags: 0-core, 1-high, 2-normal, 3-low # ============================================ # --- CONFIGURATION (Customize these for your environment) --- CORE_TAG="0-core" CORE_WEIGHT="2048" CORE_IO_PRI="7" HIGH_TAG="1-high" HIGH_WEIGHT="1024" HIGH_IO_PRI="7" NORMAL_TAG="2-normal" NORMAL_WEIGHT="256" NORMAL_IO_PRI="4" LOW_TAG="3-low" LOW_WEIGHT="128" LOW_IO_PRI="1" LOW_QOS_KBPS="100000" # 100Mbps cap for noisy neighbors # --- CORE CRITICAL VMs --- echo "=== Applying $CORE_TAG CPU & I/O Priority ===" xe vm-list tags:contains="$CORE_TAG" --minimal | tr ',' '\n' | while read uuid; do [ -z "$uuid" ] && continue xe vm-param-set uuid=$uuid VCPUs-params:weight=$CORE_WEIGHT xe vm-param-set uuid=$uuid other-config:sched-pri=$CORE_IO_PRI echo "CORE CRITICAL priority applied: $uuid" done # --- HIGH PRIORITY VMs --- echo "=== Applying $HIGH_TAG CPU & I/O Priority ===" xe vm-list tags:contains="$HIGH_TAG" --minimal | tr ',' '\n' | while read uuid; do [ -z "$uuid" ] && continue xe vm-param-set uuid=$uuid VCPUs-params:weight=$HIGH_WEIGHT xe vm-param-set uuid=$uuid other-config:sched-pri=$HIGH_IO_PRI echo "HIGH priority applied: $uuid" done # --- NORMAL PRIORITY VMs --- echo "=== Applying $NORMAL_TAG CPU & I/O Priority ===" xe vm-list tags:contains="$NORMAL_TAG" --minimal | tr ',' '\n' | while read uuid; do [ -z "$uuid" ] && continue xe vm-param-set uuid=$uuid VCPUs-params:weight=$NORMAL_WEIGHT xe vm-param-set uuid=$uuid other-config:sched-pri=$NORMAL_IO_PRI echo "NORMAL priority applied: $uuid" done # --- LOW PRIORITY VMs (with Network QoS cap) --- echo "=== Applying $LOW_TAG CPU & I/O Priority ===" xe vm-list tags:contains="$LOW_TAG" --minimal | tr ',' '\n' | while read uuid; do [ -z "$uuid" ] && continue xe vm-param-set uuid=$uuid VCPUs-params:weight=$LOW_WEIGHT xe vm-param-set uuid=$uuid other-config:sched-pri=$LOW_IO_PRI echo "LOW priority applied: $uuid" done echo "=== Performance Tuning Complete! ===" How to Deploy: 1 Upload script bash # Copy to your pool master scp set-performace.sh root@your-pool-master:/usr/local/bin/ chmod +x /usr/local/bin/set-performace.sh 2 Add to crontab # Add to crontab (runs hourly) */60 * * * * root /usr/local/bin/set-performance.sh >> /var/log/set-performance.log 2>&1 3 Test # Test it manually /usr/local/bin/set-performace.sh
  • Our hyperconverged storage solution

    47 Topics
    745 Posts
    J
    @Mathieu-L linstor n l was included in my original post. All nodes were updated to May 2026 Security and Maintenance Updates for XCP-ng 8.3 LTS, all nodes were restarted. May 2026 Updates #2 for XCP-ng 8.3 LTS was released, and a couple days later I installed on all hosts. No host restarted. When xen04 was restarted, that is when this issue happened. I had used systemctl restart linstor-controller here (https://xcp-ng.org/forum/post/105309) to restart the controller.
  • 35 Topics
    113 Posts
    olivierlambertO
    Ah excellente nouvelle Je passe le sujet en résolu !