XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login
    1. Home
    2. henri9813
    3. Posts
    Offline
    • Profile
    • Following 1
    • Followers 0
    • Topics 11
    • Posts 48
    • Groups 0

    Posts

    Recent Best Controversial
    • RE: VM metadata import fail & stuck

      Hello @nikade.

      I agree but this is my case.

      Try to migrate a running VM: error 500
      Try to migrate an halted VM: error 500.
      Warm migrate: It's okay.

      I don't understand myself the difference except it doesn't transfer the "VM" but recreate the VM and import the VDI, ( so , the same things ), but there may be a light difference. I don't know how the "warm migration" works under the hood

      posted in Xen Orchestra
      henri9813H
      henri9813
    • RE: VM metadata import fail & stuck

      Hello,

      I was able to perform "Warm" migrate from either the slave or the master.

      Yes the master was rebooted.

      @henri9813 Did you reboot the master after it was updated? If yes, I think you should be able to migrate back the VM's to the master, and then continue patcting the rest of the hosts.

      No, i would migrate back vm from MASTER (updated ) to slave ( not updated ), but wasn't working.

      Only warm migrations works.

      posted in Xen Orchestra
      henri9813H
      henri9813
    • RE: VM metadata import fail & stuck

      Hello, @Danp

      My pool is in upgrade, not all nodes are updated.

      1. I evacuate the master hosts's vms to the slave node except 3-4 vms which are "very not important and can be shut for few minutes
      2. I upgrade the master node with success
      3. I WANT to move all vms to the master, but it deosn't have enought disk space, so i tried the following:
        a. Migrate the very not important VM to the slave node ( not updated ) to have enought space to move the "important" vms.
        b. Move vms of the updated master to ANOTHER pool.

      I tried both:

      • VM running
      • VM halted

      Thanks !

      posted in Xen Orchestra
      henri9813H
      henri9813
    • VM metadata import fail & stuck

      I have an updated pool.

      I'm evacuating an host for update to another pool.

      But the operation fail, and remains in the XO tasks

      VM metadata import (on new-hypervisor) 0%
      

      and in the VM.migrate api call detail.

      vm.migrate
      {
        "vm": "30bb4942-c7fd-a3b3-2690-ae6152d272c5",
        "mapVifsNetworks": {
          "7e1ad49f-d4df-d9d7-2a74-0d00486ae5ff": "b3204067-a3fd-bd19-7214-7856e637d076"
        },
        "migrationNetwork": "e31e7aea-37de-2819-83fe-01bd33509855",
        "sr": "3070cc36-b869-a51f-38ee-bd5de5e4cb6c",
        "targetHost": "36a07da2-7493-454d-836d-df8ada5b958f"
      }
      {
        "code": "INTERNAL_ERROR",
        "params": [
          "Http_client.Http_error(\"500\", \"{ frame = false; method = GET; uri = /export_metadata?export_snapshots=true&ref=OpaqueRef:75e166f7-5056-a662-f7ff-25c09aee5bec; query = [  ]; content_length = [  ]; transfer encoding = ; version = 1.0; cookie = [ (value filtered) ]; task = ; subtask_of = OpaqueRef:9976b5f2-3381-e79e-a6dd-0c7a20621501; content-type = ; host = ; user_agent = xapi/25.33; }\")"
        ],
        "task": {
          "uuid": "9c87e615-5dca-c714-0c55-5da571ad8fa5",
          "name_label": "Async.VM.assert_can_migrate",
          "name_description": "",
          "allowed_operations": [],
          "current_operations": {},
          "created": "20260131T08:16:14Z",
          "finished": "20260131T08:16:14Z",
          "status": "failure",
          "resident_on": "OpaqueRef:37858c1b-fa8c-5733-ed66-dcd4fc7ae88c",
          "progress": 1,
          "type": "<none/>",
          "result": "",
          "error_info": [
            "INTERNAL_ERROR",
            "Http_client.Http_error(\"500\", \"{ frame = false; method = GET; uri = /export_metadata?export_snapshots=true&ref=OpaqueRef:75e166f7-5056-a662-f7ff-25c09aee5bec; query = [  ]; content_length = [  ]; transfer encoding = ; version = 1.0; cookie = [ (value filtered) ]; task = ; subtask_of = OpaqueRef:9976b5f2-3381-e79e-a6dd-0c7a20621501; content-type = ; host = ; user_agent = xapi/25.33; }\")"
          ],
          "other_config": {},
          "subtask_of": "OpaqueRef:NULL",
          "subtasks": [],
          "backtrace": "(((process xapi)(filename ocaml/libs/http-lib/http_client.ml)(line 215))((process xapi)(filename ocaml/libs/http-lib/http_client.ml)(line 228))((process xapi)(filename ocaml/libs/http-lib/xmlrpc_client.ml)(line 375))((process xapi)(filename ocaml/libs/xapi-stdext/lib/xapi-stdext-pervasives/pervasiveext.ml)(line 24))((process xapi)(filename ocaml/libs/xapi-stdext/lib/xapi-stdext-pervasives/pervasiveext.ml)(line 39))((process xapi)(filename ocaml/xapi/importexport.ml)(line 313))((process xapi)(filename ocaml/libs/xapi-stdext/lib/xapi-stdext-pervasives/pervasiveext.ml)(line 24))((process xapi)(filename ocaml/libs/xapi-stdext/lib/xapi-stdext-pervasives/pervasiveext.ml)(line 39))((process xapi)(filename ocaml/libs/xapi-stdext/lib/xapi-stdext-pervasives/pervasiveext.ml)(line 24))((process xapi)(filename ocaml/libs/xapi-stdext/lib/xapi-stdext-pervasives/pervasiveext.ml)(line 39))((process xapi)(filename ocaml/xapi/xapi_vm_migrate.ml)(line 1920))((process xapi)(filename ocaml/xapi/message_forwarding.ml)(line 2551))((process xapi)(filename ocaml/xapi/rbac.ml)(line 229))((process xapi)(filename ocaml/xapi/rbac.ml)(line 239))((process xapi)(filename ocaml/xapi/server_helpers.ml)(line 78)))"
        },
        "message": "INTERNAL_ERROR(Http_client.Http_error(\"500\", \"{ frame = false; method = GET; uri = /export_metadata?export_snapshots=true&ref=OpaqueRef:75e166f7-5056-a662-f7ff-25c09aee5bec; query = [  ]; content_length = [  ]; transfer encoding = ; version = 1.0; cookie = [ (value filtered) ]; task = ; subtask_of = OpaqueRef:9976b5f2-3381-e79e-a6dd-0c7a20621501; content-type = ; host = ; user_agent = xapi/25.33; }\"))",
        "name": "XapiError",
        "stack": "XapiError: INTERNAL_ERROR(Http_client.Http_error(\"500\", \"{ frame = false; method = GET; uri = /export_metadata?export_snapshots=true&ref=OpaqueRef:75e166f7-5056-a662-f7ff-25c09aee5bec; query = [  ]; content_length = [  ]; transfer encoding = ; version = 1.0; cookie = [ (value filtered) ]; task = ; subtask_of = OpaqueRef:9976b5f2-3381-e79e-a6dd-0c7a20621501; content-type = ; host = ; user_agent = xapi/25.33; }\"))
          at XapiError.wrap (file:///etc/xen-orchestra/packages/xen-api/_XapiError.mjs:16:12)
          at default (file:///etc/xen-orchestra/packages/xen-api/_getTaskResult.mjs:13:29)
          at Xapi._addRecordToCache (file:///etc/xen-orchestra/packages/xen-api/index.mjs:1078:24)
          at file:///etc/xen-orchestra/packages/xen-api/index.mjs:1112:14
          at Array.forEach (<anonymous>)
          at Xapi._processEvents (file:///etc/xen-orchestra/packages/xen-api/index.mjs:1102:12)
          at Xapi._watchEvents (file:///etc/xen-orchestra/packages/xen-api/index.mjs:1275:14)"
      }
      

      My XO is up to date.

      I already update the master, i'm processing the slave.

      But neither the master ( updated ) nor the slave ( not updated ) can migrate VM to an updated pool.

      Do you have an idea ?

      posted in Xen Orchestra
      henri9813H
      henri9813
    • RE: Lock file is already being held whereas no backup are running.

      Hello @florent ,

      You mean, "check" right ? because if it's doesn't check, the job is finished before the merge and the merge run in background in xen orchestra.

      Best regards,

      posted in Backup
      henri9813H
      henri9813
    • RE: Lock file is already being held whereas no backup are running.

      Hello, @Pilow

      Thanks for the tips !

      But it would be great to see it in the Task UI to find more easily the trick.

      posted in Backup
      henri9813H
      henri9813
    • Lock file is already being held whereas no backup are running.

      Hello,

      Sometimes, i have on some jobs ( which run every 4 hours ).

      30c3a82e-1e48-4d09-8e36-514066e11352-image.png

      Howerver, the previous job is finished in success
      2bc2ed36-391c-4e02-b8d7-a40e9e7513a3-image.png

      Is it possible that has a relation with the mergeWorker of backups which could be running ? if it doesn't finished his operations ?

      Example of logs:

      xen-orchestra      | 2025-11-27T00:44:43.487Z xo:backups:mergeWorker INFO merge in progress {
      xen-orchestra      |   done: 2057,
      xen-orchestra      |   parent: '/xo-vm-backups/f37e259d-beaa-7617-e6f1-be814f21e056/vdis/29e0185e-2f67-44d4-bb9e-ee2a772e2543/b09c0230-219f-4ddf-8e19-bfed1464014f/20251126T070610Z.vhd',
      xen-orchestra      |   progress: 25,
      xen-orchestra      |   total: 8128
      xen-orchestra      | }
      

      Is it possible to have this in the XO Tasks sections ? it's interessant to see this.

      Best regards

      posted in Backup
      henri9813H
      henri9813
    • RE: XOSTOR hyperconvergence preview

      Hello @ronan-a ,

      but how recover from this situation ?

      Thanks !

      posted in XOSTOR
      henri9813H
      henri9813
    • RE: Native Ceph RBD SM driver for XCP-ng

      Hello,

      Thanks for your work !

      We have some hypervisors of tests at Gladhost, we can use them with pleasure to test your work on xcp-ng 8.3 !

      Best regards

      posted in Development
      henri9813H
      henri9813
    • RE: XOSTOR hyperconvergence preview

      Hello,

      I got my whole xostor destroyed, i don't know how precisely.

      I found some errors in sattelite

      Error context:
              An error occurred while processing resource 'Node: 'host', Rsc: 'xcp-volume-e011c043-8751-45e6-be06-4ce9f8807cad''
      ErrorContext:
        Details:     Command 'lvcreate --config 'devices { filter=['"'"'a|/dev/md127|'"'"','"'"'a|/dev/md126p3|'"'"','"'"'r|.*|'"'"'] }' --virtualsize 52543488k linstor_primary --thinpool thin_device --name xcp-volume-e011c043-8751-45e6-be06-4ce9f8807cad_00000' returned with exitcode 5. 
      
      Standard out: 
      
      
      Error message: 
        WARNING: Remaining free space in metadata of thin pool linstor_primary/thin_device is too low (98.06% >= 96.30%). Resize is recommended.
        Cannot create new thin volume, free space in thin pool linstor_primary/thin_device reached threshold.
      

      of course, i checked, my SR was not full
      aa2774a4-c2d4-4dd1-be52-3c6e418c9083-image.png

      And the controller crashed, and i couldn't make it works.

      Here is the error i got

      ==========
      
      Category:                           RuntimeException
      Class name:                         IllegalStateException
      Class canonical name:               java.lang.IllegalStateException
      Generated at:                       Method 'newIllegalStateException', Source file 'DataUtils.java', Line #870
      
      Error message:                      Reading from nio:/var/lib/linstor/linstordb.mv.db failed; file length 2293760 read length 384 at 2445540 [1.4.197/1]
      

      So i deduce the database was fucked-up, i tried to open the file as explained in the documentation, but the linstor schema was "not found" in the file, event if using cat i see data about it.

      for now, i leave xostor and i'm back to localstorage until we know what to do when this issue occured with a "solution path".

      posted in XOSTOR
      henri9813H
      henri9813
    • RE: Multiple disks groups

      Hello, @DustinB

      The https://vates.tech/xostor/ says:

      The maximum size of any single Virtual Disk Image (VDI) will always be limited by the smallest disk in your cluster.

      But in this case, maybe it can be stored in the "2TB disks" ? Maybe others can answer, i didn't test it.

      posted in XOSTOR
      henri9813H
      henri9813
    • RE: Multiple disks groups

      hello @DustinB.

      Yes you right, i would perform this to be able to have VDI with more than 1TB disk. ( which will not be possible because my smallest disk is 1TB (so, 879GB )...

      posted in XOSTOR
      henri9813H
      henri9813
    • XOSTOR Global network disruption test

      Summary

      This test permit to cover the following scenario:

      • Storage network is down
      • All networks are down

      Impact:

      • All hosts cannot see each others
      • Linstor DRDB replication can no longer works ( everyone will be in readonly ).

      Expected results:

      • All write operation on VM fails.
      • a reboot solve the issue.

      Environments

      • 3 hypervisors
        • Node 1
        • Node 2
        • Node 3 ( Master )
      • 13 vms ( mixed of Windows Server and Rocky Linux VMs on XFS, No kubernetes in this section ).
      • VM observation point is : VM1

      We didn't tests other filesystem than XFS for Linux based operating system because we use only XFS.

      Procedure

      • Unplug network cable from all non-master node:
        • Node 1
        • Node 2
      • Keep networks only on the master xcpng node to keep management and observe the behavior
      • Access to a VM located in the master node ( which is still reacheable ).
      • Try to write on VM1, ensure that you have "I/O" error.
      • Wait 5 minutes
      • Re-add back node 1 & Node 2
      • Check states of all VMs.
        • Reboot them if needed.

      Execution

      • Cable disconnected from node 1 and node 2
      • From VM1, we have
      [hdevigne@VM1 ~]$ htop^C
      [hdevigne@VM1 ~]$ echo "coucou" > test
      -bash: test: Input/output error
      [hdevigne@VM1 ~]$ dmesg
      -bash: /usr/bin/dmesg: Input/output error
      [hdevigne@VM1 ~]$ d^C
      [hdevigne@VM1 ~]$ sudo -i
      -bash: sudo: command not found
      [hdevigne@VM1 ~]$ dm^C
      [hdevigne@VM1 ~]$ sudo -i
      -bash: sudo: command not found
      [hdevigne@VM1 ~]$ dmesg
      -bash: /usr/bin/dmesg: Input/output error
      [hdevigne@VM1 ~]$ mount
      -bash: mount: command not found
      [hdevigne@VM1 ~]$ sud o-i
      -bash: sud: command not found
      [hdevigne@VM1 ~]$ sudo -i
      

      ✅ As we predicted it, the vm is completly fucked-up 😄

      • Windows VM crash and reboot in loop.

      • Linstor controller was on node 1, so we will not be able to see linstor nodes status, but we supposed they are in "disconnected" and in "pending eviction", but that doesn't matter a lot, disks are in read only, vm are fucked up after writing, it was our expected bevahior.

      • Re-plug node 1 and node 2.

      • Windows boot normally
        481f9594-2469-4793-9a84-6bd5082caf78-image.png

      • Linux VM stays in a "broken state"

      ➜  ~ ssh VM1 
      suConnection closed by UNKNOWN port 65535
      
      • Force rebooting all VMs from Xen-orchestra permit to revert all vms to a correct state

      Limitation of the test

      We didn't test a duration up to the eviction states of linstor nodes, but the documentation show that a linstor node restore would works ( see https://docs.xcp-ng.org/xostor/#what-to-do-when-a-node-is-in-an-evicted-state )
      We didn't use HA at this time in the cluster, that could helped a bit in the recovery process. but in a precedent experience that i didn't "historize" like this one, the HA was completely down because it was not able to mount a file, i will probably write another topic on the forum to bring my results public.

      Important notes

      Having HA change the criticity of the following note.

      • This test show us that while we don't have HA, all Management components should NOT be placed in XOSTOR to avoid loosing access to it uppon reboot of the VM.
        • If we maintain the idea to put Management component ( XO, Firewall etc... ) in the xostor without HA, we aim to increase recovery time because the recovery will be "manual" from IPMI.
      • Maybe should we simply force reboot nodes after network recovery ? but a bit violent, HA works like this.

      Credit

      Thanks to @olivierlambert, @ronan and other people on the discord canal for answering to daily question which permit to this kind of tests to be made. As promissed, i put my result online 🙂

      Thanks for XOSTOR.

      Futher tests to do: Retry with HA

      posted in XOSTOR
      henri9813H
      henri9813
    • RE: Migration compression is not available on this pool

      Hello,

      We tried the compression feature.

      You "can see" a benefit only if you have a shared storage. ( and again, the migration between 2 nodes is already very fast, we don't see major difference, but maybe a VM will a lot of ram ( >32GB ) can see a difference.

      If you don't have a shared storage ( like XOSTOR, NFS, ISCSI ), then you will not see any difference because there is a limitation of 30MB/s-40MB/s ( see here: https://xcp-ng.org/forum/topic/9389/backup-migration-performance )

      Best regards,

      posted in XCP-ng
      henri9813H
      henri9813
    • RE: Multiple disks groups

      Hello,

      From my test, the result is having multiple xostor is not possible at this time. it's blocked.

      ( i didn't save the precise error message, but the error was clear: cannot have more than one XOSTOR in the pool ).

      posted in XOSTOR
      henri9813H
      henri9813
    • RE: Full or not?

      Hello, @burbilog

      I think theses backup doesn't come from your backup plan.

      Is there maybe snapshots on the VM ? 2022 is "very old"

      Your full backups include memory which is not "classical" in a incremental backups.

      posted in Backup
      henri9813H
      henri9813
    • RE: Unable to add new node to pool using XOSTOR

      Hello,

      I tried on a new pool.

      a little different scenario since i don't create xostor for now, on my previous example, i tried to add a node as replacement of an existing one..

      I just run the install script only on node 1.

      When i try make node2 join the pool, i reproduce the incompatible sm error i got previously.

      The things which is "bizarre", is i don't have the license issue i got on Xen-orchestra. ( maybe it was finally not related ? )

      Here is the complete logs.

      pool.mergeInto
      {
        "sources": [
          "17510fe0-db23-9414-f3df-2941bd34f8dc"
        ],
        "target": "cc91fcdc-c7a8-a44c-65b3-a76dced49252",
        "force": true
      }
      {
        "code": "POOL_JOINING_SM_FEATURES_INCOMPATIBLE",
        "params": [
          "OpaqueRef:090b8da1-9654-066c-84f9-7ab15cb101fd",
          ""
        ],
        "call": {
          "duration": 1061,
          "method": "pool.join_force",
          "params": [
            "* session id *",
            "<MASTER_IP>",
            "root",
            "* obfuscated *"
          ]
        },
        "message": "POOL_JOINING_SM_FEATURES_INCOMPATIBLE(OpaqueRef:090b8da1-9654-066c-84f9-7ab15cb101fd, )",
        "name": "XapiError",
        "stack": "XapiError: POOL_JOINING_SM_FEATURES_INCOMPATIBLE(OpaqueRef:090b8da1-9654-066c-84f9-7ab15cb101fd, )
          at Function.wrap (file:///etc/xen-orchestra/packages/xen-api/_XapiError.mjs:16:12)
          at file:///etc/xen-orchestra/packages/xen-api/transports/json-rpc.mjs:38:21
          at runNextTicks (node:internal/process/task_queues:60:5)
          at processImmediate (node:internal/timers:454:9)
          at process.callbackTrampoline (node:internal/async_hooks:130:17)"
      }```
      posted in XOSTOR
      henri9813H
      henri9813
    • Multiple disks groups

      Hello,

      we have nodes with multiple disk group.

      Does someone experienced multiple xostor on different local group ?

      In our case

      • 2x1TB & 2x2TB

      It should works easily because we gonna register 2 xostor, but maybe should we put all disks in the same VG ? but in this case, the max vdi size will be lower ?

      I purpose to edit the install script to provide group name selection to permit this use case easily.

      It's ok ?

      posted in XOSTOR
      henri9813H
      henri9813
    • RE: Unable to add new node to pool using XOSTOR

      Hello @ronan-a

      Just to be sure, you want logs of the node which want to join or the master ?

      Bonne journée

      posted in XOSTOR
      henri9813H
      henri9813
    • RE: Backup fail whereas xostor cluster is "healthy"

      Hello @ronan-a

      I will reproduce the case, i will re-destroy one hypervisor and retrigger the case.

      Thank you @ronan-a et @olivierlambert 🙂

      If you need me to tests some special case don't hesit, we have a pool dedicated for this

      posted in XOSTOR
      henri9813H
      henri9813