XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login
    1. Home
    2. AudleyElwine
    3. Posts
    A
    Offline
    • Profile
    • Following 0
    • Followers 0
    • Topics 0
    • Posts 13
    • Groups 0

    Posts

    Recent Best Controversial
    • RE: XOSTOR hyperconvergence preview

      Hey @ronan-a ,

      What should I do to lower the chance of something in the past installation of xostor to affect my new installation?
      lsblk is still showing the linstor volumes, vgs is also showing linstor_group.
      Will a wipefs -af be enough? Or is the "Destroying SR" button in XOA is enough?

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      those two UUIDs are in the ./linstor-kv-tool-modified --dump-volumes -u 192.168.0.108 -g xcp-sr-linstor_group_thin_device device output

      ./linstor-kv-tool-modified --dump-volumes -u 192.168.0.108 -g xcp-sr-linstor_group_thin_device
      {
        "xcp/sr/journal/clone/0fb10e9f-b9ef-4b59-8b31-9330f0785514": "86b1b2af-8f1d-4155-9961-d06bbacbb7aa_0e121812-fcae-4d70-960f-ac440b3927e3",
        "xcp/sr/journal/clone/14131ee4-2956-47b7-8728-c9790764f71a": "dfb43813-91eb-46b8-9d56-22c8dbb485fc_917177d5-d03b-495c-b2db-fd62d3d25b86",
        "xcp/sr/journal/clone/45537c14-0125-4f6c-a1ad-476552888087": "36a23780-2025-4f3f-bade-03c410e63368_3e419764-9c8c-4539-9a42-be96f92e5c2a",
        "xcp/sr/journal/clone/54ec7009-2424-4299-a9ad-fb015600b88c": "af89f0fc-7d5a-4236-b249-8d9408f5fb6d_f32f2e8f-a43f-43f5-824b-f673a5cbd988",
        "xcp/sr/journal/clone/558220bc-a900-4408-a62e-a71a4bb4fd7b": "d9294359-c395-4bed-ac3a-bf4027c92bd9_0e18bf3d-78f0-4843-9e8f-ee11c6ebbf5a",
        "xcp/sr/journal/clone/c41e0d47-5c1a-45c3-9404-01f3b5735c0d": "e191eb57-2478-4e3b-be9d-e8eaba8f9efe_41eae673-a280-439b-a4c6-f3afe2390fde",
        "xcp/sr/journal/relink/50170fa2-2ca9-4218-8217-5c99ac31f10b": "1"
      }
      

      So I basically deleted all of the keys here, Maybe I should not have done that, but when I did, eva plugged in correctly to the SR and I was able to finally destroying the SR from XOA. So yeah happy ending. Will try the next beta version. Thank you @ronan-a for your work.

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      Figured out the issue, when I tried to update it from the cli instead. the /var/log partition was full due to /var/log/linstor-controller having something like 3.5G+ data (90% of the /var/log volume). maybe it is due to the past errors it accumulated. I deleted these logs and mike updated normally.

      Now regarding plugging the PBD to eva (the one host that is not connecting to it). it says the following error.

      Error code: SR_BACKEND_FAILURE_202
      Error parameters: , General backend error [opterr=Base copy 36a23780-2025-4f3f-bade-03c410e63368 not present, but no original 45537c14-0125-4f6c-a1ad-476552888087 found],
      

      this is what linstor resource-definition is showing

      [03:59 eva ~]# linstor --controllers=192.168.0.108 resource-definition list -p
      +---------------------------------------------------------------------------+
      | ResourceName            | Port | ResourceGroup                    | State |
      |===========================================================================|
      | xcp-persistent-database | 7000 | xcp-sr-linstor_group_thin_device | ok    |
      +---------------------------------------------------------------------------+
      

      And here is the KV store for linstor from that script

      [04:01 phoebe ~]# mountpoint /var/lib/linstor
      /var/lib/linstor is a mountpoint
      [04:01 phoebe ~]# ./linstor-kv-tool-modified --dump-volumes -u 192.168.0.108 -g xcp-sr-xcp-sr-linstor_group_thin_device
      {
        "xcp/sr/metadata": "{\"name_description\": \"\", \"name_label\": \"XOSTOR\"}"
      }
      [04:01 phoebe ~]# ./linstor-kv-tool-modified --dump-volumes -u 192.168.0.108 -g xcp-sr-linstor_group_thin_device
      {
        "xcp/sr/journal/clone/0fb10e9f-b9ef-4b59-8b31-9330f0785514": "86b1b2af-8f1d-4155-9961-d06bbacbb7aa_0e121812-fcae-4d70-960f-ac440b3927e3",
        "xcp/sr/journal/clone/14131ee4-2956-47b7-8728-c9790764f71a": "dfb43813-91eb-46b8-9d56-22c8dbb485fc_917177d5-d03b-495c-b2db-fd62d3d25b86",
        "xcp/sr/journal/clone/45537c14-0125-4f6c-a1ad-476552888087": "36a23780-2025-4f3f-bade-03c410e63368_3e419764-9c8c-4539-9a42-be96f92e5c2a",
        "xcp/sr/journal/clone/54ec7009-2424-4299-a9ad-fb015600b88c": "af89f0fc-7d5a-4236-b249-8d9408f5fb6d_f32f2e8f-a43f-43f5-824b-f673a5cbd988",
        "xcp/sr/journal/clone/558220bc-a900-4408-a62e-a71a4bb4fd7b": "d9294359-c395-4bed-ac3a-bf4027c92bd9_0e18bf3d-78f0-4843-9e8f-ee11c6ebbf5a",
        "xcp/sr/journal/clone/c41e0d47-5c1a-45c3-9404-01f3b5735c0d": "e191eb57-2478-4e3b-be9d-e8eaba8f9efe_41eae673-a280-439b-a4c6-f3afe2390fde",
        "xcp/sr/journal/relink/50170fa2-2ca9-4218-8217-5c99ac31f10b": "1"
      }
      

      I destroyed the PBD and then recreated it to make it just connect so I can destroy the SR, but the same error happened when I tried to connect with the new PBD that has the same config as the other PBD

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      @ronan-a I started updating xcp-ng so it can both restart and update on my four nodes (eva, phoebe, mike, ozly).
      The nodes were updated with the rolling method, and all three node updated fine, except the forth (mike) (different that the ones that refuses to connect the PBD(eva)) since it is task was stuck at 0.000 progress for 3 hours, so i restarted the toolstack for it(mike) but it didnt do anything, so i restarted the master(eva) node stack. Then when I went to manually update it from XOA, it gives me this error.

      -1(global name 'commmand' is not defined, , Traceback (most recent call last):
        File "/etc/xapi.d/plugins/xcpngutils/__init__.py", line 101, in wrapper
          return func(*args, **kwds)
        File "/etc/xapi.d/plugins/updater.py", line 96, in decorator
          return func(*args, **kwargs)
        File "/etc/xapi.d/plugins/updater.py", line 157, in update
          raise error
      NameError: global name 'commmand' is not defined
      )
      

      The good news is, the linstor controller have moved to a different node(phoebe) from the old one(mike) and I was able to delete all volumes in the linstor --controllers=... resource-definition list except for the database, yet the PBD(eva) could not be connected. And the XOA still shows me a lot of disk, and when I scan it I get this error SR_HAS_NO_PBDS.

      So now mike server cant update, and eva server cant connect its PBDs while all the other servers are connected. Note eva was the server that I started my linstor installation on.

      Do you have any thoughts on what I can do to fix this without reinstalling xcp-ng on mike?

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      @ronan-a Thank you for your fast support.

      I did these changes

      diff -u linstor-kv-tool linstor-kv-tool-modified
      --- linstor-kv-tool	2022-11-17 18:57:00.941259380 +0800
      +++ linstor-kv-tool-modified	2022-11-17 19:04:15.957504667 +0800
      @@ -33,7 +33,7 @@
           kv = linstor.KV(
               group_name,
               uri=controller_uri,
      -        namespace='/xcp/volume/{}'.format(vdi_name)
      +        namespace='/volume/{}'.format(vdi_name)
           )
      
           for key, value in list(kv.items()):
      @@ -46,11 +46,11 @@
               uri=controller_uri,
               namespace='/'
           )
      -
           for key, value in list(kv.items()):
      -        if key.startswith('xcp/volume/'):
      +        if key.startswith('volume/'):
                   size = key.rindex('/')
                   kv.namespace = key[:size]
      +            print("key is {}".format(repr(key[size + 1:])))
                   del kv[key[size + 1:]]
      

      and I got the following error.

      ./linstor-kv-tool-modified --remove-all-volumes -u 192.168.0.106 -g xcp-sr-xcp-sr-linstor_group_thin_device
      key is u'metadata'
      Traceback (most recent call last):
        File "./linstor-kv-tool-modified", line 78, in <module>
          main()
        File "./linstor-kv-tool-modified", line 74, in main
          remove_all_volumes(args.uri, args.group_name)
        File "./linstor-kv-tool-modified", line 54, in remove_all_volumes
          del kv[key[size + 1:]]
        File "/usr/lib/python2.7/site-packages/linstor/kv.py", line 151, in __delitem__
          self._del_linstor_kv(k)
        File "/usr/lib/python2.7/site-packages/linstor/kv.py", line 89, in _del_linstor_kv
          raise linstor.LinstorError('Could not delete kv({}): {}'.format(k, rs[0]))
      linstor.errors.LinstorError: Error: Could not delete kv(/volume/aec2104e-e501-4d7d-b0fb-95a80e843e0a/metadata): ERRO:Exception thrown.
      

      and I can confirm the volume exist when I dump all of them

      "volume/aec2104e-e501-4d7d-b0fb-95a80e843e0a/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": \"\", \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/aec2104e-e501-4d7d-b0fb-95a80e843e0a/not-exists": "0",
        "volume/aec2104e-e501-4d7d-b0fb-95a80e843e0a/volume-name": "xcp-volume-b1748285-7cda-429f-b230-50dfba161e9c",
      

      May I ask what do you recommend me to do? And thank you for your continues support.

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      @ronan-a Thank you for the detailed steps.

      I get the following output when i dump my volumes using the xcp-sr-linstor_group_thin_device group

      {
        "xcp/sr/journal/clone/0fb10e9f-b9ef-4b59-8b31-9330f0785514": "86b1b2af-8f1d-4155-9961-d06bbacbb7aa_0e121812-fcae-4d70-960f-ac440b3927e3",
        "xcp/sr/journal/clone/14131ee4-2956-47b7-8728-c9790764f71a": "dfb43813-91eb-46b8-9d56-22c8dbb485fc_917177d5-d03b-495c-b2db-fd62d3d25b86",
        "xcp/sr/journal/clone/45537c14-0125-4f6c-a1ad-476552888087": "36a23780-2025-4f3f-bade-03c410e63368_3e419764-9c8c-4539-9a42-be96f92e5c2a",
        "xcp/sr/journal/clone/54ec7009-2424-4299-a9ad-fb015600b88c": "af89f0fc-7d5a-4236-b249-8d9408f5fb6d_f32f2e8f-a43f-43f5-824b-f673a5cbd988",
        "xcp/sr/journal/clone/558220bc-a900-4408-a62e-a71a4bb4fd7b": "d9294359-c395-4bed-ac3a-bf4027c92bd9_0e18bf3d-78f0-4843-9e8f-ee11c6ebbf5a",
        "xcp/sr/journal/clone/c41e0d47-5c1a-45c3-9404-01f3b5735c0d": "e191eb57-2478-4e3b-be9d-e8eaba8f9efe_41eae673-a280-439b-a4c6-f3afe2390fde",
        "xcp/sr/journal/relink/50170fa2-2ca9-4218-8217-5c99ac31f10b": "1"
      }
      

      but the --remove-all-volumes does not delete them because they dont start with xcp/volume/.

      Also when i placed xcp-sr-xcp-sr-linstor_group_thin_device a lot of volumes appeared similar to the following

      {
        "volume/00897d74-53c9-41b4-8f5f-73132e4a9af7/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/00897d74-53c9-41b4-8f5f-73132e4a9af7/not-exists": "0",
        "volume/00897d74-53c9-41b4-8f5f-73132e4a9af7/volume-name": "xcp-volume-2892500d-d80a-4978-aa87-ab2b39ace9e9",
        "volume/00b0dbb5-2dfa-4fd5-baf4-81065afa2431/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/00b0dbb5-2dfa-4fd5-baf4-81065afa2431/not-exists": "0",
      ...
      ...
      ...
        "volume/fcbcd0dc-8d90-441e-8d03-e435ac417b96/not-exists": "0",
        "volume/fcbcd0dc-8d90-441e-8d03-e435ac417b96/volume-name": "xcp-volume-f3748b88-1b25-4f18-8f63-4017b09f2ac6",
        "volume/fce3b2e0-1025-4c94-9473-e71562ca11bd/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/fce3b2e0-1025-4c94-9473-e71562ca11bd/not-exists": "0",
        "volume/fce3b2e0-1025-4c94-9473-e71562ca11bd/volume-name": "xcp-volume-08f1fb0b-d6a3-47eb-893b-6c8b08417726",
        "volume/fe6bc8fd-4211-4b4a-8ee5-ba55a7641053/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/fe6bc8fd-4211-4b4a-8ee5-ba55a7641053/not-exists": "0",
        "volume/fe6bc8fd-4211-4b4a-8ee5-ba55a7641053/volume-name": "xcp-volume-7a46e0f4-0f61-4a37-b235-1d2bd9eaf033",
        "volume/fe8dc6e6-a2c6-449a-8858-255a37cc8f98/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": \"\", \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/fe8dc6e6-a2c6-449a-8858-255a37cc8f98/not-exists": "0",
        "volume/fe8dc6e6-a2c6-449a-8858-255a37cc8f98/volume-name": "xcp-volume-0290c420-9f14-43ae-9af5-fe333b60c7dc",
        "volume/feadfc8d-5aeb-429c-8335-4530aa24cc86/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/feadfc8d-5aeb-429c-8335-4530aa24cc86/not-exists": "0",
        "volume/feadfc8d-5aeb-429c-8335-4530aa24cc86/volume-name": "xcp-volume-6a37cf38-d6e4-4af3-90d7-84bec3938b20",
        "xcp/sr/metadata": "{\"name_description\": \"\", \"name_label\": \"XOSTOR\"}"
      }
      

      and the --remove-all-volumes also does not work on them.
      I did the following with and without the xcp-sr prefix and it produced an empty json when specifiying the namesapce as /xcp/volume to match the startwith in the delete thingy.

      ./linstor-kv-tool --dump-volumes -u 192.168.0.106 -g xcp-sr-linstor_group_thin_device -n /xcp/volume
      

      What do you think I should do?

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      @ronan-a I did not forget the SR so yeah it is the PBDs.
      I tried to plug them back with

      xe pbd-plug uuid=...
      

      taking the uuid from the

      xe pbd-list sr-uuid=xostor-uuid
      

      I was able to plug three hosts, however the last forth host says the following.

      Error code: SR_BACKEND_FAILURE_1200
      Error parameters: , Cannot update volume uuid 36a23780-2025-4f3f-bade-03c410e63368 to 45537c14-0125-4f6c-a1ad-476552888087: this last one is not empty,
      

      What do you think I should do to make the forth host pbd connect to delete the SR correctly?

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      Hey @ronan-a
      I'm trying to remove the SR via XOA with "Remove this SR" button but im getting this error.

      SR_NOT_EMPTY()
      

      And there is no VM connected to it. So i tried to delete the disks manually in it but I get this error

      SR_HAS_NO_PBDS(OpaqueRef:6d7520e0-60fa-4b93-9dfe-aa7ceb3b17d2)
      

      Could you help me how to remove it? I dont care about its content so any way is okay for me since I want to reinstall it after i install the updates.

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      Hey @ronan-a , Now all the VDIs on the device are broken, I tried to migrate them but i get errors such as.

      SR_BACKEND_FAILURE_1200(, Cannot update volume uuid 36a23780-2025-4f3f-bade-03c410e63368 to 45537c14-0125-4f6c-a1ad-476552888087: this last one is not empty, )
      
      SR_BACKEND_FAILURE_78(, VDI Creation failed [opterr=error Error: Could not set kv(/volume/9cdc83cc-0fd8-490e-a3af-2ca40c95f398/not-exists:2): ERRO:Exception thrown.], )
      
      SR_BACKEND_FAILURE_46(, The VDI is not available [opterr=Plugin linstor-manager failed], )
      

      I dont care about the broken VDIs content so no worries.
      It was fun experimenting with it, but I need more storage and will move the SSDs to my NAS and run my VMs on NFS there instead.
      Before I do so I thought you might be interested in debugging the issues and getting my logs if that will help the project. Just let me know what files I need to send and will be happy to do so.

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      @ronan-a My appoiliges for replying late. The issue happened again and remembered this thread.
      I tried cat /sys/kernel/debug/drbd/resources/xcp-volume-{UUID}/volumes/0/openers and it is empty across all hosts for both the old broken VDI and the new one.
      The hosts are:

      • eva (master)
      • phoebe
      • mike (linstor controller)
      • ozly

      I also have scheduled backups snapshots so not sure if this will affect the vdi removal.
      Here is the log SMlog.zip.txt The file is not a .txt it is just a .zip (the forum doesnt allow .zip).
      The file is filled with bad volume and idk what to do to fix it.

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      I exported a snapshot for the vm so the data is safe. It is just now I'm worried about restarting other nodes on the cluster.

      I'm happy to debug it I just need some gudiance as I havent done anything special to the VM since it is creation, so not sure why the VDI error appeared.

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      Hi,

      I restarted one of my vms but I couldint get it up due to this error.

      SR_BACKEND_FAILURE_46(, The VDI is not available [opterr=Plugin linstor-manager failed], )
      

      I really dont know why, I tried my best to debug it but I dont know from where to start.

      I did a grep -r {VID-UUID} /var/log, hopefully there is something here we can debug.

      xostor-log-vdi.txt

      Note it was working normally before August 1 3pm in the logs, I cant migrate the data from this XOSTOR to local SR due to another error but I have a backup of the vms.

      I'm on the latest version of the xostor; ./xostor --update-only says im on the latest version.

      Thank you

      posted in XOSTOR
      A
      AudleyElwine
    • RE: XOSTOR hyperconvergence preview

      Hello,

      Have been using XOSTOR for a week. On Xen Orchestra the free space and used space calculations are off.
      09a9da4c-15fb-49a0-967e-462ab90bb1f4-image.png
      173.86/446.91 = 39%, the green lines at the bottom do not look like they are filling up 39%

      Also the free/used space in xen are not accurate, when I do linstor storage-pool list i see a 40GiB used space while Xen Orchestra says 173.86GiB. Note I'm running delta backups on the vms so not sure if this is the cause.

      ╭─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
      ┊ StoragePool                      ┊ Node   ┊ Driver   ┊ PoolName                  ┊ FreeCapacity ┊ TotalCapacity ┊ CanSnapshots ┊ State ┊ SharedName ┊
      ╞═════════════════════════════════════════════════════════════════════════════════════════════════════════════════════════════════════════════════════╡
      ┊ DfltDisklessStorPool             ┊ eva    ┊ DISKLESS ┊                           ┊              ┊               ┊ False        ┊ Ok    ┊            ┊
      ┊ DfltDisklessStorPool             ┊ mike   ┊ DISKLESS ┊                           ┊              ┊               ┊ False        ┊ Ok    ┊            ┊
      ┊ DfltDisklessStorPool             ┊ ozly   ┊ DISKLESS ┊                           ┊              ┊               ┊ False        ┊ Ok    ┊            ┊
      ┊ DfltDisklessStorPool             ┊ phoebe ┊ DISKLESS ┊                           ┊              ┊               ┊ False        ┊ Ok    ┊            ┊
      ┊ xcp-sr-linstor_group_thin_device ┊ eva    ┊ LVM_THIN ┊ linstor_group/thin_device ┊   403.47 GiB ┊    446.91 GiB ┊ True         ┊ Ok    ┊            ┊
      ┊ xcp-sr-linstor_group_thin_device ┊ mike   ┊ LVM_THIN ┊ linstor_group/thin_device ┊   403.47 GiB ┊    446.91 GiB ┊ True         ┊ Ok    ┊            ┊
      ┊ xcp-sr-linstor_group_thin_device ┊ ozly   ┊ LVM_THIN ┊ linstor_group/thin_device ┊   403.47 GiB ┊    446.91 GiB ┊ True         ┊ Ok    ┊            ┊
      ┊ xcp-sr-linstor_group_thin_device ┊ phoebe ┊ LVM_THIN ┊ linstor_group/thin_device ┊   403.38 GiB ┊    446.91 GiB ┊ True         ┊ Ok    ┊            ┊
      ╰─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
      

      should I only focus on linstor storage-pool list and ignore the numbers on Xen orchestra?

      Thank you,

      posted in XOSTOR
      A
      AudleyElwine