XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    XOSTOR hyperconvergence preview

    Scheduled Pinned Locked Moved XOSTOR
    446 Posts 47 Posters 481.2k Views 48 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • A Offline
      AudleyElwine @ronan-a
      last edited by AudleyElwine

      @ronan-a Thank you for the detailed steps.

      I get the following output when i dump my volumes using the xcp-sr-linstor_group_thin_device group

      {
        "xcp/sr/journal/clone/0fb10e9f-b9ef-4b59-8b31-9330f0785514": "86b1b2af-8f1d-4155-9961-d06bbacbb7aa_0e121812-fcae-4d70-960f-ac440b3927e3",
        "xcp/sr/journal/clone/14131ee4-2956-47b7-8728-c9790764f71a": "dfb43813-91eb-46b8-9d56-22c8dbb485fc_917177d5-d03b-495c-b2db-fd62d3d25b86",
        "xcp/sr/journal/clone/45537c14-0125-4f6c-a1ad-476552888087": "36a23780-2025-4f3f-bade-03c410e63368_3e419764-9c8c-4539-9a42-be96f92e5c2a",
        "xcp/sr/journal/clone/54ec7009-2424-4299-a9ad-fb015600b88c": "af89f0fc-7d5a-4236-b249-8d9408f5fb6d_f32f2e8f-a43f-43f5-824b-f673a5cbd988",
        "xcp/sr/journal/clone/558220bc-a900-4408-a62e-a71a4bb4fd7b": "d9294359-c395-4bed-ac3a-bf4027c92bd9_0e18bf3d-78f0-4843-9e8f-ee11c6ebbf5a",
        "xcp/sr/journal/clone/c41e0d47-5c1a-45c3-9404-01f3b5735c0d": "e191eb57-2478-4e3b-be9d-e8eaba8f9efe_41eae673-a280-439b-a4c6-f3afe2390fde",
        "xcp/sr/journal/relink/50170fa2-2ca9-4218-8217-5c99ac31f10b": "1"
      }
      

      but the --remove-all-volumes does not delete them because they dont start with xcp/volume/.

      Also when i placed xcp-sr-xcp-sr-linstor_group_thin_device a lot of volumes appeared similar to the following

      {
        "volume/00897d74-53c9-41b4-8f5f-73132e4a9af7/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/00897d74-53c9-41b4-8f5f-73132e4a9af7/not-exists": "0",
        "volume/00897d74-53c9-41b4-8f5f-73132e4a9af7/volume-name": "xcp-volume-2892500d-d80a-4978-aa87-ab2b39ace9e9",
        "volume/00b0dbb5-2dfa-4fd5-baf4-81065afa2431/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/00b0dbb5-2dfa-4fd5-baf4-81065afa2431/not-exists": "0",
      ...
      ...
      ...
        "volume/fcbcd0dc-8d90-441e-8d03-e435ac417b96/not-exists": "0",
        "volume/fcbcd0dc-8d90-441e-8d03-e435ac417b96/volume-name": "xcp-volume-f3748b88-1b25-4f18-8f63-4017b09f2ac6",
        "volume/fce3b2e0-1025-4c94-9473-e71562ca11bd/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/fce3b2e0-1025-4c94-9473-e71562ca11bd/not-exists": "0",
        "volume/fce3b2e0-1025-4c94-9473-e71562ca11bd/volume-name": "xcp-volume-08f1fb0b-d6a3-47eb-893b-6c8b08417726",
        "volume/fe6bc8fd-4211-4b4a-8ee5-ba55a7641053/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/fe6bc8fd-4211-4b4a-8ee5-ba55a7641053/not-exists": "0",
        "volume/fe6bc8fd-4211-4b4a-8ee5-ba55a7641053/volume-name": "xcp-volume-7a46e0f4-0f61-4a37-b235-1d2bd9eaf033",
        "volume/fe8dc6e6-a2c6-449a-8858-255a37cc8f98/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": \"\", \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/fe8dc6e6-a2c6-449a-8858-255a37cc8f98/not-exists": "0",
        "volume/fe8dc6e6-a2c6-449a-8858-255a37cc8f98/volume-name": "xcp-volume-0290c420-9f14-43ae-9af5-fe333b60c7dc",
        "volume/feadfc8d-5aeb-429c-8335-4530aa24cc86/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": null, \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
        "volume/feadfc8d-5aeb-429c-8335-4530aa24cc86/not-exists": "0",
        "volume/feadfc8d-5aeb-429c-8335-4530aa24cc86/volume-name": "xcp-volume-6a37cf38-d6e4-4af3-90d7-84bec3938b20",
        "xcp/sr/metadata": "{\"name_description\": \"\", \"name_label\": \"XOSTOR\"}"
      }
      

      and the --remove-all-volumes also does not work on them.
      I did the following with and without the xcp-sr prefix and it produced an empty json when specifiying the namesapce as /xcp/volume to match the startwith in the delete thingy.

      ./linstor-kv-tool --dump-volumes -u 192.168.0.106 -g xcp-sr-linstor_group_thin_device -n /xcp/volume
      

      What do you think I should do?

      ronan-aR 1 Reply Last reply Reply Quote 0
      • ronan-aR Offline
        ronan-a Vates 🪐 XCP-ng Team @AudleyElwine
        last edited by

        @AudleyElwine said in XOSTOR hyperconvergence preview:

        but the --remove-all-volumes does not delete them because they dont start with xcp/volume/.

        Right, another problem already fixed, but I forgot to put an adapted version on my gist, sorry, you can modify the script to use volume/ instead of xcp/volume. šŸ™‚

        A 1 Reply Last reply Reply Quote 0
        • A Offline
          AudleyElwine @ronan-a
          last edited by

          @ronan-a Thank you for your fast support.

          I did these changes

          diff -u linstor-kv-tool linstor-kv-tool-modified
          --- linstor-kv-tool	2022-11-17 18:57:00.941259380 +0800
          +++ linstor-kv-tool-modified	2022-11-17 19:04:15.957504667 +0800
          @@ -33,7 +33,7 @@
               kv = linstor.KV(
                   group_name,
                   uri=controller_uri,
          -        namespace='/xcp/volume/{}'.format(vdi_name)
          +        namespace='/volume/{}'.format(vdi_name)
               )
          
               for key, value in list(kv.items()):
          @@ -46,11 +46,11 @@
                   uri=controller_uri,
                   namespace='/'
               )
          -
               for key, value in list(kv.items()):
          -        if key.startswith('xcp/volume/'):
          +        if key.startswith('volume/'):
                       size = key.rindex('/')
                       kv.namespace = key[:size]
          +            print("key is {}".format(repr(key[size + 1:])))
                       del kv[key[size + 1:]]
          

          and I got the following error.

          ./linstor-kv-tool-modified --remove-all-volumes -u 192.168.0.106 -g xcp-sr-xcp-sr-linstor_group_thin_device
          key is u'metadata'
          Traceback (most recent call last):
            File "./linstor-kv-tool-modified", line 78, in <module>
              main()
            File "./linstor-kv-tool-modified", line 74, in main
              remove_all_volumes(args.uri, args.group_name)
            File "./linstor-kv-tool-modified", line 54, in remove_all_volumes
              del kv[key[size + 1:]]
            File "/usr/lib/python2.7/site-packages/linstor/kv.py", line 151, in __delitem__
              self._del_linstor_kv(k)
            File "/usr/lib/python2.7/site-packages/linstor/kv.py", line 89, in _del_linstor_kv
              raise linstor.LinstorError('Could not delete kv({}): {}'.format(k, rs[0]))
          linstor.errors.LinstorError: Error: Could not delete kv(/volume/aec2104e-e501-4d7d-b0fb-95a80e843e0a/metadata): ERRO:Exception thrown.
          

          and I can confirm the volume exist when I dump all of them

          "volume/aec2104e-e501-4d7d-b0fb-95a80e843e0a/metadata": "{\"read_only\": true, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": \"\", \"name_label\": \"base copy\", \"name_description\": \"\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}",
            "volume/aec2104e-e501-4d7d-b0fb-95a80e843e0a/not-exists": "0",
            "volume/aec2104e-e501-4d7d-b0fb-95a80e843e0a/volume-name": "xcp-volume-b1748285-7cda-429f-b230-50dfba161e9c",
          

          May I ask what do you recommend me to do? And thank you for your continues support.

          ronan-aR 1 Reply Last reply Reply Quote 0
          • ronan-aR Offline
            ronan-a Vates 🪐 XCP-ng Team @AudleyElwine
            last edited by

            @AudleyElwine said in XOSTOR hyperconvergence preview:
            Really strange... Maybe there is a lock or another issue with LINSTOR. In the worst case you can retry after a reboot of all hosts. If it's always stuck I can take a look using a support tunnel, I'm not sure to understand why you have this error.

            A 1 Reply Last reply Reply Quote 0
            • A Offline
              AudleyElwine @ronan-a
              last edited by

              @ronan-a I started updating xcp-ng so it can both restart and update on my four nodes (eva, phoebe, mike, ozly).
              The nodes were updated with the rolling method, and all three node updated fine, except the forth (mike) (different that the ones that refuses to connect the PBD(eva)) since it is task was stuck at 0.000 progress for 3 hours, so i restarted the toolstack for it(mike) but it didnt do anything, so i restarted the master(eva) node stack. Then when I went to manually update it from XOA, it gives me this error.

              -1(global name 'commmand' is not defined, , Traceback (most recent call last):
                File "/etc/xapi.d/plugins/xcpngutils/__init__.py", line 101, in wrapper
                  return func(*args, **kwds)
                File "/etc/xapi.d/plugins/updater.py", line 96, in decorator
                  return func(*args, **kwargs)
                File "/etc/xapi.d/plugins/updater.py", line 157, in update
                  raise error
              NameError: global name 'commmand' is not defined
              )
              

              The good news is, the linstor controller have moved to a different node(phoebe) from the old one(mike) and I was able to delete all volumes in the linstor --controllers=... resource-definition list except for the database, yet the PBD(eva) could not be connected. And the XOA still shows me a lot of disk, and when I scan it I get this error SR_HAS_NO_PBDS.

              So now mike server cant update, and eva server cant connect its PBDs while all the other servers are connected. Note eva was the server that I started my linstor installation on.

              Do you have any thoughts on what I can do to fix this without reinstalling xcp-ng on mike?

              A 1 Reply Last reply Reply Quote 0
              • A Offline
                AudleyElwine @AudleyElwine
                last edited by AudleyElwine

                Figured out the issue, when I tried to update it from the cli instead. the /var/log partition was full due to /var/log/linstor-controller having something like 3.5G+ data (90% of the /var/log volume). maybe it is due to the past errors it accumulated. I deleted these logs and mike updated normally.

                Now regarding plugging the PBD to eva (the one host that is not connecting to it). it says the following error.

                Error code: SR_BACKEND_FAILURE_202
                Error parameters: , General backend error [opterr=Base copy 36a23780-2025-4f3f-bade-03c410e63368 not present, but no original 45537c14-0125-4f6c-a1ad-476552888087 found],
                

                this is what linstor resource-definition is showing

                [03:59 eva ~]# linstor --controllers=192.168.0.108 resource-definition list -p
                +---------------------------------------------------------------------------+
                | ResourceName            | Port | ResourceGroup                    | State |
                |===========================================================================|
                | xcp-persistent-database | 7000 | xcp-sr-linstor_group_thin_device | ok    |
                +---------------------------------------------------------------------------+
                

                And here is the KV store for linstor from that script

                [04:01 phoebe ~]# mountpoint /var/lib/linstor
                /var/lib/linstor is a mountpoint
                [04:01 phoebe ~]# ./linstor-kv-tool-modified --dump-volumes -u 192.168.0.108 -g xcp-sr-xcp-sr-linstor_group_thin_device
                {
                  "xcp/sr/metadata": "{\"name_description\": \"\", \"name_label\": \"XOSTOR\"}"
                }
                [04:01 phoebe ~]# ./linstor-kv-tool-modified --dump-volumes -u 192.168.0.108 -g xcp-sr-linstor_group_thin_device
                {
                  "xcp/sr/journal/clone/0fb10e9f-b9ef-4b59-8b31-9330f0785514": "86b1b2af-8f1d-4155-9961-d06bbacbb7aa_0e121812-fcae-4d70-960f-ac440b3927e3",
                  "xcp/sr/journal/clone/14131ee4-2956-47b7-8728-c9790764f71a": "dfb43813-91eb-46b8-9d56-22c8dbb485fc_917177d5-d03b-495c-b2db-fd62d3d25b86",
                  "xcp/sr/journal/clone/45537c14-0125-4f6c-a1ad-476552888087": "36a23780-2025-4f3f-bade-03c410e63368_3e419764-9c8c-4539-9a42-be96f92e5c2a",
                  "xcp/sr/journal/clone/54ec7009-2424-4299-a9ad-fb015600b88c": "af89f0fc-7d5a-4236-b249-8d9408f5fb6d_f32f2e8f-a43f-43f5-824b-f673a5cbd988",
                  "xcp/sr/journal/clone/558220bc-a900-4408-a62e-a71a4bb4fd7b": "d9294359-c395-4bed-ac3a-bf4027c92bd9_0e18bf3d-78f0-4843-9e8f-ee11c6ebbf5a",
                  "xcp/sr/journal/clone/c41e0d47-5c1a-45c3-9404-01f3b5735c0d": "e191eb57-2478-4e3b-be9d-e8eaba8f9efe_41eae673-a280-439b-a4c6-f3afe2390fde",
                  "xcp/sr/journal/relink/50170fa2-2ca9-4218-8217-5c99ac31f10b": "1"
                }
                

                I destroyed the PBD and then recreated it to make it just connect so I can destroy the SR, but the same error happened when I tried to connect with the new PBD that has the same config as the other PBD

                A 1 Reply Last reply Reply Quote 0
                • A Offline
                  AudleyElwine @AudleyElwine
                  last edited by

                  those two UUIDs are in the ./linstor-kv-tool-modified --dump-volumes -u 192.168.0.108 -g xcp-sr-linstor_group_thin_device device output

                  ./linstor-kv-tool-modified --dump-volumes -u 192.168.0.108 -g xcp-sr-linstor_group_thin_device
                  {
                    "xcp/sr/journal/clone/0fb10e9f-b9ef-4b59-8b31-9330f0785514": "86b1b2af-8f1d-4155-9961-d06bbacbb7aa_0e121812-fcae-4d70-960f-ac440b3927e3",
                    "xcp/sr/journal/clone/14131ee4-2956-47b7-8728-c9790764f71a": "dfb43813-91eb-46b8-9d56-22c8dbb485fc_917177d5-d03b-495c-b2db-fd62d3d25b86",
                    "xcp/sr/journal/clone/45537c14-0125-4f6c-a1ad-476552888087": "36a23780-2025-4f3f-bade-03c410e63368_3e419764-9c8c-4539-9a42-be96f92e5c2a",
                    "xcp/sr/journal/clone/54ec7009-2424-4299-a9ad-fb015600b88c": "af89f0fc-7d5a-4236-b249-8d9408f5fb6d_f32f2e8f-a43f-43f5-824b-f673a5cbd988",
                    "xcp/sr/journal/clone/558220bc-a900-4408-a62e-a71a4bb4fd7b": "d9294359-c395-4bed-ac3a-bf4027c92bd9_0e18bf3d-78f0-4843-9e8f-ee11c6ebbf5a",
                    "xcp/sr/journal/clone/c41e0d47-5c1a-45c3-9404-01f3b5735c0d": "e191eb57-2478-4e3b-be9d-e8eaba8f9efe_41eae673-a280-439b-a4c6-f3afe2390fde",
                    "xcp/sr/journal/relink/50170fa2-2ca9-4218-8217-5c99ac31f10b": "1"
                  }
                  

                  So I basically deleted all of the keys here, Maybe I should not have done that, but when I did, eva plugged in correctly to the SR and I was able to finally destroying the SR from XOA. So yeah happy ending. Will try the next beta version. Thank you @ronan-a for your work.

                  ronan-aR 1 Reply Last reply Reply Quote 0
                  • TheiLLeniumStudiosT Offline
                    TheiLLeniumStudios @ronan-a
                    last edited by

                    @ronan-a I tried following the guide that you posted to remove the linstor volumes manually but the resource-definition list command already showed a bunch of resources in a "DELETING" state.

                    [22:24 xcp-ng-node-1 ~]# linstor --controllers=192.168.10.211 resource-definition list
                    ╭──────────────────────────────────────────────────────────────────────────────────────────────────────╮
                    ā”Š ResourceName                                    ā”Š Port ā”Š ResourceGroup                    ā”Š State    ā”Š
                    ā•žā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•ā•”
                    ā”Š xcp-persistent-database                         ā”Š 7000 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-persistent-ha-statefile                     ā”Š 7001 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-persistent-redo-log                         ā”Š 7002 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-volume-13a94a7a-d433-4426-8232-812e3c6dc52e ā”Š 7004 ā”Š xcp-sr-linstor_group_thin_device ā”Š DELETING ā”Š
                    ā”Š xcp-volume-4b70d69b-9cca-4aa3-842f-09366ac76901 ā”Š 7006 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-volume-50aa2e9f-caf0-4b0d-82f3-35893987e53b ā”Š 7010 ā”Š xcp-sr-linstor_group_thin_device ā”Š DELETING ā”Š
                    ā”Š xcp-volume-55c5c3fb-6782-46d6-8a81-f4a5f7cca691 ā”Š 7012 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-volume-5ebca692-6a61-47ec-8cac-e4fa0b6cc38a ā”Š 7016 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-volume-668bcb64-1150-43ac-baaa-db7b92331506 ā”Š 7014 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-volume-6f5235da-8f01-4057-a172-5e68bcb3f423 ā”Š 7007 ā”Š xcp-sr-linstor_group_thin_device ā”Š DELETING ā”Š
                    ā”Š xcp-volume-70bf80a2-a008-469a-a7db-0ea92fcfc392 ā”Š 7009 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-volume-92d4d363-ef03-4d3c-9d47-bef5cb1ca181 ā”Š 7015 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-volume-9a413b51-2625-407a-b05c-62bff025b947 ā”Š 7005 ā”Š xcp-sr-linstor_group_thin_device ā”Š ok       ā”Š
                    ā”Š xcp-volume-a02d160d-34fc-4fd6-957d-c7f3f9206ae2 ā”Š 7008 ā”Š xcp-sr-linstor_group_thin_device ā”Š DELETING ā”Š
                    ā”Š xcp-volume-ed04ffda-b379-4be7-8935-4f534f969a3f ā”Š 7003 ā”Š xcp-sr-linstor_group_thin_device ā”Š DELETING ā”Š
                    ╰──────────────────────────────────────────────────────────────────────────────────────────────────────╯
                    

                    Executing resource-definition delete has no impact on them. I just get the following output:

                    [22:24 xcp-ng-node-1 ~]# linstor resource-definition delete xcp-volume-13a94a7a-d433-4426-8232-812e3c6dc52e
                    SUCCESS:
                    Description:
                        Resource definition 'xcp-volume-13a94a7a-d433-4426-8232-812e3c6dc52e' marked for deletion.
                    Details:
                        Resource definition 'xcp-volume-13a94a7a-d433-4426-8232-812e3c6dc52e' UUID is: 52aceda9-b19b-461a-a119-f62931ba1af9
                    WARNING:
                    Description:
                        No active connection to satellite 'xcp-ng-node-3'
                    Details:
                        The controller is trying to (re-) establish a connection to the satellite. The controller stored the changes and as soon the satellite is connected, it will receive this update.
                    SUCCESS:
                        Resource 'xcp-volume-13a94a7a-d433-4426-8232-812e3c6dc52e' on 'xcp-ng-node-1' deleted
                    SUCCESS:
                        Resource 'xcp-volume-13a94a7a-d433-4426-8232-812e3c6dc52e' on 'xcp-ng-node-2' deleted
                    

                    I can confirm that node-1 can reach node-3 which it is complaining about for some reason. And I can see node-3 in XO as well and can run VMs on them.

                    ronan-aR 1 Reply Last reply Reply Quote 0
                    • ronan-aR Offline
                      ronan-a Vates 🪐 XCP-ng Team @TheiLLeniumStudios
                      last edited by

                      @TheiLLeniumStudios In this case, if DRBD is completely stuck, you can reboot your hosts. There is probably a lock or processes that have a lock on them. šŸ˜‰

                      TheiLLeniumStudiosT 1 Reply Last reply Reply Quote 0
                      • ronan-aR Offline
                        ronan-a Vates 🪐 XCP-ng Team @AudleyElwine
                        last edited by

                        @AudleyElwine Thank you for your feedbacks, I will update the script to handle the journal cases. šŸ™‚

                        1 Reply Last reply Reply Quote 0
                        • TheiLLeniumStudiosT Offline
                          TheiLLeniumStudios @ronan-a
                          last edited by

                          @ronan-a 2 of the nodes broke after restarting. I just kept getting the blinking cursor at the top left of the screen for hours. I'm going to have to reprovision all the nodes again sadly šŸ˜ž

                          ronan-aR 1 Reply Last reply Reply Quote 0
                          • A Offline
                            AudleyElwine @ronan-a
                            last edited by

                            Hey @ronan-a ,

                            What should I do to lower the chance of something in the past installation of xostor to affect my new installation?
                            lsblk is still showing the linstor volumes, vgs is also showing linstor_group.
                            Will a wipefs -af be enough? Or is the "Destroying SR" button in XOA is enough?

                            ronan-aR 1 Reply Last reply Reply Quote 0
                            • ronan-aR Offline
                              ronan-a Vates 🪐 XCP-ng Team @AudleyElwine
                              last edited by

                              @AudleyElwine The PVs/VGs are kept after a SR.destroy call but it's totally safe to reuse them for a new installation. The content of /var/lib/linstor is not removed after a destroy call, but it's normally not used because the linstor database is shared between hosts using a DRBD volume and mounted in this directory by the running controller. So you don't have manual steps to execute here. šŸ™‚

                              Of course if you want to reuse your disks for another thing, wipefs is nice for that.

                              1 Reply Last reply Reply Quote 0
                              • ronan-aR Offline
                                ronan-a Vates 🪐 XCP-ng Team @TheiLLeniumStudios
                                last edited by

                                @TheiLLeniumStudios There are always a solution to repair nodes, what's the output of linstor --controllers=<HOST_IPS> node list?

                                (Use a comma separated values for HOST_IPS.)

                                1 Reply Last reply Reply Quote 0
                                • Maelstrom96M Offline
                                  Maelstrom96
                                  last edited by

                                  We just hit a weird issue that we managed to fix, but wasn't really clear at first what was wrong, and might be a good idea for you guys to add some type of healthcheck / error handling to catch this and fix it.

                                  What happened was that for some unknown reason, our /var/lib/linstor mount (xcp-persistent-database) became read only, but everything kinda kept working-ish, but some stuff would randomly fail, like attempting to delete a ressource. Upon looking at the logs, saw this :

                                  Error message:                      The database is read only; SQL statement:
                                   UPDATE SEC_ACL_MAP SET ACCESS_TYPE = ?  WHERE OBJECT_PATH = ? AND        ROLE_NAME = ? [90097-197]
                                  

                                  We did a quick write test on the mount /var/lib/linstor and saw that it was indeed in RO mode. We also noticed that the last update time on the db file was 2 days ago.

                                  Unmounting the mount and remounting it had the controller start back again, but the first time, some nodes were missing from the node list, so we restarted the linstor-controller service again and everything is now up and healthy.

                                  ronan-aR 1 Reply Last reply Reply Quote 0
                                  • olivierlambertO Offline
                                    olivierlambert Vates 🪐 Co-Founder CEO
                                    last edited by

                                    I think we have a relatively important number of updates coming, fixing various potential bugs šŸ™‚ Stay tuned!

                                    SwenS 1 Reply Last reply Reply Quote 0
                                    • SwenS Offline
                                      Swen @olivierlambert
                                      last edited by

                                      @olivierlambert any eta for the next beta release? Looking forward to test it. šŸ™‚

                                      1 Reply Last reply Reply Quote 0
                                      • olivierlambertO Offline
                                        olivierlambert Vates 🪐 Co-Founder CEO
                                        last edited by

                                        Very soon©

                                        1 Reply Last reply Reply Quote 1
                                        • ronan-aR Offline
                                          ronan-a Vates 🪐 XCP-ng Team @Maelstrom96
                                          last edited by

                                          @Maelstrom96 said in XOSTOR hyperconvergence preview:

                                          What happened was that for some unknown reason, our /var/lib/linstor mount (xcp-persistent-database) became read only

                                          There is already a protection against that: https://github.com/xcp-ng/sm/commit/55779a64593df9407f861c3132ab85863b4f7e46 (2021-10-21)

                                          So I don't understand how it's possible to have a new time this issue. Without the log files I can't say what's the source of this issue, can you share them?

                                          Did you launch a controller manually before having this problem or not? There is a daemon to automatically mount and start a controller: minidrbdcluster. All actions related to the controllers must be executed by this program.

                                          Another idea, the problem can be related to: https://github.com/xcp-ng/sm/commit/a6385091370c6b358c7466944cc9b63f8c337c0d
                                          But this commit should be present in the last release.

                                          0 Wescoeur committed to xcp-ng/sm
                                          fix(var-lib-linstor.mount): ensure we always mount database with RW flags
                                          
                                          Sometimes systemd fallback to read only FS if the volume can't be mounted, we must
                                          forbid that. It's probably a DRBD error.
                                          
                                          Signed-off-by: Ronan Abhamon <ronan.abhamon@vates.fr>
                                          0 Wescoeur committed to xcp-ng/sm
                                          fix(minidrbdcluster): ensure SIGINT is handled correctly
                                          
                                          This patch is here to make sure no LINSTOR controller survives when
                                          systemd asks to minidrbdcluster to stop with `SIGINT`.
                                          
                                          - Remove `os.system`, it's totally unsafe, all signals are ignored with it.
                                          - Use `subprocess.Popen` instead and catch correctly signal exceptions, it works
                                            because `wait` call doesn't hide the signals.
                                          - Ensure `SIGINT` is only sent to the main process, not to the subprocesses.
                                          - Ensure `SIGKILL` is NEVER sent to minidrbdcluster.
                                          
                                          Signed-off-by: Ronan Abhamon <ronan.abhamon@vates.fr>
                                          Maelstrom96M 1 Reply Last reply Reply Quote 0
                                          • Maelstrom96M Offline
                                            Maelstrom96 @ronan-a
                                            last edited by Maelstrom96

                                            @ronan-a Is there a way to easily check if the process is managed by the daemon and not a manual start? We might have some point restarted the controller manually.

                                            Edit :

                                            ā— minidrbdcluster.service - Minimalistic high-availability cluster resource manager
                                               Loaded: loaded (/usr/lib/systemd/system/minidrbdcluster.service; enabled; vendor preset: disabled)
                                               Active: active (running) since Wed 2023-01-25 15:58:01 EST; 1 weeks 0 days ago
                                             Main PID: 2738 (python2)
                                               CGroup: /system.slice/minidrbdcluster.service
                                                       ā”œā”€2738 python2 /opt/xensource/libexec/minidrbdcluster
                                                       ā”œā”€2902 /usr/sbin/dmeventd
                                                       └─2939 drbdsetup events2
                                            
                                            [11:58 ovbh-pprod-xen10 system]# systemctl status var-lib-linstor.service
                                            ā— var-lib-linstor.service - Mount filesystem for the LINSTOR controller
                                               Loaded: loaded (/etc/systemd/system/var-lib-linstor.service; static; vendor p                                                                                                                                                                                                                                                                                                                                                           reset: disabled)
                                               Active: active (exited) since Wed 2023-01-25 15:58:03 EST; 1 weeks 0 days ago
                                              Process: 2947 ExecStart=/bin/mount -w /dev/drbd/by-res/xcp-persistent-database                                                                                                                                                                                                                                                                                                                                                           /0 /var/lib/linstor (code=exited, status=0/SUCCESS)
                                             Main PID: 2947 (code=exited, status=0/SUCCESS)
                                               CGroup: /system.slice/var-lib-linstor.service
                                            

                                            Also, what logs would you like to have?

                                            Edit2 : Also, I don't believe that service would've actually caught what happened, since it was mounted first using RW, but seems like DRBD had an issue while the mount was active and changed it to RO. The controller service was still healthy and active, just impacted on DB writes.

                                            ronan-aR 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post