XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    XOSTOR hyperconvergence preview

    Scheduled Pinned Locked Moved XOSTOR
    465 Posts 51 Posters 804.5k Views 54 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • J Offline
      JeffBerntsen Top contributor @dthenot
      last edited by

      @dthenot said in XOSTOR hyperconvergence preview:

      @JeffBerntsen That's why I meant, the way to install written in the first post still work in 8.3, the script still work as expected also, it basically only create the VG/LV needed on hosts before you create the SR.

      Got it. Thanks!

      henri9813H 1 Reply Last reply Reply Quote 0
      • henri9813H Offline
        henri9813 @JeffBerntsen
        last edited by henri9813

        Hello,

        I plan to install my XOSTOR cluster on a pool of 7 nodes with 3 replicas, but not all nodes at once because disks are in use.
        consider:

        • node1
        • node2
        • node ...
        • node 5
        • node 6
        • node 7.

        with 2 disks on each

        • sda: 128GB for the OS
        • sdb: 1TB for local sr ( for now 😄 )

        I emptied node 6 & 7.

        so, here is what i plan to do:

        • On ALL NODES: setup linstor packages

        Run the install script on node 6 & 7 to add their disks
        so:

        node6# install.sh --disks /dev/sdb
        node7# install.sh --disks /dev/sdb
        

        Then, configure the SR and the linstor plugin manager as the following

        xe sr-create \ 
            type=linstor name-label=pool-01 \
            host-uuid=XXXX \
            device-config:group-name=linstor_group/thin_device device-config:redundancy=3 shared=true device-config:provisioning=thin
        

        Normally, i should have a linstor cluster running of 2 nodes ( 2 satellite and one controller randomly placed ) with only 2 disks and then, only 2/3 working replicas.

        The cluster SHOULD be usable ( i'm right on this point ? )

        The next step, would be to move VM from node 5 on it to evacuate node 5. and then add it to the cluster by the following

        node5# install.sh --disks /dev/sdb
        node5# xe host-call-plugin \
          host-uuid=node5-uuid \
          plugin=linstor-manager \
          fn=addHost args:groupName=linstor_group/thin_device
        

        That should deploy satelite on node 5 and add the disk.

        I normally should have 3/3 working replicas and can start to deploy others nodes progressively.

        I'm right on the process ?

        aS mentionned in the discord, i will post my feedbacks and results from my setup once i finalized it. ( maybe thought a blog post somewhere ).

        Thanks to provide xostor in opensource, it's clearly the missing piece for this virtualization stack in opensource ( vs proxmox )

        1 Reply Last reply Reply Quote 0
        • henri9813H henri9813 referenced this topic on
        • J Offline
          Jonathon @Jonathon
          last edited by

          I have amazing news!

          After the upgrade to xcp-ng 8.3, I retested velero backup, and it all just works 😁

          Completed Backup

          jonathon@jonathon-framework:~$ velero --kubeconfig k8s_configs/production.yaml backup describe grafana-test
          Name:         grafana-test
          Namespace:    velero
          Labels:       objectset.rio.cattle.io/hash=c2b5f500ab5d9b8ffe14f2c70bf3742291df565c
                        velero.io/storage-location=default
          Annotations:  objectset.rio.cattle.io/applied=H4sIAAAAAAAA/4SSQW/bPgzFvwvPtv9OajeJj/8N22HdBqxFL0MPlEQlWmTRkOhgQ5HvPsixE2yH7iji8ffIJ74CDu6ZYnIcoIMTeYpcOf7vtIICji4Y6OB/1MdxgAJ6EjQoCN0rYAgsKI5Dyk9WP0hLIqmi40qjiKfMcRlAq7pBY+py26qmbEi15a5p78vtaqe0oqbVVsO5AI+K/Ju4A6YDdKDXqrVtXaNqzU5traVVY9d6Uyt7t2nW693K2Pa+naABe4IO9hEtBiyFksClmgbUdN06a9NAOtvr5B4DDunA8uR64lGgg7u6rxMUYMji6OWZ/dhTeuIPaQ6os+gTFUA/tR8NmXd+TELxUfNA5hslHqOmBN13OF16ZwvNQShIqpZClYQj7qk6blPlGF5uzC/L3P+kvok7MB9z0OcCXPiLPLHmuLLWCfVfB4rTZ9/iaA5zHovNZz7R++k6JI50q89BXcuXYR5YT0DolkChABEPHWzW9cK+rPQx8jgsH/KQj+QT/frzXCdduc/Ca9u1Y7aaFvMu5Ang5Xz+HQAA//8X7Fu+/QIAAA
                        objectset.rio.cattle.io/id=e104add0-85b4-4eb5-9456-819bcbe45cfc
                        velero.io/resource-timeout=10m0s
                        velero.io/source-cluster-k8s-gitversion=v1.33.4+rke2r1
                        velero.io/source-cluster-k8s-major-version=1
                        velero.io/source-cluster-k8s-minor-version=33
          
          Phase:  Completed
          
          
          Namespaces:
            Included:  grafana
            Excluded:  <none>
          
          Resources:
            Included cluster-scoped:    <none>
            Excluded cluster-scoped:    volumesnapshotcontents.snapshot.storage.k8s.io
            Included namespace-scoped:  *
            Excluded namespace-scoped:  volumesnapshots.snapshot.storage.k8s.io
          
          Label selector:  <none>
          
          Or label selector:  <none>
          
          Storage Location:  default
          
          Velero-Native Snapshot PVs:  true
          Snapshot Move Data:          true
          Data Mover:                  velero
          
          TTL:  720h0m0s
          
          CSISnapshotTimeout:    30m0s
          ItemOperationTimeout:  4h0m0s
          
          Hooks:  <none>
          
          Backup Format Version:  1.1.0
          
          Started:    2025-10-15 15:29:52 -0700 PDT
          Completed:  2025-10-15 15:31:25 -0700 PDT
          
          Expiration:  2025-11-14 14:29:52 -0800 PST
          
          Total items to be backed up:  35
          Items backed up:              35
          
          Backup Item Operations:  1 of 1 completed successfully, 0 failed (specify --details for more information)
          Backup Volumes:
            Velero-Native Snapshots: <none included>
          
            CSI Snapshots:
              grafana/central-grafana:
                Data Movement: included, specify --details for more information
          
            Pod Volume Backups: <none included>
          
          HooksAttempted:  0
          HooksFailed:     0
          

          Completed Restore

          jonathon@jonathon-framework:~$ velero --kubeconfig k8s_configs/production.yaml restore describe restore-grafana-test --details
          Name:         restore-grafana-test
          Namespace:    velero
          Labels:       objectset.rio.cattle.io/hash=252addb3ed156c52d9fa9b8c045b47a55d66c0af
          Annotations:  objectset.rio.cattle.io/applied=H4sIAAAAAAAA/3yRTW7zIBBA7zJrO5/j35gzfE2rtsomymIM45jGBgTjbKLcvaKJm6qL7kDwnt7ABdDpHfmgrQEBZxrJ25W2/85rSOCkjQIBrxTYeoIEJmJUyAjiAmiMZWRtTYhb232Q5EC88tquJDKPFEU6GlpUG5UVZdpUdZ6WZZ+niOtNWtR1SypvqC8buCYwYkfjn7oBwwAC8ipHpbqC1LqqZZWrtse228isrLqywapSdS0z7KPU4EQgwN+mSI8eezSYMgWG22lwKOl7/MgERzJmdChPs9veDL9IGfSbQRcGy+96IjszCCiyCRLQRo6zIrVd5AHEfuHhkIBmmp4d+a/3e9Dl8LPoCZ3T5hg7FvQRcR8nxt6XL7sAgv1MCZztOE+01P23cvmnPYzaxNtwuF4/AwAA//8k6OwC/QEAAA
                        objectset.rio.cattle.io/id=9ad8d034-7562-44f2-aa18-3669ed27ef47
          
          Phase:                       Completed
          Total items to be restored:  33
          Items restored:              33
          
          Started:    2025-10-15 15:35:26 -0700 PDT
          Completed:  2025-10-15 15:36:34 -0700 PDT
          
          Warnings:
            Velero:     <none>
            Cluster:    <none>
            Namespaces:
              grafana-restore:  could not restore, ConfigMap:elasticsearch-es-transport-ca-internal already exists. Warning: the in-cluster version is different than the backed-up version
                                could not restore, ConfigMap:kube-root-ca.crt already exists. Warning: the in-cluster version is different than the backed-up version
          
          Backup:  grafana-test
          
          Namespaces:
            Included:  grafana
            Excluded:  <none>
          
          Resources:
            Included:        *
            Excluded:        nodes, events, events.events.k8s.io, backups.velero.io, restores.velero.io, resticrepositories.velero.io, csinodes.storage.k8s.io, volumeattachments.storage.k8s.io, backuprepositories.velero.io
            Cluster-scoped:  auto
          
          Namespace mappings:  grafana=grafana-restore
          
          Label selector:  <none>
          
          Or label selector:  <none>
          
          Restore PVs:  true
          
          CSI Snapshot Restores:
            grafana-restore/central-grafana:
              Data Movement:
                Operation ID: dd-ffa56e1c-9fd0-44b4-a8bb-8163f40a49e9.330b82fc-ca6a-423217ee5
                Data Mover: velero
                Uploader Type: kopia
          
          Existing Resource Policy:   <none>
          ItemOperationTimeout:       4h0m0s
          
          Preserve Service NodePorts:  auto
          
          Restore Item Operations:
            Operation for persistentvolumeclaims grafana-restore/central-grafana:
              Restore Item Action Plugin:  velero.io/csi-pvc-restorer
              Operation ID:                dd-ffa56e1c-9fd0-44b4-a8bb-8163f40a49e9.330b82fc-ca6a-423217ee5
              Phase:                       Completed
              Progress:                    856284762 of 856284762 complete (Bytes)
              Progress description:        Completed
              Created:                     2025-10-15 15:35:28 -0700 PDT
              Started:                     2025-10-15 15:36:06 -0700 PDT
              Updated:                     2025-10-15 15:36:26 -0700 PDT
          
          HooksAttempted:   0
          HooksFailed:      0
          
          Resource List:
            apps/v1/Deployment:
              - grafana-restore/central-grafana(created)
              - grafana-restore/grafana-debug(created)
            apps/v1/ReplicaSet:
              - grafana-restore/central-grafana-5448b9f65(created)
              - grafana-restore/central-grafana-56887c6cb6(created)
              - grafana-restore/central-grafana-56ddd4f497(created)
              - grafana-restore/central-grafana-5f4757844b(created)
              - grafana-restore/central-grafana-5f69f86c85(created)
              - grafana-restore/central-grafana-64545dcdc(created)
              - grafana-restore/central-grafana-69c66c54d9(created)
              - grafana-restore/central-grafana-6c8d6f65b8(created)
              - grafana-restore/central-grafana-7b479f79ff(created)
              - grafana-restore/central-grafana-bc7d96cdd(created)
              - grafana-restore/central-grafana-cb88bd49c(created)
              - grafana-restore/grafana-debug-556845ff7b(created)
              - grafana-restore/grafana-debug-6fb594cb5f(created)
              - grafana-restore/grafana-debug-8f66bfbf6(created)
            discovery.k8s.io/v1/EndpointSlice:
              - grafana-restore/central-grafana-hkgd5(created)
            networking.k8s.io/v1/Ingress:
              - grafana-restore/central-grafana(created)
            rbac.authorization.k8s.io/v1/Role:
              - grafana-restore/central-grafana(created)
            rbac.authorization.k8s.io/v1/RoleBinding:
              - grafana-restore/central-grafana(created)
            v1/ConfigMap:
              - grafana-restore/central-grafana(created)
              - grafana-restore/elasticsearch-es-transport-ca-internal(failed)
              - grafana-restore/kube-root-ca.crt(failed)
            v1/Endpoints:
              - grafana-restore/central-grafana(created)
            v1/PersistentVolume:
              - pvc-e3f6578f-08b2-4e79-85f0-76bbf8985b55(skipped)
            v1/PersistentVolumeClaim:
              - grafana-restore/central-grafana(created)
            v1/Pod:
              - grafana-restore/central-grafana-cb88bd49c-fc5br(created)
            v1/Secret:
              - grafana-restore/fpinfra-net-cf-cert(created)
              - grafana-restore/grafana(created)
            v1/Service:
              - grafana-restore/central-grafana(created)
            v1/ServiceAccount:
              - grafana-restore/central-grafana(created)
              - grafana-restore/default(skipped)
            velero.io/v2alpha1/DataUpload:
              - velero/grafana-test-nw7zj(skipped)
          

          Image of working restore pod, with correct data in PV
          34d87db1-19ae-4348-8d4e-6599375d7634-image.png

          Velero installed from helm: https://vmware-tanzu.github.io/helm-charts
          Version: velero:11.1.0
          Values

          ---
          image:
            repository: velero/velero
            tag: v1.17.0
          
          # Whether to deploy the restic daemonset.
          deployNodeAgent: true
          
          initContainers:
             - name: velero-plugin-for-aws
               image: velero/velero-plugin-for-aws:latest
               imagePullPolicy: IfNotPresent
               volumeMounts:
                 - mountPath: /target
                   name: plugins
          
          configuration:
            defaultItemOperationTimeout: 2h
            features: EnableCSI
            defaultSnapshotMoveData: true
          
            backupStorageLocation:
              - name: default
                provider: aws
                bucket: velero
                config:
                  region: us-east-1
                  s3ForcePathStyle: true
                  s3Url: https://s3.location
          
            # Destination VSL points to LINSTOR snapshot class
            volumeSnapshotLocation:
              - name: linstor
                provider: velero.io/csi
                config:
                  snapshotClass: linstor-vsc
          
          credentials:
            useSecret: true
            existingSecret: velero-user
          
          
          metrics:
            enabled: true
          
            serviceMonitor:
              enabled: true
          
            prometheusRule:
              enabled: true
              # Additional labels to add to deployed PrometheusRule
              additionalLabels: {}
              # PrometheusRule namespace. Defaults to Velero namespace.
              # namespace: ""
              # Rules to be deployed
              spec:
                - alert: VeleroBackupPartialFailures
                  annotations:
                    message: Velero backup {{ $labels.schedule }} has {{ $value | humanizePercentage }} partialy failed backups.
                  expr: |-
                    velero_backup_partial_failure_total{schedule!=""} / velero_backup_attempt_total{schedule!=""} > 0.25
                  for: 15m
                  labels:
                    severity: warning
                - alert: VeleroBackupFailures
                  annotations:
                    message: Velero backup {{ $labels.schedule }} has {{ $value | humanizePercentage }} failed backups.
                  expr: |-
                    velero_backup_failure_total{schedule!=""} / velero_backup_attempt_total{schedule!=""} > 0.25
                  for: 15m
                  labels:
                    severity: warning
          

          Also create the following.

          apiVersion: snapshot.storage.k8s.io/v1
          kind: VolumeSnapshotClass
          metadata:
            name: linstor-vsc
            labels:
              velero.io/csi-volumesnapshot-class: "true"
          driver: linstor.csi.linbit.com
          deletionPolicy: Delete
          

          We are using Piraeus operator to use xostor in k8s
          https://github.com/piraeusdatastore/piraeus-operator.git
          Version: v2.9.1
          Values:

          ---
          operator: 
            resources:
              requests:
                cpu: 250m
                memory: 500Mi
              limits:
                memory: 1Gi
          installCRDs: true
          imageConfigOverride:
          - base: quay.io/piraeusdatastore
            components:
              linstor-satellite:
                image: piraeus-server
                tag: v1.29.0
          tls:
            certManagerIssuerRef:
              name: step-issuer
              kind: StepClusterIssuer
              group: certmanager.step.sm
          

          Then we just connect to the xostor cluster like external linstor controller.

          henri9813H 1 Reply Last reply Reply Quote 1
          • henri9813H Offline
            henri9813 @Jonathon
            last edited by henri9813

            Hello,

            I got my whole xostor destroyed, i don't know how precisely.

            I found some errors in sattelite

            Error context:
                    An error occurred while processing resource 'Node: 'host', Rsc: 'xcp-volume-e011c043-8751-45e6-be06-4ce9f8807cad''
            ErrorContext:
              Details:     Command 'lvcreate --config 'devices { filter=['"'"'a|/dev/md127|'"'"','"'"'a|/dev/md126p3|'"'"','"'"'r|.*|'"'"'] }' --virtualsize 52543488k linstor_primary --thinpool thin_device --name xcp-volume-e011c043-8751-45e6-be06-4ce9f8807cad_00000' returned with exitcode 5. 
            
            Standard out: 
            
            
            Error message: 
              WARNING: Remaining free space in metadata of thin pool linstor_primary/thin_device is too low (98.06% >= 96.30%). Resize is recommended.
              Cannot create new thin volume, free space in thin pool linstor_primary/thin_device reached threshold.
            

            of course, i checked, my SR was not full
            aa2774a4-c2d4-4dd1-be52-3c6e418c9083-image.png

            And the controller crashed, and i couldn't make it works.

            Here is the error i got

            ==========
            
            Category:                           RuntimeException
            Class name:                         IllegalStateException
            Class canonical name:               java.lang.IllegalStateException
            Generated at:                       Method 'newIllegalStateException', Source file 'DataUtils.java', Line #870
            
            Error message:                      Reading from nio:/var/lib/linstor/linstordb.mv.db failed; file length 2293760 read length 384 at 2445540 [1.4.197/1]
            

            So i deduce the database was fucked-up, i tried to open the file as explained in the documentation, but the linstor schema was "not found" in the file, event if using cat i see data about it.

            for now, i leave xostor and i'm back to localstorage until we know what to do when this issue occured with a "solution path".

            ronan-aR 1 Reply Last reply Reply Quote 0
            • ronan-aR Offline
              ronan-a Vates 🪐 XCP-ng Team @henri9813
              last edited by

              @henri9813 said in XOSTOR hyperconvergence preview:

              of course, i checked, my SR was not full

              The visual representation of used space is for informational purposes only; it's an approximation that takes into account replication, disks in use, etc. For more information: https://docs.xcp-ng.org/xostor/#how-a-linstor-sr-capacity-is-calculated

              We plan to display a complete view of each physical disk space on each host someday to provide a more detailed overview. In any case, if you use "lvs"/"vgs" on each machine, you should indeed see the actual disk space used.

              henri9813H 1 Reply Last reply Reply Quote 0
              • henri9813H Offline
                henri9813 @ronan-a
                last edited by

                Hello @ronan-a ,

                but how recover from this situation ?

                Thanks !

                1 Reply Last reply Reply Quote 0
                • snk33S Offline
                  snk33 @ronan-a
                  last edited by

                  @ronan-a said in XOSTOR hyperconvergence preview:

                  @peter_webbird We've already had feedback on CBT and LINSTOR/DRBD, we don't necessarily recommend enabling it. We have a blocking dev card regarding a bug with LVM lvchange command that may fail on CBT volumes used by a XOSTOR SR. We also have other issues related to migration with CBT.

                  Is the problem still occurring on latest XCP-ng / XOSTOR version ? Not being able to use CBT on XOSTOR is a big issue for backup/replication.

                  snk33S 1 Reply Last reply Reply Quote 0
                  • snk33S Offline
                    snk33 @snk33
                    last edited by

                    Any update from someone ? Not having the ability to enable CBT on XOSTOR is really complicated to make replication / DR scenarios.

                    At least to know if this is work in progress or not. If it's not we'll need to pause our PoC and post-pone production schedule waiting for significant update.

                    ronan-aR 1 Reply Last reply Reply Quote 0
                    • ronan-aR Offline
                      ronan-a Vates 🪐 XCP-ng Team @snk33
                      last edited by

                      @snk33 In fact we have limitations with CBT and XOSTOR. And this won't be resolved until we get back into SMAPIv3.

                      Not being able to use CBT on XOSTOR is a big issue for backup/replication.

                      Using it is more likely to cause problems than improve your situation. What bothers you about not using CBT here?

                      snk33S 1 Reply Last reply Reply Quote 1
                      • snk33S Offline
                        snk33 @ronan-a
                        last edited by

                        @ronan-a we need to have CBT to use 3rd party backup & replication solution such as Veeam.

                        On the other hand, XOSTOR with VHD disk might be ok using Xen Orchestra backup & replication features but it needs more testing on our end to make sure it works good enough.
                        One big question is about the snapshot chain. Coming from VMware world where keeping snapshots may impact I/O performances, we're not very comfortable with it.

                        Also, we need to find solution for low-RPO replication. On our VMware infrastructure we use Zerto DR for this but XOA being more like Veeam using snapshots, low-RPO will be tough to achieve (unless making snapshots doesn't affect performances at all ?).

                        1 Reply Last reply Reply Quote 0
                        • First post
                          Last post