XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Our future backup code: test it!

    Scheduled Pinned Locked Moved Backup
    128 Posts 10 Posters 8.0k Views 11 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • florentF Offline
      florent Vates 🪐 XO Team
      last edited by

      thanks for your tests . I will recheck NBD backup

      florentF 1 Reply Last reply Reply Quote 0
      • florentF Offline
        florent Vates 🪐 XO Team @florent
        last edited by

        ih @Tristis-Oris and @Davidj-0 I pushed an update to better handle NBD error, please keep us informed

        Tristis OrisT 1 Reply Last reply Reply Quote 0
        • Tristis OrisT Offline
          Tristis Oris Top contributor @florent
          last edited by Tristis Oris

          @florent i'm not sure how to update from this branch(

          git checkout .
          git pull --ff-only
          yarn
          yarn build
          
          Updating 04be60321..6641580e6
          Fast-forward
           @xen-orchestra/xapi/disks/Xapi.mjs             | 17 ++++++++++-------
           @xen-orchestra/xapi/disks/XapiVhdCbt.mjs       |  3 ++-
           @xen-orchestra/xapi/disks/XapiVhdStreamNbd.mjs |  3 ---
           @xen-orchestra/xapi/disks/utils.mjs            | 29 ++++++++++++++++-------------
           4 files changed, 28 insertions(+), 24 deletions(-)
          
          git checkout feat_generator_backups
          Already on 'feat_generator_backups'
          Your branch is up to date with 'origin/feat_generator_backups'.
          
          
          git pull origin feat_generator_backups
          From https://github.com/vatesfr/xen-orchestra
           * branch                feat_generator_backups -> FETCH_HEAD
          Already up to date.
          

          but still no NBD at backup.

          1 Reply Last reply Reply Quote 0
          • Tristis OrisT Offline
            Tristis Oris Top contributor
            last edited by

            i see significant speed difference from master branch.

            prod
            NBD
            160.71 MiB
            no NBD
            169

            test
            NBD (broken)
            110-130 MiB
            no NBD
            115

            florentF 1 Reply Last reply Reply Quote 0
            • florentF Offline
              florent Vates 🪐 XO Team @Tristis Oris
              last edited by

              @Tristis-Oris ouch that is quite costly
              Can you describe which backup you run ?
              Can you check if the duration is different (maybe this is a measurement error and not a slow speed) ?

              Tristis OrisT 1 Reply Last reply Reply Quote 0
              • Tristis OrisT Offline
                Tristis Oris Top contributor @florent
                last edited by

                @florent

                1 vm, 1 storage, NBD connections: 1. delta, first full.

                Duration: 3 minutes
                Size: 26.54 GiB
                Speed: 160.71 MiB/s

                Duration: 4 minutes
                Size: 26.53 GiB
                Speed: 113.74 MiB/s

                florentF 1 Reply Last reply Reply Quote 1
                • florentF Offline
                  florent Vates 🪐 XO Team @Tristis Oris
                  last edited by

                  @Tristis-Oris Do you have the same performance without NBD ?
                  Does your storage use blocks ?

                  Tristis OrisT 1 Reply Last reply Reply Quote 0
                  • Tristis OrisT Offline
                    Tristis Oris Top contributor @florent
                    last edited by

                    @florent blocks! i forgot about that checkbox)

                    better but not same.

                    Duration: 3 minutes
                    Size: 26.53 GiB
                    Speed: 143.7 MiB/s

                    Speed: 146.09 MiB/s

                    florentF 1 Reply Last reply Reply Quote 0
                    • florentF Offline
                      florent Vates 🪐 XO Team @Tristis Oris
                      last edited by

                      @Tristis-Oris I made a little change, can you update (like the last time ) and retest ?

                      Tristis OrisT 1 Reply Last reply Reply Quote 0
                      • Tristis OrisT Offline
                        Tristis Oris Top contributor @florent
                        last edited by

                        @florent
                        same speed after fix.
                        142.56 MiB/s - 145.63 MiB/s

                        maybe i miss something else? same database as prod, only another LUN connected for backups.

                        florentF 1 Reply Last reply Reply Quote 0
                        • florentF Offline
                          florent Vates 🪐 XO Team @Tristis Oris
                          last edited by florent

                          @Tristis-Oris no it's on our end

                          Could you retry nbd + target a block based directory ?
                          ON my test setup, with the latest changes I get better speed than master ( 190MB/s per disk vs 130-170 depending on the run and settings on master)

                          I got quite a huge variation between the same runs (40MB/s)

                          Tristis OrisT 1 Reply Last reply Reply Quote 0
                          • Tristis OrisT Offline
                            Tristis Oris Top contributor @florent
                            last edited by

                            @florent yep, now it equal. Maybe my hw bottleneck? i can also check with ssd storage to see max speed.

                            Duration: 3 minutes
                            Size: 26.53 GiB
                            Speed: 157.78 MiB/s
                            Speed: 149.39 MiB/s
                            Speed: 163.76 MiB/s

                            No more errors incorrect backup size in metadata.

                            But still no NBD(

                            florentF 1 Reply Last reply Reply Quote 0
                            • florentF Offline
                              florent Vates 🪐 XO Team @Tristis Oris
                              last edited by

                              @Tristis-Oris that is already a good news.

                              I pushed an additional fix : the NBD info was not shown on the UI
                              79c6d837-2b12-4b66-b66d-59f887c088a5-image.png

                              Tristis OrisT 1 Reply Last reply Reply Quote 0
                              • Tristis OrisT Offline
                                Tristis Oris Top contributor @florent
                                last edited by

                                @florent haha, now it works too good)

                                60174b41-7179-4e31-8d02-bf0c9ec405a6-изображение.png

                                39e6a5fa-942f-4895-933d-51bd765ddde5-изображение.png

                                983f9e40-8f8b-417e-8360-85d195aeba2b-изображение.png

                                Tristis OrisT 1 Reply Last reply Reply Quote 0
                                • Tristis OrisT Offline
                                  Tristis Oris Top contributor
                                  last edited by

                                  well, that was my CPU bottleneck. XO live at most stable DC, but oldest one.

                                  • Intel(R) Xeon(R) CPU E5-2690 v4 @ 2.60GHz

                                  flash:
                                  Speed: 151.36 MiB/s
                                  summary: { duration: '3m', cpuUsage: '131%', memoryUsage: '162.19 MiB' }
                                  hdd:
                                  Speed: 152 MiB/s
                                  summary: { duration: '3m', cpuUsage: '201%', memoryUsage: '314.1 MiB' }

                                  • Intel(R) Xeon(R) Gold 5215 CPU @ 2.50GHz

                                  flash:
                                  Speed: 196.78 MiB/s
                                  summary: { duration: '3m', cpuUsage: '129%', memoryUsage: '170.8 MiB' }
                                  hdd:
                                  Speed: 184.72 MiB/s
                                  summary: { duration: '3m', cpuUsage: '198%', memoryUsage: '321.06 MiB' }

                                  • Intel(R) Xeon(R) Platinum 8260 CPU @ 2.40GHz

                                  flash:
                                  Speed: 222.32 MiB/s
                                  Speed: 220 MiB/s
                                  summary: { duration: '2m', cpuUsage: '155%', memoryUsage: '183.77 MiB' }

                                  hdd:
                                  Speed: 185.63 MiB/s
                                  Speed: 185.21 MiB/s
                                  summary: { duration: '3m', cpuUsage: '196%', memoryUsage: '315.87 MiB' }

                                  Look at high memory usage with hdd.

                                  sometimes i still got errors.

                                            "id": "1744875242122:0",
                                            "message": "export",
                                            "start": 1744875242122,
                                            "status": "success",
                                            "tasks": [
                                              {
                                                "id": "1744875245258",
                                                "message": "transfer",
                                                "start": 1744875245258,
                                                "status": "success",
                                                "end": 1744875430762,
                                                "result": {
                                                  "size": 28489809920
                                                }
                                              },
                                              {
                                                "id": "1744875432586",
                                                "message": "clean-vm",
                                                "start": 1744875432586,
                                                "status": "success",
                                                "warnings": [
                                                  {
                                                    "data": {
                                                      "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T073405Z.json",
                                                      "actual": 28489809920,
                                                      "expected": 28496828928
                                                    },
                                                    "message": "cleanVm: incorrect backup size in metadata"
                                                  }
                                  
                                            "id": "1744876967012:0",
                                            "message": "export",
                                            "start": 1744876967012,
                                            "status": "success",
                                            "tasks": [
                                              {
                                                "id": "1744876970075",
                                                "message": "transfer",
                                                "start": 1744876970075,
                                                "status": "success",
                                                "end": 1744877108146,
                                                "result": {
                                                  "size": 28489809920
                                                }
                                              },
                                              {
                                                "id": "1744877119430",
                                                "message": "clean-vm",
                                                "start": 1744877119430,
                                                "status": "success",
                                                "warnings": [
                                                  {
                                                    "data": {
                                                      "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T080250Z.json",
                                                      "actual": 28489809920,
                                                      "expected": 28496828928
                                                    },
                                                    "message": "cleanVm: incorrect backup size in metadata"
                                                  }
                                  
                                  1 Reply Last reply Reply Quote 0
                                  • Tristis OrisT Offline
                                    Tristis Oris Top contributor
                                    last edited by

                                    i tried to move tests to another vm, but again can't build it with same commands(

                                    yarn start
                                    yarn run v1.22.22
                                    $ node dist/cli.mjs
                                    node:internal/modules/esm/resolve:275
                                        throw new ERR_MODULE_NOT_FOUND(
                                              ^
                                    
                                    Error [ERR_MODULE_NOT_FOUND]: Cannot find module '/opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs' imported from /opt/xen-orchestra/@xen-orchestra/xapi/disks/Xapi.mjs
                                        at finalizeResolution (node:internal/modules/esm/resolve:275:11)
                                        at moduleResolve (node:internal/modules/esm/resolve:860:10)
                                        at defaultResolve (node:internal/modules/esm/resolve:984:11)
                                        at ModuleLoader.defaultResolve (node:internal/modules/esm/loader:685:12)
                                        at #cachedDefaultResolve (node:internal/modules/esm/loader:634:25)
                                        at ModuleLoader.resolve (node:internal/modules/esm/loader:617:38)
                                        at ModuleLoader.getModuleJobForImport (node:internal/modules/esm/loader:273:38)
                                        at ModuleJob._link (node:internal/modules/esm/module_job:135:49) {
                                      code: 'ERR_MODULE_NOT_FOUND',
                                      url: 'file:///opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs'
                                    }
                                    
                                    Node.js v22.14.0
                                    error Command failed with exit code 1.
                                    info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
                                    
                                    florentF 1 Reply Last reply Reply Quote 0
                                    • florentF Offline
                                      florent Vates 🪐 XO Team @Tristis Oris
                                      last edited by

                                      @Tristis-Oris thanks , I missed a file
                                      I pushed it just now

                                      A 1 Reply Last reply Reply Quote 1
                                      • A Offline
                                        Andrew Top contributor @florent
                                        last edited by

                                        @florent I finally got the new code running and I tested a Delta Backup (full first run) with NBD x3 enabled and it's leaving NBD transfer (on xcp1) 99% connected after a run. The backup does complete but the task is stuck.

                                        florentF 1 Reply Last reply Reply Quote 1
                                        • florentF Offline
                                          florent Vates 🪐 XO Team @Andrew
                                          last edited by

                                          @Andrew nice catch andrew I will look into it
                                          is it keeping disk attached to dom0 ? (in dashboard -> health )

                                          A 1 Reply Last reply Reply Quote 0
                                          • A Offline
                                            Andrew Top contributor @florent
                                            last edited by

                                            @florent No. The dashboard health is clean. No VDIs attached to control domain

                                            1 Reply Last reply Reply Quote 1
                                            • First post
                                              Last post