XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Our future backup code: test it!

    Scheduled Pinned Locked Moved Backup
    60 Posts 9 Posters 2.2k Views 9 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Tristis OrisT Offline
      Tristis Oris Top contributor @florent
      last edited by Tristis Oris

      @florent i'm not sure how to update from this branch(

      git checkout .
      git pull --ff-only
      yarn
      yarn build
      
      Updating 04be60321..6641580e6
      Fast-forward
       @xen-orchestra/xapi/disks/Xapi.mjs             | 17 ++++++++++-------
       @xen-orchestra/xapi/disks/XapiVhdCbt.mjs       |  3 ++-
       @xen-orchestra/xapi/disks/XapiVhdStreamNbd.mjs |  3 ---
       @xen-orchestra/xapi/disks/utils.mjs            | 29 ++++++++++++++++-------------
       4 files changed, 28 insertions(+), 24 deletions(-)
      
      git checkout feat_generator_backups
      Already on 'feat_generator_backups'
      Your branch is up to date with 'origin/feat_generator_backups'.
      
      
      git pull origin feat_generator_backups
      From https://github.com/vatesfr/xen-orchestra
       * branch                feat_generator_backups -> FETCH_HEAD
      Already up to date.
      

      but still no NBD at backup.

      1 Reply Last reply Reply Quote 0
      • Tristis OrisT Offline
        Tristis Oris Top contributor
        last edited by

        i see significant speed difference from master branch.

        prod
        NBD
        160.71 MiB
        no NBD
        169

        test
        NBD (broken)
        110-130 MiB
        no NBD
        115

        florentF 1 Reply Last reply Reply Quote 0
        • florentF Offline
          florent Vates 🪐 XO Team @Tristis Oris
          last edited by

          @Tristis-Oris ouch that is quite costly
          Can you describe which backup you run ?
          Can you check if the duration is different (maybe this is a measurement error and not a slow speed) ?

          Tristis OrisT 1 Reply Last reply Reply Quote 0
          • Tristis OrisT Offline
            Tristis Oris Top contributor @florent
            last edited by

            @florent

            1 vm, 1 storage, NBD connections: 1. delta, first full.

            Duration: 3 minutes
            Size: 26.54 GiB
            Speed: 160.71 MiB/s

            Duration: 4 minutes
            Size: 26.53 GiB
            Speed: 113.74 MiB/s

            florentF 1 Reply Last reply Reply Quote 1
            • florentF Offline
              florent Vates 🪐 XO Team @Tristis Oris
              last edited by

              @Tristis-Oris Do you have the same performance without NBD ?
              Does your storage use blocks ?

              Tristis OrisT 1 Reply Last reply Reply Quote 0
              • Tristis OrisT Offline
                Tristis Oris Top contributor @florent
                last edited by

                @florent blocks! i forgot about that checkbox)

                better but not same.

                Duration: 3 minutes
                Size: 26.53 GiB
                Speed: 143.7 MiB/s

                Speed: 146.09 MiB/s

                florentF 1 Reply Last reply Reply Quote 0
                • florentF Offline
                  florent Vates 🪐 XO Team @Tristis Oris
                  last edited by

                  @Tristis-Oris I made a little change, can you update (like the last time ) and retest ?

                  Tristis OrisT 1 Reply Last reply Reply Quote 0
                  • Tristis OrisT Offline
                    Tristis Oris Top contributor @florent
                    last edited by

                    @florent
                    same speed after fix.
                    142.56 MiB/s - 145.63 MiB/s

                    maybe i miss something else? same database as prod, only another LUN connected for backups.

                    florentF 1 Reply Last reply Reply Quote 0
                    • florentF Offline
                      florent Vates 🪐 XO Team @Tristis Oris
                      last edited by florent

                      @Tristis-Oris no it's on our end

                      Could you retry nbd + target a block based directory ?
                      ON my test setup, with the latest changes I get better speed than master ( 190MB/s per disk vs 130-170 depending on the run and settings on master)

                      I got quite a huge variation between the same runs (40MB/s)

                      Tristis OrisT 1 Reply Last reply Reply Quote 0
                      • Tristis OrisT Offline
                        Tristis Oris Top contributor @florent
                        last edited by

                        @florent yep, now it equal. Maybe my hw bottleneck? i can also check with ssd storage to see max speed.

                        Duration: 3 minutes
                        Size: 26.53 GiB
                        Speed: 157.78 MiB/s
                        Speed: 149.39 MiB/s
                        Speed: 163.76 MiB/s

                        No more errors incorrect backup size in metadata.

                        But still no NBD(

                        florentF 1 Reply Last reply Reply Quote 0
                        • florentF Offline
                          florent Vates 🪐 XO Team @Tristis Oris
                          last edited by

                          @Tristis-Oris that is already a good news.

                          I pushed an additional fix : the NBD info was not shown on the UI
                          79c6d837-2b12-4b66-b66d-59f887c088a5-image.png

                          Tristis OrisT 1 Reply Last reply Reply Quote 0
                          • Tristis OrisT Offline
                            Tristis Oris Top contributor @florent
                            last edited by

                            @florent haha, now it works too good)

                            60174b41-7179-4e31-8d02-bf0c9ec405a6-изображение.png

                            39e6a5fa-942f-4895-933d-51bd765ddde5-изображение.png

                            983f9e40-8f8b-417e-8360-85d195aeba2b-изображение.png

                            1 Reply Last reply Reply Quote 0
                            • Tristis OrisT Offline
                              Tristis Oris Top contributor
                              last edited by

                              well, that was my CPU bottleneck. XO live at most stable DC, but oldest one.

                              • Intel(R) Xeon(R) CPU E5-2690 v4 @ 2.60GHz

                              flash:
                              Speed: 151.36 MiB/s
                              summary: { duration: '3m', cpuUsage: '131%', memoryUsage: '162.19 MiB' }
                              hdd:
                              Speed: 152 MiB/s
                              summary: { duration: '3m', cpuUsage: '201%', memoryUsage: '314.1 MiB' }

                              • Intel(R) Xeon(R) Gold 5215 CPU @ 2.50GHz

                              flash:
                              Speed: 196.78 MiB/s
                              summary: { duration: '3m', cpuUsage: '129%', memoryUsage: '170.8 MiB' }
                              hdd:
                              Speed: 184.72 MiB/s
                              summary: { duration: '3m', cpuUsage: '198%', memoryUsage: '321.06 MiB' }

                              • Intel(R) Xeon(R) Platinum 8260 CPU @ 2.40GHz

                              flash:
                              Speed: 222.32 MiB/s
                              Speed: 220 MiB/s
                              summary: { duration: '2m', cpuUsage: '155%', memoryUsage: '183.77 MiB' }

                              hdd:
                              Speed: 185.63 MiB/s
                              Speed: 185.21 MiB/s
                              summary: { duration: '3m', cpuUsage: '196%', memoryUsage: '315.87 MiB' }

                              Look at high memory usage with hdd.

                              sometimes i still got errors.

                                        "id": "1744875242122:0",
                                        "message": "export",
                                        "start": 1744875242122,
                                        "status": "success",
                                        "tasks": [
                                          {
                                            "id": "1744875245258",
                                            "message": "transfer",
                                            "start": 1744875245258,
                                            "status": "success",
                                            "end": 1744875430762,
                                            "result": {
                                              "size": 28489809920
                                            }
                                          },
                                          {
                                            "id": "1744875432586",
                                            "message": "clean-vm",
                                            "start": 1744875432586,
                                            "status": "success",
                                            "warnings": [
                                              {
                                                "data": {
                                                  "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T073405Z.json",
                                                  "actual": 28489809920,
                                                  "expected": 28496828928
                                                },
                                                "message": "cleanVm: incorrect backup size in metadata"
                                              }
                              
                                        "id": "1744876967012:0",
                                        "message": "export",
                                        "start": 1744876967012,
                                        "status": "success",
                                        "tasks": [
                                          {
                                            "id": "1744876970075",
                                            "message": "transfer",
                                            "start": 1744876970075,
                                            "status": "success",
                                            "end": 1744877108146,
                                            "result": {
                                              "size": 28489809920
                                            }
                                          },
                                          {
                                            "id": "1744877119430",
                                            "message": "clean-vm",
                                            "start": 1744877119430,
                                            "status": "success",
                                            "warnings": [
                                              {
                                                "data": {
                                                  "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T080250Z.json",
                                                  "actual": 28489809920,
                                                  "expected": 28496828928
                                                },
                                                "message": "cleanVm: incorrect backup size in metadata"
                                              }
                              
                              1 Reply Last reply Reply Quote 0
                              • Tristis OrisT Offline
                                Tristis Oris Top contributor
                                last edited by

                                i tried to move tests to another vm, but again can't build it with same commands(

                                yarn start
                                yarn run v1.22.22
                                $ node dist/cli.mjs
                                node:internal/modules/esm/resolve:275
                                    throw new ERR_MODULE_NOT_FOUND(
                                          ^
                                
                                Error [ERR_MODULE_NOT_FOUND]: Cannot find module '/opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs' imported from /opt/xen-orchestra/@xen-orchestra/xapi/disks/Xapi.mjs
                                    at finalizeResolution (node:internal/modules/esm/resolve:275:11)
                                    at moduleResolve (node:internal/modules/esm/resolve:860:10)
                                    at defaultResolve (node:internal/modules/esm/resolve:984:11)
                                    at ModuleLoader.defaultResolve (node:internal/modules/esm/loader:685:12)
                                    at #cachedDefaultResolve (node:internal/modules/esm/loader:634:25)
                                    at ModuleLoader.resolve (node:internal/modules/esm/loader:617:38)
                                    at ModuleLoader.getModuleJobForImport (node:internal/modules/esm/loader:273:38)
                                    at ModuleJob._link (node:internal/modules/esm/module_job:135:49) {
                                  code: 'ERR_MODULE_NOT_FOUND',
                                  url: 'file:///opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs'
                                }
                                
                                Node.js v22.14.0
                                error Command failed with exit code 1.
                                info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
                                
                                florentF 1 Reply Last reply Reply Quote 0
                                • florentF Offline
                                  florent Vates 🪐 XO Team @Tristis Oris
                                  last edited by

                                  @Tristis-Oris thanks , I missed a file
                                  I pushed it just now

                                  A 1 Reply Last reply Reply Quote 1
                                  • A Offline
                                    Andrew Top contributor @florent
                                    last edited by

                                    @florent I finally got the new code running and I tested a Delta Backup (full first run) with NBD x3 enabled and it's leaving NBD transfer (on xcp1) 99% connected after a run. The backup does complete but the task is stuck.

                                    florentF 1 Reply Last reply Reply Quote 1
                                    • florentF Offline
                                      florent Vates 🪐 XO Team @Andrew
                                      last edited by

                                      @Andrew nice catch andrew I will look into it
                                      is it keeping disk attached to dom0 ? (in dashboard -> health )

                                      A 1 Reply Last reply Reply Quote 0
                                      • A Offline
                                        Andrew Top contributor @florent
                                        last edited by

                                        @florent No. The dashboard health is clean. No VDIs attached to control domain

                                        1 Reply Last reply Reply Quote 1
                                        • florentF Offline
                                          florent Vates 🪐 XO Team
                                          last edited by

                                          so that is probably only a off by one error in the task code
                                          Thanks andrew

                                          1 Reply Last reply Reply Quote 0
                                          • First post
                                            Last post