XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Our future backup code: test it!

    Scheduled Pinned Locked Moved Backup
    128 Posts 10 Posters 10.7k Views 11 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Tristis OrisT Offline
      Tristis Oris Top contributor
      last edited by

      delta, first full.

      • no NBD
      2025-04-11T12:19:15.169Z xo:backups:worker INFO starting backup
      2025-04-11T12:22:29.509Z xo:backups:MixinBackupWriter WARN cleanVm: incorrect backup size in metadata {
        path: '/xo-vm-backups/07e5d245-7ed3-b690-e386-3461556ad41f/20250411T121931Z.json',
        actual: 27271364608,
        expected: 27278188544
      }
      

      2 vms but 1 log at cli.

      delta

      [2025-04-11T12:27:41.428Z xo:backups:worker INFO starting backup
      2025-04-11T12:27:48.644Z xo:backups:MixinBackupWriter WARN cleanVm: incorrect backup size in metadata {
        path: '/xo-vm-backups/f3cfdc19-0e6a-04d1-5651-82e1dbb25191/20250411T122746Z.json',
        actual: 44040192,
        expected: 46314496
      }
      2025-04-11T12:28:03.055Z xo:backups:MixinBackupWriter WARN cleanVm: incorrect backup size in metadata {
        path: '/xo-vm-backups/07e5d245-7ed3-b690-e386-3461556ad41f/20250411T122801Z.json',
        actual: 46137344,
        expected: 48412160
      }
      2025-04-11T12:28:03.081Z xo:backups:worker INFO backup has ended
      2025-04-11T12:28:03.181Z xo:backups:worker INFO process will exit {
        duration: 21752192,
        exitCode: 0,
        resourceUsage: {
          userCPUTime: 3794057,
          systemCPUTime: 468414,
          maxRSS: 69592,
          sharedMemorySize: 0,
          unsharedDataSize: 0,
          unsharedStackSize: 0,
          minorPageFault: 42277,
          majorPageFault: 0,
          swappedOut: 0,
          fsRead: 13624,
          fsWrite: 185840,
          ipcSent: 0,
          ipcReceived: 0,
          signalsCount: 0,
          voluntaryContextSwitches: 3591,
          involuntaryContextSwitches: 441
        },
        summary: { duration: '22s', cpuUsage: '20%', memoryUsage: '67.96 MiB' }
      }
      

      can't get NBD to work. It enabled on nic, but not used for backups. XO settings copied from prod instance.

      1 Reply Last reply Reply Quote 0
      • D Offline
        Davidj 0
        last edited by

        Like @Tristis-Oris , my first tests were done using a clone of my production XO. They failed with "No NBD" errors.

        Creating a new backup works.

        Tristis OrisT 1 Reply Last reply Reply Quote 0
        • Tristis OrisT Offline
          Tristis Oris Top contributor @Davidj 0
          last edited by

          @Davidj-0 No NBD usage even with new task( Main XO is on same host, same nic.

          D 1 Reply Last reply Reply Quote 0
          • D Offline
            Davidj 0 @Tristis Oris
            last edited by

            @Tristis-Oris My setup is similar to yours, and I get similar errors on any existing backup job.

            However, if I create a new backup job, then it works without any error.

            ping @florent , maybe these data points are useful.

            1 Reply Last reply Reply Quote 0
            • florentF Offline
              florent Vates πŸͺ XO Team
              last edited by

              thanks for your tests . I will recheck NBD backup

              florentF 1 Reply Last reply Reply Quote 0
              • florentF Offline
                florent Vates πŸͺ XO Team @florent
                last edited by

                ih @Tristis-Oris and @Davidj-0 I pushed an update to better handle NBD error, please keep us informed

                Tristis OrisT 1 Reply Last reply Reply Quote 0
                • Tristis OrisT Offline
                  Tristis Oris Top contributor @florent
                  last edited by Tristis Oris

                  @florent i'm not sure how to update from this branch(

                  git checkout .
                  git pull --ff-only
                  yarn
                  yarn build
                  
                  Updating 04be60321..6641580e6
                  Fast-forward
                   @xen-orchestra/xapi/disks/Xapi.mjs             | 17 ++++++++++-------
                   @xen-orchestra/xapi/disks/XapiVhdCbt.mjs       |  3 ++-
                   @xen-orchestra/xapi/disks/XapiVhdStreamNbd.mjs |  3 ---
                   @xen-orchestra/xapi/disks/utils.mjs            | 29 ++++++++++++++++-------------
                   4 files changed, 28 insertions(+), 24 deletions(-)
                  
                  git checkout feat_generator_backups
                  Already on 'feat_generator_backups'
                  Your branch is up to date with 'origin/feat_generator_backups'.
                  
                  
                  git pull origin feat_generator_backups
                  From https://github.com/vatesfr/xen-orchestra
                   * branch                feat_generator_backups -> FETCH_HEAD
                  Already up to date.
                  

                  but still no NBD at backup.

                  1 Reply Last reply Reply Quote 0
                  • Tristis OrisT Offline
                    Tristis Oris Top contributor
                    last edited by

                    i see significant speed difference from master branch.

                    prod
                    NBD
                    160.71 MiB
                    no NBD
                    169

                    test
                    NBD (broken)
                    110-130 MiB
                    no NBD
                    115

                    florentF 1 Reply Last reply Reply Quote 0
                    • florentF Offline
                      florent Vates πŸͺ XO Team @Tristis Oris
                      last edited by

                      @Tristis-Oris ouch that is quite costly
                      Can you describe which backup you run ?
                      Can you check if the duration is different (maybe this is a measurement error and not a slow speed) ?

                      Tristis OrisT 1 Reply Last reply Reply Quote 0
                      • Tristis OrisT Offline
                        Tristis Oris Top contributor @florent
                        last edited by

                        @florent

                        1 vm, 1 storage, NBD connections: 1. delta, first full.

                        Duration: 3 minutes
                        Size: 26.54 GiB
                        Speed: 160.71 MiB/s

                        Duration: 4 minutes
                        Size: 26.53 GiB
                        Speed: 113.74 MiB/s

                        florentF 1 Reply Last reply Reply Quote 1
                        • florentF Offline
                          florent Vates πŸͺ XO Team @Tristis Oris
                          last edited by

                          @Tristis-Oris Do you have the same performance without NBD ?
                          Does your storage use blocks ?

                          Tristis OrisT 1 Reply Last reply Reply Quote 0
                          • Tristis OrisT Offline
                            Tristis Oris Top contributor @florent
                            last edited by

                            @florent blocks! i forgot about that checkbox)

                            better but not same.

                            Duration: 3 minutes
                            Size: 26.53 GiB
                            Speed: 143.7 MiB/s

                            Speed: 146.09 MiB/s

                            florentF 1 Reply Last reply Reply Quote 0
                            • florentF Offline
                              florent Vates πŸͺ XO Team @Tristis Oris
                              last edited by

                              @Tristis-Oris I made a little change, can you update (like the last time ) and retest ?

                              Tristis OrisT 1 Reply Last reply Reply Quote 0
                              • Tristis OrisT Offline
                                Tristis Oris Top contributor @florent
                                last edited by

                                @florent
                                same speed after fix.
                                142.56 MiB/s - 145.63 MiB/s

                                maybe i miss something else? same database as prod, only another LUN connected for backups.

                                florentF 1 Reply Last reply Reply Quote 0
                                • florentF Offline
                                  florent Vates πŸͺ XO Team @Tristis Oris
                                  last edited by florent

                                  @Tristis-Oris no it's on our end

                                  Could you retry nbd + target a block based directory ?
                                  ON my test setup, with the latest changes I get better speed than master ( 190MB/s per disk vs 130-170 depending on the run and settings on master)

                                  I got quite a huge variation between the same runs (40MB/s)

                                  Tristis OrisT 1 Reply Last reply Reply Quote 0
                                  • Tristis OrisT Offline
                                    Tristis Oris Top contributor @florent
                                    last edited by

                                    @florent yep, now it equal. Maybe my hw bottleneck? i can also check with ssd storage to see max speed.

                                    Duration: 3 minutes
                                    Size: 26.53 GiB
                                    Speed: 157.78 MiB/s
                                    Speed: 149.39 MiB/s
                                    Speed: 163.76 MiB/s

                                    No more errors incorrect backup size in metadata.

                                    But still no NBD(

                                    florentF 1 Reply Last reply Reply Quote 0
                                    • florentF Offline
                                      florent Vates πŸͺ XO Team @Tristis Oris
                                      last edited by

                                      @Tristis-Oris that is already a good news.

                                      I pushed an additional fix : the NBD info was not shown on the UI
                                      79c6d837-2b12-4b66-b66d-59f887c088a5-image.png

                                      Tristis OrisT 1 Reply Last reply Reply Quote 0
                                      • Tristis OrisT Offline
                                        Tristis Oris Top contributor @florent
                                        last edited by

                                        @florent haha, now it works too good)

                                        60174b41-7179-4e31-8d02-bf0c9ec405a6-ΠΈΠ·ΠΎΠ±Ρ€Π°ΠΆΠ΅Π½ΠΈΠ΅.png

                                        39e6a5fa-942f-4895-933d-51bd765ddde5-ΠΈΠ·ΠΎΠ±Ρ€Π°ΠΆΠ΅Π½ΠΈΠ΅.png

                                        983f9e40-8f8b-417e-8360-85d195aeba2b-ΠΈΠ·ΠΎΠ±Ρ€Π°ΠΆΠ΅Π½ΠΈΠ΅.png

                                        Tristis OrisT 1 Reply Last reply Reply Quote 0
                                        • Tristis OrisT Offline
                                          Tristis Oris Top contributor
                                          last edited by

                                          well, that was my CPU bottleneck. XO live at most stable DC, but oldest one.

                                          • Intel(R) Xeon(R) CPU E5-2690 v4 @ 2.60GHz

                                          flash:
                                          Speed: 151.36 MiB/s
                                          summary: { duration: '3m', cpuUsage: '131%', memoryUsage: '162.19 MiB' }
                                          hdd:
                                          Speed: 152 MiB/s
                                          summary: { duration: '3m', cpuUsage: '201%', memoryUsage: '314.1 MiB' }

                                          • Intel(R) Xeon(R) Gold 5215 CPU @ 2.50GHz

                                          flash:
                                          Speed: 196.78 MiB/s
                                          summary: { duration: '3m', cpuUsage: '129%', memoryUsage: '170.8 MiB' }
                                          hdd:
                                          Speed: 184.72 MiB/s
                                          summary: { duration: '3m', cpuUsage: '198%', memoryUsage: '321.06 MiB' }

                                          • Intel(R) Xeon(R) Platinum 8260 CPU @ 2.40GHz

                                          flash:
                                          Speed: 222.32 MiB/s
                                          Speed: 220 MiB/s
                                          summary: { duration: '2m', cpuUsage: '155%', memoryUsage: '183.77 MiB' }

                                          hdd:
                                          Speed: 185.63 MiB/s
                                          Speed: 185.21 MiB/s
                                          summary: { duration: '3m', cpuUsage: '196%', memoryUsage: '315.87 MiB' }

                                          Look at high memory usage with hdd.

                                          sometimes i still got errors.

                                                    "id": "1744875242122:0",
                                                    "message": "export",
                                                    "start": 1744875242122,
                                                    "status": "success",
                                                    "tasks": [
                                                      {
                                                        "id": "1744875245258",
                                                        "message": "transfer",
                                                        "start": 1744875245258,
                                                        "status": "success",
                                                        "end": 1744875430762,
                                                        "result": {
                                                          "size": 28489809920
                                                        }
                                                      },
                                                      {
                                                        "id": "1744875432586",
                                                        "message": "clean-vm",
                                                        "start": 1744875432586,
                                                        "status": "success",
                                                        "warnings": [
                                                          {
                                                            "data": {
                                                              "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T073405Z.json",
                                                              "actual": 28489809920,
                                                              "expected": 28496828928
                                                            },
                                                            "message": "cleanVm: incorrect backup size in metadata"
                                                          }
                                          
                                                    "id": "1744876967012:0",
                                                    "message": "export",
                                                    "start": 1744876967012,
                                                    "status": "success",
                                                    "tasks": [
                                                      {
                                                        "id": "1744876970075",
                                                        "message": "transfer",
                                                        "start": 1744876970075,
                                                        "status": "success",
                                                        "end": 1744877108146,
                                                        "result": {
                                                          "size": 28489809920
                                                        }
                                                      },
                                                      {
                                                        "id": "1744877119430",
                                                        "message": "clean-vm",
                                                        "start": 1744877119430,
                                                        "status": "success",
                                                        "warnings": [
                                                          {
                                                            "data": {
                                                              "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T080250Z.json",
                                                              "actual": 28489809920,
                                                              "expected": 28496828928
                                                            },
                                                            "message": "cleanVm: incorrect backup size in metadata"
                                                          }
                                          
                                          1 Reply Last reply Reply Quote 0
                                          • Tristis OrisT Offline
                                            Tristis Oris Top contributor
                                            last edited by

                                            i tried to move tests to another vm, but again can't build it with same commands(

                                            yarn start
                                            yarn run v1.22.22
                                            $ node dist/cli.mjs
                                            node:internal/modules/esm/resolve:275
                                                throw new ERR_MODULE_NOT_FOUND(
                                                      ^
                                            
                                            Error [ERR_MODULE_NOT_FOUND]: Cannot find module '/opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs' imported from /opt/xen-orchestra/@xen-orchestra/xapi/disks/Xapi.mjs
                                                at finalizeResolution (node:internal/modules/esm/resolve:275:11)
                                                at moduleResolve (node:internal/modules/esm/resolve:860:10)
                                                at defaultResolve (node:internal/modules/esm/resolve:984:11)
                                                at ModuleLoader.defaultResolve (node:internal/modules/esm/loader:685:12)
                                                at #cachedDefaultResolve (node:internal/modules/esm/loader:634:25)
                                                at ModuleLoader.resolve (node:internal/modules/esm/loader:617:38)
                                                at ModuleLoader.getModuleJobForImport (node:internal/modules/esm/loader:273:38)
                                                at ModuleJob._link (node:internal/modules/esm/module_job:135:49) {
                                              code: 'ERR_MODULE_NOT_FOUND',
                                              url: 'file:///opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs'
                                            }
                                            
                                            Node.js v22.14.0
                                            error Command failed with exit code 1.
                                            info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
                                            
                                            florentF 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post