XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Our future backup code: test it!

    Scheduled Pinned Locked Moved Backup
    128 Posts 10 Posters 49.2k Views 11 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Tristis OrisT Offline
      Tristis Oris Top contributor
      last edited by

      delta, first full.

      • no NBD
      2025-04-11T12:19:15.169Z xo:backups:worker INFO starting backup
      2025-04-11T12:22:29.509Z xo:backups:MixinBackupWriter WARN cleanVm: incorrect backup size in metadata {
        path: '/xo-vm-backups/07e5d245-7ed3-b690-e386-3461556ad41f/20250411T121931Z.json',
        actual: 27271364608,
        expected: 27278188544
      }
      

      2 vms but 1 log at cli.

      delta

      [2025-04-11T12:27:41.428Z xo:backups:worker INFO starting backup
      2025-04-11T12:27:48.644Z xo:backups:MixinBackupWriter WARN cleanVm: incorrect backup size in metadata {
        path: '/xo-vm-backups/f3cfdc19-0e6a-04d1-5651-82e1dbb25191/20250411T122746Z.json',
        actual: 44040192,
        expected: 46314496
      }
      2025-04-11T12:28:03.055Z xo:backups:MixinBackupWriter WARN cleanVm: incorrect backup size in metadata {
        path: '/xo-vm-backups/07e5d245-7ed3-b690-e386-3461556ad41f/20250411T122801Z.json',
        actual: 46137344,
        expected: 48412160
      }
      2025-04-11T12:28:03.081Z xo:backups:worker INFO backup has ended
      2025-04-11T12:28:03.181Z xo:backups:worker INFO process will exit {
        duration: 21752192,
        exitCode: 0,
        resourceUsage: {
          userCPUTime: 3794057,
          systemCPUTime: 468414,
          maxRSS: 69592,
          sharedMemorySize: 0,
          unsharedDataSize: 0,
          unsharedStackSize: 0,
          minorPageFault: 42277,
          majorPageFault: 0,
          swappedOut: 0,
          fsRead: 13624,
          fsWrite: 185840,
          ipcSent: 0,
          ipcReceived: 0,
          signalsCount: 0,
          voluntaryContextSwitches: 3591,
          involuntaryContextSwitches: 441
        },
        summary: { duration: '22s', cpuUsage: '20%', memoryUsage: '67.96 MiB' }
      }
      

      can't get NBD to work. It enabled on nic, but not used for backups. XO settings copied from prod instance.

      1 Reply Last reply Reply Quote 0
      • D Offline
        Davidj 0
        last edited by

        Like @Tristis-Oris , my first tests were done using a clone of my production XO. They failed with "No NBD" errors.

        Creating a new backup works.

        Tristis OrisT 1 Reply Last reply Reply Quote 0
        • Tristis OrisT Offline
          Tristis Oris Top contributor @Davidj 0
          last edited by

          @Davidj-0 No NBD usage even with new task( Main XO is on same host, same nic.

          D 1 Reply Last reply Reply Quote 0
          • D Offline
            Davidj 0 @Tristis Oris
            last edited by

            @Tristis-Oris My setup is similar to yours, and I get similar errors on any existing backup job.

            However, if I create a new backup job, then it works without any error.

            ping @florent , maybe these data points are useful.

            1 Reply Last reply Reply Quote 0
            • florentF Online
              florent Vates πŸͺ XO Team
              last edited by

              thanks for your tests . I will recheck NBD backup

              florentF 1 Reply Last reply Reply Quote 0
              • florentF Online
                florent Vates πŸͺ XO Team @florent
                last edited by

                ih @Tristis-Oris and @Davidj-0 I pushed an update to better handle NBD error, please keep us informed

                Tristis OrisT 1 Reply Last reply Reply Quote 0
                • Tristis OrisT Offline
                  Tristis Oris Top contributor @florent
                  last edited by Tristis Oris

                  @florent i'm not sure how to update from this branch(

                  git checkout .
                  git pull --ff-only
                  yarn
                  yarn build
                  
                  Updating 04be60321..6641580e6
                  Fast-forward
                   @xen-orchestra/xapi/disks/Xapi.mjs             | 17 ++++++++++-------
                   @xen-orchestra/xapi/disks/XapiVhdCbt.mjs       |  3 ++-
                   @xen-orchestra/xapi/disks/XapiVhdStreamNbd.mjs |  3 ---
                   @xen-orchestra/xapi/disks/utils.mjs            | 29 ++++++++++++++++-------------
                   4 files changed, 28 insertions(+), 24 deletions(-)
                  
                  git checkout feat_generator_backups
                  Already on 'feat_generator_backups'
                  Your branch is up to date with 'origin/feat_generator_backups'.
                  
                  
                  git pull origin feat_generator_backups
                  From https://github.com/vatesfr/xen-orchestra
                   * branch                feat_generator_backups -> FETCH_HEAD
                  Already up to date.
                  

                  but still no NBD at backup.

                  1 Reply Last reply Reply Quote 0
                  • Tristis OrisT Offline
                    Tristis Oris Top contributor
                    last edited by

                    i see significant speed difference from master branch.

                    prod
                    NBD
                    160.71 MiB
                    no NBD
                    169

                    test
                    NBD (broken)
                    110-130 MiB
                    no NBD
                    115

                    florentF 1 Reply Last reply Reply Quote 0
                    • florentF Online
                      florent Vates πŸͺ XO Team @Tristis Oris
                      last edited by

                      @Tristis-Oris ouch that is quite costly
                      Can you describe which backup you run ?
                      Can you check if the duration is different (maybe this is a measurement error and not a slow speed) ?

                      Tristis OrisT 1 Reply Last reply Reply Quote 0
                      • Tristis OrisT Offline
                        Tristis Oris Top contributor @florent
                        last edited by

                        @florent

                        1 vm, 1 storage, NBD connections: 1. delta, first full.

                        Duration: 3 minutes
                        Size: 26.54 GiB
                        Speed: 160.71 MiB/s

                        Duration: 4 minutes
                        Size: 26.53 GiB
                        Speed: 113.74 MiB/s

                        florentF 1 Reply Last reply Reply Quote 1
                        • florentF Online
                          florent Vates πŸͺ XO Team @Tristis Oris
                          last edited by

                          @Tristis-Oris Do you have the same performance without NBD ?
                          Does your storage use blocks ?

                          Tristis OrisT 1 Reply Last reply Reply Quote 0
                          • Tristis OrisT Offline
                            Tristis Oris Top contributor @florent
                            last edited by

                            @florent blocks! i forgot about that checkbox)

                            better but not same.

                            Duration: 3 minutes
                            Size: 26.53 GiB
                            Speed: 143.7 MiB/s

                            Speed: 146.09 MiB/s

                            florentF 1 Reply Last reply Reply Quote 0
                            • florentF Online
                              florent Vates πŸͺ XO Team @Tristis Oris
                              last edited by

                              @Tristis-Oris I made a little change, can you update (like the last time ) and retest ?

                              Tristis OrisT 1 Reply Last reply Reply Quote 0
                              • Tristis OrisT Offline
                                Tristis Oris Top contributor @florent
                                last edited by

                                @florent
                                same speed after fix.
                                142.56 MiB/s - 145.63 MiB/s

                                maybe i miss something else? same database as prod, only another LUN connected for backups.

                                florentF 1 Reply Last reply Reply Quote 0
                                • florentF Online
                                  florent Vates πŸͺ XO Team @Tristis Oris
                                  last edited by florent

                                  @Tristis-Oris no it's on our end

                                  Could you retry nbd + target a block based directory ?
                                  ON my test setup, with the latest changes I get better speed than master ( 190MB/s per disk vs 130-170 depending on the run and settings on master)

                                  I got quite a huge variation between the same runs (40MB/s)

                                  Tristis OrisT 1 Reply Last reply Reply Quote 0
                                  • Tristis OrisT Offline
                                    Tristis Oris Top contributor @florent
                                    last edited by

                                    @florent yep, now it equal. Maybe my hw bottleneck? i can also check with ssd storage to see max speed.

                                    Duration: 3 minutes
                                    Size: 26.53 GiB
                                    Speed: 157.78 MiB/s
                                    Speed: 149.39 MiB/s
                                    Speed: 163.76 MiB/s

                                    No more errors incorrect backup size in metadata.

                                    But still no NBD(

                                    florentF 1 Reply Last reply Reply Quote 0
                                    • florentF Online
                                      florent Vates πŸͺ XO Team @Tristis Oris
                                      last edited by

                                      @Tristis-Oris that is already a good news.

                                      I pushed an additional fix : the NBD info was not shown on the UI
                                      79c6d837-2b12-4b66-b66d-59f887c088a5-image.png

                                      Tristis OrisT 1 Reply Last reply Reply Quote 0
                                      • Tristis OrisT Offline
                                        Tristis Oris Top contributor @florent
                                        last edited by

                                        @florent haha, now it works too good)

                                        60174b41-7179-4e31-8d02-bf0c9ec405a6-ΠΈΠ·ΠΎΠ±Ρ€Π°ΠΆΠ΅Π½ΠΈΠ΅.png

                                        39e6a5fa-942f-4895-933d-51bd765ddde5-ΠΈΠ·ΠΎΠ±Ρ€Π°ΠΆΠ΅Π½ΠΈΠ΅.png

                                        983f9e40-8f8b-417e-8360-85d195aeba2b-ΠΈΠ·ΠΎΠ±Ρ€Π°ΠΆΠ΅Π½ΠΈΠ΅.png

                                        Tristis OrisT 1 Reply Last reply Reply Quote 0
                                        • Tristis OrisT Offline
                                          Tristis Oris Top contributor
                                          last edited by

                                          well, that was my CPU bottleneck. XO live at most stable DC, but oldest one.

                                          • Intel(R) Xeon(R) CPU E5-2690 v4 @ 2.60GHz

                                          flash:
                                          Speed: 151.36 MiB/s
                                          summary: { duration: '3m', cpuUsage: '131%', memoryUsage: '162.19 MiB' }
                                          hdd:
                                          Speed: 152 MiB/s
                                          summary: { duration: '3m', cpuUsage: '201%', memoryUsage: '314.1 MiB' }

                                          • Intel(R) Xeon(R) Gold 5215 CPU @ 2.50GHz

                                          flash:
                                          Speed: 196.78 MiB/s
                                          summary: { duration: '3m', cpuUsage: '129%', memoryUsage: '170.8 MiB' }
                                          hdd:
                                          Speed: 184.72 MiB/s
                                          summary: { duration: '3m', cpuUsage: '198%', memoryUsage: '321.06 MiB' }

                                          • Intel(R) Xeon(R) Platinum 8260 CPU @ 2.40GHz

                                          flash:
                                          Speed: 222.32 MiB/s
                                          Speed: 220 MiB/s
                                          summary: { duration: '2m', cpuUsage: '155%', memoryUsage: '183.77 MiB' }

                                          hdd:
                                          Speed: 185.63 MiB/s
                                          Speed: 185.21 MiB/s
                                          summary: { duration: '3m', cpuUsage: '196%', memoryUsage: '315.87 MiB' }

                                          Look at high memory usage with hdd.

                                          sometimes i still got errors.

                                                    "id": "1744875242122:0",
                                                    "message": "export",
                                                    "start": 1744875242122,
                                                    "status": "success",
                                                    "tasks": [
                                                      {
                                                        "id": "1744875245258",
                                                        "message": "transfer",
                                                        "start": 1744875245258,
                                                        "status": "success",
                                                        "end": 1744875430762,
                                                        "result": {
                                                          "size": 28489809920
                                                        }
                                                      },
                                                      {
                                                        "id": "1744875432586",
                                                        "message": "clean-vm",
                                                        "start": 1744875432586,
                                                        "status": "success",
                                                        "warnings": [
                                                          {
                                                            "data": {
                                                              "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T073405Z.json",
                                                              "actual": 28489809920,
                                                              "expected": 28496828928
                                                            },
                                                            "message": "cleanVm: incorrect backup size in metadata"
                                                          }
                                          
                                                    "id": "1744876967012:0",
                                                    "message": "export",
                                                    "start": 1744876967012,
                                                    "status": "success",
                                                    "tasks": [
                                                      {
                                                        "id": "1744876970075",
                                                        "message": "transfer",
                                                        "start": 1744876970075,
                                                        "status": "success",
                                                        "end": 1744877108146,
                                                        "result": {
                                                          "size": 28489809920
                                                        }
                                                      },
                                                      {
                                                        "id": "1744877119430",
                                                        "message": "clean-vm",
                                                        "start": 1744877119430,
                                                        "status": "success",
                                                        "warnings": [
                                                          {
                                                            "data": {
                                                              "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T080250Z.json",
                                                              "actual": 28489809920,
                                                              "expected": 28496828928
                                                            },
                                                            "message": "cleanVm: incorrect backup size in metadata"
                                                          }
                                          
                                          1 Reply Last reply Reply Quote 0
                                          • Tristis OrisT Offline
                                            Tristis Oris Top contributor
                                            last edited by

                                            i tried to move tests to another vm, but again can't build it with same commands(

                                            yarn start
                                            yarn run v1.22.22
                                            $ node dist/cli.mjs
                                            node:internal/modules/esm/resolve:275
                                                throw new ERR_MODULE_NOT_FOUND(
                                                      ^
                                            
                                            Error [ERR_MODULE_NOT_FOUND]: Cannot find module '/opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs' imported from /opt/xen-orchestra/@xen-orchestra/xapi/disks/Xapi.mjs
                                                at finalizeResolution (node:internal/modules/esm/resolve:275:11)
                                                at moduleResolve (node:internal/modules/esm/resolve:860:10)
                                                at defaultResolve (node:internal/modules/esm/resolve:984:11)
                                                at ModuleLoader.defaultResolve (node:internal/modules/esm/loader:685:12)
                                                at #cachedDefaultResolve (node:internal/modules/esm/loader:634:25)
                                                at ModuleLoader.resolve (node:internal/modules/esm/loader:617:38)
                                                at ModuleLoader.getModuleJobForImport (node:internal/modules/esm/loader:273:38)
                                                at ModuleJob._link (node:internal/modules/esm/module_job:135:49) {
                                              code: 'ERR_MODULE_NOT_FOUND',
                                              url: 'file:///opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs'
                                            }
                                            
                                            Node.js v22.14.0
                                            error Command failed with exit code 1.
                                            info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
                                            
                                            florentF 1 Reply Last reply Reply Quote 0

                                            Hello! It looks like you're interested in this conversation, but you don't have an account yet.

                                            Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.

                                            With your input, this post could be even better πŸ’—

                                            Register Login
                                            • First post
                                              Last post