XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Our future backup code: test it!

    Scheduled Pinned Locked Moved Backup
    60 Posts 9 Posters 2.2k Views 9 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Tristis OrisT Offline
      Tristis Oris Top contributor @florent
      last edited by

      @florent

      1 vm, 1 storage, NBD connections: 1. delta, first full.

      Duration: 3 minutes
      Size: 26.54 GiB
      Speed: 160.71 MiB/s

      Duration: 4 minutes
      Size: 26.53 GiB
      Speed: 113.74 MiB/s

      florentF 1 Reply Last reply Reply Quote 1
      • florentF Offline
        florent Vates 🪐 XO Team @Tristis Oris
        last edited by

        @Tristis-Oris Do you have the same performance without NBD ?
        Does your storage use blocks ?

        Tristis OrisT 1 Reply Last reply Reply Quote 0
        • Tristis OrisT Offline
          Tristis Oris Top contributor @florent
          last edited by

          @florent blocks! i forgot about that checkbox)

          better but not same.

          Duration: 3 minutes
          Size: 26.53 GiB
          Speed: 143.7 MiB/s

          Speed: 146.09 MiB/s

          florentF 1 Reply Last reply Reply Quote 0
          • florentF Offline
            florent Vates 🪐 XO Team @Tristis Oris
            last edited by

            @Tristis-Oris I made a little change, can you update (like the last time ) and retest ?

            Tristis OrisT 1 Reply Last reply Reply Quote 0
            • Tristis OrisT Offline
              Tristis Oris Top contributor @florent
              last edited by

              @florent
              same speed after fix.
              142.56 MiB/s - 145.63 MiB/s

              maybe i miss something else? same database as prod, only another LUN connected for backups.

              florentF 1 Reply Last reply Reply Quote 0
              • florentF Offline
                florent Vates 🪐 XO Team @Tristis Oris
                last edited by florent

                @Tristis-Oris no it's on our end

                Could you retry nbd + target a block based directory ?
                ON my test setup, with the latest changes I get better speed than master ( 190MB/s per disk vs 130-170 depending on the run and settings on master)

                I got quite a huge variation between the same runs (40MB/s)

                Tristis OrisT 1 Reply Last reply Reply Quote 0
                • Tristis OrisT Offline
                  Tristis Oris Top contributor @florent
                  last edited by

                  @florent yep, now it equal. Maybe my hw bottleneck? i can also check with ssd storage to see max speed.

                  Duration: 3 minutes
                  Size: 26.53 GiB
                  Speed: 157.78 MiB/s
                  Speed: 149.39 MiB/s
                  Speed: 163.76 MiB/s

                  No more errors incorrect backup size in metadata.

                  But still no NBD(

                  florentF 1 Reply Last reply Reply Quote 0
                  • florentF Offline
                    florent Vates 🪐 XO Team @Tristis Oris
                    last edited by

                    @Tristis-Oris that is already a good news.

                    I pushed an additional fix : the NBD info was not shown on the UI
                    79c6d837-2b12-4b66-b66d-59f887c088a5-image.png

                    Tristis OrisT 1 Reply Last reply Reply Quote 0
                    • Tristis OrisT Offline
                      Tristis Oris Top contributor @florent
                      last edited by

                      @florent haha, now it works too good)

                      60174b41-7179-4e31-8d02-bf0c9ec405a6-изображение.png

                      39e6a5fa-942f-4895-933d-51bd765ddde5-изображение.png

                      983f9e40-8f8b-417e-8360-85d195aeba2b-изображение.png

                      1 Reply Last reply Reply Quote 0
                      • Tristis OrisT Offline
                        Tristis Oris Top contributor
                        last edited by

                        well, that was my CPU bottleneck. XO live at most stable DC, but oldest one.

                        • Intel(R) Xeon(R) CPU E5-2690 v4 @ 2.60GHz

                        flash:
                        Speed: 151.36 MiB/s
                        summary: { duration: '3m', cpuUsage: '131%', memoryUsage: '162.19 MiB' }
                        hdd:
                        Speed: 152 MiB/s
                        summary: { duration: '3m', cpuUsage: '201%', memoryUsage: '314.1 MiB' }

                        • Intel(R) Xeon(R) Gold 5215 CPU @ 2.50GHz

                        flash:
                        Speed: 196.78 MiB/s
                        summary: { duration: '3m', cpuUsage: '129%', memoryUsage: '170.8 MiB' }
                        hdd:
                        Speed: 184.72 MiB/s
                        summary: { duration: '3m', cpuUsage: '198%', memoryUsage: '321.06 MiB' }

                        • Intel(R) Xeon(R) Platinum 8260 CPU @ 2.40GHz

                        flash:
                        Speed: 222.32 MiB/s
                        Speed: 220 MiB/s
                        summary: { duration: '2m', cpuUsage: '155%', memoryUsage: '183.77 MiB' }

                        hdd:
                        Speed: 185.63 MiB/s
                        Speed: 185.21 MiB/s
                        summary: { duration: '3m', cpuUsage: '196%', memoryUsage: '315.87 MiB' }

                        Look at high memory usage with hdd.

                        sometimes i still got errors.

                                  "id": "1744875242122:0",
                                  "message": "export",
                                  "start": 1744875242122,
                                  "status": "success",
                                  "tasks": [
                                    {
                                      "id": "1744875245258",
                                      "message": "transfer",
                                      "start": 1744875245258,
                                      "status": "success",
                                      "end": 1744875430762,
                                      "result": {
                                        "size": 28489809920
                                      }
                                    },
                                    {
                                      "id": "1744875432586",
                                      "message": "clean-vm",
                                      "start": 1744875432586,
                                      "status": "success",
                                      "warnings": [
                                        {
                                          "data": {
                                            "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T073405Z.json",
                                            "actual": 28489809920,
                                            "expected": 28496828928
                                          },
                                          "message": "cleanVm: incorrect backup size in metadata"
                                        }
                        
                                  "id": "1744876967012:0",
                                  "message": "export",
                                  "start": 1744876967012,
                                  "status": "success",
                                  "tasks": [
                                    {
                                      "id": "1744876970075",
                                      "message": "transfer",
                                      "start": 1744876970075,
                                      "status": "success",
                                      "end": 1744877108146,
                                      "result": {
                                        "size": 28489809920
                                      }
                                    },
                                    {
                                      "id": "1744877119430",
                                      "message": "clean-vm",
                                      "start": 1744877119430,
                                      "status": "success",
                                      "warnings": [
                                        {
                                          "data": {
                                            "path": "/xo-vm-backups/d4950e88-f6aa-dbc1-e6fe-e3c73ebe9904/20250417T080250Z.json",
                                            "actual": 28489809920,
                                            "expected": 28496828928
                                          },
                                          "message": "cleanVm: incorrect backup size in metadata"
                                        }
                        
                        1 Reply Last reply Reply Quote 0
                        • Tristis OrisT Offline
                          Tristis Oris Top contributor
                          last edited by

                          i tried to move tests to another vm, but again can't build it with same commands(

                          yarn start
                          yarn run v1.22.22
                          $ node dist/cli.mjs
                          node:internal/modules/esm/resolve:275
                              throw new ERR_MODULE_NOT_FOUND(
                                    ^
                          
                          Error [ERR_MODULE_NOT_FOUND]: Cannot find module '/opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs' imported from /opt/xen-orchestra/@xen-orchestra/xapi/disks/Xapi.mjs
                              at finalizeResolution (node:internal/modules/esm/resolve:275:11)
                              at moduleResolve (node:internal/modules/esm/resolve:860:10)
                              at defaultResolve (node:internal/modules/esm/resolve:984:11)
                              at ModuleLoader.defaultResolve (node:internal/modules/esm/loader:685:12)
                              at #cachedDefaultResolve (node:internal/modules/esm/loader:634:25)
                              at ModuleLoader.resolve (node:internal/modules/esm/loader:617:38)
                              at ModuleLoader.getModuleJobForImport (node:internal/modules/esm/loader:273:38)
                              at ModuleJob._link (node:internal/modules/esm/module_job:135:49) {
                            code: 'ERR_MODULE_NOT_FOUND',
                            url: 'file:///opt/xen-orchestra/@xen-orchestra/xapi/disks/XapiProgress.mjs'
                          }
                          
                          Node.js v22.14.0
                          error Command failed with exit code 1.
                          info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
                          
                          florentF 1 Reply Last reply Reply Quote 0
                          • florentF Offline
                            florent Vates 🪐 XO Team @Tristis Oris
                            last edited by

                            @Tristis-Oris thanks , I missed a file
                            I pushed it just now

                            A 1 Reply Last reply Reply Quote 1
                            • A Offline
                              Andrew Top contributor @florent
                              last edited by

                              @florent I finally got the new code running and I tested a Delta Backup (full first run) with NBD x3 enabled and it's leaving NBD transfer (on xcp1) 99% connected after a run. The backup does complete but the task is stuck.

                              florentF 1 Reply Last reply Reply Quote 1
                              • florentF Offline
                                florent Vates 🪐 XO Team @Andrew
                                last edited by

                                @Andrew nice catch andrew I will look into it
                                is it keeping disk attached to dom0 ? (in dashboard -> health )

                                A 1 Reply Last reply Reply Quote 0
                                • A Offline
                                  Andrew Top contributor @florent
                                  last edited by

                                  @florent No. The dashboard health is clean. No VDIs attached to control domain

                                  1 Reply Last reply Reply Quote 1
                                  • florentF Offline
                                    florent Vates 🪐 XO Team
                                    last edited by

                                    so that is probably only a off by one error in the task code
                                    Thanks andrew

                                    1 Reply Last reply Reply Quote 0
                                    • First post
                                      Last post