XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    backblaze b2 / amazon s3 as remote in xoa

    Scheduled Pinned Locked Moved Xen Orchestra
    59 Posts 9 Posters 15.3k Views 3 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • olivierlambertO Online
      olivierlambert Vates πŸͺ Co-Founder CEO
      last edited by

      @nraynaud will be happy to hear that πŸ™‚

      1 Reply Last reply Reply Quote 1
      • nraynaudN Offline
        nraynaud XCP-ng Team @shorian
        last edited by

        @shorian Thank you.

        I will look into the memory issue, that seems to be the clue into a hidden issue, thank you.

        1 Reply Last reply Reply Quote 2
        • nraynaudN Offline
          nraynaud XCP-ng Team @shorian
          last edited by

          @shorian I think I found the memory hog in S3 thanks to your message: https://github.com/vatesfr/xen-orchestra/pull/5456

          nraynaud opened this pull request in vatesfr/xen-orchestra

          closed WiP: fix(backup/s3): maybe fix S3 timeout #5456

          K 1 Reply Last reply Reply Quote 0
          • julien-fJ Offline
            julien-f Vates πŸͺ Co-Founder XO Team
            last edited by

            @akurzawa @DodgeThis @shorian Any of you able to test this branch to see whether it fixes the issue?

            D 3 Replies Last reply Reply Quote 0
            • D Offline
              DodgeThis @julien-f
              last edited by

              @julien-f No problem. glad to test it.πŸ˜€
              πŸ‘

              1 Reply Last reply Reply Quote 1
              • D Offline
                DodgeThis @julien-f
                last edited by DodgeThis

                @julien-f
                Still no sucess,
                This vm that i'm trying to backups is about 290 Gb

                Xen Orchestra is on a VM with:
                Debian 10
                4 Gb Ram
                85 GB free hd
                1CPU

                xo-server 5.72.1
                xo-web 5.75.0

                {
                  "data": {
                    "mode": "full",
                    "reportWhen": "failure"
                  },
                  "id": "1608035511007",
                  "jobId": "695f3dfb-783f-4cb9-85ef-6805d75c7d26",
                  "jobName": "AmazonTest",
                  "message": "backup",
                  "scheduleId": "75c619c3-8b6d-4c72-b110-131b01992b3e",
                  "start": 1608035511007,
                  "status": "failure",
                  "tasks": [
                    {
                      "data": {
                        "type": "VM",
                        "id": "93b88999-8b61-7207-138b-1ef389491e71"
                      },
                      "id": "1608035511027",
                      "message": "Starting backup of MysqlDB. (695f3dfb-783f-4cb9-85ef-6805d75c7d26)",
                      "start": 1608035511027,
                      "status": "failure",
                      "tasks": [
                        {
                          "id": "1608035511031",
                          "message": "snapshot",
                          "start": 1608035511031,
                          "status": "success",
                          "end": 1608035512857,
                          "result": "e612bb3b-9141-45ef-a901-ac8e4f1927c2"
                        },
                        {
                          "id": "1608035512859",
                          "message": "add metadata to snapshot",
                          "start": 1608035512859,
                          "status": "success",
                          "end": 1608035512869
                        },
                        {
                          "id": "1608035513055",
                          "message": "waiting for uptodate snapshot record",
                          "start": 1608035513055,
                          "status": "success",
                          "end": 1608035513259
                        },
                        {
                          "id": "1608035513261",
                          "message": "start VM export",
                          "start": 1608035513261,
                          "status": "success",
                          "end": 1608035513302
                        },
                        {
                          "data": {
                            "id": "72341e61-56db-43f5-8389-78baf364516f",
                            "type": "remote"
                          },
                          "id": "1608035513304",
                          "message": "export",
                          "start": 1608035513304,
                          "status": "failure",
                          "tasks": [
                            {
                              "id": "1608035514389",
                              "message": "transfer",
                              "start": 1608035514389,
                              "status": "failure",
                              "end": 1608035839218,
                              "result": {
                                "message": "write EPIPE",
                                "errno": "EPIPE",
                                "code": "NetworkingError",
                                "syscall": "write",
                                "region": "eu-central-1",
                                "hostname": "s3.eu-central-1.amazonaws.com",
                                "retryable": true,
                                "time": "2020-12-15T12:37:18.736Z",
                                "statusCode": 400,
                                "name": "NetworkingError",
                                "stack": "Error: write EPIPE\n    at WriteWrap.onWriteComplete [as oncomplete] (internal/stream_base_commons.js:94:16)"
                              }
                            }
                          ],
                          "end": 1608035839357,
                          "result": {
                            "message": "write EPIPE",
                            "errno": "EPIPE",
                            "code": "NetworkingError",
                            "syscall": "write",
                            "region": "eu-central-1",
                            "hostname": "s3.eu-central-1.amazonaws.com",
                            "retryable": true,
                            "time": "2020-12-15T12:37:18.736Z",
                            "statusCode": 400,
                            "name": "NetworkingError",
                            "stack": "Error: write EPIPE\n    at WriteWrap.onWriteComplete [as oncomplete] (internal/stream_base_commons.js:94:16)"
                          }
                        },
                        {
                          "id": "1608035839607",
                          "message": "set snapshot.other_config[xo:backup:exported]",
                          "start": 1608035839607,
                          "status": "success",
                          "end": 1608035839831
                        }
                      ],
                      "end": 1608039896381
                    }
                  ],
                  "end": 1608039896383
                }
                
                
                1 Reply Last reply Reply Quote 0
                • D Offline
                  DodgeThis
                  last edited by

                  Another one with a VM with +/-600 Gb

                  {
                    "data": {
                      "mode": "full",
                      "reportWhen": "failure"
                    },
                    "id": "1608043681979",
                    "jobId": "abfed8aa-b445-4c94-be5c-b6ed723792b8",
                    "jobName": "amazontest",
                    "message": "backup",
                    "scheduleId": "44840d13-2302-45ae-9460-53ed451183ed",
                    "start": 1608043681979,
                    "status": "failure",
                    "infos": [
                      {
                        "data": {
                          "vms": [
                            "fbea839a-72ac-ca3d-51e5-d4248e363d72"
                          ]
                        },
                        "message": "vms"
                      }
                    ],
                    "tasks": [
                      {
                        "data": {
                          "type": "VM",
                          "id": "fbea839a-72ac-ca3d-51e5-d4248e363d72"
                        },
                        "id": "1608043681986:0",
                        "message": "Starting backup of DESKTOPGREEN. (abfed8aa-b445-4c94-be5c-b6ed723792b8)",
                        "start": 1608043681986,
                        "status": "failure",
                        "tasks": [
                          {
                            "id": "1608043681990",
                            "message": "snapshot",
                            "start": 1608043681990,
                            "status": "success",
                            "end": 1608043683927,
                            "result": "42f38a1a-5037-107d-1ceb-921fed6910fa"
                          },
                          {
                            "id": "1608043683929",
                            "message": "add metadata to snapshot",
                            "start": 1608043683929,
                            "status": "success",
                            "end": 1608043683938
                          },
                          {
                            "id": "1608043684127",
                            "message": "waiting for uptodate snapshot record",
                            "start": 1608043684127,
                            "status": "success",
                            "end": 1608043684330
                          },
                          {
                            "id": "1608043684332",
                            "message": "start VM export",
                            "start": 1608043684332,
                            "status": "success",
                            "end": 1608043684348
                          },
                          {
                            "data": {
                              "id": "7847ac06-d46f-476a-84c1-608f9302cb67",
                              "type": "remote"
                            },
                            "id": "1608043684352",
                            "message": "export",
                            "start": 1608043684352,
                            "status": "failure",
                            "tasks": [
                              {
                                "id": "1608043684783",
                                "message": "transfer",
                                "start": 1608043684783,
                                "status": "failure",
                                "end": 1608043987240,
                                "result": {
                                  "message": "Your socket connection to the server was not read from or written to within the timeout period. Idle connections will be closed.",
                                  "code": "RequestTimeout",
                                  "region": null,
                                  "time": "2020-12-15T14:53:06.821Z",
                                  "requestId": "E47562F2C7A82C45",
                                  "extendedRequestId": "5nM1888rfrbLHHgUrlsuF5cISTd2kj6oVpWfz5Yhtpj4GtfW7wkX9HnTq/Tj8uU37Opp8FNjlFU=",
                                  "statusCode": 400,
                                  "retryable": true,
                                  "name": "RequestTimeout",
                                  "stack": "RequestTimeout: Your socket connection to the server was not read from or written to within the timeout period. Idle connections will be closed.\n    at Request.extractError (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/services/s3.js:700:35)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:106:20)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:78:10)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:688:14)\n    at Request.transition (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:22:10)\n    at AcceptorStateMachine.runTo (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:14:12)\n    at /opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:26:10\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:38:9)\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:690:12)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:116:18)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:78:10)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:688:14)\n    at Request.transition (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:22:10)\n    at AcceptorStateMachine.runTo (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:14:12)\n    at /opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:26:10\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:38:9)\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:690:12)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:116:18)\n    at callNextListener (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:96:12)\n    at IncomingMessage.onEnd (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/event_listeners.js:313:13)\n    at IncomingMessage.apply (events.js:326:22)\n    at IncomingMessage.patchedEmit (/opt/xo/xo-builds/xen-orchestra-202012150908/@xen-orchestra/log/src/configure.js:92:17)"
                                }
                              }
                            ],
                            "end": 1608043987241,
                            "result": {
                              "message": "Your socket connection to the server was not read from or written to within the timeout period. Idle connections will be closed.",
                              "code": "RequestTimeout",
                              "region": null,
                              "time": "2020-12-15T14:53:06.821Z",
                              "requestId": "E47562F2C7A82C45",
                              "extendedRequestId": "5nM1888rfrbLHHgUrlsuF5cISTd2kj6oVpWfz5Yhtpj4GtfW7wkX9HnTq/Tj8uU37Opp8FNjlFU=",
                              "statusCode": 400,
                              "retryable": true,
                              "name": "RequestTimeout",
                              "stack": "RequestTimeout: Your socket connection to the server was not read from or written to within the timeout period. Idle connections will be closed.\n    at Request.extractError (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/services/s3.js:700:35)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:106:20)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:78:10)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:688:14)\n    at Request.transition (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:22:10)\n    at AcceptorStateMachine.runTo (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:14:12)\n    at /opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:26:10\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:38:9)\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:690:12)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:116:18)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:78:10)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:688:14)\n    at Request.transition (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:22:10)\n    at AcceptorStateMachine.runTo (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:14:12)\n    at /opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:26:10\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:38:9)\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:690:12)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:116:18)\n    at callNextListener (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:96:12)\n    at IncomingMessage.onEnd (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/event_listeners.js:313:13)\n    at IncomingMessage.apply (events.js:326:22)\n    at IncomingMessage.patchedEmit (/opt/xo/xo-builds/xen-orchestra-202012150908/@xen-orchestra/log/src/configure.js:92:17)"
                            }
                          },
                          {
                            "id": "1608043987670",
                            "message": "set snapshot.other_config[xo:backup:exported]",
                            "start": 1608043987670,
                            "status": "success",
                            "end": 1608043989451
                          }
                        ],
                        "end": 1608051207208
                      }
                    ],
                    "end": 1608051207238
                  }
                  
                  1 Reply Last reply Reply Quote 0
                  • D Offline
                    DodgeThis @julien-f
                    last edited by DodgeThis

                    @julien-f shorian like @shorian says increasing mem to 16 Gb fixed the problem

                    julien-fJ nraynaudN 2 Replies Last reply Reply Quote 1
                    • julien-fJ Offline
                      julien-f Vates πŸͺ Co-Founder XO Team @DodgeThis
                      last edited by

                      @nraynaud ping

                      1 Reply Last reply Reply Quote 0
                      • nraynaudN Offline
                        nraynaud XCP-ng Team @DodgeThis
                        last edited by

                        @dodgethis thank you, I will look harder into this.

                        1 Reply Last reply Reply Quote 0
                        • D Offline
                          DodgeThis
                          last edited by

                          @nraynaud glad to help somehow 😊

                          1 Reply Last reply Reply Quote 0
                          • K Offline
                            klou @nraynaud
                            last edited by

                            @nraynaud

                            Confirmed bumping to 12 Gb RAM allowed my "larger" VM's (60 GB+) to complete.

                            xo-server 5.73.0
                            xo-web 5.76.0

                            1 Reply Last reply Reply Quote 0
                            • nraynaudN Offline
                              nraynaud XCP-ng Team
                              last edited by

                              Can someone please test the issue with the current master branch please?

                              S 1 Reply Last reply Reply Quote 0
                              • S Offline
                                shorian @nraynaud
                                last edited by

                                @nraynaud Will aim to do this over the weekend

                                S 1 Reply Last reply Reply Quote 0
                                • S Offline
                                  shorian @shorian
                                  last edited by shorian

                                  I cannot get the backup to run; I have now tried about 12 times to run a backup that should complete in under 2 hours; without a timeout it runs on (stopped it after 10 hours), with a timeout of 3 hours it consistently fails with 'interrupted' without explanation.

                                  Looking through the logs, we've got a lot of buffer allocation errors, it starts with this:

                                  xo:xo-server ERROR uncaught exception {
                                    error: Error [RangeError]: Array buffer allocation failed
                                        at new ArrayBuffer (<anonymous>)
                                        at new Uint8Array (<anonymous>)
                                        at new FastBuffer (internal/buffer.js:950:1)
                                        at createUnsafeBuffer (buffer.js:149:12)
                                        at allocate (buffer.js:419:10)
                                        at Function.allocUnsafe (buffer.js:384:10)
                                        at BufferList._getBuffer (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/internal/streams/buffer_list.js:166:24)
                                        at BufferList.consume (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/internal/streams/buffer_list.js:119:54)
                                        at fromList (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/_stream_readable.js:1073:24)
                                        at Transform.Readable.read (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/_stream_readable.js:465:20)
                                        at ManagedUpload.fillStream (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:422:25)
                                        at Response.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:574:24)
                                        at Request.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:369:18)
                                        at Request.callListeners (/opt/xen-orchestra/node_modules/aws-sdk/lib/sequential_executor.js:106:20)
                                        at Request.emit (/opt/xen-orchestra/node_modules/aws-sdk/lib/sequential_executor.js:78:10)
                                        at Request.emit (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:688:14)
                                        at Request.transition (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:22:10)
                                        at AcceptorStateMachine.runTo (/opt/xen-orchestra/node_modules/aws-sdk/lib/state_machine.js:14:12)
                                        at /opt/xen-orchestra/node_modules/aws-sdk/lib/state_machine.js:26:10
                                        at Request.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:38:9)
                                        at Request.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:690:12)
                                        at Request.callListeners (/opt/xen-orchestra/node_modules/aws-sdk/lib/sequential_executor.js:116:18) {
                                      code: 'RangeError',
                                      time: 2021-01-16T22:07:11.590Z
                                    }
                                  }
                                  

                                  ...followed by this repeating numerous times:

                                  xo:xo-server ERROR uncaught exception {
                                    error: RangeError: Array buffer allocation failed
                                        at new ArrayBuffer (<anonymous>)
                                        at new Uint8Array (<anonymous>)
                                        at new FastBuffer (internal/buffer.js:950:1)
                                        at createUnsafeBuffer (buffer.js:149:12)
                                        at allocate (buffer.js:419:10)
                                        at Function.allocUnsafe (buffer.js:384:10)
                                        at Function.concat (buffer.js:564:25)
                                        at ManagedUpload.fillStream (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:433:38)
                                        at Transform.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:188:44)
                                        at Transform.apply (events.js:314:20)
                                        at Transform.patchedEmit (/opt/xen-orchestra/@xen-orchestra/log/src/configure.js:92:17)
                                        at Transform.EventEmitter.emit (domain.js:483:12)
                                        at emitReadable_ (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/_stream_readable.js:538:12)
                                        at processTicksAndRejections (internal/process/task_queues.js:83:21)
                                  }
                                  

                                  Going to try whacking the memory right up ; will report back later today.

                                  S 1 Reply Last reply Reply Quote 0
                                  • S Offline
                                    shorian @shorian
                                    last edited by

                                    Ok, been running this all weekend. In positive news, I've not had the timeout error once. In less positive news, I've yet to get a successful backup 😞

                                    With lots of memory and unlimited time, I'm afraid we consistently fail with HANDLE_INVALID. I'm unsure where to go with things from here - @nraynaud let me know what I can do to help rather than throw problems at you. I'm at your disposal.

                                    nraynaudN 1 Reply Last reply Reply Quote 0
                                    • nraynaudN Offline
                                      nraynaud XCP-ng Team @shorian
                                      last edited by

                                      @shorian thank you for your help. can you tell me what kind of backup you were launching and at what phase the error happened ?

                                      S 1 Reply Last reply Reply Quote 0
                                      • S Offline
                                        shorian @nraynaud
                                        last edited by shorian

                                        @nraynaud Normal backup to B2, concurrency set to 1, compression set to zsd - no other settings (ie no 'delete before' or 'with memory' - just a standard backup with zstd compression). I have also tried without compression without any variation in outcome.

                                        Error occurs immediately after snapshot (about 5 mins later, but its the next line in the debug log), but we don't fail out for another 18 minutes; we get the bad allocation error every second until we finally die with a terminal 'bad allocation' error and completely fall over with an auto-restart of the script:

                                        2021-01-16T22:20:00.164Z xo:xo-server ERROR uncaught exception {
                                          error: RangeError: Array buffer allocation failed
                                              at new ArrayBuffer (<anonymous>)
                                              at new Uint8Array (<anonymous>)
                                              at new FastBuffer (internal/buffer.js:950:1)
                                              at createUnsafeBuffer (buffer.js:149:12)
                                              at allocate (buffer.js:419:10)
                                              at Function.allocUnsafe (buffer.js:384:10)
                                              at Function.concat (buffer.js:564:25)
                                              at ManagedUpload.fillStream (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:433:38)
                                              at Transform.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:188:44)
                                              at Transform.apply (events.js:314:20)
                                              at Transform.patchedEmit (/opt/xen-orchestra/@xen-orchestra/log/src/configure.js:92:17)
                                              at Transform.EventEmitter.emit (domain.js:483:12)
                                              at emitReadable_ (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/_stream_readable.js:538:12)
                                              at processTicksAndRejections (internal/process/task_queues.js:83:21)
                                        }
                                        terminate called after throwing an instance of 'std::bad_alloc'
                                          what():  std::bad_alloc
                                        error: Forever detected script was killed by signal: SIGABRT
                                        error: Script restart attempt #2
                                        2021-01-16T22:20:17.950Z app-conf /opt/xen-orchestra/packages/xo-server/config.toml
                                        2021-01-16T22:20:17.956Z app-conf /root/.config/xo-server/config.toml
                                        2021-01-16T22:20:17.957Z app-conf /root/.config/xo-server/config.z-auto.json
                                        2021-01-16T22:20:17.960Z xo:main INFO Configuration loaded.
                                        2021-01-16T22:20:17.963Z xo:main INFO Web server listening on http://[::]:80
                                        

                                        Interestingly I've left the backup process to keep running every few hours, and we're now getting some success - I've yet to have the 3 VMs in question all back up in one pass, but the last couple of passes have resulted in one or two of them completing successfully. See sample logs:

                                        2021-01-18T09_00_00.011Z - backup NG.log.txt

                                        2021-01-18T03_00_00.009Z - backup NG.log.txt

                                        In summary, we are now we are getting some timeouts - different than the 12,000ms that we were seeing originally, but now with the new errors in the attached logs of 'no object with UUID or opaque ref'.

                                        If I lower the memory back down to 4gb we revert back to the buffer allocation errors.

                                        I'm very happy to send over more extensive logs but would prefer to do so 1:1 rather than into a public forum.

                                        nraynaudN 2 Replies Last reply Reply Quote 0
                                        • nraynaudN Offline
                                          nraynaud XCP-ng Team @shorian
                                          last edited by

                                          @shorian thank you I will look at all that.

                                          1 Reply Last reply Reply Quote 1
                                          • nraynaudN Offline
                                            nraynaud XCP-ng Team @shorian
                                            last edited by

                                            @shorian Capture d’écran 2021-01-19 aΜ€ 14.07.27.png

                                            I just did a test with this configuration, would you mind checking that it as close as possible as yours? I got a backup to work with this configuration. That suggest we have to search for a limit and not plainly search for a bug.

                                            If it's the same backup configuration, we'll have to look for a difference in the VMs (mine transfers 11.77 GiB of data).

                                            Can you describe your VM please? in particular the number and size of the disks.

                                            thank you for your help.

                                            S 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post