XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    backblaze b2 / amazon s3 as remote in xoa

    Scheduled Pinned Locked Moved Xen Orchestra
    59 Posts 9 Posters 15.2k Views 3 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • nraynaudN Offline
      nraynaud XCP-ng Team @shorian
      last edited by

      shorian Thank you.

      I will look into the memory issue, that seems to be the clue into a hidden issue, thank you.

      1 Reply Last reply Reply Quote 2
      • nraynaudN Offline
        nraynaud XCP-ng Team @shorian
        last edited by

        shorian I think I found the memory hog in S3 thanks to your message: https://github.com/vatesfr/xen-orchestra/pull/5456

        nraynaud opened this pull request in vatesfr/xen-orchestra

        closed WiP: fix(backup/s3): maybe fix S3 timeout #5456

        K 1 Reply Last reply Reply Quote 0
        • julien-fJ Offline
          julien-f Vates ๐Ÿช Co-Founder XO Team
          last edited by

          akurzawa DodgeThis shorian Any of you able to test this branch to see whether it fixes the issue?

          D 3 Replies Last reply Reply Quote 0
          • D Offline
            DodgeThis @julien-f
            last edited by

            julien-f No problem. glad to test it.๐Ÿ˜€
            ๐Ÿ‘

            1 Reply Last reply Reply Quote 1
            • D Offline
              DodgeThis @julien-f
              last edited by DodgeThis

              julien-f
              Still no sucess,
              This vm that i'm trying to backups is about 290 Gb

              Xen Orchestra is on a VM with:
              Debian 10
              4 Gb Ram
              85 GB free hd
              1CPU

              xo-server 5.72.1
              xo-web 5.75.0

              {
                "data": {
                  "mode": "full",
                  "reportWhen": "failure"
                },
                "id": "1608035511007",
                "jobId": "695f3dfb-783f-4cb9-85ef-6805d75c7d26",
                "jobName": "AmazonTest",
                "message": "backup",
                "scheduleId": "75c619c3-8b6d-4c72-b110-131b01992b3e",
                "start": 1608035511007,
                "status": "failure",
                "tasks": [
                  {
                    "data": {
                      "type": "VM",
                      "id": "93b88999-8b61-7207-138b-1ef389491e71"
                    },
                    "id": "1608035511027",
                    "message": "Starting backup of MysqlDB. (695f3dfb-783f-4cb9-85ef-6805d75c7d26)",
                    "start": 1608035511027,
                    "status": "failure",
                    "tasks": [
                      {
                        "id": "1608035511031",
                        "message": "snapshot",
                        "start": 1608035511031,
                        "status": "success",
                        "end": 1608035512857,
                        "result": "e612bb3b-9141-45ef-a901-ac8e4f1927c2"
                      },
                      {
                        "id": "1608035512859",
                        "message": "add metadata to snapshot",
                        "start": 1608035512859,
                        "status": "success",
                        "end": 1608035512869
                      },
                      {
                        "id": "1608035513055",
                        "message": "waiting for uptodate snapshot record",
                        "start": 1608035513055,
                        "status": "success",
                        "end": 1608035513259
                      },
                      {
                        "id": "1608035513261",
                        "message": "start VM export",
                        "start": 1608035513261,
                        "status": "success",
                        "end": 1608035513302
                      },
                      {
                        "data": {
                          "id": "72341e61-56db-43f5-8389-78baf364516f",
                          "type": "remote"
                        },
                        "id": "1608035513304",
                        "message": "export",
                        "start": 1608035513304,
                        "status": "failure",
                        "tasks": [
                          {
                            "id": "1608035514389",
                            "message": "transfer",
                            "start": 1608035514389,
                            "status": "failure",
                            "end": 1608035839218,
                            "result": {
                              "message": "write EPIPE",
                              "errno": "EPIPE",
                              "code": "NetworkingError",
                              "syscall": "write",
                              "region": "eu-central-1",
                              "hostname": "s3.eu-central-1.amazonaws.com",
                              "retryable": true,
                              "time": "2020-12-15T12:37:18.736Z",
                              "statusCode": 400,
                              "name": "NetworkingError",
                              "stack": "Error: write EPIPE\n    at WriteWrap.onWriteComplete [as oncomplete] (internal/stream_base_commons.js:94:16)"
                            }
                          }
                        ],
                        "end": 1608035839357,
                        "result": {
                          "message": "write EPIPE",
                          "errno": "EPIPE",
                          "code": "NetworkingError",
                          "syscall": "write",
                          "region": "eu-central-1",
                          "hostname": "s3.eu-central-1.amazonaws.com",
                          "retryable": true,
                          "time": "2020-12-15T12:37:18.736Z",
                          "statusCode": 400,
                          "name": "NetworkingError",
                          "stack": "Error: write EPIPE\n    at WriteWrap.onWriteComplete [as oncomplete] (internal/stream_base_commons.js:94:16)"
                        }
                      },
                      {
                        "id": "1608035839607",
                        "message": "set snapshot.other_config[xo:backup:exported]",
                        "start": 1608035839607,
                        "status": "success",
                        "end": 1608035839831
                      }
                    ],
                    "end": 1608039896381
                  }
                ],
                "end": 1608039896383
              }
              
              
              1 Reply Last reply Reply Quote 0
              • D Offline
                DodgeThis
                last edited by

                Another one with a VM with +/-600 Gb

                {
                  "data": {
                    "mode": "full",
                    "reportWhen": "failure"
                  },
                  "id": "1608043681979",
                  "jobId": "abfed8aa-b445-4c94-be5c-b6ed723792b8",
                  "jobName": "amazontest",
                  "message": "backup",
                  "scheduleId": "44840d13-2302-45ae-9460-53ed451183ed",
                  "start": 1608043681979,
                  "status": "failure",
                  "infos": [
                    {
                      "data": {
                        "vms": [
                          "fbea839a-72ac-ca3d-51e5-d4248e363d72"
                        ]
                      },
                      "message": "vms"
                    }
                  ],
                  "tasks": [
                    {
                      "data": {
                        "type": "VM",
                        "id": "fbea839a-72ac-ca3d-51e5-d4248e363d72"
                      },
                      "id": "1608043681986:0",
                      "message": "Starting backup of DESKTOPGREEN. (abfed8aa-b445-4c94-be5c-b6ed723792b8)",
                      "start": 1608043681986,
                      "status": "failure",
                      "tasks": [
                        {
                          "id": "1608043681990",
                          "message": "snapshot",
                          "start": 1608043681990,
                          "status": "success",
                          "end": 1608043683927,
                          "result": "42f38a1a-5037-107d-1ceb-921fed6910fa"
                        },
                        {
                          "id": "1608043683929",
                          "message": "add metadata to snapshot",
                          "start": 1608043683929,
                          "status": "success",
                          "end": 1608043683938
                        },
                        {
                          "id": "1608043684127",
                          "message": "waiting for uptodate snapshot record",
                          "start": 1608043684127,
                          "status": "success",
                          "end": 1608043684330
                        },
                        {
                          "id": "1608043684332",
                          "message": "start VM export",
                          "start": 1608043684332,
                          "status": "success",
                          "end": 1608043684348
                        },
                        {
                          "data": {
                            "id": "7847ac06-d46f-476a-84c1-608f9302cb67",
                            "type": "remote"
                          },
                          "id": "1608043684352",
                          "message": "export",
                          "start": 1608043684352,
                          "status": "failure",
                          "tasks": [
                            {
                              "id": "1608043684783",
                              "message": "transfer",
                              "start": 1608043684783,
                              "status": "failure",
                              "end": 1608043987240,
                              "result": {
                                "message": "Your socket connection to the server was not read from or written to within the timeout period. Idle connections will be closed.",
                                "code": "RequestTimeout",
                                "region": null,
                                "time": "2020-12-15T14:53:06.821Z",
                                "requestId": "E47562F2C7A82C45",
                                "extendedRequestId": "5nM1888rfrbLHHgUrlsuF5cISTd2kj6oVpWfz5Yhtpj4GtfW7wkX9HnTq/Tj8uU37Opp8FNjlFU=",
                                "statusCode": 400,
                                "retryable": true,
                                "name": "RequestTimeout",
                                "stack": "RequestTimeout: Your socket connection to the server was not read from or written to within the timeout period. Idle connections will be closed.\n    at Request.extractError (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/services/s3.js:700:35)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:106:20)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:78:10)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:688:14)\n    at Request.transition (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:22:10)\n    at AcceptorStateMachine.runTo (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:14:12)\n    at /opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:26:10\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:38:9)\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:690:12)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:116:18)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:78:10)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:688:14)\n    at Request.transition (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:22:10)\n    at AcceptorStateMachine.runTo (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:14:12)\n    at /opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:26:10\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:38:9)\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:690:12)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:116:18)\n    at callNextListener (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:96:12)\n    at IncomingMessage.onEnd (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/event_listeners.js:313:13)\n    at IncomingMessage.apply (events.js:326:22)\n    at IncomingMessage.patchedEmit (/opt/xo/xo-builds/xen-orchestra-202012150908/@xen-orchestra/log/src/configure.js:92:17)"
                              }
                            }
                          ],
                          "end": 1608043987241,
                          "result": {
                            "message": "Your socket connection to the server was not read from or written to within the timeout period. Idle connections will be closed.",
                            "code": "RequestTimeout",
                            "region": null,
                            "time": "2020-12-15T14:53:06.821Z",
                            "requestId": "E47562F2C7A82C45",
                            "extendedRequestId": "5nM1888rfrbLHHgUrlsuF5cISTd2kj6oVpWfz5Yhtpj4GtfW7wkX9HnTq/Tj8uU37Opp8FNjlFU=",
                            "statusCode": 400,
                            "retryable": true,
                            "name": "RequestTimeout",
                            "stack": "RequestTimeout: Your socket connection to the server was not read from or written to within the timeout period. Idle connections will be closed.\n    at Request.extractError (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/services/s3.js:700:35)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:106:20)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:78:10)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:688:14)\n    at Request.transition (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:22:10)\n    at AcceptorStateMachine.runTo (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:14:12)\n    at /opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:26:10\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:38:9)\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:690:12)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:116:18)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:78:10)\n    at Request.emit (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:688:14)\n    at Request.transition (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:22:10)\n    at AcceptorStateMachine.runTo (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:14:12)\n    at /opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/state_machine.js:26:10\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:38:9)\n    at Request.<anonymous> (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/request.js:690:12)\n    at Request.callListeners (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:116:18)\n    at callNextListener (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/sequential_executor.js:96:12)\n    at IncomingMessage.onEnd (/opt/xo/xo-builds/xen-orchestra-202012150908/node_modules/aws-sdk/lib/event_listeners.js:313:13)\n    at IncomingMessage.apply (events.js:326:22)\n    at IncomingMessage.patchedEmit (/opt/xo/xo-builds/xen-orchestra-202012150908/@xen-orchestra/log/src/configure.js:92:17)"
                          }
                        },
                        {
                          "id": "1608043987670",
                          "message": "set snapshot.other_config[xo:backup:exported]",
                          "start": 1608043987670,
                          "status": "success",
                          "end": 1608043989451
                        }
                      ],
                      "end": 1608051207208
                    }
                  ],
                  "end": 1608051207238
                }
                
                1 Reply Last reply Reply Quote 0
                • D Offline
                  DodgeThis @julien-f
                  last edited by DodgeThis

                  julien-f shorian like shorian says increasing mem to 16 Gb fixed the problem

                  julien-fJ nraynaudN 2 Replies Last reply Reply Quote 1
                  • julien-fJ Offline
                    julien-f Vates ๐Ÿช Co-Founder XO Team @DodgeThis
                    last edited by

                    nraynaud ping

                    1 Reply Last reply Reply Quote 0
                    • nraynaudN Offline
                      nraynaud XCP-ng Team @DodgeThis
                      last edited by

                      DodgeThis thank you, I will look harder into this.

                      1 Reply Last reply Reply Quote 0
                      • D Offline
                        DodgeThis
                        last edited by

                        nraynaud glad to help somehow ๐Ÿ˜Š

                        1 Reply Last reply Reply Quote 0
                        • K Offline
                          klou @nraynaud
                          last edited by

                          nraynaud

                          Confirmed bumping to 12 Gb RAM allowed my "larger" VM's (60 GB+) to complete.

                          xo-server 5.73.0
                          xo-web 5.76.0

                          1 Reply Last reply Reply Quote 0
                          • nraynaudN Offline
                            nraynaud XCP-ng Team
                            last edited by

                            Can someone please test the issue with the current master branch please?

                            S 1 Reply Last reply Reply Quote 0
                            • S Offline
                              shorian @nraynaud
                              last edited by

                              nraynaud Will aim to do this over the weekend

                              S 1 Reply Last reply Reply Quote 0
                              • S Offline
                                shorian @shorian
                                last edited by shorian

                                I cannot get the backup to run; I have now tried about 12 times to run a backup that should complete in under 2 hours; without a timeout it runs on (stopped it after 10 hours), with a timeout of 3 hours it consistently fails with 'interrupted' without explanation.

                                Looking through the logs, we've got a lot of buffer allocation errors, it starts with this:

                                xo:xo-server ERROR uncaught exception {
                                  error: Error [RangeError]: Array buffer allocation failed
                                      at new ArrayBuffer (<anonymous>)
                                      at new Uint8Array (<anonymous>)
                                      at new FastBuffer (internal/buffer.js:950:1)
                                      at createUnsafeBuffer (buffer.js:149:12)
                                      at allocate (buffer.js:419:10)
                                      at Function.allocUnsafe (buffer.js:384:10)
                                      at BufferList._getBuffer (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/internal/streams/buffer_list.js:166:24)
                                      at BufferList.consume (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/internal/streams/buffer_list.js:119:54)
                                      at fromList (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/_stream_readable.js:1073:24)
                                      at Transform.Readable.read (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/_stream_readable.js:465:20)
                                      at ManagedUpload.fillStream (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:422:25)
                                      at Response.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:574:24)
                                      at Request.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:369:18)
                                      at Request.callListeners (/opt/xen-orchestra/node_modules/aws-sdk/lib/sequential_executor.js:106:20)
                                      at Request.emit (/opt/xen-orchestra/node_modules/aws-sdk/lib/sequential_executor.js:78:10)
                                      at Request.emit (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:688:14)
                                      at Request.transition (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:22:10)
                                      at AcceptorStateMachine.runTo (/opt/xen-orchestra/node_modules/aws-sdk/lib/state_machine.js:14:12)
                                      at /opt/xen-orchestra/node_modules/aws-sdk/lib/state_machine.js:26:10
                                      at Request.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:38:9)
                                      at Request.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/request.js:690:12)
                                      at Request.callListeners (/opt/xen-orchestra/node_modules/aws-sdk/lib/sequential_executor.js:116:18) {
                                    code: 'RangeError',
                                    time: 2021-01-16T22:07:11.590Z
                                  }
                                }
                                

                                ...followed by this repeating numerous times:

                                xo:xo-server ERROR uncaught exception {
                                  error: RangeError: Array buffer allocation failed
                                      at new ArrayBuffer (<anonymous>)
                                      at new Uint8Array (<anonymous>)
                                      at new FastBuffer (internal/buffer.js:950:1)
                                      at createUnsafeBuffer (buffer.js:149:12)
                                      at allocate (buffer.js:419:10)
                                      at Function.allocUnsafe (buffer.js:384:10)
                                      at Function.concat (buffer.js:564:25)
                                      at ManagedUpload.fillStream (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:433:38)
                                      at Transform.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:188:44)
                                      at Transform.apply (events.js:314:20)
                                      at Transform.patchedEmit (/opt/xen-orchestra/@xen-orchestra/log/src/configure.js:92:17)
                                      at Transform.EventEmitter.emit (domain.js:483:12)
                                      at emitReadable_ (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/_stream_readable.js:538:12)
                                      at processTicksAndRejections (internal/process/task_queues.js:83:21)
                                }
                                

                                Going to try whacking the memory right up ; will report back later today.

                                S 1 Reply Last reply Reply Quote 0
                                • S Offline
                                  shorian @shorian
                                  last edited by

                                  Ok, been running this all weekend. In positive news, I've not had the timeout error once. In less positive news, I've yet to get a successful backup ๐Ÿ˜ž

                                  With lots of memory and unlimited time, I'm afraid we consistently fail with HANDLE_INVALID. I'm unsure where to go with things from here - nraynaud let me know what I can do to help rather than throw problems at you. I'm at your disposal.

                                  nraynaudN 1 Reply Last reply Reply Quote 0
                                  • nraynaudN Offline
                                    nraynaud XCP-ng Team @shorian
                                    last edited by

                                    shorian thank you for your help. can you tell me what kind of backup you were launching and at what phase the error happened ?

                                    S 1 Reply Last reply Reply Quote 0
                                    • S Offline
                                      shorian @nraynaud
                                      last edited by shorian

                                      nraynaud Normal backup to B2, concurrency set to 1, compression set to zsd - no other settings (ie no 'delete before' or 'with memory' - just a standard backup with zstd compression). I have also tried without compression without any variation in outcome.

                                      Error occurs immediately after snapshot (about 5 mins later, but its the next line in the debug log), but we don't fail out for another 18 minutes; we get the bad allocation error every second until we finally die with a terminal 'bad allocation' error and completely fall over with an auto-restart of the script:

                                      2021-01-16T22:20:00.164Z xo:xo-server ERROR uncaught exception {
                                        error: RangeError: Array buffer allocation failed
                                            at new ArrayBuffer (<anonymous>)
                                            at new Uint8Array (<anonymous>)
                                            at new FastBuffer (internal/buffer.js:950:1)
                                            at createUnsafeBuffer (buffer.js:149:12)
                                            at allocate (buffer.js:419:10)
                                            at Function.allocUnsafe (buffer.js:384:10)
                                            at Function.concat (buffer.js:564:25)
                                            at ManagedUpload.fillStream (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:433:38)
                                            at Transform.<anonymous> (/opt/xen-orchestra/node_modules/aws-sdk/lib/s3/managed_upload.js:188:44)
                                            at Transform.apply (events.js:314:20)
                                            at Transform.patchedEmit (/opt/xen-orchestra/@xen-orchestra/log/src/configure.js:92:17)
                                            at Transform.EventEmitter.emit (domain.js:483:12)
                                            at emitReadable_ (/opt/xen-orchestra/@xen-orchestra/fs/node_modules/readable-stream/lib/_stream_readable.js:538:12)
                                            at processTicksAndRejections (internal/process/task_queues.js:83:21)
                                      }
                                      terminate called after throwing an instance of 'std::bad_alloc'
                                        what():  std::bad_alloc
                                      error: Forever detected script was killed by signal: SIGABRT
                                      error: Script restart attempt #2
                                      2021-01-16T22:20:17.950Z app-conf /opt/xen-orchestra/packages/xo-server/config.toml
                                      2021-01-16T22:20:17.956Z app-conf /root/.config/xo-server/config.toml
                                      2021-01-16T22:20:17.957Z app-conf /root/.config/xo-server/config.z-auto.json
                                      2021-01-16T22:20:17.960Z xo:main INFO Configuration loaded.
                                      2021-01-16T22:20:17.963Z xo:main INFO Web server listening on http://[::]:80
                                      

                                      Interestingly I've left the backup process to keep running every few hours, and we're now getting some success - I've yet to have the 3 VMs in question all back up in one pass, but the last couple of passes have resulted in one or two of them completing successfully. See sample logs:

                                      2021-01-18T09_00_00.011Z - backup NG.log.txt

                                      2021-01-18T03_00_00.009Z - backup NG.log.txt

                                      In summary, we are now we are getting some timeouts - different than the 12,000ms that we were seeing originally, but now with the new errors in the attached logs of 'no object with UUID or opaque ref'.

                                      If I lower the memory back down to 4gb we revert back to the buffer allocation errors.

                                      I'm very happy to send over more extensive logs but would prefer to do so 1:1 rather than into a public forum.

                                      nraynaudN 2 Replies Last reply Reply Quote 0
                                      • nraynaudN Offline
                                        nraynaud XCP-ng Team @shorian
                                        last edited by

                                        shorian thank you I will look at all that.

                                        1 Reply Last reply Reply Quote 1
                                        • nraynaudN Offline
                                          nraynaud XCP-ng Team @shorian
                                          last edited by

                                          shorian Capture dโ€™eฬcran 2021-01-19 aฬ€ 14.07.27.png

                                          I just did a test with this configuration, would you mind checking that it as close as possible as yours? I got a backup to work with this configuration. That suggest we have to search for a limit and not plainly search for a bug.

                                          If it's the same backup configuration, we'll have to look for a difference in the VMs (mine transfers 11.77 GiB of data).

                                          Can you describe your VM please? in particular the number and size of the disks.

                                          thank you for your help.

                                          S 1 Reply Last reply Reply Quote 0
                                          • S Offline
                                            shorian @nraynaud
                                            last edited by

                                            nraynaud Identical setup, except using B2 not S3.

                                            Initially looking at backing up 3 VMs, each with a single disk sized 100gb, 150gb and 25gb respectively, SR is thin provisioned (ext4), total backup size (compressed) is about 76gb over a 1gbps connection; takes about 90 mins for a clean backup.

                                            Individually they back up fine, as a group they don't. (In total have about 1.2tb of VMs to back up but not trying for that one just yet).

                                            S 1 Reply Last reply Reply Quote 0
                                            • First post
                                              Last post