XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login
    1. Home
    2. klou
    3. Posts
    K
    Offline
    • Profile
    • Following 0
    • Followers 0
    • Topics 0
    • Posts 16
    • Groups 0

    Posts

    Recent Best Controversial
    • RE: Wasabi just wont work...

      5.79.3 / 5.81.0

      All of my backups work (now), and speeds range between 5 MB/s to 20 MB/s.

      Most of my VM's are smaller (zstd helps a lot), though the largest archive is about 80 GB.

      Also, all of my backups are sequential -- I seem to remember a previous note where somebody had problems with concurrent uploads.

      My XOCE VM runs with 12 GB, which was the number that "just works" according to some of my earlier problems (and I haven't bothered to chase down or fine tune further).

      posted in Xen Orchestra
      K
      klou
    • RE: backblaze b2 / amazon s3 as remote in xoa

      @nraynaud

      Confirmed bumping to 12 Gb RAM allowed my "larger" VM's (60 GB+) to complete.

      xo-server 5.73.0
      xo-web 5.76.0

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      @nraynaud Sorry about dropping this for a while. If you want to pick it back up, let me know and I'll try to test changes as well.

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      @nraynaud

      If it helps, I'm reasonably certain that I was able to backup and upload similarly sized VMs previously without memory issues, before the 50GB+ chunking changes.

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      @nraynaud

      Yes, Full Backup, target is S3, both with or without compression.

      (Side note: I didn't realize that Delta Backups were possible with S3. This could be significant in storage space. But I also assume that this is *nix VMs only?)

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      @Danp

      Just did, reduced to 2GB RAM. The pfSense Backup was about the same, except the post-backup idle state was around 900MB usage.

      The 2nd VM bombed out with an "Interrupted" transfer status (similar to a few posts above).

      b15efd75-9f70-4089-a7c0-84e0445a3ba1-image.png

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      @olivierlambert

      OK, DL'ed/Registered XOA, bumped it to 6 GB just in case (VM only, not the node process).

      Updated to XOA latest channel, 5.51.1 (5.68.0/5.72.0). (p.s. thanks for the config backup/restore function. Needed to restart the VM or xo-server, but retained my S3 remote settings.)

      First is a pfSense VM, ~600 MB after zstd. The initial 5+GB usage is VM bootup.
      808b296f-54a7-4fd3-906b-6dd18e543210-image.png

      Next is the same VM that I used for the previous tests. ~7GB after ztsd. The job started around 9:30, where the initial ramp-up occurred (during snapshot and transfer start). Then it jumped further to 5+ GB.
      e513c294-0b05-4e73-8b5b-21f9d4af944e-image.png

      That's about as clear as I can get in the 10 minute window. It finished, dropped down to 3.5GB, and then eventually back to 1.4GB.

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      @olivierlambert From sources. Latest pull was last Friday, so 5.68.0/5.72.0.

      Memory usage is relatively stable around 1.4 GB (most of this morning, with --max-old-space-size=2560) , balloons during a S3 backup job, and then goes back down to 1.4 GB when transfer is complete.

      a4944018-7ad0-4b75-b540-c2a103228fc1-image.png

      edit: The above was a backup without compression.

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      @olivierlambert

      Again, thanks for the help with this.

      Whether or not the node process is actually using that much (I've been reading that by default, it maxes around 1.4 GB, but I just increased that with --max-old-space-size=2560), larger S3 backups/transfers are still only successful if I increase XO VM memory to > 5 GB.

      A recent backup run showed overall usage ballooned to 5.5+ GB during the backup, and then went back to ~1.4GB afterwards.

      I don't know if this is intended behavior or if you want to finetune it later, but leaving the VM at 6 GB works for me.

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      Thanks for the hint. The upload still eventually times out now (after 12000ms of inactivity somewhere), and I still have sporadic TimeOutErrors (see above) during the transfer.

      However, the xo-server process doesn't obviously crash anymore after I increased VM memory from 3GB to 4GB. I see that XOA defaults to 2GB -- what's recommended at this point (~30 VMs, 3 hosts)?

      With both 3GB or 4GB, top shows the node process taking ~90% of memory during a transfer. I wonder if it's buffering the entire upload chunk in memory?

      With that in mind, I increased to 5.5GB, since the largest upload chunk should be 5GB. And it completed the upload successfully, though still using 90% memory throughout the process. This ended up being a 6GB upload, after zstd.

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      Updated today. Here's the snippet from syslog.

      Oct 14 14:42:00 chaz xo-server[101910]: [load-balancer]Execute plans!
      Oct 14 14:43:00 chaz xo-server[101910]: [load-balancer]Execute plans!
      Oct 14 14:44:00 chaz xo-server[101910]: [load-balancer]Execute plans!
      Oct 14 14:44:16 chaz xo-server[101910]: 2020-10-14T21:44:16.127Z xo:main INFO + WebSocket connection (::ffff:10.0.0.15)
      Oct 14 14:44:59 chaz xo-server[101910]: 2020-10-14T21:44:59.644Z xo:xapi DEBUG Snapshotting VM Calculon as [XO Backup Calculon-Onetime] Calculon
      Oct 14 14:45:00 chaz xo-server[101910]: [load-balancer]Execute plans!
      Oct 14 14:45:02 chaz xo-server[101910]: 2020-10-14T21:45:02.694Z xo:xapi DEBUG Deleting VM [XO Backup Calculon-Onetime] Calculon
      Oct 14 14:45:02 chaz xo-server[101910]: 2020-10-14T21:45:02.916Z xo:xapi DEBUG Deleting VDI OpaqueRef:67b407aa-14a4-4452-9d7c-852b645bbd90
      Oct 14 14:46:01 chaz xo-server[101910]: [load-balancer]Execute plans!
      Oct 14 14:46:59 chaz xo-server[101910]: _watchEvents TimeoutError: operation timed out
      Oct 14 14:46:59 chaz xo-server[101910]:     at Promise.call (/opt/xen-orchestra/node_modules/promise-toolbox/timeout.js:13:16)
      Oct 14 14:46:59 chaz xo-server[101910]:     at Xapi._call (/opt/xen-orchestra/packages/xen-api/src/index.js:666:37)
      Oct 14 14:46:59 chaz xo-server[101910]:     at Xapi._watchEvents (/opt/xen-orchestra/packages/xen-api/src/index.js:1012:31)
      Oct 14 14:46:59 chaz xo-server[101910]:     at runMicrotasks (<anonymous>)
      Oct 14 14:46:59 chaz xo-server[101910]:     at processTicksAndRejections (internal/process/task_queues.js:97:5) {
      Oct 14 14:46:59 chaz xo-server[101910]:   call: {
      Oct 14 14:46:59 chaz xo-server[101910]:     method: 'event.from',
      Oct 14 14:46:59 chaz xo-server[101910]:     params: [ [Array], '00000000000012051192,00000000000011907676', 60.1 ]
      Oct 14 14:46:59 chaz xo-server[101910]:   }
      Oct 14 14:46:59 chaz xo-server[101910]: }
      Oct 14 14:47:00 chaz xo-server[101910]: [load-balancer]Execute plans!
      Oct 14 14:48:05 chaz xo-server[101910]: [load-balancer]Execute plans!
      Oct 14 14:48:06 chaz xo-server[101910]: terminate called after throwing an instance of 'std::bad_alloc'
      Oct 14 14:48:06 chaz xo-server[101910]:   what():  std::bad_alloc
      Oct 14 14:48:06 chaz systemd[1]: xo-server.service: Main process exited, code=killed, status=6/ABRT
      Oct 14 14:48:06 chaz systemd[1]: xo-server.service: Failed with result 'signal'.
      Oct 14 14:48:07 chaz systemd[1]: xo-server.service: Service RestartSec=100ms expired, scheduling restart.
      Oct 14 14:48:07 chaz systemd[1]: xo-server.service: Scheduled restart job, restart counter is at 1.
      Oct 14 14:48:07 chaz systemd[1]: Stopped XO Server.
      

      Transfer (Interrupted) started at 14:45:03.

      {
        "data": {
          "mode": "full",
          "reportWhen": "always"
        },
        "id": "1602711899558",
        "jobId": "c686003e-d71a-47f1-950d-ca57afe31923",
        "jobName": "Calculon-Onetime",
        "message": "backup",
        "scheduleId": "9d9c2499-b476-4e82-a9a5-505032da49a7",
        "start": 1602711899558,
        "status": "interrupted",
        "tasks": [
          {
            "data": {
              "type": "VM",
              "id": "8e56876a-e4cf-6583-d8de-36dba6dfad9e"
            },
            "id": "1602711899609",
            "message": "Starting backup of Calculon. (c686003e-d71a-47f1-950d-ca57afe31923)",
            "start": 1602711899609,
            "status": "interrupted",
            "tasks": [
              {
                "id": "1602711899614",
                "message": "snapshot",
                "start": 1602711899614,
                "status": "success",
                "end": 1602711902288,
                "result": "ae8968f8-01bb-2316-4d45-9a78c42c4e95"
              },
              {
                "id": "1602711902294",
                "message": "add metadata to snapshot",
                "start": 1602711902294,
                "status": "success",
                "end": 1602711902308
              },
              {
                "id": "1602711903334",
                "message": "waiting for uptodate snapshot record",
                "start": 1602711903334,
                "status": "success",
                "end": 1602711903545
              },
              {
                "id": "1602711903548",
                "message": "start VM export",
                "start": 1602711903548,
                "status": "success",
                "end": 1602711903571
              },
              {
                "data": {
                  "id": "8f721d6a-0d67-4c67-a860-e648bed3a458",
                  "type": "remote"
                },
                "id": "1602711903573",
                "message": "export",
                "start": 1602711903573,
                "status": "interrupted",
                "tasks": [
                  {
                    "id": "1602711905594",
                    "message": "transfer",
                    "start": 1602711905594,
                    "status": "interrupted"
                  }
                ]
              }
            ]
          }
        ]
      }
      
      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      Also running into the same "Interrupted" message, with VM's that are larger than 7+GB or so (this is not a hard limit, just what I've tested). I've tried with zstd compression and none, both are "interrupted."

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      @nraynaud

      130 GB Windows VM (before zstd) backup to Wasabi failed during the upload:

      Error: Part number must be an integer between 1 and 10000, inclusive
      

      A quick Google turned out something about multipart chunk sizes ...

      Anything I can do to help pinpoint this?

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      Update:

      Eliminating Directory via header edits actually didn't work. The backup/export seems to have completed, but the files don't exist. It may be a Wasabi thing, not the end of the world.

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      I just saw that! (and am playing with it).

      Quick question: The configurator seems to require a Directory underneath the bucket. What if we want to just touch the root of the bucket?

      (I'm not an S3 guru, so please ignore any ... ignorance).

      But, it seems I can put a dummy value in the directory, and eliminate it later via "header" edits. Not so much at the "detail" level.

      posted in Xen Orchestra
      K
      klou
    • RE: S3 / Wasabi Backup

      I also mounted wasabi via s3fs-fuse in fstab, but directly to xcp-ng.

      Most of my VM's were snapshotted, exported (with zstd), and correctly uploaded. However, 2 didn't, and I have this in XCP-ng Center:

      "File System on Control Domain Full","Disk usage for the Control Domain on server 'xcp-<hypervisor>' has reached 100.0%. XCP-ng's performance will be critically affected if this disk becomes full. Log files or other non-essential (user created) files should be removed.","xcp-<hypervisor>","Jul 26, 2020 2:28 AM",""
      

      Local filesystem is not even close to full, same as above.

      posted in Xen Orchestra
      K
      klou