XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    S3 Chunk Size

    Scheduled Pinned Locked Moved Backup
    12 Posts 4 Posters 269 Views 4 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • florentF Offline
      florent Vates 🪐 XO Team @rizaemet 0
      last edited by

      @rizaemet-0 this is not configurable .
      before qcow2 the S3 blocks were aligned with the VHD blocks, then compressed, so changing it would have been a lot of work
      now with qcow2 support almost reaching stable, we have some of the tools available to resize the block size on the file ( since qcow2 are based on 64KB blocks) , so it's not impossible, but it's still quite complex to do reliably

      Could you name the provider ?

      1 Reply Last reply Reply Quote 0
      • rizaemet 0R Offline
        rizaemet 0
        last edited by

        @florent We are a university in Türkiye. The S3 service is provided to us by the National Academic Network and Information Center (ULAKBİM), an official institution of Türkiye.

        1 Reply Last reply Reply Quote 0
        • olivierlambertO Online
          olivierlambert Vates 🪐 Co-Founder CEO
          last edited by

          Do you know what software are they using? Ceph? Minio? Garage?

          rizaemet 0R 1 Reply Last reply Reply Quote 0
          • rizaemet 0R Offline
            rizaemet 0 @olivierlambert
            last edited by rizaemet 0

            @olivierlambert Ceph. There are a few configuration examples. That's how I learned it's Ceph.

            Edit: When I asked the AI ​​some questions, it said something like this: "If the chunk size is too small → the risk of a 502 increase". Seeing this, I ran a few tests. A backup of a virtual machine with 80 GB of disk space (backup size: 70 GB) went through without any problems. However, a backup of a virtual machine with 16 GB of disk space (backup size: 3 GB) resulted in a failure. It seems the 502 error occurred during the clean-vm phase of the backup. However, the backup appears to have been created and it was working when I restored. I was backing up virtual machines with large disk sizes and had never encountered this error before.
            This section in the log exists both before the snapshot and after the export:

            ...
            {
              "id": "1768672615028",
              "message": "clean-vm",
              "start": 1768672615028,
              "status": "failure",
              "end": 1768673350279,
              "result": {
                "$metadata": {
                  "httpStatusCode": 502,
                  "clockSkewCorrected": true,
                  "attempts": 3,
                  "totalRetryDelay": 112
                },
                "message": "Expected closing tag 'hr' (opened in line 9, col 1) instead of closing tag 'body'.:11:1
              Deserialization error: to see the raw response, inspect the hidden field {error}.$response on this object.",
                "name": "Error",
                "stack": "Error: Expected closing tag 'hr' (opened in line 9, col 1) instead of closing tag 'body'.:11:1
              Deserialization error: to see the raw response, inspect the hidden field {error}.$response on this object.
                at st.parse (/opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/fast-xml-parser/lib/fxp.cjs:1:20727)
                at parseXML (/opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/@aws-sdk/xml-builder/dist-cjs/xml-parser.js:17:19)
                at /opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/@aws-sdk/core/dist-cjs/submodules/protocols/index.js:1454:52
                at process.processTicksAndRejections (node:internal/process/task_queues:103:5)
                at async parseXmlErrorBody (/opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/@aws-sdk/core/dist-cjs/submodules/protocols/index.js:1475:17)
                at async de_CommandError (/opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/@aws-sdk/client-s3/dist-cjs/index.js:5154:11)
                at async /opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/@smithy/middleware-serde/dist-cjs/index.js:8:24
                at async /opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/@aws-sdk/middleware-sdk-s3/dist-cjs/index.js:488:18
                at async /opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/@smithy/middleware-retry/dist-cjs/index.js:254:46
                at async /opt/xo/xo-builds/xen-orchestra-202601171930/node_modules/@aws-sdk/middleware-flexible-checksums/dist-cjs/index.js:318:18"
              }
            }
            ...
            
            florentF 1 Reply Last reply Reply Quote 0
            • florentF Offline
              florent Vates 🪐 XO Team @rizaemet 0
              last edited by

              @rizaemet-0 the cleanVM is the most demanding part of the backup job (mostly listing, moving and deleting blocks )

              1 Reply Last reply Reply Quote 1
              • olivierlambertO Online
                olivierlambert Vates 🪐 Co-Founder CEO
                last edited by

                Yeah, it's not surprising. Ceph S3 implementation is known to be "average" and failure-prone in those cases.

                J 1 Reply Last reply Reply Quote 0
                • rizaemet 0R Offline
                  rizaemet 0
                  last edited by

                  Could you please share which version of aws-sdk Xen-Orchestra is currently using? I will share this with our S3 service provider.

                  florentF 1 Reply Last reply Reply Quote 0
                  • florentF Offline
                    florent Vates 🪐 XO Team @rizaemet 0
                    last edited by

                    @rizaemet-0
                    sure : "@aws-sdk/client-s3": "^3.54.0",

                    1 Reply Last reply Reply Quote 0
                    • J Offline
                      john.c @olivierlambert
                      last edited by john.c

                      @olivierlambert said in S3 Chunk Size:

                      Yeah, it's not surprising. Ceph S3 implementation is known to be "average" and failure-prone in those cases.

                      @olivierlambert @florent Well you’re likely to see more use of Ceph by providers, following MinIO entering maintenance mode. Also Canonical are going to be doing more development and selling what it calls Micro Ceph. It’s blog post gives more details.

                      https://ubuntu.com/blog/microceph-why-its-the-superior-minio-alternative

                      florentF 1 Reply Last reply Reply Quote 0
                      • florentF Offline
                        florent Vates 🪐 XO Team @john.c
                        last edited by

                        @john.c yes for sure, and maybe we will be able to set the chunk size on a future date ( especially since we did some of the ground work for vhd / qcow2)

                        1 Reply Last reply Reply Quote 0
                        • olivierlambertO Online
                          olivierlambert Vates 🪐 Co-Founder CEO
                          last edited by

                          I heard better feedback from Garage or RustFS than Ceph for a successor after Minio.

                          1 Reply Last reply Reply Quote 0
                          • First post
                            Last post