XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    backup mail report says INTERRUPTED but it's not ?

    Scheduled Pinned Locked Moved Backup
    115 Posts 10 Posters 7.5k Views 12 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • M Offline
      MajorP93 @florent
      last edited by MajorP93

      After implementing the --max-old-space-size Node parameter as recommended by @pilow it took longer time for the VM to hit the issue.
      Still: backups went into interrupted status.
      Memory leak seems to be still there.
      With each subsequent backup run the memory usage rises and rises. After backup run the memory usage does not fully go back to "normal".

      6ad321a1-2e39-4bca-9285-062e502a17b2-grafik.png

      After adding the node parameter there was no heap size error on Node anymore since the heap size got increased. The system went into various OOM errors in kernel log (dmesg) despite not all RAM (8GB) being used.

      This is what htop looks like with 3 backup jobs running:
      68db77eb-26f2-4dbb-b1db-2273984eabb3-grafik.png

      1 Reply Last reply Reply Quote 0
      • A Offline
        acebmxer
        last edited by acebmxer

        While working last night i noticed one of my backups/pools did this. Got the email that it was interupted but when i looked the tasks were still running and moving data it untill it porcess all vms in that backup job.

        Edit - note my backup job was run via proxy on the specific pool/job.

        2026-02-19T03_00_00.028Z - backup NG.txt

        Edit 2 - homelab same last backup was interupted.

        2026-02-19T05_00_00.011Z - backup NG.txt

        M 1 Reply Last reply Reply Quote 0
        • M Offline
          MajorP93 @acebmxer
          last edited by MajorP93

          I wonder if this PR https://github.com/vatesfr/xen-orchestra/pull/9506 aims to solve the issue that was discussed in this thread.
          To me it looks like it's the case as the issue seems to be related to RAM used by backup jobs not being freed correctly and the PR seems to add some garbage collection to backup jobs.
          I hope that it will fix the issue and if needed I can test a branch.

          b-Nollet opened this pull request in vatesfr/xen-orchestra

          closed Backup tasks gc #9506

          Bastien NolletB 1 Reply Last reply Reply Quote 0
          • Bastien NolletB Offline
            Bastien Nollet Vates 🪐 XO Team @MajorP93
            last edited by

            Hi @MajorP93,

            This PR is only about changing the way we delete old logs (linked to a bigger work of making backups use XO tasks instead of their own task system), it won't fix the issue discussed in this topic.

            M 1 Reply Last reply Reply Quote 0
            • M Offline
              MajorP93 @Bastien Nollet
              last edited by

              Hi @Bastien-Nollet,

              oh okay, thanks for clarifying!

              F 1 Reply Last reply Reply Quote 0
              • P Pilow referenced this topic
              • F Offline
                flakpyro @MajorP93
                last edited by flakpyro

                I have been having the same issue and have been watching it for the last couple weeks. Initially my XOA only had 8GB of ram assigned, i have bumped it up to 16 to try an alleviate the issue. Seems to be some sort of memory leak. This is the official XO Appliance too not XO CE.

                I changed the systemd file to make use of the extra memory as per the docs,

                ExecStart=/usr/local/bin/node --max-old-space-size=12288 /usr/local/bin/xo-server
                

                It seems that over time it will just consume all of its memory until it crashes and restarts no matter how much i assign.

                d7c736e8-475a-401a-acce-e22d4c8688d7-image.png

                M 1 Reply Last reply Reply Quote 1
                • M Offline
                  MajorP93 @flakpyro
                  last edited by MajorP93

                  During the past month my backups failed (status interrupted) 1-2 times per week due to this memory leak.
                  When increasing heap size (node old space size) it takes longer but the backup fails when RAM usage eventually hits 100%.
                  I guess I’ll go with @Pilow ‘s workaround for now and create a cronjob for rebooting XO VM right before backups start.

                  1 Reply Last reply Reply Quote 0
                  • olivierlambertO Offline
                    olivierlambert Vates 🪐 Co-Founder CEO
                    last edited by

                    All of you are using a Node 24 LTS version? Do you still have the issue with Node 20 or Node 22?

                    M 1 Reply Last reply Reply Quote 0
                    • M Offline
                      MajorP93 @olivierlambert
                      last edited by MajorP93

                      @olivierlambert No.
                      I reverted to Node 20 as previously mentioned.
                      I was using Node 24 before but reverted to Node 20 as I hoped it would "fix" the issue.
                      Using Node 20 it takes longer for these issues to arise but in the end they arise.

                      3 of the users in this thread that encounter the issue said that they are using XOA.
                      As XOA also uses Node 20, I think most people that reported this issue actually use Node 20.

                      1 Reply Last reply Reply Quote 0
                      • olivierlambertO Offline
                        olivierlambert Vates 🪐 Co-Founder CEO
                        last edited by

                        Thanks for the recap!

                        F 1 Reply Last reply Reply Quote 0
                        • F Offline
                          flakpyro @olivierlambert
                          last edited by

                          @olivierlambert Using the prebuilt XOA appliance which reports:

                          [08:39 23] xoa@xoa:~$ node  --version
                          v20.18.3
                          
                          J 1 Reply Last reply Reply Quote 0
                          • J Offline
                            john.c @flakpyro
                            last edited by john.c

                            @flakpyro said in backup mail report says INTERRUPTED but it's not ?:

                            @olivierlambert Using the prebuilt XOA appliance which reports:

                            [08:39 23] xoa@xoa:~$ node  --version
                            v20.18.3
                            

                            @majorp93 @pilow Can you please capture some heap snapshots from during backup runs of XOA via NodeJS?

                            Then compare them to each other, they need to be in the following order:-

                            1. Snapshot before backup
                            2. Snapshot following first backup
                            3. Snapshot following second backup
                            4. Snapshot following third backup
                            5. Snapshot following subsequent backups to get to Node.js OOM (or as close as you’re willing to risk)

                            These will require that XOA (or XOCE) is started with Node.js heap snapshots enabled. Then open in a Chromium based browser the following url:-
                            chrome://inspect

                            The above URL will require using the browser’s DevTools features!

                            Another option is to integrate and enable use of Clinic.js (clinic heapprofiler), or configure node to use node-heapdump when it reaches a threshold amount.

                            Once your got those heap dumps your looking for the following:-

                            • Object types that grow massively between the snapshots.
                            • Large arrays or maps of backup-related objects (VMs, snapshots, jobs, tasks, etc.).
                            • Retained objects whose “retainers” point to long-lived structures (global, caches, singletons).

                            These will likely help to pin down what and where in the backup code, the memory leak is located.

                            Once have these a heap snapshot diff showing which object type (or types) growing by a stated size per backup will finally help the Vates developers fix this issue.

                            @florent I left the above for the original reporters of the memory leak issue, and/or yourselves.

                            M 1 Reply Last reply Reply Quote 0
                            • M Offline
                              MajorP93 @john.c
                              last edited by

                              @john.c If a Xen Orchestra developer asks for logs / heap snapshots I will be happy to provide them.
                              Ideally they should tell us what is needed from us to debug this.

                              That being said I am not entirely sure they are currently working on this as XO team has not been very vocal here.

                              Looking at Github repository commits indicates that other things are currently being prioritized over this crucial backup stability issue.

                              1 Reply Last reply Reply Quote 0
                              • florentF Offline
                                florent Vates 🪐 XO Team
                                last edited by florent

                                we are trying to reproduce it here, without success for now
                                It would be good for us too if we can identify what cause this before the upcoming release ( thursday )

                                for example here my xoa at home , seems quite ok running by itself :

                                68cb6a77-fc52-4a8a-91e6-88c46d0f55cd-image.png

                                to be fair the debug is not helped by the fact that we have one big process, with so much object that changes that exploring and pinpointing exactly what is growing is tricky ( or we may not look at the right place)

                                P 1 Reply Last reply Reply Quote 0
                                • P Offline
                                  Pilow @florent
                                  last edited by

                                  @florent do you have (heavy?) use of backup jobs on your XOA ?
                                  we all noticed this because of correlation between backup jobs & memory growth

                                  I do have all my backup jobs through XO PROXIES though, and XOA still explodes in RAM Usage...

                                  MathieuRAM florentF 2 Replies Last reply Reply Quote 1
                                  • MathieuRAM Offline
                                    MathieuRA Vates 🪐 XO Team @Pilow
                                    last edited by MathieuRA

                                    Could you try disabling XO6?
                                    In a configuration file, enable XO5 as the default version and ensure that XO6 is inaccessible (for example, by using an invalid path or removing the corresponding entry in the configuration). This will help determine if the issue is related to XO6/REST API.

                                    https://docs.xen-orchestra.com/configuration#using-xo-5-as-the-default-interface

                                    1 Reply Last reply Reply Quote 0
                                    • florentF Offline
                                      florent Vates 🪐 XO Team @Pilow
                                      last edited by

                                      @Pilow said in backup mail report says INTERRUPTED but it's not ?:

                                      @florent do you have (heavy?) use of backup jobs on your XOA ?
                                      we all noticed this because of correlation between backup jobs & memory growth

                                      I do have all my backup jobs through XO PROXIES though, and XOA still explodes in RAM Usage...

                                      I back up a only few VM on this one , it was to illustrate that we try to reproduce it with a scope small enough to improve the situation

                                      is the memory stable on the proxy ?

                                      M P 2 Replies Last reply Reply Quote 0
                                      • M Offline
                                        MajorP93 @florent
                                        last edited by MajorP93

                                        @florent In my case I have 4 backup jobs. 1 full, 2 delta (1 onsite, 1 offsite), 1 mirror.
                                        106 VMs are being backed up by these jobs.
                                        In my case that setup clearly shows the RAM issue. I do not know if it can be replicated in a very small setup.
                                        Sometimes multiple backup jobs run at the same time. It appears that backup jobs running in parallel make the issue occur faster.

                                        @mathieura Thanks for the hint! I can imagine that this could possibly change something in this regard as I did not have any issues with this exact backup job configuration some time ago. The issues started to appear somewhere in december so it could be related to the XO6 release. I will try what you suggested and report back.

                                        1 Reply Last reply Reply Quote 0
                                        • P Offline
                                          ph7
                                          last edited by

                                          I also have this problem in my XO-CE (ronivay-script) at home
                                          I get the mail report after 4 days
                                          A reboot resets the memory

                                          the XO-CE have 3.1 GB and the Control domain memory have 2 GB
                                          Node v24.13.1
                                          Running Continuous Replication and Delta Backups

                                          1 Reply Last reply Reply Quote 0
                                          • P Offline
                                            Pilow @florent
                                            last edited by

                                            @florent said in backup mail report says INTERRUPTED but it's not ?:

                                            is the memory stable on the proxy ?

                                            can't really say, as i reboot them everyday with my XOA
                                            4cfc8e9e-2334-447d-a073-56c5362cd27a-image.png

                                            but when I do not reboot them, they dont seem to be so impacted as I need to reboot them.
                                            I will detach them from the reboot job and will report RAM consumption on proxies

                                            florentF 1 Reply Last reply Reply Quote 0

                                            Hello! It looks like you're interested in this conversation, but you don't have an account yet.

                                            Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.

                                            With your input, this post could be even better 💗

                                            Register Login
                                            • First post
                                              Last post