XCP-ng
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    VM, missing disk

    Scheduled Pinned Locked Moved Xen Orchestra
    26 Posts 6 Posters 3.6k Views 4 Watching
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • DarkbeldinD Offline
      Darkbeldin Vates 🪐 Pro Support Team @MRisberg
      last edited by

      MRisberg You take no risk at trying it you can keep the VM and restore it with another name.

      1 Reply Last reply Reply Quote 0
      • M Offline
        MRisberg
        last edited by MRisberg

        While troubleshooting something else, backups also struggling, me and my friend found out that the host doesn't seem to coalesce successfully on the SR Stor 15K:

        Sep 22 22:27:16 ninja SMGC: [1630] In cleanup
        Sep 22 22:27:16 ninja SMGC: [1630] SR 35ec ('Stor 15K') (44 VDIs in 5 VHD trees):
        Sep 22 22:27:16 ninja SMGC: [1630]         *3fc6e297(40.000G/85.000K)
        Sep 22 22:27:16 ninja SMGC: [1630]         *4cce0cbb(100.000G/88.195G)
        Sep 22 22:27:16 ninja SMGC: [1630]         *81341eac(40.000G/85.000K)
        Sep 22 22:27:16 ninja SMGC: [1630]         *0377b389(40.000G/85.000K)
        Sep 22 22:27:16 ninja SMGC: [1630]         *b891bcb0(40.000G/85.000K)
        Sep 22 22:27:16 ninja SMGC: [1630]
        Sep 22 22:27:16 ninja SMGC: [1630] *~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*
        Sep 22 22:27:16 ninja SMGC: [1630]          ***********************
        Sep 22 22:27:16 ninja SMGC: [1630]          *  E X C E P T I O N  *
        Sep 22 22:27:16 ninja SMGC: [1630]          ***********************
        Sep 22 22:27:16 ninja SMGC: [1630] gc: EXCEPTION <class 'util.SMException'>, Parent VDI 6b13ce6a-b809-4a50-81b2-508be9dc606a of d36a5cb0-564a-4abb-920e-a6367741                            574a not found
        Sep 22 22:27:16 ninja SMGC: [1630]   File "/opt/xensource/sm/cleanup.py", line 3379, in gc
        Sep 22 22:27:16 ninja SMGC: [1630]     _gc(None, srUuid, dryRun)
        Sep 22 22:27:16 ninja SMGC: [1630]   File "/opt/xensource/sm/cleanup.py", line 3264, in _gc
        Sep 22 22:27:16 ninja SMGC: [1630]     _gcLoop(sr, dryRun)
        Sep 22 22:27:16 ninja SMGC: [1630]   File "/opt/xensource/sm/cleanup.py", line 3174, in _gcLoop
        Sep 22 22:27:16 ninja SMGC: [1630]     sr.scanLocked()
        Sep 22 22:27:16 ninja SMGC: [1630]   File "/opt/xensource/sm/cleanup.py", line 1606, in scanLocked
        Sep 22 22:27:16 ninja SMGC: [1630]     self.scan(force)
        Sep 22 22:27:16 ninja SMGC: [1630]   File "/opt/xensource/sm/cleanup.py", line 2357, in scan
        Sep 22 22:27:16 ninja SMGC: [1630]     self._buildTree(force)
        Sep 22 22:27:16 ninja SMGC: [1630]   File "/opt/xensource/sm/cleanup.py", line 2313, in _buildTree
        Sep 22 22:27:16 ninja SMGC: [1630]     "found" % (vdi.parentUuid, vdi.uuid))
        Sep 22 22:27:16 ninja SMGC: [1630]
        Sep 22 22:27:16 ninja SMGC: [1630] *~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*
        Sep 22 22:27:16 ninja SMGC: [1630] * * * * * SR 35ecc7ae-4e98-58c1-5ed3-2c22c649bd32: ERROR
        
        

        Maybe my problem is related to the host rather than XO? Back to troubleshooting .. and any input is appreciated.

        DarkbeldinD 1 Reply Last reply Reply Quote 0
        • DarkbeldinD Offline
          Darkbeldin Vates 🪐 Pro Support Team @MRisberg
          last edited by

          MRisberg That's mostly the issue apparently you lost a VDI from a chain there.
          So the fact that the disk is now seen as a snashop can come from there yes.
          But solving coalesce issue is not easy, how much do you love this VM?

          M 1 Reply Last reply Reply Quote 0
          • M Offline
            MRisberg @Darkbeldin
            last edited by MRisberg

            Darkbeldin
            I actually solved the coalesce issue yesterday. After removing a certain VDI (not related) and a coalesce lock file, SR Stor 15K managed to process the coalesce backlog. After that the backups started to process again.

            Due to poor performance I choose to postpone exporting or backing up the million VM. When I do, I'll see if I can restore it with both XVDA and XVDB.

            I'll copy the important contents from the million VM before I do anything. That way I can improvise and dare to move forward in anyway i need to - even if I'll have to reinstall the VM. But first I always try to make an attempt at resolving the problem rather than just throwing away and redo from start. That way I learn more about the platform.

            AnonabharA 1 Reply Last reply Reply Quote 0
            • AnonabharA Offline
              Anonabhar @MRisberg
              last edited by

              MRisberg Just wondering.. Where is the lock file? I might try that in the future for myself..

              DarkbeldinD M 2 Replies Last reply Reply Quote 0
              • DarkbeldinD Offline
                Darkbeldin Vates 🪐 Pro Support Team @Anonabhar
                last edited by

                Anonabhar Lock file should be at the root of the SR mount directory.

                1 Reply Last reply Reply Quote 0
                • M Offline
                  MRisberg @Anonabhar
                  last edited by MRisberg

                  Anonabhar

                  For me it was here:
                  On my (only) host: /run/sr-mount/35ecc7ae-4e98-58c1-5ed3-2c22c649bd32
                  (where 35ecc etc is one of my SRs)

                  There was a file named coalesce_ followed by a VDI UUID.

                  I moved this file out of the directory, rather than deleting it .. moving slowly. 🙂

                  1 Reply Last reply Reply Quote 0
                  • M Offline
                    MRisberg
                    last edited by

                    After export and import I can actually see a reference to both drives. Although they seem to be disconnected.

                    Screenshot 2022-09-24 mising disk 7.png

                    I'll see what I can do about it.

                    DanpD 1 Reply Last reply Reply Quote 0
                    • DanpD Offline
                      Danp Pro Support Team @MRisberg
                      last edited by

                      MRisberg That's normal behavior when the VM isn't running.

                      M 1 Reply Last reply Reply Quote 0
                      • M Offline
                        MRisberg @Danp
                        last edited by MRisberg

                        Danp

                        Thanks, yes of course. I got excited about fixing my problems and reported too early 🙂

                        Well, the VM has been working fine since the export / import. I can see both drives.

                        Although the VM running fine, since the last post I had to address two related VDIs that had been connected to the Control Domain. One being million_D and the other million_D (with a camera icon). These probably ended up here because of a aborted / faulty snapshot or backup. I disconnected them .. checked a few things ... and then forgot them. Everything is running fine now.

                        For future googlers: During the problem solving, I found out some pretty helpful info using tail -f /var/log/SMlog.


                        To me, this issue is solved. Thank you all for the feedback.

                        1 Reply Last reply Reply Quote 0
                        • S stevezemlicka referenced this topic on
                        • First post
                          Last post