Categories

  • All news regarding Xen and XCP-ng ecosystem

    142 Topics
    4k Posts
    bvitnikB
    @benapetr just a small notice, at least in 0.0.5 version. When adding an NFS ISO library SR, there is no choice for NFS version (v3 and v4) like in XCP-ng Center. The XAPI then defaults to v3 which did not work in my environment.
  • Everything related to the virtualization platform

    1k Topics
    15k Posts
    A
    That is a frustrating loop to be in, especially with TrueNAS. Usually, when the VM fails to start after a change, it’s because XCP-ng is trying to pass through a PCI device (like an HBA) that isn't being released properly by the host. Have you checked if the "hide" parameters in your grub config are still correct? Sometimes an update can reset those, and the host grabs the controller before the VM can. Another thing to try is toggling the BIOS/UEFI mode in the VM settings - TrueNAS can be picky about that depending on which version you’re running.
  • 3k Topics
    28k Posts
    P
    Re: FILE RESTORE / overlapping loop device exists we still had this annoying probleme where we couldn't FLR our backup, source VM could be Windows or linux, with or without LVM... (see error logs in previous thread, overlapping loop...) latest XOA is installed. on a secondaray XO CE, I mounted the same remote, and NO PROBLEMS AT ALL... soooo I thought my main XOA is probably beaten up somewhere... but that was not it. all remotes on main XOA are attached to a XO PROXY (official, not CE) ! on the XO CE that is in another datacenter, I mounted the S3 remotes directly WITHOUT the proxy use, and it worked. On main XOA, we offload all backup tasks to proxies, so all remotes are attached to a proxy... I duplicated a remote, same conf, but without proxy, so that XOA would do the restore job, not the proxy, and voila, all is working as intended, FLR is okay on main datacenter. TLDR : FLR is not working with proxy attached remotes, if you duplicate the remote without the XO PROXY, all good @florent @bastien-nollet I think there is something to look into. probably a race condition, the XOA and the XO PROXY both try to mount the loop device ?
  • Our hyperconverged storage solution

    44 Topics
    731 Posts
    olivierlambertO
    Different use cases: Ceph is better with more hosts (at least 6 or 7 minimum) while XOSTOR is better between 3 to 7/8. We might have better Ceph support in the future for large clusters.
  • 34 Topics
    101 Posts
    B
    @AtaxyaNetwork Merci becuoup pour ce retour. Je vais compléter l'article dans ce sens.