<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0"><channel><title><![CDATA[Switching to XCP-NG, want to hear your problems]]></title><description><![CDATA[<p dir="auto">This is a duplicate of my reddit post so I apologize. The XCP-NG subreddit doesn't seem to have a ton of traffic so I wanted to spread this post a bit wider.</p>
<p dir="auto"><a href="https://www.reddit.com/r/xcpng/comments/1czm1oi/switching_to_xcpng_want_to_hear_your_problems/" target="_blank" rel="noopener noreferrer nofollow ugc">https://www.reddit.com/r/xcpng/comments/1czm1oi/switching_to_xcpng_want_to_hear_your_problems/</a></p>
<p dir="auto">My org is strongly considering switching to XCP-NG as our hypervisor ASAP. We need to make a decision on this in the next couple weeks. Our environment is probably medium-sized, with over 1000 VMs. I've done a lot of testing and it seems like a good product so far. We're worried about what we might be missing, though.</p>
<p dir="auto">What issues have you encountered? What has thrown a serious wrench in your plans or been hard to deal with?</p>
<p dir="auto">A couple things I've run into that I'm curious to hear about more:</p>
<ol>
<li>
<p dir="auto">Seems there are issues with Secure Boot on Windows: <a href="https://docs.xcp-ng.org/guides/guest-UEFI-Secure-Boot/#:~:text=Enabling%20Secure%20Boot%20for%20the,Citrix%20work%20with%20Secure%20Boot" target="_blank" rel="noopener noreferrer nofollow ugc">https://docs.xcp-ng.org/guides/guest-UEFI-Secure-Boot/#:~:text=Enabling Secure Boot for the,Citrix work with Secure Boot</a>. It sounds like we need to have Citrix guest tools installed? I've done that but Secure Boot still fails on Windows VMs. Or are the "Citrix VP drivers" a different thing that I need to look into?</p>
</li>
<li>
<p dir="auto">We use iSCSI with SANs for storage. This doesn't support this provision in XCP-NG, which we absolutely need. Our SANs do support thin provision. In my limited testing the SAN thin provision seems to work, but I'm wondering if anybody has used this long term. If we start running out of storage in 1 year because relying on Dell SAN thin provision isn't good, that's not going to be pretty.</p>
</li>
<li>
<p dir="auto">RAM access performance of VMs seems to be much lower on XCP-NG compared to ESXi and Hyper-V (over 10x slower when testing via sysbench. Windows doesn't show this speed decrease). Overall, XCP-NG performance seems mostly on par with ESXi (save a couple test cases), so I'm not really worried about that, but wondering if anybody knows why RAM access is so much slower.</p>
</li>
<li>
<p dir="auto">Importing VMs from ESXi via Xen Orchestra works, but it is very slow and the live import is broken. StarWind has a converter that works with some other hypervisors, and it allows you to import a VM while it's running through snapshots. This feature seems to be broken in Xen Orchestra, and would be very nice to have.</p>
</li>
</ol>
]]></description><link>https://xcp-ng.org/forum/topic/9093/switching-to-xcp-ng-want-to-hear-your-problems</link><generator>RSS for Node</generator><lastBuildDate>Fri, 17 Apr 2026 03:00:06 GMT</lastBuildDate><atom:link href="https://xcp-ng.org/forum/topic/9093.rss" rel="self" type="application/rss+xml"/><pubDate>Fri, 24 May 2024 18:14:15 GMT</pubDate><ttl>60</ttl><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 14 Sep 2024 10:56:07 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> said in <a href="/forum/post/82625">Switching to XCP-NG, want to hear your problems</a>:</p>
<blockquote>
<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/codemercenary" aria-label="Profile: CodeMercenary">@<bdi>CodeMercenary</bdi></a> I am using V4 on the XO-Server to our backup remotes and it seems to work just fine. However using V4 as a storage SR was nothing but problems, as <a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/nikade" aria-label="Profile: nikade">@<bdi>nikade</bdi></a> mentioned we had tons of NFS Server not responding issues which would lock up hosts and VMs causing downtime. Since moving to v3 that hasn't happened.</p>
<p dir="auto">Checking a host's NFS retransmissions stats after 9 days of uptime i see we have had some retransmissions but they have not caused any downtime or even any timeout messages to appear in dmesg on the host.</p>
<pre><code>[xcpng-prd-02 ~]# nfsstat -rc
Client rpc stats:
calls      retrans    authrefrsh
268513028   169        268537542

</code></pre>
<p dir="auto">From what a gather from this blog post from redhat (<a href="https://www.redhat.com/sysadmin/using-nfsstat-nfsiostat" target="_blank" rel="noopener noreferrer nofollow ugc">https://www.redhat.com/sysadmin/using-nfsstat-nfsiostat</a>) it seems like that amount of retransmissions is VERY low and not an issue.</p>
</blockquote>
<p dir="auto">Thats fine, we've got a lot more and I haven't seen any "nfs server not responding" in dmesg yet.<br />
Using NFS v3 for both SR and backups now for a couple of years and it's been great, I think I had issues once or twice in like 5-6 years on the backup SR where the vhd file got locked by dom0, Vates helped out there as always and it was resolved quickly.</p>
]]></description><link>https://xcp-ng.org/forum/post/82634</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/82634</guid><dc:creator><![CDATA[nikade]]></dc:creator><pubDate>Sat, 14 Sep 2024 10:56:07 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Fri, 13 Sep 2024 17:13:32 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/codemercenary" aria-label="Profile: CodeMercenary">@<bdi>CodeMercenary</bdi></a> I am using V4 on the XO-Server to our backup remotes and it seems to work just fine. However using V4 as a storage SR was nothing but problems, as <a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/nikade" aria-label="Profile: nikade">@<bdi>nikade</bdi></a> mentioned we had tons of NFS Server not responding issues which would lock up hosts and VMs causing downtime. Since moving to v3 that hasn't happened.</p>
<p dir="auto">Checking a host's NFS retransmissions stats after 9 days of uptime i see we have had some retransmissions but they have not caused any downtime or even any timeout messages to appear in dmesg on the host.</p>
<pre><code>[xcpng-prd-02 ~]# nfsstat -rc
Client rpc stats:
calls      retrans    authrefrsh
268513028   169        268537542

</code></pre>
<p dir="auto">From what a gather from this blog post from redhat (<a href="https://www.redhat.com/sysadmin/using-nfsstat-nfsiostat" target="_blank" rel="noopener noreferrer nofollow ugc">https://www.redhat.com/sysadmin/using-nfsstat-nfsiostat</a>) it seems like that amount of retransmissions is VERY low and not an issue.</p>
]]></description><link>https://xcp-ng.org/forum/post/82625</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/82625</guid><dc:creator><![CDATA[flakpyro]]></dc:creator><pubDate>Fri, 13 Sep 2024 17:13:32 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Fri, 13 Sep 2024 16:51:32 GMT]]></title><description><![CDATA[<p dir="auto">Interesting to know that v3 seems to be more reliable that v4. I had repeated problems with using NFS for a backup remote and those problems only went away when I changed the remotes to use SMB. I know NFS would be better to use but a backup that happens through an inferior protocol is way better than one that fails using a better protocol. <img src="https://xcp-ng.org/forum/assets/plugins/nodebb-plugin-emoji/emoji/android/1f642.png?v=a78c449d9ac" class="not-responsive emoji emoji-android emoji--slightly_smiling_face" style="height:23px;width:auto;vertical-align:middle" title=":-)" alt="🙂" /></p>
<p dir="auto">Maybe I should give NFS another chance but force it to use v3.</p>
<p dir="auto">In my case, I'd get backups working on NFS and then several days later a backup would fail. Then backups fail every day until I intervene, usually by rebooting the XO VM. Sometimes I'd then have to do cleanup, like releasing a VDI or something. Then it may or may not start working again but if it did start working I'd have another failure a few days later. It's been 2.5 weeks since I switched it to SMB and have had no failures. That's definitely the longest I've gone without a failure from a delta backup to a networked drive.</p>
<p dir="auto">Note, I also have backups going to a local drive mounted in XO so with all those remote failures I always had a clean backup somewhere. This was in the process of trying to decide if I could trust sending delta backups to a network remote rather than using full backups to a local remote. My initial feelings were that the delta backups didn't work reliably but now I believe the issue was with NFS, not with deltas specifically.</p>
]]></description><link>https://xcp-ng.org/forum/post/82623</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/82623</guid><dc:creator><![CDATA[CodeMercenary]]></dc:creator><pubDate>Fri, 13 Sep 2024 16:51:32 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Fri, 13 Sep 2024 10:56:51 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> said in <a href="/forum/post/82598">Switching to XCP-NG, want to hear your problems</a>:</p>
<blockquote>
<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/rfx77" aria-label="Profile: rfx77">@<bdi>rfx77</bdi></a> As a follow up v4 did cause more downtime for us, i switched over everything to v3 which has been much better so far. Going to be curious to see how a controller failover goes during a firmware update as v3 is stateless vs v4 and iscsi being stateful protocols.</p>
</blockquote>
<p dir="auto">We're also using v3 and failover on our Dell Powerstore's are seamless.<br />
Haven't tried v4 since we had A LOT of "nfs server not responding" issues with it and immediately went back to v3.</p>
]]></description><link>https://xcp-ng.org/forum/post/82609</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/82609</guid><dc:creator><![CDATA[nikade]]></dc:creator><pubDate>Fri, 13 Sep 2024 10:56:51 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Fri, 13 Sep 2024 05:38:23 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/rfx77" aria-label="Profile: rfx77">@<bdi>rfx77</bdi></a> As a follow up v4 did cause more downtime for us, i switched over everything to v3 which has been much better so far. Going to be curious to see how a controller failover goes during a firmware update as v3 is stateless vs v4 and iscsi being stateful protocols.</p>
]]></description><link>https://xcp-ng.org/forum/post/82598</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/82598</guid><dc:creator><![CDATA[flakpyro]]></dc:creator><pubDate>Fri, 13 Sep 2024 05:38:23 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Mon, 19 Aug 2024 07:02:08 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a></p>
<p dir="auto">Our Issue with commvault was that despite it mounts the snapshots in the proxy VM it has to read them as a whole. So Inc Backups take nearly as long as Full backups. When you want to backup 20TB+ Incs every night, Thats not acceptable. So the Lack of CBT is really a big deal when doing VM Backups.</p>
<p dir="auto">Performance with CommVault would be not that big of a problem since you can do multiple streams and multiple VMs simultanously but we could not get past 600MB/s which is a Xen Problem as of our Testing. With HyperV we see Performance in access of 1.6GB/s in the same Scenarios. But also this performance would be much to slow to do Incs every night.</p>
<p dir="auto">We tested XO CBT but it was not stable and for what i read in the posts there seems to be problems with CBT and Live-Migration so that the CBT state seems to be lost. Also a big Problem for us.</p>
<p dir="auto">When we used CommVault to backup multiple VMs we ran into Blue-Screens of the Xen Toolstack in the Windows VMs when attaching or detaching Snapshot-VDIs to the Proxy VMs. There clearly is a bug in the Xen Windows Drivers. So we had to reduce concurrency which reduced backup speed.</p>
<p dir="auto">We didn't do a short test with xen, we migrated out internal production Cluster to it from VMWare and used it for about 4 month now (30+ VMs on iSCSI Storage with 3 Nodes) and after we ran into more and more Problems we had many discussions with our team and we had to ask ourselfes what Xen brings to the table that is worth the drawbacks. The only scenario where it fits for us is where we have to map physical hardware into VMs.</p>
<p dir="auto">We decided to come back in some time to see how the SMAPIv3 drivers work out and if there is a better support for shared storage.</p>
]]></description><link>https://xcp-ng.org/forum/post/81692</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81692</guid><dc:creator><![CDATA[rfx77]]></dc:creator><pubDate>Mon, 19 Aug 2024 07:02:08 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sun, 18 Aug 2024 19:22:01 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> I never  tried commvault so I cant really tell, but I've tried Acronis and Quadric but performance wasn't too great, atleast not better than XOA so there was no point.</p>
]]></description><link>https://xcp-ng.org/forum/post/81682</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81682</guid><dc:creator><![CDATA[nikade]]></dc:creator><pubDate>Sun, 18 Aug 2024 19:22:01 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sun, 18 Aug 2024 17:07:58 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/nikade" aria-label="Profile: nikade">@<bdi>nikade</bdi></a> Yeah Veeam was very set and forget, we backed up around 100 VMs a night locally and to our DR site and it just reliably worked and i never really had to think about it. Once CBT stabilizes in XCP-NG i think that will go a long way in helping but i don't think its quite production ready yet.</p>
<p dir="auto">Im hoping to eventually get to the same point with XCP-NG, be it with XOA backups or with something like commvault. Our NFS mounts last dropped Thursday night during backups and have been fine since so its VERY intermittent. I think if it happens again i will begin moving to NFS3 backed SRs and hope that solves it.</p>
<p dir="auto">What issues with commvault did you run into due to its lack of native Xen CBT? Was their own internal change block tracking not reliable? I have a feeling the xen portion of the product does not see a lot of development attention from reading their documentation.</p>
]]></description><link>https://xcp-ng.org/forum/post/81674</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81674</guid><dc:creator><![CDATA[flakpyro]]></dc:creator><pubDate>Sun, 18 Aug 2024 17:07:58 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sun, 18 Aug 2024 15:52:05 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> said in <a href="/forum/post/81669">Switching to XCP-NG, want to hear your problems</a>:</p>
<blockquote>
<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/rfx77" aria-label="Profile: rfx77">@<bdi>rfx77</bdi></a> Makes sense. I was thinking of giving the Commvault trial a try with XCP-NG since it looked like while they don't use Xens' CBT, they still tracked changed blocks within their helper VM while still doing dedupe.</p>
<p dir="auto">Backups have been the biggest set back in our move from VMware, i knew going in i would miss Veeam more than i'd miss Vmware itself.</p>
<p dir="auto">As for NFS 4.1 vs 3, if the timeouts return this week i think i will give v3 a try if it worked more reliability for you.</p>
</blockquote>
<p dir="auto">Yeah, Veeam really is the king of backups. We're backing up about 50 VM's with Veeam from our vmware clusters and man it is sooo fast and reliable, i've seen backups go at 7Gbit/s which is incredible.</p>
]]></description><link>https://xcp-ng.org/forum/post/81672</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81672</guid><dc:creator><![CDATA[nikade]]></dc:creator><pubDate>Sun, 18 Aug 2024 15:52:05 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sun, 18 Aug 2024 15:50:28 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> said in <a href="/forum/post/81635">Switching to XCP-NG, want to hear your problems</a>:</p>
<blockquote>
<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/nikade" aria-label="Profile: nikade">@<bdi>nikade</bdi></a> I am not using any custom mount options other than "hard" to do a hard nfs mount to prevent data lose when drops like these happen.</p>
<p dir="auto">Did you have to use custom mount options with V3 as well then or just with V4? I may try moving VMs over to a V3 mount from V4 to see if that helps stabilize things.</p>
</blockquote>
<p dir="auto">In XOA we got some mount options from <a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/yannik" aria-label="Profile: yannik">@<bdi>yannik</bdi></a> but in XCP we have not had to use any special options when mounting the NFS SR (as long as we're using NFS 3).</p>
]]></description><link>https://xcp-ng.org/forum/post/81671</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81671</guid><dc:creator><![CDATA[nikade]]></dc:creator><pubDate>Sun, 18 Aug 2024 15:50:28 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sun, 18 Aug 2024 15:19:47 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/rfx77" aria-label="Profile: rfx77">@<bdi>rfx77</bdi></a> Makes sense. I was thinking of giving the Commvault trial a try with XCP-NG since it looked like while they don't use Xens' CBT, they still tracked changed blocks within their helper VM while still doing dedupe.</p>
<p dir="auto">Backups have been the biggest set back in our move from VMware, i knew going in i would miss Veeam more than i'd miss Vmware itself.</p>
<p dir="auto">As for NFS 4.1 vs 3, if the timeouts return this week i think i will give v3 a try if it worked more reliability for you.</p>
]]></description><link>https://xcp-ng.org/forum/post/81669</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81669</guid><dc:creator><![CDATA[flakpyro]]></dc:creator><pubDate>Sun, 18 Aug 2024 15:19:47 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sun, 18 Aug 2024 14:14:42 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a><br />
To be honest we gave up on Xen for our systems where we have shared storage. We swiched to HyperV.</p>
<p dir="auto">The Lack of CBT Support in CommVault was a major problem. We cannot use XO for backups because it lacks 90% of the features we now have with CommVault (Dedup, Tape, Agents, IntelliSnap of VM disks, AUX Copys,..) and we dont want to combine it with CommVault agents since this complicates our Backups and dies not make real sense.</p>
<p dir="auto">HyperV with Clustering is free in our setup so it was not a big discussion there. We also can utilize the full potential (performance and feature wise) of our SAN and Network which always was a major issue with Xen.</p>
<p dir="auto">We keep XCP-NG for now on some specialized installs where we need to map physical hardware into a vm.</p>
]]></description><link>https://xcp-ng.org/forum/post/81666</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81666</guid><dc:creator><![CDATA[rfx77]]></dc:creator><pubDate>Sun, 18 Aug 2024 14:14:42 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 20:06:11 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/nikade" aria-label="Profile: nikade">@<bdi>nikade</bdi></a> I am not using any custom mount options other than "hard" to do a hard nfs mount to prevent data lose when drops like these happen.</p>
<p dir="auto">Did you have to use custom mount options with V3 as well then or just with V4? I may try moving VMs over to a V3 mount from V4 to see if that helps stabilize things.</p>
]]></description><link>https://xcp-ng.org/forum/post/81635</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81635</guid><dc:creator><![CDATA[flakpyro]]></dc:creator><pubDate>Sat, 17 Aug 2024 20:06:11 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 19:47:35 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> said in <a href="/forum/post/81632">Switching to XCP-NG, want to hear your problems</a>:</p>
<blockquote>
<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/nikade" aria-label="Profile: nikade">@<bdi>nikade</bdi></a> Interesting to hear we are not the only ones experiencing timeouts with V4 We are on NFS 4.1 with the Pure array on the latest firmware. I initially wanted to use v4 as it is a stateful protocol and thought it may handle controller failovers better due to that. Perhaps i should try V3 and see if it fairs better. We had okay luck with V4 on TrueNAS but never really ran it under any extreme load. It will run for 2 or 3 days without issue then suddenly NFS drops appear in dmesg on the hosts.</p>
<p dir="auto">For example:</p>
<pre><code>dmesg -T from the host shows the following with 10.174.199.25 being the array. 

[Thu Aug 15 01:28:30 2024] nfs: server 10.174.199.25 not responding, still trying
[Thu Aug 15 01:28:30 2024] nfs: server 10.174.199.25 not responding, still trying
[Thu Aug 15 01:28:30 2024] nfs: server 10.174.199.25 not responding, still trying
[Thu Aug 15 01:28:30 2024] nfs: server 10.174.199.25 not responding, still trying

Followed by recovery after some time:

[Thu Aug 15 01:29:49 2024] nfs: server 10.174.199.25 OK
[Thu Aug 15 01:29:49 2024] nfs: server 10.174.199.25 OK
[Thu Aug 15 01:29:49 2024] nfs: server 10.174.199.25 OK
[Thu Aug 15 01:29:49 2024] nfs: server 10.174.199.25 OK

</code></pre>
<p dir="auto">Is that what you were seeing on your powerstore with V4 as well?</p>
</blockquote>
<p dir="auto">We actually had some timeouts like these on our "slower" TrueNAS, we got help from <a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/yannik" aria-label="Profile: yannik">@<bdi>yannik</bdi></a> I think to tweak the mount options when our backups failed.<br />
Ever since they tweaked the mount options it has been rock solid and we havent seen any timeouts in dmesg.</p>
]]></description><link>https://xcp-ng.org/forum/post/81634</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81634</guid><dc:creator><![CDATA[nikade]]></dc:creator><pubDate>Sat, 17 Aug 2024 19:47:35 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 19:46:00 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> Thanks for replying!<br />
Yeah we've not had much luck with the NFS 4... but we've been using NFS 3 for years (Since 2017 I think) and it has been rock solid, failover between controllers are not even causing a single timeout in dmesg on the hosts.</p>
<p dir="auto">I've tried to tweak the mount options, but I didn't have much luck, thats the reason why I asked if you had played around with it.</p>
]]></description><link>https://xcp-ng.org/forum/post/81633</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81633</guid><dc:creator><![CDATA[nikade]]></dc:creator><pubDate>Sat, 17 Aug 2024 19:46:00 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 19:44:58 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/nikade" aria-label="Profile: nikade">@<bdi>nikade</bdi></a> Interesting to hear we are not the only ones experiencing timeouts with V4 We are on NFS 4.1 with the Pure array on the latest firmware. I initially wanted to use v4 as it is a stateful protocol and thought it may handle controller failovers better due to that. Perhaps i should try V3 and see if it fairs better. We had okay luck with V4 on TrueNAS but never really ran it under any extreme load. It will run for 2 or 3 days without issue then suddenly NFS drops appear in dmesg on the hosts.</p>
<p dir="auto">For example:</p>
<pre><code>dmesg -T from the host shows the following with 10.174.199.25 being the array. 

[Thu Aug 15 01:28:30 2024] nfs: server 10.174.199.25 not responding, still trying
[Thu Aug 15 01:28:30 2024] nfs: server 10.174.199.25 not responding, still trying
[Thu Aug 15 01:28:30 2024] nfs: server 10.174.199.25 not responding, still trying
[Thu Aug 15 01:28:30 2024] nfs: server 10.174.199.25 not responding, still trying

Followed by recovery after some time:

[Thu Aug 15 01:29:49 2024] nfs: server 10.174.199.25 OK
[Thu Aug 15 01:29:49 2024] nfs: server 10.174.199.25 OK
[Thu Aug 15 01:29:49 2024] nfs: server 10.174.199.25 OK
[Thu Aug 15 01:29:49 2024] nfs: server 10.174.199.25 OK

</code></pre>
<p dir="auto">Is that what you were seeing on your powerstore with V4 as well?</p>
]]></description><link>https://xcp-ng.org/forum/post/81632</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81632</guid><dc:creator><![CDATA[flakpyro]]></dc:creator><pubDate>Sat, 17 Aug 2024 19:44:58 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 19:31:18 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> what NFS version are you using?<br />
We're backing up 50-60 VM's every night (without CBT) and the coalescale is pretty quick, i've tried NFS 4 but we had some timeout issues so we went back to NFS 3 which seems stable.</p>
<p dir="auto">Have you tried to experiment with the mount options?<br />
Just curious, since we didnt have any luck with NFS 4 on both TrueNAS and our Dell Powerstore, the latter is NVME allflash and the TrueNAS boxes are 10K SAS with SSD as cache.</p>
]]></description><link>https://xcp-ng.org/forum/post/81631</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81631</guid><dc:creator><![CDATA[nikade]]></dc:creator><pubDate>Sat, 17 Aug 2024 19:31:18 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 19:13:23 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> ah i understand, i have seen issues with nfs as well, doing it with iscsi did give us much better results.</p>
<p dir="auto">Also i would recommend you to look into alike a3 backup, we use it with high io vms with 2tb disk size, no issues there and u need only one snapshot per vm.</p>
]]></description><link>https://xcp-ng.org/forum/post/81629</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81629</guid><dc:creator><![CDATA[rtjdamen]]></dc:creator><pubDate>Sat, 17 Aug 2024 19:13:23 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 18:30:32 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/rtjdamen" aria-label="Profile: rtjdamen">@<bdi>rtjdamen</bdi></a> The Pure storage array is all flash. I did see your comments about increasing those values, i did do that and while it did help it still can be hit or miss for us. For example our sharepoint server took 2-3 hours the other night to coalesce when using full CBT with snapshot delete enabled. Running it without snapshot delete the traditional snapshot coalesce only takes maybe 10-15 mins.</p>
<p dir="auto">With that said we are also having issues with Pure's NFS implementation and how it interacts with XCP-NG, causing storage timeouts for us. According to them the array, when under load is disconnecting hosts due to "expired NFS leases" we are currently working with them to stabilize that, perhaps then i can revisit full CBT backups. Weirdly enough the disconnects do not appear to happen under regular VM operations even under high load, we mostly run into these during backup runs with CBT enabled runs increasing the chance of it happening.</p>
<p dir="auto">What we ideally will end up with is local backups, offsite backups and archival backups. This is what we had with Veeam prior. Doing this with traditional XOA backups would result in 3 snapshots per VM (1 for each job) which lead me to looking into commvault since it does things a bit differently. CBT would also solve this as well i assume.</p>
]]></description><link>https://xcp-ng.org/forum/post/81628</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81628</guid><dc:creator><![CDATA[flakpyro]]></dc:creator><pubDate>Sat, 17 Aug 2024 18:30:32 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 17:24:35 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/flakpyro" aria-label="Profile: flakpyro">@<bdi>flakpyro</bdi></a> are u using ssd storage? Of so i would recommend increasing the leaf coalesce parameters. This will prevent this loops from occuring. We had issues with this this prior to changing it. Now we do not see this anymore.</p>
]]></description><link>https://xcp-ng.org/forum/post/81624</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81624</guid><dc:creator><![CDATA[rtjdamen]]></dc:creator><pubDate>Sat, 17 Aug 2024 17:24:35 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Sat, 17 Aug 2024 16:53:33 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/rfx77" aria-label="Profile: rfx77">@<bdi>rfx77</bdi></a> How has your experience been with Commvault and XCP-NG for backups? We are bumping into similar issues migrating away from VMware/Veeam. XO backups work but the snapshots that remain on larger VMs can grow quite large between nightly backup runs.</p>
<p dir="auto">Trying CBT has been a mixed bag with snapshot coalesces taking anywhere from seconds up to 2-3 hours on larger VMs (with looping timeouts in SMlog) and backup failures if a VM has moved between hosts in the same pool.,and in the worst cases causing our NFS based storage (Pure storage) to expire all NFS leases to the pool master trying to coalesce causing NFS timeouts and storage outages. We have Pure looking into why this is happening for us now but until that's figured out we are back to CBT-less backups that leave a snapshots behind.</p>
<p dir="auto">From what i have read Commvault uses their proxy VM to track changed blocks without leaving the snapshot in place? Sounds like it may be worth checking out.</p>
]]></description><link>https://xcp-ng.org/forum/post/81621</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/81621</guid><dc:creator><![CDATA[flakpyro]]></dc:creator><pubDate>Sat, 17 Aug 2024 16:53:33 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Fri, 26 Jul 2024 07:39:35 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/olivierlambert" aria-label="Profile: olivierlambert">@<bdi>olivierlambert</bdi></a> i do agree, from our experience cbt did solve this issues. Snapshots are removed in seconds, this was 10 to 15 minutes prior to using cbt. There is still some work to be done with cbt but it is usable at this time.</p>
]]></description><link>https://xcp-ng.org/forum/post/80849</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/80849</guid><dc:creator><![CDATA[rtjdamen]]></dc:creator><pubDate>Fri, 26 Jul 2024 07:39:35 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Wed, 24 Jul 2024 08:53:24 GMT]]></title><description><![CDATA[<p dir="auto">So I advise to test CBT, this might solve your coalesce issues, since the snapshot will be removed a lot faster <img src="https://xcp-ng.org/forum/assets/plugins/nodebb-plugin-emoji/emoji/android/1f642.png?v=a78c449d9ac" class="not-responsive emoji emoji-android emoji--slightly_smiling_face" style="height:23px;width:auto;vertical-align:middle" title=":)" alt="🙂" /></p>
]]></description><link>https://xcp-ng.org/forum/post/80769</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/80769</guid><dc:creator><![CDATA[olivierlambert]]></dc:creator><pubDate>Wed, 24 Jul 2024 08:53:24 GMT</pubDate></item><item><title><![CDATA[Reply to Switching to XCP-NG, want to hear your problems on Wed, 24 Jul 2024 08:39:21 GMT]]></title><description><![CDATA[<p dir="auto"><a class="plugin-mentions-user plugin-mentions-a" href="/forum/user/olivierlambert" aria-label="Profile: olivierlambert">@<bdi>olivierlambert</bdi></a> i have running a XO installation which does backup a small part of our infrastructure, which works well. it is a single host with local thin-provisioned storage.</p>
<p dir="auto">we also tried it on our main pool with iscsi storage and had mostly the same problems like others above.</p>
<p dir="auto">we decided to wait a litte longer before we give it another try on our main production system mainly due to some hanging snapshots,...</p>
<p dir="auto">in the future we are planning to use it as main vm-level backup solution for standalone hosts in combination with commvault agents in the vm.</p>
<p dir="auto">our current test status i that we try to use XO as a docker container which would fit better in out existing backup infrastructure as we could easily integrate it in our linux backup appliances (see it like a linux NAS). The docker tests are running very well and we dont have any issues at the moment. even file-level-recovery is working well in this scenario.</p>
]]></description><link>https://xcp-ng.org/forum/post/80766</link><guid isPermaLink="true">https://xcp-ng.org/forum/post/80766</guid><dc:creator><![CDATA[rfx77]]></dc:creator><pubDate>Wed, 24 Jul 2024 08:39:21 GMT</pubDate></item></channel></rss>