@Danp, glusterfs.
Anyway I found the problem. It was a problem with cache size and the client fuse didn't mount it.
Thank you really much for helping me.
@Danp, glusterfs.
Anyway I found the problem. It was a problem with cache size and the client fuse didn't mount it.
Thank you really much for helping me.
@Danp, thanks for the reply.
No, but why I have to restart all hosts?
Anyway, it is the second time that i have this error on snapshot.
First time i restored the vm and all worked fine.
What is the cause?
Hi everyone,
I am encountering an issue when attempting to create a snapshot of a VM in XCP-ng. The error message I receive is:
SR_BACKEND_FAILURE_82(, Failed to snapshot VDI [opterr=['MAP_DUPLICATE_KEY', 'VDI', 'sm_config', 'OpaqueRef:34990f8e-f7dd-4974-8ad7-a814de1b809f', 'paused']], )
Steps Taken:
Questions:
MAP_DUPLICATE_KEY
error in the snapshot process?OpaqueRef
mentioned in the error message?Any insights or suggestions would be greatly appreciated. Thank you in advance for your help!
Can anyone provide a comprehensive guide for configuring rsyslog on my host? I've checked the official guide but couldn't find a complete configuration. Any help would be greatly appreciated!
@Danp, glusterfs.
Anyway I found the problem. It was a problem with cache size and the client fuse didn't mount it.
Thank you really much for helping me.
@Danp said in Issue with Two-Node HA Cluster: XAPI Failing to Log In and HA Disablement:
You could try running the following command on each host to disable HA on each one --
xe host-emergency-ha-disable --force
FWIW, HA requires three nodes to run properly, and not every environment needs HA even if you think you do.
Thank you for your input!
The XAPI is now functioning again after disable HA. However, I'm encountering an issue while reattaching the storage to each node. Here are the commands and outputs:
List Hosts:
[21:53 hv01-xcp-mo ~]# xe host-list
uuid ( RO) : 0a38ea70-8529-4d30-bf44-8f01e1e4101b
name-label ( RW): hv01-xcp-mo
name-description ( RW):
uuid ( RO) : 67c5d00a-977d-46e2-98ee-0aa620e94db0
name-label ( RW): hv02-xcp-mo
name-description ( RW):
List PBDs for SR:
[21:50 hv01-xcp-mo ~]# xe pbd-list sr-uuid=d4e8e42c-e758-dd47-800e-ce2aaae3abdc
uuid ( RO) : ccc3067f-afc2-a344-028b-3815da0b5afe
host-uuid ( RO): 0a38ea70-8529-4d30-bf44-8f01e1e4101b
sr-uuid ( RO): d4e8e42c-e758-dd47-800e-ce2aaae3abdc
device-config (MRO): backupservers: 10.1.80.11:/san; server: 10.1.80.12:/san
currently-attached ( RO): false
uuid ( RO) : 280f2c70-96e3-0404-2d20-61789c113356
host-uuid ( RO): 67c5d00a-977d-46e2-98ee-0aa620e94db0
sr-uuid ( RO): d4e8e42c-e758-dd47-800e-ce2aaae3abdc
device-config (MRO): backupservers: 10.1.80.11:/san; server: 10.1.80.12:/san
currently-attached ( RO): false
Attempt to Plug PBD:
[21:57 hv01-xcp-mo ~]# xe pbd-plug uuid=ccc3067f-afc2-a344-028b-3815da0b5afe
Error code: SR_BACKEND_FAILURE_12
Error parameters: , mount failed with return code 1,
Can you help me?
Thanks again for your support!
from
May 20 19:01:05 hv02-xcp-mo xapi: [ warn||0 |Checking HA configuration D:6662d56d1422|static_vdis] Attempt to reattach static VDIs via 'attach-static-vdis start' failed: INTERNAL_ERROR: [ Subprocess exited with unexpected code 1; stdout = [ ]; stderr = [ Redirecting to /bin/systemctl start attach-static-vdis.service\x0AJob for attach-static-vdis.service failed because the control process exited with error code. See "systemctl status attach-static-vdis.service" and "journalctl -xe" for details.\x0A ] ]
May 20 19:01:05 hv02-xcp-mo xapi: [debug||0 |Checking HA configuratio/var/log/xensource.log
Hi everyone,
I'm currently facing an issue with my two-node HA cluster. I don't have access to the REST API, and I'm receiving the following error messages:
Broadcast message from systemd-journald@hv02-xcp-mo (Mon 2024-05-20 18:38:07 CEST):
xapi-nbd[9548]: main: Failed to log in via xapi's Unix domain socket in 300.000000 seconds
Broadcast message from systemd-journald@hv02-xcp-mo (Mon 2024-05-20 18:38:07 CEST):
xapi-nbd[9548]: main: Caught unexpected exception: (Failure
Broadcast message from systemd-journald@hv02-xcp-mo (Mon 2024-05-20 18:38:07 CEST):
xapi-nbd[9548]: main: "Failed to log in via xapi's Unix domain socket in 300.000000 seconds")
Due to this, nothing seems to be working correctly. I am unable to manage the cluster or access any of the services that rely on XAPI.
Additionally, I am trying to disable HA from the CLI but encounter the following error:
[18:47 hv02-xcp-mo d4e8e42c-e758-dd47-800e-ce2aaae3abdc]# xe pool-ha-disable
The server could not join the liveset because the HA daemon could not access the heartbeat disk.
I have the disk on storage, but I can't mount it.
How can I disable HA from the CLI, even with these problems?
Please, this is urgent. Any help or guidance would be greatly appreciated!
Thank you.
Environment Summary:
Host 1:
Host 2:
Both hosts are part of a High Availability (HA) pool and are connected to shared SAN storage that serves as the Storage Repository (SR) for heartbeats, configured with:
xe pool-ha-enable heartbeat-sr-uuids="SAN storage UUID"
Scenario:
Imagine a situation where both hosts simultaneously lose their management network interfaces. However, communication with the SAN storage via the storage network interfaces remains intact. Each host is running some VM.
Questions:
@olivierlambert, but after a rescan I have to reattach the disk to vm:
so when i go to (VM/disk/attach disk) i have to select the correct disk.
I have 14 vm but many vhd in the SR.
What are base copies?
What happens if I delete them and only keep the VHD of the discs?
When I do the rescan, I will have several discs without labels and I will have to figure out which one is the correct one by taking it from the VDI identifier.
is this the only way?
Hello @olivierlambert,
1 QUESTION
I have a follow-up question regarding the storage repository (SR) UUID management. Here is the current UUID of my SR (ee8b6a3-0c17-9475-ba12-fba53a16ac29).
On the node:
[hv02-xcp-mo 9ee8b6a3-0c17-9475-ba12-fba53a16ac29]# df -h
10.1.80.12:/gv1 6.6T 783G 5.8T 12% /run/sr-mount/GlusterFS/10.1.80.12/9ee8b6a3-0c17-9475-ba12-fba53a16ac29
Based on this output, can I assume that the all VHD files of the VM are stored within this directory?
Additionally, when examining the contents of this directory, I observed other files:
[hv02-xcp-mo 10.1.80.12]# cd /run/sr-mount/GlusterFS/10.1.80.12/9ee8b6a3-0c17-9475-ba12-fba53a16ac29
[hv02-xcp-mo 9ee8b6a3-0c17-9475-ba12-fba53a16ac29]# ls -la
total 16
drwxrwxrwx 8 root root 4096 Mar 1 11:48 .
drwxr-xr-x 3 root root 60 Apr 7 21:15 ..
drwxr-xr-x 2 root root 8192 Apr 19 05:00 9ee8b6a3-0c17-9475-ba12-fba53a16ac29
drwxr-xr-x 2 root root 10 Feb 27 15:35 bf2ff74a-24ce-bfa3-74e6-3c3dae509fa3
drwxr-xr-x 2 root root 4096 Feb 27 17:23 cca38cb9-058a-d685-6025-6610eddb92a6
drwxr-xr-x 2 root root 10 Feb 26 23:09 def21a89-70fb-cfff-dd5d-c473f2123eeb
drwxr-xr-x 2 root root 189 Apr 18 23:02 e7fc85ca-b4bf-5990-b0bf-47a84be101f6
Given the listing, is it safe to remove all other directory except "the actual SR" 9ee8b6a3-0c17-9475-ba12-fba53a16ac29
?
2 QUESTION
Considering that I have numerous files in the current Storage Repository (SR), including base copies, etc., I have a question: After rescanning the SR, when I attempt to reattach the disk to the virtual machine, will the correct disk appear automatically?
Your insights would be greatly appreciated.
Thank you!