-
@fred974 And
rpm -qa | grep sm? Because the sm LINSTOR package update is in our repo. So I suppose you already installed it using koji URLs. -
@ronan-a said in XOSTOR hyperconvergence preview:
@fred974 And
sudo rpm -qa | grep sm? Because the sm LINSTOR package update is in our repo. So I suppose you already installed it using koji URLs.microsemi-smartpqi-1.2.10_025-2.xcpng8.2.x86_64 smartmontools-6.5-1.el7.x86_64 sm-rawhba-2.30.7-1.3.0.linstor.7.xcpng8.2.x86_64 ssmtp-2.64-14.el7.x86_64 sm-cli-0.23.0-7.xcpng8.2.x86_64 sm-2.30.7-1.3.0.linstor.7.xcpng8.2.x86_64 libsmbclient-4.10.16-15.el7_9.x86_64 psmisc-22.20-15.el7.x86_64Yes, I installed it from the koji URLs before seeing your rely
-
@fred974 I just repaired your pool, there was a small error in the conf that I gave in my previous post.
-
@ronan-a said in XOSTOR hyperconvergence preview:
I just repaired your pool, there was a small error in the conf that I gave in my previous post.
Thank you very much. I really appreciate you fixing this for me

-
@ronan-a said in XOSTOR hyperconvergence preview:
@Swen said in XOSTOR hyperconvergence preview:
@ronan-a perfect, thx! Is this the new release Olivier was talking about? Can you provide some information when to expect the first stable release?
If we don't have a new critical bug, normally in few weeks.
Fingers crossed!

-
@ronan-a: After doing the installation from scratch with new installed xcp-ng hosts, all uptodate, I need to do a repair (via xcp-ng center) of the SR after doing the xe sr-create, because the SR is in state: Broken and the pool-master is in state: Unplugged.
I am not really sure waht xcp-ng center is doing when I click repair, but it works.
I can reproduce this issue, it happens every installation.
regards,
Swen -
I don't remember if
sr-createis also pluging the PBD by default
Repair is just a
xe pbd-plugIIRC. -
@olivierlambert it looks like sr-create is doing it, because on all other nodes the SR is attached, only on pool-master (or maybe the node you do the sr-create from) the pdb-plug does not work.
-
What's the error message when you try to plug it?
-
@olivierlambert I need to me more clear about this: When doing the sr-create for the linstor storage no error is shown, but the pbd will not be plugged at the pool-master. On every other host in the cluster it works automatically. After doing a pdb-plug for the pool-master the SR will be plugged. No error is shown at all.
-
Okay I see, thanks

-
Is there an easy way to map the linstor resource volume to the virtual disk on xcp-ng? When doing linstor volume list I get a Resource name back from linstor like this:
xcp-volume-23d07d99-9990-4046-8e7d-020bd61c1883The last part looks like an uuid to me, but I am unable to find this uuid when using some xe commands.
-
@ronan-a: I am playing around with xcp-ng, linstor and Cloudstack. Sometimes when I create a new VM I run into this error: The VDI is not available
CS is trying it again after this error automatically and than it works and the new VM is starting. CS is using a template which is also on the linstor SR to create new VMs.
I attached the SMlog of the host.
SMlog.txt -
@Swen After doing the installation from scratch with new installed xcp-ng hosts, all uptodate, I need to do a repair (via xcp-ng center) of the SR after doing the xe sr-create, because the SR is in state: Broken and the pool-master is in state: Unplugged.
I am not really sure waht xcp-ng center is doing when I click repair, but it works.
It's just a PBD plug call I suppose. Can you share your logs please?
-
@Swen said in XOSTOR hyperconvergence preview:
Is there an easy way to map the linstor resource volume to the virtual disk on xcp-ng? When doing linstor volume list I get a Resource name back from linstor like this:
xcp-volume-23d07d99-9990-4046-8e7d-020bd61c1883The last part looks like an uuid to me, but I am unable to find this uuid when using some xe commands.
There is a tool installed by our RPMs to do that

For example on my host:linstor-kv-tool -u xostor-2 -g xcp-sr-linstor_group_thin_device --dump-volumes -n xcp/volume { "7ca7b184-ec9e-40bd-addc-082483f8e420/metadata": "{\"read_only\": false, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": \"\", \"name_label\": \"debian 11 hub disk\", \"name_description\": \"Created by XO\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}", "7ca7b184-ec9e-40bd-addc-082483f8e420/not-exists": "0", "7ca7b184-ec9e-40bd-addc-082483f8e420/volume-name": "xcp-volume-12571cf9-1c3b-4ee9-8f93-f4d2f7ea6bd8" } -
@Swen said in XOSTOR hyperconvergence preview:
@ronan-a: I am playing around with xcp-ng, linstor and Cloudstack. Sometimes when I create a new VM I run into this error: The VDI is not available
CS is trying it again after this error automatically and than it works and the new VM is starting. CS is using a template which is also on the linstor SR to create new VMs.
I attached the SMlog of the host.
SMlog.txtCan you share the log files of the other hosts please?
-
@ronan-a sure, which logs exactly do you need?
-
@ronan-a said in XOSTOR hyperconvergence preview:
There is a tool installed by our RPMs to do that

For example on my host:linstor-kv-tool -u xostor-2 -g xcp-sr-linstor_group_thin_device --dump-volumes -n xcp/volume { "7ca7b184-ec9e-40bd-addc-082483f8e420/metadata": "{\"read_only\": false, \"snapshot_time\": \"\", \"vdi_type\": \"vhd\", \"snapshot_of\": \"\", \"name_label\": \"debian 11 hub disk\", \"name_description\": \"Created by XO\", \"type\": \"user\", \"metadata_of_pool\": \"\", \"is_a_snapshot\": false}", "7ca7b184-ec9e-40bd-addc-082483f8e420/not-exists": "0", "7ca7b184-ec9e-40bd-addc-082483f8e420/volume-name": "xcp-volume-12571cf9-1c3b-4ee9-8f93-f4d2f7ea6bd8" }Great to know, thx for the info. Is there a reason not to use the same uuid in xcp-ng and linstor? Does it make sense to add the vdi and/or vbd uuid to the output of the command?
-
@Swen said in XOSTOR hyperconvergence preview:
@ronan-a sure, which logs exactly do you need?
SMlog files of each host

-
@Swen said in XOSTOR hyperconvergence preview:
Great to know, thx for the info. Is there a reason not to use the same uuid in xcp-ng and linstor? Does it make sense to add the vdi and/or vbd uuid to the output of the command?
The main reason is that you cannot rename a LINSTOR resource once it has been created. And we need to be able to do this to implement the snapshot feature. To workaround that, a shared dictionary is used to map XAPI UUIDs to the LINSTOR resources.
It's a non-sense for the readability to use the XAPI UUIDs for the LINSTOR resources due to VDI UUID renaming when a snapshot is created.
I don't see a good reason to add the VDB UUIDs in the dictionary. You already have the VDIs, you can use xe commands to fetch the other infos.
Hello! It looks like you're interested in this conversation, but you don't have an account yet.
Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.
With your input, this post could be even better 💗
Register Login