Import from VMware fails after upgrade to XOA 5.91
-
I started experiencing this exact same error after upgrading to 5.91 yesterday. It seems to affect one of my VMs that has 2 VMDK disks, the first (OS) disk copies over successfully, the 2nd (data) disk fails.
I've tried making a new VMDK and copying my data over, then destroying the old one in case there was a problem with that file but it still hangs around 50% then the task fails.
-
Might worth some investigation on our side, yes.
-
My VM doesn't have multiple vmdks, just the primary 16GB thin provisioned OS drive. Its actual size is only 4.77GB.
1vCPU
2GB RAM
16GB Disk
1 NIC
No dangling snapshots
powered offtruly nothing out of the ordinary.
-
Yeah it seems simple, I wonder what's going on, @florent will take a look. Ideally, if you can open a support ticket so we can take a look remotely, that would be helpful

-
ok, I am testing it with more disks , but the ENOSPC puzzle me for now
-
@jasonmap , @rmaclachlan , can you test this branch : fix_xva_import_thin ?
from this PR : https://github.com/vatesfr/xen-orchestra/pull/7354 -
I would be happy to test - but this is where that "hand holding" would come into play.
-
@jasonmap said in Import from VMware fails after upgrade to XOA 5.91:
I would be happy to test - but this is where that "hand holding" would come into play.
no problem . If you have a XOA , please open a support tunnel, I can patch your install
If not:- go to your XO install folder, inside, you should have a lot of files, and also a
@vates, a@xen-orchestraand apackagesfolder git checkout fix_xva_import_thinyarn && yarn build
4 restart yourxo-server
- go to your XO install folder, inside, you should have a lot of files, and also a
-
@florent
In the FWIW department, I get the same errors when trying to import with 5.91:result succeeded {} message "no opaque ref found" name "Error" stack "Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)"The journalctl errors are as follows :
Feb 02 09:30:19 xoa xo-server[1178375]: 2024-02-02T14:30:19.498Z xo:xo-server WARN possibly unhandled rejection { Feb 02 09:30:19 xoa xo-server[1178375]: error: Error: already finalized or destroyed Feb 02 09:30:19 xoa xo-server[1178375]: at Pack.entry (/usr/local/lib/node_modules/xo-server/node_modules/tar-stream/pack.js:138:51) Feb 02 09:30:19 xoa xo-server[1178375]: at Pack.resolver (/usr/local/lib/node_modules/xo-server/node_modules/promise-toolbox/fromCallback.js:5:6) Feb 02 09:30:19 xoa xo-server[1178375]: at Promise._execute (/usr/local/lib/node_modules/xo-server/node_modules/bluebird/js/release/debuggability.js:384:9) Feb 02 09:30:19 xoa xo-server[1178375]: at Promise._resolveFromExecutor (/usr/local/lib/node_modules/xo-server/node_modules/bluebird/js/release/promise.js:518:18) Feb 02 09:30:19 xoa xo-server[1178375]: at new Promise (/usr/local/lib/node_modules/xo-server/node_modules/bluebird/js/release/promise.js:103:10) Feb 02 09:30:19 xoa xo-server[1178375]: at Pack.fromCallback (/usr/local/lib/node_modules/xo-server/node_modules/promise-toolbox/fromCallback.js:9:10) Feb 02 09:30:19 xoa xo-server[1178375]: at writeBlock (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/_writeDisk.mjs:11:22) Feb 02 09:30:19 xoa xo-server[1178375]: } Feb 02 09:30:21 xoa xo-server[1178375]: root@xx.xx.xx.xx Xapi#putResource /import/ XapiError: IMPORT_ERROR(INTERNAL_ERROR: [ Unix.Unix_error(Unix.ENOSPC, "write", "") ]) Feb 02 09:30:21 xoa xo-server[1178375]: at Function.wrap (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/_XapiError.mjs:16:12) Feb 02 09:30:21 xoa xo-server[1178375]: at default (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/_getTaskResult.mjs:11:29) Feb 02 09:30:21 xoa xo-server[1178375]: at Xapi._addRecordToCache (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/index.mjs:1006:24) Feb 02 09:30:21 xoa xo-server[1178375]: at file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/index.mjs:1040:14 Feb 02 09:30:21 xoa xo-server[1178375]: at Array.forEach (<anonymous>) Feb 02 09:30:21 xoa xo-server[1178375]: at Xapi._processEvents (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/index.mjs:1030:12) Feb 02 09:30:21 xoa xo-server[1178375]: at Xapi._watchEvents (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/index.mjs:1203:14) Feb 02 09:30:21 xoa xo-server[1178375]: at runNextTicks (node:internal/process/task_queues:60:5) Feb 02 09:30:21 xoa xo-server[1178375]: at processImmediate (node:internal/timers:447:9) Feb 02 09:30:21 xoa xo-server[1178375]: at process.callbackTrampoline (node:internal/async_hooks:130:17) { Feb 02 09:30:21 xoa xo-server[1178375]: code: 'IMPORT_ERROR', Feb 02 09:30:21 xoa xo-server[1178375]: params: [ 'INTERNAL_ERROR: [ Unix.Unix_error(Unix.ENOSPC, "write", "") ]' ], Feb 02 09:30:21 xoa xo-server[1178375]: call: undefined, Feb 02 09:30:21 xoa xo-server[1178375]: url: undefined, Feb 02 09:30:21 xoa xo-server[1178375]: task: task { Feb 02 09:30:21 xoa xo-server[1178375]: uuid: '37b70fee-d769-1b2a-864b-7e18bd6fcff6', Feb 02 09:30:21 xoa xo-server[1178375]: name_label: '[XO] VM import', Feb 02 09:30:21 xoa xo-server[1178375]: name_description: '', Feb 02 09:30:21 xoa xo-server[1178375]: allowed_operations: [], Feb 02 09:30:21 xoa xo-server[1178375]: current_operations: {}, Feb 02 09:30:21 xoa xo-server[1178375]: created: '20240202T13:57:21Z', Feb 02 09:30:21 xoa xo-server[1178375]: finished: '20240202T14:30:21Z', Feb 02 09:30:21 xoa xo-server[1178375]: status: 'failure', Feb 02 09:30:21 xoa xo-server[1178375]: resident_on: 'OpaqueRef:a482c281-bc5c-6aed-e5cc-91d43c02dc46', Feb 02 09:30:21 xoa xo-server[1178375]: progress: 1, Feb 02 09:30:21 xoa xo-server[1178375]: type: '<none/>', Feb 02 09:30:21 xoa xo-server[1178375]: result: '', Feb 02 09:30:21 xoa xo-server[1178375]: error_info: [ Feb 02 09:30:21 xoa xo-server[1178375]: 'IMPORT_ERROR', Feb 02 09:30:21 xoa xo-server[1178375]: 'INTERNAL_ERROR: [ Unix.Unix_error(Unix.ENOSPC, "write", "") ]' Feb 02 09:30:21 xoa xo-server[1178375]: ], Feb 02 09:30:21 xoa xo-server[1178375]: other_config: { object_creation: 'complete' }, Feb 02 09:30:21 xoa xo-server[1178375]: subtask_of: 'OpaqueRef:NULL', Feb 02 09:30:21 xoa xo-server[1178375]: subtasks: [], Feb 02 09:30:21 xoa xo-server[1178375]: backtrace: '(((process xapi)(filename lib/backtrace.ml)(line 210))((process xapi)(filename ocaml/xapi/import.ml)(line 2131))((process xapi)(filename ocaml/xapi/server_helpers.ml)(line 72)))' Feb 02 09:30:21 xoa xo-server[1178375]: } Feb 02 09:30:21 xoa xo-server[1178375]: } -
@florent said in Import from VMware fails after upgrade to XOA 5.91:
XO install folder
Where is the "XO install folder"? I did an ssh into the xoa vm and I don't see what you refer to?
I tried running "git checkout fix_xva_import_thin" but it says "fatal: not a git repository (or any of the parent directories): .git" -
@archw said in Import from VMware fails after upgrade to XOA 5.91:
@florent said in Import from VMware fails after upgrade to XOA 5.91:
XO install folder
Where is the "XO install folder"? I did an ssh into the xoa vm and I don't see what you refer to?
I tried running "git checkout fix_xva_import_thin" but it says "fatal: not a git repository (or any of the parent directories): .git"if it's an XOA, even a starter/free one, open a support tunnel, I will take care of it
If not, how did you install your XO from source ? -
@florent
The one I showed above what the starter version. Since I made that post, I also tried running it on the one I built from source and it is version "Master, commit 09247". It has the same errors.I just opened a support tunnel.
-
@florent I just sent you tunnell id
-
@archw said in Import from VMware fails after upgrade to XOA 5.91:
@florent I just sent you tunnell id
the patch is applied
-
@florent
Cool...I'll give it a shot...takes about twenty minutes. (it always dies in the last minute). -
Hi, I tried after the patch applied but it still failed. The task progress didn't seem to hang so I thought it would complete but in the end it failed and removed the VM from XOA.
{ "id": "0ls4sww64", "properties": { "name": "importing vms 239", "userId": "2af1207e-7ddd-4ce8-a13e-1c6ede07d0c9", "total": 1, "done": 0, "progress": 0 }, "start": 1706887780060, "status": "failure", "updatedAt": 1706888683479, "tasks": [ { "id": "cyz9dh2453h", "properties": { "name": "importing vm 239", "done": 1, "progress": 100 }, "start": 1706887780061, "status": "failure", "tasks": [ { "id": "gqw5xqq9ke5", "properties": { "name": "connecting to 10.193.240.111" }, "start": 1706887780062, "status": "success", "end": 1706887780170, "result": { "_events": {}, "_eventsCount": 0 } }, { "id": "d07ygv9vyjl", "properties": { "name": "get metadata of 239" }, "start": 1706887780171, "status": "success", "end": 1706887780569, "result": { "name_label": "BusCam", "memory": 4294967296, "nCpus": 2, "guestToolsInstalled": false, "firmware": "bios", "powerState": "poweredOff", "disks": [ { "capacity": 42949672960, "isFull": true, "uid": "698f084e", "fileName": "BusCam_2-flat.vmdk", "parentId": "ffffffff", "vmdFormat": "VMFS", "nameLabel": "BusCam_2-flat.vmdk", "datastore": "TS Storage 2", "path": "BusCam", "descriptionLabel": " from esxi", "node": "scsi0:0" }, { "capacity": 42949672960, "isFull": true, "uid": "1d9d8b5e", "fileName": "BusCam_1-flat.vmdk", "parentId": "ffffffff", "vmdFormat": "VMFS", "nameLabel": "BusCam_1-flat.vmdk", "datastore": "TS Storage 2", "path": "BusCam", "descriptionLabel": " from esxi", "node": "scsi0:1" } ], "networks": [ { "macAddress": "00:50:56:b9:55:dc", "isGenerated": false } ] } }, { "id": "ngimkpdl7e", "properties": { "name": "build disks and snapshots chains for 239" }, "start": 1706887780569, "status": "success", "end": 1706887780569, "result": { "scsi0:0": [ { "capacity": 42949672960, "isFull": true, "uid": "698f084e", "fileName": "BusCam_2-flat.vmdk", "parentId": "ffffffff", "vmdFormat": "VMFS", "nameLabel": "BusCam_2-flat.vmdk", "datastore": "TS Storage 2", "path": "BusCam", "descriptionLabel": " from esxi", "node": "scsi0:0" } ], "scsi0:1": [ { "capacity": 42949672960, "isFull": true, "uid": "1d9d8b5e", "fileName": "BusCam_1-flat.vmdk", "parentId": "ffffffff", "vmdFormat": "VMFS", "nameLabel": "BusCam_1-flat.vmdk", "datastore": "TS Storage 2", "path": "BusCam", "descriptionLabel": " from esxi", "node": "scsi0:1" } ] } }, { "id": "i8w7p4xgicn", "properties": { "name": "creating MV on XCP side" }, "start": 1706887780570, "status": "success", "end": 1706887780692, "result": { "uuid": "09d2d428-4555-f70d-b246-80f62edabde7", "allowed_operations": [ "create_vtpm", "changing_NVRAM", "changing_dynamic_range", "changing_shadow_memory", "changing_static_range", "make_into_template", "migrate_send", "destroy", "export", "start_on", "start", "clone", "copy", "snapshot" ], "current_operations": {}, "name_label": "BusCam", "name_description": "from esxi", "power_state": "Halted", "user_version": 1, "is_a_template": false, "is_default_template": false, "suspend_VDI": "OpaqueRef:NULL", "resident_on": "OpaqueRef:NULL", "scheduled_to_be_resident_on": "OpaqueRef:NULL", "affinity": "OpaqueRef:NULL", "memory_overhead": 37748736, "memory_target": 0, "memory_static_max": 4294967296, "memory_dynamic_max": 4294967296, "memory_dynamic_min": 4294967296, "memory_static_min": 4294967296, "VCPUs_params": {}, "VCPUs_max": 2, "VCPUs_at_startup": 2, "actions_after_softreboot": "soft_reboot", "actions_after_shutdown": "destroy", "actions_after_reboot": "restart", "actions_after_crash": "restart", "consoles": [], "VIFs": [], "VBDs": [], "VUSBs": [], "crash_dumps": [], "VTPMs": [], "PV_bootloader": "", "PV_kernel": "", "PV_ramdisk": "", "PV_args": "", "PV_bootloader_args": "", "PV_legacy_args": "", "HVM_boot_policy": "BIOS order", "HVM_boot_params": { "order": "cdn" }, "HVM_shadow_multiplier": 1, "platform": { "timeoffset": "0", "nx": "true", "acpi": "1", "apic": "true", "pae": "true", "hpet": "true", "viridian": "true" }, "PCI_bus": "", "other_config": { "mac_seed": "9e80dc56-0498-2f80-4fb8-545ce0366a26", "vgpu_pci": "", "base_template_name": "Other install media", "install-methods": "cdrom" }, "domid": -1, "domarch": "", "last_boot_CPU_flags": {}, "is_control_domain": false, "metrics": "OpaqueRef:6252fb53-d094-107d-efca-0ba6694e2d87", "guest_metrics": "OpaqueRef:NULL", "last_booted_record": "", "recommendations": "<restrictions><restriction field=\"memory-static-max\" max=\"137438953472\" /><restriction field=\"vcpus-max\" max=\"32\" /><restriction property=\"number-of-vbds\" max=\"255\" /><restriction property=\"number-of-vifs\" max=\"7\" /><restriction field=\"has-vendor-device\" value=\"false\" /></restrictions>", "xenstore_data": {}, "ha_always_run": false, "ha_restart_priority": "", "is_a_snapshot": false, "snapshot_of": "OpaqueRef:NULL", "snapshots": [], "snapshot_time": "19700101T00:00:00Z", "transportable_snapshot_id": "", "blobs": {}, "tags": [], "blocked_operations": {}, "snapshot_info": {}, "snapshot_metadata": "", "parent": "OpaqueRef:NULL", "children": [], "bios_strings": {}, "protection_policy": "OpaqueRef:NULL", "is_snapshot_from_vmpp": false, "snapshot_schedule": "OpaqueRef:NULL", "is_vmss_snapshot": false, "appliance": "OpaqueRef:NULL", "start_delay": 0, "shutdown_delay": 0, "order": 0, "VGPUs": [], "attached_PCIs": [], "suspend_SR": "OpaqueRef:NULL", "version": 0, "generation_id": "0:0", "hardware_platform_version": 0, "has_vendor_device": false, "requires_reboot": false, "reference_label": "", "domain_type": "hvm", "NVRAM": {}, "pending_guidances": [] } }, { "id": "iacdn87dv3", "properties": { "name": "Cold import of disks scsi0:0" }, "start": 1706887780693, "status": "failure", "end": 1706888683036, "result": { "message": "no opaque ref found", "name": "Error", "stack": "Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)" } }, { "id": "ku5ckax1wr", "properties": { "name": "Cold import of disks scsi0:1" }, "start": 1706887780696, "status": "success", "end": 1706888650415, "result": { "ref": "Ref:002", "label": "BusCam_1-flat.vmdk" } } ], "end": 1706888683479, "result": { "message": "no opaque ref found", "name": "Error", "stack": "Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)" } } ], "end": 1706888683479, "result": { "succeeded": {}, "message": "no opaque ref found", "name": "Error", "stack": "Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)" } } -
I tried after you patched it...same error:
Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)" -
@archw said in Import from VMware fails after upgrade to XOA 5.91:
I tried after you patched it...same error:
Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)"you had two pool with the same ip, that seems like a good candidates for xo losing the objects refs. I applied the same patches as rmaclachlan
Can you retry ? -
@florent
I's running it right now! -
@florent tunnel is open!

Hello! It looks like you're interested in this conversation, but you don't have an account yet.
Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.
With your input, this post could be even better 💗
Register Login