Import from VMware fails after upgrade to XOA 5.91
-
[12:25 01] xoa:log$ df -h Filesystem Size Used Avail Use% Mounted on udev 954M 0 954M 0% /dev tmpfs 195M 528K 195M 1% /run /dev/xvda1 19G 3.1G 15G 18% / tmpfs 973M 0 973M 0% /dev/shm tmpfs 5.0M 0 5.0M 0% /run/lock tmpfs 195M 0 195M 0% /run/user/1000
-
Weird, it's like the tar stream with ended because of not enough space But it doesn't make sense reading your available space.
-
I started experiencing this exact same error after upgrading to 5.91 yesterday. It seems to affect one of my VMs that has 2 VMDK disks, the first (OS) disk copies over successfully, the 2nd (data) disk fails.
I've tried making a new VMDK and copying my data over, then destroying the old one in case there was a problem with that file but it still hangs around 50% then the task fails.
-
Might worth some investigation on our side, yes.
-
My VM doesn't have multiple vmdks, just the primary 16GB thin provisioned OS drive. Its actual size is only 4.77GB.
1vCPU
2GB RAM
16GB Disk
1 NIC
No dangling snapshots
powered offtruly nothing out of the ordinary.
-
Yeah it seems simple, I wonder what's going on, @florent will take a look. Ideally, if you can open a support ticket so we can take a look remotely, that would be helpful
-
ok, I am testing it with more disks , but the ENOSPC puzzle me for now
-
@jasonmap , @rmaclachlan , can you test this branch : fix_xva_import_thin ?
from this PR : https://github.com/vatesfr/xen-orchestra/pull/7354 -
I would be happy to test - but this is where that "hand holding" would come into play.
-
@jasonmap said in Import from VMware fails after upgrade to XOA 5.91:
I would be happy to test - but this is where that "hand holding" would come into play.
no problem . If you have a XOA , please open a support tunnel, I can patch your install
If not:- go to your XO install folder, inside, you should have a lot of files, and also a
@vates
, a@xen-orchestra
and apackages
folder git checkout fix_xva_import_thin
yarn && yarn build
4 restart yourxo-server
- go to your XO install folder, inside, you should have a lot of files, and also a
-
@florent
In the FWIW department, I get the same errors when trying to import with 5.91:result succeeded {} message "no opaque ref found" name "Error" stack "Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)"
The journalctl errors are as follows :
Feb 02 09:30:19 xoa xo-server[1178375]: 2024-02-02T14:30:19.498Z xo:xo-server WARN possibly unhandled rejection { Feb 02 09:30:19 xoa xo-server[1178375]: error: Error: already finalized or destroyed Feb 02 09:30:19 xoa xo-server[1178375]: at Pack.entry (/usr/local/lib/node_modules/xo-server/node_modules/tar-stream/pack.js:138:51) Feb 02 09:30:19 xoa xo-server[1178375]: at Pack.resolver (/usr/local/lib/node_modules/xo-server/node_modules/promise-toolbox/fromCallback.js:5:6) Feb 02 09:30:19 xoa xo-server[1178375]: at Promise._execute (/usr/local/lib/node_modules/xo-server/node_modules/bluebird/js/release/debuggability.js:384:9) Feb 02 09:30:19 xoa xo-server[1178375]: at Promise._resolveFromExecutor (/usr/local/lib/node_modules/xo-server/node_modules/bluebird/js/release/promise.js:518:18) Feb 02 09:30:19 xoa xo-server[1178375]: at new Promise (/usr/local/lib/node_modules/xo-server/node_modules/bluebird/js/release/promise.js:103:10) Feb 02 09:30:19 xoa xo-server[1178375]: at Pack.fromCallback (/usr/local/lib/node_modules/xo-server/node_modules/promise-toolbox/fromCallback.js:9:10) Feb 02 09:30:19 xoa xo-server[1178375]: at writeBlock (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/_writeDisk.mjs:11:22) Feb 02 09:30:19 xoa xo-server[1178375]: } Feb 02 09:30:21 xoa xo-server[1178375]: root@xx.xx.xx.xx Xapi#putResource /import/ XapiError: IMPORT_ERROR(INTERNAL_ERROR: [ Unix.Unix_error(Unix.ENOSPC, "write", "") ]) Feb 02 09:30:21 xoa xo-server[1178375]: at Function.wrap (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/_XapiError.mjs:16:12) Feb 02 09:30:21 xoa xo-server[1178375]: at default (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/_getTaskResult.mjs:11:29) Feb 02 09:30:21 xoa xo-server[1178375]: at Xapi._addRecordToCache (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/index.mjs:1006:24) Feb 02 09:30:21 xoa xo-server[1178375]: at file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/index.mjs:1040:14 Feb 02 09:30:21 xoa xo-server[1178375]: at Array.forEach (<anonymous>) Feb 02 09:30:21 xoa xo-server[1178375]: at Xapi._processEvents (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/index.mjs:1030:12) Feb 02 09:30:21 xoa xo-server[1178375]: at Xapi._watchEvents (file:///usr/local/lib/node_modules/xo-server/node_modules/xen-api/index.mjs:1203:14) Feb 02 09:30:21 xoa xo-server[1178375]: at runNextTicks (node:internal/process/task_queues:60:5) Feb 02 09:30:21 xoa xo-server[1178375]: at processImmediate (node:internal/timers:447:9) Feb 02 09:30:21 xoa xo-server[1178375]: at process.callbackTrampoline (node:internal/async_hooks:130:17) { Feb 02 09:30:21 xoa xo-server[1178375]: code: 'IMPORT_ERROR', Feb 02 09:30:21 xoa xo-server[1178375]: params: [ 'INTERNAL_ERROR: [ Unix.Unix_error(Unix.ENOSPC, "write", "") ]' ], Feb 02 09:30:21 xoa xo-server[1178375]: call: undefined, Feb 02 09:30:21 xoa xo-server[1178375]: url: undefined, Feb 02 09:30:21 xoa xo-server[1178375]: task: task { Feb 02 09:30:21 xoa xo-server[1178375]: uuid: '37b70fee-d769-1b2a-864b-7e18bd6fcff6', Feb 02 09:30:21 xoa xo-server[1178375]: name_label: '[XO] VM import', Feb 02 09:30:21 xoa xo-server[1178375]: name_description: '', Feb 02 09:30:21 xoa xo-server[1178375]: allowed_operations: [], Feb 02 09:30:21 xoa xo-server[1178375]: current_operations: {}, Feb 02 09:30:21 xoa xo-server[1178375]: created: '20240202T13:57:21Z', Feb 02 09:30:21 xoa xo-server[1178375]: finished: '20240202T14:30:21Z', Feb 02 09:30:21 xoa xo-server[1178375]: status: 'failure', Feb 02 09:30:21 xoa xo-server[1178375]: resident_on: 'OpaqueRef:a482c281-bc5c-6aed-e5cc-91d43c02dc46', Feb 02 09:30:21 xoa xo-server[1178375]: progress: 1, Feb 02 09:30:21 xoa xo-server[1178375]: type: '<none/>', Feb 02 09:30:21 xoa xo-server[1178375]: result: '', Feb 02 09:30:21 xoa xo-server[1178375]: error_info: [ Feb 02 09:30:21 xoa xo-server[1178375]: 'IMPORT_ERROR', Feb 02 09:30:21 xoa xo-server[1178375]: 'INTERNAL_ERROR: [ Unix.Unix_error(Unix.ENOSPC, "write", "") ]' Feb 02 09:30:21 xoa xo-server[1178375]: ], Feb 02 09:30:21 xoa xo-server[1178375]: other_config: { object_creation: 'complete' }, Feb 02 09:30:21 xoa xo-server[1178375]: subtask_of: 'OpaqueRef:NULL', Feb 02 09:30:21 xoa xo-server[1178375]: subtasks: [], Feb 02 09:30:21 xoa xo-server[1178375]: backtrace: '(((process xapi)(filename lib/backtrace.ml)(line 210))((process xapi)(filename ocaml/xapi/import.ml)(line 2131))((process xapi)(filename ocaml/xapi/server_helpers.ml)(line 72)))' Feb 02 09:30:21 xoa xo-server[1178375]: } Feb 02 09:30:21 xoa xo-server[1178375]: }
-
@florent said in Import from VMware fails after upgrade to XOA 5.91:
XO install folder
Where is the "XO install folder"? I did an ssh into the xoa vm and I don't see what you refer to?
I tried running "git checkout fix_xva_import_thin" but it says "fatal: not a git repository (or any of the parent directories): .git" -
@archw said in Import from VMware fails after upgrade to XOA 5.91:
@florent said in Import from VMware fails after upgrade to XOA 5.91:
XO install folder
Where is the "XO install folder"? I did an ssh into the xoa vm and I don't see what you refer to?
I tried running "git checkout fix_xva_import_thin" but it says "fatal: not a git repository (or any of the parent directories): .git"if it's an XOA, even a starter/free one, open a support tunnel, I will take care of it
If not, how did you install your XO from source ? -
@florent
The one I showed above what the starter version. Since I made that post, I also tried running it on the one I built from source and it is version "Master, commit 09247". It has the same errors.I just opened a support tunnel.
-
@florent I just sent you tunnell id
-
@archw said in Import from VMware fails after upgrade to XOA 5.91:
@florent I just sent you tunnell id
the patch is applied
-
@florent
Cool...I'll give it a shot...takes about twenty minutes. (it always dies in the last minute). -
Hi, I tried after the patch applied but it still failed. The task progress didn't seem to hang so I thought it would complete but in the end it failed and removed the VM from XOA.
{ "id": "0ls4sww64", "properties": { "name": "importing vms 239", "userId": "2af1207e-7ddd-4ce8-a13e-1c6ede07d0c9", "total": 1, "done": 0, "progress": 0 }, "start": 1706887780060, "status": "failure", "updatedAt": 1706888683479, "tasks": [ { "id": "cyz9dh2453h", "properties": { "name": "importing vm 239", "done": 1, "progress": 100 }, "start": 1706887780061, "status": "failure", "tasks": [ { "id": "gqw5xqq9ke5", "properties": { "name": "connecting to 10.193.240.111" }, "start": 1706887780062, "status": "success", "end": 1706887780170, "result": { "_events": {}, "_eventsCount": 0 } }, { "id": "d07ygv9vyjl", "properties": { "name": "get metadata of 239" }, "start": 1706887780171, "status": "success", "end": 1706887780569, "result": { "name_label": "BusCam", "memory": 4294967296, "nCpus": 2, "guestToolsInstalled": false, "firmware": "bios", "powerState": "poweredOff", "disks": [ { "capacity": 42949672960, "isFull": true, "uid": "698f084e", "fileName": "BusCam_2-flat.vmdk", "parentId": "ffffffff", "vmdFormat": "VMFS", "nameLabel": "BusCam_2-flat.vmdk", "datastore": "TS Storage 2", "path": "BusCam", "descriptionLabel": " from esxi", "node": "scsi0:0" }, { "capacity": 42949672960, "isFull": true, "uid": "1d9d8b5e", "fileName": "BusCam_1-flat.vmdk", "parentId": "ffffffff", "vmdFormat": "VMFS", "nameLabel": "BusCam_1-flat.vmdk", "datastore": "TS Storage 2", "path": "BusCam", "descriptionLabel": " from esxi", "node": "scsi0:1" } ], "networks": [ { "macAddress": "00:50:56:b9:55:dc", "isGenerated": false } ] } }, { "id": "ngimkpdl7e", "properties": { "name": "build disks and snapshots chains for 239" }, "start": 1706887780569, "status": "success", "end": 1706887780569, "result": { "scsi0:0": [ { "capacity": 42949672960, "isFull": true, "uid": "698f084e", "fileName": "BusCam_2-flat.vmdk", "parentId": "ffffffff", "vmdFormat": "VMFS", "nameLabel": "BusCam_2-flat.vmdk", "datastore": "TS Storage 2", "path": "BusCam", "descriptionLabel": " from esxi", "node": "scsi0:0" } ], "scsi0:1": [ { "capacity": 42949672960, "isFull": true, "uid": "1d9d8b5e", "fileName": "BusCam_1-flat.vmdk", "parentId": "ffffffff", "vmdFormat": "VMFS", "nameLabel": "BusCam_1-flat.vmdk", "datastore": "TS Storage 2", "path": "BusCam", "descriptionLabel": " from esxi", "node": "scsi0:1" } ] } }, { "id": "i8w7p4xgicn", "properties": { "name": "creating MV on XCP side" }, "start": 1706887780570, "status": "success", "end": 1706887780692, "result": { "uuid": "09d2d428-4555-f70d-b246-80f62edabde7", "allowed_operations": [ "create_vtpm", "changing_NVRAM", "changing_dynamic_range", "changing_shadow_memory", "changing_static_range", "make_into_template", "migrate_send", "destroy", "export", "start_on", "start", "clone", "copy", "snapshot" ], "current_operations": {}, "name_label": "BusCam", "name_description": "from esxi", "power_state": "Halted", "user_version": 1, "is_a_template": false, "is_default_template": false, "suspend_VDI": "OpaqueRef:NULL", "resident_on": "OpaqueRef:NULL", "scheduled_to_be_resident_on": "OpaqueRef:NULL", "affinity": "OpaqueRef:NULL", "memory_overhead": 37748736, "memory_target": 0, "memory_static_max": 4294967296, "memory_dynamic_max": 4294967296, "memory_dynamic_min": 4294967296, "memory_static_min": 4294967296, "VCPUs_params": {}, "VCPUs_max": 2, "VCPUs_at_startup": 2, "actions_after_softreboot": "soft_reboot", "actions_after_shutdown": "destroy", "actions_after_reboot": "restart", "actions_after_crash": "restart", "consoles": [], "VIFs": [], "VBDs": [], "VUSBs": [], "crash_dumps": [], "VTPMs": [], "PV_bootloader": "", "PV_kernel": "", "PV_ramdisk": "", "PV_args": "", "PV_bootloader_args": "", "PV_legacy_args": "", "HVM_boot_policy": "BIOS order", "HVM_boot_params": { "order": "cdn" }, "HVM_shadow_multiplier": 1, "platform": { "timeoffset": "0", "nx": "true", "acpi": "1", "apic": "true", "pae": "true", "hpet": "true", "viridian": "true" }, "PCI_bus": "", "other_config": { "mac_seed": "9e80dc56-0498-2f80-4fb8-545ce0366a26", "vgpu_pci": "", "base_template_name": "Other install media", "install-methods": "cdrom" }, "domid": -1, "domarch": "", "last_boot_CPU_flags": {}, "is_control_domain": false, "metrics": "OpaqueRef:6252fb53-d094-107d-efca-0ba6694e2d87", "guest_metrics": "OpaqueRef:NULL", "last_booted_record": "", "recommendations": "<restrictions><restriction field=\"memory-static-max\" max=\"137438953472\" /><restriction field=\"vcpus-max\" max=\"32\" /><restriction property=\"number-of-vbds\" max=\"255\" /><restriction property=\"number-of-vifs\" max=\"7\" /><restriction field=\"has-vendor-device\" value=\"false\" /></restrictions>", "xenstore_data": {}, "ha_always_run": false, "ha_restart_priority": "", "is_a_snapshot": false, "snapshot_of": "OpaqueRef:NULL", "snapshots": [], "snapshot_time": "19700101T00:00:00Z", "transportable_snapshot_id": "", "blobs": {}, "tags": [], "blocked_operations": {}, "snapshot_info": {}, "snapshot_metadata": "", "parent": "OpaqueRef:NULL", "children": [], "bios_strings": {}, "protection_policy": "OpaqueRef:NULL", "is_snapshot_from_vmpp": false, "snapshot_schedule": "OpaqueRef:NULL", "is_vmss_snapshot": false, "appliance": "OpaqueRef:NULL", "start_delay": 0, "shutdown_delay": 0, "order": 0, "VGPUs": [], "attached_PCIs": [], "suspend_SR": "OpaqueRef:NULL", "version": 0, "generation_id": "0:0", "hardware_platform_version": 0, "has_vendor_device": false, "requires_reboot": false, "reference_label": "", "domain_type": "hvm", "NVRAM": {}, "pending_guidances": [] } }, { "id": "iacdn87dv3", "properties": { "name": "Cold import of disks scsi0:0" }, "start": 1706887780693, "status": "failure", "end": 1706888683036, "result": { "message": "no opaque ref found", "name": "Error", "stack": "Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)" } }, { "id": "ku5ckax1wr", "properties": { "name": "Cold import of disks scsi0:1" }, "start": 1706887780696, "status": "success", "end": 1706888650415, "result": { "ref": "Ref:002", "label": "BusCam_1-flat.vmdk" } } ], "end": 1706888683479, "result": { "message": "no opaque ref found", "name": "Error", "stack": "Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)" } } ], "end": 1706888683479, "result": { "succeeded": {}, "message": "no opaque ref found", "name": "Error", "stack": "Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)" } }
-
I tried after you patched it...same error:
Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)"
-
@archw said in Import from VMware fails after upgrade to XOA 5.91:
I tried after you patched it...same error:
Error: no opaque ref found\n at importVm (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVm.mjs:28:19)\n at processTicksAndRejections (node:internal/process/task_queues:95:5)\n at importVdi (file:///usr/local/lib/node_modules/xo-server/node_modules/@xen-orchestra/xva/importVdi.mjs:6:17)\n at file:///usr/local/lib/node_modules/xo-server/src/xo-mixins/migrate-vm.mjs:260:21\n at Task.runInside (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:158:22)\n at Task.run (/usr/local/lib/node_modules/xo-server/node_modules/@vates/task/index.js:141:20)"
you had two pool with the same ip, that seems like a good candidates for xo losing the objects refs. I applied the same patches as rmaclachlan
Can you retry ?