Thanks for replies.
Here is where it is at:
# Two nodes think no VMs exist
[root@odin ~]# vdsm-client Host getVMList
[]
#One showing one VM but down
[root@medusa ~]# vdsm-client Host getVMList
[
{
"status": "Down",
"statusTime": "2153886148",
"vmId": "69ab4f82-1a53-42c8-afca-210a3a2715f1"
}
]
[root@medusa ~]# vdsm-client Host getAllVmStats
[
{
"exitCode": 1,
"exitMessage": "VM terminated with error",
"exitReason": 1,
"status": "Down",
"statusTime": "2153916276",
"vmId": "69ab4f82-1a53-42c8-afca-210a3a2715f1"
}
]
[root@medusa ~]# vdsm-client VM cont
vmID="69ab4f82-1a53-42c8-afca-210a3a2715f1"
vdsm-client: Command VM.cont with args {'vmID':
'69ab4f82-1a53-42c8-afca-210a3a2715f1'} failed:
(code=16, message=Unexpected exception)
# Assuming that ID represents the hosted-engine I tried to start it
[root@medusa ~]# hosted-engine --vm-start
The hosted engine configuration has not been retrieved from shared storage. Please ensure
that ovirt-ha-agent is running and the storage server is reachable.
# Back to ovirt-ha-agent being fubar and stoping things.
I have about 8 or so VMs on the cluster. Two are my IDM nodes which has DNS and other core
services.. which is what I am really trying to get up .. even if manual until I figure out
oVirt issue. I think you are correct. "engine" volume is for just the engine.
Data is where the other VMs are
[root@medusa images]# tree
.
├── 335c6b1a-d8a5-4664-9a9c-39744d511af8
│ ├── 579323ad-bf7b-479b-b682-6e1e234a7908
│ ├── 579323ad-bf7b-479b-b682-6e1e234a7908.lease
│ └── 579323ad-bf7b-479b-b682-6e1e234a7908.meta
├── d318cb8f-743a-461b-b246-75ffcde6bc5a
│ ├── c16877d0-eb23-42ef-a06e-a3221ea915fc
│ ├── c16877d0-eb23-42ef-a06e-a3221ea915fc.lease
│ └── c16877d0-eb23-42ef-a06e-a3221ea915fc.meta
└── junk
├── 296163f2-846d-4a2c-9a4e-83a58640b907
│ ├── 376b895f-e0f2-4387-b038-fbef4705fbcc
│ ├── 376b895f-e0f2-4387-b038-fbef4705fbcc.lease
│ └── 376b895f-e0f2-4387-b038-fbef4705fbcc.meta
├── 45a478d7-4c1b-43e8-b106-7acc75f066fa
│ ├── b5249e6c-0ba6-4302-8e53-b74d2b919d20
│ ├── b5249e6c-0ba6-4302-8e53-b74d2b919d20.lease
│ └── b5249e6c-0ba6-4302-8e53-b74d2b919d20.meta
├── d8b708c1-5762-4215-ae1f-0e57444c99ad
│ ├── 2536ca6d-3254-4cdc-bbd8-349ec1b8a0e9
│ ├── 2536ca6d-3254-4cdc-bbd8-349ec1b8a0e9.lease
│ └── 2536ca6d-3254-4cdc-bbd8-349ec1b8a0e9.meta
└── eaf12f3c-301f-4b61-b5a1-0c6d0b0a7f7b
├── fbf3bf59-a23a-4c6f-b66e-71369053b406
├── fbf3bf59-a23a-4c6f-b66e-71369053b406.lease
└── fbf3bf59-a23a-4c6f-b66e-71369053b406.meta
7 directories, 18 files
[root@medusa images]# cd /media/engine/
[root@medusa engine]# ls
3afc47ba-afb9-413f-8de5-8d9a2f45ecde
[root@medusa engine]# tree
.
└── 3afc47ba-afb9-413f-8de5-8d9a2f45ecde
├── dom_md
│ ├── ids
│ ├── inbox
│ ├── leases
│ ├── metadata
│ ├── outbox
│ └── xleases
├── ha_agent
├── images
│ ├── 1dc69552-dcc6-484d-8149-86c93ff4b8cc
│ │ ├── e4e26573-09a5-43fa-91ec-37d12de46480
│ │ ├── e4e26573-09a5-43fa-91ec-37d12de46480.lease
│ │ └── e4e26573-09a5-43fa-91ec-37d12de46480.meta
│ ├── 375d2483-ee83-4cad-b421-a5a70ec06ba6
│ │ ├── f936d4be-15e3-4983-8bf0-9ba5b97e638a
│ │ ├── f936d4be-15e3-4983-8bf0-9ba5b97e638a.lease
│ │ └── f936d4be-15e3-4983-8bf0-9ba5b97e638a.meta
│ ├── 6023f2b1-ea6e-485b-9ac2-8decd5f7820d
│ │ ├── b38a5e37-fac4-4c23-a0c4-7359adff619c
│ │ ├── b38a5e37-fac4-4c23-a0c4-7359adff619c.lease
│ │ └── b38a5e37-fac4-4c23-a0c4-7359adff619c.meta
│ ├── 685309b1-1ae9-45f3-90c3-d719a594482d
│ │ ├── 9eddcf51-fd15-4de5-a4b6-a83a9082dee0
│ │ ├── 9eddcf51-fd15-4de5-a4b6-a83a9082dee0.lease
│ │ └── 9eddcf51-fd15-4de5-a4b6-a83a9082dee0.meta
│ ├── 74f1b2e7-2483-4e4d-8301-819bcd99129e
│ │ ├── c1888b6a-c48e-46ce-9677-02e172ef07af
│ │ ├── c1888b6a-c48e-46ce-9677-02e172ef07af.lease
│ │ └── c1888b6a-c48e-46ce-9677-02e172ef07af.meta
│ └── 77082dd8-7cb5-41cc-a69f-0f4c0380db23
│ ├── 38d552c5-689d-47b7-9eea-adb308da8027
│ ├── 38d552c5-689d-47b7-9eea-adb308da8027.lease
│ └── 38d552c5-689d-47b7-9eea-adb308da8027.meta
└── master
├── tasks
│ ├── 150927c5-bae6-45e4-842c-a7ba229fc3ba
│ │ └── 150927c5-bae6-45e4-842c-a7ba229fc3ba.job.0
│ ├── 21bba697-26e6-4fd8-ac7c-76f86b458368.temp
│ ├── 26c580b8-cdb2-4d21-9bea-96e0788025e6.temp
│ ├── 2e0e347c-fd01-404f-9459-ef175c82c354.backup
│ │ └── 2e0e347c-fd01-404f-9459-ef175c82c354.task
│ ├── 43f17022-e003-4e9f-81ec-4a01582223bd.backup
│ │ └── 43f17022-e003-4e9f-81ec-4a01582223bd.task
│ ├── 5055f61a-4cc8-459f-8fe5-19427b74a4f2.temp
│ ├── 6826c8f5-b9df-498e-a576-af0c4e7fe69c
│ │ └── 6826c8f5-b9df-498e-a576-af0c4e7fe69c.task
│ ├── 78ed90b0-2a87-4c48-8204-03d4b0bd7694
│ │ └── 78ed90b0-2a87-4c48-8204-03d4b0bd7694.job.0
│ ├── 7c7799a5-d28e-4b42-86ee-84bb8822e82f.temp
│ ├── 95d29b8c-23d9-4d1a-b995-2ba364970893
│ ├── 95d29b8c-23d9-4d1a-b995-2ba364970893.temp
│ ├── a1fa934a-5ea7-4160-ab8c-7e3476dc2676.backup
│ │ └── a1fa934a-5ea7-4160-ab8c-7e3476dc2676.task
│ ├── bcee8725-efde-4848-a108-01c262625aaa
│ │ └── bcee8725-efde-4848-a108-01c262625aaa.job.0
│ ├── c0b5a032-c4a9-4648-b348-c2a5cf4d6cad.temp
│ ├── ce7e2ebf-2c28-435d-b359-14d0da2e9011
│ └── ce7e2ebf-2c28-435d-b359-14d0da2e9011.temp
└── vms
29 directories, 31 files
# Finding the XML file for hosted-engine VM
root@medusa /]# cd /gluster_bricks/vmstore/vmstore/qemu/
[root@medusa qemu]# ls
ns01.xml ns02.xml
[root@medusa qemu]# ls -alh
total 36K
drwxr-xr-x. 2 root root 38 Sep 17 10:19 .
drwxr-xr-x. 8 vdsm kvm 8.0K Jan 15 11:26 ..
-rw-------. 2 qemu qemu 4.7K Sep 17 07:19 ns01.xml
-rw-------. 2 root root 4.7K Sep 17 10:19 ns02.xml
[root@medusa qemu]# cat ns01.xml
<!--
WARNING: THIS IS AN AUTO-GENERATED FILE. CHANGES TO IT ARE LIKELY TO BE
OVERWRITTEN AND LOST. Changes to this xml configuration should be made using:
virsh edit ns01
or other application using the libvirt API.
-->
<domain type='kvm'>
<name>ns01</name>
<uuid>0bfd4ad4-b405-4154-94da-ac3261ebc17e</uuid>
<title>ns01</title>
<memory unit='KiB'>16777216</memory>
<currentMemory unit='KiB'>4184064</currentMemory>
<vcpu placement='static'>4</vcpu>
<os>
<type arch='x86_64'
machine='pc-i440fx-rhel7.6.0'>hvm</type>
<boot dev='hd'/>
</os>
<features>
<acpi/>
<apic/>
<vmport state='off'/>
</features>
<cpu mode='host-model' check='partial'>
<model fallback='allow'/>
</cpu>
<clock offset='utc'>
<timer name='rtc' tickpolicy='catchup'/>
<timer name='pit' tickpolicy='delay'/>
<timer name='hpet' present='no'/>
</clock>
<on_poweroff>destroy</on_poweroff>
<on_reboot>restart</on_reboot>
<on_crash>destroy</on_crash>
<pm>
<suspend-to-mem enabled='no'/>
<suspend-to-disk enabled='no'/>
</pm>
<devices>
<emulator>/usr/libexec/qemu-kvm</emulator>
<disk type='file' device='disk'>
<driver name='qemu' type='qcow2'/>
<source file='/media/vmstore/ns01.qcow2'/>
<target dev='vda' bus='virtio'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x07' function='0x0'/>
</disk>
<disk type='file' device='disk'>
<driver name='qemu' type='qcow2'/>
<source file='/media/vmstore/ns01_var.qcow2'/>
<target dev='vdb' bus='virtio'/>
<snip>
# Try to start VM via use of this XML file
[root@medusa qemu]# cp /gluster_bricks/vmstore/vmstore/qemu/ns01.xml /tmp
[root@medusa qemu]# virsh define /tmp/ns01.xml
Please enter your authentication name: admin
Please enter your password:
Domain ns01 defined from /tmp/ns01.xml
[root@medusa qemu]# virsh start /tmp/ns01.xml
Please enter your authentication name: admin
Please enter your password:
error: failed to get domain '/tmp/ns01.xml'
[root@medusa qemu]#
Hmm... working on it.... but if if you have suggestion / example .. that would help