On Fri, Jan 15, 2021, 22:04 penguin pages <jeremey.wise@gmail.com> wrote:

Thanks for replies.

Here is where it is at:

# Two nodes think no VMs exist
[root@odin ~]# vdsm-client Host getVMList
[]

#One showing one VM but down
[root@medusa ~]# vdsm-client Host getVMList
[
    {
        "status": "Down",
        "statusTime": "2153886148",
        "vmId": "69ab4f82-1a53-42c8-afca-210a3a2715f1"
    }
]
[root@medusa ~]# vdsm-client Host getAllVmStats
[
    {
        "exitCode": 1,
        "exitMessage": "VM terminated with error",
        "exitReason": 1,
        "status": "Down",
        "statusTime": "2153916276",
        "vmId": "69ab4f82-1a53-42c8-afca-210a3a2715f1"
    }
]
[root@medusa ~]# vdsm-client VM cont vmID="69ab4f82-1a53-42c8-afca-210a3a2715f1"
vdsm-client: Command VM.cont with args {'vmID': '69ab4f82-1a53-42c8-afca-210a3a2715f1'} failed:
(code=16, message=Unexpected exception)


# Assuming that ID represents the hosted-engine I tried to start it
[root@medusa ~]# hosted-engine --vm-start
The hosted engine configuration has not been retrieved from shared storage. Please ensure that ovirt-ha-agent is running and the storage server is reachable.

# Back to ovirt-ha-agent being fubar and stoping things.

I have about 8 or so VMs on the cluster. Two are my IDM nodes which has DNS and other core services.. which is what I am really trying to get up .. even if manual until I figure out oVirt issue.  I think you are correct. "engine" volume is for just the engine.  Data is where the other VMs are

[root@medusa images]# tree
.
├── 335c6b1a-d8a5-4664-9a9c-39744d511af8
│   ├── 579323ad-bf7b-479b-b682-6e1e234a7908
│   ├── 579323ad-bf7b-479b-b682-6e1e234a7908.lease
│   └── 579323ad-bf7b-479b-b682-6e1e234a7908.meta
├── d318cb8f-743a-461b-b246-75ffcde6bc5a
│   ├── c16877d0-eb23-42ef-a06e-a3221ea915fc
│   ├── c16877d0-eb23-42ef-a06e-a3221ea915fc.lease
│   └── c16877d0-eb23-42ef-a06e-a3221ea915fc.meta
└── junk
    ├── 296163f2-846d-4a2c-9a4e-83a58640b907
    │   ├── 376b895f-e0f2-4387-b038-fbef4705fbcc
    │   ├── 376b895f-e0f2-4387-b038-fbef4705fbcc.lease
    │   └── 376b895f-e0f2-4387-b038-fbef4705fbcc.meta
    ├── 45a478d7-4c1b-43e8-b106-7acc75f066fa
    │   ├── b5249e6c-0ba6-4302-8e53-b74d2b919d20
    │   ├── b5249e6c-0ba6-4302-8e53-b74d2b919d20.lease
    │   └── b5249e6c-0ba6-4302-8e53-b74d2b919d20.meta
    ├── d8b708c1-5762-4215-ae1f-0e57444c99ad
    │   ├── 2536ca6d-3254-4cdc-bbd8-349ec1b8a0e9
    │   ├── 2536ca6d-3254-4cdc-bbd8-349ec1b8a0e9.lease
    │   └── 2536ca6d-3254-4cdc-bbd8-349ec1b8a0e9.meta
    └── eaf12f3c-301f-4b61-b5a1-0c6d0b0a7f7b
        ├── fbf3bf59-a23a-4c6f-b66e-71369053b406
        ├── fbf3bf59-a23a-4c6f-b66e-71369053b406.lease
        └── fbf3bf59-a23a-4c6f-b66e-71369053b406.meta

7 directories, 18 files
[root@medusa images]# cd /media/engine/
[root@medusa engine]# ls
3afc47ba-afb9-413f-8de5-8d9a2f45ecde
[root@medusa engine]# tree
.
└── 3afc47ba-afb9-413f-8de5-8d9a2f45ecde
    ├── dom_md
    │   ├── ids
    │   ├── inbox
    │   ├── leases
    │   ├── metadata
    │   ├── outbox
    │   └── xleases
    ├── ha_agent
    ├── images
    │   ├── 1dc69552-dcc6-484d-8149-86c93ff4b8cc
    │   │   ├── e4e26573-09a5-43fa-91ec-37d12de46480
    │   │   ├── e4e26573-09a5-43fa-91ec-37d12de46480.lease
    │   │   └── e4e26573-09a5-43fa-91ec-37d12de46480.meta
    │   ├── 375d2483-ee83-4cad-b421-a5a70ec06ba6
    │   │   ├── f936d4be-15e3-4983-8bf0-9ba5b97e638a
    │   │   ├── f936d4be-15e3-4983-8bf0-9ba5b97e638a.lease
    │   │   └── f936d4be-15e3-4983-8bf0-9ba5b97e638a.meta
    │   ├── 6023f2b1-ea6e-485b-9ac2-8decd5f7820d
    │   │   ├── b38a5e37-fac4-4c23-a0c4-7359adff619c
    │   │   ├── b38a5e37-fac4-4c23-a0c4-7359adff619c.lease
    │   │   └── b38a5e37-fac4-4c23-a0c4-7359adff619c.meta
    │   ├── 685309b1-1ae9-45f3-90c3-d719a594482d
    │   │   ├── 9eddcf51-fd15-4de5-a4b6-a83a9082dee0
    │   │   ├── 9eddcf51-fd15-4de5-a4b6-a83a9082dee0.lease
    │   │   └── 9eddcf51-fd15-4de5-a4b6-a83a9082dee0.meta
    │   ├── 74f1b2e7-2483-4e4d-8301-819bcd99129e
    │   │   ├── c1888b6a-c48e-46ce-9677-02e172ef07af
    │   │   ├── c1888b6a-c48e-46ce-9677-02e172ef07af.lease
    │   │   └── c1888b6a-c48e-46ce-9677-02e172ef07af.meta
    │   └── 77082dd8-7cb5-41cc-a69f-0f4c0380db23
    │       ├── 38d552c5-689d-47b7-9eea-adb308da8027
    │       ├── 38d552c5-689d-47b7-9eea-adb308da8027.lease
    │       └── 38d552c5-689d-47b7-9eea-adb308da8027.meta
    └── master
        ├── tasks
        │   ├── 150927c5-bae6-45e4-842c-a7ba229fc3ba
        │   │   └── 150927c5-bae6-45e4-842c-a7ba229fc3ba.job.0
        │   ├── 21bba697-26e6-4fd8-ac7c-76f86b458368.temp
        │   ├── 26c580b8-cdb2-4d21-9bea-96e0788025e6.temp
        │   ├── 2e0e347c-fd01-404f-9459-ef175c82c354.backup
        │   │   └── 2e0e347c-fd01-404f-9459-ef175c82c354.task
        │   ├── 43f17022-e003-4e9f-81ec-4a01582223bd.backup
        │   │   └── 43f17022-e003-4e9f-81ec-4a01582223bd.task
        │   ├── 5055f61a-4cc8-459f-8fe5-19427b74a4f2.temp
        │   ├── 6826c8f5-b9df-498e-a576-af0c4e7fe69c
        │   │   └── 6826c8f5-b9df-498e-a576-af0c4e7fe69c.task
        │   ├── 78ed90b0-2a87-4c48-8204-03d4b0bd7694
        │   │   └── 78ed90b0-2a87-4c48-8204-03d4b0bd7694.job.0
        │   ├── 7c7799a5-d28e-4b42-86ee-84bb8822e82f.temp
        │   ├── 95d29b8c-23d9-4d1a-b995-2ba364970893
        │   ├── 95d29b8c-23d9-4d1a-b995-2ba364970893.temp
        │   ├── a1fa934a-5ea7-4160-ab8c-7e3476dc2676.backup
        │   │   └── a1fa934a-5ea7-4160-ab8c-7e3476dc2676.task
        │   ├── bcee8725-efde-4848-a108-01c262625aaa
        │   │   └── bcee8725-efde-4848-a108-01c262625aaa.job.0
        │   ├── c0b5a032-c4a9-4648-b348-c2a5cf4d6cad.temp
        │   ├── ce7e2ebf-2c28-435d-b359-14d0da2e9011
        │   └── ce7e2ebf-2c28-435d-b359-14d0da2e9011.temp
        └── vms

29 directories, 31 files


# Finding the XML file for hosted-engine VM
root@medusa /]# cd /gluster_bricks/vmstore/vmstore/qemu/
[root@medusa qemu]# ls
ns01.xml  ns02.xml
[root@medusa qemu]# ls -alh
total 36K
drwxr-xr-x. 2 root root   38 Sep 17 10:19 .
drwxr-xr-x. 8 vdsm kvm  8.0K Jan 15 11:26 ..
-rw-------. 2 qemu qemu 4.7K Sep 17 07:19 ns01.xml
-rw-------. 2 root root 4.7K Sep 17 10:19 ns02.xml
[root@medusa qemu]# cat ns01.xml
<!--
WARNING: THIS IS AN AUTO-GENERATED FILE. CHANGES TO IT ARE LIKELY TO BE
OVERWRITTEN AND LOST. Changes to this xml configuration should be made using:
  virsh edit ns01
or other application using the libvirt API.
-->

<domain type='kvm'>
  <name>ns01</name>
  <uuid>0bfd4ad4-b405-4154-94da-ac3261ebc17e</uuid>
  <title>ns01</title>
  <memory unit='KiB'>16777216</memory>
  <currentMemory unit='KiB'>4184064</currentMemory>
  <vcpu placement='static'>4</vcpu>
  <os>
    <type arch='x86_64' machine='pc-i440fx-rhel7.6.0'>hvm</type>
    <boot dev='hd'/>
  </os>
  <features>
    <acpi/>
    <apic/>
    <vmport state='off'/>
  </features>
  <cpu mode='host-model' check='partial'>
    <model fallback='allow'/>
  </cpu>
  <clock offset='utc'>
    <timer name='rtc' tickpolicy='catchup'/>
    <timer name='pit' tickpolicy='delay'/>
    <timer name='hpet' present='no'/>
  </clock>
  <on_poweroff>destroy</on_poweroff>
  <on_reboot>restart</on_reboot>
  <on_crash>destroy</on_crash>
  <pm>
    <suspend-to-mem enabled='no'/>
    <suspend-to-disk enabled='no'/>
  </pm>
  <devices>
    <emulator>/usr/libexec/qemu-kvm</emulator>
    <disk type='file' device='disk'>
      <driver name='qemu' type='qcow2'/>
      <source file='/media/vmstore/ns01.qcow2'/>
      <target dev='vda' bus='virtio'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0'/>
    </disk>
    <disk type='file' device='disk'>
      <driver name='qemu' type='qcow2'/>
      <source file='/media/vmstore/ns01_var.qcow2'/>
      <target dev='vdb' bus='virtio'/>
<snip>

# Try to start VM via use of this XML file
[root@medusa qemu]# cp /gluster_bricks/vmstore/vmstore/qemu/ns01.xml /tmp
[root@medusa qemu]# virsh define /tmp/ns01.xml
Please enter your authentication name: admin
Please enter your password:
Domain ns01 defined from /tmp/ns01.xml
Seems that it was defined

[root@medusa qemu]# virsh start /tmp/ns01.xml
Please enter your authentication name: admin
Please enter your password:
error: failed to get domain '/tmp/ns01.xml'
You need to start the defined vm not the xml. Did you try to connect with:

virsh -c qemu:///system?authfile=/etc/ovirt-hosted-
engine/virsh_auth.conf

Otherwise you can configure also libvirt to authenticate with a sasl account. Example:

saslpasswd2 -a libvirt fred

At /etc/libvirt/auth.conf:

[credentials-test]
authname=fred
password=123456

[auth-libvirt-medusa]
credentials=test

Check https://libvirt.org/auth.html#ACL_server_username for details. 

[root@medusa qemu]#


Hmm... working on it.... but if if you have suggestion / example .. that would help
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/XM2RKVE2OYNX7HCM5HK6GSTFNNAWFWXR/