Looks normal, hosted engine uses some extra files/leases to track some of it’s stuff.
Looks like you might have a gluster problem though, that IO error appears to be on your
hosted engines disk image. Check for split brains and try and initiate a heal on the
files, see what you get.
-Darrell
On Jul 2, 2015, at 11:33 AM, Christopher Young
<mexigabacho(a)gmail.com> wrote:
If anyone has an experience of the various IDs in hosted-engine.conf
and vm.conf for the Hosted Engine, I believe I need to just verify
everything. I tried a couple of changes, but I feel like I'm just
making this worse, so I've reverted them.
One thing I do not understand well is how a gluster-based (NFS) storage
domain for the hosted-engine has so many entries:
-------------
[root@orldc-dev-vnode02 ovirt-gluster-nfs:_engine]# find . -type f |
xargs ls -lah
ls: cannot access ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
-4df6-8856-fdb14df260e3: Input/output error
-rw-rw----. 1 vdsm kvm 1.0M Jul 2 11:20 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/ids
-rw-rw----. 1 vdsm kvm 16M Jul 1 19:54 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/inbox
-rw-rw----. 1 vdsm kvm 2.0M Jul 1 19:50 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/leases
-rw-r--r--. 1 vdsm kvm 482 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/metadata
-rw-rw----. 1 vdsm kvm 16M Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/outbox
-rw-rw----. 1 vdsm kvm 1.0M Jul 2 11:32 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
-4f74-a91f-3a3a52f255bf
-rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
-4f74-a91f-3a3a52f255bf.lease
-rw-r--r--. 1 vdsm kvm 284 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
-4f74-a91f-3a3a52f255bf.meta
-rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161
-4fa0-aab1-3cb52305dba7.lease
-rw-r--r--. 1 vdsm kvm 283 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161
-4fa0-aab1-3cb52305dba7.meta
-rw-rw----. 1 vdsm kvm 25G Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
-4d54-91a7-3dd87377c362
-rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
-4d54-91a7-3dd87377c362.lease
-rw-r--r--. 1 vdsm kvm 278 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
-4d54-91a7-3dd87377c362.meta
-rw-rw----. 1 vdsm kvm 1.0M Jul 1 17:50 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
-4df6-8856-fdb14df260e3.lease
-rw-r--r--. 1 vdsm kvm 278 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
-4df6-8856-fdb14df260e3.meta
-rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
-4226-9a00-5383d8937a83
-rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
-4226-9a00-5383d8937a83.lease
-rw-r--r--. 1 vdsm kvm 284 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
-4226-9a00-5383d8937a83.meta
-rw-rw----. 1 vdsm kvm 1004K Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
-46ac-8978-576a8a4a0399
-rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
-46ac-8978-576a8a4a0399.lease
-rw-r--r--. 1 vdsm kvm 283 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
-46ac-8978-576a8a4a0399.meta
-rw-r--r--. 1 vdsm kvm 384 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.job.0
-rw-r--r--. 1 vdsm kvm 277 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.0
-rw-r--r--. 1 vdsm kvm 417 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.1
-rw-r--r--. 1 vdsm kvm 107 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.result
-rw-r--r--. 1 vdsm kvm 299 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.task
-rwxr-xr-x. 1 vdsm kvm 0 Jul 2 11:22 ./__DIRECT_IO_TEST__
-------------
On Thu, 2015-07-02 at 11:05 -0400, Christopher Young wrote:
> That's actually a local NFS implementation (running on Gluster) so I
> wouldn't expect that it would factor in, but I just verified that all
> gluster hosts are in Permissive mode. Like I mentioned, I think I
> may
> have some ids wrong or bad links or something since this had
> previously
> been working fine, though I must be missing it.
>
> On Thu, 2015-07-02 at 16:08 +0200, Sandro Bonazzola wrote:
>> Il 02/07/2015 16:04, Christopher Young ha scritto:
>>> [root@orldc-dev-vnode02 ~]# getenforce
>>> Permissive
>>>
>>> It looks like it isn't SELinux.
>>
>> Checked also on the storage server ovirt-gluster-nfs?
>>
>>
>>>
>>> On Thu, 2015-07-02 at 09:53 +0200, Sandro Bonazzola wrote:
>>>> Il 02/07/2015 02:36, Christopher Young ha scritto:
>>>>> I'm sure I have worked through this before, but I've been
>>>>> banging
>>>>> my head against this one for a while now, and I think I'm too
>>>>>
>>>>> close
>>>>> to the issue.
>>>>>
>>>>> Basically, my hosted-engine won't start anymore. I do recall
>>>>>
>>>>> attempting to migrate it to a new gluster-based NFS share
>>>>> recently,
>>>>> but I could have
>>>>> sworn that was successful and working. I _think_ I have some
>>>>>
>>>>> sort
>>>>> of id issue with storage/volume/whatever id's, but I need
>>>>> some
>>>>> help
>>>>> digging through
>>>>> it if someone would be so kind.
>>>>>
>>>>> I have the following error in
>>>>> /var/log/libvirt/qemu/HostedEngine.log:
>>>>>
>>>>> ----------
>>>>> 2015-07-02T00:01:13.080952Z qemu-kvm: -drive
>>>>> file=/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
>>>>> -7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
>>>>> -a432
>>>>> -4df6-8856-fdb14df260e3,if=none,id=drive-virtio
>>>>> -disk0,format=raw,serial=5ead7b5d-50e8-4d6c-a0e5
>>>>> -bbe6d93dd836,cache=none,werror=stop,rerror=stop,aio=threads:
>>>>> could not open disk image
>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
>>>>> -7057fcee412c/5ead7b5d
>>>>> -50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856
>>>>> -fdb14df260e3:
>>>>> Could not
>>>>> refresh total sector count: Input/output error
>>>>> ----------
>>>>
>>>>
>>>> Please check selinux: ausearch -m avc
>>>> it might be selinux preventing access to the disk image.
>>>>
>>>>
>>>>
>>>>>
>>>>> I also am including a few command outputs that I'm sure might
>>>>>
>>>>> help:
>>>>>
>>>>> ----------
>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>> -7057fcee412c]#
>>>>> ls
>>>>> -la /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
>>>>> -7057fcee412c/
>>>>> total 8
>>>>> drwxr-xr-x. 2 vdsm kvm 80 Jul 1 20:04 .
>>>>> drwxr-xr-x. 3 vdsm kvm 60 Jul 1 20:04 ..
>>>>> lrwxrwxrwx. 1 vdsm kvm 128 Jul 1 20:04 1d80a60c-8f26-4448
>>>>> -9460
>>>>> -2c7b00ff75bf ->
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>> -4ad1
>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
>>>>> -2c7b00ff75bf
>>>>> lrwxrwxrwx. 1 vdsm kvm 128 Jul 1 20:04 23ac8897-b0c7-41d6
>>>>> -a7de
>>>>> -19f46ed78400 ->
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>> -4ad1
>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
>>>>> -19f46ed78400
>>>>>
>>>>> ----------
>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>> -7057fcee412c]#
>>>>> cat
>>>>> /etc/ovirt-hosted-engine/vm.conf
>>>>> vmId=6b7329f9-518a-4488-b1c4-2cd809f2f580
>>>>> memSize=5120
>>>>> display=vnc
>>>>> devices={index:2,iface:ide,address:{ controller:0,
>>>>> target:0,unit:0,
>>>>> bus:1,
>>>>> type:drive},specParams:{},readonly:true,deviceId:77924fc2
>>>>> -c5c9
>>>>> -408b
>>>>> -97d3-cd0b0d11a62c,path:/home/tmp/CentOS-6.6-x86_64
>>>>> -minimal.iso,device:cdrom,shared:false,type:disk}
>>>>> devices={index:0,iface:virtio,format:raw,poolID:00000000-0000
>>>>> -0000
>>>>> -0000-000000000000,volumeID:eeb2d821-a432-4df6-8856
>>>>> -fdb14df260e3,imageID:5ead7b5d-50e8-4d6c-a0e5
>>>>> -bbe6d93dd836,specParams:{},readonly:false,domainID:4e3017eb
>>>>> -d062
>>>>> -4ad1-9df8-7057fcee412c,optional:false,deviceId:5ead7b5d-50e8
>>>>> -4d6c
>>>>> -a0e5-bbe6d93dd836,address:{bus:0x00,
>>>>> slot:0x06, domain:0x0000, type:pci,
>>>>> function:0x0},device:disk,shared:exclusive,propagateErrors:of
>>>>> f,
>>>>> type
>>>>> :disk,bootOrder:1}
>>>>> devices={device:scsi,model:virtio-scsi,type:controller}
>>>>> devices={nicModel:pv,macAddr:00:16:3e:0e:d0:68,linkActive:tru
>>>>> e,
>>>>> netw
>>>>> ork:ovirtmgmt,filter:vdsm-no-mac
>>>>> -spoofing,specParams:{},deviceId:f70ba622-6ac8-4c06-a005
>>>>> -0ebd940a15b2,address:{bus:0x00,
>>>>> slot:0x03, domain:0x0000, type:pci,
>>>>> function:0x0},device:bridge,type:interface}
>>>>> devices={device:console,specParams:{},type:console,deviceId:9
>>>>> 83
>>>>> 86e6
>>>>> c-ae56-4b6d-9bfb-c72bbd299ad1,alias:console0}
>>>>> vmName=HostedEngine
>>>>> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,
>>>>> sr
>>>>> ecor
>>>>> d,ssmartcard,susbredir
>>>>> smp=2
>>>>> cpuType=Westmere
>>>>> emulatedMachine=pc
>>>>> ----------
>>>>>
>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>> -7057fcee412c]#
>>>>> cat
>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf
>>>>> fqdn=orldc-dev-vengine01.*******
>>>>> vm_disk_id=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
>>>>> vmid=6b7329f9-518a-4488-b1c4-2cd809f2f580
>>>>> storage=ovirt-gluster-nfs:/engine
>>>>> conf=/etc/ovirt-hosted-engine/vm.conf
>>>>> service_start_time=0
>>>>> host_id=1
>>>>> console=vnc
>>>>> domainType=nfs3
>>>>> spUUID=379cf161-d5b1-4c20-bb71-e3ca5d2ccd6b
>>>>> sdUUID=4e3017eb-d062-4ad1-9df8-7057fcee412c
>>>>> connectionUUID=0d1b50ac-cf3f-4cd7-90df-3c3a6d11a984
>>>>> ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
>>>>> ca_subject="C=EN, L=Test, O=Test, CN=Test"
>>>>> vdsm_use_ssl=true
>>>>> gateway=10.16.3.1
>>>>> bridge=ovirtmgmt
>>>>> metadata_volume_UUID=dd9f373c-d161-4fa0-aab1-3cb52305dba7
>>>>> metadata_image_UUID=23ac8897-b0c7-41d6-a7de-19f46ed78400
>>>>> lockspace_volume_UUID=d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
>>>>> lockspace_image_UUID=1d80a60c-8f26-4448-9460-2c7b00ff75bf
>>>>>
>>>>> # The following are used only for iSCSI storage
>>>>> iqn=
>>>>> portal=
>>>>> user=
>>>>> password=
>>>>> port=
>>>>> ----------
>>>>>
>>>>> (mount output for the NFS share this should be running on
>>>>> (gluster
>>>>> -based):
>>>>>
>>>>> ovirt-gluster-nfs:/engine on /rhev/data-center/mnt/ovirt
>>>>> -gluster
>>>>> -nfs:_engine type nfs
>>>>> (rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>> ft
>>>>> ,nol
>>>>> ock,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>> dd
>>>>> r=10
>>>>> .16.3.30,mountvers=3,mountport=38465,mountproto=tcp,local_loc
>>>>> k=
>>>>> all,
>>>>> addr=10.16.3.30)
>>>>>
>>>>> ----------
>>>>>
>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>> -7057fcee412c]#
>>>>> tree
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs\:_engine/
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/
>>>>> ├── 4e3017eb-d062-4ad1-9df8-7057fcee412c
>>>>> │ ├── dom_md
>>>>> │ │ ├── ids
>>>>> │ │ ├── inbox
>>>>> │ │ ├── leases
>>>>> │ │ ├── metadata
>>>>> │ │ └── outbox
>>>>> │ ├── ha_agent
>>>>> │ │ ├── hosted-engine.lockspace ->
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>> -4ad1
>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
>>>>> -2c7b00ff75bf/d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
>>>>> │ │ └── hosted-engine.metadata ->
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>> -4ad1
>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
>>>>> -19f46ed78400/dd9f373c-d161-4fa0-aab1-3cb52305dba7
>>>>> │ ├── images
>>>>> │ │ ├── 1d80a60c-8f26-4448-9460-2c7b00ff75bf
>>>>> │ │ │ ├── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
>>>>> │ │ │ ├── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf.lease
>>>>> │ │ │ └── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf.meta
>>>>> │ │ ├── 23ac8897-b0c7-41d6-a7de-19f46ed78400
>>>>> │ │ │ ├── dd9f373c-d161-4fa0-aab1-3cb52305dba7 ->
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>> -4ad1
>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
>>>>> -19f46ed78400
>>>>> │ │ │ ├── dd9f373c-d161-4fa0-aab1-3cb52305dba7.lease
>>>>> │ │ │ └── dd9f373c-d161-4fa0-aab1-3cb52305dba7.meta
>>>>> │ │ ├── 3278c444-d92a-4cb9-87d6-9669c6e4993e
>>>>> │ │ │ ├── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362
>>>>> │ │ │ ├── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362.lease
>>>>> │ │ │ └── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362.meta
>>>>> │ │ ├── 5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
>>>>> │ │ │ ├── eeb2d821-a432-4df6-8856-fdb14df260e3.lease
>>>>> │ │ │ └── eeb2d821-a432-4df6-8856-fdb14df260e3.meta
>>>>> │ │ ├── 6064179f-2720-4db9-a7c4-a97e044c2238
>>>>> │ │ │ ├── 05afaa26-95af-4226-9a00-5383d8937a83
>>>>> │ │ │ ├── 05afaa26-95af-4226-9a00-5383d8937a83.lease
>>>>> │ │ │ └── 05afaa26-95af-4226-9a00-5383d8937a83.meta
>>>>> │ │ └── bb9d9a37-4f91-4973-ba9e-72ee81aed0b6
>>>>> │ │ ├── 5acb27b3-62c5-46ac-8978-576a8a4a0399
>>>>> │ │ ├── 5acb27b3-62c5-46ac-8978-576a8a4a0399.lease
>>>>> │ │ └── 5acb27b3-62c5-46ac-8978-576a8a4a0399.meta
>>>>> │ └── master
>>>>> │ ├── tasks
>>>>> │ │ └── fef13299-0e7f-4c7a-a399-092a1235faab
>>>>> │ │ ├── fef13299-0e7f-4c7a-a399
>>>>> -092a1235faab.job.0
>>>>> │ │ ├── fef13299-0e7f-4c7a-a399
>>>>> -092a1235faab.recover.0
>>>>> │ │ ├── fef13299-0e7f-4c7a-a399
>>>>> -092a1235faab.recover.1
>>>>> │ │ ├── fef13299-0e7f-4c7a-a399
>>>>> -092a1235faab.result
>>>>> │ │ └── fef13299-0e7f-4c7a-a399-092a1235faab.task
>>>>> │ └── vms
>>>>> └── __DIRECT_IO_TEST__
>>>>>
>>>>> 16 directories, 28 files
>>>>> ----------
>>>>>
>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>> -7057fcee412c]#
>>>>> tree
>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c
>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c
>>>>> ├── 1d80a60c-8f26-4448-9460-2c7b00ff75bf ->
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>> -4ad1
>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
>>>>> -2c7b00ff75bf
>>>>> └── 23ac8897-b0c7-41d6-a7de-19f46ed78400 ->
>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>> -4ad1
>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
>>>>> -19f46ed78400
>>>>>
>>>>> 2 directories, 0 files
>>>>> ----------
>>>>>
>>>>>
>>>>> Can you anyone point me in the right direction here?
>>>>> Eventually,
>>>>> I'd like to get this hosted-engine running in HA mode (I had
>>>>> started this process
>>>>> with ctdb/gluster/etc. via the well-known guide out there),
>>>>> but
>>>>> I'm
>>>>> thinking that at this point, I should get this up and running
>>>>>
>>>>> again, export
>>>>> everything, and perhaps just build a new hosted-engine and
>>>>> import
>>>>> things into (if that is possible). Right now, though, I just
>>>>>
>>>>> need
>>>>> to get this
>>>>> hosted-engine run.
>>>>>
>>>>> Many many thanks!
>>>>>
>>>>> Chris
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> Users(a)ovirt.org
>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>
>>>>
>>>>
>>
>>
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users