One more thing to add, I've also been seeing a lot of this in the syslog as
well:
Jan 12 10:10:49 cultivar2 journal: vdsm jsonrpc.JsonRpcServer ERROR
Internal server error#012Traceback (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 606, in
_handle_request#012 res = method(**params)#012 File
"/usr/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 197, in
_dynamicMethod#012 result = fn(*methodArgs)#012 File "<string>", line
2, in getAllVmIoTunePolicies#012 File
"/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in
method#012 ret = func(*args, **kwargs)#012 File
"/usr/lib/python2.7/site-packages/vdsm/API.py", line 1354, in
getAllVmIoTunePolicies#012 io_tune_policies_dict =
self._cif.getAllVmIoTunePolicies()#012 File
"/usr/lib/python2.7/site-packages/vdsm/clientIF.py", line 524, in
getAllVmIoTunePolicies#012 'current_values': v.getIoTune()}#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 3481, in
getIoTune#012 result = self.getIoTuneResponse()#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 3500, in
getIoTuneResponse#012 res = self._dom.blockIoTune(#012 File
"/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 47, in
__getattr__#012 % self.vmid)#012NotConnectedError: VM
'4013c829-c9d7-4b72-90d5-6fe58137504c' was not defined yet or was undefined
On Fri, Jan 12, 2018 at 10:09 AM, Jayme <jaymef(a)gmail.com> wrote:
Thanks for the help thus far. Storage could be related but all other
VMs
on same storage are running ok. The storage is mounted via NFS from within
one of the three hosts, I realize this is not ideal. This was setup by a
previous admin more as a proof of concept and VMs were put on there that
should not have been placed in a proof of concept environment.. it was
intended to be rebuilt with proper storage down the road.
So the storage is on HOST0 and the other hosts mount NFS
cultivar0.grove.silverorange.com:/exports/data 4861742080
1039352832 3822389248 22% /rhev/data-center/mnt/cultivar
0.grove.silverorange.com:_exports_data
cultivar0.grove.silverorange.com:/exports/iso 4861742080
1039352832 3822389248 22% /rhev/data-center/mnt/cultivar
0.grove.silverorange.com:_exports_iso
cultivar0.grove.silverorange.com:/exports/import_export 4861742080
1039352832 3822389248 22% /rhev/data-center/mnt/cultivar
0.grove.silverorange.com:_exports_import__export
cultivar0.grove.silverorange.com:/exports/hosted_engine 4861742080
1039352832 3822389248 22% /rhev/data-center/mnt/cultivar
0.grove.silverorange.com:_exports_hosted__engine
Like I said, the VM data storage itself seems to be working ok, as all
other VMs appear to be running.
I'm curious why the broker log says this file is not found when it is
correct and I can see the file at that path:
RequestError: failed to read metadata: [Errno 2] No such file or
directory: '/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/
14a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8'
ls -al /var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/1
4a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8
-rw-rw----. 1 vdsm kvm 1028096 Jan 12 09:59 /var/run/vdsm/storage/248f46f0
-d793-4581-9810-c9d965e2f286/14a20941-1b84-4b82-be8f-ace38d7
c037a/8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8
Is this due to the symlink problem you guys are referring to that was
addressed in RC1 or something else? Could there possibly be a permissions
problem somewhere?
Assuming that all three hosts have 4.2 rpms installed and the host engine
will not start is it safe for me to update hosts to 4.2 RC1 rpms? Or
perhaps install that repo and *only* update the ovirt HA packages?
Assuming that I cannot yet apply the same updates to the inaccessible
hosted engine VM.
I should also mention one more thing. I originally upgraded the engine VM
first using new RPMS then engine-setup. It failed due to not being in
global maintenance, so I set global maintenance and ran it again, which
appeared to complete as intended but never came back up after. Just in
case this might have anything at all to do with what could have happened.
Thanks very much again, I very much appreciate the help!
- Jayme
On Fri, Jan 12, 2018 at 8:44 AM, Simone Tiraboschi <stirabos(a)redhat.com>
wrote:
>
>
> On Fri, Jan 12, 2018 at 11:11 AM, Martin Sivak <msivak(a)redhat.com> wrote:
>
>> Hi,
>>
>> the hosted engine agent issue might be fixed by restarting
>> ovirt-ha-broker or updating to newest ovirt-hosted-engine-ha and
>> -setup. We improved handling of the missing symlink.
>>
>
> Available just in oVirt 4.2.1 RC1
>
>
>>
>> All the other issues seem to point to some storage problem I am afraid.
>>
>> You said you started the VM, do you see it in virsh -r list?
>>
>> Best regards
>>
>> Martin Sivak
>>
>> On Thu, Jan 11, 2018 at 10:00 PM, Jayme <jaymef(a)gmail.com> wrote:
>> > Please help, I'm really not sure what else to try at this point.
>> Thank you
>> > for reading!
>> >
>> >
>> > I'm still working on trying to get my hosted engine running after a
>> botched
>> > upgrade to 4.2. Storage is NFS mounted from within one of the hosts.
>> Right
>> > now I have 3 centos7 hosts that are fully updated with yum packages
>> from
>> > ovirt 4.2, the engine was fully updated with yum packages and failed
>> to come
>> > up after reboot. As of right now, everything should have full yum
>> updates
>> > and all having 4.2 rpms. I have global maintenance mode on right now
>> and
>> > started hosted-engine on one of the three host and the status is
>> currently:
>> > Engine status : {"reason": "failed liveliness check”;
"health": "bad",
>> "vm":
>> > "up", "detail": "Up"}
>> >
>> >
>> > this is what I get when trying to enter hosted-vm --console
>> >
>> >
>> > The engine VM is running on this host
>> >
>> > error: failed to get domain 'HostedEngine'
>> >
>> > error: Domain not found: no domain with matching name
'HostedEngine'
>> >
>> >
>> > Here are logs from various sources when I start the VM on HOST3:
>> >
>> >
>> > hosted-engine --vm-start
>> >
>> > Command VM.getStats with args {'vmID':
>> > '4013c829-c9d7-4b72-90d5-6fe58137504c'} failed:
>> >
>> > (code=1, message=Virtual machine does not exist: {'vmId':
>> > u'4013c829-c9d7-4b72-90d5-6fe58137504c'})
>> >
>> >
>> > Jan 11 16:55:57 cultivar3 systemd-machined: New machine
>> qemu-110-Cultivar.
>> >
>> > Jan 11 16:55:57 cultivar3 systemd: Started Virtual Machine
>> > qemu-110-Cultivar.
>> >
>> > Jan 11 16:55:57 cultivar3 systemd: Starting Virtual Machine
>> > qemu-110-Cultivar.
>> >
>> > Jan 11 16:55:57 cultivar3 kvm: 3 guests now active
>> >
>> >
>> > ==> /var/log/vdsm/vdsm.log <==
>> >
>> > File "/usr/lib/python2.7/site-packages/vdsm/common/api.py",
line
>> 48, in
>> > method
>> >
>> > ret = func(*args, **kwargs)
>> >
>> > File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py",
line
>> 2718, in
>> > getStorageDomainInfo
>> >
>> > dom = self.validateSdUUID(sdUUID)
>> >
>> > File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py",
line
>> 304, in
>> > validateSdUUID
>> >
>> > sdDom.validate()
>> >
>> > File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py",
>> line 515,
>> > in validate
>> >
>> > raise se.StorageDomainAccessError(self.sdUUID)
>> >
>> > StorageDomainAccessError: Domain is either partially accessible or
>> entirely
>> > inaccessible: (u'248f46f0-d793-4581-9810-c9d965e2f286',)
>> >
>> > jsonrpc/2::ERROR::2018-01-11
>> > 16:55:16,144::dispatcher::82::storage.Dispatcher::(wrapper) FINISH
>> > getStorageDomainInfo error=Domain is either partially accessible or
>> entirely
>> > inaccessible: (u'248f46f0-d793-4581-9810-c9d965e2f286',)
>> >
>> >
>> > ==> /var/log/libvirt/qemu/Cultivar.log <==
>> >
>> > LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
>> > QEMU_AUDIO_DRV=spice /usr/libexec/qemu-kvm -name
>> > guest=Cultivar,debug-threads=on -S -object
>> > secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/d
>> omain-108-Cultivar/master-key.aes
>> > -machine pc-i440fx-rhel7.3.0,accel=kvm,usb=off,dump-guest-core=off
>> -cpu
>> > Conroe -m 8192 -realtime mlock=off -smp
>> > 2,maxcpus=16,sockets=16,cores=1,threads=1 -uuid
>> > 4013c829-c9d7-4b72-90d5-6fe58137504c -smbios
>> > 'type=1,manufacturer=oVirt,product=oVirt
>> > Node,version=7-4.1708.el7.centos,serial=44454C4C-4300-1034-8
>> 035-CAC04F424331,uuid=4013c829-c9d7-4b72-90d5-6fe58137504c'
>> > -no-user-config -nodefaults -chardev
>> > socket,id=charmonitor,path=/var/lib/libvirt/qemu/domain-108-
>> Cultivar/monitor.sock,server,nowait
>> > -mon chardev=charmonitor,id=monitor,mode=control -rtc
>> > base=2018-01-11T20:33:19,driftfix=slew -global
>> > kvm-pit.lost_tick_policy=delay -no-hpet -no-reboot -boot strict=on
>> -device
>> > piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
>> > virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
>> > file=/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f
>> 286/c2dde892-f978-4dfc-a421-c8e04cf387f9/23aa0a66-fa6c-4967-
>> a1e5-fbe47c0cd705,format=raw,if=none,id=drive-virtio-disk0,
>> serial=c2dde892-f978-4dfc-a421-c8e04cf387f9,cache=none,
>> werror=stop,rerror=stop,aio=threads
>> > -device
>> > virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virti
>> o-disk0,id=virtio-disk0,bootindex=1
>> > -drive if=none,id=drive-ide0-1-0,readonly=on -device
>> > ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -netdev
>> > tap,fd=30,id=hostnet0,vhost=on,vhostfd=32 -device
>> > virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:7f:d6:83
>> ,bus=pci.0,addr=0x3
>> > -chardev
>> > socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.com.redhat.rhevm.vdsm,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel
>> 0,id=channel0,name=com.redhat.rhevm.vdsm
>> > -chardev
>> > socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.org.qemu.guest_agent.0,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel
>> 1,id=channel1,name=org.qemu.guest_agent.0
>> > -chardev spicevmc,id=charchannel2,name=vdagent -device
>> > virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel
>> 2,id=channel2,name=com.redhat.spice.0
>> > -chardev
>> > socket,id=charchannel3,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.org.ovirt.hosted-engine-
>> setup.0,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=4,chardev=charchannel
>> 3,id=channel3,name=org.ovirt.hosted-engine-setup.0
>> > -chardev pty,id=charconsole0 -device
>> > virtconsole,chardev=charconsole0,id=console0 -spice
>> > tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tl
>> s-channel=default,seamless-migration=on
>> > -device cirrus-vga,id=video0,bus=pci.0,addr=0x2 -object
>> > rng-random,id=objrng0,filename=/dev/urandom -device
>> > virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x5 -msg
>> timestamp=on
>> >
>> > 2018-01-11T20:33:19.699999Z qemu-kvm: -chardev pty,id=charconsole0:
>> char
>> > device redirected to /dev/pts/2 (label charconsole0)
>> >
>> > 2018-01-11 20:38:11.640+0000: shutting down, reason=shutdown
>> >
>> > 2018-01-11 20:39:02.122+0000: starting up libvirt version: 3.2.0,
>> package:
>> > 14.el7_4.7 (CentOS BuildSystem <
http://bugs.centos.org>,
>> > 2018-01-04-19:31:34,
c1bm.rdu2.centos.org), qemu version:
>> > 2.9.0(qemu-kvm-ev-2.9.0-16.el7_4.13.1), hostname: cultivar3
>> >
>> > LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
>> > QEMU_AUDIO_DRV=spice /usr/libexec/qemu-kvm -name
>> > guest=Cultivar,debug-threads=on -S -object
>> > secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/d
>> omain-109-Cultivar/master-key.aes
>> > -machine pc-i440fx-rhel7.3.0,accel=kvm,usb=off,dump-guest-core=off
>> -cpu
>> > Conroe -m 8192 -realtime mlock=off -smp
>> > 2,maxcpus=16,sockets=16,cores=1,threads=1 -uuid
>> > 4013c829-c9d7-4b72-90d5-6fe58137504c -smbios
>> > 'type=1,manufacturer=oVirt,product=oVirt
>> > Node,version=7-4.1708.el7.centos,serial=44454C4C-4300-1034-8
>> 035-CAC04F424331,uuid=4013c829-c9d7-4b72-90d5-6fe58137504c'
>> > -no-user-config -nodefaults -chardev
>> > socket,id=charmonitor,path=/var/lib/libvirt/qemu/domain-109-
>> Cultivar/monitor.sock,server,nowait
>> > -mon chardev=charmonitor,id=monitor,mode=control -rtc
>> > base=2018-01-11T20:39:02,driftfix=slew -global
>> > kvm-pit.lost_tick_policy=delay -no-hpet -no-reboot -boot strict=on
>> -device
>> > piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
>> > virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
>> > file=/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f
>> 286/c2dde892-f978-4dfc-a421-c8e04cf387f9/23aa0a66-fa6c-4967-
>> a1e5-fbe47c0cd705,format=raw,if=none,id=drive-virtio-disk0,
>> serial=c2dde892-f978-4dfc-a421-c8e04cf387f9,cache=none,
>> werror=stop,rerror=stop,aio=threads
>> > -device
>> > virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virti
>> o-disk0,id=virtio-disk0,bootindex=1
>> > -drive if=none,id=drive-ide0-1-0,readonly=on -device
>> > ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -netdev
>> > tap,fd=30,id=hostnet0,vhost=on,vhostfd=32 -device
>> > virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:7f:d6:83
>> ,bus=pci.0,addr=0x3
>> > -chardev
>> > socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.com.redhat.rhevm.vdsm,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel
>> 0,id=channel0,name=com.redhat.rhevm.vdsm
>> > -chardev
>> > socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.org.qemu.guest_agent.0,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel
>> 1,id=channel1,name=org.qemu.guest_agent.0
>> > -chardev spicevmc,id=charchannel2,name=vdagent -device
>> > virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel
>> 2,id=channel2,name=com.redhat.spice.0
>> > -chardev
>> > socket,id=charchannel3,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.org.ovirt.hosted-engine-
>> setup.0,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=4,chardev=charchannel
>> 3,id=channel3,name=org.ovirt.hosted-engine-setup.0
>> > -chardev pty,id=charconsole0 -device
>> > virtconsole,chardev=charconsole0,id=console0 -spice
>> > tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tl
>> s-channel=default,seamless-migration=on
>> > -device cirrus-vga,id=video0,bus=pci.0,addr=0x2 -object
>> > rng-random,id=objrng0,filename=/dev/urandom -device
>> > virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x5 -msg
>> timestamp=on
>> >
>> > 2018-01-11T20:39:02.380773Z qemu-kvm: -chardev pty,id=charconsole0:
>> char
>> > device redirected to /dev/pts/2 (label charconsole0)
>> >
>> > 2018-01-11 20:53:11.407+0000: shutting down, reason=shutdown
>> >
>> > 2018-01-11 20:55:57.210+0000: starting up libvirt version: 3.2.0,
>> package:
>> > 14.el7_4.7 (CentOS BuildSystem <
http://bugs.centos.org>,
>> > 2018-01-04-19:31:34,
c1bm.rdu2.centos.org), qemu version:
>> > 2.9.0(qemu-kvm-ev-2.9.0-16.el7_4.13.1), hostname:
>> >
cultivar3.grove.silverorange.com
>> >
>> > LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
>> > QEMU_AUDIO_DRV=spice /usr/libexec/qemu-kvm -name
>> > guest=Cultivar,debug-threads=on -S -object
>> > secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/d
>> omain-110-Cultivar/master-key.aes
>> > -machine pc-i440fx-rhel7.3.0,accel=kvm,usb=off,dump-guest-core=off
>> -cpu
>> > Conroe -m 8192 -realtime mlock=off -smp
>> > 2,maxcpus=16,sockets=16,cores=1,threads=1 -uuid
>> > 4013c829-c9d7-4b72-90d5-6fe58137504c -smbios
>> > 'type=1,manufacturer=oVirt,product=oVirt
>> > Node,version=7-4.1708.el7.centos,serial=44454C4C-4300-1034-8
>> 035-CAC04F424331,uuid=4013c829-c9d7-4b72-90d5-6fe58137504c'
>> > -no-user-config -nodefaults -chardev
>> > socket,id=charmonitor,path=/var/lib/libvirt/qemu/domain-110-
>> Cultivar/monitor.sock,server,nowait
>> > -mon chardev=charmonitor,id=monitor,mode=control -rtc
>> > base=2018-01-11T20:55:57,driftfix=slew -global
>> > kvm-pit.lost_tick_policy=delay -no-hpet -no-reboot -boot strict=on
>> -device
>> > piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
>> > virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
>> > file=/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f
>> 286/c2dde892-f978-4dfc-a421-c8e04cf387f9/23aa0a66-fa6c-4967-
>> a1e5-fbe47c0cd705,format=raw,if=none,id=drive-virtio-disk0,
>> serial=c2dde892-f978-4dfc-a421-c8e04cf387f9,cache=none,
>> werror=stop,rerror=stop,aio=threads
>> > -device
>> > virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virti
>> o-disk0,id=virtio-disk0,bootindex=1
>> > -drive if=none,id=drive-ide0-1-0,readonly=on -device
>> > ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -netdev
>> > tap,fd=30,id=hostnet0,vhost=on,vhostfd=32 -device
>> > virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:7f:d6:83
>> ,bus=pci.0,addr=0x3
>> > -chardev
>> > socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.com.redhat.rhevm.vdsm,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel
>> 0,id=channel0,name=com.redhat.rhevm.vdsm
>> > -chardev
>> > socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.org.qemu.guest_agent.0,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel
>> 1,id=channel1,name=org.qemu.guest_agent.0
>> > -chardev spicevmc,id=charchannel2,name=vdagent -device
>> > virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel
>> 2,id=channel2,name=com.redhat.spice.0
>> > -chardev
>> > socket,id=charchannel3,path=/var/lib/libvirt/qemu/channels/4
>> 013c829-c9d7-4b72-90d5-6fe58137504c.org.ovirt.hosted-engine-
>> setup.0,server,nowait
>> > -device
>> > virtserialport,bus=virtio-serial0.0,nr=4,chardev=charchannel
>> 3,id=channel3,name=org.ovirt.hosted-engine-setup.0
>> > -chardev pty,id=charconsole0 -device
>> > virtconsole,chardev=charconsole0,id=console0 -spice
>> > tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tl
>> s-channel=default,seamless-migration=on
>> > -device cirrus-vga,id=video0,bus=pci.0,addr=0x2 -object
>> > rng-random,id=objrng0,filename=/dev/urandom -device
>> > virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x5 -msg
>> timestamp=on
>> >
>> > 2018-01-11T20:55:57.468037Z qemu-kvm: -chardev pty,id=charconsole0:
>> char
>> > device redirected to /dev/pts/2 (label charconsole0)
>> >
>> >
>> > ==> /var/log/ovirt-hosted-engine-ha/broker.log <==
>> >
>> > File
>> > "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/bro
>> ker/storage_broker.py",
>> > line 151, in get_raw_stats
>> >
>> > f = os.open(path, direct_flag | os.O_RDONLY | os.O_SYNC)
>> >
>> > OSError: [Errno 2] No such file or directory:
>> > '/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/
>> 14a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-
>> 9f1e-f5b7ec1f1cf8'
>> >
>> > StatusStorageThread::ERROR::2018-01-11
>> > 16:55:15,761::status_broker::92::ovirt_hosted_engine_ha.brok
>> er.status_broker.StatusBroker.Update::(run)
>> > Failed to read state.
>> >
>> > Traceback (most recent call last):
>> >
>> > File
>> > "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/bro
>> ker/status_broker.py",
>> > line 88, in run
>> >
>> > self._storage_broker.get_raw_stats()
>> >
>> > File
>> > "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/bro
>> ker/storage_broker.py",
>> > line 162, in get_raw_stats
>> >
>> > .format(str(e)))
>> >
>> > RequestError: failed to read metadata: [Errno 2] No such file or
>> directory:
>> > '/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/
>> 14a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-
>> 9f1e-f5b7ec1f1cf8'
>> >
>> >
>> > ==> /var/log/ovirt-hosted-engine-ha/agent.log <==
>> >
>> > result = refresh_method()
>> >
>> > File
>> > "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/env
>> /config.py",
>> > line 519, in refresh_vm_conf
>> >
>> > content = self._get_file_content_from_shared_storage(VM)
>> >
>> > File
>> > "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/env
>> /config.py",
>> > line 484, in _get_file_content_from_shared_storage
>> >
>> > config_volume_path = self._get_config_volume_path()
>> >
>> > File
>> > "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/env
>> /config.py",
>> > line 188, in _get_config_volume_path
>> >
>> > conf_vol_uuid
>> >
>> > File
>> > "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib
>> /heconflib.py",
>> > line 358, in get_volume_path
>> >
>> > root=envconst.SD_RUN_DIR,
>> >
>> > RuntimeError: Path to volume 4838749f-216d-406b-b245-98d0343fcf7f not
>> found
>> > in /run/vdsm/storag
>> >
>> >
>> >
>> > ==> /var/log/vdsm/vdsm.log <==
>> >
>> > periodic/42::ERROR::2018-01-11
>> > 16:56:11,446::vmstats::260::virt.vmstats::(send_metrics) VM metrics
>> > collection failed
>> >
>> > Traceback (most recent call last):
>> >
>> > File "/usr/lib/python2.7/site-packages/vdsm/virt/vmstats.py",
line
>> 197, in
>> > send_metrics
>> >
>> > data[prefix + '.cpu.usage'] = stat['cpuUsage']
>> >
>> > KeyError: 'cpuUsage'
>> >
>> >
>> > _______________________________________________
>> > Users mailing list
>> > Users(a)ovirt.org
>> >
http://lists.ovirt.org/mailman/listinfo/users
>> >
>> _______________________________________________
>> Users mailing list
>> Users(a)ovirt.org
>>
http://lists.ovirt.org/mailman/listinfo/users
>>
>
>