Sure not a problem. For the first issue regarding agent and broker crashing. Again the hosted engine VM is up and running at this time, I have no idea why the logs are saying volume doesn't exist and why file /var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/14a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8 does not exist when the file actually does exist in that path.
I assume this problem is most likely also related or causing my other problems when accessing hosted vm snapshot section of web gui as well.vdsm log:jsonrpc/0::ERROR::2018-01-14 09:48:09,302::task::875::storage.TaskManager.Task::(_setError) (Task='37eba553-9c13-4e69-90f7-d0c987cc694c') Unexpected errorTraceback (most recent call last):File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in _runreturn fn(*args, **kargs)File "<string>", line 2, in prepareImageFile "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in methodret = func(*args, **kwargs)File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 3162, in prepareImageraise se.VolumeDoesNotExist(leafUUID)VolumeDoesNotExist: Volume does not exist: (u'8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8',)jsonrpc/0::ERROR::2018-01-14 09:48:09,303::dispatcher::82::storage.Dispatcher::(wrapper) FINISH prepareImage error=Volume does not exist: (u'8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8',)agent log:MainThread::ERROR::2018-01-14 09:49:26,546::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Trying to restart agentMainThread::ERROR::2018-01-14 09:49:37,782::hosted_engine::538::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker) Failed to start necessary monitorsMainThread::ERROR::2018-01-14 09:49:37,783::agent::144::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Traceback (most recent call last):File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 131, in _run_agentreturn action(he)File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 55, in action_properreturn he.start_monitoring()File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 416, in start_monitoringself._initialize_broker()File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 535, in _initialize_brokerm.get('options', {}))File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 83, in start_monitor.format(type, options, e))RequestError: Failed to start monitor ping, options {'addr': '192.168.0.1'}: [Errno 2] No such file or directoryMainThread::ERROR::2018-01-14 09:49:37,783::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Trying to restart agentbroker log:StatusStorageThread::ERROR::2018-01-12 14:03:57,629::status_broker::85::ovirt_hosted_engine_ha.broker.status_broker.StatusBroker.Update::(run) Failed to update state.Traceback (most recent call last):File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/status_broker.py", line 81, in runentry.dataFile "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py", line 212, in put_stats.format(str(e)))RequestError: failed to write metadata: [Errno 2] No such file or directory: '/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/14a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8'StatusStorageThread::ERROR::2018-01-12 14:03:57,629::storage_broker::160::ovirt_hosted_engine_ha.broker.storage_broker.StorageBroker::(get_raw_stats) Failed to read metadata from /var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/14a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8Traceback (most recent call last):File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py", line 151, in get_raw_statsf = os.open(path, direct_flag | os.O_RDONLY | os.O_SYNC)OSError: [Errno 2] No such file or directory: '/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/14a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8'StatusStorageThread::ERROR::2018-01-12 14:03:57,630::status_broker::92::ovirt_hosted_engine_ha.broker.status_broker.StatusBroker.Update::(run) Failed to read state.Traceback (most recent call last):File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/status_broker.py", line 88, in runself._storage_broker.get_raw_stats()File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py", line 162, in get_raw_stats.format(str(e)))RequestError: failed to read metadata: [Errno 2] No such file or directory: '/var/run/vdsm/storage/248f46f0-d793-4581-9810-c9d965e2f286/14a20941-1b84-4b82-be8f-ace38d7c037a/8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8'Syslog:Jan 12 16:52:34 cultivar0 journal: vdsm storage.Dispatcher ERROR FINISH prepareImage error=Volume does not exist: (u'8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8',)Jan 12 16:52:34 cultivar0 python: detected unhandled Python exception in '/usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker'Jan 12 16:52:34 cultivar0 abrt-server: Not saving repeating crash in '/usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker'Jan 12 16:52:34 cultivar0 systemd: ovirt-ha-broker.service: main process exited, code=exited, status=1/FAILUREJan 12 16:52:34 cultivar0 systemd: Unit ovirt-ha-broker.service entered failed state.Jan 12 16:52:34 cultivar0 systemd: ovirt-ha-broker.service failed.Jan 12 16:52:34 cultivar0 systemd: ovirt-ha-broker.service holdoff time over, scheduling restart.Jan 12 16:52:34 cultivar0 systemd: Cannot add dependency job for unit lvm2-lvmetad.socket, ignoring: Unit is masked.Jan 12 16:52:34 cultivar0 systemd: Started oVirt Hosted Engine High Availability Communications Broker.Jan 12 16:52:34 cultivar0 systemd: Starting oVirt Hosted Engine High Availability Communications Broker...Jan 12 16:52:36 cultivar0 journal: vdsm storage.TaskManager.Task ERROR (Task='73141dec-9d8f-4164-9c4e-67c43a102eff') Unexpected error#012Traceback (most recent call last):#012 File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in _run#012 return fn(*args, **kargs)#012 File "<string>", line 2, in prepareImage#012 File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in method#012 ret = func(*args, **kwargs)#012 File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 3162, in prepareImage#012 raise se.VolumeDoesNotExist(leafUUID)#012VolumeDoesNotExist: Volume does not exist: (u'8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8',)Jan 12 16:52:36 cultivar0 journal: vdsm storage.Dispatcher ERROR FINISH prepareImage error=Volume does not exist: (u'8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8',)Jan 12 16:52:36 cultivar0 python: detected unhandled Python exception in '/usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker'Jan 12 16:52:36 cultivar0 abrt-server: Not saving repeating crash in '/usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker'Jan 12 16:52:36 cultivar0 systemd: ovirt-ha-broker.service: main process exited, code=exited, status=1/FAILUREJan 12 16:52:36 cultivar0 systemd: Unit ovirt-ha-broker.service entered failed state.Jan 12 16:52:36 cultivar0 systemd: ovirt-ha-broker.service failed.Jan 12 16:52:36 cultivar0 systemd: ovirt-ha-broker.service holdoff time over, scheduling restart.Jan 12 16:52:36 cultivar0 systemd: Cannot add dependency job for unit lvm2-lvmetad.socket, ignoring: Unit is masked.Jan 12 16:52:36 cultivar0 systemd: Started oVirt Hosted Engine High Availability Communications Broker.Jan 12 16:52:36 cultivar0 systemd: Starting oVirt Hosted Engine High Availability Communications Broker...Jan 12 16:52:37 cultivar0 journal: vdsm storage.TaskManager.Task ERROR (Task='bc7af1e2-0ab2-4164-ae88-d2bee03500f9') Unexpected error#012Traceback (most recent call last):#012 File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in _run#012 return fn(*args, **kargs)#012 File "<string>", line 2, in prepareImage#012 File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in method#012 ret = func(*args, **kwargs)#012 File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 3162, in prepareImage#012 raise se.VolumeDoesNotExist(leafUUID)#012VolumeDoesNotExist: Volume does not exist: (u'8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8',)Jan 12 16:52:37 cultivar0 journal: vdsm storage.Dispatcher ERROR FINISH prepareImage error=Volume does not exist: (u'8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8',)Jan 12 16:52:37 cultivar0 python: detected unhandled Python exception in '/usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker'Jan 12 16:52:38 cultivar0 abrt-server: Not saving repeating crash in '/usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker'Jan 12 16:52:38 cultivar0 systemd: ovirt-ha-broker.service: main process exited, code=exited, status=1/FAILUREJan 12 16:52:38 cultivar0 systemd: Unit ovirt-ha-broker.service entered failed state.Jan 12 16:52:38 cultivar0 systemd: ovirt-ha-broker.service failed.Jan 12 16:52:38 cultivar0 systemd: ovirt-ha-broker.service holdoff time over, scheduling restart.Jan 12 16:52:38 cultivar0 systemd: Cannot add dependency job for unit lvm2-lvmetad.socket, ignoring: Unit is masked.Jan 12 16:52:38 cultivar0 systemd: start request repeated too quickly for ovirt-ha-broker.serviceJan 12 16:52:38 cultivar0 systemd: Failed to start oVirt Hosted Engine High Availability Communications Broker.Jan 12 16:52:38 cultivar0 systemd: Unit ovirt-ha-broker.service entered failed state.Jan 12 16:52:38 cultivar0 systemd: ovirt-ha-broker.service failed.Jan 12 16:52:40 cultivar0 systemd: ovirt-ha-agent.service holdoff time over, scheduling restart.Jan 12 16:52:40 cultivar0 systemd: Cannot add dependency job for unit lvm2-lvmetad.socket, ignoring: Unit is masked.Jan 12 16:52:40 cultivar0 systemd: Started oVirt Hosted Engine High Availability Communications Broker.Jan 12 16:52:40 cultivar0 systemd: Starting oVirt Hosted Engine High Availability Communications Broker...Jan 12 16:52:40 cultivar0 systemd: Started oVirt Hosted Engine High Availability Monitoring Agent.Jan 12 16:52:40 cultivar0 systemd: Starting oVirt Hosted Engine High Availability Monitoring Agent...Jan 12 16:52:41 cultivar0 journal: ovirt-ha-agent ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine ERROR Failed to start necessary monitorsJan 12 16:52:41 cultivar0 journal: ovirt-ha-agent ovirt_hosted_engine_ha.agent.agent.Agent ERROR Traceback (most recent call last):#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 131, in _run_agent#012 return action(he)#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 55, in action_proper#012 return he.start_monitoring()#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 416, in start_monitoring#012 self._initialize_broker()#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 535, in _initialize_broker#012 m.get('options', {}))#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 83, in start_monitor#012 .format(type, options, e))#012RequestError: Failed to start monitor ping, options {'addr': '192.168.0.1'}: [Errno 2] No such file or directoryJan 12 16:52:41 cultivar0 journal: ovirt-ha-agent ovirt_hosted_engine_ha.agent.agent.Agent ERROR Trying to restart agentJan 12 16:52:42 cultivar0 systemd: ovirt-ha-agent.service: main process exited, code=exited, status=157/n/aJan 12 16:52:42 cultivar0 systemd: Unit ovirt-ha-agent.service entered failed state.Jan 12 16:52:42 cultivar0 systemd: ovirt-ha-agent.service failed._______________________________________________On Sun, Jan 14, 2018 at 9:46 AM, Yedidyah Bar David <didi@redhat.com> wrote:On Sun, Jan 14, 2018 at 3:37 PM, Jayme <jaymef@gmail.com> wrote:
> First, apologies for all the posts to this list lately, I've been having a
> heck of a time after 4.2 upgrade and you've been helpful, I appreciate that.
>
> Since 4.2 upgrade I'm experiencing a few problems that I'm trying to debug.
>
> Current status is engine and all hosts are upgraded to 4.2, and cluster and
> domain set to 4.2 compatibility. Hosted Engine VM is running and ui
> accessible etc, all VMs on hosts are running but no HA service. Web UI is
> giving a few errors when checking network and snapshots on the hosted engine
> VM only, it doesn't give errors on any of the others VMs that I spot
> checked.
>
> 1. HA-agent and HA-broker are continually crashing on all three hosts over
> and over every few seconds. I sent an email to users list with more details
> on this problem but unfortunately haven't heard anything back yet. The
> general error in the logs seems to be:
> VolumeDoesNotExist(leafUUID)#012VolumeDoesNotExist: Volume does not exist:
> (u'8582bdfc-ef54-47af-9f1e-f5b7ec1f1cf8',) -- What? Volume doesn't exist,
> why not?
If agent/broker logs do not reveal this, the next step is usually checking
vdsm logs and/or system logs. Can you please check/share these? Thanks.
>
> 2. Error when clicking "network interfaces" in the web gui for the hosted VM
> engine.
>
> 3. Similar to #2 above an error is given when clicking "snapshots" in the
> web gui for the hosted engine VM.
>
> The errors for #2 and #3 are generic "cannot read property 'a' of null".
> I've read previous postings on ovirt-mailing list that suggest you can
> install debug-info package to get a human readable error.. but this package
> does not seem to be compatible with 4.2, it expects 4.1: Requires:
> "ovirt-engine-webadmin-portal = 4.1.2.2-1.el7.centos" -- Perhaps this
> package is no longer required? I do see some additional details in the
> ui.log that I can post if helpful.
>
> There is obviously something odd going on here with the hosted engine VM.
> All three errors appear to related to a problem with it, although it is
> indeed up and running. I'd really like to get HA broker and agent back up
> and running, and fix these GUI errors related to hosted engine VM. All
> three problems may be connected to one common issue?
>
> Thanks in advance!
>
>
>
> _______________________________________________
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
--
Didi
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users