[ovirt-users] This VM is not managed by the engine

Mark Steele msteele at telvue.com
Mon Jul 13 11:25:38 UTC 2015


Thank you Roy!

BTW - I went to the IRC channel - it was just me, a bot and one other
person who did not respond.

I'll keep an eye on the chat in the future.


***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
800.885.8886 x128 | msteele at telvue.com | http://www.telvue.com
twitter: http://twitter.com/telvue | facebook:
https://www.facebook.com/telvue

On Sun, Jul 12, 2015 at 10:46 PM, Roy Golan <rgolan at redhat.com> wrote:

>  On 07/13/2015 01:36 AM, Mark Steele wrote:
>
>  Roy - Success!
>
>  I was able to restart vdsmd
>
>  'service vdsmd restart'
>
>  From there - I was able to issue the stop destroy command:
>
>  'vdsClient -s 0 destroy 41703d5c-6cdb-42b4-93df-d78be2776e2b'
>
>  Once that was done, I was able to REMOVE the VM from the GUI.
>
>  Thank you for all your help AND your patience with a noob!
>
>
> You're welcome. And btw,next time try also the #ovirt irc channel at
> oftc.net,  you might get quicker response.
>
>
>  ***
> *Mark Steele*
>  CIO / VP Technical Operations | TelVue Corporation
> TelVue - We Share Your Vision
> 800.885.8886 x128 | msteele at telvue.com |  <http://www.telvue.com/>
> http://www.telvue.com
> twitter:  <http://twitter.com/telvue>http://twitter.com/telvue |
> facebook:  <https://www.facebook.com/telvue>
> https://www.facebook.com/telvue
>
> On Sun, Jul 12, 2015 at 3:16 PM, Mark Steele <msteele at telvue.com> wrote:
>
>>  No joy - the shutdown appears to start - but eventually the VM shows as
>> running again. There is NO process running that shows up with ps command.
>>
>>  I'm not sure what to do next
>>
>>
>>
>>  ***
>> *Mark Steele*
>>  CIO / VP Technical Operations | TelVue Corporation
>> TelVue - We Share Your Vision
>>  800.885.8886 x128 | msteele at telvue.com |  <http://www.telvue.com/>
>> http://www.telvue.com
>> twitter:  <http://twitter.com/telvue>http://twitter.com/telvue |
>> facebook:  <https://www.facebook.com/telvue>
>> https://www.facebook.com/telvue
>>
>> On Sun, Jul 12, 2015 at 10:13 AM, Artyom Lukianov < <alukiano at redhat.com>
>> alukiano at redhat.com> wrote:
>>
>>> Also provide to us vdsm log from host(/var/log/vdsm/vdsm.log) for future
>>> investigation.
>>> Thanks
>>>
>>> ----- Original Message -----
>>> From: "Mark Steele" < <msteele at telvue.com>msteele at telvue.com>
>>> To: "Roy Golan" < <rgolan at redhat.com>rgolan at redhat.com>
>>> Cc: "Artyom Lukianov" < <alukiano at redhat.com>alukiano at redhat.com>,
>>> <users at ovirt.org>users at ovirt.org
>>> Sent: Sunday, July 12, 2015 4:21:34 PM
>>> Subject: Re: [ovirt-users] This VM is not managed by the engine
>>>
>>> [root at hv-02 etc]# vdsClient -s 0 destroy
>>> 41703d5c-6cdb-42b4-93df-d78be2776e2b
>>>
>>> Unexpected exception
>>>
>>> Not sure I'm getting any closer :-)
>>>
>>>
>>>
>>>
>>> ***
>>> *Mark Steele*
>>> CIO / VP Technical Operations | TelVue Corporation
>>> TelVue - We Share Your Vision
>>> 800.885.8886 x128 <800.885.8886%20x128> | msteele at telvue.com |
>>> <http://www.telvue.com>http://www.telvue.com
>>> twitter: http://twitter.com/telvue | facebook:
>>> https://www.facebook.com/telvue
>>>
>>>  On Sun, Jul 12, 2015 at 9:13 AM, Mark Steele <msteele at telvue.com>
>>> wrote:
>>>
>>> > OK - I think I'm getting closer - now that I'm on the correct box.
>>> >
>>> > here is the output of the vdsClient command - which is the device id -
>>> the
>>> > first line?
>>> >
>>> > [root at hv-02 etc]# vdsClient -s 0 list
>>> >
>>> > 41703d5c-6cdb-42b4-93df-d78be2776e2b
>>> > Status = Up
>>> > acpiEnable = true
>>> > emulatedMachine = rhel6.5.0
>>> > afterMigrationStatus =
>>> > pid = 27304
>>> > memGuaranteedSize = 2048
>>> > transparentHugePages = true
>>> > displaySecurePort = 5902
>>> > spiceSslCipherSuite = DEFAULT
>>> > cpuType = SandyBridge
>>> > smp = 2
>>> > numaTune = {'nodeset': '0,1', 'mode': 'interleave'}
>>> > custom =
>>> >
>>> {'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72':
>>> > 'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>>> > deviceId=ebd4c73d-12c4-435e-8cc5-f180d8f20a72, device=unix,
>>> type=CHANNEL,
>>> > bootOrder=0, specParams={}, address={bus=0, controller=0,
>>> > type=virtio-serial, port=2}, managed=false, plugged=true,
>>> readOnly=false,
>>> > deviceAlias=channel1, customProperties={}, snapshotId=null}',
>>> >
>>> 'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72device_ffd2796f-7644-4008-b920-5f0970b0ef0e':
>>> > 'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>>> > deviceId=ffd2796f-7644-4008-b920-5f0970b0ef0e, device=unix,
>>> type=CHANNEL,
>>> > bootOrder=0, specParams={}, address={bus=0, controller=0,
>>> > type=virtio-serial, port=1}, managed=false, plugged=true,
>>> readOnly=false,
>>> > deviceAlias=channel0, customProperties={}, snapshotId=null}',
>>> > 'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6': 'VmDevice
>>> > {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>>> > deviceId=86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6, device=ide,
>>> type=CONTROLLER,
>>> > bootOrder=0, specParams={}, address={slot=0x01, bus=0x00,
>>> domain=0x0000,
>>> > type=pci, function=0x1}, managed=false, plugged=true, readOnly=false,
>>> > deviceAlias=ide0, customProperties={}, snapshotId=null}',
>>> >
>>> 'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72device_ffd2796f-7644-4008-b920-5f0970b0ef0edevice_6693d023-9c1f-433c-870e-e9771be8474b':
>>> > 'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>>> > deviceId=6693d023-9c1f-433c-870e-e9771be8474b, device=spicevmc,
>>> > type=CHANNEL, bootOrder=0, specParams={}, address={bus=0, controller=0,
>>> > type=virtio-serial, port=3}, managed=false, plugged=true,
>>> readOnly=false,
>>> > deviceAlias=channel2, customProperties={}, snapshotId=null}'}
>>> > vmType = kvm
>>> > memSize = 2048
>>> > smpCoresPerSocket = 1
>>> > vmName = connect-turbo-stage-03
>>> > nice = 0
>>> > bootMenuEnable = false
>>> > copyPasteEnable = true
>>> > displayIp = 10.1.90.161
>>> > displayPort = -1
>>> > smartcardEnable = false
>>> > clientIp =
>>> > fileTransferEnable = true
>>> > nicModel = rtl8139,pv
>>> > keyboardLayout = en-us
>>> > kvmEnable = true
>>> > pitReinjection = false
>>> > displayNetwork = ovirtmgmt
>>> > devices = [{'target': 2097152, 'specParams': {'model': 'none'},
>>> 'alias':
>>> > 'balloon0', 'deviceType': 'balloon', 'device': 'memballoon', 'type':
>>> > 'balloon'}, {'device': 'unix', 'alias': 'channel0', 'address': {'bus':
>>> '0',
>>> > 'controller': '0', 'type': 'virtio-serial', 'port': '1'}, 'deviceType':
>>> > 'channel', 'type': 'channel'}, {'device': 'unix', 'alias': 'channel1',
>>> > 'address': {'bus': '0', 'controller': '0', 'type': 'virtio-serial',
>>> 'port':
>>> > '2'}, 'deviceType': 'channel', 'type': 'channel'}, {'device':
>>> 'spicevmc',
>>> > 'alias': 'channel2', 'address': {'bus': '0', 'controller': '0', 'type':
>>> > 'virtio-serial', 'port': '3'}, 'deviceType': 'channel', 'type':
>>> 'channel'},
>>> > {'index': '0', 'alias': 'scsi0', 'specParams': {}, 'deviceType':
>>> > 'controller', 'deviceId': '88db8cb9-0960-4797-bd41-1694bf14b8a9',
>>> > 'address': {'slot': '0x04', 'bus': '0x00', 'domain': '0x0000', 'type':
>>> > 'pci', 'function': '0x0'}, 'device': 'scsi', 'model': 'virtio-scsi',
>>> > 'type': 'controller'}, {'alias': 'virtio-serial0', 'specParams': {},
>>> > 'deviceType': 'controller', 'deviceId':
>>> > '4bb9c112-e027-4e7d-8b1c-32f99c7040ee', 'address': {'slot': '0x05',
>>> 'bus':
>>> > '0x00', 'domain': '0x0000', 'type': 'pci', 'function': '0x0'},
>>> 'device':
>>> > 'virtio-serial', 'type': 'controller'}, {'device': 'usb', 'alias':
>>> 'usb0',
>>> > 'address': {'slot': '0x01', 'bus': '0x00', 'domain': '0x0000', 'type':
>>> > 'pci', 'function': '0x2'}, 'deviceType': 'controller', 'type':
>>> > 'controller'}, {'device': 'ide', 'alias': 'ide0', 'address': {'slot':
>>> > '0x01', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci', 'function':
>>> > '0x1'}, 'deviceType': 'controller', 'type': 'controller'}, {'alias':
>>> > 'video0', 'specParams': {'vram': '32768', 'ram': '65536', 'heads':
>>> '1'},
>>> > 'deviceType': 'video', 'deviceId':
>>> '23634541-3b7e-460d-9580-39504392ba36',
>>> > 'address': {'slot': '0x02', 'bus': '0x00', 'domain': '0x0000', 'type':
>>> > 'pci', 'function': '0x0'}, 'device': 'qxl', 'type': 'video'},
>>> {'device':
>>> > 'spice', 'specParams': {'displayNetwork': 'ovirtmgmt',
>>> > 'spiceSecureChannels':
>>> >
>>> 'smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard',
>>> > 'keyMap': 'en-us', 'displayIp': '10.1.90.161', 'copyPasteEnable':
>>> 'true'},
>>> > 'deviceType': 'graphics', 'tlsPort': '5902', 'type': 'graphics'},
>>> > {'nicModel': 'pv', 'macAddr': '00:01:a4:a2:b4:30', 'linkActive': True,
>>> > 'network': 'ovirtmgmt', 'alias': 'net0', 'filter':
>>> 'vdsm-no-mac-spoofing',
>>> > 'specParams': {'inbound': {}, 'outbound': {}}, 'deviceType':
>>> 'interface',
>>> > 'deviceId': '63651662-2ddf-4611-b988-1a58d05982f6', 'address': {'slot':
>>> > '0x03', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci', 'function':
>>> > '0x0'}, 'device': 'bridge', 'type': 'interface', 'name': 'vnet5'},
>>> > {'nicModel': 'pv', 'macAddr': '00:01:a4:a2:b4:31', 'linkActive': True,
>>> > 'network': 'storage', 'alias': 'net1', 'filter':
>>> 'vdsm-no-mac-spoofing',
>>> > 'specParams': {'inbound': {}, 'outbound': {}}, 'deviceType':
>>> 'interface',
>>> > 'deviceId': 'b112d9c6-5144-4b67-912b-dcc27aabdae9', 'address': {'slot':
>>> > '0x07', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci', 'function':
>>> > '0x0'}, 'device': 'bridge', 'type': 'interface', 'name': 'vnet6'},
>>> > {'index': '3', 'iface': 'ide', 'name': 'hdd', 'alias': 'ide0-1-1',
>>> > 'specParams': {'vmPayload': {'volId': 'config-2', 'file':
>>> > {'openstack/latest/meta_data.json':
>>> >
>>> 'ewogICJsYXVuY2hfaW5kZXgiIDogIjAiLAogICJhdmFpbGFiaWxpdHlfem9uZSIgOiAibm92YSIs\nCiAgIm5hbWUiIDogImNvbm5lY3QtdHVyYm8tc3RhZ2UtMDMiLAogICJob3N0bmFtZSIgOiAiY29u\nbmVjdC10dXJiby1zdGFnZS0wMyIsCiAgInV1aWQiIDogImJiNmIwMzdhLTZkY2ItNGZmZS04MjUw\nLTMwYjlkOWE0ZTlmZCIsCiAgIm1ldGEiIDogewogICAgImVzc2VudGlhbCIgOiAiZmFsc2UiLAog\nICAgInJvbGUiIDogInNlcnZlciIsCiAgICAiZHNtb2RlIiA6ICJsb2NhbCIKICB9Cn0=\n',
>>> > 'openstack/latest/user_data':
>>> >
>>> 'I2Nsb3VkLWNvbmZpZwpzc2hfcHdhdXRoOiB0cnVlCmRpc2FibGVfcm9vdDogMApvdXRwdXQ6CiAg\nYWxsOiAnPj4gL3Zhci9sb2cvY2xvdWQtaW5pdC1vdXRwdXQubG9nJwpjaHBhc3N3ZDoKICBleHBp\ncmU6IGZhbHNlCnJ1bmNtZDoKLSAnc2VkIC1pICcnL15kYXRhc291cmNlX2xpc3Q6IC9kJycgL2V0\nYy9jbG91ZC9jbG91ZC5jZmc7IGVjaG8gJydkYXRhc291cmNlX2xpc3Q6CiAgWyJOb0Nsb3VkIiwg\nIkNvbmZpZ0RyaXZlIl0nJyA+PiAvZXRjL2Nsb3VkL2Nsb3VkLmNmZycK\n'}}},
>>> > 'readonly': 'True', 'deviceType': 'disk', 'deviceId':
>>> > '6de890b2-6454-4377-9a71-bea2e46d50a8', 'address': {'bus': '1',
>>> > 'controller': '0', 'type': 'drive', 'target': '0', 'unit': '1'},
>>> 'device':
>>> > 'cdrom', 'shared': 'false', 'path': '', 'type': 'disk'}, {'index': '2',
>>> > 'iface': 'ide', 'name': 'hdd', 'alias': 'ide0-1-1', 'specParams':
>>> {'path':
>>> > ''}, 'readonly': 'True', 'deviceType': 'disk', 'deviceId':
>>> > '2763a41b-6576-4135-b349-4fe402c31246', 'address': {'bus': '1',
>>> > 'controller': '0', 'type': 'drive', 'target': '0', 'unit': '1'},
>>> 'device':
>>> > 'cdrom', 'shared': 'false', 'path': '', 'type': 'disk'}, {'device':
>>> 'file',
>>> > 'alias': 'ide0-1-0', 'address': {'bus': '1', 'controller': '0', 'type':
>>> > 'drive', 'target': '0', 'unit': '0'}, 'deviceType': 'disk', 'type':
>>> 'disk'}]
>>> > timeOffset = -891891
>>> > maxVCpus = 16
>>> > spiceSecureChannels =
>>> > smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard
>>> > display = qxl
>>> > [root at hv-02 etc]#
>>> >
>>> >
>>> > ***
>>>  > *Mark Steele*
>>> > CIO / VP Technical Operations | TelVue Corporation
>>> > TelVue - We Share Your Vision
>>> > 800.885.8886 x128 <800.885.8886%20x128> | msteele at telvue.com |
>>> <http://www.telvue.com>http://www.telvue.com
>>> > twitter: http://twitter.com/telvue | facebook:
>>> > https://www.facebook.com/telvue
>>> >
>>> > On Sun, Jul 12, 2015 at 9:06 AM, Mark Steele <msteele at telvue.com>
>>> wrote:
>>> >
>>> >> I think I may have not given you all the information.
>>> >>
>>> >> I am not logging into the host - I am logging into the ovirt
>>> management.
>>> >>
>>> >> Let me try logging into the host and checking
>>> >>
>>> >>
>>> >>
>>> >> ***
>>> >> *Mark Steele*
>>> >> CIO / VP Technical Operations | TelVue Corporation
>>> >> TelVue - We Share Your Vision
>>> >> 800.885.8886 x128 <800.885.8886%20x128> | msteele at telvue.com |
>>> <http://www.telvue.com>http://www.telvue.com
>>> >> twitter: http://twitter.com/telvue | facebook:
>>> >> https://www.facebook.com/telvue
>>> >>
>>> >> On Sun, Jul 12, 2015 at 9:03 AM, Roy Golan <rgolan at redhat.com> wrote:
>>> >>
>>> >>>  On 07/12/2015 03:52 PM, Mark Steele wrote:
>>> >>>
>>> >>>  That command returns nothing - I don't think qemu is running?
>>> >>>
>>> >>>  Not sure how to start it on CentOS
>>> >>>
>>> >>>  [root at ovirt-01 ~]# ps -ef | grep qemu
>>> >>>
>>> >>> root     23279 23130  0 08:51 pts/0    00:00:00 grep qemu
>>> >>>
>>> >>>
>>> >>>   that mean you don't have vm running on that host. so you can
>>> restart
>>> >>> vdsm
>>> >>>
>>> >>>
>>> >>>
>>> >>>  ***
>>> >>> *Mark Steele*
>>> >>>  CIO / VP Technical Operations | TelVue Corporation
>>> >>> TelVue - We Share Your Vision
>>> >>> 800.885.8886 x128 <800.885.8886%20x128> | msteele at telvue.com |  <
>>> http://www.telvue.com/>
>>> >>> http://www.telvue.com
>>> >>> twitter:  < <http://twitter.com/telvue>http://twitter.com/telvue>
>>> <http://twitter.com/telvue>http://twitter.com/telvue |
>>> >>> facebook:  < <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue>
>>> >>> https://www.facebook.com/telvue
>>> >>>
>>> >>> On Sun, Jul 12, 2015 at 8:45 AM, Roy Golan <rgolan at redhat.com>
>>> wrote:
>>> >>>
>>> >>>>  On 07/12/2015 03:42 PM, Mark Steele wrote:
>>> >>>>
>>> >>>>  I run into the same issue - I am unable to completely go into
>>> >>>> maintenance mode because this VM is still on it - it cannot be
>>> migrated
>>> >>>> because it is not managed.
>>> >>>>
>>> >>>>  find you qemu process:
>>> >>>>   pgrep -an qemu-kvm | grep external
>>> >>>>
>>> >>>> and kill the process
>>> >>>>
>>> >>>>
>>> >>>>
>>> >>>>  [image: Inline image 1]
>>> >>>>
>>> >>>>
>>> >>>>  ***
>>> >>>> *Mark Steele*
>>> >>>>  CIO / VP Technical Operations | TelVue Corporation
>>> >>>> TelVue - We Share Your Vision
>>> >>>> 800.885.8886 x128 <800.885.8886%20x128> |  <msteele at telvue.com>
>>> <msteele at telvue.com>msteele at telvue.com |
>>> >>>> <http://www.telvue.com> <http://www.telvue.com>
>>> http://www.telvue.com
>>> >>>> twitter:  < <http://twitter.com/telvue>http://twitter.com/telvue>
>>> <http://twitter.com/telvue>http://twitter.com/telvue |
>>> >>>> facebook:  < <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue>
>>>  >>>> https://www.facebook.com/telvue
>>> >>>>
>>> >>>> On Sun, Jul 12, 2015 at 8:25 AM, Roy Golan < < <rgolan at redhat.com>
>>> rgolan at redhat.com>
>>> >>>> rgolan at redhat.com> wrote:
>>> >>>>
>>> >>>>>  On 07/12/2015 03:12 PM, Mark Steele wrote:
>>> >>>>>
>>> >>>>>  I think I may have found the problem:
>>> >>>>>
>>> >>>>>  [root at ovirt-01 pki]# ls -lah
>>> >>>>> total 48K
>>> >>>>> drwxr-xr-x.  10 root  root  4.0K Nov 14  2014 .
>>> >>>>> drwxr-xr-x. 118 root  root   12K Jul 12 03:35 ..
>>> >>>>> drwxr-xr-x.   6 root  root  4.0K Nov 14  2014 CA
>>> >>>>> drwxr-xr-x.   4 root  root  4.0K Nov 14  2014 ca-trust
>>> >>>>> drwxr-xr-x.   2 root  root  4.0K Nov 14  2014 java
>>> >>>>> drwxr-xr-x.   2 root  root  4.0K Jul 12 07:03 nssdb
>>> >>>>> drwxr-xr-x.   6 ovirt ovirt 4.0K Nov 19  2014 ovirt-engine
>>> >>>>> drwxr-xr-x.   2 root  root  4.0K Nov 14  2014 rpm-gpg
>>> >>>>> drwx------.   2 root  root  4.0K Nov 22  2013 rsyslog
>>> >>>>> drwxr-xr-x.   5 root  root  4.0K Nov 14  2014 tls
>>> >>>>> [root at ovirt-01 pki]#
>>> >>>>>
>>> >>>>>  There is no vsdm directory under /etc/pki
>>> >>>>>
>>> >>>>>  This is an ovirt node. Version of software is 3.5.0.1-1.el6 from
>>> the
>>> >>>>> ovirt management console.
>>> >>>>>
>>> >>>>>  I'd like to add that I am not the person who originally installed
>>> >>>>> this instance - and am not entirely familiar with how it is setup
>>> and
>>> >>>>> installed - so I may ask ignorant questions from time to time.
>>> >>>>>
>>> >>>>>
>>> >>>>>  not urgent but at this point it looks like it would be good to
>>> >>>>> reinstall this host from the webadmin. if you have the capacity,
>>> >>>>> you can put the host to maintenance, that will migrate vms to other
>>> >>>>> hosts, and then choose "reinstall" once its in "maintenance'
>>> >>>>>
>>> >>>>>
>>> >>>>>  ***
>>>  >>>>> *Mark Steele*
>>> >>>>>  CIO / VP Technical Operations | TelVue Corporation
>>> >>>>> TelVue - We Share Your Vision
>>> >>>>> 800.885.8886 x128 <800.885.8886%20x128> |  <msteele at telvue.com>
>>> <msteele at telvue.com>msteele at telvue.com |
>>> >>>>> <http://www.telvue.com> <http://www.telvue.com>
>>> http://www.telvue.com
>>>  >>>>> twitter: <http://twitter.com/telvue>http://twitter.com/telvue |
>>> facebook:
>>> >>>>> <https://www.facebook.com/telvue>https://www.facebook.com/telvue
>>> >>>>>
>>> >>>>> On Sun, Jul 12, 2015 at 8:02 AM, Roy Golan < < <rgolan at redhat.com>
>>> rgolan at redhat.com>
>>> >>>>> <rgolan at redhat.com>rgolan at redhat.com> wrote:
>>> >>>>>
>>> >>>>>>  On 07/12/2015 02:07 PM, Mark Steele wrote:
>>> >>>>>>
>>> >>>>>>  Thank you Roy,
>>> >>>>>>
>>> >>>>>>  I installed the client but am getting a permissions error when I
>>> >>>>>> run it
>>> >>>>>>
>>> >>>>>>  [root at ovirt-01 ~]# vdsClient -s 0 list
>>> >>>>>> Traceback (most recent call last):
>>> >>>>>>   File "/usr/share/vdsm/vdsClient.py", line 2678, in <module>
>>> >>>>>>     serv.do_connect(hostPort)
>>> >>>>>>   File "/usr/share/vdsm/vdsClient.py", line 136, in do_connect
>>> >>>>>>     self.s = vdscli.connect(hostPort, self.useSSL,
>>> self.truststore)
>>> >>>>>>   File "/usr/lib/python2.6/site-packages/vdsm/vdscli.py", line
>>> 110,
>>> >>>>>> in connect
>>> >>>>>>     raise Exception("No permission to read file: %s" % f)
>>> >>>>>> Exception: No permission to read file:
>>> /etc/pki/vdsm/keys/vdsmkey.pem
>>> >>>>>>
>>> >>>>>>
>>> >>>>>>  This should work. something isn't right with your setup
>>> >>>>>> is your host an ovirt-node? could be that you hit [1] . let me
>>> know
>>> >>>>>> what version are you running.
>>> >>>>>>
>>> >>>>>> please try the same with user vdsm. it should have permissions to
>>> >>>>>> /etc/pki/vdsm
>>> >>>>>>
>>> >>>>>> [1]  < <https://gerrit.ovirt.org/#/c/27779/>
>>> https://gerrit.ovirt.org/#/c/27779/>
>>> >>>>>> <https://gerrit.ovirt.org/#/c/27779/>
>>> https://gerrit.ovirt.org/#/c/27779/
>>> >>>>>>
>>> >>>>>>
>>> >>>>>>   If I restart vdsm, will that cause any issues with running VM's
>>> on
>>> >>>>>> this ovirt installation? This is our production environment.
>>> >>>>>>
>>> >>>>>>
>>> >>>>>>  Generatlly the answer is no but lets avoid it if  we can for
>>> this is
>>> >>>>>> a minor cosmetic issue I guess.
>>> >>>>>>
>>> >>>>>> just as FYI - vdsm only reconnects to the socket exposed by
>>> libvirt
>>> >>>>>> to control the vm lifecycle. VDSM doesn't manadate the lifecycle
>>> of a VM
>>> >>>>>> unless the engine tells is so. Storage wise there could be some
>>> operations
>>> >>>>>> but i'm almost sure they must not have effect on running VMs.
>>> >>>>>>
>>> >>>>>>
>>> >>>>>>  Thank you
>>> >>>>>>
>>> >>>>>>
>>> >>>>>>  ***
>>>  >>>>>> *Mark Steele*
>>> >>>>>>  CIO / VP Technical Operations | TelVue Corporation
>>> >>>>>> TelVue - We Share Your Vision
>>> >>>>>> 800.885.8886 x128 <800.885.8886%20x128> |  <msteele at telvue.com>
>>> <msteele at telvue.com>msteele at telvue.com |
>>> >>>>>> < <http://www.telvue.com>http://www.telvue.com>
>>> <http://www.telvue.com>http://www.telvue.com
>>> >>>>>> twitter: <http://twitter.com/telvue>http://twitter.com/telvue |
>>> facebook:
>>> >>>>>> < <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue> <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue
>>> >>>>>>
>>> >>>>>> On Sun, Jul 12, 2015 at 4:09 AM, Roy Golan < <
>>> <rgolan at redhat.com>rgolan at redhat.com>
>>> >>>>>> <rgolan at redhat.com>rgolan at redhat.com> wrote:
>>> >>>>>>
>>> >>>>>>>  On 07/09/2015 06:34 PM, Mark Steele wrote:
>>> >>>>>>>
>>> >>>>>>>  Yes,
>>> >>>>>>>
>>> >>>>>>>  It is displayed in the engine:
>>> >>>>>>>
>>> >>>>>>>  [image: Inline image 1]
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>>  the vdsm on that host reports it back to the engine . since
>>> this vm
>>> >>>>>>> isn't in the engine DB it is concidered as EXTERNAL (thus the
>>> error 400
>>> >>>>>>> from the API)
>>> >>>>>>>
>>> >>>>>>> do yo know if the qemu-kvm proccess isn't running anymore?
>>> >>>>>>>
>>> >>>>>>> if the process isn't running then vdsm must clean its cache
>>> >>>>>>>
>>> >>>>>>> try to:
>>> >>>>>>>
>>> >>>>>>> yum install vdsm-cli
>>> >>>>>>> vdsClient -s 0 list
>>> >>>>>>> vdsClient -s 0 destroy {vmId}
>>> >>>>>>>
>>> >>>>>>> alternativly a vdsm restart will work (if the qemu proccess isn't
>>> >>>>>>> running)
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>>  The VM is not really running - the IP addresses that are being
>>> >>>>>>> reported are from another VM that was recently removed. All
>>> attempts to
>>> >>>>>>> control the VM have failed. It does not have any NICS or disk
>>> associated
>>> >>>>>>> with it - so this seems to be a ghost in the machine. I
>>> attempted to unlock
>>> >>>>>>> it using the unlock_entity.sh script - it reports successful,
>>> however I
>>> >>>>>>> still cannot do anything with the VM.
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>>  ***
>>> >>>>>>> *Mark Steele*
>>> >>>>>>>  CIO / VP Technical Operations | TelVue Corporation
>>> >>>>>>> TelVue - We Share Your Vision
>>> >>>>>>> 800.885.8886 x128 <800.885.8886%20x128> |  <msteele at telvue.com>
>>> <msteele at telvue.com>msteele at telvue.com |
>>> >>>>>>> < <http://www.telvue.com>http://www.telvue.com>
>>> <http://www.telvue.com>http://www.telvue.com
>>> >>>>>>> twitter: <http://twitter.com/telvue>http://twitter.com/telvue |
>>> facebook:
>>> >>>>>>> < <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue> <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue
>>> >>>>>>>
>>> >>>>>>> On Thu, Jul 9, 2015 at 11:11 AM, Artyom Lukianov <
>>> >>>>>>> < <alukiano at redhat.com>alukiano at redhat.com>
>>> <alukiano at redhat.com>alukiano at redhat.com> wrote:
>>> >>>>>>>
>>> >>>>>>>> Can you sea via engine, on what host run VM?
>>> >>>>>>>> Anyway if you have really run VM on host you can try to figure
>>> it
>>> >>>>>>>> with 'ps aux | grep qemu', if it will return you some process,
>>> you can just
>>> >>>>>>>> kill process via 'kill pid'.
>>> >>>>>>>> I hope it will help you.
>>> >>>>>>>>
>>> >>>>>>>> ----- Original Message -----
>>> >>>>>>>> From: "Mark Steele" < < <msteele at telvue.com>msteele at telvue.com>
>>> <msteele at telvue.com>msteele at telvue.com>
>>> >>>>>>>> To: "Artyom Lukianov" < <alukiano at redhat.com>
>>> <alukiano at redhat.com>alukiano at redhat.com>
>>> >>>>>>>> Cc: < <users at ovirt.org>users at ovirt.org> <users at ovirt.org>
>>> users at ovirt.org
>>> >>>>>>>> Sent: Thursday, July 9, 2015 5:42:20 PM
>>> >>>>>>>> Subject: Re: [ovirt-users] This VM is not managed by the engine
>>> >>>>>>>>
>>> >>>>>>>> Artyom,
>>> >>>>>>>>
>>> >>>>>>>> Thank you - I don't have vdsClient installed - can you point me
>>> to
>>> >>>>>>>> the
>>> >>>>>>>> download?
>>> >>>>>>>>
>>> >>>>>>>>
>>> >>>>>>>> ***
>>> >>>>>>>> *Mark Steele*
>>> >>>>>>>>  CIO / VP Technical Operations | TelVue Corporation
>>> >>>>>>>> TelVue - We Share Your Vision
>>> >>>>>>>> 800.885.8886 x128 <800.885.8886%20x128> <800.885.8886%20x128>
>>> | < <msteele at telvue.com>msteele at telvue.com>
>>> >>>>>>>> <msteele at telvue.com>msteele at telvue.com | <http://www.telvue.com
>>> > <http://www.telvue.com>http://www.telvue.com
>>> >>>>>>>> twitter: < <http://twitter.com/telvue>http://twitter.com/telvue
>>> > <http://twitter.com/telvue>http://twitter.com/telvue |
>>> >>>>>>>> facebook:
>>> >>>>>>>>  < <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue> <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue
>>> >>>>>>>>
>>> >>>>>>>> On Thu, Jul 9, 2015 at 10:11 AM, Artyom Lukianov <
>>> >>>>>>>> < <alukiano at redhat.com>alukiano at redhat.com>
>>> <alukiano at redhat.com>alukiano at redhat.com>
>>> >>>>>>>> wrote:
>>> >>>>>>>>
>>> >>>>>>>> > Please check host where VM run(vdsClient -s 0 list table), and
>>> >>>>>>>> you can
>>> >>>>>>>> > destroy it via vdsClient(vdsClient -s 0 destroy vm_id).
>>> >>>>>>>> > Thanks
>>> >>>>>>>> >
>>> >>>>>>>> > ----- Original Message -----
>>> >>>>>>>> > From: "Mark Steele" < < <msteele at telvue.com>
>>> msteele at telvue.com> <msteele at telvue.com>msteele at telvue.com>
>>> >>>>>>>> > To: < <users at ovirt.org>users at ovirt.org> <users at ovirt.org>
>>> users at ovirt.org
>>> >>>>>>>> > Sent: Thursday, July 9, 2015 4:38:32 PM
>>> >>>>>>>> > Subject: [ovirt-users] This VM is not managed by the engine
>>> >>>>>>>> >
>>> >>>>>>>> > I have a VM that was not started and is now showing as
>>> running.
>>> >>>>>>>> When I
>>> >>>>>>>> > attempt to suspend or stop it in the ovirt-shell, I get the
>>> >>>>>>>> message:
>>> >>>>>>>> >
>>> >>>>>>>> > status: 400
>>> >>>>>>>> > reason: bad request
>>> >>>>>>>> > detail: Cannot hibernate VM. This VM is not managed by the
>>> engine.
>>> >>>>>>>> >
>>> >>>>>>>> > Not sure how the VM was initially created on the ovirt
>>> manager.
>>> >>>>>>>> This VM is
>>> >>>>>>>> > not needed - how can I 'shutdown' and remove this VM?
>>> >>>>>>>> >
>>> >>>>>>>> > Thanks
>>> >>>>>>>> >
>>> >>>>>>>> > ***
>>> >>>>>>>> > Mark Steele
>>> >>>>>>>> > CIO / VP Technical Operations | TelVue Corporation
>>> >>>>>>>> > TelVue - We Share Your Vision
>>> >>>>>>>> > 800.885.8886 x128 <800.885.8886%20x128>
>>> <800.885.8886%20x128> | < <msteele at telvue.com>msteele at telvue.com>
>>> >>>>>>>> <msteele at telvue.com>msteele at telvue.com | <http://www.telvue.com
>>> > <http://www.telvue.com>http://www.telvue.com
>>> >>>>>>>> > twitter: <http://twitter.com/telvue>
>>> <http://twitter.com/telvue>http://twitter.com/telvue |
>>> >>>>>>>> facebook:
>>> >>>>>>>> > < <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue> <https://www.facebook.com/telvue>
>>> https://www.facebook.com/telvue
>>> >>>>>>>> >
>>> >>>>>>>> > _______________________________________________
>>> >>>>>>>> > Users mailing list
>>> >>>>>>>> > < <Users at ovirt.org>Users at ovirt.org> <Users at ovirt.org>
>>> Users at ovirt.org
>>> >>>>>>>> > < <http://lists.ovirt.org/mailman/listinfo/users>
>>> http://lists.ovirt.org/mailman/listinfo/users>
>>>  >>>>>>>> <http://lists.ovirt.org/mailman/listinfo/users>
>>> http://lists.ovirt.org/mailman/listinfo/users
>>> >>>>>>>> >
>>> >>>>>>>>
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>> _______________________________________________
>>> >>>>>>> Users mailing listUsers at ovirt.orghttp://
>>> lists.ovirt.org/mailman/listinfo/users
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>>
>>> >>>>>>
>>> >>>>>>
>>> >>>>>
>>> >>>>>
>>> >>>>
>>> >>>>
>>> >>>
>>> >>>
>>> >>
>>> >
>>>
>>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150713/1898ec19/attachment-0001.html>


More information about the Users mailing list