[ovirt-users] This VM is not managed by the engine

Roy Golan rgolan at redhat.com
Mon Jul 13 02:46:44 UTC 2015


On 07/13/2015 01:36 AM, Mark Steele wrote:
> Roy - Success!
>
> I was able to restart vdsmd
>
> 'service vdsmd restart'
>
> From there - I was able to issue the stop destroy command:
>
> 'vdsClient -s 0 destroy 41703d5c-6cdb-42b4-93df-d78be2776e2b'
>
> Once that was done, I was able to REMOVE the VM from the GUI.
>
> Thank you for all your help AND your patience with a noob!
>

You're welcome. And btw,next time try also the #ovirt irc channel at 
oftc.net,  you might get quicker response.
>
> ***
> *Mark Steele*
> CIO / VP Technical Operations | TelVue Corporation
> TelVue - We Share Your Vision
> 800.885.8886 x128 | msteele at telvue.com <mailto:msteele at telvue.com> | 
> http://www.telvue.com <http://www.telvue.com/>
> twitter: http://twitter.com/telvue | facebook: 
> https://www.facebook.com/telvue
>
> On Sun, Jul 12, 2015 at 3:16 PM, Mark Steele <msteele at telvue.com 
> <mailto:msteele at telvue.com>> wrote:
>
>     No joy - the shutdown appears to start - but eventually the VM
>     shows as running again. There is NO process running that shows up
>     with ps command.
>
>     I'm not sure what to do next
>
>
>
>     ***
>     *Mark Steele*
>     CIO / VP Technical Operations | TelVue Corporation
>     TelVue - We Share Your Vision
>     800.885.8886 x128 <tel:800.885.8886%20x128> | msteele at telvue.com
>     <mailto:msteele at telvue.com> | http://www.telvue.com
>     <http://www.telvue.com/>
>     twitter: http://twitter.com/telvue | facebook:
>     https://www.facebook.com/telvue
>
>     On Sun, Jul 12, 2015 at 10:13 AM, Artyom Lukianov
>     <alukiano at redhat.com <mailto:alukiano at redhat.com>> wrote:
>
>         Also provide to us vdsm log from host(/var/log/vdsm/vdsm.log)
>         for future investigation.
>         Thanks
>
>         ----- Original Message -----
>         From: "Mark Steele" <msteele at telvue.com
>         <mailto:msteele at telvue.com>>
>         To: "Roy Golan" <rgolan at redhat.com <mailto:rgolan at redhat.com>>
>         Cc: "Artyom Lukianov" <alukiano at redhat.com
>         <mailto:alukiano at redhat.com>>, users at ovirt.org
>         <mailto:users at ovirt.org>
>         Sent: Sunday, July 12, 2015 4:21:34 PM
>         Subject: Re: [ovirt-users] This VM is not managed by the engine
>
>         [root at hv-02 etc]# vdsClient -s 0 destroy
>         41703d5c-6cdb-42b4-93df-d78be2776e2b
>
>         Unexpected exception
>
>         Not sure I'm getting any closer :-)
>
>
>
>
>         ***
>         *Mark Steele*
>         CIO / VP Technical Operations | TelVue Corporation
>         TelVue - We Share Your Vision
>         800.885.8886 x128 <tel:800.885.8886%20x128> |
>         msteele at telvue.com <mailto:msteele at telvue.com> |
>         http://www.telvue.com
>         twitter: http://twitter.com/telvue | facebook:
>         https://www.facebook.com/telvue
>
>         On Sun, Jul 12, 2015 at 9:13 AM, Mark Steele
>         <msteele at telvue.com <mailto:msteele at telvue.com>> wrote:
>
>         > OK - I think I'm getting closer - now that I'm on the
>         correct box.
>         >
>         > here is the output of the vdsClient command - which is the
>         device id - the
>         > first line?
>         >
>         > [root at hv-02 etc]# vdsClient -s 0 list
>         >
>         > 41703d5c-6cdb-42b4-93df-d78be2776e2b
>         > Status = Up
>         > acpiEnable = true
>         > emulatedMachine = rhel6.5.0
>         > afterMigrationStatus =
>         > pid = 27304
>         > memGuaranteedSize = 2048
>         > transparentHugePages = true
>         > displaySecurePort = 5902
>         > spiceSslCipherSuite = DEFAULT
>         > cpuType = SandyBridge
>         > smp = 2
>         > numaTune = {'nodeset': '0,1', 'mode': 'interleave'}
>         > custom =
>         >
>         {'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72':
>         > 'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>         > deviceId=ebd4c73d-12c4-435e-8cc5-f180d8f20a72, device=unix,
>         type=CHANNEL,
>         > bootOrder=0, specParams={}, address={bus=0, controller=0,
>         > type=virtio-serial, port=2}, managed=false, plugged=true,
>         readOnly=false,
>         > deviceAlias=channel1, customProperties={}, snapshotId=null}',
>         >
>         'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72device_ffd2796f-7644-4008-b920-5f0970b0ef0e':
>         > 'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>         > deviceId=ffd2796f-7644-4008-b920-5f0970b0ef0e, device=unix,
>         type=CHANNEL,
>         > bootOrder=0, specParams={}, address={bus=0, controller=0,
>         > type=virtio-serial, port=1}, managed=false, plugged=true,
>         readOnly=false,
>         > deviceAlias=channel0, customProperties={}, snapshotId=null}',
>         > 'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6': 'VmDevice
>         > {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>         > deviceId=86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6, device=ide,
>         type=CONTROLLER,
>         > bootOrder=0, specParams={}, address={slot=0x01, bus=0x00,
>         domain=0x0000,
>         > type=pci, function=0x1}, managed=false, plugged=true,
>         readOnly=false,
>         > deviceAlias=ide0, customProperties={}, snapshotId=null}',
>         >
>         'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72device_ffd2796f-7644-4008-b920-5f0970b0ef0edevice_6693d023-9c1f-433c-870e-e9771be8474b':
>         > 'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>         > deviceId=6693d023-9c1f-433c-870e-e9771be8474b, device=spicevmc,
>         > type=CHANNEL, bootOrder=0, specParams={}, address={bus=0,
>         controller=0,
>         > type=virtio-serial, port=3}, managed=false, plugged=true,
>         readOnly=false,
>         > deviceAlias=channel2, customProperties={}, snapshotId=null}'}
>         > vmType = kvm
>         > memSize = 2048
>         > smpCoresPerSocket = 1
>         > vmName = connect-turbo-stage-03
>         > nice = 0
>         > bootMenuEnable = false
>         > copyPasteEnable = true
>         > displayIp = 10.1.90.161
>         > displayPort = -1
>         > smartcardEnable = false
>         > clientIp =
>         > fileTransferEnable = true
>         > nicModel = rtl8139,pv
>         > keyboardLayout = en-us
>         > kvmEnable = true
>         > pitReinjection = false
>         > displayNetwork = ovirtmgmt
>         > devices = [{'target': 2097152, 'specParams': {'model':
>         'none'}, 'alias':
>         > 'balloon0', 'deviceType': 'balloon', 'device': 'memballoon',
>         'type':
>         > 'balloon'}, {'device': 'unix', 'alias': 'channel0',
>         'address': {'bus': '0',
>         > 'controller': '0', 'type': 'virtio-serial', 'port': '1'},
>         'deviceType':
>         > 'channel', 'type': 'channel'}, {'device': 'unix', 'alias':
>         'channel1',
>         > 'address': {'bus': '0', 'controller': '0', 'type':
>         'virtio-serial', 'port':
>         > '2'}, 'deviceType': 'channel', 'type': 'channel'},
>         {'device': 'spicevmc',
>         > 'alias': 'channel2', 'address': {'bus': '0', 'controller':
>         '0', 'type':
>         > 'virtio-serial', 'port': '3'}, 'deviceType': 'channel',
>         'type': 'channel'},
>         > {'index': '0', 'alias': 'scsi0', 'specParams': {}, 'deviceType':
>         > 'controller', 'deviceId':
>         '88db8cb9-0960-4797-bd41-1694bf14b8a9',
>         > 'address': {'slot': '0x04', 'bus': '0x00', 'domain':
>         '0x0000', 'type':
>         > 'pci', 'function': '0x0'}, 'device': 'scsi', 'model':
>         'virtio-scsi',
>         > 'type': 'controller'}, {'alias': 'virtio-serial0',
>         'specParams': {},
>         > 'deviceType': 'controller', 'deviceId':
>         > '4bb9c112-e027-4e7d-8b1c-32f99c7040ee', 'address': {'slot':
>         '0x05', 'bus':
>         > '0x00', 'domain': '0x0000', 'type': 'pci', 'function':
>         '0x0'}, 'device':
>         > 'virtio-serial', 'type': 'controller'}, {'device': 'usb',
>         'alias': 'usb0',
>         > 'address': {'slot': '0x01', 'bus': '0x00', 'domain':
>         '0x0000', 'type':
>         > 'pci', 'function': '0x2'}, 'deviceType': 'controller', 'type':
>         > 'controller'}, {'device': 'ide', 'alias': 'ide0', 'address':
>         {'slot':
>         > '0x01', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>         'function':
>         > '0x1'}, 'deviceType': 'controller', 'type': 'controller'},
>         {'alias':
>         > 'video0', 'specParams': {'vram': '32768', 'ram': '65536',
>         'heads': '1'},
>         > 'deviceType': 'video', 'deviceId':
>         '23634541-3b7e-460d-9580-39504392ba36',
>         > 'address': {'slot': '0x02', 'bus': '0x00', 'domain':
>         '0x0000', 'type':
>         > 'pci', 'function': '0x0'}, 'device': 'qxl', 'type':
>         'video'}, {'device':
>         > 'spice', 'specParams': {'displayNetwork': 'ovirtmgmt',
>         > 'spiceSecureChannels':
>         >
>         'smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard',
>         > 'keyMap': 'en-us', 'displayIp': '10.1.90.161',
>         'copyPasteEnable': 'true'},
>         > 'deviceType': 'graphics', 'tlsPort': '5902', 'type':
>         'graphics'},
>         > {'nicModel': 'pv', 'macAddr': '00:01:a4:a2:b4:30',
>         'linkActive': True,
>         > 'network': 'ovirtmgmt', 'alias': 'net0', 'filter':
>         'vdsm-no-mac-spoofing',
>         > 'specParams': {'inbound': {}, 'outbound': {}}, 'deviceType':
>         'interface',
>         > 'deviceId': '63651662-2ddf-4611-b988-1a58d05982f6',
>         'address': {'slot':
>         > '0x03', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>         'function':
>         > '0x0'}, 'device': 'bridge', 'type': 'interface', 'name':
>         'vnet5'},
>         > {'nicModel': 'pv', 'macAddr': '00:01:a4:a2:b4:31',
>         'linkActive': True,
>         > 'network': 'storage', 'alias': 'net1', 'filter':
>         'vdsm-no-mac-spoofing',
>         > 'specParams': {'inbound': {}, 'outbound': {}}, 'deviceType':
>         'interface',
>         > 'deviceId': 'b112d9c6-5144-4b67-912b-dcc27aabdae9',
>         'address': {'slot':
>         > '0x07', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>         'function':
>         > '0x0'}, 'device': 'bridge', 'type': 'interface', 'name':
>         'vnet6'},
>         > {'index': '3', 'iface': 'ide', 'name': 'hdd', 'alias':
>         'ide0-1-1',
>         > 'specParams': {'vmPayload': {'volId': 'config-2', 'file':
>         > {'openstack/latest/meta_data.json':
>         >
>         'ewogICJsYXVuY2hfaW5kZXgiIDogIjAiLAogICJhdmFpbGFiaWxpdHlfem9uZSIgOiAibm92YSIs\nCiAgIm5hbWUiIDogImNvbm5lY3QtdHVyYm8tc3RhZ2UtMDMiLAogICJob3N0bmFtZSIgOiAiY29u\nbmVjdC10dXJiby1zdGFnZS0wMyIsCiAgInV1aWQiIDogImJiNmIwMzdhLTZkY2ItNGZmZS04MjUw\nLTMwYjlkOWE0ZTlmZCIsCiAgIm1ldGEiIDogewogICAgImVzc2VudGlhbCIgOiAiZmFsc2UiLAog\nICAgInJvbGUiIDogInNlcnZlciIsCiAgICAiZHNtb2RlIiA6ICJsb2NhbCIKICB9Cn0=\n',
>         > 'openstack/latest/user_data':
>         >
>         'I2Nsb3VkLWNvbmZpZwpzc2hfcHdhdXRoOiB0cnVlCmRpc2FibGVfcm9vdDogMApvdXRwdXQ6CiAg\nYWxsOiAnPj4gL3Zhci9sb2cvY2xvdWQtaW5pdC1vdXRwdXQubG9nJwpjaHBhc3N3ZDoKICBleHBp\ncmU6IGZhbHNlCnJ1bmNtZDoKLSAnc2VkIC1pICcnL15kYXRhc291cmNlX2xpc3Q6IC9kJycgL2V0\nYy9jbG91ZC9jbG91ZC5jZmc7IGVjaG8gJydkYXRhc291cmNlX2xpc3Q6CiAgWyJOb0Nsb3VkIiwg\nIkNvbmZpZ0RyaXZlIl0nJyA+PiAvZXRjL2Nsb3VkL2Nsb3VkLmNmZycK\n'}}},
>         > 'readonly': 'True', 'deviceType': 'disk', 'deviceId':
>         > '6de890b2-6454-4377-9a71-bea2e46d50a8', 'address': {'bus': '1',
>         > 'controller': '0', 'type': 'drive', 'target': '0', 'unit':
>         '1'}, 'device':
>         > 'cdrom', 'shared': 'false', 'path': '', 'type': 'disk'},
>         {'index': '2',
>         > 'iface': 'ide', 'name': 'hdd', 'alias': 'ide0-1-1',
>         'specParams': {'path':
>         > ''}, 'readonly': 'True', 'deviceType': 'disk', 'deviceId':
>         > '2763a41b-6576-4135-b349-4fe402c31246', 'address': {'bus': '1',
>         > 'controller': '0', 'type': 'drive', 'target': '0', 'unit':
>         '1'}, 'device':
>         > 'cdrom', 'shared': 'false', 'path': '', 'type': 'disk'},
>         {'device': 'file',
>         > 'alias': 'ide0-1-0', 'address': {'bus': '1', 'controller':
>         '0', 'type':
>         > 'drive', 'target': '0', 'unit': '0'}, 'deviceType': 'disk',
>         'type': 'disk'}]
>         > timeOffset = -891891
>         > maxVCpus = 16
>         > spiceSecureChannels =
>         >
>         smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard
>         > display = qxl
>         > [root at hv-02 etc]#
>         >
>         >
>         > ***
>         > *Mark Steele*
>         > CIO / VP Technical Operations | TelVue Corporation
>         > TelVue - We Share Your Vision
>         > 800.885.8886 x128 <tel:800.885.8886%20x128> |
>         msteele at telvue.com <mailto:msteele at telvue.com> |
>         http://www.telvue.com
>         > twitter: http://twitter.com/telvue | facebook:
>         > https://www.facebook.com/telvue
>         >
>         > On Sun, Jul 12, 2015 at 9:06 AM, Mark Steele <msteele at telvue.com <mailto:msteele at telvue.com>>
>         wrote:
>         >
>         >> I think I may have not given you all the information.
>         >>
>         >> I am not logging into the host - I am logging into the
>         ovirt management.
>         >>
>         >> Let me try logging into the host and checking
>         >>
>         >>
>         >>
>         >> ***
>         >> *Mark Steele*
>         >> CIO / VP Technical Operations | TelVue Corporation
>         >> TelVue - We Share Your Vision
>         >> 800.885.8886 x128 <tel:800.885.8886%20x128> |
>         msteele at telvue.com <mailto:msteele at telvue.com> |
>         http://www.telvue.com
>         >> twitter: http://twitter.com/telvue | facebook:
>         >> https://www.facebook.com/telvue
>         >>
>         >> On Sun, Jul 12, 2015 at 9:03 AM, Roy Golan <rgolan at redhat.com <mailto:rgolan at redhat.com>>
>         wrote:
>         >>
>         >>>  On 07/12/2015 03:52 PM, Mark Steele wrote:
>         >>>
>         >>>  That command returns nothing - I don't think qemu is running?
>         >>>
>         >>>  Not sure how to start it on CentOS
>         >>>
>         >>>  [root at ovirt-01 ~]# ps -ef | grep qemu
>         >>>
>         >>> root     23279 23130  0 08:51 pts/0    00:00:00 grep qemu
>         >>>
>         >>>
>         >>>   that mean you don't have vm running on that host. so you
>         can restart
>         >>> vdsm
>         >>>
>         >>>
>         >>>
>         >>>  ***
>         >>> *Mark Steele*
>         >>>  CIO / VP Technical Operations | TelVue Corporation
>         >>> TelVue - We Share Your Vision
>         >>> 800.885.8886 x128 <tel:800.885.8886%20x128> |
>         msteele at telvue.com <mailto:msteele at telvue.com> | 
>         <http://www.telvue.com/>
>         >>> http://www.telvue.com
>         >>> twitter: 
>         <http://twitter.com/telvue>http://twitter.com/telvue |
>         >>> facebook:  <https://www.facebook.com/telvue>
>         >>>https://www.facebook.com/telvue
>         >>>
>         >>> On Sun, Jul 12, 2015 at 8:45 AM, Roy Golan
>         <rgolan at redhat.com <mailto:rgolan at redhat.com>> wrote:
>         >>>
>         >>>>  On 07/12/2015 03:42 PM, Mark Steele wrote:
>         >>>>
>         >>>>  I run into the same issue - I am unable to completely go
>         into
>         >>>> maintenance mode because this VM is still on it - it
>         cannot be migrated
>         >>>> because it is not managed.
>         >>>>
>         >>>>  find you qemu process:
>         >>>>   pgrep -an qemu-kvm | grep external
>         >>>>
>         >>>> and kill the process
>         >>>>
>         >>>>
>         >>>>
>         >>>>  [image: Inline image 1]
>         >>>>
>         >>>>
>         >>>>  ***
>         >>>> *Mark Steele*
>         >>>>  CIO / VP Technical Operations | TelVue Corporation
>         >>>> TelVue - We Share Your Vision
>         >>>> 800.885.8886 x128 <tel:800.885.8886%20x128> | 
>         <msteele at telvue.com
>         <mailto:msteele at telvue.com>>msteele at telvue.com
>         <mailto:msteele at telvue.com> |
>         >>>> <http://www.telvue.com>http://www.telvue.com
>         >>>> twitter: 
>         <http://twitter.com/telvue>http://twitter.com/telvue |
>         >>>> facebook:  <https://www.facebook.com/telvue>
>         >>>> https://www.facebook.com/telvue
>         >>>>
>         >>>> On Sun, Jul 12, 2015 at 8:25 AM, Roy Golan <
>         <rgolan at redhat.com <mailto:rgolan at redhat.com>>
>         >>>> rgolan at redhat.com <mailto:rgolan at redhat.com>> wrote:
>         >>>>
>         >>>>>  On 07/12/2015 03:12 PM, Mark Steele wrote:
>         >>>>>
>         >>>>>  I think I may have found the problem:
>         >>>>>
>         >>>>>  [root at ovirt-01 pki]# ls -lah
>         >>>>> total 48K
>         >>>>> drwxr-xr-x.  10 root root  4.0K Nov 14  2014 .
>         >>>>> drwxr-xr-x. 118 root root   12K Jul 12 03:35 ..
>         >>>>> drwxr-xr-x.   6 root root  4.0K Nov 14  2014 CA
>         >>>>> drwxr-xr-x.   4 root root  4.0K Nov 14  2014 ca-trust
>         >>>>> drwxr-xr-x.   2 root root  4.0K Nov 14  2014 java
>         >>>>> drwxr-xr-x.   2 root root  4.0K Jul 12 07:03 nssdb
>         >>>>> drwxr-xr-x.   6 ovirt ovirt 4.0K Nov 19  2014 ovirt-engine
>         >>>>> drwxr-xr-x.   2 root root  4.0K Nov 14  2014 rpm-gpg
>         >>>>> drwx------.   2 root root  4.0K Nov 22  2013 rsyslog
>         >>>>> drwxr-xr-x.   5 root root  4.0K Nov 14  2014 tls
>         >>>>> [root at ovirt-01 pki]#
>         >>>>>
>         >>>>>  There is no vsdm directory under /etc/pki
>         >>>>>
>         >>>>>  This is an ovirt node. Version of software is
>         3.5.0.1-1.el6 from the
>         >>>>> ovirt management console.
>         >>>>>
>         >>>>>  I'd like to add that I am not the person who originally
>         installed
>         >>>>> this instance - and am not entirely familiar with how it
>         is setup and
>         >>>>> installed - so I may ask ignorant questions from time to
>         time.
>         >>>>>
>         >>>>>
>         >>>>>  not urgent but at this point it looks like it would be
>         good to
>         >>>>> reinstall this host from the webadmin. if you have the
>         capacity,
>         >>>>> you can put the host to maintenance, that will migrate
>         vms to other
>         >>>>> hosts, and then choose "reinstall" once its in "maintenance'
>         >>>>>
>         >>>>>
>         >>>>>  ***
>         >>>>> *Mark Steele*
>         >>>>>  CIO / VP Technical Operations | TelVue Corporation
>         >>>>> TelVue - We Share Your Vision
>         >>>>> 800.885.8886 x128 <tel:800.885.8886%20x128> | 
>         <msteele at telvue.com
>         <mailto:msteele at telvue.com>>msteele at telvue.com
>         <mailto:msteele at telvue.com> |
>         >>>>> <http://www.telvue.com>http://www.telvue.com
>         >>>>> twitter: http://twitter.com/telvue | facebook:
>         >>>>> https://www.facebook.com/telvue
>         >>>>>
>         >>>>> On Sun, Jul 12, 2015 at 8:02 AM, Roy Golan <
>         <rgolan at redhat.com <mailto:rgolan at redhat.com>>
>         >>>>> rgolan at redhat.com <mailto:rgolan at redhat.com>> wrote:
>         >>>>>
>         >>>>>>  On 07/12/2015 02:07 PM, Mark Steele wrote:
>         >>>>>>
>         >>>>>>  Thank you Roy,
>         >>>>>>
>         >>>>>>  I installed the client but am getting a permissions
>         error when I
>         >>>>>> run it
>         >>>>>>
>         >>>>>>  [root at ovirt-01 ~]# vdsClient -s 0 list
>         >>>>>> Traceback (most recent call last):
>         >>>>>>   File "/usr/share/vdsm/vdsClient.py", line 2678, in
>         <module>
>         >>>>>>  serv.do_connect(hostPort)
>         >>>>>>   File "/usr/share/vdsm/vdsClient.py", line 136, in
>         do_connect
>         >>>>>>     self.s = vdscli.connect(hostPort, self.useSSL,
>         self.truststore)
>         >>>>>>   File
>         "/usr/lib/python2.6/site-packages/vdsm/vdscli.py", line 110,
>         >>>>>> in connect
>         >>>>>>     raise Exception("No permission to read file: %s" % f)
>         >>>>>> Exception: No permission to read file:
>         /etc/pki/vdsm/keys/vdsmkey.pem
>         >>>>>>
>         >>>>>>
>         >>>>>>  This should work. something isn't right with your setup
>         >>>>>> is your host an ovirt-node? could be that you hit [1] .
>         let me know
>         >>>>>> what version are you running.
>         >>>>>>
>         >>>>>> please try the same with user vdsm. it should have
>         permissions to
>         >>>>>> /etc/pki/vdsm
>         >>>>>>
>         >>>>>> [1]  <https://gerrit.ovirt.org/#/c/27779/>
>         >>>>>> https://gerrit.ovirt.org/#/c/27779/
>         >>>>>>
>         >>>>>>
>         >>>>>>   If I restart vdsm, will that cause any issues with
>         running VM's on
>         >>>>>> this ovirt installation? This is our production
>         environment.
>         >>>>>>
>         >>>>>>
>         >>>>>>  Generatlly the answer is no but lets avoid it if  we
>         can for this is
>         >>>>>> a minor cosmetic issue I guess.
>         >>>>>>
>         >>>>>> just as FYI - vdsm only reconnects to the socket
>         exposed by libvirt
>         >>>>>> to control the vm lifecycle. VDSM doesn't manadate the
>         lifecycle of a VM
>         >>>>>> unless the engine tells is so. Storage wise there could
>         be some operations
>         >>>>>> but i'm almost sure they must not have effect on
>         running VMs.
>         >>>>>>
>         >>>>>>
>         >>>>>>  Thank you
>         >>>>>>
>         >>>>>>
>         >>>>>>  ***
>         >>>>>> *Mark Steele*
>         >>>>>>  CIO / VP Technical Operations | TelVue Corporation
>         >>>>>> TelVue - We Share Your Vision
>         >>>>>> 800.885.8886 x128 <tel:800.885.8886%20x128> | 
>         <msteele at telvue.com
>         <mailto:msteele at telvue.com>>msteele at telvue.com
>         <mailto:msteele at telvue.com> |
>         >>>>>> <http://www.telvue.com>http://www.telvue.com
>         >>>>>> twitter: http://twitter.com/telvue | facebook:
>         >>>>>>
>         <https://www.facebook.com/telvue>https://www.facebook.com/telvue
>         >>>>>>
>         >>>>>> On Sun, Jul 12, 2015 at 4:09 AM, Roy Golan <
>         <rgolan at redhat.com <mailto:rgolan at redhat.com>>
>         >>>>>> rgolan at redhat.com <mailto:rgolan at redhat.com>> wrote:
>         >>>>>>
>         >>>>>>>  On 07/09/2015 06:34 PM, Mark Steele wrote:
>         >>>>>>>
>         >>>>>>>  Yes,
>         >>>>>>>
>         >>>>>>>  It is displayed in the engine:
>         >>>>>>>
>         >>>>>>>  [image: Inline image 1]
>         >>>>>>>
>         >>>>>>>
>         >>>>>>>  the vdsm on that host reports it back to the engine .
>         since this vm
>         >>>>>>> isn't in the engine DB it is concidered as EXTERNAL
>         (thus the error 400
>         >>>>>>> from the API)
>         >>>>>>>
>         >>>>>>> do yo know if the qemu-kvm proccess isn't running anymore?
>         >>>>>>>
>         >>>>>>> if the process isn't running then vdsm must clean its
>         cache
>         >>>>>>>
>         >>>>>>> try to:
>         >>>>>>>
>         >>>>>>> yum install vdsm-cli
>         >>>>>>> vdsClient -s 0 list
>         >>>>>>> vdsClient -s 0 destroy {vmId}
>         >>>>>>>
>         >>>>>>> alternativly a vdsm restart will work (if the qemu
>         proccess isn't
>         >>>>>>> running)
>         >>>>>>>
>         >>>>>>>
>         >>>>>>>  The VM is not really running - the IP addresses that
>         are being
>         >>>>>>> reported are from another VM that was recently
>         removed. All attempts to
>         >>>>>>> control the VM have failed. It does not have any NICS
>         or disk associated
>         >>>>>>> with it - so this seems to be a ghost in the machine.
>         I attempted to unlock
>         >>>>>>> it using the unlock_entity.sh script - it reports
>         successful, however I
>         >>>>>>> still cannot do anything with the VM.
>         >>>>>>>
>         >>>>>>>
>         >>>>>>>  ***
>         >>>>>>> *Mark Steele*
>         >>>>>>>  CIO / VP Technical Operations | TelVue Corporation
>         >>>>>>> TelVue - We Share Your Vision
>         >>>>>>> 800.885.8886 x128 <tel:800.885.8886%20x128> | 
>         <msteele at telvue.com
>         <mailto:msteele at telvue.com>>msteele at telvue.com
>         <mailto:msteele at telvue.com> |
>         >>>>>>> <http://www.telvue.com>http://www.telvue.com
>         >>>>>>> twitter: http://twitter.com/telvue | facebook:
>         >>>>>>>
>         <https://www.facebook.com/telvue>https://www.facebook.com/telvue
>         >>>>>>>
>         >>>>>>> On Thu, Jul 9, 2015 at 11:11 AM, Artyom Lukianov <
>         >>>>>>> <alukiano at redhat.com
>         <mailto:alukiano at redhat.com>>alukiano at redhat.com
>         <mailto:alukiano at redhat.com>> wrote:
>         >>>>>>>
>         >>>>>>>> Can you sea via engine, on what host run VM?
>         >>>>>>>> Anyway if you have really run VM on host you can try
>         to figure it
>         >>>>>>>> with 'ps aux | grep qemu', if it will return you some
>         process, you can just
>         >>>>>>>> kill process via 'kill pid'.
>         >>>>>>>> I hope it will help you.
>         >>>>>>>>
>         >>>>>>>> ----- Original Message -----
>         >>>>>>>> From: "Mark Steele" < <msteele at telvue.com
>         <mailto:msteele at telvue.com>>msteele at telvue.com
>         <mailto:msteele at telvue.com>>
>         >>>>>>>> To: "Artyom Lukianov" < <alukiano at redhat.com
>         <mailto:alukiano at redhat.com>>alukiano at redhat.com
>         <mailto:alukiano at redhat.com>>
>         >>>>>>>> Cc: <users at ovirt.org
>         <mailto:users at ovirt.org>>users at ovirt.org <mailto:users at ovirt.org>
>         >>>>>>>> Sent: Thursday, July 9, 2015 5:42:20 PM
>         >>>>>>>> Subject: Re: [ovirt-users] This VM is not managed by
>         the engine
>         >>>>>>>>
>         >>>>>>>> Artyom,
>         >>>>>>>>
>         >>>>>>>> Thank you - I don't have vdsClient installed - can
>         you point me to
>         >>>>>>>> the
>         >>>>>>>> download?
>         >>>>>>>>
>         >>>>>>>>
>         >>>>>>>> ***
>         >>>>>>>> *Mark Steele*
>         >>>>>>>>  CIO / VP Technical Operations | TelVue Corporation
>         >>>>>>>> TelVue - We Share Your Vision
>         >>>>>>>> 800.885.8886 x128 <tel:800.885.8886%20x128>
>         <800.885.8886%20x128> | <msteele at telvue.com
>         <mailto:msteele at telvue.com>>
>         >>>>>>>> msteele at telvue.com <mailto:msteele at telvue.com> |
>         <http://www.telvue.com>http://www.telvue.com
>         >>>>>>>> twitter:
>         <http://twitter.com/telvue>http://twitter.com/telvue |
>         >>>>>>>> facebook:
>         >>>>>>>> 
>         <https://www.facebook.com/telvue>https://www.facebook.com/telvue
>         >>>>>>>>
>         >>>>>>>> On Thu, Jul 9, 2015 at 10:11 AM, Artyom Lukianov <
>         >>>>>>>> <alukiano at redhat.com
>         <mailto:alukiano at redhat.com>>alukiano at redhat.com
>         <mailto:alukiano at redhat.com>>
>         >>>>>>>> wrote:
>         >>>>>>>>
>         >>>>>>>> > Please check host where VM run(vdsClient -s 0 list
>         table), and
>         >>>>>>>> you can
>         >>>>>>>> > destroy it via vdsClient(vdsClient -s 0 destroy vm_id).
>         >>>>>>>> > Thanks
>         >>>>>>>> >
>         >>>>>>>> > ----- Original Message -----
>         >>>>>>>> > From: "Mark Steele" < <msteele at telvue.com
>         <mailto:msteele at telvue.com>>msteele at telvue.com
>         <mailto:msteele at telvue.com>>
>         >>>>>>>> > To: <users at ovirt.org
>         <mailto:users at ovirt.org>>users at ovirt.org <mailto:users at ovirt.org>
>         >>>>>>>> > Sent: Thursday, July 9, 2015 4:38:32 PM
>         >>>>>>>> > Subject: [ovirt-users] This VM is not managed by
>         the engine
>         >>>>>>>> >
>         >>>>>>>> > I have a VM that was not started and is now showing
>         as running.
>         >>>>>>>> When I
>         >>>>>>>> > attempt to suspend or stop it in the ovirt-shell, I
>         get the
>         >>>>>>>> message:
>         >>>>>>>> >
>         >>>>>>>> > status: 400
>         >>>>>>>> > reason: bad request
>         >>>>>>>> > detail: Cannot hibernate VM. This VM is not managed
>         by the engine.
>         >>>>>>>> >
>         >>>>>>>> > Not sure how the VM was initially created on the
>         ovirt manager.
>         >>>>>>>> This VM is
>         >>>>>>>> > not needed - how can I 'shutdown' and remove this VM?
>         >>>>>>>> >
>         >>>>>>>> > Thanks
>         >>>>>>>> >
>         >>>>>>>> > ***
>         >>>>>>>> > Mark Steele
>         >>>>>>>> > CIO / VP Technical Operations | TelVue Corporation
>         >>>>>>>> > TelVue - We Share Your Vision
>         >>>>>>>> > 800.885.8886 x128 <tel:800.885.8886%20x128>
>         <800.885.8886%20x128> | <msteele at telvue.com
>         <mailto:msteele at telvue.com>>
>         >>>>>>>> msteele at telvue.com <mailto:msteele at telvue.com> |
>         <http://www.telvue.com>http://www.telvue.com
>         >>>>>>>> > twitter:
>         <http://twitter.com/telvue>http://twitter.com/telvue |
>         >>>>>>>> facebook:
>         >>>>>>>> >
>         <https://www.facebook.com/telvue>https://www.facebook.com/telvue
>         >>>>>>>> >
>         >>>>>>>> > _______________________________________________
>         >>>>>>>> > Users mailing list
>         >>>>>>>> > <Users at ovirt.org
>         <mailto:Users at ovirt.org>>Users at ovirt.org <mailto:Users at ovirt.org>
>         >>>>>>>> > <http://lists.ovirt.org/mailman/listinfo/users>
>         >>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>         >>>>>>>> >
>         >>>>>>>>
>         >>>>>>>
>         >>>>>>>
>         >>>>>>>
>         >>>>>>> _______________________________________________
>         >>>>>>> Users mailing
>         listUsers at ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>         <http://lists.ovirt.org/mailman/listinfo/users>
>         >>>>>>>
>         >>>>>>>
>         >>>>>>>
>         >>>>>>
>         >>>>>>
>         >>>>>
>         >>>>>
>         >>>>
>         >>>>
>         >>>
>         >>>
>         >>
>         >
>
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150713/c4deaad5/attachment-0001.html>


More information about the Users mailing list