[ovirt-users] This VM is not managed by the engine

Roy Golan rgolan at redhat.com
Sun Jul 12 13:58:39 UTC 2015


On 07/12/2015 04:21 PM, Mark Steele wrote:
>
> [root at hv-02 etc]# vdsClient -s 0 destroy 
> 41703d5c-6cdb-42b4-93df-d78be2776e2b
>
> Unexpected exception
>
> Not sure I'm getting any closer :-)
>
>


start with a shutdown - if it won't go now search and kill the process

vdsClient -s 0 shutdown41703d5c-6cdb-42b4-93df-d78be2776e2b

>
>
> ***
> *Mark Steele*
> CIO / VP Technical Operations | TelVue Corporation
> TelVue - We Share Your Vision
> 800.885.8886 x128 | msteele at telvue.com <mailto:msteele at telvue.com> | 
> http://www.telvue.com <http://www.telvue.com/>
> twitter: http://twitter.com/telvue | facebook: 
> https://www.facebook.com/telvue
>
> On Sun, Jul 12, 2015 at 9:13 AM, Mark Steele <msteele at telvue.com 
> <mailto:msteele at telvue.com>> wrote:
>
>     OK - I think I'm getting closer - now that I'm on the correct box.
>
>     here is the output of the vdsClient command - which is the device
>     id - the first line?
>
>     [root at hv-02 etc]# vdsClient -s 0 list
>
>     41703d5c-6cdb-42b4-93df-d78be2776e2b
>     Status = Up
>     acpiEnable = true
>     emulatedMachine = rhel6.5.0
>     afterMigrationStatus =
>     pid = 27304
>     memGuaranteedSize = 2048
>     transparentHugePages = true
>     displaySecurePort = 5902
>     spiceSslCipherSuite = DEFAULT
>     cpuType = SandyBridge
>     smp = 2
>     numaTune = {'nodeset': '0,1', 'mode': 'interleave'}
>     custom =
>     {'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72':
>     'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>     deviceId=ebd4c73d-12c4-435e-8cc5-f180d8f20a72, device=unix,
>     type=CHANNEL, bootOrder=0, specParams={}, address={bus=0,
>     controller=0, type=virtio-serial, port=2}, managed=false,
>     plugged=true, readOnly=false, deviceAlias=channel1,
>     customProperties={}, snapshotId=null}',
>     'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72device_ffd2796f-7644-4008-b920-5f0970b0ef0e':
>     'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>     deviceId=ffd2796f-7644-4008-b920-5f0970b0ef0e, device=unix,
>     type=CHANNEL, bootOrder=0, specParams={}, address={bus=0,
>     controller=0, type=virtio-serial, port=1}, managed=false,
>     plugged=true, readOnly=false, deviceAlias=channel0,
>     customProperties={}, snapshotId=null}',
>     'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6': 'VmDevice
>     {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>     deviceId=86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6, device=ide,
>     type=CONTROLLER, bootOrder=0, specParams={}, address={slot=0x01,
>     bus=0x00, domain=0x0000, type=pci, function=0x1}, managed=false,
>     plugged=true, readOnly=false, deviceAlias=ide0,
>     customProperties={}, snapshotId=null}',
>     'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72device_ffd2796f-7644-4008-b920-5f0970b0ef0edevice_6693d023-9c1f-433c-870e-e9771be8474b':
>     'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
>     deviceId=6693d023-9c1f-433c-870e-e9771be8474b, device=spicevmc,
>     type=CHANNEL, bootOrder=0, specParams={}, address={bus=0,
>     controller=0, type=virtio-serial, port=3}, managed=false,
>     plugged=true, readOnly=false, deviceAlias=channel2,
>     customProperties={}, snapshotId=null}'}
>     vmType = kvm
>     memSize = 2048
>     smpCoresPerSocket = 1
>     vmName = connect-turbo-stage-03
>     nice = 0
>     bootMenuEnable = false
>     copyPasteEnable = true
>     displayIp = 10.1.90.161
>     displayPort = -1
>     smartcardEnable = false
>     clientIp =
>     fileTransferEnable = true
>     nicModel = rtl8139,pv
>     keyboardLayout = en-us
>     kvmEnable = true
>     pitReinjection = false
>     displayNetwork = ovirtmgmt
>     devices = [{'target': 2097152, 'specParams': {'model': 'none'},
>     'alias': 'balloon0', 'deviceType': 'balloon', 'device':
>     'memballoon', 'type': 'balloon'}, {'device': 'unix', 'alias':
>     'channel0', 'address': {'bus': '0', 'controller': '0', 'type':
>     'virtio-serial', 'port': '1'}, 'deviceType': 'channel', 'type':
>     'channel'}, {'device': 'unix', 'alias': 'channel1', 'address':
>     {'bus': '0', 'controller': '0', 'type': 'virtio-serial', 'port':
>     '2'}, 'deviceType': 'channel', 'type': 'channel'}, {'device':
>     'spicevmc', 'alias': 'channel2', 'address': {'bus': '0',
>     'controller': '0', 'type': 'virtio-serial', 'port': '3'},
>     'deviceType': 'channel', 'type': 'channel'}, {'index': '0',
>     'alias': 'scsi0', 'specParams': {}, 'deviceType': 'controller',
>     'deviceId': '88db8cb9-0960-4797-bd41-1694bf14b8a9', 'address':
>     {'slot': '0x04', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>     'function': '0x0'}, 'device': 'scsi', 'model': 'virtio-scsi',
>     'type': 'controller'}, {'alias': 'virtio-serial0', 'specParams':
>     {}, 'deviceType': 'controller', 'deviceId':
>     '4bb9c112-e027-4e7d-8b1c-32f99c7040ee', 'address': {'slot':
>     '0x05', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>     'function': '0x0'}, 'device': 'virtio-serial', 'type':
>     'controller'}, {'device': 'usb', 'alias': 'usb0', 'address':
>     {'slot': '0x01', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>     'function': '0x2'}, 'deviceType': 'controller', 'type':
>     'controller'}, {'device': 'ide', 'alias': 'ide0', 'address':
>     {'slot': '0x01', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>     'function': '0x1'}, 'deviceType': 'controller', 'type':
>     'controller'}, {'alias': 'video0', 'specParams': {'vram': '32768',
>     'ram': '65536', 'heads': '1'}, 'deviceType': 'video', 'deviceId':
>     '23634541-3b7e-460d-9580-39504392ba36', 'address': {'slot':
>     '0x02', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>     'function': '0x0'}, 'device': 'qxl', 'type': 'video'}, {'device':
>     'spice', 'specParams': {'displayNetwork': 'ovirtmgmt',
>     'spiceSecureChannels':
>     'smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard',
>     'keyMap': 'en-us', 'displayIp': '10.1.90.161', 'copyPasteEnable':
>     'true'}, 'deviceType': 'graphics', 'tlsPort': '5902', 'type':
>     'graphics'}, {'nicModel': 'pv', 'macAddr': '00:01:a4:a2:b4:30',
>     'linkActive': True, 'network': 'ovirtmgmt', 'alias': 'net0',
>     'filter': 'vdsm-no-mac-spoofing', 'specParams': {'inbound': {},
>     'outbound': {}}, 'deviceType': 'interface', 'deviceId':
>     '63651662-2ddf-4611-b988-1a58d05982f6', 'address': {'slot':
>     '0x03', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>     'function': '0x0'}, 'device': 'bridge', 'type': 'interface',
>     'name': 'vnet5'}, {'nicModel': 'pv', 'macAddr':
>     '00:01:a4:a2:b4:31', 'linkActive': True, 'network': 'storage',
>     'alias': 'net1', 'filter': 'vdsm-no-mac-spoofing', 'specParams':
>     {'inbound': {}, 'outbound': {}}, 'deviceType': 'interface',
>     'deviceId': 'b112d9c6-5144-4b67-912b-dcc27aabdae9', 'address':
>     {'slot': '0x07', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci',
>     'function': '0x0'}, 'device': 'bridge', 'type': 'interface',
>     'name': 'vnet6'}, {'index': '3', 'iface': 'ide', 'name': 'hdd',
>     'alias': 'ide0-1-1', 'specParams': {'vmPayload': {'volId':
>     'config-2', 'file': {'openstack/latest/meta_data.json':
>     'ewogICJsYXVuY2hfaW5kZXgiIDogIjAiLAogICJhdmFpbGFiaWxpdHlfem9uZSIgOiAibm92YSIs\nCiAgIm5hbWUiIDogImNvbm5lY3QtdHVyYm8tc3RhZ2UtMDMiLAogICJob3N0bmFtZSIgOiAiY29u\nbmVjdC10dXJiby1zdGFnZS0wMyIsCiAgInV1aWQiIDogImJiNmIwMzdhLTZkY2ItNGZmZS04MjUw\nLTMwYjlkOWE0ZTlmZCIsCiAgIm1ldGEiIDogewogICAgImVzc2VudGlhbCIgOiAiZmFsc2UiLAog\nICAgInJvbGUiIDogInNlcnZlciIsCiAgICAiZHNtb2RlIiA6ICJsb2NhbCIKICB9Cn0=\n',
>     'openstack/latest/user_data':
>     'I2Nsb3VkLWNvbmZpZwpzc2hfcHdhdXRoOiB0cnVlCmRpc2FibGVfcm9vdDogMApvdXRwdXQ6CiAg\nYWxsOiAnPj4gL3Zhci9sb2cvY2xvdWQtaW5pdC1vdXRwdXQubG9nJwpjaHBhc3N3ZDoKICBleHBp\ncmU6IGZhbHNlCnJ1bmNtZDoKLSAnc2VkIC1pICcnL15kYXRhc291cmNlX2xpc3Q6IC9kJycgL2V0\nYy9jbG91ZC9jbG91ZC5jZmc7IGVjaG8gJydkYXRhc291cmNlX2xpc3Q6CiAgWyJOb0Nsb3VkIiwg\nIkNvbmZpZ0RyaXZlIl0nJyA+PiAvZXRjL2Nsb3VkL2Nsb3VkLmNmZycK\n'}}},
>     'readonly': 'True', 'deviceType': 'disk', 'deviceId':
>     '6de890b2-6454-4377-9a71-bea2e46d50a8', 'address': {'bus': '1',
>     'controller': '0', 'type': 'drive', 'target': '0', 'unit': '1'},
>     'device': 'cdrom', 'shared': 'false', 'path': '', 'type': 'disk'},
>     {'index': '2', 'iface': 'ide', 'name': 'hdd', 'alias': 'ide0-1-1',
>     'specParams': {'path': ''}, 'readonly': 'True', 'deviceType':
>     'disk', 'deviceId': '2763a41b-6576-4135-b349-4fe402c31246',
>     'address': {'bus': '1', 'controller': '0', 'type': 'drive',
>     'target': '0', 'unit': '1'}, 'device': 'cdrom', 'shared': 'false',
>     'path': '', 'type': 'disk'}, {'device': 'file', 'alias':
>     'ide0-1-0', 'address': {'bus': '1', 'controller': '0', 'type':
>     'drive', 'target': '0', 'unit': '0'}, 'deviceType': 'disk',
>     'type': 'disk'}]
>     timeOffset = -891891
>     maxVCpus = 16
>     spiceSecureChannels =
>     smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard
>     display = qxl
>     [root at hv-02 etc]#
>
>
>     ***
>     *Mark Steele*
>     CIO / VP Technical Operations | TelVue Corporation
>     TelVue - We Share Your Vision
>     800.885.8886 x128 <tel:800.885.8886%20x128> | msteele at telvue.com
>     <mailto:msteele at telvue.com> | http://www.telvue.com
>     <http://www.telvue.com/>
>     twitter: http://twitter.com/telvue | facebook:
>     https://www.facebook.com/telvue
>
>     On Sun, Jul 12, 2015 at 9:06 AM, Mark Steele <msteele at telvue.com
>     <mailto:msteele at telvue.com>> wrote:
>
>         I think I may have not given you all the information.
>
>         I am not logging into the host - I am logging into the ovirt
>         management.
>
>         Let me try logging into the host and checking
>
>
>
>         ***
>         *Mark Steele*
>         CIO / VP Technical Operations | TelVue Corporation
>         TelVue - We Share Your Vision
>         800.885.8886 x128 <tel:800.885.8886%20x128> |
>         msteele at telvue.com <mailto:msteele at telvue.com> |
>         http://www.telvue.com <http://www.telvue.com/>
>         twitter: http://twitter.com/telvue | facebook:
>         https://www.facebook.com/telvue
>
>         On Sun, Jul 12, 2015 at 9:03 AM, Roy Golan <rgolan at redhat.com
>         <mailto:rgolan at redhat.com>> wrote:
>
>             On 07/12/2015 03:52 PM, Mark Steele wrote:
>>             That command returns nothing - I don't think qemu is running?
>>
>>             Not sure how to start it on CentOS
>>
>>             [root at ovirt-01 ~]# ps -ef | grep qemu
>>
>>             root     23279 23130 0 08:51 pts/0    00:00:00 grep qemu
>>
>>
>             that mean you don't have vm running on that host. so you
>             can restart vdsm
>
>>
>>
>>             ***
>>             *Mark Steele*
>>             CIO / VP Technical Operations | TelVue Corporation
>>             TelVue - We Share Your Vision
>>             800.885.8886 <tel:800.885.8886> x128 | msteele at telvue.com
>>             <mailto:msteele at telvue.com> | http://www.telvue.com
>>             twitter: http://twitter.com/telvue | facebook:
>>             https://www.facebook.com/telvue
>>
>>             On Sun, Jul 12, 2015 at 8:45 AM, Roy Golan
>>             <rgolan at redhat.com <mailto:rgolan at redhat.com>> wrote:
>>
>>                 On 07/12/2015 03:42 PM, Mark Steele wrote:
>>>                 I run into the same issue - I am unable to
>>>                 completely go into maintenance mode because this VM
>>>                 is still on it - it cannot be migrated because it is
>>>                 not managed.
>>                 find you qemu process:
>>                   pgrep -an qemu-kvm | grep external
>>
>>                 and kill the process
>>
>>
>>>
>>>                 Inline image 1
>>>
>>>
>>>                 ***
>>>                 *Mark Steele*
>>>                 CIO / VP Technical Operations | TelVue Corporation
>>>                 TelVue - We Share Your Vision
>>>                 800.885.8886 <tel:800.885.8886> x128 |
>>>                 msteele at telvue.com <mailto:msteele at telvue.com> |
>>>                 http://www.telvue.com
>>>                 twitter: http://twitter.com/telvue | facebook:
>>>                 https://www.facebook.com/telvue
>>>
>>>                 On Sun, Jul 12, 2015 at 8:25 AM, Roy Golan
>>>                 <rgolan at redhat.com <mailto:rgolan at redhat.com>> wrote:
>>>
>>>                     On 07/12/2015 03:12 PM, Mark Steele wrote:
>>>>                     I think I may have found the problem:
>>>>
>>>>                     [root at ovirt-01 pki]# ls -lah
>>>>                     total 48K
>>>>                     drwxr-xr-x.  10 root  root  4.0K Nov 14  2014 .
>>>>                     drwxr-xr-x. 118 root  root   12K Jul 12 03:35 ..
>>>>                     drwxr-xr-x.   6 root  root  4.0K Nov 14  2014 CA
>>>>                     drwxr-xr-x.   4 root  root  4.0K Nov 14  2014
>>>>                     ca-trust
>>>>                     drwxr-xr-x.   2 root  root  4.0K Nov 14  2014 java
>>>>                     drwxr-xr-x.   2 root  root  4.0K Jul 12 07:03 nssdb
>>>>                     drwxr-xr-x.   6 ovirt ovirt 4.0K Nov 19  2014
>>>>                     ovirt-engine
>>>>                     drwxr-xr-x.   2 root  root  4.0K Nov 14  2014
>>>>                     rpm-gpg
>>>>                     drwx------.   2 root  root  4.0K Nov 22  2013
>>>>                     rsyslog
>>>>                     drwxr-xr-x.   5 root  root  4.0K Nov 14  2014 tls
>>>>                     [root at ovirt-01 pki]#
>>>>
>>>>                     There is no vsdm directory under /etc/pki
>>>>
>>>>                     This is an ovirt node. Version of software is
>>>>                     3.5.0.1-1.el6 from the ovirt management console.
>>>>
>>>>                     I'd like to add that I am not the person who
>>>>                     originally installed this instance - and am not
>>>>                     entirely familiar with how it is setup and
>>>>                     installed - so I may ask ignorant questions
>>>>                     from time to time.
>>>>
>>>
>>>                     not urgent but at this point it looks like it
>>>                     would be good to reinstall this host from the
>>>                     webadmin. if you have the capacity,
>>>                     you can put the host to maintenance, that will
>>>                     migrate vms to other hosts, and then choose
>>>                     "reinstall" once its in "maintenance'
>>>
>>>>
>>>>                     ***
>>>>                     *Mark Steele*
>>>>                     CIO / VP Technical Operations | TelVue Corporation
>>>>                     TelVue - We Share Your Vision
>>>>                     800.885.8886 <tel:800.885.8886> x128 |
>>>>                     msteele at telvue.com
>>>>                     <mailto:msteele at telvue.com> | http://www.telvue.com
>>>>                     twitter: http://twitter.com/telvue | facebook:
>>>>                     https://www.facebook.com/telvue
>>>>
>>>>                     On Sun, Jul 12, 2015 at 8:02 AM, Roy Golan
>>>>                     <rgolan at redhat.com <mailto:rgolan at redhat.com>>
>>>>                     wrote:
>>>>
>>>>                         On 07/12/2015 02:07 PM, Mark Steele wrote:
>>>>>                         Thank you Roy,
>>>>>
>>>>>                         I installed the client but am getting a
>>>>>                         permissions error when I run it
>>>>>
>>>>>                         [root at ovirt-01 ~]# vdsClient -s 0 list
>>>>>                         Traceback (most recent call last):
>>>>>                         File "/usr/share/vdsm/vdsClient.py", line
>>>>>                         2678, in <module>
>>>>>                         serv.do_connect(hostPort)
>>>>>                         File "/usr/share/vdsm/vdsClient.py", line
>>>>>                         136, in do_connect
>>>>>                           self.s = vdscli.connect(hostPort,
>>>>>                         self.useSSL, self.truststore)
>>>>>                         File
>>>>>                         "/usr/lib/python2.6/site-packages/vdsm/vdscli.py",
>>>>>                         line 110, in connect
>>>>>                           raise Exception("No permission to read
>>>>>                         file: %s" % f)
>>>>>                         Exception: No permission to read file:
>>>>>                         /etc/pki/vdsm/keys/vdsmkey.pem
>>>>>
>>>>
>>>>                         This should work. something isn't right
>>>>                         with your setup
>>>>                         is your host an ovirt-node? could be that
>>>>                         you hit [1] . let me know what version are
>>>>                         you running.
>>>>
>>>>                         please try the same with user vdsm. it
>>>>                         should have permissions to /etc/pki/vdsm
>>>>
>>>>                         [1] https://gerrit.ovirt.org/#/c/27779/
>>>>
>>>>
>>>>>                         If I restart vdsm, will that cause any
>>>>>                         issues with running VM's on this ovirt
>>>>>                         installation? This is our production
>>>>>                         environment.
>>>>
>>>>                         Generatlly the answer is no but lets avoid
>>>>                         it if  we can for this is a minor cosmetic
>>>>                         issue I guess.
>>>>
>>>>                         just as FYI - vdsm only reconnects to the
>>>>                         socket exposed by libvirt to control the vm
>>>>                         lifecycle. VDSM doesn't manadate the
>>>>                         lifecycle of a VM unless the engine tells
>>>>                         is so. Storage wise there could be some
>>>>                         operations but i'm almost sure they must
>>>>                         not have effect on running VMs.
>>>>>
>>>>>                         Thank you
>>>>>
>>>>>
>>>>>                         ***
>>>>>                         *Mark Steele*
>>>>>                         CIO / VP Technical Operations | TelVue
>>>>>                         Corporation
>>>>>                         TelVue - We Share Your Vision
>>>>>                         800.885.8886 <tel:800.885.8886> x128 |
>>>>>                         msteele at telvue.com
>>>>>                         <mailto:msteele at telvue.com> |
>>>>>                         http://www.telvue.com
>>>>>                         twitter: http://twitter.com/telvue |
>>>>>                         facebook: https://www.facebook.com/telvue
>>>>>
>>>>>                         On Sun, Jul 12, 2015 at 4:09 AM, Roy Golan
>>>>>                         <rgolan at redhat.com
>>>>>                         <mailto:rgolan at redhat.com>> wrote:
>>>>>
>>>>>                             On 07/09/2015 06:34 PM, Mark Steele wrote:
>>>>>>                             Yes,
>>>>>>
>>>>>>                             It is displayed in the engine:
>>>>>>
>>>>>>                             Inline image 1
>>>>>
>>>>>                             the vdsm on that host reports it back
>>>>>                             to the engine . since this vm isn't in
>>>>>                             the engine DB it is concidered as
>>>>>                             EXTERNAL (thus the error 400 from the API)
>>>>>
>>>>>                             do yo know if the qemu-kvm proccess
>>>>>                             isn't running anymore?
>>>>>
>>>>>                             if the process isn't running then vdsm
>>>>>                             must clean its cache
>>>>>
>>>>>                             try to:
>>>>>
>>>>>                             yum install vdsm-cli
>>>>>                             vdsClient -s 0 list
>>>>>                             vdsClient -s 0 destroy {vmId}
>>>>>
>>>>>                             alternativly a vdsm restart will work
>>>>>                             (if the qemu proccess isn't running)
>>>>>
>>>>>>
>>>>>>                             The VM is not really running - the IP
>>>>>>                             addresses that are being reported are
>>>>>>                             from another VM that was recently
>>>>>>                             removed. All attempts to control the
>>>>>>                             VM have failed. It does not have any
>>>>>>                             NICS or disk associated with it - so
>>>>>>                             this seems to be a ghost in the
>>>>>>                             machine. I attempted to unlock it
>>>>>>                             using the unlock_entity.sh script -
>>>>>>                             it reports successful, however I
>>>>>>                             still cannot do anything with the VM.
>>>>>>
>>>>>>
>>>>>>                             ***
>>>>>>                             *Mark Steele*
>>>>>>                             CIO / VP Technical Operations |
>>>>>>                             TelVue Corporation
>>>>>>                             TelVue - We Share Your Vision
>>>>>>                             800.885.8886 <tel:800.885.8886> x128
>>>>>>                             | msteele at telvue.com
>>>>>>                             <mailto:msteele at telvue.com> |
>>>>>>                             http://www.telvue.com
>>>>>>                             twitter: http://twitter.com/telvue |
>>>>>>                             facebook: https://www.facebook.com/telvue
>>>>>>
>>>>>>                             On Thu, Jul 9, 2015 at 11:11 AM,
>>>>>>                             Artyom Lukianov <alukiano at redhat.com
>>>>>>                             <mailto:alukiano at redhat.com>> wrote:
>>>>>>
>>>>>>                                 Can you sea via engine, on what
>>>>>>                                 host run VM?
>>>>>>                                 Anyway if you have really run VM
>>>>>>                                 on host you can try to figure it
>>>>>>                                 with 'ps aux | grep qemu', if it
>>>>>>                                 will return you some process, you
>>>>>>                                 can just kill process via 'kill pid'.
>>>>>>                                 I hope it will help you.
>>>>>>
>>>>>>                                 ----- Original Message -----
>>>>>>                                 From: "Mark Steele"
>>>>>>                                 <msteele at telvue.com
>>>>>>                                 <mailto:msteele at telvue.com>>
>>>>>>                                 To: "Artyom Lukianov"
>>>>>>                                 <alukiano at redhat.com
>>>>>>                                 <mailto:alukiano at redhat.com>>
>>>>>>                                 Cc: users at ovirt.org
>>>>>>                                 <mailto:users at ovirt.org>
>>>>>>                                 Sent: Thursday, July 9, 2015
>>>>>>                                 5:42:20 PM
>>>>>>                                 Subject: Re: [ovirt-users] This
>>>>>>                                 VM is not managed by the engine
>>>>>>
>>>>>>                                 Artyom,
>>>>>>
>>>>>>                                 Thank you - I don't have
>>>>>>                                 vdsClient installed - can you
>>>>>>                                 point me to the
>>>>>>                                 download?
>>>>>>
>>>>>>
>>>>>>                                 ***
>>>>>>                                 *Mark Steele*
>>>>>>                                 CIO / VP Technical Operations |
>>>>>>                                 TelVue Corporation
>>>>>>                                 TelVue - We Share Your Vision
>>>>>>                                 800.885.8886 x128
>>>>>>                                 <tel:800.885.8886%20x128> |
>>>>>>                                 msteele at telvue.com
>>>>>>                                 <mailto:msteele at telvue.com> |
>>>>>>                                 http://www.telvue.com
>>>>>>                                 twitter:
>>>>>>                                 http://twitter.com/telvue | facebook:
>>>>>>                                 https://www.facebook.com/telvue
>>>>>>
>>>>>>                                 On Thu, Jul 9, 2015 at 10:11 AM,
>>>>>>                                 Artyom Lukianov
>>>>>>                                 <alukiano at redhat.com
>>>>>>                                 <mailto:alukiano at redhat.com>>
>>>>>>                                 wrote:
>>>>>>
>>>>>>                                 > Please check host where VM
>>>>>>                                 run(vdsClient -s 0 list table),
>>>>>>                                 and you can
>>>>>>                                 > destroy it via
>>>>>>                                 vdsClient(vdsClient -s 0 destroy
>>>>>>                                 vm_id).
>>>>>>                                 > Thanks
>>>>>>                                 >
>>>>>>                                 > ----- Original Message -----
>>>>>>                                 > From: "Mark Steele"
>>>>>>                                 <msteele at telvue.com
>>>>>>                                 <mailto:msteele at telvue.com>>
>>>>>>                                 > To: users at ovirt.org
>>>>>>                                 <mailto:users at ovirt.org>
>>>>>>                                 > Sent: Thursday, July 9, 2015
>>>>>>                                 4:38:32 PM
>>>>>>                                 > Subject: [ovirt-users] This VM
>>>>>>                                 is not managed by the engine
>>>>>>                                 >
>>>>>>                                 > I have a VM that was not
>>>>>>                                 started and is now showing as
>>>>>>                                 running. When I
>>>>>>                                 > attempt to suspend or stop it
>>>>>>                                 in the ovirt-shell, I get the
>>>>>>                                 message:
>>>>>>                                 >
>>>>>>                                 > status: 400
>>>>>>                                 > reason: bad request
>>>>>>                                 > detail: Cannot hibernate VM.
>>>>>>                                 This VM is not managed by the engine.
>>>>>>                                 >
>>>>>>                                 > Not sure how the VM was
>>>>>>                                 initially created on the ovirt
>>>>>>                                 manager. This VM is
>>>>>>                                 > not needed - how can I
>>>>>>                                 'shutdown' and remove this VM?
>>>>>>                                 >
>>>>>>                                 > Thanks
>>>>>>                                 >
>>>>>>                                 > ***
>>>>>>                                 > Mark Steele
>>>>>>                                 > CIO / VP Technical Operations |
>>>>>>                                 TelVue Corporation
>>>>>>                                 > TelVue - We Share Your Vision
>>>>>>                                 > 800.885.8886 x128
>>>>>>                                 <tel:800.885.8886%20x128> |
>>>>>>                                 msteele at telvue.com
>>>>>>                                 <mailto:msteele at telvue.com> |
>>>>>>                                 http://www.telvue.com
>>>>>>                                 > twitter:
>>>>>>                                 http://twitter.com/telvue | facebook:
>>>>>>                                 > https://www.facebook.com/telvue
>>>>>>                                 >
>>>>>>                                 >
>>>>>>                                 _______________________________________________
>>>>>>                                 > Users mailing list
>>>>>>                                 > Users at ovirt.org
>>>>>>                                 <mailto:Users at ovirt.org>
>>>>>>                                 >
>>>>>>                                 http://lists.ovirt.org/mailman/listinfo/users
>>>>>>                                 >
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>                             _______________________________________________
>>>>>>                             Users mailing list
>>>>>>                             Users at ovirt.org <mailto:Users at ovirt.org>
>>>>>>                             http://lists.ovirt.org/mailman/listinfo/users
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>
>>
>
>
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150712/61e40666/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: image/png
Size: 166519 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150712/61e40666/attachment-0002.png>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: image/png
Size: 25389 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150712/61e40666/attachment-0003.png>


More information about the Users mailing list