Hi,
it's grayed out on web interface, is there any other way? Trying to
detach gives error
VDSM command DetachStorageDomainVDS failed: Storage domain does not
exist: (u'c44343af-cc4a-4bb7-a548-0c6f609d60d5',)
Failed to detach Storage Domain hosted_storage from Data Center
Default. (User: admin@internal-authz)
Best regards,
Misak Khachatryan
On Thu, Aug 31, 2017 at 4:22 PM, Martin Sivak <msivak(a)redhat.com> wrote:
Hi,
you can remote the hosted engine storage domain from the engine as
well. It should also be re-imported.
We had cases where destroying the domain ended up with a locked SD,
but removing the SD and re-importing is the proper way here.
Best regards
PS: Re-adding the mailing list, we should really set a proper Reply-To header..
Martin Sivak
On Thu, Aug 31, 2017 at 2:07 PM, Misak Khachatryan <kmisak(a)gmail.com> wrote:
> Hi,
>
> I would love to, but:
>
> Error while executing action:
>
> HostedEngine:
>
> Cannot remove VM. The relevant Storage Domain's status is Inactive.
>
> it seems i should somehow fix storage domain first ...
>
> engine=# update storage_domain_static set id =
> '2e2820f3-8c3d-487d-9a56-1b8cd278ec6c' where id =
> 'c44343af-cc4a-4bb7-a548-0c6f609d60d5';
> ERROR: update or delete on table "storage_domain_static" violates
> foreign key constraint "disk_profiles_storage_domain_id_fkey" on table
> "disk_profiles"
> DETAIL: Key (id)=(c44343af-cc4a-4bb7-a548-0c6f609d60d5) is still
> referenced from table "disk_profiles".
>
> engine=# update disk_profiles set storage_domain_id =
> '2e2820f3-8c3d-487d-9a56-1b8cd278ec6c' where id =
> 'a6d71571-a13a-415b-9f97-635f17cbe67d';
> ERROR: insert or update on table "disk_profiles" violates foreign key
> constraint "disk_profiles_storage_domain_id_fkey"
> DETAIL: Key (storage_domain_id)=(2e2820f3-8c3d-487d-9a56-1b8cd278ec6c)
> is not present in table "storage_domain_static".
>
> engine=# select * from storage_domain_static;
> id | storage
> | storage_name | storage_domain_type | storage_type |
> storage_domain_format_type | _create_date |
> _update_date | recoverable | last_time_used_as_maste
> r | storage_description | storage_comment | wipe_after_delete |
> warning_low_space_indicator | critical_space_action_blocker |
> first_metadata_device | vg_metadata_device | discard_after_delete
>
--------------------------------------+--------------------------------------+------------------------+---------------------+--------------+----------------------------+-------------------------------+-------------------------------+-------------+------------------------
>
--+---------------------+-----------------+-------------------+-----------------------------+-------------------------------+-----------------------+--------------------+----------------------
> 072fbaa1-08f3-4a40-9f34-a5ca22dd1d74 |
> ceab03af-7220-4d42-8f5c-9b557f5d29af | ovirt-image-repository |
> 4 | 8 | 0 | 2016-11-02
> 21:27:22.118586+04 | | t |
> | | | f |
> | |
> | | f
> 51c903f6-df83-4510-ac69-c164742ca6e7 |
> 34b72ce0-6ad7-4180-a8a1-2acfd45824d7 | iso |
> 2 | 7 | 0 | 2016-11-02
> 23:26:21.296635+04 | | t |
> 0 | | | f |
> 10 | 5 |
> | | f
> ece1f05c-97c9-4482-a1a5-914397cddd35 |
> dd38f31f-7bdc-463c-9ae4-fcd4dc8c99fd | export |
> 3 | 1 | 0 | 2016-12-14
> 11:28:15.736746+04 | 2016-12-14 11:33:12.872562+04 | t |
> 0 | Export | | f |
> 10 | 5 |
> | | f
> 07ea2089-a82b-4ca1-9c8b-54e3895b2ed4 |
> d1e9e3c8-aaf3-43de-ae80-101e5bd2574f | data |
> 0 | 7 | 4 | 2016-11-02
> 23:24:43.402629+04 | 2017-02-22 17:20:42.721092+04 | t |
> 0 | | | f |
> 10 | 5 |
> | | f
> c44343af-cc4a-4bb7-a548-0c6f609d60d5 |
> 8b54ce35-3187-4fba-a2c7-6b604d077f5b | hosted_storage |
> 1 | 7 | 4 | 2016-11-02
> 23:26:13.165435+04 | 2017-02-22 17:20:42.721092+04 | t |
> 0 | | | f |
> 10 | 5 |
> | | f
> 004ca4dd-c621-463d-b514-ccfe07ef99d7 |
> b31a7de9-e789-4ece-9f99-4b150bf581db | virt4-Local |
> 0 | 4 | 4 | 2017-03-23
> 09:02:26.37006+04 | 2017-03-23 09:02:31.887534+04 | t |
> 0 | | | f |
> 10 | 5 |
> | | f
> (6 rows)
>
> engine=# select * from storage_domain_dynamic;
> id | available_disk_size |
> used_disk_size | _update_date | external_status
>
--------------------------------------+---------------------+----------------+-------------------------------+-----------------
> 072fbaa1-08f3-4a40-9f34-a5ca22dd1d74 | |
> | | 0
> 07ea2089-a82b-4ca1-9c8b-54e3895b2ed4 | 1102 |
> 313 | 2017-08-31 14:20:47.444292+04 | 0
> 51c903f6-df83-4510-ac69-c164742ca6e7 | 499 |
> 0 | 2017-08-31 14:20:47.45047+04 | 0
> ece1f05c-97c9-4482-a1a5-914397cddd35 | 9669 |
> 6005 | 2017-08-31 14:20:47.454629+04 | 0
> c44343af-cc4a-4bb7-a548-0c6f609d60d5 | |
> | 2017-08-31 14:18:37.199062+04 | 0
> 004ca4dd-c621-463d-b514-ccfe07ef99d7 | 348 |
> 1 | 2017-08-31 14:20:42.671688+04 | 0
> (6 rows)
>
>
> engine=# select * from disk_profiles;
> id | name |
> storage_domain_id | qos_id | description |
> _create_date | _update_date
>
--------------------------------------+----------------+--------------------------------------+--------+-------------+-------------------------------+--------------
> 04257bff-e95d-4380-b120-adcbe46ae213 | data |
> 07ea2089-a82b-4ca1-9c8b-54e3895b2ed4 | | |
> 2016-11-02 23:24:43.528982+04 |
> a6d71571-a13a-415b-9f97-635f17cbe67d | hosted_storage |
> c44343af-cc4a-4bb7-a548-0c6f609d60d5 | | |
> 2016-11-02 23:26:13.178791+04 |
> 0f9ecdb7-4fca-45e7-9b5c-971b50d4c12e | virt4-Local |
> 004ca4dd-c621-463d-b514-ccfe07ef99d7 | | |
> 2017-03-23 09:02:26.409574+04 |
> (3 rows)
>
>
> Best regards,
> Misak Khachatryan
>
>
> On Thu, Aug 31, 2017 at 3:33 PM, Martin Sivak <msivak(a)redhat.com> wrote:
>> Hi,
>>
>> I would not touch the database in this case. I would just delete the
>> old hosted engine VM from the webadmin and wait for it to reimport
>> itself.
>>
>> But I haven't played with this mechanism for some time.
>>
>> Best regards
>>
>> Martin Sivak
>>
>> On Thu, Aug 31, 2017 at 1:17 PM, Misak Khachatryan <kmisak(a)gmail.com>
wrote:
>>> Hi,
>>>
>>> Yesterday someone powered off our storage, and all my 3 hosts lose
>>> their disks. After 2 days of recovering i managed to bring back
>>> everything, except engine VM, which is online but not visible to
>>> itself.
>>>
>>> I did new deployment of VM, restored backup and started engine setup.
>>> After manual database updates, my all VMs and hosts are OK now, but
>>> engine. I have engine VM with different VM id running than in
>>> database.
>>>
>>> I've tried this with no luck.
>>>
>>> engine=# update vm_static set vm_guid =
>>> '75072b32-6f93-4c38-8f18-825004072c1a' where vm_guid =(select
>>> vm_guid from vm_static where vm_name = 'HostedEngine');
>>> ERROR: update or delete on table "vm_static" violates foreign key
>>> constraint "fk_disk_vm_element_vm_static" on table
"disk_vm_element"
>>> DETAIL: Key (vm_guid)=(d81ccb53-2594-49db-b69a-04c73b504c59) is still
>>> referenced from table "disk_vm_element".
>>>
>>>
>>> Right now i've deployed engine on all 3 hosts but see this picture:
>>>
>>> [root@virt3 ~]# hosted-engine --vm-status
>>>
>>>
>>> !! Cluster is in GLOBAL MAINTENANCE mode !!
>>>
>>>
>>>
>>>
>>> [root@virt3 ~]# vdsClient -s 0 list
>>>
>>> 75072b32-6f93-4c38-8f18-825004072c1a
>>> Status = Up
>>> statusTime = 4397337690
>>> kvmEnable = true
>>> emulatedMachine = pc
>>> afterMigrationStatus =
>>> pid = 5280
>>> devices = [{'device': 'console', 'specParams':
{}, 'type':
>>> 'console', 'deviceId':
'2b6b0e87-c86a-4144-ad39-40d5bfe25df1',
>>> 'alias': 'console0'}, {'device':
'memballoon', 'specParams': {'model':
>>> 'none'}, 'type': 'balloon', 'target':
16777216, 'alias': 'balloon0'},
>>> {'specParams': {'source': 'random'}, 'alias':
'rng0', 'address':
>>> {'slot': '0x07', 'bus': '0x00',
'domain': '0x0000', 'type': 'pci',
>>> 'function': '0x0'}, 'device': 'virtio',
'model': 'virtio', 'type':
>>> 'rng'}, {'device': 'unix', 'alias':
'channel0', 'type': 'channel',
>>> 'addr
>>> ess': {'bus': '0', 'controller': '0',
'type': 'virtio-serial', 'port':
>>> '1'}}, {'device': 'unix', 'alias':
'channel1', 'type': 'channel',
>>> 'address': {'bus': '0', 'controller':
'0', 'type': 'virtio-serial',
>>> 'port': '2'}}, {'device': 'unix',
'alias': 'channel2', 'type': 'ch
>>> annel', 'address': {'bus': '0',
'controller': '0', 'type':
>>> 'virtio-serial', 'port': '3'}}, {'device':
'scsi', 'alias': 'scsi0',
>>> 'model': 'virtio-scsi', 'type': 'controller',
'address': {'slot':
>>> '0x04', 'bus': '0x00', 'domain':
'0x0000', 'type': 'pci', 'function':
>>> '0x0'}}
>>> , {'device': 'usb', 'alias': 'usb',
'type': 'controller', 'address':
>>> {'slot': '0x01', 'bus': '0x00',
'domain': '0x0000', 'type': 'pci',
>>> 'function': '0x2'}}, {'device': 'ide',
'alias': 'ide', 'type':
>>> 'controller', 'address': {'slot': '0x01',
'bus': '0x00', 'domain':
>>> '0x00
>>> 00', 'type': 'pci', 'function': '0x1'}},
{'device': 'virtio-serial',
>>> 'alias': 'virtio-serial0', 'type':
'controller', 'address': {'slot':
>>> '0x05', 'bus': '0x00', 'domain':
'0x0000', 'type': 'pci', 'function':
>>> '0x0'}}, {'device': 'vga', 'alias':
'video0', 'type': 'video',
>>> 'address': {'slot': '0x02', 'bus':
'0x00', 'domain': '0x0000', 'type':
>>> 'pci', 'function': '0x0'}}, {'device':
'vnc', 'type': 'graphics',
>>> 'port': '5900'}, {'nicModel': 'pv',
'macAddr': '00:16:3e:01:29:95',
>>> 'linkActive': True, 'network': 'ovirtmgmt',
'alias': 'net0', 'spec
>>> Params': {}, 'deviceId':
'd348a068-063b-4a40-9119-a3d34f6c7db4',
>>> 'address': {'slot': '0x03', 'bus':
'0x00', 'domain': '0x0000', 'type':
>>> 'pci', 'function': '0x0'}, 'device':
'bridge', 'type': 'interface',
>>> 'name': 'vnet0'}, {'index': '2',
'iface': 'ide', 'name': 'hdc', 'al
>>> ias': 'ide0-1-0', 'specParams': {}, 'readonly':
'True', 'deviceId':
>>> 'e738b50b-c200-4429-8489-4519325339c7', 'address':
{'bus': '1',
>>> 'controller': '0', 'type': 'drive',
'target': '0', 'unit': '0'},
>>> 'device': 'cdrom', 'shared': 'false',
'path': '', 'type': 'disk'},
>>> {'poolI
>>> D': '00000000-0000-0000-0000-000000000000', 'volumeInfo':
{'path':
>>>
'engine/2e2820f3-8c3d-487d-9a56-1b8cd278ec6c/images/5deeac2d-18d7-4622-9371-ebf965d2bd6b/60aa51b7-32eb-41a9-940d-9489b0375a3d',
>>> 'protocol': 'gluster', 'hosts': [{'port':
'0', 'transport': 'tcp',
>>> 'name': '
>>> virt1'}, {'port': '0', 'transport':
'tcp', 'name': 'virt2'}, {'port':
>>> '0', 'transport': 'tcp', 'name':
'virt3'}]}, 'index': '0', 'iface':
>>> 'virtio', 'apparentsize': '62277025792',
'specParams': {}, 'imageID':
>>> '5deeac2d-18d7-4622-9371-ebf965d2bd6b', 'readonly':
'False', 's
>>> hared': 'exclusive', 'truesize': '3255476224',
'type': 'disk',
>>> 'domainID': '2e2820f3-8c3d-487d-9a56-1b8cd278ec6c',
'reqsize': '0',
>>> 'format': 'raw', 'deviceId':
'5deeac2d-18d7-4622-9371-ebf965d2bd6b',
>>> 'address': {'slot': '0x06', 'bus':
'0x00', 'domain': '0x0000', 'type':
>>> 'pci', 'function': '0x0'}, 'device':
'disk', 'path':
>>>
'/var/run/vdsm/storage/2e2820f3-8c3d-487d-9a56-1b8cd278ec6c/5deeac2d-18d7-4622-9371-ebf965d2bd6b/60aa51b7-32eb-41a9-940d-9489b0375a3d',
>>> 'propagateErrors': 'off', 'optional':
'false', 'name': 'vda',
>>> 'bootOrder': '1', 'v
>>> olumeID': '60aa51b7-32eb-41a9-940d-9489b0375a3d',
'alias':
>>> 'virtio-disk0', 'volumeChain': [{'domainID':
>>> '2e2820f3-8c3d-487d-9a56-1b8cd278ec6c', 'leaseOffset': 0,
'volumeID':
>>> '60aa51b7-32eb-41a9-940d-9489b0375a3d', 'leasePath':
>>> '/rhev/data-center/mnt/glusterSD/virt1:_engi
>>>
ne/2e2820f3-8c3d-487d-9a56-1b8cd278ec6c/images/5deeac2d-18d7-4622-9371-ebf965d2bd6b/60aa51b7-32eb-41a9-940d-9489b0375a3d.lease',
>>> 'imageID': '5deeac2d-18d7-4622-9371-ebf965d2bd6b',
'path':
>>>
'/rhev/data-center/mnt/glusterSD/virt1:_engine/2e2820f3-8c3d-487d-9a56-1b8cd278ec6c
>>>
/images/5deeac2d-18d7-4622-9371-ebf965d2bd6b/60aa51b7-32eb-41a9-940d-9489b0375a3d'}]}]
>>> guestDiskMapping = {'5deeac2d-18d7-4622-9': {'name':
>>> '/dev/vda'}, 'QEMU_DVD-ROM_QM00003': {'name':
'/dev/sr0'}}
>>> vmType = kvm
>>> display = vnc
>>> memSize = 16384
>>> cpuType = Westmere
>>> spiceSecureChannels =
>>> smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir
>>> smp = 4
>>> vmName = HostedEngine
>>> clientIp =
>>> maxVCpus = 16
>>> [root@virt3 ~]#
>>>
>>> [root@virt3 ~]# hosted-engine --vm-status
>>>
>>>
>>> !! Cluster is in GLOBAL MAINTENANCE mode !!
>>>
>>>
>>>
>>> --== Host 1 status ==--
>>>
>>> conf_on_shared_storage : True
>>> Status up-to-date : True
>>> Hostname : virt1.management.gnc.am
>>> Host ID : 1
>>> Engine status : {"reason": "vm not
running on
>>> this host", "health": "bad", "vm":
"down", "detail": "unknown"}
>>> Score : 3400
>>> stopped : False
>>> Local maintenance : False
>>> crc32 : ef49e5b4
>>> local_conf_timestamp : 7515
>>> Host timestamp : 7512
>>> Extra metadata (valid at timestamp):
>>> metadata_parse_version=1
>>> metadata_feature_version=1
>>> timestamp=7512 (Thu Aug 31 15:14:59 2017)
>>> host-id=1
>>> score=3400
>>> vm_conf_refresh_time=7515 (Thu Aug 31 15:15:01 2017)
>>> conf_on_shared_storage=True
>>> maintenance=False
>>> state=GlobalMaintenance
>>> stopped=False
>>>
>>>
>>> --== Host 3 status ==--
>>>
>>> conf_on_shared_storage : True
>>> Status up-to-date : True
>>> Hostname : virt3
>>> Host ID : 3
>>> Engine status : {"health": "good",
"vm": "up",
>>> "detail": "up"}
>>> Score : 3400
>>> stopped : False
>>> Local maintenance : False
>>> crc32 : 4a85111c
>>> local_conf_timestamp : 102896
>>> Host timestamp : 102893
>>> Extra metadata (valid at timestamp):
>>> metadata_parse_version=1
>>> metadata_feature_version=1
>>> timestamp=102893 (Thu Aug 31 15:14:46 2017)
>>> host-id=3
>>> score=3400
>>> vm_conf_refresh_time=102896 (Thu Aug 31 15:14:49 2017)
>>> conf_on_shared_storage=True
>>> maintenance=False
>>> state=GlobalMaintenance
>>> stopped=False
>>>
>>>
>>> !! Cluster is in GLOBAL MAINTENANCE mode !!
>>>
>>> Also my storage domain for hosted engine is inactive, can't activate
>>> it it gives this error in web console:
>>>
>>> VDSM command GetImagesListVDS failed: Storage domain does not exist:
>>> (u'c44343af-cc4a-4bb7-a548-0c6f609d60d5',)
>>>
>>>
>>> It seems I should fiddle with database a bit more, but is't scary thing
for me.
>>>
>>> Any help?
>>>
>>>
>>>
>>> Best regards,
>>> Misak Khachatryan
>>> _______________________________________________
>>> Users mailing list
>>> Users(a)ovirt.org
>>>
http://lists.ovirt.org/mailman/listinfo/users