On Thu, Feb 6, 2020 at 1:07 PM Jorick Astrego <jorick@netbulae.eu> wrote:
Here you go, this is from the activation I just did a couple of minutes ago.
I was hoping to see how it was first connected to host, but it doesn't go that far back. Anyway, the storage domain type is set from engine and vdsm never try to guess it as far as I saw.
I put the host in maintenance and activated it again, this should
give you some more info. See attached log.
Could you query the engine db about the misbehaving domain and paste the results?
# su - postgres
Last login: Thu Feb 6 07:17:52 EST 2020 on pts/0
-bash-4.2$ LD_LIBRARY_PATH=/opt/rh/rh-postgresql10/root/lib64/ /opt/rh/rh-postgresql10/root/usr/bin/psql enginepsql (10.6)
Type "help" for help.
engine=# select * from storage_domain_static where id = 'f5d2f7c6-093f-46d6-a844-224d92db5ef9' ;
engine=# select * from storage_domain_static where id = 'f5d2f7c6-093f-46d6-a844-224d92db5ef9' ;
id | storage | storage_name | storage_domain_type | storage_type | storage_domain_format_type | _create_date | _update_date | recoverable | la
st_time_used_as_master | storage_description | storage_comment | wipe_after_delete | warning_low_space_indicator | critical_space_action_blocker | first_metadata_device | vg_metadata_device | discard_after_delete | backup | warning_low_co
nfirmed_space_indicator | block_size
--------------------------------------+--------------------------------------+--------------+---------------------+--------------+----------------------------+-------------------------------+-----------------------------+-------------+---
-----------------------+---------------------+-----------------+-------------------+-----------------------------+-------------------------------+-----------------------+--------------------+----------------------+--------+---------------
------------------------+------------
f5d2f7c6-093f-46d6-a844-224d92db5ef9 | b8b456f0-27c3-49b9-b5e9-9fa81fb3cdaa | backupnfs | 1 | 1 | 4 | 2018-01-19 13:31:25.899738+01 | 2019-02-14 14:36:22.3171+01 | t |
1530772724454 | | | f | 10 | 5 | | | f | f |
0 | 512
(1 row)
Regards,
Jorick
On 2/6/20 11:23 AM, Amit Bawer wrote:
On Thu, Feb 6, 2020 at 11:07 AM Jorick Astrego <jorick@netbulae.eu> wrote:
Hi,
Something weird is going on with our ovirt node 4.3.8 install mounting a nfs share.
We have a NFS domain for a couple of backup disks and we have a couple of 4.2 nodes connected to it.
Now I'm adding a fresh cluster of 4.3.8 nodes and the backupnfs mount doesn't work.
(annoying you cannot copy the text from the events view)
The domain is up and working
ID:f5d2f7c6-093f-46d6-a844-224d92db5ef9Size: 10238 GiBAvailable:2491 GiBUsed:7747 GiBAllocated: 3302 GiBOver Allocation Ratio:37%Images:7Path:*.*.*.*:/data/ovirtNFS Version: AUTOWarning Low Space Indicator:10% (1023 GiB)Critical Space Action Blocker:5 GiB
But somehow the node appears to thin thinks it's an LVM volume? It tries to find the VGs volume group but fails... which is not so strange as it is an NFS volume:
Could you provide full vdsm.log file with this flow?
_______________________________________________2020-02-05 14:17:54,190+0000 WARN (monitor/f5d2f7c) [storage.LVM] Reloading VGs failed (vgs=[u'f5d2f7c6-093f-46d6-a844-224d92db5ef9'] rc=5 out=[] err=[' Volume group "f5d2f7c6-093f-46d6-a844-224d92db5ef9" not found', ' Cannot process volume group f5d2f7c6-093f-46d6-a844-224d92db5ef9']) (lvm:470)
2020-02-05 14:17:54,201+0000 ERROR (monitor/f5d2f7c) [storage.Monitor] Setting up monitor for f5d2f7c6-093f-46d6-a844-224d92db5ef9 failed (monitor:330)
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", line 327, in _setupLoop
self._setupMonitor()
File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", line 349, in _setupMonitor
self._produceDomain()
File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 159, in wrapper
value = meth(self, *a, **kw)
File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", line 367, in _produceDomain
self.domain = sdCache.produce(self.sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 110, in produce
domain.getRealDomain()
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 51, in getRealDomain
return self._cache._realProduce(self._sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 134, in _realProduce
domain = self._findDomain(sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 151, in _findDomain
return findMethod(sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 176, in _findUnfetchedDomain
raise se.StorageDomainDoesNotExist(sdUUID)
StorageDomainDoesNotExist: Storage domain does not exist: (u'f5d2f7c6-093f-46d6-a844-224d92db5ef9',)
The volume is actually mounted fine on the node:
On NFS server
Feb 5 15:47:09 back1en rpc.mountd[4899]: authenticated mount request from *.*.*.*:673 for /data/ovirt (/data/ovirt)
On the host
mount|grep nfs
*.*.*.*:/data/ovirt on /rhev/data-center/mnt/*.*.*.*:_data_ovirt type nfs (rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,soft,nolock,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mountaddr=*.*.*.*,mountvers=3,mountport=20048,mountproto=udp,local_lock=all,addr=*.*.*.*)
And I can see the files:ls -alrt /rhev/data-center/mnt/*.*.*.*:_data_ovirt
total 4
drwxr-xr-x. 5 vdsm kvm 61 Oct 26 2016 1ed0a635-67ee-4255-aad9-b70822350706
-rwxr-xr-x. 1 vdsm kvm 0 Feb 5 14:37 __DIRECT_IO_TEST__
drwxrwxrwx. 3 root root 86 Feb 5 14:37 .
drwxr-xr-x. 5 vdsm kvm 4096 Feb 5 14:37 ..
Met vriendelijke groet, With kind regards,
Jorick Astrego
Netbulae Virtualization Experts
Tel: 053 20 30 270 info@netbulae.eu Staalsteden 4-3A KvK 08198180 Fax: 053 20 30 271 www.netbulae.eu 7547 TA Enschede BTW NL821234584B01
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/IFTO5WBLVLGTVWKYN3BGLOHAC453UBD5/
Met vriendelijke groet, With kind regards,
Jorick Astrego
Netbulae Virtualization Experts
Tel: 053 20 30 270 info@netbulae.eu Staalsteden 4-3A KvK 08198180 Fax: 053 20 30 271 www.netbulae.eu 7547 TA Enschede BTW NL821234584B01
Tel: 053 20 30 270 | info@netbulae.eu | Staalsteden 4-3A | KvK 08198180 |
Fax: 053 20 30 271 | www.netbulae.eu | 7547 TA Enschede | BTW NL821234584B01 |