# mount NetApp_NFS:/path/to/export /mnt
# chown -R 36.36 /mnt
# chmod -R 755 /mnt
# umount /mnt
Which is quite similar to the steps you've mentioned, except the last step of unmounting:Hi Shani,
Thank you for your reply, butHow do I do that?Reason why I am asking is following:Hosts 2,3,4 do not have that issue. Host 1 and 5 do.What I learned previously is that when using Netapp based NFS, which we are, it’s required to before provisioning SHE and/or just adding a Host to a pool, it’s required to execute following steps:
Create random dir on a host:- mkdir /mnt/rhevstoreMount netapp volume to the dir- mount -o sec=sys -t nfs 10.214.13.64:/ovirt_production /mnt/rhevstoreSet ownership to vdsm:kvm (36:36):- chown -R vdsm:kvm /mnt/rhevstore/*Unmount the 10.214.13.64:/ovirt_production
I do not expect the above ownership actions need to be done initially on each host, before starting the deployment, otherwise it would be practically impossible to expand the Host pool.
All 5 hosts are provisioned in same way. How? I am using foreman to provision these servers, so they are built of same kickstart hostgroup template.
I even installed ovirt-hosted-engine-setup package to make sure all required packages, users and groups are in place before adding host to oVirt via UI or Ansible.
Is it possible that we if I am already using or heavily using the mentioned volume via Hosts already added to oVirt pool, that ownership actions executed,on host about to be added to the pool, will fail to complete setting ownership on all required files on the volume?
To repeat the question above: How do I make sure Host can read metadata file of the storage volume?
Kindly awaiting your reply.
All best,Marko VrgoticSent from my iPhoneHi Marko,Is seems that there's a connectivity problem with host 10.210.13.64.Can you please make sure the metadata under /rhev/data-center/mnt/10.210.13.64:_ovirt__production/6effda5e-1a0d-4312-bf93-d97fa9eb5aee/dom_md/metadata is accessible?
Regards,Shani Leviim
On Sat, Aug 10, 2019 at 2:57 AM Vrgotic, Marko <M.Vrgotic@activevideo.com> wrote:
Log files from ovirt engine and ovirt-sj-05 vdsm attached.
Its related to host named: ovirt-sj-05.ictv.com
Kindly awaiting your reply.
— — —
Met vriendelijke groet / Kind regards,
Marko Vrgotic
From: "Vrgotic, Marko" <M.Vrgotic@activevideo.com>
Date: Thursday, 8 August 2019 at 17:02
To: Shani Leviim <sleviim@redhat.com>
Cc: "users@ovirt.org" <users@ovirt.org>
Subject: Re: [ovirt-users] Re: oVirt 4.3.5 potential issue with NFS storage
Hey Shanii,
Thank you for the reply.
Sure, I will attach the full logs asap.
What do you mean by “flow you are doing”?
Kindly awaiting your reply.
Marko Vrgotic
From: Shani Leviim <sleviim@redhat.com>
Date: Thursday, 8 August 2019 at 00:01
To: "Vrgotic, Marko" <M.Vrgotic@activevideo.com>
Cc: "users@ovirt.org" <users@ovirt.org>
Subject: Re: [ovirt-users] Re: oVirt 4.3.5 potential issue with NFS storage
Hi,
Can you please clarify the flow you're doing?
Also, can you please attach full vdsm and engine logs?
Regards,
Shani Leviim
On Thu, Aug 8, 2019 at 6:25 AM Vrgotic, Marko <M.Vrgotic@activevideo.com> wrote:
Log line form VDSM:
“[root@ovirt-sj-05 ~]# tail -f /var/log/vdsm/vdsm.log | grep WARN
2019-08-07 09:40:03,556-0700 WARN (check/loop) [storage.check] Checker u'/rhev/data-center/mnt/10.210.13.64:_ovirt__production/bda97276-a399-448f-9113-017972f6b55a/dom_md/metadata' is blocked for 20.00 seconds (check:282)
2019-08-07 09:40:47,132-0700 WARN (monitor/bda9727) [storage.Monitor] Host id for domain bda97276-a399-448f-9113-017972f6b55a was released (id: 5) (monitor:445)
2019-08-07 09:44:53,564-0700 WARN (check/loop) [storage.check] Checker u'/rhev/data-center/mnt/10.210.13.64:_ovirt__production/bda97276-a399-448f-9113-017972f6b55a/dom_md/metadata' is blocked for 20.00 seconds (check:282)
2019-08-07 09:46:38,604-0700 WARN (monitor/bda9727) [storage.Monitor] Host id for domain bda97276-a399-448f-9113-017972f6b55a was released (id: 5) (monitor:445)”
From: "Vrgotic, Marko" <M.Vrgotic@activevideo.com>
Date: Wednesday, 7 August 2019 at 09:09
To: "users@ovirt.org" <users@ovirt.org>
Subject: oVirt 4.3.5 potential issue with NFS storage
Dear oVIrt,
This is my third oVirt platform in the company, but first time I am seeing following logs:
“2019-08-07 16:00:16,099Z INFO [org.ovirt.engine.core.bll.provider.network.SyncNetworkProviderCommand] (EE-ManagedThreadFactory-engineScheduled-Thread-51) [1b85e637] Lock freed to object 'EngineLock:{exclusiveLocks='[2350ee82-94ed-4f90-9366-451e0104d1d6=PROVIDER]', sharedLocks=''}'
2019-08-07 16:00:25,618Z WARN [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] (EE-ManagedThreadFactory-engine-Thread-37723) [] domain 'bda97276-a399-448f-9113-017972f6b55a:ovirt_production' in problem 'PROBLEMATIC'. vds: 'ovirt-sj-05.ictv.com'
2019-08-07 16:00:40,630Z INFO [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] (EE-ManagedThreadFactory-engine-Thread-37735) [] Domain 'bda97276-a399-448f-9113-017972f6b55a:ovirt_production' recovered from problem. vds: 'ovirt-sj-05.ictv.com'
2019-08-07 16:00:40,652Z INFO [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] (EE-ManagedThreadFactory-engine-Thread-37737) [] Domain 'bda97276-a399-448f-9113-017972f6b55a:ovirt_production' recovered from problem. vds: 'ovirt-sj-01.ictv.com'
2019-08-07 16:00:40,652Z INFO [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] (EE-ManagedThreadFactory-engine-Thread-37737) [] Domain 'bda97276-a399-448f-9113-017972f6b55a:ovirt_production' has recovered from problem. No active host in the DC is reporting it as problematic, so clearing the domain recovery timer.”
Can you help me understanding why is this being reported?
This setup is:
5HOSTS, 3 in HA
SelfHostedEngine
Version 4.3.5
NFS based Netapp storage, version 4.1
“10.210.13.64:/ovirt_hosted_engine on /rhev/data-center/mnt/10.210.13.64:_ovirt__hosted__engine type nfs4 (rw,relatime,vers=4.1,rsize=65536,wsize=65536,namlen=255,soft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,clientaddr=10.210.11.14,local_lock=none,addr=10.210.13.64)
10.210.13.64:/ovirt_production on /rhev/data-center/mnt/10.210.13.64:_ovirt__production type nfs4 (rw,relatime,vers=4.1,rsize=65536,wsize=65536,namlen=255,soft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,clientaddr=10.210.11.14,local_lock=none,addr=10.210.13.64)
tmpfs on /run/user/0 type tmpfs (rw,nosuid,nodev,relatime,seclabel,size=9878396k,mode=700)”
First mount is SHE dedicated storage.
Second mount “ovirt_produciton” is for other VM Guests.
Kindly awaiting your reply.
Marko Vrgotic
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ICRKHD3GXTPQEZN2T6LJBS6YIVLER6TP/