Basically, I meant to verify the access by ssh, but I want to verify
something following your detailed reply:
According to [1], in order to set a NetApp NFS server, the required steps
should look like this:
# mount NetApp_NFS:/path/to/export /mnt
# chown -R 36.36 /mnt
# chmod -R 755 /mnt
# umount /mnt
Which is quite similar to the steps you've mentioned, except the last step
of unmounting:
Unmount the 10.214.13.64:/ovirt_production
I think that you had to unmount /mnt/rhevstore instead.
Can you please verify?
[1]
*Regards,*
*Shani Leviim*
On Sun, Aug 11, 2019 at 10:57 PM Vrgotic, Marko <M.Vrgotic(a)activevideo.com>
wrote:
Hi Shani,
Thank you for your reply, but
How do I do that?
Reason why I am asking is following:
Hosts 2,3,4 do not have that issue. Host 1 and 5 do.
What I learned previously is that when using Netapp based NFS, which we
are, it’s required to before provisioning SHE and/or just adding a Host to
a pool, it’s required to execute following steps:
Create random dir on a host:
- mkdir /mnt/rhevstore
Mount netapp volume to the dir
- mount -o sec=sys -t nfs 10.214.13.64:/ovirt_production /mnt/rhevstore
Set ownership to vdsm:kvm (36:36):
- chown -R vdsm:kvm /mnt/rhevstore/*
Unmount the 10.214.13.64:/ovirt_production
I do not expect the above ownership actions need to be done initially on
each host, before starting the deployment, otherwise it would be
practically impossible to expand the Host pool.
All 5 hosts are provisioned in same way. How? I am using foreman to
provision these servers, so they are built of same kickstart hostgroup
template.
I even installed ovirt-hosted-engine-setup package to make sure all
required packages, users and groups are in place before adding host to
oVirt via UI or Ansible.
Is it possible that we if I am already using or heavily using the
mentioned volume via Hosts already added to oVirt pool, that ownership
actions executed,on host about to be added to the pool, will fail to
complete setting ownership on all required files on the volume?
To repeat the question above: How do I make sure Host can read metadata
file of the storage volume?
Kindly awaiting your reply.
All best,
Marko Vrgotic
Sent from my iPhone
On 11 Aug 2019, at 01:19, Shani Leviim <sleviim(a)redhat.com> wrote:
Hi Marko,
Is seems that there's a connectivity problem with host 10.210.13.64.
Can you please make sure the metadata under
/rhev/data-center/mnt/10.210.13.64:_ovirt__production/6effda5e-1a0d-4312-bf93-d97fa9eb5aee/dom_md/metadata
is accessible?
*Regards, *
*Shani Leviim *
On Sat, Aug 10, 2019 at 2:57 AM Vrgotic, Marko <M.Vrgotic(a)activevideo.com>
wrote:
> Log files from ovirt engine and ovirt-sj-05 vdsm attached.
>
>
>
> Its related to host named:
ovirt-sj-05.ictv.com
>
>
>
> Kindly awaiting your reply.
>
>
>
>
>
> — — —
> Met vriendelijke groet / Kind regards,
>
> *Marko Vrgotic*
>
>
>
>
>
>
>
> *From: *"Vrgotic, Marko" <M.Vrgotic(a)activevideo.com>
> *Date: *Thursday, 8 August 2019 at 17:02
> *To: *Shani Leviim <sleviim(a)redhat.com>
> *Cc: *"users(a)ovirt.org" <users(a)ovirt.org>
> *Subject: *Re: [ovirt-users] Re: oVirt 4.3.5 potential issue with NFS
> storage
>
>
>
> Hey Shanii,
>
>
>
> Thank you for the reply.
>
> Sure, I will attach the full logs asap.
>
> What do you mean by “flow you are doing”?
>
>
>
> Kindly awaiting your reply.
>
>
>
> Marko Vrgotic
>
>
>
> *From: *Shani Leviim <sleviim(a)redhat.com>
> *Date: *Thursday, 8 August 2019 at 00:01
> *To: *"Vrgotic, Marko" <M.Vrgotic(a)activevideo.com>
> *Cc: *"users(a)ovirt.org" <users(a)ovirt.org>
> *Subject: *Re: [ovirt-users] Re: oVirt 4.3.5 potential issue with NFS
> storage
>
>
>
> Hi,
>
> Can you please clarify the flow you're doing?
>
> Also, can you please attach full vdsm and engine logs?
>
>
> *Regards,*
>
> *Shani Leviim*
>
>
>
>
>
> On Thu, Aug 8, 2019 at 6:25 AM Vrgotic, Marko <M.Vrgotic(a)activevideo.com>
> wrote:
>
> Log line form VDSM:
>
>
>
> “[root@ovirt-sj-05 ~]# tail -f /var/log/vdsm/vdsm.log | grep WARN
>
> 2019-08-07 09:40:03,556-0700 WARN (check/loop) [storage.check] Checker
>
u'/rhev/data-center/mnt/10.210.13.64:_ovirt__production/bda97276-a399-448f-9113-017972f6b55a/dom_md/metadata'
> is blocked for 20.00 seconds (check:282)
>
> 2019-08-07 09:40:47,132-0700 WARN (monitor/bda9727) [storage.Monitor]
> Host id for domain bda97276-a399-448f-9113-017972f6b55a was released (id:
> 5) (monitor:445)
>
> 2019-08-07 09:44:53,564-0700 WARN (check/loop) [storage.check] Checker
>
u'/rhev/data-center/mnt/10.210.13.64:_ovirt__production/bda97276-a399-448f-9113-017972f6b55a/dom_md/metadata'
> is blocked for 20.00 seconds (check:282)
>
> 2019-08-07 09:46:38,604-0700 WARN (monitor/bda9727) [storage.Monitor]
> Host id for domain bda97276-a399-448f-9113-017972f6b55a was released (id:
> 5) (monitor:445)”
>
>
>
>
>
>
>
> *From: *"Vrgotic, Marko" <M.Vrgotic(a)activevideo.com>
> *Date: *Wednesday, 7 August 2019 at 09:09
> *To: *"users(a)ovirt.org" <users(a)ovirt.org>
> *Subject: *oVirt 4.3.5 potential issue with NFS storage
>
>
>
> Dear oVIrt,
>
>
>
> This is my third oVirt platform in the company, but first time I am
> seeing following logs:
>
>
>
> “2019-08-07 16:00:16,099Z INFO
> [org.ovirt.engine.core.bll.provider.network.SyncNetworkProviderCommand]
> (EE-ManagedThreadFactory-engineScheduled-Thread-51) [1b85e637] Lock freed
> to object
>
'EngineLock:{exclusiveLocks='[2350ee82-94ed-4f90-9366-451e0104d1d6=PROVIDER]',
> sharedLocks=''}'
>
> 2019-08-07 16:00:25,618Z WARN
> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy]
> (EE-ManagedThreadFactory-engine-Thread-37723) [] domain
> 'bda97276-a399-448f-9113-017972f6b55a:ovirt_production' in problem
> 'PROBLEMATIC'. vds: 'ovirt-sj-05.ictv.com'
>
> 2019-08-07 16:00:40,630Z INFO
> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy]
> (EE-ManagedThreadFactory-engine-Thread-37735) [] Domain
> 'bda97276-a399-448f-9113-017972f6b55a:ovirt_production' recovered from
> problem. vds: 'ovirt-sj-05.ictv.com'
>
> 2019-08-07 16:00:40,652Z INFO
> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy]
> (EE-ManagedThreadFactory-engine-Thread-37737) [] Domain
> 'bda97276-a399-448f-9113-017972f6b55a:ovirt_production' recovered from
> problem. vds: 'ovirt-sj-01.ictv.com'
>
> 2019-08-07 16:00:40,652Z INFO
> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy]
> (EE-ManagedThreadFactory-engine-Thread-37737) [] Domain
> 'bda97276-a399-448f-9113-017972f6b55a:ovirt_production' has recovered from
> problem. No active host in the DC is reporting it as problematic, so
> clearing the domain recovery timer.”
>
>
>
> Can you help me understanding why is this being reported?
>
>
>
> This setup is:
>
>
>
> 5HOSTS, 3 in HA
>
> SelfHostedEngine
>
> Version 4.3.5
>
> NFS based Netapp storage, version 4.1
>
> “10.210.13.64:/ovirt_hosted_engine on
/rhev/data-center/mnt/10.210.13.64:_ovirt__hosted__engine
> type nfs4
>
(rw,relatime,vers=4.1,rsize=65536,wsize=65536,namlen=255,soft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,clientaddr=10.210.11.14,local_lock=none,addr=10.210.13.64)
>
>
>
> 10.210.13.64:/ovirt_production on
/rhev/data-center/mnt/10.210.13.64:_ovirt__production
> type nfs4
>
(rw,relatime,vers=4.1,rsize=65536,wsize=65536,namlen=255,soft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,clientaddr=10.210.11.14,local_lock=none,addr=10.210.13.64)
>
> tmpfs on /run/user/0 type tmpfs
> (rw,nosuid,nodev,relatime,seclabel,size=9878396k,mode=700)”
>
>
>
> First mount is SHE dedicated storage.
>
> Second mount “ovirt_produciton” is for other VM Guests.
>
>
>
> Kindly awaiting your reply.
>
>
>
> Marko Vrgotic
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ICRKHD3GXTP...
>
>