[ovirt-users] Adding posix compliant FS

Fred Rolland frolland at redhat.com
Wed Mar 22 07:40:57 UTC 2017


Is it a local disk ? If you want to use a local disk, it is not the way to
do it.
POSIX Storage Domain should be accessible from all the hosts in the cluster.

On Tue, Mar 21, 2017 at 9:36 PM, carl langlois <crl.langlois at gmail.com>
wrote:

> Okey, i have manager to use the POSIX compliant FS.
>
> First thing i did was to remove any multipath stuff from the disk and have
> a standard parttiton table i.e /dev/sdb1 (but i do not think that realy
> help)
> change block device(/dev/sdb1) group and owner to vdsm:kvm (did not do the
> trick either got still permission denied)
> create a directory in /rhev/data-center/mnt/_dev_sdb1 and set owner and
> group to vdsm:kvm (this did the trick)
>
> So why did i had to create the last directory by hand to make it work?..i
> my missing something?
>
> Thanks
> Carl
>
>
>
>
> On Tue, Mar 21, 2017 at 10:50 AM, Fred Rolland <frolland at redhat.com>
> wrote:
>
>> Can you try :
>>
>> chown -R vdsm:kvm /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>
>> On Tue, Mar 21, 2017 at 4:32 PM, carl langlois <crl.langlois at gmail.com>
>> wrote:
>>
>>>
>>> jsonrpc.Executor/7::WARNING::2017-03-21 09:27:40,099::outOfProcess::19
>>> 3::Storage.oop::(validateAccess) Permission denied for directory:
>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
>>> with permissions:7
>>> jsonrpc.Executor/7::INFO::2017-03-21 09:27:40,099::mount::233::storage.Mount::(umount)
>>> unmounting /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5
>>> 0026B726804F13B1
>>> jsonrpc.Executor/7::DEBUG::2017-03-21 09:27:40,104::utils::871::storage.Mount::(stopwatch)
>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
>>> unmounted: 0.00 seconds
>>> jsonrpc.Executor/7::ERROR::2017-03-21 09:27:40,104::hsm::2403::Storage.HSM::(connectStorageServer)
>>> Could not connect to storageServer
>>> Traceback (most recent call last):
>>>   File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>> connectStorageServer
>>>     conObj.connect()
>>>   File "/usr/share/vdsm/storage/storageServer.py", line 249, in connect
>>>     six.reraise(t, v, tb)
>>>   File "/usr/share/vdsm/storage/storageServer.py", line 242, in connect
>>>     self.getMountObj().getRecord().fs_file)
>>>   File "/usr/share/vdsm/storage/fileSD.py", line 81, in
>>> validateDirAccess
>>>     raise se.StorageServerAccessPermissionError(dirPath)
>>> StorageServerAccessPermissionError: Permission settings on the
>>> specified path do not allow access to the storage. Verify permission
>>> settings on the specified storage path.: 'path =
>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5
>>> 0026B726804F13B1'
>>> jsonrpc.Executor/7::DEBUG::201
>>>
>>> Thanks again.
>>>
>>>
>>> On Tue, Mar 21, 2017 at 10:14 AM, Fred Rolland <frolland at redhat.com>
>>> wrote:
>>>
>>>> Can you share the VDSM log again ?
>>>>
>>>> On Tue, Mar 21, 2017 at 4:08 PM, carl langlois <crl.langlois at gmail.com>
>>>> wrote:
>>>>
>>>>> Interesting, when i'm using /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>> now the UI give error on the permission setting..
>>>>>
>>>>> root at ovhost4 ~]# ls -al /dev/mapper/KINGSTON_SV300S37A
>>>>> 240G_50026B726804F13B1
>>>>> lrwxrwxrwx 1 root root 7 Mar 18 08:28 /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>> -> ../dm-3
>>>>>
>>>>> and the permission on the dm-3
>>>>>
>>>>> [root at ovhost4 ~]# ls -al /dev/dm-3
>>>>> brw-rw---- 1 vdsm kvm 253, 3 Mar 18 08:28 /dev/dm-3
>>>>>
>>>>>
>>>>> how do i change the permission on the sym link..
>>>>>
>>>>> Thanks
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Tue, Mar 21, 2017 at 10:00 AM, Fred Rolland <frolland at redhat.com>
>>>>> wrote:
>>>>>
>>>>>> Can you try to use /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>>> in the UI.
>>>>>> It seems the kernel change the path that we use to mount and then we
>>>>>> cannot validate that the mount exists.
>>>>>>
>>>>>> It should be anyway better as the mapping could change after reboot.
>>>>>>
>>>>>> On Tue, Mar 21, 2017 at 2:20 PM, carl langlois <
>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>
>>>>>>> Here is the /proc/mounts
>>>>>>>
>>>>>>> rootfs / rootfs rw 0 0
>>>>>>> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0
>>>>>>> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0
>>>>>>> devtmpfs /dev devtmpfs rw,nosuid,size=65948884k,nr_inodes=16487221,mode=755
>>>>>>> 0 0
>>>>>>> securityfs /sys/kernel/security securityfs
>>>>>>> rw,nosuid,nodev,noexec,relatime 0 0
>>>>>>> tmpfs /dev/shm tmpfs rw,nosuid,nodev 0 0
>>>>>>> devpts /dev/pts devpts rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000
>>>>>>> 0 0
>>>>>>> tmpfs /run tmpfs rw,nosuid,nodev,mode=755 0 0
>>>>>>> tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0
>>>>>>> cgroup /sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatim
>>>>>>> e,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd
>>>>>>> 0 0
>>>>>>> pstore /sys/fs/pstore pstore rw,nosuid,nodev,noexec,relatime 0 0
>>>>>>> cgroup /sys/fs/cgroup/cpu,cpuacct cgroup
>>>>>>> rw,nosuid,nodev,noexec,relatime,cpuacct,cpu 0 0
>>>>>>> cgroup /sys/fs/cgroup/net_cls,net_prio cgroup
>>>>>>> rw,nosuid,nodev,noexec,relatime,net_prio,net_cls 0 0
>>>>>>> cgroup /sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids
>>>>>>> 0 0
>>>>>>> cgroup /sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices
>>>>>>> 0 0
>>>>>>> cgroup /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset
>>>>>>> 0 0
>>>>>>> cgroup /sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio
>>>>>>> 0 0
>>>>>>> cgroup /sys/fs/cgroup/perf_event cgroup
>>>>>>> rw,nosuid,nodev,noexec,relatime,perf_event 0 0
>>>>>>> cgroup /sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory
>>>>>>> 0 0
>>>>>>> cgroup /sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer
>>>>>>> 0 0
>>>>>>> cgroup /sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb
>>>>>>> 0 0
>>>>>>> configfs /sys/kernel/config configfs rw,relatime 0 0
>>>>>>> /dev/mapper/cl_ovhost1-root / xfs rw,relatime,attr2,inode64,noquota
>>>>>>> 0 0
>>>>>>> systemd-1 /proc/sys/fs/binfmt_misc autofs
>>>>>>> rw,relatime,fd=35,pgrp=1,timeout=300,minproto=5,maxproto=5,direct 0
>>>>>>> 0
>>>>>>> mqueue /dev/mqueue mqueue rw,relatime 0 0
>>>>>>> debugfs /sys/kernel/debug debugfs rw,relatime 0 0
>>>>>>> hugetlbfs /dev/hugepages hugetlbfs rw,relatime 0 0
>>>>>>> tmpfs /tmp tmpfs rw 0 0
>>>>>>> nfsd /proc/fs/nfsd nfsd rw,relatime 0 0
>>>>>>> /dev/mapper/cl_ovhost1-home /home xfs rw,relatime,attr2,inode64,noquota
>>>>>>> 0 0
>>>>>>> /dev/sda1 /boot xfs rw,relatime,attr2,inode64,noquota 0 0
>>>>>>> sunrpc /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0
>>>>>>> tmpfs /run/user/42 tmpfs rw,nosuid,nodev,relatime,size=
>>>>>>> 13192948k,mode=700,uid=42,gid=42 0 0
>>>>>>> gvfsd-fuse /run/user/42/gvfs fuse.gvfsd-fuse
>>>>>>> rw,nosuid,nodev,relatime,user_id=42,group_id=42 0 0
>>>>>>> fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0
>>>>>>> ovhost2:/home/exports/defaultdata /rhev/data-center/mnt/ovhost2:_home_exports_defaultdata
>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>>> ovhost2:/home/exports/ISO /rhev/data-center/mnt/ovhost2:_home_exports_ISO
>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>>> ovhost2:/home/exports/data /rhev/data-center/mnt/ovhost2:_home_exports_data
>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>>> tmpfs /run/user/0 tmpfs rw,nosuid,nodev,relatime,size=13192948k,mode=700
>>>>>>> 0 0
>>>>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>>>> /rhev/data-center/mnt/_dev_dm-3 ext4 rw,nosuid,relatime,data=ordered
>>>>>>> 0 0
>>>>>>>
>>>>>>> Thanks you for your help.
>>>>>>>
>>>>>>> Carl
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Mar 21, 2017 at 6:31 AM, Fred Rolland <frolland at redhat.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Can you provide the content of /proc/mounts after it has being
>>>>>>>> mounted by VDSM ?
>>>>>>>>
>>>>>>>> On Tue, Mar 21, 2017 at 12:28 PM, carl langlois <
>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Here is the vdsm.log
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> jsonrpc.Executor/0::ERROR::2017-03-18
>>>>>>>>> 08:23:48,317::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>>>>> Could not connect to storageServer
>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>   File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>>>>> connectStorageServer
>>>>>>>>>     conObj.connect()
>>>>>>>>>   File "/usr/share/vdsm/storage/storageServer.py", line 242, in
>>>>>>>>> connect
>>>>>>>>>     self.getMountObj().getRecord().fs_file)
>>>>>>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>>> line 260, in getRecord
>>>>>>>>>     (self.fs_spec, self.fs_file))
>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> thanks
>>>>>>>>>
>>>>>>>>> On Fri, Mar 17, 2017 at 3:06 PM, Fred Rolland <frolland at redhat.com
>>>>>>>>> > wrote:
>>>>>>>>>
>>>>>>>>>> Please send Vdsm log.
>>>>>>>>>> Thanks
>>>>>>>>>>
>>>>>>>>>> On Fri, Mar 17, 2017 at 8:46 PM, carl langlois <
>>>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> The link that you send is for NFS strorage but i am trying to
>>>>>>>>>>> add a POSIX compliant.
>>>>>>>>>>>
>>>>>>>>>>> [image: Inline image 1]
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> when i press okey it mount the disk to :
>>>>>>>>>>>
>>>>>>>>>>> [root at ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/
>>>>>>>>>>> total 28
>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .
>>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm  4096 Mar 17 13:35 ..
>>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found
>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .Trash-0
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> and doing a touch with vdsm user work
>>>>>>>>>>>
>>>>>>>>>>> [root at ovhost4 ~]# sudo -u vdsm touch
>>>>>>>>>>>  /rhev/data-center/mnt/_dev_dm-4/test
>>>>>>>>>>> [root at ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/
>>>>>>>>>>> total 28
>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 17 13:44 .
>>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm  4096 Mar 17 13:35 ..
>>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found
>>>>>>>>>>> -rw-r--r--. 1 vdsm kvm     0 Mar 17 13:44 test
>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .Trash-0
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> But it fail with a general exception error and the storage does
>>>>>>>>>>> not exist in ovirt
>>>>>>>>>>>
>>>>>>>>>>> any help would be appreciated.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Which log you need to see?
>>>>>>>>>>>
>>>>>>>>>>> Thanks
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Le jeu. 16 mars 2017 17:02, Fred Rolland <frolland at redhat.com>
>>>>>>>>>>> a écrit :
>>>>>>>>>>>
>>>>>>>>>>>> Hi,
>>>>>>>>>>>>
>>>>>>>>>>>> Can you check if the folder permissions are OK ?
>>>>>>>>>>>> Check [1] for more details.
>>>>>>>>>>>>
>>>>>>>>>>>> Can you share more of the log ?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> [1] https://www.ovirt.org/document
>>>>>>>>>>>> ation/how-to/troubleshooting/troubleshooting-nfs-storage-iss
>>>>>>>>>>>> ues/
>>>>>>>>>>>>
>>>>>>>>>>>> On Thu, Mar 16, 2017 at 7:49 PM, carl langlois <
>>>>>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Hi Guys,
>>>>>>>>>>>>
>>>>>>>>>>>> I am trying to add a posix FS on one of my host. Ovirt in
>>>>>>>>>>>> actually mounting it but fail with "Error while executing
>>>>>>>>>>>> action Add Storage Connection: General Exception"
>>>>>>>>>>>>
>>>>>>>>>>>> If i look in the vdsm.log i cant see
>>>>>>>>>>>>
>>>>>>>>>>>> sonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>> 12:39:28,248::fileUtils::209::Storage.fileUtils::(createdir)
>>>>>>>>>>>> Creating directory: /rhev/data-center/mnt/_dev_dm-3 mode: None
>>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>> 12:39:28,248::fileUtils::218::Storage.fileUtils::(createdir)
>>>>>>>>>>>> Using existing directory: /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>>>>> jsonrpc.Executor/7::INFO::2017-03-16
>>>>>>>>>>>> 12:39:28,248::mount::226::storage.Mount::(mount) mounting
>>>>>>>>>>>> /dev/dm-3 at /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>> 12:39:28,270::utils::871::storage.Mount::(stopwatch)
>>>>>>>>>>>> /rhev/data-center/mnt/_dev_dm-3 mounted: 0.02 seconds
>>>>>>>>>>>> jsonrpc.Executor/7::ERROR::2017-03-16
>>>>>>>>>>>> 12:39:28,271::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>>>>>>>> Could not connect to storageServer
>>>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>>>>   File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>>>>>>>> connectStorageServer
>>>>>>>>>>>>     conObj.connect()
>>>>>>>>>>>>   File "/usr/share/vdsm/storage/storageServer.py", line 242,
>>>>>>>>>>>> in connect
>>>>>>>>>>>>     self.getMountObj().getRecord().fs_file)
>>>>>>>>>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>>>>>> line 260, in getRecord
>>>>>>>>>>>>     (self.fs_spec, self.fs_file))
>>>>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> any help would be appreciated.
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks
>>>>>>>>>>>>
>>>>>>>>>>>> CL
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> Users mailing list
>>>>>>>>>>>> Users at ovirt.org
>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170322/f2bcd4fe/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image.png
Type: image/png
Size: 27276 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170322/f2bcd4fe/attachment-0001.png>


More information about the Users mailing list