[ovirt-users] Adding posix compliant FS

carl langlois crl.langlois at gmail.com
Tue Mar 21 19:36:09 UTC 2017


Okey, i have manager to use the POSIX compliant FS.

First thing i did was to remove any multipath stuff from the disk and have
a standard parttiton table i.e /dev/sdb1 (but i do not think that realy
help)
change block device(/dev/sdb1) group and owner to vdsm:kvm (did not do the
trick either got still permission denied)
create a directory in /rhev/data-center/mnt/_dev_sdb1 and set owner and
group to vdsm:kvm (this did the trick)

So why did i had to create the last directory by hand to make it work?..i
my missing something?

Thanks
Carl




On Tue, Mar 21, 2017 at 10:50 AM, Fred Rolland <frolland at redhat.com> wrote:

> Can you try :
>
> chown -R vdsm:kvm /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>
> On Tue, Mar 21, 2017 at 4:32 PM, carl langlois <crl.langlois at gmail.com>
> wrote:
>
>>
>> jsonrpc.Executor/7::WARNING::2017-03-21 09:27:40,099::outOfProcess::19
>> 3::Storage.oop::(validateAccess) Permission denied for directory:
>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
>> with permissions:7
>> jsonrpc.Executor/7::INFO::2017-03-21 09:27:40,099::mount::233::storage.Mount::(umount)
>> unmounting /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5
>> 0026B726804F13B1
>> jsonrpc.Executor/7::DEBUG::2017-03-21 09:27:40,104::utils::871::storage.Mount::(stopwatch)
>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
>> unmounted: 0.00 seconds
>> jsonrpc.Executor/7::ERROR::2017-03-21 09:27:40,104::hsm::2403::Storage.HSM::(connectStorageServer)
>> Could not connect to storageServer
>> Traceback (most recent call last):
>>   File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>> connectStorageServer
>>     conObj.connect()
>>   File "/usr/share/vdsm/storage/storageServer.py", line 249, in connect
>>     six.reraise(t, v, tb)
>>   File "/usr/share/vdsm/storage/storageServer.py", line 242, in connect
>>     self.getMountObj().getRecord().fs_file)
>>   File "/usr/share/vdsm/storage/fileSD.py", line 81, in validateDirAccess
>>     raise se.StorageServerAccessPermissionError(dirPath)
>> StorageServerAccessPermissionError: Permission settings on the specified
>> path do not allow access to the storage. Verify permission settings on the
>> specified storage path.: 'path = /rhev/data-center/mnt/_dev_map
>> per_KINGSTON__SV300S37A240G__50026B726804F13B1'
>> jsonrpc.Executor/7::DEBUG::201
>>
>> Thanks again.
>>
>>
>> On Tue, Mar 21, 2017 at 10:14 AM, Fred Rolland <frolland at redhat.com>
>> wrote:
>>
>>> Can you share the VDSM log again ?
>>>
>>> On Tue, Mar 21, 2017 at 4:08 PM, carl langlois <crl.langlois at gmail.com>
>>> wrote:
>>>
>>>> Interesting, when i'm using /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>> now the UI give error on the permission setting..
>>>>
>>>> root at ovhost4 ~]# ls -al /dev/mapper/KINGSTON_SV300S37A
>>>> 240G_50026B726804F13B1
>>>> lrwxrwxrwx 1 root root 7 Mar 18 08:28 /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>> -> ../dm-3
>>>>
>>>> and the permission on the dm-3
>>>>
>>>> [root at ovhost4 ~]# ls -al /dev/dm-3
>>>> brw-rw---- 1 vdsm kvm 253, 3 Mar 18 08:28 /dev/dm-3
>>>>
>>>>
>>>> how do i change the permission on the sym link..
>>>>
>>>> Thanks
>>>>
>>>>
>>>>
>>>>
>>>> On Tue, Mar 21, 2017 at 10:00 AM, Fred Rolland <frolland at redhat.com>
>>>> wrote:
>>>>
>>>>> Can you try to use /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>> in the UI.
>>>>> It seems the kernel change the path that we use to mount and then we
>>>>> cannot validate that the mount exists.
>>>>>
>>>>> It should be anyway better as the mapping could change after reboot.
>>>>>
>>>>> On Tue, Mar 21, 2017 at 2:20 PM, carl langlois <crl.langlois at gmail.com
>>>>> > wrote:
>>>>>
>>>>>> Here is the /proc/mounts
>>>>>>
>>>>>> rootfs / rootfs rw 0 0
>>>>>> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0
>>>>>> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0
>>>>>> devtmpfs /dev devtmpfs rw,nosuid,size=65948884k,nr_inodes=16487221,mode=755
>>>>>> 0 0
>>>>>> securityfs /sys/kernel/security securityfs
>>>>>> rw,nosuid,nodev,noexec,relatime 0 0
>>>>>> tmpfs /dev/shm tmpfs rw,nosuid,nodev 0 0
>>>>>> devpts /dev/pts devpts rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000
>>>>>> 0 0
>>>>>> tmpfs /run tmpfs rw,nosuid,nodev,mode=755 0 0
>>>>>> tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0
>>>>>> cgroup /sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatim
>>>>>> e,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd
>>>>>> 0 0
>>>>>> pstore /sys/fs/pstore pstore rw,nosuid,nodev,noexec,relatime 0 0
>>>>>> cgroup /sys/fs/cgroup/cpu,cpuacct cgroup
>>>>>> rw,nosuid,nodev,noexec,relatime,cpuacct,cpu 0 0
>>>>>> cgroup /sys/fs/cgroup/net_cls,net_prio cgroup
>>>>>> rw,nosuid,nodev,noexec,relatime,net_prio,net_cls 0 0
>>>>>> cgroup /sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb
>>>>>> 0 0
>>>>>> configfs /sys/kernel/config configfs rw,relatime 0 0
>>>>>> /dev/mapper/cl_ovhost1-root / xfs rw,relatime,attr2,inode64,noquota
>>>>>> 0 0
>>>>>> systemd-1 /proc/sys/fs/binfmt_misc autofs
>>>>>> rw,relatime,fd=35,pgrp=1,timeout=300,minproto=5,maxproto=5,direct 0 0
>>>>>> mqueue /dev/mqueue mqueue rw,relatime 0 0
>>>>>> debugfs /sys/kernel/debug debugfs rw,relatime 0 0
>>>>>> hugetlbfs /dev/hugepages hugetlbfs rw,relatime 0 0
>>>>>> tmpfs /tmp tmpfs rw 0 0
>>>>>> nfsd /proc/fs/nfsd nfsd rw,relatime 0 0
>>>>>> /dev/mapper/cl_ovhost1-home /home xfs rw,relatime,attr2,inode64,noquota
>>>>>> 0 0
>>>>>> /dev/sda1 /boot xfs rw,relatime,attr2,inode64,noquota 0 0
>>>>>> sunrpc /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0
>>>>>> tmpfs /run/user/42 tmpfs rw,nosuid,nodev,relatime,size=
>>>>>> 13192948k,mode=700,uid=42,gid=42 0 0
>>>>>> gvfsd-fuse /run/user/42/gvfs fuse.gvfsd-fuse
>>>>>> rw,nosuid,nodev,relatime,user_id=42,group_id=42 0 0
>>>>>> fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0
>>>>>> ovhost2:/home/exports/defaultdata /rhev/data-center/mnt/ovhost2:_home_exports_defaultdata
>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>> ovhost2:/home/exports/ISO /rhev/data-center/mnt/ovhost2:_home_exports_ISO
>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>> ovhost2:/home/exports/data /rhev/data-center/mnt/ovhost2:_home_exports_data
>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>> tmpfs /run/user/0 tmpfs rw,nosuid,nodev,relatime,size=13192948k,mode=700
>>>>>> 0 0
>>>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>>> /rhev/data-center/mnt/_dev_dm-3 ext4 rw,nosuid,relatime,data=ordered
>>>>>> 0 0
>>>>>>
>>>>>> Thanks you for your help.
>>>>>>
>>>>>> Carl
>>>>>>
>>>>>>
>>>>>> On Tue, Mar 21, 2017 at 6:31 AM, Fred Rolland <frolland at redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Can you provide the content of /proc/mounts after it has being
>>>>>>> mounted by VDSM ?
>>>>>>>
>>>>>>> On Tue, Mar 21, 2017 at 12:28 PM, carl langlois <
>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>
>>>>>>>> Here is the vdsm.log
>>>>>>>>
>>>>>>>>
>>>>>>>> jsonrpc.Executor/0::ERROR::2017-03-18
>>>>>>>> 08:23:48,317::hsm::2403::Storage.HSM::(connectStorageServer) Could
>>>>>>>> not connect to storageServer
>>>>>>>> Traceback (most recent call last):
>>>>>>>>   File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>>>> connectStorageServer
>>>>>>>>     conObj.connect()
>>>>>>>>   File "/usr/share/vdsm/storage/storageServer.py", line 242, in
>>>>>>>> connect
>>>>>>>>     self.getMountObj().getRecord().fs_file)
>>>>>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>> line 260, in getRecord
>>>>>>>>     (self.fs_spec, self.fs_file))
>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>>>
>>>>>>>>
>>>>>>>> thanks
>>>>>>>>
>>>>>>>> On Fri, Mar 17, 2017 at 3:06 PM, Fred Rolland <frolland at redhat.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Please send Vdsm log.
>>>>>>>>> Thanks
>>>>>>>>>
>>>>>>>>> On Fri, Mar 17, 2017 at 8:46 PM, carl langlois <
>>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> The link that you send is for NFS strorage but i am trying to add
>>>>>>>>>> a POSIX compliant.
>>>>>>>>>>
>>>>>>>>>> [image: Inline image 1]
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> when i press okey it mount the disk to :
>>>>>>>>>>
>>>>>>>>>> [root at ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/
>>>>>>>>>> total 28
>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .
>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm  4096 Mar 17 13:35 ..
>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found
>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .Trash-0
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> and doing a touch with vdsm user work
>>>>>>>>>>
>>>>>>>>>> [root at ovhost4 ~]# sudo -u vdsm touch
>>>>>>>>>>  /rhev/data-center/mnt/_dev_dm-4/test
>>>>>>>>>> [root at ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/
>>>>>>>>>> total 28
>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 17 13:44 .
>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm  4096 Mar 17 13:35 ..
>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found
>>>>>>>>>> -rw-r--r--. 1 vdsm kvm     0 Mar 17 13:44 test
>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .Trash-0
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> But it fail with a general exception error and the storage does
>>>>>>>>>> not exist in ovirt
>>>>>>>>>>
>>>>>>>>>> any help would be appreciated.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Which log you need to see?
>>>>>>>>>>
>>>>>>>>>> Thanks
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Le jeu. 16 mars 2017 17:02, Fred Rolland <frolland at redhat.com> a
>>>>>>>>>> écrit :
>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> Can you check if the folder permissions are OK ?
>>>>>>>>>>> Check [1] for more details.
>>>>>>>>>>>
>>>>>>>>>>> Can you share more of the log ?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> [1] https://www.ovirt.org/documentation/how-to/troubleshooting/t
>>>>>>>>>>> roubleshooting-nfs-storage-issues/
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Mar 16, 2017 at 7:49 PM, carl langlois <
>>>>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>> Hi Guys,
>>>>>>>>>>>
>>>>>>>>>>> I am trying to add a posix FS on one of my host. Ovirt in
>>>>>>>>>>> actually mounting it but fail with "Error while executing
>>>>>>>>>>> action Add Storage Connection: General Exception"
>>>>>>>>>>>
>>>>>>>>>>> If i look in the vdsm.log i cant see
>>>>>>>>>>>
>>>>>>>>>>> sonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>> 12:39:28,248::fileUtils::209::Storage.fileUtils::(createdir)
>>>>>>>>>>> Creating directory: /rhev/data-center/mnt/_dev_dm-3 mode: None
>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>> 12:39:28,248::fileUtils::218::Storage.fileUtils::(createdir)
>>>>>>>>>>> Using existing directory: /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>>>> jsonrpc.Executor/7::INFO::2017-03-16
>>>>>>>>>>> 12:39:28,248::mount::226::storage.Mount::(mount) mounting
>>>>>>>>>>> /dev/dm-3 at /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>> 12:39:28,270::utils::871::storage.Mount::(stopwatch)
>>>>>>>>>>> /rhev/data-center/mnt/_dev_dm-3 mounted: 0.02 seconds
>>>>>>>>>>> jsonrpc.Executor/7::ERROR::2017-03-16
>>>>>>>>>>> 12:39:28,271::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>>>>>>> Could not connect to storageServer
>>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>>>   File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>>>>>>> connectStorageServer
>>>>>>>>>>>     conObj.connect()
>>>>>>>>>>>   File "/usr/share/vdsm/storage/storageServer.py", line 242, in
>>>>>>>>>>> connect
>>>>>>>>>>>     self.getMountObj().getRecord().fs_file)
>>>>>>>>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>>>>> line 260, in getRecord
>>>>>>>>>>>     (self.fs_spec, self.fs_file))
>>>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> any help would be appreciated.
>>>>>>>>>>>
>>>>>>>>>>> Thanks
>>>>>>>>>>>
>>>>>>>>>>> CL
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> Users mailing list
>>>>>>>>>>> Users at ovirt.org
>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170321/a4fc1a5d/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image.png
Type: image/png
Size: 27276 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170321/a4fc1a5d/attachment-0001.png>


More information about the Users mailing list