[ovirt-users] Adding posix compliant FS

carl langlois crl.langlois at gmail.com
Tue Mar 21 14:32:53 UTC 2017


jsonrpc.Executor/7::WARNING::2017-03-21
09:27:40,099::outOfProcess::193::Storage.oop::(validateAccess) Permission
denied for directory:
/rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
with permissions:7
jsonrpc.Executor/7::INFO::2017-03-21
09:27:40,099::mount::233::storage.Mount::(umount) unmounting
/rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
jsonrpc.Executor/7::DEBUG::2017-03-21
09:27:40,104::utils::871::storage.Mount::(stopwatch)
/rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
unmounted: 0.00 seconds
jsonrpc.Executor/7::ERROR::2017-03-21
09:27:40,104::hsm::2403::Storage.HSM::(connectStorageServer) Could not
connect to storageServer
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/hsm.py", line 2400, in connectStorageServer
    conObj.connect()
  File "/usr/share/vdsm/storage/storageServer.py", line 249, in connect
    six.reraise(t, v, tb)
  File "/usr/share/vdsm/storage/storageServer.py", line 242, in connect
    self.getMountObj().getRecord().fs_file)
  File "/usr/share/vdsm/storage/fileSD.py", line 81, in validateDirAccess
    raise se.StorageServerAccessPermissionError(dirPath)
StorageServerAccessPermissionError: Permission settings on the specified
path do not allow access to the storage. Verify permission settings on the
specified storage path.: 'path =
/rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1'
jsonrpc.Executor/7::DEBUG::201

Thanks again.


On Tue, Mar 21, 2017 at 10:14 AM, Fred Rolland <frolland at redhat.com> wrote:

> Can you share the VDSM log again ?
>
> On Tue, Mar 21, 2017 at 4:08 PM, carl langlois <crl.langlois at gmail.com>
> wrote:
>
>> Interesting, when i'm using /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>> now the UI give error on the permission setting..
>>
>> root at ovhost4 ~]# ls -al /dev/mapper/KINGSTON_SV300S37A
>> 240G_50026B726804F13B1
>> lrwxrwxrwx 1 root root 7 Mar 18 08:28 /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>> -> ../dm-3
>>
>> and the permission on the dm-3
>>
>> [root at ovhost4 ~]# ls -al /dev/dm-3
>> brw-rw---- 1 vdsm kvm 253, 3 Mar 18 08:28 /dev/dm-3
>>
>>
>> how do i change the permission on the sym link..
>>
>> Thanks
>>
>>
>>
>>
>> On Tue, Mar 21, 2017 at 10:00 AM, Fred Rolland <frolland at redhat.com>
>> wrote:
>>
>>> Can you try to use /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>> in the UI.
>>> It seems the kernel change the path that we use to mount and then we
>>> cannot validate that the mount exists.
>>>
>>> It should be anyway better as the mapping could change after reboot.
>>>
>>> On Tue, Mar 21, 2017 at 2:20 PM, carl langlois <crl.langlois at gmail.com>
>>> wrote:
>>>
>>>> Here is the /proc/mounts
>>>>
>>>> rootfs / rootfs rw 0 0
>>>> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0
>>>> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0
>>>> devtmpfs /dev devtmpfs rw,nosuid,size=65948884k,nr_inodes=16487221,mode=755
>>>> 0 0
>>>> securityfs /sys/kernel/security securityfs
>>>> rw,nosuid,nodev,noexec,relatime 0 0
>>>> tmpfs /dev/shm tmpfs rw,nosuid,nodev 0 0
>>>> devpts /dev/pts devpts rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000
>>>> 0 0
>>>> tmpfs /run tmpfs rw,nosuid,nodev,mode=755 0 0
>>>> tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0
>>>> cgroup /sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatim
>>>> e,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd
>>>> 0 0
>>>> pstore /sys/fs/pstore pstore rw,nosuid,nodev,noexec,relatime 0 0
>>>> cgroup /sys/fs/cgroup/cpu,cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpuacct,cpu
>>>> 0 0
>>>> cgroup /sys/fs/cgroup/net_cls,net_prio cgroup
>>>> rw,nosuid,nodev,noexec,relatime,net_prio,net_cls 0 0
>>>> cgroup /sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids
>>>> 0 0
>>>> cgroup /sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices
>>>> 0 0
>>>> cgroup /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset
>>>> 0 0
>>>> cgroup /sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio
>>>> 0 0
>>>> cgroup /sys/fs/cgroup/perf_event cgroup rw,nosuid,nodev,noexec,relatime,perf_event
>>>> 0 0
>>>> cgroup /sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory
>>>> 0 0
>>>> cgroup /sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer
>>>> 0 0
>>>> cgroup /sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb
>>>> 0 0
>>>> configfs /sys/kernel/config configfs rw,relatime 0 0
>>>> /dev/mapper/cl_ovhost1-root / xfs rw,relatime,attr2,inode64,noquota 0 0
>>>> systemd-1 /proc/sys/fs/binfmt_misc autofs rw,relatime,fd=35,pgrp=1,timeo
>>>> ut=300,minproto=5,maxproto=5,direct 0 0
>>>> mqueue /dev/mqueue mqueue rw,relatime 0 0
>>>> debugfs /sys/kernel/debug debugfs rw,relatime 0 0
>>>> hugetlbfs /dev/hugepages hugetlbfs rw,relatime 0 0
>>>> tmpfs /tmp tmpfs rw 0 0
>>>> nfsd /proc/fs/nfsd nfsd rw,relatime 0 0
>>>> /dev/mapper/cl_ovhost1-home /home xfs rw,relatime,attr2,inode64,noquota
>>>> 0 0
>>>> /dev/sda1 /boot xfs rw,relatime,attr2,inode64,noquota 0 0
>>>> sunrpc /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0
>>>> tmpfs /run/user/42 tmpfs rw,nosuid,nodev,relatime,size=
>>>> 13192948k,mode=700,uid=42,gid=42 0 0
>>>> gvfsd-fuse /run/user/42/gvfs fuse.gvfsd-fuse
>>>> rw,nosuid,nodev,relatime,user_id=42,group_id=42 0 0
>>>> fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0
>>>> ovhost2:/home/exports/defaultdata /rhev/data-center/mnt/ovhost2:_home_exports_defaultdata
>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>> local_lock=none,addr=10.8.236.162 0 0
>>>> ovhost2:/home/exports/ISO /rhev/data-center/mnt/ovhost2:_home_exports_ISO
>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>> local_lock=none,addr=10.8.236.162 0 0
>>>> ovhost2:/home/exports/data /rhev/data-center/mnt/ovhost2:_home_exports_data
>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>> local_lock=none,addr=10.8.236.162 0 0
>>>> tmpfs /run/user/0 tmpfs rw,nosuid,nodev,relatime,size=13192948k,mode=700
>>>> 0 0
>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>> /rhev/data-center/mnt/_dev_dm-3 ext4 rw,nosuid,relatime,data=ordered 0
>>>> 0
>>>>
>>>> Thanks you for your help.
>>>>
>>>> Carl
>>>>
>>>>
>>>> On Tue, Mar 21, 2017 at 6:31 AM, Fred Rolland <frolland at redhat.com>
>>>> wrote:
>>>>
>>>>> Can you provide the content of /proc/mounts after it has being mounted
>>>>> by VDSM ?
>>>>>
>>>>> On Tue, Mar 21, 2017 at 12:28 PM, carl langlois <
>>>>> crl.langlois at gmail.com> wrote:
>>>>>
>>>>>> Here is the vdsm.log
>>>>>>
>>>>>>
>>>>>> jsonrpc.Executor/0::ERROR::2017-03-18 08:23:48,317::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>> Could not connect to storageServer
>>>>>> Traceback (most recent call last):
>>>>>>   File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>> connectStorageServer
>>>>>>     conObj.connect()
>>>>>>   File "/usr/share/vdsm/storage/storageServer.py", line 242, in
>>>>>> connect
>>>>>>     self.getMountObj().getRecord().fs_file)
>>>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>> line 260, in getRecord
>>>>>>     (self.fs_spec, self.fs_file))
>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>
>>>>>>
>>>>>> thanks
>>>>>>
>>>>>> On Fri, Mar 17, 2017 at 3:06 PM, Fred Rolland <frolland at redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Please send Vdsm log.
>>>>>>> Thanks
>>>>>>>
>>>>>>> On Fri, Mar 17, 2017 at 8:46 PM, carl langlois <
>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> The link that you send is for NFS strorage but i am trying to add a
>>>>>>>> POSIX compliant.
>>>>>>>>
>>>>>>>> [image: Inline image 1]
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> when i press okey it mount the disk to :
>>>>>>>>
>>>>>>>> [root at ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/
>>>>>>>> total 28
>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .
>>>>>>>> drwxr-xr-x. 6 vdsm kvm  4096 Mar 17 13:35 ..
>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found
>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .Trash-0
>>>>>>>>
>>>>>>>>
>>>>>>>> and doing a touch with vdsm user work
>>>>>>>>
>>>>>>>> [root at ovhost4 ~]# sudo -u vdsm touch  /rhev/data-center/mnt/_dev_dm
>>>>>>>> -4/test
>>>>>>>> [root at ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/
>>>>>>>> total 28
>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 17 13:44 .
>>>>>>>> drwxr-xr-x. 6 vdsm kvm  4096 Mar 17 13:35 ..
>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found
>>>>>>>> -rw-r--r--. 1 vdsm kvm     0 Mar 17 13:44 test
>>>>>>>> drwxr-xr-x. 4 vdsm kvm  4096 Mar 16 12:12 .Trash-0
>>>>>>>>
>>>>>>>>
>>>>>>>> But it fail with a general exception error and the storage does not
>>>>>>>> exist in ovirt
>>>>>>>>
>>>>>>>> any help would be appreciated.
>>>>>>>>
>>>>>>>>
>>>>>>>> Which log you need to see?
>>>>>>>>
>>>>>>>> Thanks
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Le jeu. 16 mars 2017 17:02, Fred Rolland <frolland at redhat.com> a
>>>>>>>> écrit :
>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> Can you check if the folder permissions are OK ?
>>>>>>>>> Check [1] for more details.
>>>>>>>>>
>>>>>>>>> Can you share more of the log ?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> [1] https://www.ovirt.org/documentation/how-to/troubleshooting/t
>>>>>>>>> roubleshooting-nfs-storage-issues/
>>>>>>>>>
>>>>>>>>> On Thu, Mar 16, 2017 at 7:49 PM, carl langlois <
>>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>>
>>>>>>>>> Hi Guys,
>>>>>>>>>
>>>>>>>>> I am trying to add a posix FS on one of my host. Ovirt in actually
>>>>>>>>> mounting it but fail with "Error while executing action Add
>>>>>>>>> Storage Connection: General Exception"
>>>>>>>>>
>>>>>>>>> If i look in the vdsm.log i cant see
>>>>>>>>>
>>>>>>>>> sonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>> 12:39:28,248::fileUtils::209::Storage.fileUtils::(createdir)
>>>>>>>>> Creating directory: /rhev/data-center/mnt/_dev_dm-3 mode: None
>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>> 12:39:28,248::fileUtils::218::Storage.fileUtils::(createdir)
>>>>>>>>> Using existing directory: /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>> jsonrpc.Executor/7::INFO::2017-03-16
>>>>>>>>> 12:39:28,248::mount::226::storage.Mount::(mount) mounting
>>>>>>>>> /dev/dm-3 at /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>> 12:39:28,270::utils::871::storage.Mount::(stopwatch)
>>>>>>>>> /rhev/data-center/mnt/_dev_dm-3 mounted: 0.02 seconds
>>>>>>>>> jsonrpc.Executor/7::ERROR::2017-03-16
>>>>>>>>> 12:39:28,271::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>>>>> Could not connect to storageServer
>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>   File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>>>>> connectStorageServer
>>>>>>>>>     conObj.connect()
>>>>>>>>>   File "/usr/share/vdsm/storage/storageServer.py", line 242, in
>>>>>>>>> connect
>>>>>>>>>     self.getMountObj().getRecord().fs_file)
>>>>>>>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>>> line 260, in getRecord
>>>>>>>>>     (self.fs_spec, self.fs_file))
>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> any help would be appreciated.
>>>>>>>>>
>>>>>>>>> Thanks
>>>>>>>>>
>>>>>>>>> CL
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list
>>>>>>>>> Users at ovirt.org
>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170321/655824e4/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image.png
Type: image/png
Size: 27276 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170321/655824e4/attachment-0001.png>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: vdsm.log
Type: application/octet-stream
Size: 14114 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170321/655824e4/attachment-0001.obj>


More information about the Users mailing list