Is it a local disk ? If you want to use a local disk, it is not the way to
do it.
POSIX Storage Domain should be accessible from all the hosts in the cluster.
On Tue, Mar 21, 2017 at 9:36 PM, carl langlois <crl.langlois(a)gmail.com>
wrote:
Okey, i have manager to use the POSIX compliant FS.
First thing i did was to remove any multipath stuff from the disk and have
a standard parttiton table i.e /dev/sdb1 (but i do not think that realy
help)
change block device(/dev/sdb1) group and owner to vdsm:kvm (did not do the
trick either got still permission denied)
create a directory in /rhev/data-center/mnt/_dev_sdb1 and set owner and
group to vdsm:kvm (this did the trick)
So why did i had to create the last directory by hand to make it work?..i
my missing something?
Thanks
Carl
On Tue, Mar 21, 2017 at 10:50 AM, Fred Rolland <frolland(a)redhat.com>
wrote:
> Can you try :
>
> chown -R vdsm:kvm /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>
> On Tue, Mar 21, 2017 at 4:32 PM, carl langlois <crl.langlois(a)gmail.com>
> wrote:
>
>>
>> jsonrpc.Executor/7::WARNING::2017-03-21 09:27:40,099::outOfProcess::19
>> 3::Storage.oop::(validateAccess) Permission denied for directory:
>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
>> with permissions:7
>> jsonrpc.Executor/7::INFO::2017-03-21
09:27:40,099::mount::233::storage.Mount::(umount)
>> unmounting /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5
>> 0026B726804F13B1
>> jsonrpc.Executor/7::DEBUG::2017-03-21
09:27:40,104::utils::871::storage.Mount::(stopwatch)
>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
>> unmounted: 0.00 seconds
>> jsonrpc.Executor/7::ERROR::2017-03-21
09:27:40,104::hsm::2403::Storage.HSM::(connectStorageServer)
>> Could not connect to storageServer
>> Traceback (most recent call last):
>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>> connectStorageServer
>> conObj.connect()
>> File "/usr/share/vdsm/storage/storageServer.py", line 249, in
connect
>> six.reraise(t, v, tb)
>> File "/usr/share/vdsm/storage/storageServer.py", line 242, in
connect
>> self.getMountObj().getRecord().fs_file)
>> File "/usr/share/vdsm/storage/fileSD.py", line 81, in
>> validateDirAccess
>> raise se.StorageServerAccessPermissionError(dirPath)
>> StorageServerAccessPermissionError: Permission settings on the
>> specified path do not allow access to the storage. Verify permission
>> settings on the specified storage path.: 'path =
>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5
>> 0026B726804F13B1'
>> jsonrpc.Executor/7::DEBUG::201
>>
>> Thanks again.
>>
>>
>> On Tue, Mar 21, 2017 at 10:14 AM, Fred Rolland <frolland(a)redhat.com>
>> wrote:
>>
>>> Can you share the VDSM log again ?
>>>
>>> On Tue, Mar 21, 2017 at 4:08 PM, carl langlois
<crl.langlois(a)gmail.com>
>>> wrote:
>>>
>>>> Interesting, when i'm using
/dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>> now the UI give error on the permission setting..
>>>>
>>>> root@ovhost4 ~]# ls -al /dev/mapper/KINGSTON_SV300S37A
>>>> 240G_50026B726804F13B1
>>>> lrwxrwxrwx 1 root root 7 Mar 18 08:28
/dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>> -> ../dm-3
>>>>
>>>> and the permission on the dm-3
>>>>
>>>> [root@ovhost4 ~]# ls -al /dev/dm-3
>>>> brw-rw---- 1 vdsm kvm 253, 3 Mar 18 08:28 /dev/dm-3
>>>>
>>>>
>>>> how do i change the permission on the sym link..
>>>>
>>>> Thanks
>>>>
>>>>
>>>>
>>>>
>>>> On Tue, Mar 21, 2017 at 10:00 AM, Fred Rolland
<frolland(a)redhat.com>
>>>> wrote:
>>>>
>>>>> Can you try to use
/dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>> in the UI.
>>>>> It seems the kernel change the path that we use to mount and then we
>>>>> cannot validate that the mount exists.
>>>>>
>>>>> It should be anyway better as the mapping could change after reboot.
>>>>>
>>>>> On Tue, Mar 21, 2017 at 2:20 PM, carl langlois <
>>>>> crl.langlois(a)gmail.com> wrote:
>>>>>
>>>>>> Here is the /proc/mounts
>>>>>>
>>>>>> rootfs / rootfs rw 0 0
>>>>>> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0
>>>>>> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0
>>>>>> devtmpfs /dev devtmpfs
rw,nosuid,size=65948884k,nr_inodes=16487221,mode=755
>>>>>> 0 0
>>>>>> securityfs /sys/kernel/security securityfs
>>>>>> rw,nosuid,nodev,noexec,relatime 0 0
>>>>>> tmpfs /dev/shm tmpfs rw,nosuid,nodev 0 0
>>>>>> devpts /dev/pts devpts
rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000
>>>>>> 0 0
>>>>>> tmpfs /run tmpfs rw,nosuid,nodev,mode=755 0 0
>>>>>> tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0
>>>>>> cgroup /sys/fs/cgroup/systemd cgroup
rw,nosuid,nodev,noexec,relatim
>>>>>>
e,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd
>>>>>> 0 0
>>>>>> pstore /sys/fs/pstore pstore rw,nosuid,nodev,noexec,relatime 0 0
>>>>>> cgroup /sys/fs/cgroup/cpu,cpuacct cgroup
>>>>>> rw,nosuid,nodev,noexec,relatime,cpuacct,cpu 0 0
>>>>>> cgroup /sys/fs/cgroup/net_cls,net_prio cgroup
>>>>>> rw,nosuid,nodev,noexec,relatime,net_prio,net_cls 0 0
>>>>>> cgroup /sys/fs/cgroup/pids cgroup
rw,nosuid,nodev,noexec,relatime,pids
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/devices cgroup
rw,nosuid,nodev,noexec,relatime,devices
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/cpuset cgroup
rw,nosuid,nodev,noexec,relatime,cpuset
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/blkio cgroup
rw,nosuid,nodev,noexec,relatime,blkio
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/perf_event cgroup
>>>>>> rw,nosuid,nodev,noexec,relatime,perf_event 0 0
>>>>>> cgroup /sys/fs/cgroup/memory cgroup
rw,nosuid,nodev,noexec,relatime,memory
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/freezer cgroup
rw,nosuid,nodev,noexec,relatime,freezer
>>>>>> 0 0
>>>>>> cgroup /sys/fs/cgroup/hugetlb cgroup
rw,nosuid,nodev,noexec,relatime,hugetlb
>>>>>> 0 0
>>>>>> configfs /sys/kernel/config configfs rw,relatime 0 0
>>>>>> /dev/mapper/cl_ovhost1-root / xfs
rw,relatime,attr2,inode64,noquota
>>>>>> 0 0
>>>>>> systemd-1 /proc/sys/fs/binfmt_misc autofs
>>>>>> rw,relatime,fd=35,pgrp=1,timeout=300,minproto=5,maxproto=5,direct
0
>>>>>> 0
>>>>>> mqueue /dev/mqueue mqueue rw,relatime 0 0
>>>>>> debugfs /sys/kernel/debug debugfs rw,relatime 0 0
>>>>>> hugetlbfs /dev/hugepages hugetlbfs rw,relatime 0 0
>>>>>> tmpfs /tmp tmpfs rw 0 0
>>>>>> nfsd /proc/fs/nfsd nfsd rw,relatime 0 0
>>>>>> /dev/mapper/cl_ovhost1-home /home xfs
rw,relatime,attr2,inode64,noquota
>>>>>> 0 0
>>>>>> /dev/sda1 /boot xfs rw,relatime,attr2,inode64,noquota 0 0
>>>>>> sunrpc /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0
>>>>>> tmpfs /run/user/42 tmpfs rw,nosuid,nodev,relatime,size=
>>>>>> 13192948k,mode=700,uid=42,gid=42 0 0
>>>>>> gvfsd-fuse /run/user/42/gvfs fuse.gvfsd-fuse
>>>>>> rw,nosuid,nodev,relatime,user_id=42,group_id=42 0 0
>>>>>> fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0
>>>>>> ovhost2:/home/exports/defaultdata
/rhev/data-center/mnt/ovhost2:_home_exports_defaultdata
>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>> ovhost2:/home/exports/ISO
/rhev/data-center/mnt/ovhost2:_home_exports_ISO
>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>> ovhost2:/home/exports/data
/rhev/data-center/mnt/ovhost2:_home_exports_data
>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>> tmpfs /run/user/0 tmpfs
rw,nosuid,nodev,relatime,size=13192948k,mode=700
>>>>>> 0 0
>>>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>>> /rhev/data-center/mnt/_dev_dm-3 ext4
rw,nosuid,relatime,data=ordered
>>>>>> 0 0
>>>>>>
>>>>>> Thanks you for your help.
>>>>>>
>>>>>> Carl
>>>>>>
>>>>>>
>>>>>> On Tue, Mar 21, 2017 at 6:31 AM, Fred Rolland
<frolland(a)redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Can you provide the content of /proc/mounts after it has
being
>>>>>>> mounted by VDSM ?
>>>>>>>
>>>>>>> On Tue, Mar 21, 2017 at 12:28 PM, carl langlois <
>>>>>>> crl.langlois(a)gmail.com> wrote:
>>>>>>>
>>>>>>>> Here is the vdsm.log
>>>>>>>>
>>>>>>>>
>>>>>>>> jsonrpc.Executor/0::ERROR::2017-03-18
>>>>>>>>
08:23:48,317::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>>>> Could not connect to storageServer
>>>>>>>> Traceback (most recent call last):
>>>>>>>> File "/usr/share/vdsm/storage/hsm.py", line
2400, in
>>>>>>>> connectStorageServer
>>>>>>>> conObj.connect()
>>>>>>>> File
"/usr/share/vdsm/storage/storageServer.py", line 242, in
>>>>>>>> connect
>>>>>>>> self.getMountObj().getRecord().fs_file)
>>>>>>>> File
"/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>> line 260, in getRecord
>>>>>>>> (self.fs_spec, self.fs_file))
>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>>>
>>>>>>>>
>>>>>>>> thanks
>>>>>>>>
>>>>>>>> On Fri, Mar 17, 2017 at 3:06 PM, Fred Rolland
<frolland(a)redhat.com
>>>>>>>> > wrote:
>>>>>>>>
>>>>>>>>> Please send Vdsm log.
>>>>>>>>> Thanks
>>>>>>>>>
>>>>>>>>> On Fri, Mar 17, 2017 at 8:46 PM, carl langlois <
>>>>>>>>> crl.langlois(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> The link that you send is for NFS strorage but i
am trying to
>>>>>>>>>> add a POSIX compliant.
>>>>>>>>>>
>>>>>>>>>> [image: Inline image 1]
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> when i press okey it mount the disk to :
>>>>>>>>>>
>>>>>>>>>> [root@ovhost4 ~]# ls -al
/rhev/data-center/mnt/_dev_dm-4/
>>>>>>>>>> total 28
>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .
>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 ..
>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42
lost+found
>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12
.Trash-0
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> and doing a touch with vdsm user work
>>>>>>>>>>
>>>>>>>>>> [root@ovhost4 ~]# sudo -u vdsm touch
>>>>>>>>>> /rhev/data-center/mnt/_dev_dm-4/test
>>>>>>>>>> [root@ovhost4 ~]# ls -al
/rhev/data-center/mnt/_dev_dm-4/
>>>>>>>>>> total 28
>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 17 13:44 .
>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 ..
>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42
lost+found
>>>>>>>>>> -rw-r--r--. 1 vdsm kvm 0 Mar 17 13:44 test
>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12
.Trash-0
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> But it fail with a general exception error and
the storage does
>>>>>>>>>> not exist in ovirt
>>>>>>>>>>
>>>>>>>>>> any help would be appreciated.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Which log you need to see?
>>>>>>>>>>
>>>>>>>>>> Thanks
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Le jeu. 16 mars 2017 17:02, Fred Rolland
<frolland(a)redhat.com>
>>>>>>>>>> a écrit :
>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> Can you check if the folder permissions are
OK ?
>>>>>>>>>>> Check [1] for more details.
>>>>>>>>>>>
>>>>>>>>>>> Can you share more of the log ?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> [1]
https://www.ovirt.org/document
>>>>>>>>>>>
ation/how-to/troubleshooting/troubleshooting-nfs-storage-iss
>>>>>>>>>>> ues/
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Mar 16, 2017 at 7:49 PM, carl
langlois <
>>>>>>>>>>> crl.langlois(a)gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>> Hi Guys,
>>>>>>>>>>>
>>>>>>>>>>> I am trying to add a posix FS on one of my
host. Ovirt in
>>>>>>>>>>> actually mounting it but fail with
"Error while executing
>>>>>>>>>>> action Add Storage Connection: General
Exception"
>>>>>>>>>>>
>>>>>>>>>>> If i look in the vdsm.log i cant see
>>>>>>>>>>>
>>>>>>>>>>> sonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>
12:39:28,248::fileUtils::209::Storage.fileUtils::(createdir)
>>>>>>>>>>> Creating directory:
/rhev/data-center/mnt/_dev_dm-3 mode: None
>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>
12:39:28,248::fileUtils::218::Storage.fileUtils::(createdir)
>>>>>>>>>>> Using existing directory:
/rhev/data-center/mnt/_dev_dm-3
>>>>>>>>>>> jsonrpc.Executor/7::INFO::2017-03-16
>>>>>>>>>>>
12:39:28,248::mount::226::storage.Mount::(mount) mounting
>>>>>>>>>>> /dev/dm-3 at /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>
12:39:28,270::utils::871::storage.Mount::(stopwatch)
>>>>>>>>>>> /rhev/data-center/mnt/_dev_dm-3 mounted: 0.02
seconds
>>>>>>>>>>> jsonrpc.Executor/7::ERROR::2017-03-16
>>>>>>>>>>>
12:39:28,271::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>>>>>>> Could not connect to storageServer
>>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>>> File
"/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>>>>>>> connectStorageServer
>>>>>>>>>>> conObj.connect()
>>>>>>>>>>> File
"/usr/share/vdsm/storage/storageServer.py", line 242,
>>>>>>>>>>> in connect
>>>>>>>>>>> self.getMountObj().getRecord().fs_file)
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>>>>> line 260, in getRecord
>>>>>>>>>>> (self.fs_spec, self.fs_file))
>>>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not
exist
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> any help would be appreciated.
>>>>>>>>>>>
>>>>>>>>>>> Thanks
>>>>>>>>>>>
>>>>>>>>>>> CL
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>> Users mailing list
>>>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>