[ovirt-users] Adding posix compliant FS
carl langlois
crl.langlois at gmail.com
Wed Mar 22 10:43:40 UTC 2017
Thanks for your help. This is what i understand after seeing all the vms
being migrate to that hosts :-). So i have created a new data center with
local storage. My question on this is is it possible to migrate VM between
the 2 data center? also is it possible to share ISO/EXPORT domain between
the 2 data center.
Thanks
On Wed, Mar 22, 2017 at 3:40 AM, Fred Rolland <frolland at redhat.com> wrote:
> Is it a local disk ? If you want to use a local disk, it is not the way to
> do it.
> POSIX Storage Domain should be accessible from all the hosts in the
> cluster.
>
> On Tue, Mar 21, 2017 at 9:36 PM, carl langlois <crl.langlois at gmail.com>
> wrote:
>
>> Okey, i have manager to use the POSIX compliant FS.
>>
>> First thing i did was to remove any multipath stuff from the disk and
>> have a standard parttiton table i.e /dev/sdb1 (but i do not think that
>> realy help)
>> change block device(/dev/sdb1) group and owner to vdsm:kvm (did not do
>> the trick either got still permission denied)
>> create a directory in /rhev/data-center/mnt/_dev_sdb1 and set owner and
>> group to vdsm:kvm (this did the trick)
>>
>> So why did i had to create the last directory by hand to make it work?..i
>> my missing something?
>>
>> Thanks
>> Carl
>>
>>
>>
>>
>> On Tue, Mar 21, 2017 at 10:50 AM, Fred Rolland <frolland at redhat.com>
>> wrote:
>>
>>> Can you try :
>>>
>>> chown -R vdsm:kvm /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>
>>> On Tue, Mar 21, 2017 at 4:32 PM, carl langlois <crl.langlois at gmail.com>
>>> wrote:
>>>
>>>>
>>>> jsonrpc.Executor/7::WARNING::2017-03-21 09:27:40,099::outOfProcess::19
>>>> 3::Storage.oop::(validateAccess) Permission denied for directory:
>>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
>>>> with permissions:7
>>>> jsonrpc.Executor/7::INFO::2017-03-21 09:27:40,099::mount::233::storage.Mount::(umount)
>>>> unmounting /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5
>>>> 0026B726804F13B1
>>>> jsonrpc.Executor/7::DEBUG::2017-03-21 09:27:40,104::utils::871::storage.Mount::(stopwatch)
>>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1
>>>> unmounted: 0.00 seconds
>>>> jsonrpc.Executor/7::ERROR::2017-03-21 09:27:40,104::hsm::2403::Storage.HSM::(connectStorageServer)
>>>> Could not connect to storageServer
>>>> Traceback (most recent call last):
>>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>> connectStorageServer
>>>> conObj.connect()
>>>> File "/usr/share/vdsm/storage/storageServer.py", line 249, in connect
>>>> six.reraise(t, v, tb)
>>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, in connect
>>>> self.getMountObj().getRecord().fs_file)
>>>> File "/usr/share/vdsm/storage/fileSD.py", line 81, in
>>>> validateDirAccess
>>>> raise se.StorageServerAccessPermissionError(dirPath)
>>>> StorageServerAccessPermissionError: Permission settings on the
>>>> specified path do not allow access to the storage. Verify permission
>>>> settings on the specified storage path.: 'path =
>>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5
>>>> 0026B726804F13B1'
>>>> jsonrpc.Executor/7::DEBUG::201
>>>>
>>>> Thanks again.
>>>>
>>>>
>>>> On Tue, Mar 21, 2017 at 10:14 AM, Fred Rolland <frolland at redhat.com>
>>>> wrote:
>>>>
>>>>> Can you share the VDSM log again ?
>>>>>
>>>>> On Tue, Mar 21, 2017 at 4:08 PM, carl langlois <crl.langlois at gmail.com
>>>>> > wrote:
>>>>>
>>>>>> Interesting, when i'm using /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>>> now the UI give error on the permission setting..
>>>>>>
>>>>>> root at ovhost4 ~]# ls -al /dev/mapper/KINGSTON_SV300S37A
>>>>>> 240G_50026B726804F13B1
>>>>>> lrwxrwxrwx 1 root root 7 Mar 18 08:28 /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>>> -> ../dm-3
>>>>>>
>>>>>> and the permission on the dm-3
>>>>>>
>>>>>> [root at ovhost4 ~]# ls -al /dev/dm-3
>>>>>> brw-rw---- 1 vdsm kvm 253, 3 Mar 18 08:28 /dev/dm-3
>>>>>>
>>>>>>
>>>>>> how do i change the permission on the sym link..
>>>>>>
>>>>>> Thanks
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Mar 21, 2017 at 10:00 AM, Fred Rolland <frolland at redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Can you try to use /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>>>> in the UI.
>>>>>>> It seems the kernel change the path that we use to mount and then we
>>>>>>> cannot validate that the mount exists.
>>>>>>>
>>>>>>> It should be anyway better as the mapping could change after reboot.
>>>>>>>
>>>>>>> On Tue, Mar 21, 2017 at 2:20 PM, carl langlois <
>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>
>>>>>>>> Here is the /proc/mounts
>>>>>>>>
>>>>>>>> rootfs / rootfs rw 0 0
>>>>>>>> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0
>>>>>>>> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0
>>>>>>>> devtmpfs /dev devtmpfs rw,nosuid,size=65948884k,nr_inodes=16487221,mode=755
>>>>>>>> 0 0
>>>>>>>> securityfs /sys/kernel/security securityfs
>>>>>>>> rw,nosuid,nodev,noexec,relatime 0 0
>>>>>>>> tmpfs /dev/shm tmpfs rw,nosuid,nodev 0 0
>>>>>>>> devpts /dev/pts devpts rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000
>>>>>>>> 0 0
>>>>>>>> tmpfs /run tmpfs rw,nosuid,nodev,mode=755 0 0
>>>>>>>> tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0
>>>>>>>> cgroup /sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatim
>>>>>>>> e,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd
>>>>>>>> 0 0
>>>>>>>> pstore /sys/fs/pstore pstore rw,nosuid,nodev,noexec,relatime 0 0
>>>>>>>> cgroup /sys/fs/cgroup/cpu,cpuacct cgroup
>>>>>>>> rw,nosuid,nodev,noexec,relatime,cpuacct,cpu 0 0
>>>>>>>> cgroup /sys/fs/cgroup/net_cls,net_prio cgroup
>>>>>>>> rw,nosuid,nodev,noexec,relatime,net_prio,net_cls 0 0
>>>>>>>> cgroup /sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids
>>>>>>>> 0 0
>>>>>>>> cgroup /sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices
>>>>>>>> 0 0
>>>>>>>> cgroup /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset
>>>>>>>> 0 0
>>>>>>>> cgroup /sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio
>>>>>>>> 0 0
>>>>>>>> cgroup /sys/fs/cgroup/perf_event cgroup
>>>>>>>> rw,nosuid,nodev,noexec,relatime,perf_event 0 0
>>>>>>>> cgroup /sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory
>>>>>>>> 0 0
>>>>>>>> cgroup /sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer
>>>>>>>> 0 0
>>>>>>>> cgroup /sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb
>>>>>>>> 0 0
>>>>>>>> configfs /sys/kernel/config configfs rw,relatime 0 0
>>>>>>>> /dev/mapper/cl_ovhost1-root / xfs rw,relatime,attr2,inode64,noquota
>>>>>>>> 0 0
>>>>>>>> systemd-1 /proc/sys/fs/binfmt_misc autofs
>>>>>>>> rw,relatime,fd=35,pgrp=1,timeout=300,minproto=5,maxproto=5,direct
>>>>>>>> 0 0
>>>>>>>> mqueue /dev/mqueue mqueue rw,relatime 0 0
>>>>>>>> debugfs /sys/kernel/debug debugfs rw,relatime 0 0
>>>>>>>> hugetlbfs /dev/hugepages hugetlbfs rw,relatime 0 0
>>>>>>>> tmpfs /tmp tmpfs rw 0 0
>>>>>>>> nfsd /proc/fs/nfsd nfsd rw,relatime 0 0
>>>>>>>> /dev/mapper/cl_ovhost1-home /home xfs rw,relatime,attr2,inode64,noquota
>>>>>>>> 0 0
>>>>>>>> /dev/sda1 /boot xfs rw,relatime,attr2,inode64,noquota 0 0
>>>>>>>> sunrpc /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0
>>>>>>>> tmpfs /run/user/42 tmpfs rw,nosuid,nodev,relatime,size=
>>>>>>>> 13192948k,mode=700,uid=42,gid=42 0 0
>>>>>>>> gvfsd-fuse /run/user/42/gvfs fuse.gvfsd-fuse
>>>>>>>> rw,nosuid,nodev,relatime,user_id=42,group_id=42 0 0
>>>>>>>> fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0
>>>>>>>> ovhost2:/home/exports/defaultdata /rhev/data-center/mnt/ovhost2:_home_exports_defaultdata
>>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>>>> ovhost2:/home/exports/ISO /rhev/data-center/mnt/ovhost2:_home_exports_ISO
>>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>>>> ovhost2:/home/exports/data /rhev/data-center/mnt/ovhost2:_home_exports_data
>>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp,
>>>>>>>> local_lock=none,addr=10.8.236.162 0 0
>>>>>>>> tmpfs /run/user/0 tmpfs rw,nosuid,nodev,relatime,size=13192948k,mode=700
>>>>>>>> 0 0
>>>>>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
>>>>>>>> /rhev/data-center/mnt/_dev_dm-3 ext4 rw,nosuid,relatime,data=ordered
>>>>>>>> 0 0
>>>>>>>>
>>>>>>>> Thanks you for your help.
>>>>>>>>
>>>>>>>> Carl
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Mar 21, 2017 at 6:31 AM, Fred Rolland <frolland at redhat.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Can you provide the content of /proc/mounts after it has being
>>>>>>>>> mounted by VDSM ?
>>>>>>>>>
>>>>>>>>> On Tue, Mar 21, 2017 at 12:28 PM, carl langlois <
>>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Here is the vdsm.log
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> jsonrpc.Executor/0::ERROR::2017-03-18
>>>>>>>>>> 08:23:48,317::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>>>>>> Could not connect to storageServer
>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>>>>>> connectStorageServer
>>>>>>>>>> conObj.connect()
>>>>>>>>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, in
>>>>>>>>>> connect
>>>>>>>>>> self.getMountObj().getRecord().fs_file)
>>>>>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>>>> line 260, in getRecord
>>>>>>>>>> (self.fs_spec, self.fs_file))
>>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> thanks
>>>>>>>>>>
>>>>>>>>>> On Fri, Mar 17, 2017 at 3:06 PM, Fred Rolland <
>>>>>>>>>> frolland at redhat.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Please send Vdsm log.
>>>>>>>>>>> Thanks
>>>>>>>>>>>
>>>>>>>>>>> On Fri, Mar 17, 2017 at 8:46 PM, carl langlois <
>>>>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi,
>>>>>>>>>>>>
>>>>>>>>>>>> The link that you send is for NFS strorage but i am trying to
>>>>>>>>>>>> add a POSIX compliant.
>>>>>>>>>>>>
>>>>>>>>>>>> [image: Inline image 1]
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> when i press okey it mount the disk to :
>>>>>>>>>>>>
>>>>>>>>>>>> [root at ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/
>>>>>>>>>>>> total 28
>>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .
>>>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 ..
>>>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found
>>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .Trash-0
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> and doing a touch with vdsm user work
>>>>>>>>>>>>
>>>>>>>>>>>> [root at ovhost4 ~]# sudo -u vdsm touch
>>>>>>>>>>>> /rhev/data-center/mnt/_dev_dm-4/test
>>>>>>>>>>>> [root at ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/
>>>>>>>>>>>> total 28
>>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 17 13:44 .
>>>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 ..
>>>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found
>>>>>>>>>>>> -rw-r--r--. 1 vdsm kvm 0 Mar 17 13:44 test
>>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .Trash-0
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> But it fail with a general exception error and the storage does
>>>>>>>>>>>> not exist in ovirt
>>>>>>>>>>>>
>>>>>>>>>>>> any help would be appreciated.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Which log you need to see?
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Le jeu. 16 mars 2017 17:02, Fred Rolland <frolland at redhat.com>
>>>>>>>>>>>> a écrit :
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>
>>>>>>>>>>>>> Can you check if the folder permissions are OK ?
>>>>>>>>>>>>> Check [1] for more details.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Can you share more of the log ?
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> [1] https://www.ovirt.org/document
>>>>>>>>>>>>> ation/how-to/troubleshooting/troubleshooting-nfs-storage-iss
>>>>>>>>>>>>> ues/
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Thu, Mar 16, 2017 at 7:49 PM, carl langlois <
>>>>>>>>>>>>> crl.langlois at gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>> Hi Guys,
>>>>>>>>>>>>>
>>>>>>>>>>>>> I am trying to add a posix FS on one of my host. Ovirt in
>>>>>>>>>>>>> actually mounting it but fail with "Error while executing
>>>>>>>>>>>>> action Add Storage Connection: General Exception"
>>>>>>>>>>>>>
>>>>>>>>>>>>> If i look in the vdsm.log i cant see
>>>>>>>>>>>>>
>>>>>>>>>>>>> sonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>>> 12:39:28,248::fileUtils::209::Storage.fileUtils::(createdir)
>>>>>>>>>>>>> Creating directory: /rhev/data-center/mnt/_dev_dm-3 mode: None
>>>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>>> 12:39:28,248::fileUtils::218::Storage.fileUtils::(createdir)
>>>>>>>>>>>>> Using existing directory: /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>>>>>> jsonrpc.Executor/7::INFO::2017-03-16
>>>>>>>>>>>>> 12:39:28,248::mount::226::storage.Mount::(mount) mounting
>>>>>>>>>>>>> /dev/dm-3 at /rhev/data-center/mnt/_dev_dm-3
>>>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16
>>>>>>>>>>>>> 12:39:28,270::utils::871::storage.Mount::(stopwatch)
>>>>>>>>>>>>> /rhev/data-center/mnt/_dev_dm-3 mounted: 0.02 seconds
>>>>>>>>>>>>> jsonrpc.Executor/7::ERROR::2017-03-16
>>>>>>>>>>>>> 12:39:28,271::hsm::2403::Storage.HSM::(connectStorageServer)
>>>>>>>>>>>>> Could not connect to storageServer
>>>>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>>>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in
>>>>>>>>>>>>> connectStorageServer
>>>>>>>>>>>>> conObj.connect()
>>>>>>>>>>>>> File "/usr/share/vdsm/storage/storageServer.py", line 242,
>>>>>>>>>>>>> in connect
>>>>>>>>>>>>> self.getMountObj().getRecord().fs_file)
>>>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py",
>>>>>>>>>>>>> line 260, in getRecord
>>>>>>>>>>>>> (self.fs_spec, self.fs_file))
>>>>>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at
>>>>>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> any help would be appreciated.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>
>>>>>>>>>>>>> CL
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>> Users mailing list
>>>>>>>>>>>>> Users at ovirt.org
>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170322/de6575fd/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image.png
Type: image/png
Size: 27276 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170322/de6575fd/attachment-0001.png>
More information about the Users
mailing list