
Thanks for your help. This is what i understand after seeing all the vms being migrate to that hosts :-). So i have created a new data center with local storage. My question on this is is it possible to migrate VM between the 2 data center? also is it possible to share ISO/EXPORT domain between the 2 data center. Thanks On Wed, Mar 22, 2017 at 3:40 AM, Fred Rolland <frolland@redhat.com> wrote:
Is it a local disk ? If you want to use a local disk, it is not the way to do it. POSIX Storage Domain should be accessible from all the hosts in the cluster.
On Tue, Mar 21, 2017 at 9:36 PM, carl langlois <crl.langlois@gmail.com> wrote:
Okey, i have manager to use the POSIX compliant FS.
First thing i did was to remove any multipath stuff from the disk and have a standard parttiton table i.e /dev/sdb1 (but i do not think that realy help) change block device(/dev/sdb1) group and owner to vdsm:kvm (did not do the trick either got still permission denied) create a directory in /rhev/data-center/mnt/_dev_sdb1 and set owner and group to vdsm:kvm (this did the trick)
So why did i had to create the last directory by hand to make it work?..i my missing something?
Thanks Carl
On Tue, Mar 21, 2017 at 10:50 AM, Fred Rolland <frolland@redhat.com> wrote:
Can you try :
chown -R vdsm:kvm /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
On Tue, Mar 21, 2017 at 4:32 PM, carl langlois <crl.langlois@gmail.com> wrote:
jsonrpc.Executor/7::WARNING::2017-03-21 09:27:40,099::outOfProcess::19 3::Storage.oop::(validateAccess) Permission denied for directory: /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1 with permissions:7 jsonrpc.Executor/7::INFO::2017-03-21 09:27:40,099::mount::233::storage.Mount::(umount) unmounting /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5 0026B726804F13B1 jsonrpc.Executor/7::DEBUG::2017-03-21 09:27:40,104::utils::871::storage.Mount::(stopwatch) /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1 unmounted: 0.00 seconds jsonrpc.Executor/7::ERROR::2017-03-21 09:27:40,104::hsm::2403::Storage.HSM::(connectStorageServer) Could not connect to storageServer Traceback (most recent call last): File "/usr/share/vdsm/storage/hsm.py", line 2400, in connectStorageServer conObj.connect() File "/usr/share/vdsm/storage/storageServer.py", line 249, in connect six.reraise(t, v, tb) File "/usr/share/vdsm/storage/storageServer.py", line 242, in connect self.getMountObj().getRecord().fs_file) File "/usr/share/vdsm/storage/fileSD.py", line 81, in validateDirAccess raise se.StorageServerAccessPermissionError(dirPath) StorageServerAccessPermissionError: Permission settings on the specified path do not allow access to the storage. Verify permission settings on the specified storage path.: 'path = /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5 0026B726804F13B1' jsonrpc.Executor/7::DEBUG::201
Thanks again.
On Tue, Mar 21, 2017 at 10:14 AM, Fred Rolland <frolland@redhat.com> wrote:
Can you share the VDSM log again ?
On Tue, Mar 21, 2017 at 4:08 PM, carl langlois <crl.langlois@gmail.com
wrote:
Interesting, when i'm using /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 now the UI give error on the permission setting..
root@ovhost4 ~]# ls -al /dev/mapper/KINGSTON_SV300S37A 240G_50026B726804F13B1 lrwxrwxrwx 1 root root 7 Mar 18 08:28 /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 -> ../dm-3
and the permission on the dm-3
[root@ovhost4 ~]# ls -al /dev/dm-3 brw-rw---- 1 vdsm kvm 253, 3 Mar 18 08:28 /dev/dm-3
how do i change the permission on the sym link..
Thanks
On Tue, Mar 21, 2017 at 10:00 AM, Fred Rolland <frolland@redhat.com> wrote:
> Can you try to use /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 > in the UI. > It seems the kernel change the path that we use to mount and then we > cannot validate that the mount exists. > > It should be anyway better as the mapping could change after reboot. > > On Tue, Mar 21, 2017 at 2:20 PM, carl langlois < > crl.langlois@gmail.com> wrote: > >> Here is the /proc/mounts >> >> rootfs / rootfs rw 0 0 >> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0 >> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0 >> devtmpfs /dev devtmpfs rw,nosuid,size=65948884k,nr_inodes=16487221,mode=755 >> 0 0 >> securityfs /sys/kernel/security securityfs >> rw,nosuid,nodev,noexec,relatime 0 0 >> tmpfs /dev/shm tmpfs rw,nosuid,nodev 0 0 >> devpts /dev/pts devpts rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000 >> 0 0 >> tmpfs /run tmpfs rw,nosuid,nodev,mode=755 0 0 >> tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0 >> cgroup /sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatim >> e,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd >> 0 0 >> pstore /sys/fs/pstore pstore rw,nosuid,nodev,noexec,relatime 0 0 >> cgroup /sys/fs/cgroup/cpu,cpuacct cgroup >> rw,nosuid,nodev,noexec,relatime,cpuacct,cpu 0 0 >> cgroup /sys/fs/cgroup/net_cls,net_prio cgroup >> rw,nosuid,nodev,noexec,relatime,net_prio,net_cls 0 0 >> cgroup /sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids >> 0 0 >> cgroup /sys/fs/cgroup/devices cgroup rw,nosuid,nodev,noexec,relatime,devices >> 0 0 >> cgroup /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset >> 0 0 >> cgroup /sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio >> 0 0 >> cgroup /sys/fs/cgroup/perf_event cgroup >> rw,nosuid,nodev,noexec,relatime,perf_event 0 0 >> cgroup /sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory >> 0 0 >> cgroup /sys/fs/cgroup/freezer cgroup rw,nosuid,nodev,noexec,relatime,freezer >> 0 0 >> cgroup /sys/fs/cgroup/hugetlb cgroup rw,nosuid,nodev,noexec,relatime,hugetlb >> 0 0 >> configfs /sys/kernel/config configfs rw,relatime 0 0 >> /dev/mapper/cl_ovhost1-root / xfs rw,relatime,attr2,inode64,noquota >> 0 0 >> systemd-1 /proc/sys/fs/binfmt_misc autofs >> rw,relatime,fd=35,pgrp=1,timeout=300,minproto=5,maxproto=5,direct >> 0 0 >> mqueue /dev/mqueue mqueue rw,relatime 0 0 >> debugfs /sys/kernel/debug debugfs rw,relatime 0 0 >> hugetlbfs /dev/hugepages hugetlbfs rw,relatime 0 0 >> tmpfs /tmp tmpfs rw 0 0 >> nfsd /proc/fs/nfsd nfsd rw,relatime 0 0 >> /dev/mapper/cl_ovhost1-home /home xfs rw,relatime,attr2,inode64,noquota >> 0 0 >> /dev/sda1 /boot xfs rw,relatime,attr2,inode64,noquota 0 0 >> sunrpc /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0 >> tmpfs /run/user/42 tmpfs rw,nosuid,nodev,relatime,size= >> 13192948k,mode=700,uid=42,gid=42 0 0 >> gvfsd-fuse /run/user/42/gvfs fuse.gvfsd-fuse >> rw,nosuid,nodev,relatime,user_id=42,group_id=42 0 0 >> fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0 >> ovhost2:/home/exports/defaultdata /rhev/data-center/mnt/ovhost2:_home_exports_defaultdata >> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >> local_lock=none,addr=10.8.236.162 0 0 >> ovhost2:/home/exports/ISO /rhev/data-center/mnt/ovhost2:_home_exports_ISO >> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >> local_lock=none,addr=10.8.236.162 0 0 >> ovhost2:/home/exports/data /rhev/data-center/mnt/ovhost2:_home_exports_data >> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >> local_lock=none,addr=10.8.236.162 0 0 >> tmpfs /run/user/0 tmpfs rw,nosuid,nodev,relatime,size=13192948k,mode=700 >> 0 0 >> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >> /rhev/data-center/mnt/_dev_dm-3 ext4 rw,nosuid,relatime,data=ordered >> 0 0 >> >> Thanks you for your help. >> >> Carl >> >> >> On Tue, Mar 21, 2017 at 6:31 AM, Fred Rolland <frolland@redhat.com> >> wrote: >> >>> Can you provide the content of /proc/mounts after it has being >>> mounted by VDSM ? >>> >>> On Tue, Mar 21, 2017 at 12:28 PM, carl langlois < >>> crl.langlois@gmail.com> wrote: >>> >>>> Here is the vdsm.log >>>> >>>> >>>> jsonrpc.Executor/0::ERROR::2017-03-18 >>>> 08:23:48,317::hsm::2403::Storage.HSM::(connectStorageServer) >>>> Could not connect to storageServer >>>> Traceback (most recent call last): >>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in >>>> connectStorageServer >>>> conObj.connect() >>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, in >>>> connect >>>> self.getMountObj().getRecord().fs_file) >>>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", >>>> line 260, in getRecord >>>> (self.fs_spec, self.fs_file)) >>>> OSError: [Errno 2] Mount of `/dev/dm-3` at >>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist >>>> >>>> >>>> thanks >>>> >>>> On Fri, Mar 17, 2017 at 3:06 PM, Fred Rolland < >>>> frolland@redhat.com> wrote: >>>> >>>>> Please send Vdsm log. >>>>> Thanks >>>>> >>>>> On Fri, Mar 17, 2017 at 8:46 PM, carl langlois < >>>>> crl.langlois@gmail.com> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> The link that you send is for NFS strorage but i am trying to >>>>>> add a POSIX compliant. >>>>>> >>>>>> [image: Inline image 1] >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> when i press okey it mount the disk to : >>>>>> >>>>>> [root@ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/ >>>>>> total 28 >>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 . >>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 .. >>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found >>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .Trash-0 >>>>>> >>>>>> >>>>>> and doing a touch with vdsm user work >>>>>> >>>>>> [root@ovhost4 ~]# sudo -u vdsm touch >>>>>> /rhev/data-center/mnt/_dev_dm-4/test >>>>>> [root@ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/ >>>>>> total 28 >>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 17 13:44 . >>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 .. >>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found >>>>>> -rw-r--r--. 1 vdsm kvm 0 Mar 17 13:44 test >>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .Trash-0 >>>>>> >>>>>> >>>>>> But it fail with a general exception error and the storage does >>>>>> not exist in ovirt >>>>>> >>>>>> any help would be appreciated. >>>>>> >>>>>> >>>>>> Which log you need to see? >>>>>> >>>>>> Thanks >>>>>> >>>>>> >>>>>> >>>>>> Le jeu. 16 mars 2017 17:02, Fred Rolland <frolland@redhat.com> >>>>>> a écrit : >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> Can you check if the folder permissions are OK ? >>>>>>> Check [1] for more details. >>>>>>> >>>>>>> Can you share more of the log ? >>>>>>> >>>>>>> >>>>>>> [1] https://www.ovirt.org/document >>>>>>> ation/how-to/troubleshooting/troubleshooting-nfs-storage-iss >>>>>>> ues/ >>>>>>> >>>>>>> On Thu, Mar 16, 2017 at 7:49 PM, carl langlois < >>>>>>> crl.langlois@gmail.com> wrote: >>>>>>> >>>>>>> Hi Guys, >>>>>>> >>>>>>> I am trying to add a posix FS on one of my host. Ovirt in >>>>>>> actually mounting it but fail with "Error while executing >>>>>>> action Add Storage Connection: General Exception" >>>>>>> >>>>>>> If i look in the vdsm.log i cant see >>>>>>> >>>>>>> sonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>> 12:39:28,248::fileUtils::209::Storage.fileUtils::(createdir) >>>>>>> Creating directory: /rhev/data-center/mnt/_dev_dm-3 mode: None >>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>> 12:39:28,248::fileUtils::218::Storage.fileUtils::(createdir) >>>>>>> Using existing directory: /rhev/data-center/mnt/_dev_dm-3 >>>>>>> jsonrpc.Executor/7::INFO::2017-03-16 >>>>>>> 12:39:28,248::mount::226::storage.Mount::(mount) mounting >>>>>>> /dev/dm-3 at /rhev/data-center/mnt/_dev_dm-3 >>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>> 12:39:28,270::utils::871::storage.Mount::(stopwatch) >>>>>>> /rhev/data-center/mnt/_dev_dm-3 mounted: 0.02 seconds >>>>>>> jsonrpc.Executor/7::ERROR::2017-03-16 >>>>>>> 12:39:28,271::hsm::2403::Storage.HSM::(connectStorageServer) >>>>>>> Could not connect to storageServer >>>>>>> Traceback (most recent call last): >>>>>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in >>>>>>> connectStorageServer >>>>>>> conObj.connect() >>>>>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, >>>>>>> in connect >>>>>>> self.getMountObj().getRecord().fs_file) >>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", >>>>>>> line 260, in getRecord >>>>>>> (self.fs_spec, self.fs_file)) >>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at >>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist >>>>>>> >>>>>>> >>>>>>> any help would be appreciated. >>>>>>> >>>>>>> Thanks >>>>>>> >>>>>>> CL >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Users mailing list >>>>>>> Users@ovirt.org >>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>> >>>>>>> >>>>>>> >>>>> >>>> >>> >> >