Thanks for your help. This is what i understand after seeing all the vms being migrate to that hosts :-). So i have created a new data center with local storage. My question on this is is it possible to migrate VM between the 2 data center? also is it possible to share ISO/EXPORT domain between the 2 data center.
Thanks On Wed, Mar 22, 2017 at 3:40 AM, Fred Rolland <[email protected]> wrote: > Is it a local disk ? If you want to use a local disk, it is not the way to > do it. > POSIX Storage Domain should be accessible from all the hosts in the > cluster. > > On Tue, Mar 21, 2017 at 9:36 PM, carl langlois <[email protected]> > wrote: > >> Okey, i have manager to use the POSIX compliant FS. >> >> First thing i did was to remove any multipath stuff from the disk and >> have a standard parttiton table i.e /dev/sdb1 (but i do not think that >> realy help) >> change block device(/dev/sdb1) group and owner to vdsm:kvm (did not do >> the trick either got still permission denied) >> create a directory in /rhev/data-center/mnt/_dev_sdb1 and set owner and >> group to vdsm:kvm (this did the trick) >> >> So why did i had to create the last directory by hand to make it work?..i >> my missing something? >> >> Thanks >> Carl >> >> >> >> >> On Tue, Mar 21, 2017 at 10:50 AM, Fred Rolland <[email protected]> >> wrote: >> >>> Can you try : >>> >>> chown -R vdsm:kvm /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>> >>> On Tue, Mar 21, 2017 at 4:32 PM, carl langlois <[email protected]> >>> wrote: >>> >>>> >>>> jsonrpc.Executor/7::WARNING::2017-03-21 09:27:40,099::outOfProcess::19 >>>> 3::Storage.oop::(validateAccess) Permission denied for directory: >>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1 >>>> with permissions:7 >>>> jsonrpc.Executor/7::INFO::2017-03-21 >>>> 09:27:40,099::mount::233::storage.Mount::(umount) >>>> unmounting /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5 >>>> 0026B726804F13B1 >>>> jsonrpc.Executor/7::DEBUG::2017-03-21 >>>> 09:27:40,104::utils::871::storage.Mount::(stopwatch) >>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1 >>>> unmounted: 0.00 seconds >>>> jsonrpc.Executor/7::ERROR::2017-03-21 >>>> 09:27:40,104::hsm::2403::Storage.HSM::(connectStorageServer) >>>> Could not connect to storageServer >>>> Traceback (most recent call last): >>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in >>>> connectStorageServer >>>> conObj.connect() >>>> File "/usr/share/vdsm/storage/storageServer.py", line 249, in connect >>>> six.reraise(t, v, tb) >>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, in connect >>>> self.getMountObj().getRecord().fs_file) >>>> File "/usr/share/vdsm/storage/fileSD.py", line 81, in >>>> validateDirAccess >>>> raise se.StorageServerAccessPermissionError(dirPath) >>>> StorageServerAccessPermissionError: Permission settings on the >>>> specified path do not allow access to the storage. Verify permission >>>> settings on the specified storage path.: 'path = >>>> /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__5 >>>> 0026B726804F13B1' >>>> jsonrpc.Executor/7::DEBUG::201 >>>> >>>> Thanks again. >>>> >>>> >>>> On Tue, Mar 21, 2017 at 10:14 AM, Fred Rolland <[email protected]> >>>> wrote: >>>> >>>>> Can you share the VDSM log again ? >>>>> >>>>> On Tue, Mar 21, 2017 at 4:08 PM, carl langlois <[email protected] >>>>> > wrote: >>>>> >>>>>> Interesting, when i'm using >>>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>>>>> now the UI give error on the permission setting.. >>>>>> >>>>>> root@ovhost4 ~]# ls -al /dev/mapper/KINGSTON_SV300S37A >>>>>> 240G_50026B726804F13B1 >>>>>> lrwxrwxrwx 1 root root 7 Mar 18 08:28 >>>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>>>>> -> ../dm-3 >>>>>> >>>>>> and the permission on the dm-3 >>>>>> >>>>>> [root@ovhost4 ~]# ls -al /dev/dm-3 >>>>>> brw-rw---- 1 vdsm kvm 253, 3 Mar 18 08:28 /dev/dm-3 >>>>>> >>>>>> >>>>>> how do i change the permission on the sym link.. >>>>>> >>>>>> Thanks >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Tue, Mar 21, 2017 at 10:00 AM, Fred Rolland <[email protected]> >>>>>> wrote: >>>>>> >>>>>>> Can you try to use /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>>>>>> in the UI. >>>>>>> It seems the kernel change the path that we use to mount and then we >>>>>>> cannot validate that the mount exists. >>>>>>> >>>>>>> It should be anyway better as the mapping could change after reboot. >>>>>>> >>>>>>> On Tue, Mar 21, 2017 at 2:20 PM, carl langlois < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> Here is the /proc/mounts >>>>>>>> >>>>>>>> rootfs / rootfs rw 0 0 >>>>>>>> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0 >>>>>>>> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0 >>>>>>>> devtmpfs /dev devtmpfs >>>>>>>> rw,nosuid,size=65948884k,nr_inodes=16487221,mode=755 >>>>>>>> 0 0 >>>>>>>> securityfs /sys/kernel/security securityfs >>>>>>>> rw,nosuid,nodev,noexec,relatime 0 0 >>>>>>>> tmpfs /dev/shm tmpfs rw,nosuid,nodev 0 0 >>>>>>>> devpts /dev/pts devpts >>>>>>>> rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000 >>>>>>>> 0 0 >>>>>>>> tmpfs /run tmpfs rw,nosuid,nodev,mode=755 0 0 >>>>>>>> tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0 >>>>>>>> cgroup /sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatim >>>>>>>> e,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd >>>>>>>> 0 0 >>>>>>>> pstore /sys/fs/pstore pstore rw,nosuid,nodev,noexec,relatime 0 0 >>>>>>>> cgroup /sys/fs/cgroup/cpu,cpuacct cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,cpuacct,cpu 0 0 >>>>>>>> cgroup /sys/fs/cgroup/net_cls,net_prio cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,net_prio,net_cls 0 0 >>>>>>>> cgroup /sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids >>>>>>>> 0 0 >>>>>>>> cgroup /sys/fs/cgroup/devices cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,devices >>>>>>>> 0 0 >>>>>>>> cgroup /sys/fs/cgroup/cpuset cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,cpuset >>>>>>>> 0 0 >>>>>>>> cgroup /sys/fs/cgroup/blkio cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,blkio >>>>>>>> 0 0 >>>>>>>> cgroup /sys/fs/cgroup/perf_event cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,perf_event 0 0 >>>>>>>> cgroup /sys/fs/cgroup/memory cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,memory >>>>>>>> 0 0 >>>>>>>> cgroup /sys/fs/cgroup/freezer cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,freezer >>>>>>>> 0 0 >>>>>>>> cgroup /sys/fs/cgroup/hugetlb cgroup >>>>>>>> rw,nosuid,nodev,noexec,relatime,hugetlb >>>>>>>> 0 0 >>>>>>>> configfs /sys/kernel/config configfs rw,relatime 0 0 >>>>>>>> /dev/mapper/cl_ovhost1-root / xfs rw,relatime,attr2,inode64,noquota >>>>>>>> 0 0 >>>>>>>> systemd-1 /proc/sys/fs/binfmt_misc autofs >>>>>>>> rw,relatime,fd=35,pgrp=1,timeout=300,minproto=5,maxproto=5,direct >>>>>>>> 0 0 >>>>>>>> mqueue /dev/mqueue mqueue rw,relatime 0 0 >>>>>>>> debugfs /sys/kernel/debug debugfs rw,relatime 0 0 >>>>>>>> hugetlbfs /dev/hugepages hugetlbfs rw,relatime 0 0 >>>>>>>> tmpfs /tmp tmpfs rw 0 0 >>>>>>>> nfsd /proc/fs/nfsd nfsd rw,relatime 0 0 >>>>>>>> /dev/mapper/cl_ovhost1-home /home xfs rw,relatime,attr2,inode64,noquota >>>>>>>> 0 0 >>>>>>>> /dev/sda1 /boot xfs rw,relatime,attr2,inode64,noquota 0 0 >>>>>>>> sunrpc /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0 >>>>>>>> tmpfs /run/user/42 tmpfs rw,nosuid,nodev,relatime,size= >>>>>>>> 13192948k,mode=700,uid=42,gid=42 0 0 >>>>>>>> gvfsd-fuse /run/user/42/gvfs fuse.gvfsd-fuse >>>>>>>> rw,nosuid,nodev,relatime,user_id=42,group_id=42 0 0 >>>>>>>> fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0 >>>>>>>> ovhost2:/home/exports/defaultdata >>>>>>>> /rhev/data-center/mnt/ovhost2:_home_exports_defaultdata >>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >>>>>>>> local_lock=none,addr=10.8.236.162 0 0 >>>>>>>> ovhost2:/home/exports/ISO >>>>>>>> /rhev/data-center/mnt/ovhost2:_home_exports_ISO >>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >>>>>>>> local_lock=none,addr=10.8.236.162 0 0 >>>>>>>> ovhost2:/home/exports/data >>>>>>>> /rhev/data-center/mnt/ovhost2:_home_exports_data >>>>>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >>>>>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >>>>>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >>>>>>>> local_lock=none,addr=10.8.236.162 0 0 >>>>>>>> tmpfs /run/user/0 tmpfs >>>>>>>> rw,nosuid,nodev,relatime,size=13192948k,mode=700 >>>>>>>> 0 0 >>>>>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>>>>>>> /rhev/data-center/mnt/_dev_dm-3 ext4 rw,nosuid,relatime,data=ordered >>>>>>>> 0 0 >>>>>>>> >>>>>>>> Thanks you for your help. >>>>>>>> >>>>>>>> Carl >>>>>>>> >>>>>>>> >>>>>>>> On Tue, Mar 21, 2017 at 6:31 AM, Fred Rolland <[email protected]> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Can you provide the content of /proc/mounts after it has being >>>>>>>>> mounted by VDSM ? >>>>>>>>> >>>>>>>>> On Tue, Mar 21, 2017 at 12:28 PM, carl langlois < >>>>>>>>> [email protected]> wrote: >>>>>>>>> >>>>>>>>>> Here is the vdsm.log >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> jsonrpc.Executor/0::ERROR::2017-03-18 >>>>>>>>>> 08:23:48,317::hsm::2403::Storage.HSM::(connectStorageServer) >>>>>>>>>> Could not connect to storageServer >>>>>>>>>> Traceback (most recent call last): >>>>>>>>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in >>>>>>>>>> connectStorageServer >>>>>>>>>> conObj.connect() >>>>>>>>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, in >>>>>>>>>> connect >>>>>>>>>> self.getMountObj().getRecord().fs_file) >>>>>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", >>>>>>>>>> line 260, in getRecord >>>>>>>>>> (self.fs_spec, self.fs_file)) >>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at >>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> thanks >>>>>>>>>> >>>>>>>>>> On Fri, Mar 17, 2017 at 3:06 PM, Fred Rolland < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> Please send Vdsm log. >>>>>>>>>>> Thanks >>>>>>>>>>> >>>>>>>>>>> On Fri, Mar 17, 2017 at 8:46 PM, carl langlois < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> The link that you send is for NFS strorage but i am trying to >>>>>>>>>>>> add a POSIX compliant. >>>>>>>>>>>> >>>>>>>>>>>> [image: Inline image 1] >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> when i press okey it mount the disk to : >>>>>>>>>>>> >>>>>>>>>>>> [root@ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/ >>>>>>>>>>>> total 28 >>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 . >>>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 .. >>>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found >>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .Trash-0 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> and doing a touch with vdsm user work >>>>>>>>>>>> >>>>>>>>>>>> [root@ovhost4 ~]# sudo -u vdsm touch >>>>>>>>>>>> /rhev/data-center/mnt/_dev_dm-4/test >>>>>>>>>>>> [root@ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/ >>>>>>>>>>>> total 28 >>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 17 13:44 . >>>>>>>>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 .. >>>>>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found >>>>>>>>>>>> -rw-r--r--. 1 vdsm kvm 0 Mar 17 13:44 test >>>>>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .Trash-0 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> But it fail with a general exception error and the storage does >>>>>>>>>>>> not exist in ovirt >>>>>>>>>>>> >>>>>>>>>>>> any help would be appreciated. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Which log you need to see? >>>>>>>>>>>> >>>>>>>>>>>> Thanks >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Le jeu. 16 mars 2017 17:02, Fred Rolland <[email protected]> >>>>>>>>>>>> a écrit : >>>>>>>>>>>> >>>>>>>>>>>>> Hi, >>>>>>>>>>>>> >>>>>>>>>>>>> Can you check if the folder permissions are OK ? >>>>>>>>>>>>> Check [1] for more details. >>>>>>>>>>>>> >>>>>>>>>>>>> Can you share more of the log ? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> [1] https://www.ovirt.org/document >>>>>>>>>>>>> ation/how-to/troubleshooting/troubleshooting-nfs-storage-iss >>>>>>>>>>>>> ues/ >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Mar 16, 2017 at 7:49 PM, carl langlois < >>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> Hi Guys, >>>>>>>>>>>>> >>>>>>>>>>>>> I am trying to add a posix FS on one of my host. Ovirt in >>>>>>>>>>>>> actually mounting it but fail with "Error while executing >>>>>>>>>>>>> action Add Storage Connection: General Exception" >>>>>>>>>>>>> >>>>>>>>>>>>> If i look in the vdsm.log i cant see >>>>>>>>>>>>> >>>>>>>>>>>>> sonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>>>>>>>> 12:39:28,248::fileUtils::209::Storage.fileUtils::(createdir) >>>>>>>>>>>>> Creating directory: /rhev/data-center/mnt/_dev_dm-3 mode: None >>>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>>>>>>>> 12:39:28,248::fileUtils::218::Storage.fileUtils::(createdir) >>>>>>>>>>>>> Using existing directory: /rhev/data-center/mnt/_dev_dm-3 >>>>>>>>>>>>> jsonrpc.Executor/7::INFO::2017-03-16 >>>>>>>>>>>>> 12:39:28,248::mount::226::storage.Mount::(mount) mounting >>>>>>>>>>>>> /dev/dm-3 at /rhev/data-center/mnt/_dev_dm-3 >>>>>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>>>>>>>> 12:39:28,270::utils::871::storage.Mount::(stopwatch) >>>>>>>>>>>>> /rhev/data-center/mnt/_dev_dm-3 mounted: 0.02 seconds >>>>>>>>>>>>> jsonrpc.Executor/7::ERROR::2017-03-16 >>>>>>>>>>>>> 12:39:28,271::hsm::2403::Storage.HSM::(connectStorageServer) >>>>>>>>>>>>> Could not connect to storageServer >>>>>>>>>>>>> Traceback (most recent call last): >>>>>>>>>>>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in >>>>>>>>>>>>> connectStorageServer >>>>>>>>>>>>> conObj.connect() >>>>>>>>>>>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, >>>>>>>>>>>>> in connect >>>>>>>>>>>>> self.getMountObj().getRecord().fs_file) >>>>>>>>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", >>>>>>>>>>>>> line 260, in getRecord >>>>>>>>>>>>> (self.fs_spec, self.fs_file)) >>>>>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at >>>>>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> any help would be appreciated. >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks >>>>>>>>>>>>> >>>>>>>>>>>>> CL >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>> [email protected] >>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

