Can you try : chown -R vdsm:kvm /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1
On Tue, Mar 21, 2017 at 4:32 PM, carl langlois <[email protected]> wrote: > > jsonrpc.Executor/7::WARNING::2017-03-21 09:27:40,099::outOfProcess:: > 193::Storage.oop::(validateAccess) Permission denied for directory: > /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1 > with permissions:7 > jsonrpc.Executor/7::INFO::2017-03-21 > 09:27:40,099::mount::233::storage.Mount::(umount) > unmounting /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__ > 50026B726804F13B1 > jsonrpc.Executor/7::DEBUG::2017-03-21 > 09:27:40,104::utils::871::storage.Mount::(stopwatch) > /rhev/data-center/mnt/_dev_mapper_KINGSTON__SV300S37A240G__50026B726804F13B1 > unmounted: 0.00 seconds > jsonrpc.Executor/7::ERROR::2017-03-21 09:27:40,104::hsm::2403:: > Storage.HSM::(connectStorageServer) Could not connect to storageServer > Traceback (most recent call last): > File "/usr/share/vdsm/storage/hsm.py", line 2400, in > connectStorageServer > conObj.connect() > File "/usr/share/vdsm/storage/storageServer.py", line 249, in connect > six.reraise(t, v, tb) > File "/usr/share/vdsm/storage/storageServer.py", line 242, in connect > self.getMountObj().getRecord().fs_file) > File "/usr/share/vdsm/storage/fileSD.py", line 81, in validateDirAccess > raise se.StorageServerAccessPermissionError(dirPath) > StorageServerAccessPermissionError: Permission settings on the specified > path do not allow access to the storage. Verify permission settings on the > specified storage path.: 'path = /rhev/data-center/mnt/_dev_ > mapper_KINGSTON__SV300S37A240G__50026B726804F13B1' > jsonrpc.Executor/7::DEBUG::201 > > Thanks again. > > > On Tue, Mar 21, 2017 at 10:14 AM, Fred Rolland <[email protected]> > wrote: > >> Can you share the VDSM log again ? >> >> On Tue, Mar 21, 2017 at 4:08 PM, carl langlois <[email protected]> >> wrote: >> >>> Interesting, when i'm using >>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>> now the UI give error on the permission setting.. >>> >>> root@ovhost4 ~]# ls -al /dev/mapper/KINGSTON_SV300S37A >>> 240G_50026B726804F13B1 >>> lrwxrwxrwx 1 root root 7 Mar 18 08:28 >>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>> -> ../dm-3 >>> >>> and the permission on the dm-3 >>> >>> [root@ovhost4 ~]# ls -al /dev/dm-3 >>> brw-rw---- 1 vdsm kvm 253, 3 Mar 18 08:28 /dev/dm-3 >>> >>> >>> how do i change the permission on the sym link.. >>> >>> Thanks >>> >>> >>> >>> >>> On Tue, Mar 21, 2017 at 10:00 AM, Fred Rolland <[email protected]> >>> wrote: >>> >>>> Can you try to use /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>>> in the UI. >>>> It seems the kernel change the path that we use to mount and then we >>>> cannot validate that the mount exists. >>>> >>>> It should be anyway better as the mapping could change after reboot. >>>> >>>> On Tue, Mar 21, 2017 at 2:20 PM, carl langlois <[email protected]> >>>> wrote: >>>> >>>>> Here is the /proc/mounts >>>>> >>>>> rootfs / rootfs rw 0 0 >>>>> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0 >>>>> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0 >>>>> devtmpfs /dev devtmpfs >>>>> rw,nosuid,size=65948884k,nr_inodes=16487221,mode=755 >>>>> 0 0 >>>>> securityfs /sys/kernel/security securityfs >>>>> rw,nosuid,nodev,noexec,relatime 0 0 >>>>> tmpfs /dev/shm tmpfs rw,nosuid,nodev 0 0 >>>>> devpts /dev/pts devpts >>>>> rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000 >>>>> 0 0 >>>>> tmpfs /run tmpfs rw,nosuid,nodev,mode=755 0 0 >>>>> tmpfs /sys/fs/cgroup tmpfs ro,nosuid,nodev,noexec,mode=755 0 0 >>>>> cgroup /sys/fs/cgroup/systemd cgroup rw,nosuid,nodev,noexec,relatim >>>>> e,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd >>>>> 0 0 >>>>> pstore /sys/fs/pstore pstore rw,nosuid,nodev,noexec,relatime 0 0 >>>>> cgroup /sys/fs/cgroup/cpu,cpuacct cgroup >>>>> rw,nosuid,nodev,noexec,relatime,cpuacct,cpu >>>>> 0 0 >>>>> cgroup /sys/fs/cgroup/net_cls,net_prio cgroup >>>>> rw,nosuid,nodev,noexec,relatime,net_prio,net_cls 0 0 >>>>> cgroup /sys/fs/cgroup/pids cgroup rw,nosuid,nodev,noexec,relatime,pids >>>>> 0 0 >>>>> cgroup /sys/fs/cgroup/devices cgroup >>>>> rw,nosuid,nodev,noexec,relatime,devices >>>>> 0 0 >>>>> cgroup /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset >>>>> 0 0 >>>>> cgroup /sys/fs/cgroup/blkio cgroup rw,nosuid,nodev,noexec,relatime,blkio >>>>> 0 0 >>>>> cgroup /sys/fs/cgroup/perf_event cgroup >>>>> rw,nosuid,nodev,noexec,relatime,perf_event >>>>> 0 0 >>>>> cgroup /sys/fs/cgroup/memory cgroup rw,nosuid,nodev,noexec,relatime,memory >>>>> 0 0 >>>>> cgroup /sys/fs/cgroup/freezer cgroup >>>>> rw,nosuid,nodev,noexec,relatime,freezer >>>>> 0 0 >>>>> cgroup /sys/fs/cgroup/hugetlb cgroup >>>>> rw,nosuid,nodev,noexec,relatime,hugetlb >>>>> 0 0 >>>>> configfs /sys/kernel/config configfs rw,relatime 0 0 >>>>> /dev/mapper/cl_ovhost1-root / xfs rw,relatime,attr2,inode64,noquota 0 >>>>> 0 >>>>> systemd-1 /proc/sys/fs/binfmt_misc autofs >>>>> rw,relatime,fd=35,pgrp=1,timeout=300,minproto=5,maxproto=5,direct 0 0 >>>>> mqueue /dev/mqueue mqueue rw,relatime 0 0 >>>>> debugfs /sys/kernel/debug debugfs rw,relatime 0 0 >>>>> hugetlbfs /dev/hugepages hugetlbfs rw,relatime 0 0 >>>>> tmpfs /tmp tmpfs rw 0 0 >>>>> nfsd /proc/fs/nfsd nfsd rw,relatime 0 0 >>>>> /dev/mapper/cl_ovhost1-home /home xfs rw,relatime,attr2,inode64,noquota >>>>> 0 0 >>>>> /dev/sda1 /boot xfs rw,relatime,attr2,inode64,noquota 0 0 >>>>> sunrpc /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0 >>>>> tmpfs /run/user/42 tmpfs rw,nosuid,nodev,relatime,size= >>>>> 13192948k,mode=700,uid=42,gid=42 0 0 >>>>> gvfsd-fuse /run/user/42/gvfs fuse.gvfsd-fuse >>>>> rw,nosuid,nodev,relatime,user_id=42,group_id=42 0 0 >>>>> fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0 >>>>> ovhost2:/home/exports/defaultdata >>>>> /rhev/data-center/mnt/ovhost2:_home_exports_defaultdata >>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >>>>> local_lock=none,addr=10.8.236.162 0 0 >>>>> ovhost2:/home/exports/ISO /rhev/data-center/mnt/ovhost2:_home_exports_ISO >>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >>>>> local_lock=none,addr=10.8.236.162 0 0 >>>>> ovhost2:/home/exports/data >>>>> /rhev/data-center/mnt/ovhost2:_home_exports_data >>>>> nfs rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >>>>> ft,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >>>>> ddr=10.8.236.162,mountvers=3,mountport=20048,mountproto=udp, >>>>> local_lock=none,addr=10.8.236.162 0 0 >>>>> tmpfs /run/user/0 tmpfs rw,nosuid,nodev,relatime,size=13192948k,mode=700 >>>>> 0 0 >>>>> /dev/mapper/KINGSTON_SV300S37A240G_50026B726804F13B1 >>>>> /rhev/data-center/mnt/_dev_dm-3 ext4 rw,nosuid,relatime,data=ordered >>>>> 0 0 >>>>> >>>>> Thanks you for your help. >>>>> >>>>> Carl >>>>> >>>>> >>>>> On Tue, Mar 21, 2017 at 6:31 AM, Fred Rolland <[email protected]> >>>>> wrote: >>>>> >>>>>> Can you provide the content of /proc/mounts after it has being >>>>>> mounted by VDSM ? >>>>>> >>>>>> On Tue, Mar 21, 2017 at 12:28 PM, carl langlois < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Here is the vdsm.log >>>>>>> >>>>>>> >>>>>>> jsonrpc.Executor/0::ERROR::2017-03-18 >>>>>>> 08:23:48,317::hsm::2403::Storage.HSM::(connectStorageServer) >>>>>>> Could not connect to storageServer >>>>>>> Traceback (most recent call last): >>>>>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in >>>>>>> connectStorageServer >>>>>>> conObj.connect() >>>>>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, in >>>>>>> connect >>>>>>> self.getMountObj().getRecord().fs_file) >>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", >>>>>>> line 260, in getRecord >>>>>>> (self.fs_spec, self.fs_file)) >>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at >>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist >>>>>>> >>>>>>> >>>>>>> thanks >>>>>>> >>>>>>> On Fri, Mar 17, 2017 at 3:06 PM, Fred Rolland <[email protected]> >>>>>>> wrote: >>>>>>> >>>>>>>> Please send Vdsm log. >>>>>>>> Thanks >>>>>>>> >>>>>>>> On Fri, Mar 17, 2017 at 8:46 PM, carl langlois < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> The link that you send is for NFS strorage but i am trying to add >>>>>>>>> a POSIX compliant. >>>>>>>>> >>>>>>>>> [image: Inline image 1] >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> when i press okey it mount the disk to : >>>>>>>>> >>>>>>>>> [root@ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/ >>>>>>>>> total 28 >>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 . >>>>>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 .. >>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found >>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .Trash-0 >>>>>>>>> >>>>>>>>> >>>>>>>>> and doing a touch with vdsm user work >>>>>>>>> >>>>>>>>> [root@ovhost4 ~]# sudo -u vdsm touch >>>>>>>>> /rhev/data-center/mnt/_dev_dm-4/test >>>>>>>>> [root@ovhost4 ~]# ls -al /rhev/data-center/mnt/_dev_dm-4/ >>>>>>>>> total 28 >>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 17 13:44 . >>>>>>>>> drwxr-xr-x. 6 vdsm kvm 4096 Mar 17 13:35 .. >>>>>>>>> drwxr-xr-x. 2 vdsm kvm 16384 Mar 16 11:42 lost+found >>>>>>>>> -rw-r--r--. 1 vdsm kvm 0 Mar 17 13:44 test >>>>>>>>> drwxr-xr-x. 4 vdsm kvm 4096 Mar 16 12:12 .Trash-0 >>>>>>>>> >>>>>>>>> >>>>>>>>> But it fail with a general exception error and the storage does >>>>>>>>> not exist in ovirt >>>>>>>>> >>>>>>>>> any help would be appreciated. >>>>>>>>> >>>>>>>>> >>>>>>>>> Which log you need to see? >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Le jeu. 16 mars 2017 17:02, Fred Rolland <[email protected]> a >>>>>>>>> écrit : >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> Can you check if the folder permissions are OK ? >>>>>>>>>> Check [1] for more details. >>>>>>>>>> >>>>>>>>>> Can you share more of the log ? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> [1] https://www.ovirt.org/documentation/how-to/troubleshooting/t >>>>>>>>>> roubleshooting-nfs-storage-issues/ >>>>>>>>>> >>>>>>>>>> On Thu, Mar 16, 2017 at 7:49 PM, carl langlois < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>> Hi Guys, >>>>>>>>>> >>>>>>>>>> I am trying to add a posix FS on one of my host. Ovirt in >>>>>>>>>> actually mounting it but fail with "Error while executing action >>>>>>>>>> Add Storage Connection: General Exception" >>>>>>>>>> >>>>>>>>>> If i look in the vdsm.log i cant see >>>>>>>>>> >>>>>>>>>> sonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>>>>> 12:39:28,248::fileUtils::209::Storage.fileUtils::(createdir) >>>>>>>>>> Creating directory: /rhev/data-center/mnt/_dev_dm-3 mode: None >>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>>>>> 12:39:28,248::fileUtils::218::Storage.fileUtils::(createdir) >>>>>>>>>> Using existing directory: /rhev/data-center/mnt/_dev_dm-3 >>>>>>>>>> jsonrpc.Executor/7::INFO::2017-03-16 >>>>>>>>>> 12:39:28,248::mount::226::storage.Mount::(mount) mounting >>>>>>>>>> /dev/dm-3 at /rhev/data-center/mnt/_dev_dm-3 >>>>>>>>>> jsonrpc.Executor/7::DEBUG::2017-03-16 >>>>>>>>>> 12:39:28,270::utils::871::storage.Mount::(stopwatch) >>>>>>>>>> /rhev/data-center/mnt/_dev_dm-3 mounted: 0.02 seconds >>>>>>>>>> jsonrpc.Executor/7::ERROR::2017-03-16 >>>>>>>>>> 12:39:28,271::hsm::2403::Storage.HSM::(connectStorageServer) >>>>>>>>>> Could not connect to storageServer >>>>>>>>>> Traceback (most recent call last): >>>>>>>>>> File "/usr/share/vdsm/storage/hsm.py", line 2400, in >>>>>>>>>> connectStorageServer >>>>>>>>>> conObj.connect() >>>>>>>>>> File "/usr/share/vdsm/storage/storageServer.py", line 242, in >>>>>>>>>> connect >>>>>>>>>> self.getMountObj().getRecord().fs_file) >>>>>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/storage/mount.py", >>>>>>>>>> line 260, in getRecord >>>>>>>>>> (self.fs_spec, self.fs_file)) >>>>>>>>>> OSError: [Errno 2] Mount of `/dev/dm-3` at >>>>>>>>>> `/rhev/data-center/mnt/_dev_dm-3` does not exist >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> any help would be appreciated. >>>>>>>>>> >>>>>>>>>> Thanks >>>>>>>>>> >>>>>>>>>> CL >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list >>>>>>>>>> [email protected] >>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

