[ceph-users] Is RDMA Worth Exploring? Howto ?
I have a working Luminous 12.2.4 cluster CentOS 7.4 connected via 10G and Mellanox Connect X-3 QDR IB and would like to know if there are any worthwhile gains to be had from enabling RDMA and if there are any good up to date docs on how to do so? Thanks. - - *Paul Kunicki* - *Systems Manager* - SproutLoud Media Networks, LLC. - 954-476-6211 ext. 144 - pkuni...@sproutloud.com - www.sproutloud.com - • • • - [image: Facebook] <http://www.facebook.com/sproutLoud> [image: Twitter] <http://twitter.com/sproutloud> [image: LinkedIn] <http://www.linkedin.com/company/440456?trk=tyah> [image: LinkedIn] <https://www.instagram.com/sproutloud/> [image: YouTube] <https://www.youtube.com/user/SproutLoudVideo> The information contained in this communication is intended solely for the use of the individual or entity to whom it is addressed and for others authorized to receive it. It may contain confidential or legally privileged information. If you are not the intended recipient, you are hereby notified that any disclosure, copying, distribution, or taking any action in reliance on these contents is strictly prohibited and may be unlawful. In the event the recipient or recipients of this communication are under a non-disclosure agreement, any and all information discussed during phone calls and online presentations fall under the agreements signed by both parties. If you received this communication in error, please notify us immediately by responding to this e-mail. <pkuni...@sproutloud.com> ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Automated Failover of CephFS Clients
Thanks for the hint Linh. I had neglected to read up on mount.fuse.ceph here: http://docs.ceph.com/docs/master/man/8/mount.fuse.ceph/ I am trying this right now. Thanks again. - - *Paul Kunicki* - *Systems Manager* - SproutLoud Media Networks, LLC. - 954-476-6211 ext. 144 - pkuni...@sproutloud.com - www.sproutloud.com - • • • - [image: Facebook] <http://www.facebook.com/sproutLoud> [image: Twitter] <http://twitter.com/sproutloud> [image: LinkedIn] <http://www.linkedin.com/company/440456?trk=tyah> [image: LinkedIn] <https://www.instagram.com/sproutloud/> [image: YouTube] <https://www.youtube.com/user/SproutLoudVideo> The information contained in this communication is intended solely for the use of the individual or entity to whom it is addressed and for others authorized to receive it. It may contain confidential or legally privileged information. If you are not the intended recipient, you are hereby notified that any disclosure, copying, distribution, or taking any action in reliance on these contents is strictly prohibited and may be unlawful. In the event the recipient or recipients of this communication are under a non-disclosure agreement, any and all information discussed during phone calls and online presentations fall under the agreements signed by both parties. If you received this communication in error, please notify us immediately by responding to this e-mail. <pkuni...@sproutloud.com> On Tue, Feb 20, 2018 at 8:35 PM, Linh Vu <v...@unimelb.edu.au> wrote: > Why are you mounting with a single monitor? What is your mount command or > /etc/fstab? Ceph-fuse should use the available mons you have on the > client's /etc/ceph/ceph.conf. > > > e.g our /etc/fstab entry: > > > none/home fuse.ceph _netdev,ceph.id=myclusterid, > ceph.client_mountpoint=/home,nonempty,defaults 0 0 > -- > *From:* ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of Paul > Kunicki <pkuni...@sproutloud.com> > *Sent:* Wednesday, 21 February 2018 10:23:37 AM > *To:* ceph-users@lists.ceph.com > *Subject:* [ceph-users] Automated Failover of CephFS Clients > > We currently have multiple CephFS fuse clients mounting the same > filesystem from a single monitor even though our cluster has several > monitors. I would like to automate the fail over from one monitor to > another. Is this possible and where should I bee looking for guidance on > accomplishing this in production? I would like to avoid involving NFS if > possible and Pacemaker seems to be overkill but we can go that route if > that is what is in fact needed? > > We are currently at 12.2.2 on Centos 7.4. > > Thanks. > > > >- > - *Paul Kunicki* > - *Systems Manager* > - SproutLoud Media Networks, LLC. > - 954-476-6211 ext. 144 > - pkuni...@sproutloud.com > - www.sproutloud.com > - • • • > - [image: Facebook] <http://www.facebook.com/sproutLoud> [image: > Twitter] <http://twitter.com/sproutloud> [image: LinkedIn] > <http://www.linkedin.com/company/440456?trk=tyah> [image: > LinkedIn] <https://www.instagram.com/sproutloud/> [image: > YouTube] <https://www.youtube.com/user/SproutLoudVideo> > > The information contained in this communication is intended solely > for the use of the individual or entity to whom it is addressed and for > others authorized to receive it. It may contain confidential or legally > privileged information. If you are not the intended recipient, you are > hereby notified that any disclosure, copying, distribution, or taking > any > action in reliance on these contents is strictly prohibited and may be > unlawful. In the event the recipient or recipients of this communication > are under a non-disclosure agreement, any and all information discussed > during phone calls and online presentations fall under the agreements > signed by both parties. If you received this communication in error, > please > notify us immediately by responding to this e-mail. > > <pkuni...@sproutloud.com> > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[ceph-users] Automated Failover of CephFS Clients
We currently have multiple CephFS fuse clients mounting the same filesystem from a single monitor even though our cluster has several monitors. I would like to automate the fail over from one monitor to another. Is this possible and where should I bee looking for guidance on accomplishing this in production? I would like to avoid involving NFS if possible and Pacemaker seems to be overkill but we can go that route if that is what is in fact needed? We are currently at 12.2.2 on Centos 7.4. Thanks. - - *Paul Kunicki* - *Systems Manager* - SproutLoud Media Networks, LLC. - 954-476-6211 ext. 144 - pkuni...@sproutloud.com - www.sproutloud.com - • • • - [image: Facebook] <http://www.facebook.com/sproutLoud> [image: Twitter] <http://twitter.com/sproutloud> [image: LinkedIn] <http://www.linkedin.com/company/440456?trk=tyah> [image: LinkedIn] <https://www.instagram.com/sproutloud/> [image: YouTube] <https://www.youtube.com/user/SproutLoudVideo> The information contained in this communication is intended solely for the use of the individual or entity to whom it is addressed and for others authorized to receive it. It may contain confidential or legally privileged information. If you are not the intended recipient, you are hereby notified that any disclosure, copying, distribution, or taking any action in reliance on these contents is strictly prohibited and may be unlawful. In the event the recipient or recipients of this communication are under a non-disclosure agreement, any and all information discussed during phone calls and online presentations fall under the agreements signed by both parties. If you received this communication in error, please notify us immediately by responding to this e-mail. <pkuni...@sproutloud.com> ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Cephfs Snapshots - Usable in Single FS per Pool Scenario ?
It appears that this issue is somewhat intermittent as it took several tries to reproduce. What follows is a session where a snapshot is used to successfully retrieve the original version of a hosts file prior to the entire etc directory being snapped. The version is available even after the entire etc directory is deleted but when I go to the snapshot to retrieve other files they are empty. [root@sl-util mnt]# ls cephfs iso [root@sl-util mnt]# cp -rapf /etc cephfs [root@sl-util mnt]# cd cephfs/.snap [root@sl-util .snap]# ls etc-snap mysnap1 mysnap2 snap3 snap6 [root@sl-util .snap]# mkdir snap10 [root@sl-util .snap]# cd ../etc [root@sl-util etc]# cat hosts #192.168.2.2 sl-nas.sproutloud.com sl-nas 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 192.168.1.147 sl-util.sproutloud.com sl-util 10.0.100.51 prod-ceph-osd-1 10.0.100.53 prod-ceph-osd-2 10.0.100.54 prod-ceph-mon-1 10.0.100.55 prod-ceph-mon-2 10.0.100.56 prod-ceph-mds-1 10.0.100.57 prod-ceph-mds-2 10.0.100.58 prod-ceph-admin 10.0.100.59 prod-ceph-client-1 10.0.100.60 prod-ceph-client-2 [root@sl-util etc]# echo "This line has been appended just after taking the first snapshot of /mnt/cephfs/etc" >> hosts [root@sl-util etc]# cat hosts #192.168.2.2 sl-nas.sproutloud.com sl-nas 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 192.168.1.147 sl-util.sproutloud.com sl-util 10.0.100.51 prod-ceph-osd-1 10.0.100.53 prod-ceph-osd-2 10.0.100.54 prod-ceph-mon-1 10.0.100.55 prod-ceph-mon-2 10.0.100.56 prod-ceph-mds-1 10.0.100.57 prod-ceph-mds-2 10.0.100.58 prod-ceph-admin 10.0.100.59 prod-ceph-client-1 10.0.100.60 prod-ceph-client-2 This line has been appended just after taking the first snapshot of /mnt/cephfs/etc \\\ [root@sl-util etc]# rm -f hosts [root@sl-util etc]# cat ../.snap/snap10/etc/hosts #192.168.2.2 sl-nas.sproutloud.com sl-nas 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 192.168.1.147 sl-util.sproutloud.com sl-util 10.0.100.51 prod-ceph-osd-1 10.0.100.53 prod-ceph-osd-2 10.0.100.54 prod-ceph-mon-1 10.0.100.55 prod-ceph-mon-2 10.0.100.56 prod-ceph-mds-1 10.0.100.57 prod-ceph-mds-2 10.0.100.58 prod-ceph-admin 10.0.100.59 prod-ceph-client-1 10.0.100.60 prod-ceph-client-2 [root@sl-util etc]# cd .. [root@sl-util cephfs]# pwd /mnt/cephfs [root@sl-util cephfs]# rm -rf etc [root@sl-util cephfs]# cat .snap/snap10/etc/hosts #192.168.2.2 sl-nas.sproutloud.com sl-nas 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 192.168.1.147 sl-util.sproutloud.com sl-util 10.0.100.51 prod-ceph-osd-1 10.0.100.53 prod-ceph-osd-2 10.0.100.54 prod-ceph-mon-1 10.0.100.55 prod-ceph-mon-2 10.0.100.56 prod-ceph-mds-1 10.0.100.57 prod-ceph-mds-2 10.0.100.58 prod-ceph-admin 10.0.100.59 prod-ceph-client-1 10.0.100.60 prod-ceph-client-2 [root@sl-util cephfs]# ls CentOS-7-x86_64-Everything-1708.iso mydir yourdir [root@sl-util cephfs]# cp -rapf .snap/snap10/etc ./ [root@sl-util cephfs]# cat etc/hosts #192.168.2.2 sl-nas.sproutloud.com sl-nas 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 192.168.1.147 sl-util.sproutloud.com sl-util 10.0.100.51 prod-ceph-osd-1 10.0.100.53 prod-ceph-osd-2 10.0.100.54 prod-ceph-mon-1 10.0.100.55 prod-ceph-mon-2 10.0.100.56 prod-ceph-mds-1 10.0.100.57 prod-ceph-mds-2 10.0.100.58 prod-ceph-admin 10.0.100.59 prod-ceph-client-1 10.0.100.60 prod-ceph-client-2 [root@sl-util cephfs]# cat etc/mke2fs.conf CentOS-7-x86_64-Everything-1708.iso etc/ mydir/ yourdir/ [root@sl-util cephfs]# cat etc/e2fsck.conf [root@sl-util cephfs]# cat etc/passwd [root@sl-util cephfs]# cat etc/hosts.allow [root@sl-util cephfs]# Paul Kunicki Systems Manager SproutLoud Media Networks, LLC. 954-476-6211 ext. 144 pkuni...@sproutloud.com www.sproutloud.com • • • The information contained in this communication is intended solely for the use of the individual or entity to whom it is addressed and for others authorized to receive it. It may contain confidential or legally privileged information. If you are not the intended recipient, you are hereby notified that any disclosure, copying, distribution, or taking any action in reliance on these contents is strictly prohibited and may be unlawful. In the event the recipient or recipients of this communication are under a non-disclosure agreem
Re: [ceph-users] Cephfs Snapshots - Usable in Single FS per Pool Scenario ?
Thanks for the fast reply. I started recording a session where I unmounted and re-mounted the file system and could not duplicate the issue. I am going to do some more testing and report back any relevant findings. For now here are some details about our setup where files contained in shapshots were either empty or contained non printable contents once their original versions were modified. [root@sl-util mnt]# cat /etc/redhat-release CentOS Linux release 7.4.1708 (Core) [root@sl-util mnt]# uname -a Linux sl-util.sproutloud.com 3.10.0-693.2.2.el7.x86_64 #1 SMP Tue Sep 12 22:26:13 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux [root@sl-util mnt]# rpm -qa | grep ceph centos-release-ceph-luminous-1.0-1.el7.centos.noarch libcephfs2-12.2.2-0.el7.x86_64 ceph-common-12.2.2-0.el7.x86_64 python-cephfs-12.2.2-0.el7.x86_64 [root@sl-util mnt]# ceph -v ceph version 12.2.2 (cf0baba3b47f9427c6c97e2144b094b7e5ba) luminous (stable) [root@sl-util mnt]# ceph health HEALTH_OK [root@sl-util mnt]# ceph fs ls name: cephfs, metadata pool: cephfs_metadata, data pools: [cephfs_data ] [root@sl-util mnt]# ceph df detail GLOBAL: SIZE AVAIL RAW USED %RAW USED OBJECTS 63487G 63477G 11015M 0.02 19456 POOLS: NAMEID QUOTA OBJECTS QUOTA BYTES USED %USED MAX AVAIL OBJECTS DIRTY READ WRITE RAW USED cephfs_data 5 N/A N/A 2858M 029680G 15733 15733 12452 378305716M cephfs_metadata 6 N/A N/A 41865k 029680G3723 3723 140 17799 83731k [root@sl-util mnt]# mount -t ceph prod-ceph-mon-1:6789:/ /mnt/cephfs -o name=admin,secretfile=/etc/ceph/client.admin.secret [root@sl-util mnt]# mount proc on /proc type proc (rw,nosuid,nodev,noexec,relatime) devtmpfs on /dev type devtmpfs (rw,nosuid,size=8118464k,nr_inodes=2029616,mode=755) securityfs on /sys/kernel/security type securityfs (rw,nosuid,nodev,noexec,relatime) tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev) devpts on /dev/pts type devpts (rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000) tmpfs on /run type tmpfs (rw,nosuid,nodev,mode=755) tmpfs on /sys/fs/cgroup type tmpfs (ro,nosuid,nodev,noexec,mode=755) cgroup on /sys/fs/cgroup/systemd type cgroup (rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd) pstore on /sys/fs/pstore type pstore (rw,nosuid,nodev,noexec,relatime) cgroup on /sys/fs/cgroup/pids type cgroup (rw,nosuid,nodev,noexec,relatime,pids) cgroup on /sys/fs/cgroup/memory type cgroup (rw,nosuid,nodev,noexec,relatime,memory) cgroup on /sys/fs/cgroup/blkio type cgroup (rw,nosuid,nodev,noexec,relatime,blkio) cgroup on /sys/fs/cgroup/cpu,cpuacct type cgroup (rw,nosuid,nodev,noexec,relatime,cpuacct,cpu) cgroup on /sys/fs/cgroup/perf_event type cgroup (rw,nosuid,nodev,noexec,relatime,perf_event) cgroup on /sys/fs/cgroup/cpuset type cgroup (rw,nosuid,nodev,noexec,relatime,cpuset) cgroup on /sys/fs/cgroup/net_cls,net_prio type cgroup (rw,nosuid,nodev,noexec,relatime,net_prio,net_cls) cgroup on /sys/fs/cgroup/devices type cgroup (rw,nosuid,nodev,noexec,relatime,devices) cgroup on /sys/fs/cgroup/hugetlb type cgroup (rw,nosuid,nodev,noexec,relatime,hugetlb) cgroup on /sys/fs/cgroup/freezer type cgroup (rw,nosuid,nodev,noexec,relatime,freezer) configfs on /sys/kernel/config type configfs (rw,relatime) /dev/mapper/centos-root on / type xfs (rw,relatime,attr2,inode64,noquota) systemd-1 on /proc/sys/fs/binfmt_misc type autofs (rw,relatime,fd=29,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=9892) hugetlbfs on /dev/hugepages type hugetlbfs (rw,relatime) mqueue on /dev/mqueue type mqueue (rw,relatime) debugfs on /sys/kernel/debug type debugfs (rw,relatime) nfsd on /proc/fs/nfsd type nfsd (rw,relatime) /dev/sdl1 on /boot type xfs (rw,relatime,attr2,inode64,noquota) /dev/mapper/centos-home on /home type xfs (rw,relatime,attr2,inode64,noquota) sysfs on /sys type sysfs (rw,nosuid,nodev,noexec,relatime) sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime) tmpfs on /run/user/10065 type tmpfs (rw,nosuid,nodev,relatime,size=1626268k,mode=700,uid=10065,gid=10027) tmpfs on /run/user/10058 type tmpfs (rw,nosuid,nodev,relatime,size=1626268k,mode=700,uid=10058,gid=10027) 10.0.100.54:6789:/ on /mnt/cephfs type ceph (rw,relatime,name=admin,secret=,acl) [root@sl-util mnt]# ceph fs set cephfs allow_new_snaps true --yes-i-really-mean-it enabled new snapshots Paul Kunicki Systems Manager SproutLoud Media Networks, LLC. 954-476-6211 ext. 144 pkuni...@sproutloud.com www.sproutloud.com • • • The information contained in this communication is intended solely for the use of the individual or entity to whom it is addressed and for others authorized to receive it. It may contain confidential or legally privileged information. If you are not the intended recipient, you are hereby notified that any
[ceph-users] Cephfs Snapshots - Usable in Single FS per Pool Scenario ?
I know that snapshots on Cephfs are experimental and that a known issue exists with multiple filesystems on one pool but I was surprised at the result of the following: I attempted to take a snapshot of a directory in a pool with a single fs on our properly configured Luminous cluster. I found that the files in the the .snap directory that I had just updated in order to test a restore were unreadable if opened with and editor like VI or simply were identical to the current version of the file when copied back making the whole snapshot operation unusable. I considered the whole method of taking a snapshot to be very straightforward but perhaps I am doing something wrong or is this behavior to be expected ? Thanks. Paul Kunicki Systems Manager SproutLoud Media Networks, LLC. 954-476-6211 ext. 144 pkuni...@sproutloud.com www.sproutloud.com • • • The information contained in this communication is intended solely for the use of the individual or entity to whom it is addressed and for others authorized to receive it. It may contain confidential or legally privileged information. If you are not the intended recipient, you are hereby notified that any disclosure, copying, distribution, or taking any action in reliance on these contents is strictly prohibited and may be unlawful. In the event the recipient or recipients of this communication are under a non-disclosure agreement, any and all information discussed during phone calls and online presentations fall under the agreements signed by both parties. If you received this communication in error, please notify us immediately by responding to this e-mail. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com