[ceph-users] Is RDMA Worth Exploring? Howto ?

2018-04-25 Thread Paul Kunicki
I have a working Luminous 12.2.4 cluster CentOS 7.4 connected via 10G and
Mellanox Connect X-3 QDR IB and would like to know if there are any
worthwhile gains to be had from enabling RDMA and if there are any good up
to date docs on how to do so?



Thanks.



   -
  - *Paul Kunicki*
 - *Systems Manager*
 - SproutLoud Media Networks, LLC.
 - 954-476-6211 ext. 144
 - pkuni...@sproutloud.com
 - www.sproutloud.com
  -  •   •   •
  - [image: Facebook] <http://www.facebook.com/sproutLoud>  [image:
 Twitter] <http://twitter.com/sproutloud>  [image: LinkedIn]
 <http://www.linkedin.com/company/440456?trk=tyah>  [image:
 LinkedIn] <https://www.instagram.com/sproutloud/>  [image: YouTube]
 <https://www.youtube.com/user/SproutLoudVideo>

  The information contained in this communication is intended solely
  for the use of the individual or entity to whom it is addressed and for
  others authorized to receive it. It may contain confidential or legally
  privileged information. If you are not the intended recipient, you are
  hereby notified that any disclosure, copying, distribution, or taking any
  action in reliance on these contents is strictly prohibited and may be
  unlawful. In the event the recipient or recipients of this communication
  are under a non-disclosure agreement, any and all information discussed
  during phone calls and online presentations fall under the agreements
  signed by both parties. If you received this communication in
error, please
  notify us immediately by responding to this e-mail.

<pkuni...@sproutloud.com>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Automated Failover of CephFS Clients

2018-02-20 Thread Paul Kunicki
Thanks for the hint Linh. I had neglected to read up on mount.fuse.ceph
here: http://docs.ceph.com/docs/master/man/8/mount.fuse.ceph/

I am trying this right now.

Thanks again.




   -
  - *Paul Kunicki*
 - *Systems Manager*
 - SproutLoud Media Networks, LLC.
 - 954-476-6211 ext. 144
 - pkuni...@sproutloud.com
 - www.sproutloud.com
  -  •   •   •
  - [image: Facebook] <http://www.facebook.com/sproutLoud>  [image:
 Twitter] <http://twitter.com/sproutloud>  [image: LinkedIn]
 <http://www.linkedin.com/company/440456?trk=tyah>  [image:
 LinkedIn] <https://www.instagram.com/sproutloud/>  [image: YouTube]
 <https://www.youtube.com/user/SproutLoudVideo>

  The information contained in this communication is intended solely
  for the use of the individual or entity to whom it is addressed and for
  others authorized to receive it. It may contain confidential or legally
  privileged information. If you are not the intended recipient, you are
  hereby notified that any disclosure, copying, distribution, or taking any
  action in reliance on these contents is strictly prohibited and may be
  unlawful. In the event the recipient or recipients of this communication
  are under a non-disclosure agreement, any and all information discussed
  during phone calls and online presentations fall under the agreements
  signed by both parties. If you received this communication in
error, please
  notify us immediately by responding to this e-mail.

<pkuni...@sproutloud.com>

On Tue, Feb 20, 2018 at 8:35 PM, Linh Vu <v...@unimelb.edu.au> wrote:

> Why are you mounting with a single monitor? What is your mount command or
> /etc/fstab? Ceph-fuse should use the available mons you have on the
> client's /etc/ceph/ceph.conf.
>
>
> e.g our /etc/fstab entry:
>
>
> none/home   fuse.ceph   _netdev,ceph.id=myclusterid,
> ceph.client_mountpoint=/home,nonempty,defaults  0   0
> --
> *From:* ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of Paul
> Kunicki <pkuni...@sproutloud.com>
> *Sent:* Wednesday, 21 February 2018 10:23:37 AM
> *To:* ceph-users@lists.ceph.com
> *Subject:* [ceph-users] Automated Failover of CephFS Clients
>
> We currently have multiple CephFS fuse clients mounting the same
> filesystem from a single monitor even though our cluster has several
> monitors. I would like to automate the fail over from one monitor to
> another. Is this possible and where should I bee looking for guidance on
> accomplishing this in production? I would like to avoid involving NFS if
> possible and Pacemaker seems to be overkill but we can go that route if
> that is what is in fact needed?
>
> We are currently at 12.2.2 on Centos 7.4.
>
> Thanks.
>
>
>
>-
>   - *Paul Kunicki*
>  - *Systems Manager*
>  - SproutLoud Media Networks, LLC.
>  - 954-476-6211 ext. 144
>  - pkuni...@sproutloud.com
>  - www.sproutloud.com
>   -  •   •   •
>   - [image: Facebook] <http://www.facebook.com/sproutLoud>  [image:
>  Twitter] <http://twitter.com/sproutloud>  [image: LinkedIn]
>  <http://www.linkedin.com/company/440456?trk=tyah>  [image:
>  LinkedIn] <https://www.instagram.com/sproutloud/>  [image:
>  YouTube] <https://www.youtube.com/user/SproutLoudVideo>
>
>   The information contained in this communication is intended solely
>   for the use of the individual or entity to whom it is addressed and for
>   others authorized to receive it. It may contain confidential or legally
>   privileged information. If you are not the intended recipient, you are
>   hereby notified that any disclosure, copying, distribution, or taking 
> any
>   action in reliance on these contents is strictly prohibited and may be
>   unlawful. In the event the recipient or recipients of this communication
>   are under a non-disclosure agreement, any and all information discussed
>   during phone calls and online presentations fall under the agreements
>   signed by both parties. If you received this communication in error, 
> please
>   notify us immediately by responding to this e-mail.
>
> <pkuni...@sproutloud.com>
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Automated Failover of CephFS Clients

2018-02-20 Thread Paul Kunicki
We currently have multiple CephFS fuse clients mounting the same filesystem
from a single monitor even though our cluster has several monitors. I would
like to automate the fail over from one monitor to another. Is this
possible and where should I bee looking for guidance on accomplishing this
in production? I would like to avoid involving NFS if possible and
Pacemaker seems to be overkill but we can go that route if that is what is
in fact needed?

We are currently at 12.2.2 on Centos 7.4.

Thanks.



   -
  - *Paul Kunicki*
 - *Systems Manager*
 - SproutLoud Media Networks, LLC.
 - 954-476-6211 ext. 144
 - pkuni...@sproutloud.com
 - www.sproutloud.com
  -  •   •   •
  - [image: Facebook] <http://www.facebook.com/sproutLoud>  [image:
 Twitter] <http://twitter.com/sproutloud>  [image: LinkedIn]
 <http://www.linkedin.com/company/440456?trk=tyah>  [image:
 LinkedIn] <https://www.instagram.com/sproutloud/>  [image: YouTube]
 <https://www.youtube.com/user/SproutLoudVideo>

  The information contained in this communication is intended solely
  for the use of the individual or entity to whom it is addressed and for
  others authorized to receive it. It may contain confidential or legally
  privileged information. If you are not the intended recipient, you are
  hereby notified that any disclosure, copying, distribution, or taking any
  action in reliance on these contents is strictly prohibited and may be
  unlawful. In the event the recipient or recipients of this communication
  are under a non-disclosure agreement, any and all information discussed
  during phone calls and online presentations fall under the agreements
  signed by both parties. If you received this communication in
error, please
  notify us immediately by responding to this e-mail.

<pkuni...@sproutloud.com>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Cephfs Snapshots - Usable in Single FS per Pool Scenario ?

2018-01-30 Thread Paul Kunicki
It appears that this issue is somewhat intermittent as it took several
tries to reproduce. What follows is a session where a snapshot is used
to successfully retrieve the original version of a hosts file prior to
the entire etc directory being snapped. The version is available even
after the entire etc directory is deleted but when I go to the
snapshot to retrieve other files they are empty.

[root@sl-util mnt]# ls
cephfs  iso
[root@sl-util mnt]# cp -rapf /etc cephfs
[root@sl-util mnt]# cd cephfs/.snap
[root@sl-util .snap]# ls
etc-snap  mysnap1  mysnap2  snap3  snap6
[root@sl-util .snap]# mkdir snap10
[root@sl-util .snap]# cd ../etc
[root@sl-util etc]# cat hosts
#192.168.2.2 sl-nas.sproutloud.com sl-nas
127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
::1 localhost localhost.localdomain localhost6 localhost6.localdomain6
192.168.1.147 sl-util.sproutloud.com sl-util


10.0.100.51 prod-ceph-osd-1
10.0.100.53 prod-ceph-osd-2
10.0.100.54 prod-ceph-mon-1
10.0.100.55 prod-ceph-mon-2
10.0.100.56 prod-ceph-mds-1
10.0.100.57 prod-ceph-mds-2
10.0.100.58 prod-ceph-admin
10.0.100.59 prod-ceph-client-1
10.0.100.60 prod-ceph-client-2

[root@sl-util etc]# echo "This line has been appended just after
taking the first snapshot of /mnt/cephfs/etc" >> hosts
[root@sl-util etc]# cat hosts
#192.168.2.2 sl-nas.sproutloud.com sl-nas
127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
::1 localhost localhost.localdomain localhost6 localhost6.localdomain6
192.168.1.147 sl-util.sproutloud.com sl-util


10.0.100.51 prod-ceph-osd-1
10.0.100.53 prod-ceph-osd-2
10.0.100.54 prod-ceph-mon-1
10.0.100.55 prod-ceph-mon-2
10.0.100.56 prod-ceph-mds-1
10.0.100.57 prod-ceph-mds-2
10.0.100.58 prod-ceph-admin
10.0.100.59 prod-ceph-client-1
10.0.100.60 prod-ceph-client-2

This line has been appended just after taking the first snapshot of
/mnt/cephfs/etc
\\\
[root@sl-util etc]# rm -f hosts
[root@sl-util etc]# cat ../.snap/snap10/etc/hosts
#192.168.2.2 sl-nas.sproutloud.com sl-nas
127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
::1 localhost localhost.localdomain localhost6 localhost6.localdomain6
192.168.1.147 sl-util.sproutloud.com sl-util


10.0.100.51 prod-ceph-osd-1
10.0.100.53 prod-ceph-osd-2
10.0.100.54 prod-ceph-mon-1
10.0.100.55 prod-ceph-mon-2
10.0.100.56 prod-ceph-mds-1
10.0.100.57 prod-ceph-mds-2
10.0.100.58 prod-ceph-admin
10.0.100.59 prod-ceph-client-1
10.0.100.60 prod-ceph-client-2

[root@sl-util etc]# cd ..
[root@sl-util cephfs]# pwd
/mnt/cephfs
[root@sl-util cephfs]# rm -rf etc
[root@sl-util cephfs]# cat .snap/snap10/etc/hosts
#192.168.2.2 sl-nas.sproutloud.com sl-nas
127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
::1 localhost localhost.localdomain localhost6 localhost6.localdomain6
192.168.1.147 sl-util.sproutloud.com sl-util


10.0.100.51 prod-ceph-osd-1
10.0.100.53 prod-ceph-osd-2
10.0.100.54 prod-ceph-mon-1
10.0.100.55 prod-ceph-mon-2
10.0.100.56 prod-ceph-mds-1
10.0.100.57 prod-ceph-mds-2
10.0.100.58 prod-ceph-admin
10.0.100.59 prod-ceph-client-1
10.0.100.60 prod-ceph-client-2


[root@sl-util cephfs]# ls
CentOS-7-x86_64-Everything-1708.iso  mydir  yourdir
[root@sl-util cephfs]# cp -rapf .snap/snap10/etc ./
[root@sl-util cephfs]# cat etc/hosts
#192.168.2.2 sl-nas.sproutloud.com sl-nas
127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
::1 localhost localhost.localdomain localhost6 localhost6.localdomain6
192.168.1.147 sl-util.sproutloud.com sl-util


10.0.100.51 prod-ceph-osd-1
10.0.100.53 prod-ceph-osd-2
10.0.100.54 prod-ceph-mon-1
10.0.100.55 prod-ceph-mon-2
10.0.100.56 prod-ceph-mds-1
10.0.100.57 prod-ceph-mds-2
10.0.100.58 prod-ceph-admin
10.0.100.59 prod-ceph-client-1
10.0.100.60 prod-ceph-client-2

[root@sl-util cephfs]# cat etc/mke2fs.conf
CentOS-7-x86_64-Everything-1708.iso  etc/
   mydir/   yourdir/
[root@sl-util cephfs]# cat etc/e2fsck.conf
[root@sl-util cephfs]# cat etc/passwd
[root@sl-util cephfs]# cat etc/hosts.allow
[root@sl-util cephfs]#


Paul Kunicki
Systems Manager
SproutLoud Media Networks, LLC.
954-476-6211 ext. 144
pkuni...@sproutloud.com
www.sproutloud.com

 •   •   •



The information contained in this communication is intended solely for
the use of the individual or entity to whom it is addressed and for
others authorized to receive it. It may contain confidential or
legally privileged information. If you are not the intended recipient,
you are hereby notified that any disclosure, copying, distribution, or
taking any action in reliance on these contents is strictly prohibited
and may be unlawful. In the event the recipient or recipients of this
communication are under a non-disclosure agreem

Re: [ceph-users] Cephfs Snapshots - Usable in Single FS per Pool Scenario ?

2018-01-30 Thread Paul Kunicki
Thanks for the fast reply. I started recording a session where I
unmounted and re-mounted the file system and could not duplicate the
issue. I am going to do some more testing and report back any relevant
findings. For now here are some details about our setup where files
contained in shapshots were either empty or contained non printable
contents once their original versions were modified.

[root@sl-util mnt]# cat /etc/redhat-release
CentOS Linux release 7.4.1708 (Core)
[root@sl-util mnt]# uname -a
Linux sl-util.sproutloud.com 3.10.0-693.2.2.el7.x86_64 #1 SMP Tue Sep
12 22:26:13 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux

[root@sl-util mnt]# rpm -qa | grep ceph
centos-release-ceph-luminous-1.0-1.el7.centos.noarch
libcephfs2-12.2.2-0.el7.x86_64
ceph-common-12.2.2-0.el7.x86_64
python-cephfs-12.2.2-0.el7.x86_64

[root@sl-util mnt]# ceph -v
ceph version 12.2.2 (cf0baba3b47f9427c6c97e2144b094b7e5ba) luminous (stable)

[root@sl-util mnt]# ceph health
HEALTH_OK

[root@sl-util mnt]# ceph fs ls
name: cephfs, metadata pool: cephfs_metadata, data pools: [cephfs_data ]

[root@sl-util mnt]# ceph df detail
GLOBAL:
SIZE   AVAIL  RAW USED %RAW USED OBJECTS
63487G 63477G   11015M  0.02   19456
POOLS:
NAMEID QUOTA OBJECTS QUOTA BYTES USED
 %USED MAX AVAIL OBJECTS DIRTY READ  WRITE
RAW USED
cephfs_data 5  N/A   N/A
2858M 029680G   15733 15733 12452
378305716M
cephfs_metadata 6  N/A   N/A
41865k 029680G3723  3723   140
17799   83731k


[root@sl-util mnt]# mount -t ceph prod-ceph-mon-1:6789:/ /mnt/cephfs
-o name=admin,secretfile=/etc/ceph/client.admin.secret
[root@sl-util mnt]# mount
proc on /proc type proc (rw,nosuid,nodev,noexec,relatime)
devtmpfs on /dev type devtmpfs
(rw,nosuid,size=8118464k,nr_inodes=2029616,mode=755)
securityfs on /sys/kernel/security type securityfs
(rw,nosuid,nodev,noexec,relatime)
tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev)
devpts on /dev/pts type devpts
(rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000)
tmpfs on /run type tmpfs (rw,nosuid,nodev,mode=755)
tmpfs on /sys/fs/cgroup type tmpfs (ro,nosuid,nodev,noexec,mode=755)
cgroup on /sys/fs/cgroup/systemd type cgroup
(rw,nosuid,nodev,noexec,relatime,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd)
pstore on /sys/fs/pstore type pstore (rw,nosuid,nodev,noexec,relatime)
cgroup on /sys/fs/cgroup/pids type cgroup (rw,nosuid,nodev,noexec,relatime,pids)
cgroup on /sys/fs/cgroup/memory type cgroup
(rw,nosuid,nodev,noexec,relatime,memory)
cgroup on /sys/fs/cgroup/blkio type cgroup
(rw,nosuid,nodev,noexec,relatime,blkio)
cgroup on /sys/fs/cgroup/cpu,cpuacct type cgroup
(rw,nosuid,nodev,noexec,relatime,cpuacct,cpu)
cgroup on /sys/fs/cgroup/perf_event type cgroup
(rw,nosuid,nodev,noexec,relatime,perf_event)
cgroup on /sys/fs/cgroup/cpuset type cgroup
(rw,nosuid,nodev,noexec,relatime,cpuset)
cgroup on /sys/fs/cgroup/net_cls,net_prio type cgroup
(rw,nosuid,nodev,noexec,relatime,net_prio,net_cls)
cgroup on /sys/fs/cgroup/devices type cgroup
(rw,nosuid,nodev,noexec,relatime,devices)
cgroup on /sys/fs/cgroup/hugetlb type cgroup
(rw,nosuid,nodev,noexec,relatime,hugetlb)
cgroup on /sys/fs/cgroup/freezer type cgroup
(rw,nosuid,nodev,noexec,relatime,freezer)
configfs on /sys/kernel/config type configfs (rw,relatime)
/dev/mapper/centos-root on / type xfs (rw,relatime,attr2,inode64,noquota)
systemd-1 on /proc/sys/fs/binfmt_misc type autofs
(rw,relatime,fd=29,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=9892)
hugetlbfs on /dev/hugepages type hugetlbfs (rw,relatime)
mqueue on /dev/mqueue type mqueue (rw,relatime)
debugfs on /sys/kernel/debug type debugfs (rw,relatime)
nfsd on /proc/fs/nfsd type nfsd (rw,relatime)
/dev/sdl1 on /boot type xfs (rw,relatime,attr2,inode64,noquota)
/dev/mapper/centos-home on /home type xfs (rw,relatime,attr2,inode64,noquota)
sysfs on /sys type sysfs (rw,nosuid,nodev,noexec,relatime)
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime)
tmpfs on /run/user/10065 type tmpfs
(rw,nosuid,nodev,relatime,size=1626268k,mode=700,uid=10065,gid=10027)
tmpfs on /run/user/10058 type tmpfs
(rw,nosuid,nodev,relatime,size=1626268k,mode=700,uid=10058,gid=10027)
10.0.100.54:6789:/ on /mnt/cephfs type ceph
(rw,relatime,name=admin,secret=,acl)

[root@sl-util mnt]# ceph fs set cephfs allow_new_snaps true
--yes-i-really-mean-it
enabled new snapshots




Paul Kunicki
Systems Manager
SproutLoud Media Networks, LLC.
954-476-6211 ext. 144
pkuni...@sproutloud.com
www.sproutloud.com

 •   •   •



The information contained in this communication is intended solely for
the use of the individual or entity to whom it is addressed and for
others authorized to receive it. It may contain confidential or
legally privileged information. If you are not the intended recipient,
you are hereby notified that any

[ceph-users] Cephfs Snapshots - Usable in Single FS per Pool Scenario ?

2018-01-29 Thread Paul Kunicki
I know that snapshots on Cephfs are experimental and that a known
issue exists with multiple filesystems on one pool but I was surprised
at the result of the following:

I attempted to take a snapshot of a directory in a pool with a single
fs on our properly configured Luminous cluster. I found that the files
in the the .snap directory that I had just updated in order to test a
restore were unreadable if opened with and editor like VI or simply
were identical to the current version of the file when copied back
making the whole snapshot operation unusable.

I considered the whole method of taking a snapshot to be very
straightforward but perhaps I am doing something wrong or is this
behavior to be expected ?

Thanks.




Paul Kunicki
Systems Manager
SproutLoud Media Networks, LLC.
954-476-6211 ext. 144
pkuni...@sproutloud.com
www.sproutloud.com

 •   •   •



The information contained in this communication is intended solely for
the use of the individual or entity to whom it is addressed and for
others authorized to receive it. It may contain confidential or
legally privileged information. If you are not the intended recipient,
you are hereby notified that any disclosure, copying, distribution, or
taking any action in reliance on these contents is strictly prohibited
and may be unlawful. In the event the recipient or recipients of this
communication are under a non-disclosure agreement, any and all
information discussed during phone calls and online presentations fall
under the agreements signed by both parties. If you received this
communication in error, please notify us immediately by responding to
this e-mail.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com