Re: [ceph-users] multiple journals on SSD

2016-07-07 Thread Nick Fisk
Hi Christian, > -Original Message- > From: Christian Balzer [mailto:ch...@gol.com] > Sent: 07 July 2016 12:57 > To: ceph-users@lists.ceph.com > Cc: Nick Fisk > Subject: Re: [ceph-users] multiple journals on SSD > > > Hello Nick, > > On Thu, 7 Jul 20

[ceph-users] RBD - Deletion / Discard - IO Impact

2016-07-07 Thread Nick Fisk
Hi All, Does anybody else see a massive (ie 10x) performance impact when either deleting a RBD or running something like mkfs.xfs against an existing RBD, which would zero/discard all blocks? In the case of deleting a 4TB RBD, I'm seeing latency in some cases rise up to 10s. It looks

Re: [ceph-users] layer3 network

2016-07-07 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Matyas Koszik > Sent: 07 July 2016 11:26 > To: ceph-users@lists.ceph.com > Subject: [ceph-users] layer3 network > > > > Hi, > > My setup uses a layer3 network, where each node has two conn

Re: [ceph-users] multiple journals on SSD

2016-07-07 Thread Nick Fisk
Just to add if you really want to go with lots of HDD's to Journals then go NVME. They are not a lot more expensive than the equivalent SATA based 3700's, but the latency is low low low. Here is an example of a node I have just commissioned with 12 HDD's to one P3700 Device: rrqm/s wrqm/

Re: [ceph-users] suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

2016-07-05 Thread Nick Fisk
> -Original Message- > From: Alex Gorbachev [mailto:a...@iss-integration.com] > Sent: 04 July 2016 22:00 > To: Nick Fisk > Cc: Oliver Dzombic ; ceph-users us...@lists.ceph.com>; mq ; Christian Balzer > > Subject: Re: [ceph-users] > suse_enterprise_storage3_rbd

Re: [ceph-users] Is anyone seeing iissues with task_numa_find_cpu?

2016-07-05 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Alex Gorbachev > Sent: 04 July 2016 20:50 > To: Campbell Steven > Cc: ceph-users ; Tim Bishop li...@bishnet.net> > Subject: Re: [ceph-users] Is anyone seeing iissues with > task_numa_find_cpu

Re: [ceph-users] suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

2016-07-04 Thread Nick Fisk
> On 2016-07-01T19:11:34, Nick Fisk wrote: > > > To summarise, > > > > LIO is just not working very well at the moment because of the ABORT > > Tasks problem, this will hopefully be fixed at some point. I'm not > > sure if SUSE works around this, but see

Re: [ceph-users] RBD mirroring between a IPv6 and IPv4 Cluster

2016-07-04 Thread Nick Fisk
gt; > > > Op 4 juli 2016 om 9:25 schreef Nick Fisk : > > > > > > Hi All, > > > > Quick question. I'm currently in the process of getting ready to > > deploy a 2nd cluster, which at some point in the next 12 months, I > > will want to enable

[ceph-users] RBD mirroring between a IPv6 and IPv4 Cluster

2016-07-04 Thread Nick Fisk
Hi All, Quick question. I'm currently in the process of getting ready to deploy a 2nd cluster, which at some point in the next 12 months, I will want to enable RBD mirroring between the new and existing clusters. I'm leaning towards deploying this new cluster with IPv6, because Wido says so ;-)

Re: [ceph-users] suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

2016-07-04 Thread Nick Fisk
> -Original Message- > From: mq [mailto:maoqi1...@126.com] > Sent: 04 July 2016 08:13 > To: Nick Fisk > Subject: Re: [ceph-users] > suse_enterprise_storage3_rbd_LIO_vmware_performance_bad > > Hi Nick > i have test NFS: since NFS cannot choose Eager Zeroed Thic

Re: [ceph-users] suse_enterprise_storage3_rbd_LIO_vmware_performance_bad

2016-07-01 Thread Nick Fisk
To summarise, LIO is just not working very well at the moment because of the ABORT Tasks problem, this will hopefully be fixed at some point. I'm not sure if SUSE works around this, but see below for other pain points with RBD + ESXi + iSCSI TGT is easy to get going, but performance isn't the b

Re: [ceph-users] Pinpointing performance bottleneck / would SSD journals help?

2016-06-27 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Daniel Schneller > Sent: 27 June 2016 17:33 > To: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Pinpointing performance bottleneck / would SSD > journals help? > > On 2016-06-27 16:01:

[ceph-users] Does flushbufs on a rbd-nbd invalidate librbd cache?

2016-06-21 Thread Nick Fisk
Hi All, Does anybody know if calling a blockdev --flushbufs on a rbd-nbd device causes the librbd read cache to be invalidated? I've done a quick test and the invalidate_cache counter doesn't increment like when you send the invalidate command via the admin socket. Thanks, Nick

Re: [ceph-users] Ceph OSD journal utilization

2016-06-20 Thread Nick Fisk
There is a journal_bytes counter for the OSD, or the journal_full counter will show if you have ever filled the journal. From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of David Turner Sent: 20 June 2016 17:02 To: EP Komarla ; ceph-users@lists.ceph.com Subject: Re: [ceph-u

Re: [ceph-users] Switches and latency

2016-06-15 Thread Nick Fisk
> -Original Message- > From: Gandalf Corvotempesta [mailto:gandalf.corvotempe...@gmail.com] > Sent: 15 June 2016 22:13 > To: n...@fisk.me.uk > Cc: ceph-us...@ceph.com > Subject: Re: [ceph-users] Switches and latency > > 2016-06-15 22:59 GMT+02:00 Nick Fisk : >

Re: [ceph-users] Switches and latency

2016-06-15 Thread Nick Fisk
> -Original Message- > From: Gandalf Corvotempesta [mailto:gandalf.corvotempe...@gmail.com] > Sent: 15 June 2016 21:33 > To: n...@fisk.me.uk > Cc: ceph-us...@ceph.com > Subject: Re: [ceph-users] Switches and latency > > 2016-06-15 22:13 GMT+02:00 Nick Fisk : >

Re: [ceph-users] Switches and latency

2016-06-15 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Gandalf Corvotempesta > Sent: 15 June 2016 17:03 > To: ceph-us...@ceph.com > Subject: [ceph-users] Switches and latency > > Let's assume a fully redundant network. > We need 4 switches, 2 for

Re: [ceph-users] strange cache tier behaviour with cephfs

2016-06-14 Thread Nick Fisk
086107 Am 14.06.2016 um 07:47 schrieb Nick Fisk: > osd_tier_promote_max_objects_sec > and > osd_tier_promote_max_bytes_sec > > is what you are looking for, I think by default its set to 5MB/s, which > would roughly correlate to why you are only seeing around 8 objects each

Re: [ceph-users] strange unfounding of PGs

2016-06-13 Thread Nick Fisk
Did you enable the sortbitwise flag as per the upgrade instructions, as there is a known bug with it? I don't know why these instructions haven't been amended in light of this bug. http://tracker.ceph.com/issues/16113 > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@

Re: [ceph-users] strange cache tier behaviour with cephfs

2016-06-13 Thread Nick Fisk
osd_tier_promote_max_objects_sec and osd_tier_promote_max_bytes_sec is what you are looking for, I think by default its set to 5MB/s, which would roughly correlate to why you are only seeing around 8 objects each time being promoted. This was done like this as too many promotions hurt performance,

Re: [ceph-users] OSPF to the host

2016-06-09 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Saverio Proto > Sent: 09 June 2016 11:38 > To: n...@fisk.me.uk > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] OSPF to the host > > > Has anybody had any experience with running th

Re: [ceph-users] OSPF to the host

2016-06-06 Thread Nick Fisk
> -Original Message- > From: Luis Periquito [mailto:periqu...@gmail.com] > Sent: 06 June 2016 14:30 > To: Nick Fisk > Cc: Ceph Users > Subject: Re: [ceph-users] OSPF to the host > > Nick, > > TL;DR: works brilliantly :) Excellent, just what I wanted to

[ceph-users] OSPF to the host

2016-06-06 Thread Nick Fisk
Hi All, Has anybody had any experience with running the network routed down all the way to the host? I know the standard way most people configured their OSD nodes is to bond the two nics which will then talk via a VRRP gateway and then probably from then on the networking is all Layer3. T

Re: [ceph-users] 2 networks vs 2 NICs

2016-06-04 Thread Nick Fisk
Yes, this is fine. I currently use 2 bonded 10G nics which have the untagged vlan as the public network and a tagged vlan as the cluster network. However, when I build my next cluster I will probably forgo the separate cluster network and just run them over the same IP, as after running the clu

Re: [ceph-users] Best Network Switches for Redundancy

2016-06-01 Thread Nick Fisk
Just a couple of points. 1. I know you said 10G was not an option, but I would really push for it. You can pick up Dell 10G-T switches (N4032) for not a lot more than a 48 port 1G switch. They make a lot more difference than just 10x the bandwidth. With Ceph latency is critical. As its 10G-T, you

Re: [ceph-users] Fwd: [Ceph-community] Wasting the Storage capacity when using Ceph based On high-end storage systems

2016-05-31 Thread Nick Fisk
rift: > > IP Interactive UG ( haftungsbeschraenkt ) Zum Sonnenberg 1-3 > 63571 Gelnhausen > > HRB 93402 beim Amtsgericht Hanau > Geschäftsführung: Oliver Dzombic > > Steuer Nr.: 35 236 3622 1 > UST ID: DE274086107 > > > Am 31.05.2016 um 13:05 schrieb Nick

Re: [ceph-users] Fwd: [Ceph-community] Wasting the Storage capacity when using Ceph based On high-end storage systems

2016-05-31 Thread Nick Fisk
involved, so make ur own picture :-) > > > -- > Mit freundlichen Gruessen / Best regards > > Oliver Dzombic > IP-Interactive > > mailto:i...@ip-interactive.de > > Anschrift: > > IP Interactive UG ( haftungsbeschraenkt ) Zum Sonnenberg 1-3 > 63571 G

Re: [ceph-users] Fwd: [Ceph-community] Wasting the Storage capacity when using Ceph based On high-end storage systems

2016-05-31 Thread Nick Fisk
involved, so make ur own picture :-) > > > -- > Mit freundlichen Gruessen / Best regards > > Oliver Dzombic > IP-Interactive > > mailto:i...@ip-interactive.de > > Anschrift: > > IP Interactive UG ( haftungsbeschraenkt ) Zum Sonnenberg 1-3 > 63571 G

Re: [ceph-users] Fwd: [Ceph-community] Wasting the Storage capacity when using Ceph based On high-end storage systems

2016-05-31 Thread Nick Fisk
Steuer Nr.: 35 236 3622 1 > UST ID: DE274086107 > > > Am 30.05.2016 um 17:13 schrieb Christian Balzer: > > > > Hello, > > > > On Mon, 30 May 2016 09:40:11 +0100 Nick Fisk wrote: > > > >> The other option is to scale out rather than scale up. I&

Re: [ceph-users] Fwd: [Ceph-community] Wasting the Storage capacity when using Ceph based On high-end storage systems

2016-05-30 Thread Nick Fisk
The other option is to scale out rather than scale up. I'm currently building nodes based on a fast Xeon E3 with 12 Drives in 1U. The MB/CPU is very attractively priced and the higher clock gives you much lower write latency if that is important. The density is slightly lower, but I guess you ga

Re: [ceph-users] RBD removal issue

2016-05-23 Thread Nick Fisk
See here: http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Adrian Saul > Sent: 23 May 2016 09:37 > To: 'ceph-users@lists.ceph.com' > Subject: [ceph-users] RBD removal issue

Re: [ceph-users] NVRAM cards as OSD journals

2016-05-23 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Heath Albritton > Sent: 23 May 2016 01:24 > To: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] NVRAM cards as OSD journals > > I'm contemplating the same thing as well. Or rather, I'

Re: [ceph-users] Erasure pool performance expectations

2016-05-16 Thread Nick Fisk
> -Original Message- > From: Peter Kerdisle [mailto:peter.kerdi...@gmail.com] > Sent: 16 May 2016 12:14 > To: Nick Fisk > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Erasure pool performance expectations > > > > On Mon, May 16, 2016

Re: [ceph-users] Erasure pool performance expectations

2016-05-16 Thread Nick Fisk
> -Original Message- > From: Peter Kerdisle [mailto:peter.kerdi...@gmail.com] > Sent: 16 May 2016 11:04 > To: Nick Fisk > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Erasure pool performance expectations > > > On Mon, May 16, 2016 at

Re: [ceph-users] Erasure pool performance expectations

2016-05-16 Thread Nick Fisk
ncy for reads to be higher than the max hit_set count. Then in theory no reads will ever cause an object to be promoted. Nick > > > > On Mon, May 16, 2016 at 11:14 AM Nick Fisk wrote: > > -Original Message- > > From: Peter Kerdisle [mailto:peter.kerdi...@gmail.co

Re: [ceph-users] Erasure pool performance expectations

2016-05-16 Thread Nick Fisk
> -Original Message- > From: Peter Kerdisle [mailto:peter.kerdi...@gmail.com] > Sent: 15 May 2016 08:04 > To: Nick Fisk > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Erasure pool performance expectations > > Hey Nick, > > I&#

Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on lockfile

2016-05-16 Thread Nick Fisk
> -Original Message- > From: Eric Eastman [mailto:eric.east...@keepertech.com] > Sent: 11 May 2016 16:02 > To: Nick Fisk > Cc: Ceph Users > Subject: Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on > lockfile > > On Wed, May 11, 2016 a

Re: [ceph-users] Multiple backend pools on the same cacher tier pool ?

2016-05-13 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Haomai Wang > Sent: 13 May 2016 15:00 > To: Florent B > Cc: Ceph Users > Subject: Re: [ceph-users] Multiple backend pools on the same cacher tier > pool ? > > On Fri, May 13, 2016 at 8:11 PM

Re: [ceph-users] Weighted Priority Queue testing

2016-05-11 Thread Nick Fisk
> -Original Message- > From: Mark Nelson [mailto:mnel...@redhat.com] > Sent: 11 May 2016 13:16 > To: Somnath Roy ; Nick Fisk > ; Ben England ; Kyle Bader > > Cc: Sage Weil ; Samuel Just ; ceph- > us...@lists.ceph.com > Subject: Re: Weighted Priority Queue testi

Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on lockfile

2016-05-11 Thread Nick Fisk
> -Original Message- > From: Eric Eastman [mailto:eric.east...@keepertech.com] > Sent: 10 May 2016 18:29 > To: Nick Fisk > Cc: Ceph Users > Subject: Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on > lockfile > > On Tue, May 10, 2016 a

Re: [ceph-users] Erasure pool performance expectations

2016-05-11 Thread Nick Fisk
Hi Peter, yes just restart the OSD for the setting to take effect. From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Peter Kerdisle Sent: 10 May 2016 19:06 To: Nick Fisk Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Erasure pool performance expectations

Re: [ceph-users] Erasure pool performance expectations

2016-05-10 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Peter Kerdisle > Sent: 10 May 2016 14:37 > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Erasure pool performance expectations > > To answer my own question it seems that you can

Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on lockfile

2016-05-10 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Nick Fisk > Sent: 10 May 2016 13:30 > To: 'Eric Eastman' > Cc: 'Ceph Users' > Subject: Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout o

Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on lockfile

2016-05-10 Thread Nick Fisk
> -Original Message- > From: Eric Eastman [mailto:eric.east...@keepertech.com] > Sent: 09 May 2016 23:09 > To: Nick Fisk > Cc: Ceph Users > Subject: Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on > lockfile > > On Mon, May 9, 2016 at 3:28 PM, Ni

Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on lockfile

2016-05-09 Thread Nick Fisk
Hi Eric, > -Original Message- > From: Eric Eastman [mailto:eric.east...@keepertech.com] > Sent: 09 May 2016 19:21 > To: Nick Fisk > Cc: Ceph Users > Subject: Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on > lockfile > > I am trying to do some

Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on lockfile

2016-05-09 Thread Nick Fisk
> -Original Message- > From: Ira Cooper [mailto:icoo...@redhat.com] > Sent: 09 May 2016 17:31 > To: Sage Weil > Cc: Nick Fisk ; ceph-users@lists.ceph.com > Subject: Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on > lockfile > - Original Message

[ceph-users] CephFS + CTDB/Samba - MDS session timeout on lockfile

2016-05-09 Thread Nick Fisk
Hi All, I've been testing an active/active Samba cluster over CephFS, performance seems really good with small files compared to Gluster. Soft reboots work beautifully with little to no interruption in file access. However when I perform a hard shutdown/reboot of one of the samba nodes, the remain

Re: [ceph-users] Erasure pool performance expectations

2016-05-03 Thread Nick Fisk
. No real way to do this yet. > > Mark > > On 05/03/2016 08:40 AM, Peter Kerdisle wrote: > > Thank you, I will attempt to play around with these settings and see > > if I can achieve better read performance. > > > > Appreciate your insights. > &g

Re: [ceph-users] Erasure pool performance expectations

2016-05-03 Thread Nick Fisk
r taking the time to answer my questions. Some in-line comments. > > On Tue, May 3, 2016 at 10:51 AM, Nick Fisk wrote: > Hi Peter, > > > > -Original Message- > > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf > Of > > Peter Kerd

Re: [ceph-users] Erasure pool performance expectations

2016-05-03 Thread Nick Fisk
Hi Peter, > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Peter Kerdisle > Sent: 02 May 2016 08:17 > To: ceph-users@lists.ceph.com > Subject: [ceph-users] Erasure pool performance expectations > > Hi guys, > > I am currently testing the

Re: [ceph-users] Slow read on RBD mount, Hammer 0.94.5

2016-04-20 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Udo Lembke > Sent: 20 April 2016 07:21 > To: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Slow read on RBD mount, Hammer 0.94.5 > > Hi Mike, > I don't have experiences with RBD moun

Re: [ceph-users] ceph cache tier clean rate too low

2016-04-20 Thread Nick Fisk
I would advise you to take a look at the osd_agent_max_ops (and osd_agent_max_ops), these should in theory dictate how many parallel threads will be used for flushing. Do a conf dump from the admin socket to see what you are currently running with and then bump them up to see if it helps. > ---

Re: [ceph-users] Deprecating ext4 support

2016-04-12 Thread Nick Fisk
Jan, I would like to echo Sage's response here. It seems you only want a subset of what Ceph offers, whereas RADOS is designed to offer a whole lot more, which requires a lot more intelligence at the lower levels. I must say I have found your attitude to both Sage and the Ceph project as a whole

Re: [ceph-users] How can I monitor current ceph operation at cluster

2016-04-11 Thread Nick Fisk
ar. > > > > Bob > > > > On Thu, Apr 7, 2016 at 2:34 AM, David Riedl > wrote: > > > Hi. > > > > > > I use this for my zabbix environment: > > > > > > https://github.com/thelan/ceph-zabbix/ > > > > > > It wor

Re: [ceph-users] How can I monitor current ceph operation at cluster

2016-04-07 Thread Nick Fisk
> Hi. > I have small question about monitoring performance at ceph cluster. > > We have cluster with 5 nodes and 8 drives on each node, and 5 monitor on > every node. For monitoring cluster we use zabbix. It asked every node for 30 > second about current ceph operation and get different result f

Re: [ceph-users] Performance counters oddities, cache tier and otherwise

2016-04-07 Thread Nick Fisk
Hi Christian, > Hello, > > Ceph 0.94.5 for the record. > > As some may remember, I phased in a 2TB cache tier 5 weeks ago. > > About now it has reached about 60% usage, which is what I have the > cache_target_dirty_ratio set to. > > And for the last 3 days I could see some writes (op_in_byte

[ceph-users] Incorrect path in /etc/init/ceph-osd.conf?

2016-03-30 Thread Nick Fisk
Hi All, I can see the path in the upstart script https://github.com/ceph/ceph/blob/master/src/upstart/ceph-osd.conf Checks for the file /etc/default/ceph and then runs it But in all the instances of ceph that I have installed that location is a directory and the actual location of t

Re: [ceph-users] Local SSD cache for ceph on each compute node.

2016-03-30 Thread Nick Fisk
> >> > >> On 03/29/2016 04:35 PM, Nick Fisk wrote: > >>> One thing I picked up on when looking at dm-cache for doing caching > >>> with RBD's is that it wasn't really designed to be used as a > >>> writeback cache for new writes, a

Re: [ceph-users] Redirect snapshot COW to alternative pool

2016-03-30 Thread Nick Fisk
> > > > > I think this is where I see slow performance. If you are doing > > > > large IO, then copying 4MB objects (assuming defaults) is maybe > > > > only 2x times the original IO to the disk. However if you are > > > > doing smaller IO from what I can see a single 4kb write would lead > > > >

Re: [ceph-users] Redirect snapshot COW to alternative pool

2016-03-29 Thread Nick Fisk
> > I think this is where I see slow performance. If you are doing large > > IO, then copying 4MB objects (assuming defaults) is maybe only 2x > > times the original IO to the disk. However if you are doing smaller IO > > from what I can see a single 4kb write would lead to a 4MB object > > being

Re: [ceph-users] Dump Historic Ops Breakdown

2016-03-29 Thread Nick Fisk
> Been a while, but... Brilliant, just what I needed to know. Thanks for the confirmation/answers. > > On Thu, Feb 25, 2016 at 9:50 AM, Nick Fisk wrote: > > I'm just trying to understand the steps each IO goes through and have > > been looking at the output dump his

Re: [ceph-users] Redirect snapshot COW to alternative pool

2016-03-29 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Gregory Farnum > Sent: 29 March 2016 18:52 > To: Nick Fisk > Cc: Ceph Users > Subject: Re: [ceph-users] Redirect snapshot COW to alternative pool > > On Sat, Ma

Re: [ceph-users] Local SSD cache for ceph on each compute node.

2016-03-29 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Ric Wheeler > Sent: 29 March 2016 14:40 > To: Nick Fisk ; 'Sage Weil' > Cc: ceph-users@lists.ceph.com; device-mapper development de...@redhat.com> > Subject

Re: [ceph-users] Local SSD cache for ceph on each compute node.

2016-03-29 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Ric Wheeler > Sent: 29 March 2016 14:07 > To: Sage Weil > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Local SSD cache for ceph on each compute node. > > On 03/29/2016 03:42 PM,

[ceph-users] Redirect snapshot COW to alternative pool

2016-03-26 Thread Nick Fisk
Evening All, I've been testing the RBD snapshot functionality and one thing that I have seen is that once you take a snapshot of a RBD and perform small random IO on the original RBD, performance is really bad due to the amount of write amplification going on doing the COW's. ie every IO to the

Re: [ceph-users] data corruption with hammer

2016-03-19 Thread Nick Fisk
There is got to be something else going on here. All that PR does is to potentially delay the promotion to hit_set_period*recency instead of just doing it on the 2nd read regardless, it's got to be uncovering another bug. Do you see the same problem if the cache is in writeback mode before you s

Re: [ceph-users] Local SSD cache for ceph on each compute node.

2016-03-19 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Daniel Niasoff > Sent: 16 March 2016 21:02 > To: Nick Fisk ; 'Van Leeuwen, Robert' > ; 'Jason Dillaman' > Cc: ceph-users@lists.ceph.com > Subjec

Re: [ceph-users] data corruption with hammer

2016-03-19 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Irek Fasikhov > Sent: 17 March 2016 13:00 > To: Sage Weil > Cc: Robert LeBlanc ; ceph-users us...@lists.ceph.com>; Nick Fisk ; William Perkins > > Subject: Re: [

Re: [ceph-users] RBD hanging on some volumes of a pool

2016-03-19 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Adrien Gillard > Sent: 17 March 2016 10:23 > To: ceph-users > Subject: [ceph-users] RBD hanging on some volumes of a pool > > Hi, > > I am facing issues with some of my rbd volumes since yes

Re: [ceph-users] SSDs for journals vs SSDs for a cache tier, which is better?

2016-03-19 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Stephen Harker > Sent: 16 March 2016 16:22 > To: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] SSDs for journals vs SSDs for a cache tier, which is > better? > > On 2016-02-17 11:07, C

Re: [ceph-users] data corruption with hammer

2016-03-16 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Christian Balzer > Sent: 16 March 2016 07:08 > To: Robert LeBlanc > Cc: Robert LeBlanc ; ceph-users us...@lists.ceph.com>; William Perkins > Subject: Re: [ceph-users] data corruption with

Re: [ceph-users] Local SSD cache for ceph on each compute node.

2016-03-16 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Daniel Niasoff > Sent: 16 March 2016 08:26 > To: Van Leeuwen, Robert ; Jason Dillaman > > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Local SSD cache for ceph on each compute n

Re: [ceph-users] how to choose EC plugins and rulesets

2016-03-10 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Yoann Moulin > Sent: 10 March 2016 08:38 > To: Nick Fisk ; ceph-us...@ceph.com > Subject: Re: [ceph-users] how to choose EC plugins and rulesets > > Le 10/03/2016

Re: [ceph-users] how to choose EC plugins and rulesets

2016-03-10 Thread Nick Fisk
What is your intended use case RBD/FS/RGW? There are no major improvements in Jewel that I am aware of. The big one will be when EC pools allow direct partial overwrites without the use of a cache tier. > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On B

Re: [ceph-users] Cache tier operation clarifications

2016-03-07 Thread Nick Fisk
Hi Christian, > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Christian Balzer > Sent: 07 March 2016 02:22 > To: ceph-users > Subject: Re: [ceph-users] Cache tier operation clarifications > > > Hello, > > I'd like to get some insights,

Re: [ceph-users] Ceph RBD latencies

2016-03-03 Thread Nick Fisk
You can also dump the historic ops from the OSD admin socket. It will give a brief overview of each step and how long each one is taking. But generally what you are seeing is not unusual. Currently best case for a RBD on a replicated pool will be somewhere between 200-500 iops. The Ceph code is

Re: [ceph-users] Cache tier weirdness

2016-03-01 Thread Nick Fisk
Interesting... see below > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Christian Balzer > Sent: 01 March 2016 08:20 > To: ceph-users@lists.ceph.com > Cc: Nick Fisk > Subject: Re: [ceph-users] Cache tier weirdness

Re: [ceph-users] Guest sync write iops so poor.

2016-02-26 Thread Nick Fisk
Thanks Jan, that is an excellent explanation. From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Jan Schermer Sent: 26 February 2016 10:07 To: Huan Zhang Cc: josh durgin ; Nick Fisk ; ceph-users Subject: Re: [ceph-users] Guest sync write iops so poor. O_DIRECT is

Re: [ceph-users] Guest sync write iops so poor.

2016-02-26 Thread Nick Fisk
s-boun...@lists.ceph.com] On Behalf Of > Huan Zhang > Sent: 26 February 2016 09:30 > To: Nick Fisk > Cc: josh durgin ; ceph-users us...@ceph.com> > Subject: Re: [ceph-users] Guest sync write iops so poor. > > Hi Nick, > DB's IO pattern depends on config, mysql for example. >

Re: [ceph-users] Cache tier weirdness

2016-02-26 Thread Nick Fisk
Hi Christian, > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Christian Balzer > Sent: 26 February 2016 09:07 > To: ceph-users@lists.ceph.com > Subject: [ceph-users] Cache tier weirdness > > > Hello, > > still my test cluster with 0.94.6

Re: [ceph-users] Guest sync write iops so poor.

2016-02-26 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Huan Zhang > Sent: 26 February 2016 06:50 > To: Jason Dillaman > Cc: josh durgin ; Nick Fisk ; > ceph-users > Subject: Re: [ceph-users] Guest sync write iops so po

[ceph-users] Dump Historic Ops Breakdown

2016-02-25 Thread Nick Fisk
I'm just trying to understand the steps each IO goes through and have been looking at the output dump historic ops command from the admin socket. There's a couple of steps I'm not quite sure what they mean and also slightly puzzled by the delay and was wondering if anybody could share some knowledg

Re: [ceph-users] List of SSDs

2016-02-25 Thread Nick Fisk
There’s two factors really 1. Suitability for use in ceph 2. Number of people using them For #1, there are a number of people using various different drives, so lots of options. The blog articled linked is a good place to start. For #2 and I think this is quite important.

Re: [ceph-users] ceph hammer : rbd info/Status : operation not supported (95) (EC+RBD tier pools)

2016-02-25 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Jason Dillaman > Sent: 25 February 2016 01:30 > To: Christian Balzer > Cc: ceph-us...@ceph.com > Subject: Re: [ceph-users] ceph hammer : rbd info/Status : operation not > supported (95) (EC+

Re: [ceph-users] Guest sync write iops so poor.

2016-02-25 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Huan Zhang > Sent: 25 February 2016 11:11 > To: josh.dur...@inktank.com > Cc: ceph-users > Subject: [ceph-users] Guest sync write iops so poor. > > Hi, >We test sync iops with fio sync=

Re: [ceph-users] OSD Journal size config

2016-02-18 Thread Nick Fisk
> -Original Message- > From: M Ranga Swami Reddy [mailto:swamire...@gmail.com] > Sent: 18 February 2016 13:44 > To: Nick Fisk > Subject: Re: [ceph-users] OSD Journal size config > > > Hello All, > > I have increased my cluster's OSD journal size from

Re: [ceph-users] OSD Journal size config

2016-02-18 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > M Ranga Swami Reddy > Sent: 18 February 2016 12:09 > To: ceph-users > Subject: [ceph-users] OSD Journal size config > > Hello All, > I have increased my cluster's OSD journal size from 2GB

[ceph-users] OSD Performance Counters

2016-02-18 Thread Nick Fisk
Hi All, Could someone please sanity check this for me please. I trying to get my head round what counter reflect what and how they correlate to end user performance. In the attached graph I am graphing averages of the counters across all OSD's on one host Blue = osd.w_op_latency Red = Max of abo

Re: [ceph-users] Recomendations for building 1PB RadosGW with Erasure Code

2016-02-17 Thread Nick Fisk
Ah typo, I meant to say 10Mhz per IO. So a 7.2k disk does around 80IOPs = ~ 800mhz which is close to the 1Ghz figure. From: John Hogenmiller [mailto:j...@hogenmiller.net] Sent: 17 February 2016 13:15 To: Nick Fisk Cc: Василий Ангапов ; ceph-users@lists.ceph.com Subject: Re: [ceph-users

Re: [ceph-users] Recomendations for building 1PB RadosGW with Erasure Code

2016-02-17 Thread Nick Fisk
ent HDD usage (replication vs. EC). > > Plus: A lot less expensive SSDs, less CPU and RAM requirements, smaller > impact in case of node failure/maintenance. > > No ideas about the stuff below. > > Christian > > Also, am I right that for 6+3 EC profile i need at least 1

Re: [ceph-users] SSDs for journals vs SSDs for a cache tier, which is better?

2016-02-17 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Christian Balzer > Sent: 17 February 2016 04:22 > To: ceph-users@lists.ceph.com > Cc: Piotr Wachowicz > Subject: Re: [ceph-users] SSDs for journals vs SSDs for a cache tier, which is > better?

Re: [ceph-users] Recomendations for building 1PB RadosGW with Erasure Code

2016-02-17 Thread Nick Fisk
Thanks for posting your experiences John, very interesting read. I think the golden rule of around 1Ghz is still a realistic goal to aim for. It looks like you probably have around 16ghz for 60OSD's, or 0.26Ghz per OSD. Do you have any idea on how much CPU you think you would need to just be abl

Re: [ceph-users] Recomendations for building 1PB RadosGW with Erasure Code

2016-02-16 Thread Nick Fisk
> -Original Message- > From: Василий Ангапов [mailto:anga...@gmail.com] > Sent: 16 February 2016 13:15 > To: Tyler Bishop > Cc: Nick Fisk ; us...@lists.ceph.com> > Subject: Re: [ceph-users] Recomendations for building 1PB RadosGW with > Erasure Code > &

Re: [ceph-users] Recomendations for building 1PB RadosGW with Erasure Code

2016-02-16 Thread Nick Fisk
get a fairly good idea about the hardware requirements from this. > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Nick Fisk > Sent: 16 February 2016 08:12 > To: 'Tyler Bishop' ; 'Василий Ангапов' > >

Re: [ceph-users] Recomendations for building 1PB RadosGW with Erasure Code

2016-02-16 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Tyler Bishop > Sent: 16 February 2016 04:20 > To: Василий Ангапов > Cc: ceph-users > Subject: Re: [ceph-users] Recomendations for building 1PB RadosGW with > Erasure Code > > You should lo

Re: [ceph-users] cls_rbd ops on rbd_id.$name objects in EC pool

2016-02-12 Thread Nick Fisk
> -Original Message- > From: Nick Fisk [mailto:n...@fisk.me.uk] > Sent: 12 February 2016 13:31 > To: 'Sage Weil' > Cc: 'Jason Dillaman' ; 'Samuel Just' > ; ceph-users@lists.ceph.com; ceph- > de...@vger.kernel.org > Subje

Re: [ceph-users] Reducing the impact of OSD restarts (noout ain't uptosnuff)

2016-02-12 Thread Nick Fisk
f you received this message erroneously, please notify the sender and > delete it, together with any attachments. > > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Nick Fisk > Sent: Friday, February 12, 2016 9:07 AM &

Re: [ceph-users] Reducing the impact of OSD restarts (noout ain't uptosnuff)

2016-02-12 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Christian Balzer > Sent: 12 February 2016 15:38 > To: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Reducing the impact of OSD restarts (noout ain't > uptosnuff) > > On Fri, 12 Feb 2

Re: [ceph-users] cls_rbd ops on rbd_id.$name objects in EC pool

2016-02-12 Thread Nick Fisk
> -Original Message- > From: ceph-devel-ow...@vger.kernel.org [mailto:ceph-devel- > ow...@vger.kernel.org] On Behalf Of Sage Weil > Sent: 12 February 2016 13:15 > To: Nick Fisk > Cc: 'Jason Dillaman' ; 'Samuel Just' > ; ceph-users@lists.ceph.com

Re: [ceph-users] ceph 9.2.0 SAMSUNG ssd performance issue?

2016-02-12 Thread Nick Fisk
I will do my best to answer, but some of the questions are starting to stretch the limit of my knowledge > -Original Message- > From: Huan Zhang [mailto:huan.zhang...@gmail.com] > Sent: 12 February 2016 12:15 > To: Nick Fisk > Cc: Irek Fasikhov ; ceph-users us...@ceph

<    1   2   3   4   5   6   7   8   >