Re: [ceph-users] Monitoring bluestore compression ratio

2018-07-02 Thread Blair Bethwaite
Oh, for some reason i thought you'd mentioned the OSD config earlier here. Gald you figured it out anyway! Are you doing any comparison benchmarks with/without compression? There is precious little (no?) info out there about performance impact... Cheers, Blair On 3 Jul. 2018 03:18, "David

Re: [ceph-users] Monitoring bluestore compression ratio

2018-06-26 Thread Blair Bethwaite
Hi, Zeros are not a great choice of data for testing a storage system unless you are specifically testing what it does with zeros. Ceph knows that other higher layers in the storage stack use zero-fill for certain things and will probably optimise for it. E.g., it's common for thin-provisioning

Re: [ceph-users] Planning all flash cluster

2018-06-20 Thread Blair Bethwaite
This is true, but misses the point that the OP is talking about old hardware already - you're not going to save much money on removing a 2nd hand CPU from a system. On Wed, 20 Jun 2018 at 22:10, Wido den Hollander wrote: > > > On 06/20/2018 02:00 PM, Robert Sander wrote: > > On 20.06.2018

Re: [ceph-users] Multi-MDS Failover

2018-05-19 Thread Blair Bethwaite
On 19 May 2018 at 09:20, Scottix wrote: > It would be nice to have an option to have all IO blocked if it hits a > degraded state until it recovers. Since you are unaware of other MDS state, > seems like that would be tough to do. I agree this would be a nice knob to have

Re: [ceph-users] Intel Xeon Scalable and CPU frequency scaling on NVMe/SSD Ceph OSDs

2018-05-16 Thread Blair Bethwaite
th that generation of CPUs. > But 7proc/cpuinfo never lies (does it ?) > > > > > On 16 May 2018 at 13:22, Blair Bethwaite <blair.bethwa...@gmail.com> > wrote: > >> On 15 May 2018 at 08:45, Wido den Hollander <w...@42on.com> wrote: >>> >>

Re: [ceph-users] Intel Xeon Scalable and CPU frequency scaling on NVMe/SSD Ceph OSDs

2018-05-15 Thread Blair Bethwaite
Sorry, bit late to get back to this... On Wed., 2 May 2018, 06:19 Nick Fisk, wrote: > 4.16 required? > Looks like it - thanks for pointing that out. Wido, I don't think you are doing anything wrong here, maybe this is a bug... I've got RHEL7 + Broadwell based Ceph nodes here

Re: [ceph-users] Intel Xeon Scalable and CPU frequency scaling on NVMe/SSD Ceph OSDs

2018-05-01 Thread Blair Bethwaite
Also curious about this over here. We've got a rack's worth of R740XDs with Xeon 4114's running RHEL 7.4 and intel-pstate isn't even active on them, though I don't believe they are any different at the OS level to our Broadwell nodes (where it is loaded). Have you tried poking the kernel's pmqos

Re: [ceph-users] Poor read performance.

2018-04-26 Thread Blair Bethwaite
On 26 April 2018 at 14:58, Jonathan D. Proulx wrote: > Those block queue scheduler tips *might* help me squeeze a bit more > till next budget starts July 1... Maybe you could pick up some cheap cache from this guy: https://xkcd.com/908/ -- Cheers, ~Blairo

Re: [ceph-users] Poor read performance.

2018-04-25 Thread Blair Bethwaite
Hi Jon, On 25 April 2018 at 21:20, Jonathan Proulx wrote: > > here's a snap of 24hr graph form one server (others are similar in > general shape): > > https://snapshot.raintank.io/dashboard/snapshot/gB3FDPl7uRGWmL17NHNBCuWKGsXdiqlt That's what, a median IOPs of about 80?

Re: [ceph-users] rbd feature overheads

2018-02-12 Thread Blair Bethwaite
Thanks Ilya, We can probably handle ~6.2MB for a 100TB volume. Is it reasonable to expect a librbd client such as QEMU to only hold one object-map per guest? Cheers, On 12 February 2018 at 21:01, Ilya Dryomov <idryo...@gmail.com> wrote: > On Mon, Feb 12, 2018 at 6:25 AM, Blair

[ceph-users] rbd feature overheads

2018-02-11 Thread Blair Bethwaite
Hi all, Wondering if anyone can clarify whether there are any significant overheads from rbd features like object-map, fast-diff, etc. I'm interested in both performance overheads from a latency and space perspective, e.g., can object-map be sanely deployed on a 100TB volume or does the client

Re: [ceph-users] OSD servers swapping despite having free memory capacity

2018-01-24 Thread Blair Bethwaite
On 25 January 2018 at 04:53, Warren Wang wrote: > The other thing I can think of is if you have OSDs locking up and getting > corrupted, there is a severe XFS bug where the kernel will throw a NULL > pointer dereference under heavy memory pressure. Again, it's due to

Re: [ceph-users] OSD servers swapping despite having free memory capacity

2018-01-23 Thread Blair Bethwaite
+1 to Warren's advice on checking for memory fragmentation. Are you seeing kmem allocation failures in dmesg on these hosts? On 24 January 2018 at 10:44, Warren Wang wrote: > Check /proc/buddyinfo for memory fragmentation. We have some pretty severe > memory frag issues

Re: [ceph-users] What is the should be the expected latency of 10Gbit network connections

2018-01-22 Thread Blair Bethwaite
Firstly, the OP's premise in asking, "Or should there be a differnce of 10x", is fundamentally incorrect. Greater bandwidth does not mean lower latency, though the latter almost always results in the former. Unfortunately, changing the speed of light remains a difficult engineering challenge :-).

Re: [ceph-users] Ceph cluster network bandwidth?

2017-11-16 Thread Blair Bethwaite
What type of SAS disks, spinners or SSD? You really need to specify the sustained write throughput of your OSD nodes if you want to figure out whether your network is sufficient/appropriate. At 3x replication if you want to sustain e.g. 1 GB/s of write traffic from clients then you will need 2

Re: [ceph-users] Ceph @ OpenStack Sydney Summit

2017-10-29 Thread Blair Bethwaite
shout out and/or add to https://etherpad.openstack.org/p/SYD-forum-Ceph-OpenStack-BoF. Also, hope to see some of the core team there! Cheers, On 7 July 2017 at 13:47, Blair Bethwaite <blair.bethwa...@gmail.com> wrote: > Hi all, > > Are there any "official" plans to h

[ceph-users] OpenStack Sydney Forum - Ceph BoF proposal

2017-09-29 Thread Blair Bethwaite
Hi all, I just submitted an OpenStack Forum proposal for a Ceph BoF session at OpenStack Sydney. If you're interested in seeing this happen then please hit up http://forumtopics.openstack.org/cfp/details/46 with your comments / +1's. -- Cheers, ~Blairo

Re: [ceph-users] Ceph release cadence

2017-09-22 Thread Blair Bethwaite
On 23 September 2017 at 11:58, Sage Weil wrote: > I'm *much* happier with 2 :) so no complaint from me. I just heard a lot > of "2 years" and 2 releases (18 months) doesn't quite cover it. Maybe > it's best to start with that, though? It's still an improvement over the >

Re: [ceph-users] Luminous BlueStore EC performance

2017-09-13 Thread Blair Bethwaite
; 823,912 5,641-76.41%4.24 >>> 16 24,587 5,643-77.05%4.36 >>> RW >>> 120,37911,166-45.21%1.83 >>> 234,246 9,525-72.19%3.60 >>> 833,195 9,300-71.98%3.57 >>> 16 31

[ceph-users] moving mons across networks

2017-09-12 Thread Blair Bethwaite
Hi all, We're looking at readdressing the mons (moving to a different subnet) on one of our clusters. Most of the existing clients are OpenStack guests on Libvirt+KVM and we have a major upgrade to do for those in coming weeks that will mean they have to go down briefly, that will give us an

Re: [ceph-users] RBD I/O errors with QEMU [luminous upgrade/osd change]

2017-09-12 Thread Blair Bethwaite
at's why did it :-) > > > Blair Bethwaite <blair.bethwa...@gmail.com> writes: > >> Great to see this issue sorted. >> >> I have to say I am quite surprised anyone would implement the >> export/import workaround mentioned here without *first* racing to this >> ML

Re: [ceph-users] Ceph release cadence

2017-09-11 Thread Blair Bethwaite
(Apologies if this is a double post - I think my phone turned it into HTML and so bounced from ceph-devel)... We currently use both upstream and distro (RHCS) versions on different clusters. Downstream releases are still free to apply their own models. I like the idea of a predictable (and more

Re: [ceph-users] Ceph release cadence

2017-09-11 Thread Blair Bethwaite
On 7 September 2017 at 01:23, Sage Weil wrote: > * Drop the odd releases, and aim for a ~9 month cadence. This splits the > difference between the current even/odd pattern we've been doing. > > + eliminate the confusing odd releases with dubious value > + waiting for the next

Re: [ceph-users] RBD I/O errors with QEMU [luminous upgrade/osd change]

2017-09-11 Thread Blair Bethwaite
Great to see this issue sorted. I have to say I am quite surprised anyone would implement the export/import workaround mentioned here without *first* racing to this ML or IRC and crying out for help. This is a valuable resource, made more so by people sharing issues. Cheers, On 12 September

Re: [ceph-users] [SOLVED] output discards (queue drops) on switchport

2017-09-11 Thread Blair Bethwaite
On 12 September 2017 at 01:15, Blair Bethwaite <blair.bethwa...@gmail.com> wrote: > Flow-control may well just mask the real problem. Did your throughput > improve? Also, does that mean flow-control is on for all ports on the > switch...? IIUC, then such "global pause&quo

Re: [ceph-users] [SOLVED] output discards (queue drops) on switchport

2017-09-11 Thread Blair Bethwaite
Flow-control may well just mask the real problem. Did your throughput improve? Also, does that mean flow-control is on for all ports on the switch...? IIUC, then such "global pause" flow-control will mean switchports with links to upstream network devices will also be paused if the switch is

[ceph-users] Luminous BlueStore EC performance

2017-09-05 Thread Blair Bethwaite
Hi all, (Sorry if this shows up twice - I got auto-unsubscribed and so first attempt was blocked) I'm keen to read up on some performance comparisons for replication versus EC on HDD+SSD based setups. So far the only recent thing I've found is Sage's Vault17 slides [1], which have a single slide

Re: [ceph-users] XFS attempt to access beyond end of device

2017-07-21 Thread Blair Bethwaite
Hi Brad, On 22 July 2017 at 09:04, Brad Hubbard wrote: > Could you share what kernel/distro you are running and also please test > whether > the error message can be triggered by running the "blkid" command? I'm seeing it on RHEL7.3 (3.10.0-514.2.2.el7.x86_64). See Red Hat

Re: [ceph-users] 答复: 答复: How's cephfs going?

2017-07-19 Thread Blair Bethwaite
alled active-standby mode? And could you give some >> information of your cephfs's usage pattern, for example, does your client >> nodes directly mount cephfs or mount it through an NFS, or something like >> it, running a directory that is mounted with cephfs and are you us

Re: [ceph-users] How's cephfs going?

2017-07-19 Thread Blair Bethwaite
Interesting. Any FUSE client data-points? On 19 July 2017 at 20:21, Дмитрий Глушенок <gl...@jet.msk.su> wrote: > RBD (via krbd) was in action at the same time - no problems. > > 19 июля 2017 г., в 12:54, Blair Bethwaite <blair.bethwa...@gmail.com> > написал(а): &g

Re: [ceph-users] How's cephfs going?

2017-07-19 Thread Blair Bethwaite
It would be worthwhile repeating the first test (crashing/killing an OSD host) again with just plain rados clients (e.g. rados bench) and/or rbd. It's not clear whether your issue is specifically related to CephFS or actually something else. Cheers, On 19 July 2017 at 19:32, Дмитрий Глушенок

Re: [ceph-users] 答复: How's cephfs going?

2017-07-17 Thread Blair Bethwaite
We are a data-intensive university, with an increasingly large fleet of scientific instruments capturing various types of data (mostly imaging of one kind or another). That data typically needs to be stored, protected, managed, shared, connected/moved to specialised compute for analysis. Given the

Re: [ceph-users] XFS attempt to access beyond end of device

2017-07-17 Thread Blair Bethwaite
Brilliant, thanks Marcus. We have just (noticed we've) hit this too and looks like your script will fix this (will test and report back...). On 18 July 2017 at 14:08, Marcus Furlong wrote: > [ 92.938882] XFS (sdi1): Mounting V5 Filesystem > [ 93.065393] XFS (sdi1): Ending

Re: [ceph-users] 答复: How's cephfs going?

2017-07-16 Thread Blair Bethwaite
so-called active-standby mode? And could you give some information of your > cephfs's usage pattern, for example, does your client nodes directly mount > cephfs or mount it through an NFS, or something like it, running a directory > that is mounted with cephfs and are you using ceph-fuse

Re: [ceph-users] How's cephfs going?

2017-07-16 Thread Blair Bethwaite
It works and can reasonably be called "production ready". However in Jewel there are still some features (e.g. directory sharding, multi active MDS, and some security constraints) that may limit widespread usage. Also note that userspace client support in e.g. nfs-ganesha and samba is a mixed bag

Re: [ceph-users] Ceph @ OpenStack Sydney Summit

2017-07-12 Thread Blair Bethwaite
t; it will be approved since I'm kind of a newcomer myself. I'd also like a Ceph > BoF. > > <3 Trilliams > > Sent from my iPhone > >> On Jul 6, 2017, at 10:50 PM, Blair Bethwaite <blair.bethwa...@gmail.com> >> wrote: >> >> Oops, this time pl

Re: [ceph-users] Ceph @ OpenStack Sydney Summit

2017-07-12 Thread Blair Bethwaite
Hi Greg, On 12 July 2017 at 03:48, Gregory Farnum wrote: > I poked at Patrick about this and it sounds like the venue is a little > smaller than usual (and community planning is a little less > planned-out for those ranges than usual) so things are still up in the > air. :/

Re: [ceph-users] Ceph @ OpenStack Sydney Summit

2017-07-06 Thread Blair Bethwaite
Oops, this time plain text... On 7 July 2017 at 13:47, Blair Bethwaite <blair.bethwa...@gmail.com> wrote: > > Hi all, > > Are there any "official" plans to have Ceph events co-hosted with OpenStack > Summit Sydney, like in Boston? > > The call for pres

[ceph-users] Ceph @ OpenStack Sydney Summit

2017-07-06 Thread Blair Bethwaite
Hi all, Are there any "official" plans to have Ceph events co-hosted with OpenStack Summit Sydney, like in Boston? The call for presentations closes in a week. The Forum will be organised throughout September and (I think) that is the most likely place to have e.g. Ceph ops sessions like we have

Re: [ceph-users] Removing very large buckets

2017-07-06 Thread Blair Bethwaite
How did you even get 60M objects into the bucket...?! The stuck requests are only likely to be impacting the PG in which the bucket index is stored. Hopefully you are not running other pools on those OSDs? You'll need to upgrade to Jewel and gain the --bypass-gc radosgw-admin flag, that speeds up

[ceph-users] risk mitigation in 2 replica clusters

2017-06-21 Thread Blair Bethwaite
Hi all, I'm doing some work to evaluate the risks involved in running 2r storage pools. On the face of it my naive disk failure calculations give me 4-5 nines for a 2r pool of 100 OSDs (no copyset awareness, i.e., secondary disk failure based purely on chance of any 1 of the remaining 99 OSDs

Re: [ceph-users] Cephalocon Cancelled

2017-05-15 Thread Blair Bethwaite
On 15 May 2017 at 23:21, Danny Al-Gaaf wrote: > What about moving the event to the next OpenStack Summit in Sydney, let > say directly following the Summit. +1! The Ceph day just gone at the Boston OpenStack Summit felt a lot like I imagined Cephalocon would be anyway,

Re: [ceph-users] Intel power tuning - 30% throughput performance increase

2017-05-04 Thread Blair Bethwaite
>> /dev/cpu_dma_latency (platform dependant) at startup and issue a >> warning, or whether documenting this would suffice? >> >> Any doc contribution would be welcomed. >> >> On Wed, May 3, 2017 at 7:18 PM, Blair Bethwaite >> <blair.bethwa...@gmail.com> wr

Re: [ceph-users] Intel power tuning - 30% throughput performance increase

2017-05-03 Thread Blair Bethwaite
On 3 May 2017 at 19:07, Dan van der Ster wrote: > Whether cpu_dma_latency should be 0 or 1, I'm not sure yet. I assume > your 30% boost was when going from throughput-performance to > dma_latency=0, right? I'm trying to understand what is the incremental > improvement from 1

Re: [ceph-users] Intel power tuning - 30% throughput performance increase

2017-05-03 Thread Blair Bethwaite
On 3 May 2017 at 18:38, Dan van der Ster wrote: > Seems to work for me, or? Yeah now that I read the code more I see it is opening and manipulating /dev/cpu_dma_latency in response to that option, so the TODO comment seems to be outdated. I verified tuned latency-performance

Re: [ceph-users] Intel power tuning - 30% throughput performance increase

2017-05-03 Thread Blair Bethwaite
On 3 May 2017 at 18:15, Dan van der Ster wrote: > It looks like el7's tuned natively supports the pmqos interface in > plugins/plugin_cpu.py. Ahha, you are right, but I'm sure I tested tuned and it did not help. Thanks for pointing out this script, I had not noticed it

Re: [ceph-users] Intel power tuning - 30% throughput performance increase

2017-05-03 Thread Blair Bethwaite
Hi Dan, On 3 May 2017 at 17:43, Dan van der Ster wrote: > We use cpu_dma_latency=1, because it was in the latency-performance profile. > And indeed by setting cpu_dma_latency=0 on one of our OSD servers, > powertop now shows the package as 100% in turbo mode. I tried both 0

[ceph-users] Intel power tuning - 30% throughput performance increase

2017-05-03 Thread Blair Bethwaite
Hi all, We recently noticed that despite having BIOS power profiles set to performance on our RHEL7 Dell R720 Ceph OSD nodes, that CPU frequencies never seemed to be getting into the top of the range, and in fact spent a lot of time in low C-states despite that BIOS option supposedly disabling

Re: [ceph-users] object store backup tool recommendations

2017-03-27 Thread Blair Bethwaite
I suppose the other option here, which I initially dismissed because Red Hat are not supporting it, is to have a CephFS dir/tree bound to a cache-tier fronted EC pool. Is anyone having luck with such a setup? On 3 March 2017 at 21:40, Blair Bethwaite <blair.bethwa...@gmail.com> wrote: >

Re: [ceph-users] object store backup tool recommendations

2017-03-27 Thread Blair Bethwaite
Thanks for the useful reply Robin and sorry for not getting back sooner... > On Fri, Mar 03, 2017 at 18:01:00 +, Robin H. Johnson wrote: > On Fri, Mar 03, 2017 at 10:55:06 +1100, Blair Bethwaite wrote: >> Does anyone have any recommendations for good tools to perform >>

Re: [ceph-users] object store backup tool recommendations

2017-03-03 Thread Blair Bethwaite
t; things could speed up, because file information is gotten from the mds > daemon, so this should save on one rsync file lookup, and we expect that > we can run more tasks in parallel. > > > > > > -----Original Message- > From: Blair Bethwaite [mailto:blair.bethwa...@gmai

[ceph-users] object store backup tool recommendations

2017-03-02 Thread Blair Bethwaite
Hi all, Does anyone have any recommendations for good tools to perform file-system/tree backups and restores to/from a RGW object store (Swift or S3 APIs)? Happy to hear about both FOSS and commercial options please. I'm interested in: 1) tools known to work or not work at all for a basic

Re: [ceph-users] crushtool mappings wrong

2017-02-16 Thread Blair Bethwaite
1 > max_size 10 > step take default > step chooseleaf firstn 0 type rack > step emit > } > > > On Thu, Feb 16, 2017 at 7:10 AM, Blair Bethwaite < > blair.bethwa...@gmail.com> wrote: > >> Am I going nuts (it is extremely late/early here), or is crushtool >&g

[ceph-users] crushtool mappings wrong

2017-02-16 Thread Blair Bethwaite
Am I going nuts (it is extremely late/early here), or is crushtool totally broken? I'm trying to configure a ruleset that will place exactly one replica into three different racks (under each of which there are 8-10 hosts). crushtool has given me empty mappings for just about every rule I've tried

Re: [ceph-users] EC pool migrations

2017-02-08 Thread Blair Bethwaite
is that you have enough space in your cluster for a replicated pool that will temporarily hold the intermediate data. > On 7 February 2017 at 23:50, Blair Bethwaite <blair.bethwa...@gmail.com> > wrote: >> 1) insert a large enough temporary replicated pool as a cache tier The cac

Re: [ceph-users] EC pool migrations

2017-02-07 Thread Blair Bethwaite
On 7 February 2017 at 23:50, Blair Bethwaite <blair.bethwa...@gmail.com> wrote: > 1) insert a large enough temporary replicated pool as a cache tier > 2) somehow force promotion of every object into the cache (don't see > any way to do that other than actually read them - bu

[ceph-users] EC pool migrations

2017-02-07 Thread Blair Bethwaite
Hi all, Wondering if anyone has come up with a quick and minimal impact way of moving data between erasure coded pools? We want to shrink an existing EC pool (also changing the EC profile at the same time) that backs our main RGW buckets. Thus far the only successful way I've found of managing

Re: [ceph-users] CephFS Samba VFS RHEL packages

2016-07-22 Thread Blair Bethwaite
Brett > > > On Fri, Jul 22, 2016 at 2:18 AM, Yan, Zheng <uker...@gmail.com> wrote: >> >> On Fri, Jul 22, 2016 at 11:15 AM, Blair Bethwaite >> <blair.bethwa...@gmail.com> wrote: >> > Thanks Zheng, >> > >> > On 22 July 2016 at 12:12,

Re: [ceph-users] CephFS Samba VFS RHEL packages

2016-07-21 Thread Blair Bethwaite
Thanks Zheng, On 22 July 2016 at 12:12, Yan, Zheng wrote: > We actively back-port fixes to RHEL 7.x kernel. When RHCS2.0 release, > the RHEL kernel should contain fixes up to 3.7 upstream kernel. You meant 4.7 right? -- Cheers, ~Blairo

Re: [ceph-users] CephFS Samba VFS RHEL packages

2016-07-21 Thread Blair Bethwaite
how it goes :) > > - Ken > > On Tue, Jul 19, 2016 at 11:45 PM, Blair Bethwaite > <blair.bethwa...@gmail.com> wrote: >> Hi all, >> >> We've started a CephFS Samba PoC on RHEL but just noticed the Samba >> Ceph VFS doesn't seem to be included with Samba on RHEL

[ceph-users] CephFS Samba VFS RHEL packages

2016-07-19 Thread Blair Bethwaite
Hi all, We've started a CephFS Samba PoC on RHEL but just noticed the Samba Ceph VFS doesn't seem to be included with Samba on RHEL, or we're not looking in the right place. Trying to avoid needing to build Samba from source if possible. Any pointers appreciated. -- Cheers, ~Blairo

Re: [ceph-users] Dramatic performance drop at certain number of objects in pool

2016-06-27 Thread Blair Bethwaite
On 25 Jun 2016 6:02 PM, "Kyle Bader" wrote: > fdatasync takes longer when you have more inodes in the slab caches, it's the double edged sword of vfs_cache_pressure. That's a bit sad when, iiuc, it's only journals doing fdatasync in the Ceph write path. I'd have expected

Re: [ceph-users] Dramatic performance drop at certain number of objects in pool

2016-06-22 Thread Blair Bethwaite
On 23 June 2016 at 12:37, Christian Balzer wrote: > Case in point, my main cluster (RBD images only) with 18 5+TB OSDs on 3 > servers (64GB RAM each) has 1.8 million 4MB RBD objects using about 7% of > the available space. > Don't think I could hit this problem before running out

Re: [ceph-users] Dramatic performance drop at certain number of objects in pool

2016-06-22 Thread Blair Bethwaite
recall the UnitedStack folks using 32MB. Cheers, On 23 June 2016 at 12:28, Christian Balzer <ch...@gol.com> wrote: > On Thu, 23 Jun 2016 12:01:38 +1000 Blair Bethwaite wrote: > >> On 23 June 2016 at 11:41, Wade Holler <wade.hol...@gmail.com> wrote: >> > Workload is nat

Re: [ceph-users] Dramatic performance drop at certain number of objects in pool

2016-06-22 Thread Blair Bethwaite
On 23 June 2016 at 11:41, Wade Holler wrote: > Workload is native librados with python. ALL 4k objects. Was that meant to be 4MB? -- Cheers, ~Blairo ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] Dramatic performance drop at certain number of objects in pool

2016-06-22 Thread Blair Bethwaite
culated threshold and a write occurs (maybe a read, I forget). >>> >> If it's a read a plain scrub might do the trick. >> >> Christian >>> Warren >>> >>> >>> From: ceph-users >>> <ceph-users-boun...@lists.ceph.com<mailto:cep

Re: [ceph-users] Dramatic performance drop at certain number of objects in pool

2016-06-20 Thread Blair Bethwaite
On 20 June 2016 at 09:21, Blair Bethwaite <blair.bethwa...@gmail.com> wrote: > slow request issues). If you watch your xfs stats you'll likely get > further confirmation. In my experience xs_dir_lookups balloons (which > means directory lookups are missing cache and going to d

Re: [ceph-users] Dramatic performance drop at certain number of objects in pool

2016-06-19 Thread Blair Bethwaite
Hi Wade, (Apologies for the slowness - AFK for the weekend). On 16 June 2016 at 23:38, Wido den Hollander wrote: > >> Op 16 juni 2016 om 14:14 schreef Wade Holler : >> >> >> Hi All, >> >> I have a repeatable condition when the object count in a pool gets to

Re: [ceph-users] Dramatic performance drop at certain number of objects in pool

2016-06-16 Thread Blair Bethwaite
Hi Wade, What IO are you seeing on the OSD devices when this happens (see e.g. iostat), are there short periods of high read IOPS where (almost) no writes occur? What does your memory usage look like (including slab)? Cheers, On 16 June 2016 at 22:14, Wade Holler wrote:

Re: [ceph-users] un-even data filled on OSDs

2016-06-09 Thread Blair Bethwaite
chable for a few mins/secs during the reblancing > task. Not sure, these low priority configurations are doing the job as > its. > > Thanks > Swami > > On Thu, Jun 9, 2016 at 5:50 PM, Blair Bethwaite > <blair.bethwa...@gmail.com> wrote: >> Swami, >> >&g

Re: [ceph-users] un-even data filled on OSDs

2016-06-09 Thread Blair Bethwaite
> 173 (0.817063 >= 0.704429) [1.00 -> 0.95] > == > > is the above scripts says to reweight 43 -> 0.95? > > Thanks > Swami > > On Wed, Jun 8, 2016 at 10:34 AM, M Ranga Swami Reddy > <swamire...@gmail.com> wrote: >> Blair - Thanks f

Re: [ceph-users] un-even data filled on OSDs

2016-06-07 Thread Blair Bethwaite
or the script...Btw, is this script has option for dry run? > > Thanks > Swami > > On Wed, Jun 8, 2016 at 6:35 AM, Blair Bethwaite > <blair.bethwa...@gmail.com> wrote: >> Swami, >> >> Try >> https://github.com/cernceph/ceph-scripts/blob/master/tools/crush-rew

Re: [ceph-users] un-even data filled on OSDs

2016-06-07 Thread Blair Bethwaite
Swami, Try https://github.com/cernceph/ceph-scripts/blob/master/tools/crush-reweight-by-utilization.py, that'll work with Firefly and allow you to only tune down weight of a specific number of overfull OSDs. Cheers, On 7 June 2016 at 23:11, M Ranga Swami Reddy wrote: >

[ceph-users] dense storage nodes

2016-05-17 Thread Blair Bethwaite
Hi all, What are the densest node configs out there, and what are your experiences with them and tuning required to make them work? If we can gather enough info here then I'll volunteer to propose some upstream docs covering this. At Monash we currently have some 32-OSD nodes (running RHEL7),

[ceph-users] reweight-by-utilization warning

2016-05-15 Thread Blair Bethwaite
Hi all, IMHO reweight-by-utilization should come with some sort of warning, it just suddenly reweights everything - no dry run, no confirmation, apparently no option to see what it's going to do. It also doesn't appear to consider pools and hence crush rulesets, which I imagine could result in it

Re: [ceph-users] threading requirements for librbd

2016-03-10 Thread Blair Bethwaite
Thanks Jason, thanks Dan, On 9 March 2016 at 01:34, Jason Dillaman wrote: > Are you interesting in the max FD count or max thread count? You mention > both in your email. True, I did mix the two somewhat incorrectly - I was sort of guessing there'd be some number of

[ceph-users] threading requirements for librbd

2016-03-08 Thread Blair Bethwaite
Hi all, Not getting very far with this query internally (RH), so hoping someone familiar with the code can spare me the C++ pain... We've hit soft thread count ulimits a couple of times with different Ceph clusters. The clients (Qemu/KVM guests on both Ubuntu and RHEL hosts) have hit the limit

[ceph-users] Keystone PKIZ token support for RadosGW

2016-01-18 Thread Blair Bethwaite
Hi all, Does anyone know if RGW supports Keystone's PKIZ tokens, or better yet know a list of the supported token types? Cheers, -- Cheers, ~Blairo ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] Kernel RBD hang on OSD Failure

2015-12-07 Thread Blair Bethwaite
Hi Matt, (CC'ing in ceph-users too - similar reports there: http://www.spinics.net/lists/ceph-users/msg23037.html) We've seen something similar for KVM [lib]RBD clients acting as NFS gateways within our OpenStack cloud, the NFS services were locking up and causing client timeouts whenever we

[ceph-users] rgw pool config with spinning cache tier

2015-07-16 Thread Blair Bethwaite
Hi all, We're setting up radosgw against a large 8/3 EC pool (across 24 nodes) with a modest 4 node cache tier in front (those 4 nodes each have 20x 10k SAS drives and 4x Intel DC S3700 journals). With the cache tiering we're not sure what the best setup is for all the various peripheral rgw

Re: [ceph-users] He8 drives

2015-07-13 Thread Blair Bethwaite
On 13 July 2015 at 21:36, Emmanuel Florac eflo...@intellique.com wrote: I've benchmarked it and found it has about exactly the same performance profile as the He6. Compared to the Seagate 6TB it draws much less power (almost half), and that's the main selling point IMO, with durability.

[ceph-users] He8 drives

2015-07-07 Thread Blair Bethwaite
Hi folks, Does anyone have any experience with the newish HGST He8 8TB Helium filled HDDs? Storagereview looked at them here: http://www.storagereview.com/hgst_ultrastar_helium_he8_8tb_enterprise_hard_drive_review. I'm torn as to the lower read performance shown there than e.g. the He6 or Seagate

Re: [ceph-users] He8 drives

2015-07-07 Thread Blair Bethwaite
:01, Christian Balzer ch...@gol.com wrote: On Wed, 8 Jul 2015 10:28:17 +1000 Blair Bethwaite wrote: Hi folks, Does anyone have any experience with the newish HGST He8 8TB Helium filled HDDs? Storagereview looked at them here: http://www.storagereview.com

[ceph-users] radosgw sync agent against aws s3

2015-06-07 Thread Blair Bethwaite
Has anyone had any luck using the radosgw-sync-agent to push or pull to/from real S3? -- Cheers, ~Blairo ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] HDFS on Ceph (RBD)

2015-05-22 Thread Blair Bethwaite
out CephFS too. Hadoop is a predictable workload that we haven't seen break at all in several years and the bindings handle data locality and such properly. :) -Greg On Thu, May 21, 2015 at 11:24 PM, Wang, Warren warren_w...@cable.comcast.com javascript:; wrote: On 5/21/15, 5:04 AM, Blair

Re: [ceph-users] HDFS on Ceph (RBD)

2015-05-21 Thread Blair Bethwaite
Hi Warren, On 20 May 2015 at 23:23, Wang, Warren warren_w...@cable.comcast.com wrote: We¹ve contemplated doing something like that, but we also realized that it would result in manual work in Ceph everytime we lose a drive or server, and a pretty bad experience for the customer when we have

[ceph-users] HDFS on Ceph (RBD)

2015-05-20 Thread Blair Bethwaite
Hi Warren, Following our brief chat after the Ceph Ops session at the Vancouver summit today, I added a few more notes to the etherpad (https://etherpad.openstack.org/p/YVR-ops-ceph). I wonder whether you'd considered setting up crush layouts so you can have multiple cinder AZs or volume-types

[ceph-users] Tiering to object storage

2015-04-20 Thread Blair Bethwaite
Hi all, I understand the present pool tiering infrastructure is intended to work for 2 layers? We're presently considering backup strategies for large pools and wondered how much of a stretch it would be to have a base tier sitting in e.g. an S3 store... I imagine a pg in the base+1 tier mapping

Re: [ceph-users] pool distribution quality report script

2015-03-04 Thread Blair Bethwaite
Hi Mark, Cool, that looks handy. Though it'd be even better if it could go a step further and recommend re-weighting values to balance things out (or increased PG counts where needed). Cheers, On 5 March 2015 at 15:11, Mark Nelson mnel...@redhat.com wrote: Hi All, Recently some folks showed

[ceph-users] multiple CephFS filesystems on the same pools

2015-02-27 Thread Blair Bethwaite
Sorry if this is actually documented somewhere, but is it possible to create and use multiple filesystems on the data data and metadata pools? I'm guessing yes, but requires multiple MDSs? -- Cheers, ~Blairo ___ ceph-users mailing list

Re: [ceph-users] who is using radosgw with civetweb?

2015-02-25 Thread Blair Bethwaite
It'd be nice to see a standard/recommended LB and HA approach for RGW with supporting documentation too. On 26 February 2015 at 06:31, Sage Weil sw...@redhat.com wrote: Hey, We are considering switching to civetweb (the embedded/standalone rgw web server) as the primary supported RGW frontend

Re: [ceph-users] wider rados namespace support?

2015-02-12 Thread Blair Bethwaite
(managed by the cinder rbd driver), that way leaking secrets from cinder gives less exposure - but I guess that would be a bit of a change from the current namespace functionality. On 13 February 2015 at 05:57, Josh Durgin josh.dur...@inktank.com wrote: On 02/10/2015 07:54 PM, Blair Bethwaite wrote

Re: [ceph-users] wider rados namespace support?

2015-02-11 Thread Blair Bethwaite
On 11 February 2015 at 20:43, John Spray john.sp...@redhat.com wrote: Namespaces in CephFS would become useful in conjunction with limiting client authorization by sub-mount -- that way subdirectories could be assigned a layout with a particular namespace, and a client could be limited to that

[ceph-users] wider rados namespace support?

2015-02-10 Thread Blair Bethwaite
Just came across this in the docs: Currently (i.e., firefly), namespaces are only useful for applications written on top of librados. Ceph clients such as block device, object storage and file system do not currently support this feature. Then found:

Re: [ceph-users] 2015 Ceph Day Planning

2014-12-03 Thread Blair Bethwaite
Patrick, This sounds really great. Let me see if we're able to rustle up a venue in Melbourne, possibly Monash Uni Clayton (but I'll look for something closer to the city first). Cheers, On 4 December 2014 at 07:55, Patrick McGarry pmcga...@redhat.com wrote: Hey cephers, I know a number of

[ceph-users] OpenStack Kilo summit followup - Build a High-Performance and High-Durability Block Storage Service Based on Ceph

2014-11-05 Thread Blair Bethwaite
Hi Haomai, Thanks for your presentation this afternoon! Would be great if you could please share your slides and perhaps go into some more detail about your modelling of copysets in crush. -- Cheers, ~Blairo ___ ceph-users mailing list

Re: [ceph-users] cephfs survey results

2014-11-03 Thread Blair Bethwaite
On 4 November 2014 01:50, Sage Weil s...@newdream.net wrote: In the Ceph session at the OpenStack summit someone asked what the CephFS survey results looked like. Thanks Sage, that was me! Here's the link: https://www.surveymonkey.com/results/SM-L5JV7WXL/ In short, people want

[ceph-users] [radosgw] object copy implementation

2014-10-16 Thread Blair Bethwaite
Hi there rgw folks, Just wondering if the server-side copy operation ties up the radosgw host to actually proxy the data or if the copy is handled transparently by rados and the backend OSDs? -- Cheers, ~Blairo ___ ceph-users mailing list

Re: [ceph-users] Micro Ceph summit during the OpenStack summit

2014-10-12 Thread Blair Bethwaite
Hi Loic, I'll be there and interested to chat with other Cephers. But your pad isn't returning any page data... Cheers, On 11 October 2014 08:48, Loic Dachary l...@dachary.org wrote: Hi Ceph, TL;DR: please register at http://pad.ceph.com/p/kilo if you're attending the OpenStack summit

Re: [ceph-users] CephFS roadmap (was Re: NAS on RBD)

2014-09-10 Thread Blair Bethwaite
On 11 September 2014 08:47, John Spray john.sp...@redhat.com wrote: I do think this is something we could think about building a tool for: lots of people will have comparatively tiny quantities of metadata so full dumps would be a nice thing to have in our back pockets. Reminds me of the way

  1   2   >