[ceph-users] pre-split causing slow requests when rebuild osd ?

2018-11-26 Thread hnuzhoulin2
Hi,guysI have a 42 nodes cluster,and I create the pool using expected_num_objects to pre-split filestore dirs.today I rebuild a osd because a disk error,it cause much slow request,filestore logs like below2018-11-26 16:49:41.003336 7f2dad075700 10

Re: [ceph-users] Poor ceph cluster performance

2018-11-26 Thread Stefan Kooman
Quoting Cody (codeology@gmail.com): > The Ceph OSD part of the cluster uses 3 identical servers with the > following specifications: > > CPU: 2 x E5-2603 @1.8GHz > RAM: 16GB > Network: 1G port shared for Ceph public and cluster traffics This will hamper throughput a lot. > Journaling

[ceph-users] Move Instance between Different Ceph and Openstack Installation

2018-11-26 Thread Danni Setiawan
Hi all, I need to move instance from one Openstack with Ceph to another different Openstack with Ceph installation. The instance use volume to boot with another volume attach for data. The instance has 200GB volume boot and attach with 1TB volume for data. From what I know, I need to

Re: [ceph-users] Journal drive recommendation

2018-11-26 Thread Amit Ghadge
On Tue, 27 Nov 2018, 10:55 Martin Verges, wrote: > Hello, > > what type of SSD data drives do you plan to use? > We have plan to using external ssd data drive. > In general, I would not recommend to use external journal on ssd OSDs, but > it is possible to squeeze out a bit more performance

Re: [ceph-users] Journal drive recommendation

2018-11-26 Thread Martin Verges
Hello, what type of SSD data drives do you plan to use? In general, I would not recommend to use external journal on ssd OSDs, but it is possible to squeeze out a bit more performance depending on your data disks. -- Martin Verges Managing director Mobile: +49 174 9335695 E-Mail:

[ceph-users] Journal drive recommendation

2018-11-26 Thread Amit Ghadge
Hi all, We have planning to use SSD data drive, so for journal drive, is there any recommendation to use same drive or separate drive? Thanks, Amit ___ ceph-users mailing list ceph-users@lists.ceph.com

[ceph-users] Poor ceph cluster performance

2018-11-26 Thread Cody
Hello, I have a Ceph cluster deployed together with OpenStack using TripleO. While the Ceph cluster shows a healthy status, its performance is painfully slow. After eliminating a possibility of network issues, I have zeroed in on the Ceph cluster itself, but have no experience in further

Re: [ceph-users] Bug: Deleting images ending with whitespace in name via dashboard

2018-11-26 Thread Lenz Grimmer
Hi Alexander, On 11/13/18 12:37 PM, Kasper, Alexander wrote: > As i am not sure howto correctly use tracker.ceph.com, i´ll post my > report here: > > Using the dashboard to delete a rbd image via gui throws an error when > the image name ends with an whitespace (user input error leaded to this

Re: [ceph-users] Should ceph build against libcurl4 for Ubuntu 18.04 and later?

2018-11-26 Thread Ken Dreyer
On Thu, Nov 22, 2018 at 11:47 AM Matthew Vernon wrote: > > On 22/11/2018 13:40, Paul Emmerich wrote: > > We've encountered the same problem on Debian Buster > > It looks to me like this could be fixed simply by building the Bionic > packages in a Bionic chroot (ditto Buster); maybe that could be

Re: [ceph-users] What could cause mon_osd_full_ratio to be exceeded?

2018-11-26 Thread Gregory Farnum
On Mon, Nov 26, 2018 at 10:28 AM Vladimir Brik wrote: > > Hello > > I am doing some Ceph testing on a near-full cluster, and I noticed that, > after I brought down a node, some OSDs' utilization reached > osd_failsafe_full_ratio (97%). Why didn't it stop at mon_osd_full_ratio > (90%) if

Re: [ceph-users] read performance, separate client CRUSH maps or limit osd read access from each client

2018-11-26 Thread Vlad Kopylov
I see. Thank you Greg. Ultimately leading to some kind of multi-primary OSD/MON setup, which will most likely add lookup overheads. Though might be a reasonable trade off for network distributed setups. Good feature for major version. With Glusterfs I solved it, funny as it sounds, by writing

Re: [ceph-users] What could cause mon_osd_full_ratio to be exceeded?

2018-11-26 Thread Vladimir Brik
> Why didn't it stop at mon_osd_full_ratio (90%) Should be 95% Vlad On 11/26/18 9:28 AM, Vladimir Brik wrote: Hello I am doing some Ceph testing on a near-full cluster, and I noticed that, after I brought down a node, some OSDs' utilization reached osd_failsafe_full_ratio (97%). Why

[ceph-users] What could cause mon_osd_full_ratio to be exceeded?

2018-11-26 Thread Vladimir Brik
Hello I am doing some Ceph testing on a near-full cluster, and I noticed that, after I brought down a node, some OSDs' utilization reached osd_failsafe_full_ratio (97%). Why didn't it stop at mon_osd_full_ratio (90%) if mon_osd_backfillfull_ratio is 90%? Thanks, Vlad

Re: [ceph-users] read performance, separate client CRUSH maps or limit osd read access from each client

2018-11-26 Thread Gregory Farnum
On Tue, Nov 20, 2018 at 9:50 PM Vlad Kopylov wrote: > I see the point, but not for the read case: > no overhead for just choosing or let Mount option choose read replica. > > This is simple feature that can be implemented, that will save many > people bandwidth in really distributed cases. >

Re: [ceph-users] will crush rule be used during object relocation in OSD failure ?

2018-11-26 Thread Gregory Farnum
On Fri, Nov 23, 2018 at 11:01 AM ST Wong (ITSC) wrote: > Hi all, > > > We've 8 osd hosts, 4 in room 1 and 4 in room2. > > A pool with size = 3 using following crush map is created, to cater for > room failure. > > > rule multiroom { > id 0 > type replicated > min_size 2 >

Re: [ceph-users] Monitor disks for SSD only cluster

2018-11-26 Thread Wido den Hollander
On 11/26/18 2:21 PM, Gregory Farnum wrote: > As the monitors limit their transaction rates, I would tend for the > higher-durability drives. I don't think any monitor throughput issues > have been reported on clusters with SSDs for storage. I can confirm that. Just make sure you have proper

Re: [ceph-users] Monitor disks for SSD only cluster

2018-11-26 Thread Gregory Farnum
As the monitors limit their transaction rates, I would tend for the higher-durability drives. I don't think any monitor throughput issues have been reported on clusters with SSDs for storage. -Greg On Mon, Nov 26, 2018 at 5:47 AM Valmar Kuristik wrote: > Hello, > > Can anyone say how important

Re: [ceph-users] No recovery when "norebalance" flag set

2018-11-26 Thread Gregory Farnum
On Sun, Nov 25, 2018 at 2:41 PM Stefan Kooman wrote: > Hi list, > > During cluster expansion (adding extra disks to existing hosts) some > OSDs failed (FAILED assert(0 == "unexpected error", _txc_add_transaction > error (39) Directory not empty not handled on operation 21 (op 1, > counting from

Re: [ceph-users] Degraded objects afte: ceph osd in $osd

2018-11-26 Thread Gregory Farnum
On Mon, Nov 26, 2018 at 3:30 AM Janne Johansson wrote: > Den sön 25 nov. 2018 kl 22:10 skrev Stefan Kooman : > > > > Hi List, > > > > Another interesting and unexpected thing we observed during cluster > > expansion is the following. After we added extra disks to the cluster, > > while

Re: [ceph-users] Degraded objects afte: ceph osd in $osd

2018-11-26 Thread Marco Gaiarin
Mandi! Janne Johansson In chel di` si favelave... > It is a slight mistake in reporting it in the same way as an error, even if > it looks to the > cluster just as if it was in error and needs fixing. I think i'm hit a similar situation, and also i'm feeling that something have to be 'fixed'.

Re: [ceph-users] No recovery when "norebalance" flag set

2018-11-26 Thread Stefan Kooman
Quoting Dan van der Ster (d...@vanderster.com): > Haven't seen that exact issue. > > One thing to note though is that if osd_max_backfills is set to 1, > then it can happen that PGs get into backfill state, taking that > single reservation on a given OSD, and therefore the recovery_wait PGs >

Re: [ceph-users] Disable intra-host replication?

2018-11-26 Thread Janne Johansson
Den mån 26 nov. 2018 kl 12:11 skrev Marco Gaiarin : > Mandi! Janne Johansson > In chel di` si favelave... > > > The default crush rules with replication=3 would only place PGs on > > separate hosts, > > so in that case it would go into degraded mode if a node goes away, > > and not place > >

Re: [ceph-users] No recovery when "norebalance" flag set

2018-11-26 Thread Dan van der Ster
Haven't seen that exact issue. One thing to note though is that if osd_max_backfills is set to 1, then it can happen that PGs get into backfill state, taking that single reservation on a given OSD, and therefore the recovery_wait PGs can't get a slot. I suppose that backfill prioritization is

Re: [ceph-users] Disable intra-host replication?

2018-11-26 Thread Marco Gaiarin
Mandi! Janne Johansson In chel di` si favelave... > The default crush rules with replication=3 would only place PGs on > separate hosts, > so in that case it would go into degraded mode if a node goes away, > and not place > replicas on different disks on the remaining hosts. 'hosts' mean

Re: [ceph-users] Low traffic Ceph cluster with consumer SSD.

2018-11-26 Thread Eneko Lacunza
Hi, El 25/11/18 a las 18:23, Виталий Филиппов escribió: Ok... That's better than previous thread with file download where the topic starter suffered from normal only-metadata-journaled fs... Thanks for the link, it would be interesting to repeat similar tests. Although I suspect it shouldn't

[ceph-users] Monitor disks for SSD only cluster

2018-11-26 Thread Valmar Kuristik
Hello, Can anyone say how important is to have fast storage on monitors for a all ssd deployment? We are planning on throwing SSDs into the monitors as well, but are at a loss about if to go for more durability or speed. Higher durability drives tend to be a lot slower for the 240GB size we'd

Re: [ceph-users] Sizing for bluestore db and wal

2018-11-26 Thread Janne Johansson
Den mån 26 nov. 2018 kl 10:10 skrev Felix Stolte : > > Hi folks, > > i upgraded our ceph cluster from jewel to luminous and want to migrate > from filestore to bluestore. Currently we use one SSD as journal for > thre 8TB Sata Drives with a journal partition size of 40GB. If my > understanding of

[ceph-users] Sizing for bluestore db and wal

2018-11-26 Thread Felix Stolte
Hi folks, i upgraded our ceph cluster from jewel to luminous and want to migrate from filestore to bluestore. Currently we use one SSD as journal for thre 8TB Sata Drives with a journal partition size of 40GB. If my understanding of the bluestore documentation is correct, i can use a wal

Re: [ceph-users] Degraded objects afte: ceph osd in $osd

2018-11-26 Thread Janne Johansson
Den mån 26 nov. 2018 kl 09:39 skrev Stefan Kooman : > > It is a slight mistake in reporting it in the same way as an error, > > even if it looks to the > > cluster just as if it was in error and needs fixing. This gives the > > new ceph admins a > > sense of urgency or danger whereas it should be

Re: [ceph-users] Degraded objects afte: ceph osd in $osd

2018-11-26 Thread Stefan Kooman
Quoting Janne Johansson (icepic...@gmail.com): > Yes, when you add a drive (or 10), some PGs decide they should have one or > more > replicas on the new drives, a new empty PG is created there, and > _then_ that replica > will make that PG get into the "degraded" mode, meaning if it had 3 > fine

Re: [ceph-users] Degraded objects afte: ceph osd in $osd

2018-11-26 Thread Janne Johansson
Den sön 25 nov. 2018 kl 22:10 skrev Stefan Kooman : > > Hi List, > > Another interesting and unexpected thing we observed during cluster > expansion is the following. After we added extra disks to the cluster, > while "norebalance" flag was set, we put the new OSDs "IN". As soon as > we did that