Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-21 Thread Stefan Priebe - Profihost AG
ed log level and dumps > them on crash with negative index starting at - up to -1 as a prefix. > > -1> 2020-01-16 01:10:13.404090 7f3350a14700 -1 rocksdb: > > > It would be great If you share several log snippets for different > crashes containing these last 100

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-19 Thread Stefan Priebe - Profihost AG
= 0x7f8001cc45c881217262'd_data.4303206b8b4567.9632!='0xfffe'o' Value size = 510) on the right size i always see 0xfffe on all failed OSDs. greets, Stefan Am 19.01.20 um 14:07 schrieb Stefan Priebe - Profihost AG: > Yes, except that this happ

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-19 Thread Stefan Priebe - Profihost AG
times this might > provide some hints. > > > Thanks, > > Igor > > >> On 1/17/2020 2:30 PM, Stefan Priebe - Profihost AG wrote: >> HI Igor, >> >>> Am 17.01.20 um 12:10 schrieb Igor Fedotov: >>> hmmm.. >>> >>> Just in ca

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-17 Thread Stefan Priebe - Profihost AG
start collecting failure-related information > (including but not limited to failure logs, perf counter dumps, system > resource reports etc) for future analysis. > > > > On 1/16/2020 11:58 PM, Stefan Priebe - Profihost AG wrote: >> Hi Igor, >> >> answers inlin

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-16 Thread Stefan Priebe - Profihost AG
; > Igor > > > > On 1/16/2020 10:04 PM, Stefan Priebe - Profihost AG wrote: >> Hi Igor, >> >> ouch sorry. Here we go: >> >>     -1> 2020-01-16 01:10:13.404090 7f3350a14700 -1 rocksdb: >> submit_transaction error: Corruption: block checksu

Re: [ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-16 Thread Stefan Priebe - Profihost AG
... > > > Thanks, > > Igor > > On 1/16/2020 11:56 AM, Stefan Priebe - Profihost AG wrote: >> Hello, >> >> does anybody know a fix for this ASSERT / crash? >> >> 2020-01-16 02:02:31.316394 7f8c3f5ab700 -1 >> /build/ceph/src/os/bluestore/BlueStore

[ceph-users] Luminous Bluestore OSDs crashing with ASSERT

2020-01-16 Thread Stefan Priebe - Profihost AG
Hello, does anybody know a fix for this ASSERT / crash? 2020-01-16 02:02:31.316394 7f8c3f5ab700 -1 /build/ceph/src/os/bluestore/BlueStore.cc: In function 'void BlueStore::_kv_sync_thread()' thread 7f8c3f5ab700 time 2020-01-16 02:02:31.304993 /build/ceph/src/os/bluestore/BlueStore.cc: 8808:

Re: [ceph-users] block db sizing and calculation

2020-01-14 Thread Stefan Priebe - Profihost AG
Hello, does anybody have real live experience with externel block db? Greets, Stefan Am 13.01.20 um 08:09 schrieb Stefan Priebe - Profihost AG: > Hello, > > i'm plannung to split the block db to a seperate flash device which i > also would like to use as an OSD for erasure cod

Re: [ceph-users] low io with enterprise SSDs ceph luminous - can we expect more? [klartext]

2020-01-13 Thread Stefan Priebe - Profihost AG
ch have a capacitor. Samsung does not. The problem is that Ceph sends a lot of flush commands which slows down drives without capacitor. You can make linux to ignore those userspace requests with the following command: echo "temporary write through" > /sys/block/sdX/device/scsi_disk/*/

[ceph-users] block db sizing and calculation

2020-01-12 Thread Stefan Priebe - Profihost AG
Hello, i'm plannung to split the block db to a seperate flash device which i also would like to use as an OSD for erasure coding metadata for rbd devices. If i want to use 14x 14TB HDDs per Node https://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#sizing recommends a

[ceph-users] Hardware selection for ceph backup on ceph

2020-01-10 Thread Stefan Priebe - Profihost AG
Hi, we‘re currently in the process of building a new ceph cluster to backup rbd images from multiple ceph clusters. We would like to start with just a single ceph cluster to backup which is about 50tb. Compression ratio of the data is around 30% while using zlib. We need to scale the backup

Re: [ceph-users] Looking for experience

2020-01-10 Thread Stefan Priebe - Profihost AG
ter to sync snapshots once a day. Important is pricing, performance todo this task and expansion. We would like to start with something around just 50tb storage. Greets, Stefan > > >> Stefan Priebe - Profihost AG < s.pri...@profihost.ag> hat am 9. Januar 2020 >> um 22:5

Re: [ceph-users] Looking for experience

2020-01-09 Thread Stefan Priebe - Profihost AG
or not? Since we started using ceph we're mostly subscribed to SSDs - so no knowlege about HDD in place. Greets, Stefan Am 09.01.20 um 16:49 schrieb Stefan Priebe - Profihost AG: > >> Am 09.01.2020 um 16:10 schrieb Wido den Hollander : >> >>  >> >>> On 1/9/20 2

Re: [ceph-users] Looking for experience

2020-01-09 Thread Stefan Priebe - Profihost AG
> Am 09.01.2020 um 16:10 schrieb Wido den Hollander : > >  > >> On 1/9/20 2:27 PM, Stefan Priebe - Profihost AG wrote: >> Hi Wido, >>> Am 09.01.20 um 14:18 schrieb Wido den Hollander: >>> >>> >>> On 1/9/20 2:07 PM, Daniel Aberger -

Re: [ceph-users] Looking for experience

2020-01-09 Thread Stefan Priebe - Profihost AG
Hi Wido, Am 09.01.20 um 14:18 schrieb Wido den Hollander: > > > On 1/9/20 2:07 PM, Daniel Aberger - Profihost AG wrote: >> >> Am 09.01.20 um 13:39 schrieb Janne Johansson: >>> >>> I'm currently trying to workout a concept for a ceph cluster which can >>> be used as a target for backups

Re: [ceph-users] ceph version 14.2.3-OSD fails

2019-10-11 Thread Stefan Priebe - Profihost AG
> Am 11.10.2019 um 14:07 schrieb Igor Fedotov : > >  > Hi! > > originally your issue looked like the ones from > https://tracker.ceph.com/issues/42223 > > And it looks like lack of some key information for FreeListManager in RocksDB. > > Once you have it present we can check the content of

Re: [ceph-users] Bluestore OSDs keep crashing in BlueStore.cc: 8808: FAILED assert(r == 0)

2019-09-16 Thread Stefan Priebe - Profihost AG
l unpatched and 12.2.13 might the first containing this fix - but no idea of an ETA. Greets, Stefan > Thanks, > > Igor > > On 9/12/2019 8:20 PM, Stefan Priebe - Profihost AG wrote: >> Hello Igor, >> >> i can now confirm that this is indeed a kernel bug. The issue does n

Re: [ceph-users] Bluestore OSDs keep crashing in BlueStore.cc: 8808: FAILED assert(r == 0)

2019-09-12 Thread Stefan Priebe - Profihost AG
wapping: > https://tracker.ceph.com/issues/22464 > > IMO memory usage worth checking as well... > > > Igor > > > On 8/27/2019 4:52 PM, Stefan Priebe - Profihost AG wrote: >> see inline >> >> Am 27.08.19 um 15:43 schrieb Igor Fedotov: >>&g

Re: [ceph-users] Bluestore OSDs keep crashing in BlueStore.cc: 8808: FAILED assert(r == 0)

2019-08-27 Thread Stefan Priebe - Profihost AG
l invalid > data reads under high memory pressure/swapping: > https://tracker.ceph.com/issues/22464 We have a current 4.19.X kernel and no memory limit. Mem avail is pretty constant at 32GB. Greets, Stefan > > IMO memory usage worth checking as well... > > > Igor > >

Re: [ceph-users] Bluestore OSDs keep crashing in BlueStore.cc: 8808: FAILED assert(r == 0)

2019-08-27 Thread Stefan Priebe - Profihost AG
see inline Am 27.08.19 um 15:43 schrieb Igor Fedotov: > see inline > > On 8/27/2019 4:41 PM, Stefan Priebe - Profihost AG wrote: >> Hi Igor, >> >> Am 27.08.19 um 14:11 schrieb Igor Fedotov: >>> Hi Stefan, >>> >>> this looks like a dupli

Re: [ceph-users] Bluestore OSDs keep crashing in BlueStore.cc: 8808: FAILED assert(r == 0)

2019-08-27 Thread Stefan Priebe - Profihost AG
you run fsck for any of broken OSDs? Any reports? Yes but no reports. > Any other errors/crashes in logs before these sort of issues happens? No > Just in case - what allocator are you using? tcmalloc Greets, Stefan > > Thanks, > > Igor > > > > On 8/27/2019 1:0

[ceph-users] Bluestore OSDs keep crashing in BlueStore.cc: 8808: FAILED assert(r == 0)

2019-08-27 Thread Stefan Priebe - Profihost AG
Hello, since some month all our bluestore OSDs keep crashing from time to time. Currently about 5 OSDs per day. All of them show the following trace: Trace: 2019-07-24 08:36:48.995397 7fb19a711700 -1 rocksdb: submit_transaction error: Corruption: block checksum mismatch code = 2 Rocksdb

[ceph-users] Bluestore OSDs keep crashing in BlueStore.cc: 8808: FAILED assert(r == 0)

2019-08-27 Thread Stefan Priebe - Profihost AG
Hello, since some month all our bluestore OSDs keep crashing from time to time. Currently about 5 OSDs per day. All of them show the following trace: Trace: 2019-07-24 08:36:48.995397 7fb19a711700 -1 rocksdb: submit_transaction error: Corruption: block checksum mismatch code = 2 Rocksdb

Re: [ceph-users] optimize bluestore for random write i/o

2019-03-06 Thread Stefan Priebe - Profihost AG
Am 06.03.19 um 14:08 schrieb Mark Nelson: > > On 3/6/19 5:12 AM, Stefan Priebe - Profihost AG wrote: >> Hi Mark, >> Am 05.03.19 um 23:12 schrieb Mark Nelson: >>> Hi Stefan, >>> >>> >>> Could you try running your random write workload a

Re: [ceph-users] optimize bluestore for random write i/o

2019-03-06 Thread Stefan Priebe - Profihost AG
_cond_wait@@GLIBC_2.3.2 () from target:/lib/x86_64-linux-gnu/libpthread.so.0 . Thread 1 "ceph-osd" received signal SIGINT, Interrupt. 0x7f917b6a615f in pthread_cond_wait@@GLIBC_2.3.2 () from target:/lib/x86_64-linux-gnu/libpthread.so.0 . Thread 1 "ceph-osd" received signal SIG

Re: [ceph-users] optimize bluestore for random write i/o

2019-03-05 Thread Stefan Priebe - Profihost AG
Am 05.03.19 um 10:05 schrieb Paul Emmerich: > This workload is probably bottlenecked by rocksdb (since the small > writes are buffered there), so that's probably what needs tuning here. while reading:

[ceph-users] optimize bluestore for random write i/o

2019-03-05 Thread Stefan Priebe - Profihost AG
Hello list, while the performance of sequential writes 4k on bluestore is very high and even higher than filestore i was wondering what i can do to optimize random pattern as well. While using: fio --rw=write --iodepth=32 --ioengine=libaio --bs=4k --numjobs=4 --filename=/tmp/test --size=10G

Re: [ceph-users] ceph osd commit latency increase over time, until restart

2019-01-30 Thread Stefan Priebe - Profihost AG
op_w_process_latency.sum"), > 1s)/non_negative_derivative(first("op_w_process_latency.avgcount"),1s) FROM > "ceph" WHERE "host" =~ /^([[host]])$/ AND collection='osd' AND "id" =~ > /^([[osd]])$/ AND $timeFilter GROUP BY time($interval), "host

Re: [ceph-users] ceph osd commit latency increase over time, until restart

2019-01-29 Thread Stefan Priebe - Profihost AG
Hi, Am 30.01.19 um 08:33 schrieb Alexandre DERUMIER: > Hi, > > here some new results, > different osd/ different cluster > > before osd restart latency was between 2-5ms > after osd restart is around 1-1.5ms > > http://odisoweb1.odiso.net/cephperf2/bad.txt (2-5ms) >

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-17 Thread Stefan Priebe - Profihost AG
ng against and using jemalloc. What happens in this case? Also i saw now - that 12.2.10 uses 1GB mem max while 12.2.8 uses 6-7GB Mem (with bluestore_cache_size = 1073741824). Greets, Stefan Am 17.01.19 um 22:59 schrieb Stefan Priebe - Profihost AG: > Hello Mark, > > for whatever reaso

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-17 Thread Stefan Priebe - Profihost AG
) That's currently all i know. Thanks a lot! Greets, Stefan Am 16.01.19 um 20:56 schrieb Stefan Priebe - Profihost AG: > i reverted the whole cluster back to 12.2.8 - recovery speed also > dropped from 300-400MB/s to 20MB/s on 12.2.10. So something is really > broken. > > Greets

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-16 Thread Stefan Priebe - Profihost AG
i reverted the whole cluster back to 12.2.8 - recovery speed also dropped from 300-400MB/s to 20MB/s on 12.2.10. So something is really broken. Greets, Stefan Am 16.01.19 um 16:00 schrieb Stefan Priebe - Profihost AG: > This is not the case with 12.2.8 - it happens with 12.2.9 as well. Af

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-16 Thread Stefan Priebe - Profihost AG
15:22 schrieb Stefan Priebe - Profihost AG: > Hello, > > while digging into this further i saw that it takes ages until all pgs > are active. After starting the OSD 3% of all pgs are inactive and it > takes minutes after they're active. > > The log of the OSD is full of: >

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-16 Thread Stefan Priebe - Profihost AG
1318356'61576253 active+rec overing+degraded m=183 snaptrimq=[ec1a0~1,ec808~1] mbc={255={(2+0)=183,(3+0)=3}}] _update_calc_stats ml 183 upset size 3 up 2 Greets, Stefan Am 16.01.19 um 09:12 schrieb Stefan Priebe - Profihost AG: > Hi, > > no ok it was not. Bug still present. It was only workin

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-16 Thread Stefan Priebe - Profihost AG
Hi, no ok it was not. Bug still present. It was only working because the osdmap was so far away that it has started backfill instead of recovery. So it happens only in the recovery case. Greets, Stefan Am 15.01.19 um 16:02 schrieb Stefan Priebe - Profihost AG: > > Am 15.01.19 um 12:45 s

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-15 Thread Stefan Priebe - Profihost AG
Hello list, > > i also tested current upstream/luminous branch and it happens as well. A > clean install works fine. It only happens on upgraded bluestore osds. > > Greets, > Stefan > > Am 14.01.19 um 20:35 schrieb Stefan Priebe - Profihost AG: >> while trying to

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-15 Thread Stefan Priebe - Profihost AG
Hello list, i also tested current upstream/luminous branch and it happens as well. A clean install works fine. It only happens on upgraded bluestore osds. Greets, Stefan Am 14.01.19 um 20:35 schrieb Stefan Priebe - Profihost AG: > while trying to upgrade a cluster from 12.2.8 to 12.2.10

Re: [ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-14 Thread Stefan Priebe - Profihost AG
Hi Paul, Am 14.01.19 um 21:39 schrieb Paul Emmerich: > What's the output of "ceph daemon osd. status" on one of the OSDs > while it's starting? { "cluster_fsid": "b338193d-39e0-40e9-baba-4965ef3868a3", "osd_fsid": "d95d0e3b-7441-4ab0-869c-fe0551d3bd52", "whoami": 2, "state":

[ceph-users] slow requests and high i/o / read rate on bluestore osds after upgrade 12.2.8 -> 12.2.10

2019-01-14 Thread Stefan Priebe - Profihost AG
Hi, while trying to upgrade a cluster from 12.2.8 to 12.2.10 i'm experience issues with bluestore osds - so i canceled the upgrade and all bluestore osds are stopped now. After starting a bluestore osd i'm seeing a lot of slow requests caused by very high read rates. Device: rrqm/s

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-19 Thread Stefan Priebe - Profihost AG
72,00 > 1512,57    14,88   22,05   24,57    6,97   1,48 100,00 > > so it reads with 500MB/s which completely saturates the osd. And it does > for > 10 minutes. > > Greets, > Stefan > > Am 13.10.2018 um 21:29 schrieb Stefan Priebe - Profihost AG:

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-13 Thread Stefan Priebe - Profihost AG
tat shows: sdi 77,00 0,00 580,00 97,00 511032,00 972,00 1512,5714,88 22,05 24,576,97 1,48 100,00 so it reads with 500MB/s which completely saturates the osd. And it does for > 10 minutes. Greets, Stefan Am 13.10.2018 um 21:29 schrieb Stefan Priebe - Profih

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-13 Thread Stefan Priebe - Profihost AG
tat shows: sdi 77,00 0,00 580,00 97,00 511032,00 972,00 1512,5714,88 22,05 24,576,97 1,48 100,00 so it reads with 500MB/s which completely saturates the osd. And it does for > 10 minutes. Greets, Stefan Am 13.10.2018 um 21:29 schrieb Stefan Priebe - Profih

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-13 Thread Stefan Priebe - Profihost AG
ods.19 is a bluestore osd on a healthy 2TB SSD. Log of osd.19 is here: https://pastebin.com/raw/6DWwhS0A Am 13.10.2018 um 21:20 schrieb Stefan Priebe - Profihost AG: > Hi David, > > i think this should be the problem - form a new log from today: > > 2018-10-13 20:57:20.36

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-13 Thread Stefan Priebe - Profihost AG
Hi David, i think this should be the problem - form a new log from today: 2018-10-13 20:57:20.367326 mon.a [WRN] Health check update: 4 osds down (OSD_DOWN) ... 2018-10-13 20:57:41.268674 mon.a [WRN] Health check update: Reduced data availability: 3 pgs peering (PG_AVAILABILITY) ... 2018-10-13

[ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-12 Thread Stefan Priebe - Profihost AG
Hi David, Am 12.10.2018 um 15:59 schrieb David Turner: > The PGs per OSD does not change unless the OSDs are marked out.  You > have noout set, so that doesn't change at all during this test.  All of > your PGs peered quickly at the beginning and then were active+undersized > the rest of the

Re: [ceph-users] Ceph and NVMe

2018-09-09 Thread Stefan Priebe - Profihost AG
his question, is anyone using Intel Optane DC P4800X on DELL > R630 ...or any other server ? > Any gotchas / feedback/ knowledge sharing will be greatly appreciated > > Steven > >> On Thu, 6 Sep 2018 at 14:59, Stefan Priebe - Profihost AG >> wrote: >> Hello

[ceph-users] Ceph and NVMe

2018-09-06 Thread Stefan Priebe - Profihost AG
Hello list, has anybody tested current NVMe performance with luminous and bluestore? Is this something which makes sense or just a waste of money? Greets, Stefan ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] ceph balancer: further optimizations?

2018-08-22 Thread Stefan Priebe - Profihost AG
Am 21.08.2018 um 17:28 schrieb Gregory Farnum: > You should be able to create issues now; we had a misconfiguration in > the tracker following the recent spam attack. > -Greg > > On Tue, Aug 21, 2018 at 3:07 AM, Stefan Priebe - Profihost AG > wrote: >> >> Am 21.0

Re: [ceph-users] ceph balancer: further optimizations?

2018-08-21 Thread Stefan Priebe - Profihost AG
Am 21.08.2018 um 12:03 schrieb Stefan Priebe - Profihost AG: > > Am 21.08.2018 um 11:56 schrieb Dan van der Ster: >> On Tue, Aug 21, 2018 at 11:54 AM Stefan Priebe - Profihost AG >> wrote: >>> >>> Am 21.08.2018 um 11:47 schrieb Dan van der Ster: >>

Re: [ceph-users] ceph balancer: further optimizations?

2018-08-21 Thread Stefan Priebe - Profihost AG
Am 21.08.2018 um 11:56 schrieb Dan van der Ster: > On Tue, Aug 21, 2018 at 11:54 AM Stefan Priebe - Profihost AG > wrote: >> >> Am 21.08.2018 um 11:47 schrieb Dan van der Ster: >>> On Mon, Aug 20, 2018 at 10:45 PM Stefan Priebe - Profihost AG >>> wrote: &

Re: [ceph-users] ceph balancer: further optimizations?

2018-08-21 Thread Stefan Priebe - Profihost AG
Am 21.08.2018 um 11:47 schrieb Dan van der Ster: > On Mon, Aug 20, 2018 at 10:45 PM Stefan Priebe - Profihost AG > wrote: >> >> >> Am 20.08.2018 um 22:38 schrieb Dan van der Ster: >>> On Mon, Aug 20, 2018 at 10:19 PM Stefan Priebe - Profihost AG >>> wr

Re: [ceph-users] ceph balancer: further optimizations?

2018-08-20 Thread Stefan Priebe - Profihost AG
Am 20.08.2018 um 22:38 schrieb Dan van der Ster: > On Mon, Aug 20, 2018 at 10:19 PM Stefan Priebe - Profihost AG > wrote: >> >> >> Am 20.08.2018 um 21:52 schrieb Sage Weil: >>> On Mon, 20 Aug 2018, Stefan Priebe - Profihost AG wrote: >>>> Hello,

Re: [ceph-users] ceph balancer: further optimizations?

2018-08-20 Thread Stefan Priebe - Profihost AG
o:s...@newdream.net>> wrote: > > On Mon, 20 Aug 2018, Stefan Priebe - Profihost AG wrote: > > Hello, > > > > since loic seems to have left ceph development and his wunderful crush > > optimization tool isn'T working anymore i'm trying to get a good >

Re: [ceph-users] ceph balancer: further optimizations?

2018-08-20 Thread Stefan Priebe - Profihost AG
Am 20.08.2018 um 21:52 schrieb Sage Weil: > On Mon, 20 Aug 2018, Stefan Priebe - Profihost AG wrote: >> Hello, >> >> since loic seems to have left ceph development and his wunderful crush >> optimization tool isn'T working anymore i'm trying to get a good >> di

[ceph-users] ceph balancer: further optimizations?

2018-08-20 Thread Stefan Priebe - Profihost AG
Hello, since loic seems to have left ceph development and his wunderful crush optimization tool isn'T working anymore i'm trying to get a good distribution with the ceph balancer. Sadly it does not work as good as i want. # ceph osd df | sort -k8 show 75 to 83% Usage which is 8% difference

Re: [ceph-users] ceph mgr balancer bad distribution

2018-03-03 Thread Stefan Priebe - Profihost AG
Hi, Am 02.03.2018 um 21:21 schrieb Stefan Priebe - Profihost AG: > Hi, > > Am 02.03.2018 um 14:29 schrieb Dan van der Ster: >> On Fri, Mar 2, 2018 at 10:12 AM, Stefan Priebe - Profihost AG >> <s.pri...@profihost.ag> wrote: >>> Thanks! Your patch works great

Re: [ceph-users] ceph mgr balancer bad distribution

2018-03-02 Thread Stefan Priebe - Profihost AG
Hi, Am 02.03.2018 um 14:29 schrieb Dan van der Ster: > On Fri, Mar 2, 2018 at 10:12 AM, Stefan Priebe - Profihost AG > <s.pri...@profihost.ag> wrote: >> Thanks! Your patch works great! > > Cool! I plan to add one more feature to allow operators to switch off > compo

Re: [ceph-users] ceph mgr balancer bad distribution

2018-03-02 Thread Stefan Priebe - Profihost AG
should evaluate the ceph health status as well. Stefan Excuse my typo sent from my mobile phone. Am 01.03.2018 um 13:12 schrieb Dan van der Ster <d...@vanderster.com <mailto:d...@vanderster.com>>: > On Thu, Mar 1, 2018 at 1:08 PM, Stefan Priebe - Profihost AG > <s.pri...@profi

Re: [ceph-users] ceph mgr balancer bad distribution

2018-03-02 Thread Stefan Priebe - Profihost AG
should evaluate the ceph health status as well. Stefan Excuse my typo sent from my mobile phone. > Am 01.03.2018 um 13:12 schrieb Dan van der Ster <d...@vanderster.com>: > > On Thu, Mar 1, 2018 at 1:08 PM, Stefan Priebe - Profihost AG > <s.pri...@profihost.ag> wrote:

Re: [ceph-users] ceph mgr balancer bad distribution

2018-03-01 Thread Stefan Priebe - Profihost AG
van der Ster <d...@vanderster.com> >> wrote: >>> On Thu, Mar 1, 2018 at 10:24 AM, Stefan Priebe - Profihost AG >>> <s.pri...@profihost.ag> wrote: >>>> >>>> Am 01.03.2018 um 09:58 schrieb Dan van der Ster: >>>>> On Thu

Re: [ceph-users] ceph mgr balancer bad distribution

2018-03-01 Thread Stefan Priebe - Profihost AG
Am 01.03.2018 um 09:58 schrieb Dan van der Ster: > On Thu, Mar 1, 2018 at 9:52 AM, Stefan Priebe - Profihost AG > <s.pri...@profihost.ag> wrote: >> Hi, >> >> Am 01.03.2018 um 09:42 schrieb Dan van der Ster: >>> On Thu, Mar 1, 2018 at 9:31 AM,

Re: [ceph-users] ceph mgr balancer bad distribution

2018-03-01 Thread Stefan Priebe - Profihost AG
Hi, Am 01.03.2018 um 09:42 schrieb Dan van der Ster: > On Thu, Mar 1, 2018 at 9:31 AM, Stefan Priebe - Profihost AG > <s.pri...@profihost.ag> wrote: >> Hi, >> Am 01.03.2018 um 09:03 schrieb Dan van der Ster: >>> Is the score improving? >>> &g

Re: [ceph-users] ceph mgr balancer bad distribution

2018-03-01 Thread Stefan Priebe - Profihost AG
l bad. >> >> It seems to balance from left to right and than back from right to left... >> >> Greets, >> Stefan >> >> Am 28.02.2018 um 13:47 schrieb Stefan Priebe - Profihost AG: >>> Hello, >>> >>> with jewel we always used the

Re: [ceph-users] ceph mgr balancer bad distribution

2018-02-28 Thread Stefan Priebe - Profihost AG
Does anybody have some more input? I keeped the balancer active for 24h now and it is rebalancing 1-3% every 30 minutes but the distribution is still bad. It seems to balance from left to right and than back from right to left... Greets, Stefan Am 28.02.2018 um 13:47 schrieb Stefan Priebe

Re: [ceph-users] ceph mgr balancer bad distribution

2018-02-28 Thread Stefan Priebe - Profihost AG
de Yes only one iteration but i set max_misplaced to 20%: "mgr/balancer/max_misplaced": "20.00", > > -- dan > > > On Wed, Feb 28, 2018 at 1:47 PM, Stefan Priebe - Profihost AG > <s.pri...@profihost.ag> wrote: >> Hello, >> >> with jewel

Re: [ceph-users] ceph mgr balancer bad distribution

2018-02-28 Thread Stefan Priebe - Profihost AG
Am 28.02.2018 um 13:59 schrieb John Spray: > On Wed, Feb 28, 2018 at 12:47 PM, Stefan Priebe - Profihost AG > <s.pri...@profihost.ag> wrote: >> Hello, >> >> with jewel we always used the python crush optimizer which gave us a >> pretty good distribution fo

[ceph-users] ceph mgr balancer bad distribution

2018-02-28 Thread Stefan Priebe - Profihost AG
Hello, with jewel we always used the python crush optimizer which gave us a pretty good distribution fo the used space. Since luminous we're using the included ceph mgr balancer but the distribution is far from perfect and much worse than the old method. Is there any way to tune the mgr

Re: [ceph-users] how to fix X is an unexpected clone

2018-02-26 Thread Stefan Priebe - Profihost AG
tool --data-path /.../osd.$OSD/ --journal-path /dev/disk/by-partlabel/journal$OSD rbd_data.$RBD remove-clone-metadata $CLONEID > > thanks > > Saverio > > > On 08.08.17 12:02, Stefan Priebe - Profihost AG wrote: >> Hello Greg, >> >> Am 08.08.2017 um 11:5

Re: [ceph-users] measure performance / latency in blustore

2018-02-07 Thread Stefan Priebe - Profihost AG
Hi, Am 14.12.2017 um 15:02 schrieb Sage Weil: > On Thu, 14 Dec 2017, Stefan Priebe - Profihost AG wrote: >> >> Am 14.12.2017 um 13:22 schrieb Sage Weil: >>> On Thu, 14 Dec 2017, Stefan Priebe - Profihost AG wrote: >>>> Hello, >>>> >>>>

[ceph-users] udev rule or script to auto add bcache devices?

2018-01-20 Thread Stefan Priebe - Profihost AG
Hello, bcache didn't supported partitions on the past so that a lot of our osds have their data directly on: /dev/bcache[0-9] But that means i can't give them the needed part type of 4fbd7e29-9d25-41b8-afd0-062c0ceff05d and that means that the activation with udev und ceph-disk does not work.

[ceph-users] how to update old pre ceph-deploy osds to current systemd way?

2018-01-17 Thread Stefan Priebe - Profihost AG
Hello, i've some osds which were created under bobtail or argonaut (pre ceph-deploy). Those are not recognized as a ceph-osd@57.service . Also they have an entry in the ceph.conf: [osd.12] host=1336 osd_data = /ceph/osd.$id/ osd_journal =

[ceph-users] manually remove problematic snapset: ceph-osd crashes

2018-01-16 Thread Stefan Priebe - Profihost AG
Hello, my ceph-osd luminous are crashing with a segmentation fault while backfilling. Is there any way to manually remove the problematic "data"? -1> 2018-01-16 20:32:50.001722 7f27d53fe700 0 osd.86 pg_epoch: 917877 pg[3.80e( v 917875'69934125 (917365'69924082,917875'69934125] lb

Re: [ceph-users] luminous: HEALTH_ERR full ratio(s) out of order

2018-01-10 Thread Stefan Priebe - Profihost AG
Am 10.01.2018 um 16:38 schrieb Sage Weil: > On Wed, 10 Jan 2018, John Spray wrote: >> On Wed, Jan 10, 2018 at 2:11 PM, Stefan Priebe - Profihost AG >> <s.pri...@profihost.ag> wrote: >>> Hello, >>> >>> since upgrading to luminous i get the follo

[ceph-users] luminous: HEALTH_ERR full ratio(s) out of order

2018-01-10 Thread Stefan Priebe - Profihost AG
Hello, since upgrading to luminous i get the following error: HEALTH_ERR full ratio(s) out of order OSD_OUT_OF_ORDER_FULL full ratio(s) out of order backfillfull_ratio (0.9) < nearfull_ratio (0.95), increased but ceph.conf has: mon_osd_full_ratio = .97

Re: [ceph-users] mon_max_pg_per_osd setting not active? too many PGs per OSD (240 > max 200)

2018-01-04 Thread Stefan Priebe - Profihost AG
Am 04.01.2018 um 18:37 schrieb Gregory Farnum: > On Thu, Jan 4, 2018 at 4:57 AM Stefan Priebe - Profihost AG > <s.pri...@profihost.ag <mailto:s.pri...@profihost.ag>> wrote: > > Hello, > > i set mon_max_pg_per_osd to 300 but the cluster keeps in

[ceph-users] mon_max_pg_per_osd setting not active? too many PGs per OSD (240 > max 200)

2018-01-04 Thread Stefan Priebe - Profihost AG
Hello, i set mon_max_pg_per_osd to 300 but the cluster keeps in warn state. # ceph -s cluster: id: 5482b798-0bf1-4adb-8d7a-1cd57bdc1905 health: HEALTH_WARN too many PGs per OSD (240 > max 200) # ceph --admin-daemon /var/run/ceph/ceph-mon.1.asok config show|grep -i

Re: [ceph-users] measure performance / latency in blustore

2017-12-14 Thread Stefan Priebe - Profihost AG
Am 14.12.2017 um 13:22 schrieb Sage Weil: > On Thu, 14 Dec 2017, Stefan Priebe - Profihost AG wrote: >> Hello, >> >> Am 21.11.2017 um 11:06 schrieb Stefan Priebe - Profihost AG: >>> Hello, >>> >>> to measure performance / latency fo

Re: [ceph-users] measure performance / latency in blustore

2017-12-13 Thread Stefan Priebe - Profihost AG
Hello, Am 21.11.2017 um 11:06 schrieb Stefan Priebe - Profihost AG: > Hello, > > to measure performance / latency for filestore we used: > filestore:apply_latency > filestore:commitcycle_latency > filestore:journal_latency > filestore:queue_transaction_latency_avg > >

[ceph-users] measure performance / latency in blustore

2017-11-21 Thread Stefan Priebe - Profihost AG
Hello, to measure performance / latency for filestore we used: filestore:apply_latency filestore:commitcycle_latency filestore:journal_latency filestore:queue_transaction_latency_avg What are the correct ones for bluestore? Greets, Stefan ___

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-11-12 Thread Stefan Priebe - Profihost AG
Am 12.11.2017 um 17:55 schrieb Sage Weil: > On Wed, 25 Oct 2017, Sage Weil wrote: >> On Wed, 25 Oct 2017, Stefan Priebe - Profihost AG wrote: >>> Hello, >>> >>> in the lumious release notes is stated that zstd is not supported by >>> bluestor due t

[ceph-users] python crush tools uses pre luminous health status

2017-11-06 Thread Stefan Priebe - Profihost AG
Hi, while trying to use python crush tools on a luminous cluster i get: crush.ceph.HealthError: expected health overall_status == HEALTH_OK but got HEALTH_WARNinstead It seems crush-1.0.35 uses the deprecated overall_status element. Greets, Stefan

[ceph-users] Kernel Version for min client luminous

2017-11-05 Thread Stefan Priebe - Profihost AG
Hello, is there already a kernel available which connects with luminous? ceph features tells for my kernel clients still release jewel. Greets, Stefan ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-11-05 Thread Stefan Priebe - Profihost AG
] Compressor/CompressorTest.decompress_16384/3 (1449 ms) [--] 64 tests from Compressor/CompressorTest (29128 ms total) Greets, Stefan Am 04.11.2017 um 21:10 schrieb Sage Weil: > On Sat, 4 Nov 2017, Stefan Priebe - Profihost AG wrote: >> Hi Sage, >> >> Am 26.10.2017 um 13

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-11-05 Thread Stefan Priebe - Profihost AG
idea why it tries to use libceph_invalid and libceph_example Stefan Am 04.11.2017 um 21:10 schrieb Sage Weil: > On Sat, 4 Nov 2017, Stefan Priebe - Profihost AG wrote: >> Hi Sage, >> >> Am 26.10.2017 um 13:58 schrieb Sage Weil: >>> On Thu, 26 Oct 2017, Stefan Prieb

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-11-05 Thread Stefan Priebe - Profihost AG
Hi, arg sorry i got it. Building the test right now. Will report results shortly. Greets, Stefan Am 04.11.2017 um 21:10 schrieb Sage Weil: > On Sat, 4 Nov 2017, Stefan Priebe - Profihost AG wrote: >> Hi Sage, >> >> Am 26.10.2017 um 13:58 schrieb Sage Weil: >>> On T

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-11-05 Thread Stefan Priebe - Profihost AG
n make: *** No rule to make target 'unittest_compression'. Stop. Stefan Am 04.11.2017 um 21:10 schrieb Sage Weil: > On Sat, 4 Nov 2017, Stefan Priebe - Profihost AG wrote: >> Hi Sage, >> >> Am 26.10.2017 um 13:58 schrieb Sage Weil: >>> On Thu, 26 Oct 2017, Stefan Priebe - Pr

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-11-04 Thread Stefan Priebe - Profihost AG
OK zstd.h needs: #define ZSTD_STATIC_LINKING_ONLY to export that one. Stefan Am 04.11.2017 um 21:23 schrieb Stefan Priebe - Profihost AG: > Thanks - not a C++ guy what's wrong with it? > > /build/ceph/src/compressor/zstd/ZstdCompressor.h: In member function > 'virtual int Zs

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-11-04 Thread Stefan Priebe - Profihost AG
from previous init; skip dictionary loading stage; zcs must be init at least once before */ Stefan Am 04.11.2017 um 21:10 schrieb Sage Weil: > On Sat, 4 Nov 2017, Stefan Priebe - Profihost AG wrote: >> Hi Sage, >> >> Am 26.10.2017 um 13:58 schrieb Sage Weil: >>> On Thu

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-11-04 Thread Stefan Priebe - Profihost AG
Hi Sage, Am 26.10.2017 um 13:58 schrieb Sage Weil: > On Thu, 26 Oct 2017, Stefan Priebe - Profihost AG wrote: >> Hi Sage, >> >> Am 25.10.2017 um 21:54 schrieb Sage Weil: >>> On Wed, 25 Oct 2017, Stefan Priebe - Profihost AG wrote: >>>> Hello, >&g

[ceph-users] Ceph and intels new scalable Xeons - replacement for E5-1650v4

2017-11-01 Thread Stefan Priebe - Profihost AG
Hello list, in the past we used the E5-1650v4 for our SSD based ceph nodes which worked fine. The new xeon generation doesn't seem to have a replacement. The only one which is still 0.2Ghz slower is the Intel Xeon Gold 6128. But the price is 3 times as high. So the question is is there any

Re: [ceph-users] crush optimize does not work

2017-10-29 Thread Stefan Priebe - Profihost AG
Hello, Am 27.10.2017 um 19:00 schrieb David Turner: > What does your crush map look like?  Also a `ceph df` output.  You're > optimizing your map for pool #5, if there are other pools with a > significant amount of data, then your going to be off on your cluster > balance. There are no other

[ceph-users] crush optimize does not work

2017-10-26 Thread Stefan Priebe - Profihost AG
Hello, while trying to optimize a ceph cluster running jewel i get the following output: 2017-10-26 10:43:27,615 argv = optimize --crushmap /home/spriebe/ceph.report --out-path /home/spriebe/optimized.crush --pool 5 --pool=5 --choose-args=5 --replication-count=3 --pg-num=4096 --pgp-num=4096

Re: [ceph-users] ceph zstd not for bluestor due to performance reasons

2017-10-26 Thread Stefan Priebe - Profihost AG
Hi Sage, Am 25.10.2017 um 21:54 schrieb Sage Weil: > On Wed, 25 Oct 2017, Stefan Priebe - Profihost AG wrote: >> Hello, >> >> in the lumious release notes is stated that zstd is not supported by >> bluestor due to performance reason. I'm wondering why btrfs instead >

[ceph-users] ceph zstd not for bluestor due to performance reasons

2017-10-25 Thread Stefan Priebe - Profihost AG
Hello, in the lumious release notes is stated that zstd is not supported by bluestor due to performance reason. I'm wondering why btrfs instead states that zstd is as fast as lz4 but compresses as good as zlib. Why is zlib than supported by bluestor? And why does btrfs / facebook behave

[ceph-users] ceph zstd not for bluestor due to performance reasons

2017-10-25 Thread Stefan Priebe - Profihost AG
Hello, in the lumious release notes is stated that zstd is not supported by bluestor due to performance reason. I'm wondering why btrfs instead states that zstd is as fast as lz4 but compresses as good as zlib. Why is zlib than supported by bluestor? And why does btrfs / facebook behave

Re: [ceph-users] PCIe journal benefit for SSD OSDs

2017-09-09 Thread Stefan Priebe - Profihost AG
18 x intel s3610 1,6TB in coming > weeks. > > I'll send results on the mailing. Thanks! Greets, Stefan > ----- Mail original - > De: "Stefan Priebe, Profihost AG" <s.pri...@profihost.ag> > À: "Christian Balzer" <ch...@gol.com>, "ceph-users

Re: [ceph-users] PCIe journal benefit for SSD OSDs

2017-09-07 Thread Stefan Priebe - Profihost AG
Am 07.09.2017 um 10:44 schrieb Christian Balzer: > > Hello, > > On Thu, 7 Sep 2017 08:03:31 +0200 Stefan Priebe - Profihost AG wrote: > >> Hello, >> Am 07.09.2017 um 03:53 schrieb Christian Balzer: >>> >>> Hello, >>> >>> On Wed,

Re: [ceph-users] PCIe journal benefit for SSD OSDs

2017-09-07 Thread Stefan Priebe - Profihost AG
f. You do this by writing the string "temporary write through" to /sys/block/sdb/device/scsi_disk/*/cache_type Greets, Stefan > > -Original Message- > From: Stefan Priebe - Profihost AG [mailto:s.pri...@profihost.ag] > Sent: donderdag 7 september 2017 8:04 > To:

Re: [ceph-users] PCIe journal benefit for SSD OSDs

2017-09-07 Thread Stefan Priebe - Profihost AG
Hello, Am 07.09.2017 um 03:53 schrieb Christian Balzer: > > Hello, > > On Wed, 6 Sep 2017 09:09:54 -0400 Alex Gorbachev wrote: > >> We are planning a Jewel filestore based cluster for a performance >> sensitive healthcare client, and the conservative OSD choice is >> Samsung SM863A. >> > >

Re: [ceph-users] how to fix X is an unexpected clone

2017-08-08 Thread Stefan Priebe - Profihost AG
Hello Greg, Am 08.08.2017 um 11:56 schrieb Gregory Farnum: > On Mon, Aug 7, 2017 at 11:55 PM Stefan Priebe - Profihost AG > <s.pri...@profihost.ag <mailto:s.pri...@profihost.ag>> wrote: > > Hello, > > how can i fix this one: > > 2017-08-08 08:42

  1   2   3   >