Re: [ceph-users] How to reduce HDD OSD flapping due to rocksdb compacting event?

2019-04-11 Thread Charles Alva
Hi Christian and Wido, I used the daily digest and lost way to reply without heavy editing the replies. I will change my subscription to be individual message later. How big is the disk? RocksDB will need to compact at some point and it > seems that the HDD can't keep up. > I've seen this with

Re: [ceph-users] How to reduce HDD OSD flapping due to rocksdb compacting event?

2019-04-11 Thread Christian Balzer
Hello Charles, On Wed, 10 Apr 2019 14:07:58 +0700 Charles Alva wrote: > Hi Ceph Users, > > Is there a way around to minimize rocksdb compacting event so that it won't > use all the spinning disk IO utilization and avoid it being marked as down > due to fail to send heartbeat to others? > >

Re: [ceph-users] BADAUTHORIZER in Nautilus

2019-04-11 Thread Shawn Edwards
Cluster is back and clean again. So I started adding plugins and such back to the mix. After adding the 'balancer' back, I got crashes in the mgr log. ceph-post-file: 0feb1562-cdc5-4a99-86ee-91006eaf6056 Turned balancer back off for now. On Tue, Apr 9, 2019 at 9:38 AM Shawn Edwards wrote: >

Re: [ceph-users] bluefs-bdev-expand experience

2019-04-11 Thread Yury Shevchuk
Hi Igor! I have upgraded from Luminous to Nautilus and now slow device expansion works indeed. The steps are shown below to round up the topic. node2# ceph osd df ID CLASS WEIGHT REWEIGHT SIZERAW USE DATAOMAPMETAAVAIL %USE VAR PGS STATUS 0 hdd 0.22739 1.0 233 GiB

Re: [ceph-users] Topology query

2019-04-11 Thread Bob Farrell
Thanks a lot, Marc - this looks similar to the post I found: http://lists.ceph.com/pipermail/ceph-users-ceph.com/2015-July/003369.html It seems to suggest that this wouldn't be an issue in more recent kernels but would be great to get confirmation on that. I'll keep researching. On Thu, 11 Apr

Re: [ceph-users] Topology query

2019-04-11 Thread Marc Roos
AFAIK you at least risk with cephfs on osd nodes this 'kernel deadlock'? I have it also, but with enough memory. Search mailing list for this. I am looking at similar setup, but with mesos and strugling with some cni plugin we have to develop. -Original Message- From: Bob Farrell

[ceph-users] Topology query

2019-04-11 Thread Bob Farrell
Hello. I am running Ceph Nautilus v14.2.0 on Ubuntu Bionic 18.04 LTS. I would like to ask if anybody could advise if there will be any potential problems with my setup as I am running a lot of services on each node. I have 8 large dedicated servers, each with two physical disks. All servers run

Re: [ceph-users] reshard list

2019-04-11 Thread Andrew Cassera
Thank you I found my problem. On Wed, Apr 10, 2019 at 10:00 PM Konstantin Shalygin wrote: > Hello, > > I am have been managing a ceph cluster running 12.2.11. This was running > 12.2.5 until the recent upgrade three months ago. We build another cluster > running 13.2.5 and synced the data

[ceph-users] mimic stability finally achieved

2019-04-11 Thread Mazzystr
I think I finally have a stable containerized mimic cluster... je zez! It was hard enough! I'm currently repopulating cephfs and cruising along at ... client: 147 MiB/s wr, 0 op/s rd, 38 op/s wr First last month I had four Seagate Barracuda drive failures at the same time with around

Re: [ceph-users] Glance client and RBD export checksum mismatch

2019-04-11 Thread Jason Dillaman
On Thu, Apr 11, 2019 at 8:49 AM Erik McCormick wrote: > > > > On Thu, Apr 11, 2019, 8:39 AM Erik McCormick > wrote: >> >> >> >> On Thu, Apr 11, 2019, 12:07 AM Brayan Perera wrote: >>> >>> Dear Jason, >>> >>> >>> Thanks for the reply. >>> >>> We are using python 2.7.5 >>> >>> Yes. script is

Re: [ceph-users] Glance client and RBD export checksum mismatch

2019-04-11 Thread Erik McCormick
On Thu, Apr 11, 2019, 8:39 AM Erik McCormick wrote: > > > On Thu, Apr 11, 2019, 12:07 AM Brayan Perera > wrote: > >> Dear Jason, >> >> >> Thanks for the reply. >> >> We are using python 2.7.5 >> >> Yes. script is based on openstack code. >> >> As suggested, we have tried chunk_size 32 and 64,

Re: [ceph-users] Glance client and RBD export checksum mismatch

2019-04-11 Thread Jason Dillaman
On Thu, Apr 11, 2019 at 12:07 AM Brayan Perera wrote: > > Dear Jason, > > > Thanks for the reply. > > We are using python 2.7.5 > > Yes. script is based on openstack code. > > As suggested, we have tried chunk_size 32 and 64, and both giving same > incorrect checksum value. > > We tried to copy

Re: [ceph-users] Glance client and RBD export checksum mismatch

2019-04-11 Thread Erik McCormick
On Thu, Apr 11, 2019, 12:07 AM Brayan Perera wrote: > Dear Jason, > > > Thanks for the reply. > > We are using python 2.7.5 > > Yes. script is based on openstack code. > > As suggested, we have tried chunk_size 32 and 64, and both giving same > incorrect checksum value. > The value of

[ceph-users] multi-site between luminous and mimic broke etag

2019-04-11 Thread Tomasz PÅ‚aza
Hi Ceph Users, In our lab on VirtualBox we have installed two Centos7 VMs. One with ceph v12.2.11 and the second one with ceph v13.2.5. We connect them using multi-site (it does not matter witch one is hosting the master zone). On master zone we: created a user, made a bucket, uploaded a

[ceph-users] Kraken - Pool storage MAX AVAIL drops by 30TB after disk failure

2019-04-11 Thread nokia ceph
Hi, We have a 5 node EC 4+1 cluster with 335 OSDs running Kraken Bluestore 11.2.0. There was a disk failure on one of the OSDs and the disk was replaced. After which it was noticed that there was a ~30TB drop in the MAX_AVAIL value for the pool storage details on output of 'ceph df' Even though