Re: [ceph-users] size of inc_osdmap vs osdmap

2019-01-02 Thread xie.xingguo
> Xie, does that sound right? yeah, looks right to me. 原始邮件 发件人:SageWeil 收件人:Sergey Dolgov ; 抄送人:Gregory Farnum ;ceph-users@lists.ceph.com ;ceph-de...@vger.kernel.org ;谢型果10072465; 日 期 :2019年01月03日 11:05 主 题 :Re: [ceph-users] size of inc_osdmap vs osdmap I think that code was

[ceph-users] Mimic 13.2.3?

2019-01-02 Thread Ashley Merrick
Have just run an apt update and have noticed there are some CEPH packages now available for update on my mimic cluster / ubuntu. Have yet to install these yet but it look's like we have the next point release of CEPH Mimic, but not able to see any release note's or official comm's yet?..

[ceph-users] [Ceph-users] Multisite-Master zone still in recover mode

2019-01-02 Thread Amit Ghadge
Hi, We following http://docs.ceph.com/docs/master/radosgw/multisite/ steps to migrate single-site to master zone and then setup secondary zone. We not delete existing data and all objects sync to secondary zone but in master zone it still showing in recovery mode, dynamic resharding is disable.

Re: [ceph-users] any way to see enabled/disabled status of bucket sync?

2019-01-02 Thread Konstantin Shalygin
I had no clue there was a bucket sync status command. This command is present with radosgw cli `radosgw-admin bucket sync status` (but hidden in `--help`, also like `radosgw-admin reshard status` command). k ___ ceph-users mailing list

Re: [ceph-users] size of inc_osdmap vs osdmap

2019-01-02 Thread Sage Weil
I think that code was broken by ea723fbb88c69bd00fefd32a3ee94bf5ce53569c and should be fixed like so: diff --git a/src/mon/OSDMonitor.cc b/src/mon/OSDMonitor.cc index 8376a40668..12f468636f 100644 --- a/src/mon/OSDMonitor.cc +++ b/src/mon/OSDMonitor.cc @@ -1006,7 +1006,8 @@ void

Re: [ceph-users] size of inc_osdmap vs osdmap

2019-01-02 Thread Sergey Dolgov
> > Well those commits made some changes, but I'm not sure what about them > you're saying is wrong? > I mean, that all pgs have "up == acting && next_up == next_acting" but at https://github.com/ceph/ceph/blob/luminous/src/mon/OSDMonitor.cc#L1009 condition "next_up != next_acting" false and we

Re: [ceph-users] size of inc_osdmap vs osdmap

2019-01-02 Thread Sergey Dolgov
Thanks Grag I dumped inc_osdmap to file ceph-dencoder type OSDMap::Incremental import ./inc\\uosdmap.1378266__0_B7F36FFA__none decode dump_json > inc_osdmap.txt There are 52330 pgs(cluster has 52332 pgs) in structure 'new_pg_temp' and for all of them osd is empty. For examle short excerpt: >

Re: [ceph-users] size of inc_osdmap vs osdmap

2019-01-02 Thread Gregory Farnum
On Thu, Dec 27, 2018 at 1:20 PM Sergey Dolgov wrote: > We investigated the issue and set debug_mon up to 20 during little change > of osdmap get many messages for all pgs of each pool (for all cluster): > >> 2018-12-25 19:28:42.426776 7f075af7d700 20 mon.1@0(leader).osd e1373789 >>

Re: [ceph-users] any way to see enabled/disabled status of bucket sync?

2019-01-02 Thread Christian Rice
I had no clue there was a bucket sync status command. And I expect that metadata get command will be useful going forward, as well. Thanks for those! From: ceph-users on behalf of Casey Bodley Date: Wednesday, January 2, 2019 at 1:04 PM To: "ceph-users@lists.ceph.com" Subject: Re:

[ceph-users] TCP qdisc + congestion control / BBR

2019-01-02 Thread Kevin Olbrich
Hi! I wonder if changing qdisc and congestion_control (for example fq with Google BBR) on Ceph servers / clients has positive effects during high load. Google BBR: https://cloud.google.com/blog/products/gcp/tcp-bbr-congestion-control-comes-to-gcp-your-internet-just-got-faster I am running a lot

Re: [ceph-users] any way to see enabled/disabled status of bucket sync?

2019-01-02 Thread Casey Bodley
Hi Christian, The easiest way to do that is probably the 'radosgw-admin bucket sync status' command, which will print "Sync is disabled for bucket ..." if disabled. Otherwise, you could use 'radosgw-admin metadata get' to inspect that flag in the bucket instance metadata. On 12/31/18 2:20

Re: [ceph-users] radosgw-admin unable to store user information

2019-01-02 Thread Casey Bodley
On 12/26/18 4:58 PM, Dilip Renkila wrote: Hi all, Some useful information >>/>> />>///What do the following return?/ >>/>> >> />>/>> >> $ radosgw-admin zone get/ /root@ctrl1:~# radosgw-admin zone get { "id": "8bfdf8a3-c165-44e9-9ed6-deff8a5d852f", "name": "default", "domain_root":

[ceph-users] Compacting omap data

2019-01-02 Thread Bryan Stillwell
Recently on one of our bigger clusters (~1,900 OSDs) running Luminous (12.2.8), we had a problem where OSDs would frequently get restarted while deep-scrubbing. After digging into it I found that a number of the OSDs had very large omap directories (50GiB+). I believe these were OSDs that had

Re: [ceph-users] Usage of devices in SSD pool vary very much

2019-01-02 Thread Konstantin Shalygin
On a medium sized cluster with device-classes, I am experiencing a problem with the SSD pool: root at adminnode :~# ceph osd df | grep ssd ID CLASS WEIGHT REWEIGHT SIZEUSE AVAIL %USE VAR PGS 2 ssd 0.43700 1.0 447GiB

Re: [ceph-users] list admin issues

2019-01-02 Thread David Galloway
On 12/28/18 4:13 AM, Ilya Dryomov wrote: > On Sat, Dec 22, 2018 at 7:18 PM Brian : wrote: >> >> Sorry to drag this one up again. >> >> Just got the unsubscribed due to excessive bounces thing. >> >> 'Your membership in the mailing list ceph-users has been disabled due >> to excessive bounces

[ceph-users] Best way to update object ACL for many files?

2019-01-02 Thread Jin Mao
Ceph Users, Updating ACL and apply s3 policy for millions of objects in a bucket using s3cmd seems to be very slow. I experienced about 3-4 objects/second when doing so. Any one know a faster way to accomplish this task either as ceph user or as ceph admin? Thank you. Jin.

Re: [ceph-users] ceph health JSON format has changed

2019-01-02 Thread Thomas Byrne - UKRI STFC
> In previous versions of Ceph, I was able to determine which PGs had > scrub errors, and then a cron.hourly script ran "ceph pg repair" for them, > provided that they were not already being scrubbed. In Luminous, the bad > PG is not visible in "ceph --status" anywhere. Should I use

Re: [ceph-users] ceph health JSON format has changed sync?

2019-01-02 Thread Konstantin Shalygin
Hello, Ceph users, I am afraid the following question is a FAQ, but I still was not able to find the answer: I use ceph --status --format=json-pretty as a source of CEPH status for my Nagios monitoring. After upgrading to Luminous, I see the following in the JSON output when the cluster

Re: [ceph-users] ceph health JSON format has changed

2019-01-02 Thread Jan Kasprzak
Thomas Byrne - UKRI STFC wrote: : I recently spent some time looking at this, I believe the 'summary' and : 'overall_status' sections are now deprecated. The 'status' and 'checks' : fields are the ones to use now. OK, thanks. : The 'status' field gives you the OK/WARN/ERR, but returning

Re: [ceph-users] ceph health JSON format has changed sync?

2019-01-02 Thread Thomas Byrne - UKRI STFC
I recently spent some time looking at this, I believe the 'summary' and 'overall_status' sections are now deprecated. The 'status' and 'checks' fields are the ones to use now. The 'status' field gives you the OK/WARN/ERR, but returning the most severe error condition from the 'checks' section

Re: [ceph-users] Balancing cluster with large disks - 10TB HHD

2019-01-02 Thread Thomas Byrne - UKRI STFC
Assuming I understand it correctly: "pg_upmap_items 6.0 [40,20]" refers to replacing (upmapping?) osd.40 with osd.20 in the acting set of the placement group '6.0'. Assuming it's a 3 replica PG, the other two OSDs in the set remain unchanged from the CRUSH calculation. "pg_upmap_items 6.6

[ceph-users] Usage of devices in SSD pool vary very much

2019-01-02 Thread Kevin Olbrich
Hi! On a medium sized cluster with device-classes, I am experiencing a problem with the SSD pool: root@adminnode:~# ceph osd df | grep ssd ID CLASS WEIGHT REWEIGHT SIZEUSE AVAIL %USE VAR PGS 2 ssd 0.43700 1.0 447GiB 254GiB 193GiB 56.77 1.28 50 3 ssd 0.43700 1.0

[ceph-users] ceph health JSON format has changed sync?

2019-01-02 Thread Jan Kasprzak
Hello, Ceph users, I am afraid the following question is a FAQ, but I still was not able to find the answer: I use ceph --status --format=json-pretty as a source of CEPH status for my Nagios monitoring. After upgrading to Luminous, I see the following in the JSON output when the cluster