[ceph-users] RuntimeError: Failed to connect any mon

2016-09-21 Thread Rens Vermeulen
Dear ceph-users, I'm preparing a Ceph cluster on my debian machines. I successfully walked true the Preflight installation guide on the website. Now i'm stuck at the STORAGE CLUSTER QUICK START, when i enter the following command: ceph-deploy mon create-initial I get the following errors: [ce

Re: [ceph-users] Ceph Rust Librados

2016-09-21 Thread Brad Hubbard
On Thu, Sep 22, 2016 at 1:42 AM, Chris Jones wrote: > Ceph-rust for librados has been released. It's an API interface in Rust for > all of librados that's a thin layer above the C APIs. There are low-level > direct access and higher level Rust helpers that make working directly with > librados sim

Re: [ceph-users] Snap delete performance impact

2016-09-21 Thread Adrian Saul
Any guidance on this? I have osd_snap_trim_sleep set to 1 and it seems to have tempered some of the issues but its still bad enough that NFS storage off RBD volumes become unavailable for over 3 minutes. It seems that the activity which the snapshot deletes are actioned triggers massive disk

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Heller, Chris
So just to put more info out there, here is what I’m seeing with a Spark/HDFS client: 2016-09-21 20:09:25.076595 7fd61c16f700 0 -- 192.168.1.157:0/634334964 >> 192.168.1.190:6802/32183 pipe(0x7fd5fcef8ca0 sd=66 :53864 s=2 pgs=50445 cs=1 l=0 c=0x7fd5fdd371d0).fault, initiating reconnect 2016-09

Re: [ceph-users] radosgw bucket name performance

2016-09-21 Thread Stas Starikevich
Felix, According to my tests there is difference in performance between usual named buckets (test, test01, test02), uuid-named buckets (like '7c9e4a81-df86-4c9d-a681-3a570de109db') or just date ('2016-09-20-16h'). Getting ~3x more upload performance (220 uploads\s vs 650 uploads\s) with SSD-bac

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Heller, Chris
I also went and bumped mds_cache_size up to 1 million… still seeing cache pressure, but I might just need to evict those clients… On 9/21/16, 9:24 PM, "Heller, Chris" wrote: What is the interesting value in ‘session ls’? Is it ‘num_leases’ or ‘num_caps’ leases appears to be, on average, 1.

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Heller, Chris
What is the interesting value in ‘session ls’? Is it ‘num_leases’ or ‘num_caps’ leases appears to be, on average, 1. But caps seems to be 16385 for many many clients! -Chris On 9/21/16, 9:22 PM, "Gregory Farnum" wrote: On Wed, Sep 21, 2016 at 6:13 PM, Heller, Chris wrote: > I’m suspe

Re: [ceph-users] rgw bucket index manual copy

2016-09-21 Thread Stas Starikevich
Ben, Works fine as far as I see: [root@273aa9f2ee9f /]# s3cmd mb s3://test Bucket 's3://test/' created [root@273aa9f2ee9f /]# s3cmd put /etc/hosts s3://test upload: '/etc/hosts' -> 's3://test/hosts' [1 of 1] 196 of 196 100% in0s 404.87 B/s done [root@273aa9f2ee9f /]# s3cmd ls s3://te

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Gregory Farnum
On Wed, Sep 21, 2016 at 6:13 PM, Heller, Chris wrote: > I’m suspecting something similar, we have millions of files and can read a > huge subset of them at a time, presently the client is Spark 1.5.2 which I > suspect is leaving the closing of file descriptors up to the garbage > collector. Tha

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Heller, Chris
I’m suspecting something similar, we have millions of files and can read a huge subset of them at a time, presently the client is Spark 1.5.2 which I suspect is leaving the closing of file descriptors up to the garbage collector. That said, I’d like to know if I could verify this theory using th

Re: [ceph-users] rgw bucket index manual copy

2016-09-21 Thread Ben Hines
Thanks. Will try it out once we get on Jewel. Just curious, does bucket deletion with --purge-objects work via radosgw-admin with the no index option? If not, i imagine rados could be used to delete them manually by prefix. On Sep 21, 2016 6:02 PM, "Stas Starikevich" wrote: > Hi Ben, > > Since

Re: [ceph-users] rgw bucket index manual copy

2016-09-21 Thread Stas Starikevich
Hi Ben, Since the 'Jewel' RadosGW supports blind buckets. To enable blind buckets configuration I used: radosgw-admin zone get --rgw-zone=default > default-zone.json #change index_type from 0 to 1 vi default-zone.json radosgw-admin zone set --rgw-zone=default --infile default-zone.json To apply

[ceph-users] Upgrading 0.94.6 -> 0.94.9 saturating mon node networking

2016-09-21 Thread Stillwell, Bryan J
While attempting to upgrade a 1200+ OSD cluster from 0.94.6 to 0.94.9 I've run into serious performance issues every time I restart an OSD. At first I thought the problem I was running into was caused by the osdmap encoding bug that Dan and Wido ran into when upgrading to 0.94.7, because I was see

[ceph-users] rgw multi-site replication issues

2016-09-21 Thread John Rowe
Hello, We have 2 Ceph clusters running in two separate data centers, each one with 3 mons, 3 rgws, and 5 osds. I am attempting to get bi-directional multi-site replication setup as described in the ceph documentation here: http://docs.ceph.com/docs/jewel/radosgw/multisite/ We are running Jewel v

Re: [ceph-users] Object lost

2016-09-21 Thread Jason Dillaman
Unfortunately, it sounds like the image's header object was lost during your corruption event. While you can manually retrieve the image data blocks from the cluster, undoubtedly many might be lost and/or corrupted as well. You'll first need to determine the internal id of your image: $ rados --po

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Gregory Farnum
On Wed, Sep 21, 2016 at 1:16 PM, Heller, Chris wrote: > Ok. I just ran into this issue again. The mds rolled after many clients were > failing to relieve cache pressure. That definitely could have had something to do with it, if say they overloaded the MDS so much it got stuck in a directory rea

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Heller, Chris
Ok. I just ran into this issue again. The mds rolled after many clients were failing to relieve cache pressure. Now here is the result of `ceph –s` # ceph -s cluster b126570e-9e7c-0bb2-991f-ecf9abe3afa0 health HEALTH_OK monmap e1: 5 mons at {a154=192.168.1.154:6789/0,a155=192.168.

Re: [ceph-users] Stat speed for objects in ceph

2016-09-21 Thread Iain Buclaw
On 21 September 2016 at 17:28, Haomai Wang wrote: > BTW, why you need to iterate so much objects. I think it should be > done by other ways to achieve the goal. > Mostly it's just a brute force way to identify objects that shouldn't exist, or objects that have been orphaned (e.g: last modific

Re: [ceph-users] rgw bucket index manual copy

2016-09-21 Thread Ben Hines
Nice, thanks! Must have missed that one. It might work well for our use case since we don't really need the index. -Ben On Wed, Sep 21, 2016 at 11:23 AM, Gregory Farnum wrote: > On Wednesday, September 21, 2016, Ben Hines wrote: > >> Yes, 200 million is way too big for a single ceph RGW bucket

Re: [ceph-users] rgw bucket index manual copy

2016-09-21 Thread Gregory Farnum
On Wednesday, September 21, 2016, Ben Hines wrote: > Yes, 200 million is way too big for a single ceph RGW bucket. We > encountered this problem early on and sharded our buckets into 20 buckets, > each which have the sharded bucket index with 20 shards. > > Unfortunately, enabling the sharded RGW

Re: [ceph-users] rgw bucket index manual copy

2016-09-21 Thread Ben Hines
Yes, 200 million is way too big for a single ceph RGW bucket. We encountered this problem early on and sharded our buckets into 20 buckets, each which have the sharded bucket index with 20 shards. Unfortunately, enabling the sharded RGW index requires recreating the bucket and all objects. The fa

Re: [ceph-users] crash of osd using cephfs jewel 10.2.2, and corruption

2016-09-21 Thread Peter Maloney
It seems the trigger for the problem is this: > 24.9141130d 1000527. [write 0~242] snapc 1=[] ondisk+write > e320) >-40> 2016-09-20 20:38:02.007942 708f67bbd700 0 > filestore(/var/lib/ceph/osd/ceph-0) write couldn't open > 24.32_head/#24:4d11884b:::1000504.:head#: (24)

[ceph-users] radosgw bucket name performance

2016-09-21 Thread Félix Barbeira
Hi, Regarding to Amazon S3 documentation, it is advised to insert a bit of random chars in the bucket name in order to gain performance. This is related to how Amazon store key names. It looks like they store an index of object key names in each region. http://docs.aws.amazon.com/AmazonS3/latest/

[ceph-users] Ceph Rust Librados

2016-09-21 Thread Chris Jones
Ceph-rust for librados has been released. It's an API interface in Rust for all of librados that's a thin layer above the C APIs. There are low-level direct access and higher level Rust helpers that make working directly with librados simple. ​The official repo is: https://github.com/ceph/ceph-rus

Re: [ceph-users] Stat speed for objects in ceph

2016-09-21 Thread Wido den Hollander
> Op 21 september 2016 om 17:23 schreef Iain Buclaw : > > > On 20 September 2016 at 19:27, Gregory Farnum wrote: > > In librados getting a stat is basically equivalent to reading a small > > object; there's not an index or anything so FileStore needs to descend its > > folder hierarchy. If look

Re: [ceph-users] Stat speed for objects in ceph

2016-09-21 Thread Haomai Wang
BTW, why you need to iterate so much objects. I think it should be done by other ways to achieve the goal. On Wed, Sep 21, 2016 at 11:23 PM, Iain Buclaw wrote: > On 20 September 2016 at 19:27, Gregory Farnum wrote: >> In librados getting a stat is basically equivalent to reading a small >> o

Re: [ceph-users] Stat speed for objects in ceph

2016-09-21 Thread Iain Buclaw
On 20 September 2016 at 19:27, Gregory Farnum wrote: > In librados getting a stat is basically equivalent to reading a small > object; there's not an index or anything so FileStore needs to descend its > folder hierarchy. If looking at metadata for all the objects in the system > efficiently is im

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Heller, Chris
Perhaps related, I was watching the active mds with debug_mds set to 5/5, when I saw this in the log: 2016-09-21 15:13:26.067698 7fbaec248700 0 -- 192.168.1.196:6802/13581 >> 192.168.1.238:0/3488321578 pipe(0x55db000 sd=49 :6802 s=2 pgs=2 cs=1 l=0 c=0x5631ce0).fault with nothing to send, going

[ceph-users] Object lost

2016-09-21 Thread Fran Barrera
Hello, I have a Ceph Jewel cluster with 4 osds and only one monitor integrated with Openstack Mitaka. Two OSD were down, with a service restart one of them was recovered. The cluster began to recover and was OK. Finally the disk of the other OSD was corrupted and the solution was a format and rec

Re: [ceph-users] Stat speed for objects in ceph

2016-09-21 Thread Iain Buclaw
On 21 September 2016 at 02:57, Haomai Wang wrote: > On Wed, Sep 21, 2016 at 2:41 AM, Wido den Hollander wrote: >> >>> Op 20 september 2016 om 20:30 schreef Haomai Wang : >>> >>> >>> On Wed, Sep 21, 2016 at 2:26 AM, Wido den Hollander wrote: >>> > >>> >> Op 20 september 2016 om 19:27 schreef Greg

[ceph-users] Fwd: Error

2016-09-21 Thread Rens Vermeulen
Dear ceph-users, I'm preparing a Ceph cluster on my debian machines. I successfully walked true the Preflight installation guide on the website. Now i'm stuck at the STORAGE CLUSTER QUICK START, when i enter the following command: ceph-deploy mon create-initial I get the following errors: [ce

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Heller, Chris
I’ll see if I can capture the output the next time this issue arises, but in general the output looks as if nothing is wrong. No OSD are down, a ‘ceph health detail’ results in HEALTH_OK, the mds server is in the up:active state, in general it’s as if nothing is wrong server side (at least from

Re: [ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Gregory Farnum
On Wed, Sep 21, 2016 at 6:30 AM, Heller, Chris wrote: > I’m running a production 0.94.7 Ceph cluster, and have been seeing a > periodic issue arise where in all my MDS clients will become stuck, and the > fix so far has been to restart the active MDS (sometimes I need to restart > the subsequent a

Re: [ceph-users] Help on RGW NFS function

2016-09-21 Thread Matt Benjamin
Hi, - Original Message - > From: "yiming xie" > To: ceph-users@lists.ceph.com > Sent: Wednesday, September 21, 2016 3:53:35 AM > Subject: [ceph-users] Help on RGW NFS function > > Hi, > I have some question about rgw nfs. > > ceph release notes: You can now access radosgw buckets via NF

[ceph-users] Faulting MDS clients, HEALTH_OK

2016-09-21 Thread Heller, Chris
I’m running a production 0.94.7 Ceph cluster, and have been seeing a periodic issue arise where in all my MDS clients will become stuck, and the fix so far has been to restart the active MDS (sometimes I need to restart the subsequent active MDS as well). These clients are using the cephfs-hado

Re: [ceph-users] cache tier on rgw index pool

2016-09-21 Thread Samuel Just
I seriously doubt that it's ever going to be a winning strategy to let rgw index objects go to a cold tier. Some practical problems: 1) We don't track omap size (the leveldb entries for an object) because it would turn writes into rmw's -- so they always show up as 0 size. Thus, the target_max_by

Re: [ceph-users] Same pg scrubbed over and over (Jewel)

2016-09-21 Thread Samuel Just
Ah, same question then. If we can get logging on the primary for one of those pgs, it should be fairly obvious. -Sam On Wed, Sep 21, 2016 at 4:08 AM, Pavan Rallabhandi wrote: > We find this as well in our fresh built Jewel clusters, and seems to happen > only with a handful of PGs from couple o

Re: [ceph-users] swiftclient call radosgw, it always response 401 Unauthorized

2016-09-21 Thread Radoslaw Zarzynski
Hi, Responded inline. On Wed, Sep 21, 2016 at 4:54 AM, Brian Chang-Chien wrote: > > > [global] > ... > debug rgw = 20 > [client.radosgw.gateway] > host = brianceph > rgw keystone url = http://10.62.13.253:35357 > rgw keystone admin token = 7bb8e26cbc714c47a26ffec3d96f246f > rgw keystone accepted

Re: [ceph-users] Same pg scrubbed over and over (Jewel)

2016-09-21 Thread Pavan Rallabhandi
We find this as well in our fresh built Jewel clusters, and seems to happen only with a handful of PGs from couple of pools. Thanks! On 9/21/16, 3:14 PM, "ceph-users on behalf of Tobias Böhm" wrote: Hi, there is an open bug in the tracker: http://tracker.ceph.com/issues/16474

Re: [ceph-users] crash of osd using cephfs jewel 10.2.2, and corruption

2016-09-21 Thread Samuel Just
Looks like the OSD didn't like an error return it got from the underlying fs. Can you reproduce with debug filestore = 20 debug osd = 20 debug ms = 1 on the osd and post the whole log? -Sam On Wed, Sep 21, 2016 at 12:10 AM, Peter Maloney wrote: > Hi, > > I created a one disk osd with data and

Re: [ceph-users] Same pg scrubbed over and over (Jewel)

2016-09-21 Thread Samuel Just
Can you reproduce with logging on the primary for that pg? debug osd = 20 debug filestore = 20 debug ms = 1 Since restarting the osd may be a workaround, can you inject the debug values without restarting the daemon? -Sam On Wed, Sep 21, 2016 at 2:44 AM, Tobias Böhm wrote: > Hi, > > there is an

Re: [ceph-users] swiftclient call radosgw, it always response 401 Unauthorized

2016-09-21 Thread Brian Chang-Chien
Hi Radoslaw Zarzynski I also test Ceph Hammer 0.94.7 for rados gateway the result still response 401 , so i post ceph.conf, radosgw.log as following ceph.conf when i call swift cmd , radosgw.log 2016-09-21 17:47:12.141783 7fc7f4bf5700 10 content=MIIBzgYJKoZIhvcNAQcCoIIBvzCCAbsCAQExDTALBglghk

Re: [ceph-users] Same pg scrubbed over and over (Jewel)

2016-09-21 Thread Tobias Böhm
Hi, there is an open bug in the tracker: http://tracker.ceph.com/issues/16474 It also suggests restarting OSDs as a workaround. We faced the same issue after increasing the number of PGs in our cluster and restarting OSDs solved it as well. Tobias > Am 21.09.2016 um 11:26 schrieb Dan van der

Re: [ceph-users] Same pg scrubbed over and over (Jewel)

2016-09-21 Thread Dan van der Ster
There was a thread about this a few days ago: http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-September/012857.html And the OP found a workaround. Looks like a bug though... (by default PGs scrub at most once per day). -- dan On Tue, Sep 20, 2016 at 10:43 PM, Martin Bureau wrote: > He

[ceph-users] Help on RGW NFS function

2016-09-21 Thread yiming xie
Hi, I have some question about rgw nfs. ceph release notes: You can now access radosgw buckets via NFS (experimental). In addition to the sentence, ceph documents does not do any explanation I don't understand the experimental implications. 1. RGW nfs functional integrity of it? If nfs function i

[ceph-users] ceph pg stuck creating

2016-09-21 Thread Yuriy Karpel
Hi, There is a cluster of ceph. To accelerate dokinuli ssd into it, the card: I ssd gathered in a separate pool and a prescribed rule. But I'm getting the following error msg: [ceph@ceph-adm test]$ ceph -s cluster a73f8be5-0fa6-4a82-aa9b-22121246398e health HEALTH_WARN 64 p

[ceph-users] crash of osd using cephfs jewel 10.2.2, and corruption

2016-09-21 Thread Peter Maloney
Hi, I created a one disk osd with data and separate journal on the same lvm volume group just for test, one mon, one mds on my desktop. I managed to crash the osd just by mounting cephfs and doing cp -a of the linux-stable git tree into it. It crashed after copying 2.1G which only covers some of