Re: [ceph-users] Consistency problems when taking RBD snapshot

2016-09-15 Thread Nikolay Borisov
On 09/15/2016 03:15 PM, Ilya Dryomov wrote: > On Thu, Sep 15, 2016 at 12:54 PM, Nikolay Borisov wrote: >> >> >> On 09/15/2016 01:24 PM, Ilya Dryomov wrote: >>> On Thu, Sep 15, 2016 at 10:22 AM, Nikolay Borisov >>> wrote: On 09/15/2016

Re: [ceph-users] OSDs thread leak during degraded cluster state

2016-09-15 Thread Kostis Fardelas
Our ceph cluster (from emperor till hammer) has made many times recoveries during host outages/network failures and threads never exceeded 10K. The thread leaks we experienced with down+peering PGs (lasting for several hours) was something that we saw for the first time. I don't see the reason to

Re: [ceph-users] Consistency problems when taking RBD snapshot

2016-09-15 Thread Ilya Dryomov
On Thu, Sep 15, 2016 at 2:43 PM, Nikolay Borisov wrote: > > [snipped] > > cat /sys/bus/rbd/devices/47/client_id > client157729 > cat /sys/bus/rbd/devices/1/client_id > client157729 > > Client client157729 is alxc13, based on correlation by the ip address > shown by the rados -p

[ceph-users] RADOSGW and LDAP

2016-09-15 Thread Andrus, Brian Contractor
All, I have been making some progress on troubleshooting this. I am seeing that when rgw is configured for LDAP, I am getting an error in my slapd log: Sep 14 06:56:21 mgmt1 slapd[23696]: conn=1762 op=0 RESULT tag=97 err=2 text=historical protocol version requested, use LDAPv3 instead Am I

Re: [ceph-users] Designing ceph cluster

2016-09-15 Thread Gaurav Goyal
Dear Ceph users, Any suggestion on this please. Regards Gaurav Goyal On Wed, Sep 14, 2016 at 2:50 PM, Gaurav Goyal wrote: > Dear Ceph Users, > > I need you help to sort out following issue with my cinder volume. > > I have created ceph as backend for cinder. Since i

[ceph-users] mds damage detected - Jewel

2016-09-15 Thread Jim Kilborn
I have a replicated cache pool and metadata pool which reside on ssd drives, with a size of 2, backed by a erasure coded data pool The cephfs filesystem was in a healthy state. I pulled an SSD drive, to perform an exercise in osd failure. The cluster recognized the ssd failure, and replicated

Re: [ceph-users] Cleanup old osdmaps after #13990 fix applied

2016-09-15 Thread Dan Van Der Ster
> On 14 Sep 2016, at 23:07, Gregory Farnum wrote: > > On Wed, Sep 14, 2016 at 7:19 AM, Dan Van Der Ster > wrote: >> Indeed, seems to be trimmed by osd_target_transaction_size (default 30) per >> new osdmap. >> Thanks a lot for your help! > >

[ceph-users] High CPU load with radosgw instances

2016-09-15 Thread lewis.geo...@innoscale.net
Hi, So, maybe someone has an idea of where to go on this. I have just setup 2 rgw instances in a multisite setup. They are working nicely. I have add a couple of test buckets and some files to make sure it works is all. The status shows both are caught up. Nobody else is accessing or using

[ceph-users] Error while searching on the mailing list archives

2016-09-15 Thread Erick Perez - Quadrian Enterprises
Hi, just to let the admins know that when searching for terms (i searched for erasure coding) in the mailing list http://lists.ceph.com/pipermail/ceph-users-ceph.com/ This error is returned in the browser at http://lists.ceph.com/mmsearch.cgi/ceph-users-ceph.com ht://Dig error htsearch detected

Re: [ceph-users] swiftclient call radosgw, it always response 401 Unauthorized

2016-09-15 Thread Brian Chang-Chien
Can anyone know this problem,please help me to watch this 2016年9月13日 下午5:58,"Brian Chang-Chien" 寫道: > Hi ,naga.b > > I use Ceph jewel 10.2.2 > my ceph.conf as follow > [global] > fsid = d056c174-2e3a-4c36-a067-cb774d176ce2 > mon_initial_members = brianceph > mon_host

Re: [ceph-users] High CPU load with radosgw instances

2016-09-15 Thread Yehuda Sadeh-Weinraub
On Thu, Sep 15, 2016 at 4:53 PM, lewis.geo...@innoscale.net wrote: > Hi, > So, maybe someone has an idea of where to go on this. > > I have just setup 2 rgw instances in a multisite setup. They are working > nicely. I have add a couple of test buckets and some files to

Re: [ceph-users] Erasure coding general information Openstack+kvm virtual machine block storage

2016-09-15 Thread Josh Durgin
On 09/16/2016 09:46 AM, Erick Perez - Quadrian Enterprises wrote: Can someone point me to a thread or site that uses ceph+erasure coding to serve block storage for Virtual Machines running with Openstack+KVM? All references that I found are using erasure coding for cold data or *not* VM block

[ceph-users] Samsung DC SV843 SSD

2016-09-15 Thread Quenten Grasso
Hi Everyone, I'm looking for some SSD's for our cluster and I came across these Samsung DC SV843 SSD's and noticed in the mailing lists from awhile back some people were talking about them. Just wondering if anyone ended up using them and how they are going? Thanks in advance, Regards,

Re: [ceph-users] Seeking your feedback on the Ceph monitoring and management functionality in openATTIC

2016-09-15 Thread John Spray
Congratulations on the release! John On Wed, Sep 14, 2016 at 4:08 PM, Lenz Grimmer wrote: > Hi, > > if you're running a Ceph cluster and would be interested in trying out a > new tool for managing/monitoring it, we've just released version 2.0.14 > of openATTIC that now

Re: [ceph-users] Consistency problems when taking RBD snapshot

2016-09-15 Thread Nikolay Borisov
On 09/14/2016 05:53 PM, Ilya Dryomov wrote: > On Wed, Sep 14, 2016 at 3:30 PM, Nikolay Borisov wrote: >> >> >> On 09/14/2016 02:55 PM, Ilya Dryomov wrote: >>> On Wed, Sep 14, 2016 at 9:01 AM, Nikolay Borisov wrote: On 09/14/2016 09:55 AM, Adrian

Re: [ceph-users] jewel blocked requests

2016-09-15 Thread Wido den Hollander
> Op 13 september 2016 om 18:54 schreef "WRIGHT, JON R (JON R)" > : > > > VM Client OS: ubuntu 14.04 > > Openstack: kilo > > libvirt: 1.2.12 > > nova-compute-kvm: 1:2015.1.4-0ubuntu2 > What librados/librbd version are you running on the client? Wido > Jon > > On

Re: [ceph-users] CephFS: Writes are faster than reads?

2016-09-15 Thread Andreas Gerstmayr
Thanks a lot for your explanation! I just increased the 'rasize' option of the kernel module and got significant better throughput for sequential reads. Thanks, Andreas 2016-09-15 0:29 GMT+02:00 Gregory Farnum : > Oh hrm, I missed the stripe count settings. I'm not sure if

Re: [ceph-users] rgw: Swift API X-Storage-Url

2016-09-15 Thread Василий Ангапов
Sorry, a revoke my question. On one node there was a duplicate RGW daemon with old config. That's why sometimes I was receiving wrong URLs. 2016-09-15 13:23 GMT+03:00 Василий Ангапов : > Hello, > > I have Ceph Jewel 10.2.1 cluster and RadosGW. Issue is that when >

[ceph-users] OSDs thread leak during degraded cluster state

2016-09-15 Thread Kostis Fardelas
Hello cephers, being in a degraded cluster state with 6/162 OSDs down ((Hammer 0.94.7, 162 OSDs, 27 "fat" nodes, 1000s of clients) ) like the below ceph cluster log indicates: 2016-09-12 06:26:08.443152 mon.0 62.217.119.14:6789/0 217309 : cluster [INF] pgmap v106027148: 28672 pgs: 2

Re: [ceph-users] cephfs/ceph-fuse: mds0: Client XXX:XXX failingtorespond to capability release

2016-09-15 Thread Wido den Hollander
> Op 14 september 2016 om 14:56 schreef "Dennis Kramer (DT)" : > > > Hi Burkhard, > > Thank you for your reply, see inline: > > On Wed, 14 Sep 2016, Burkhard Linke wrote: > > > Hi, > > > > > > On 09/14/2016 12:43 PM, Dennis Kramer (DT) wrote: > >> Hi Goncalo, > >> > >>

Re: [ceph-users] Consistency problems when taking RBD snapshot

2016-09-15 Thread Ilya Dryomov
On Thu, Sep 15, 2016 at 10:22 AM, Nikolay Borisov wrote: > > > On 09/15/2016 09:22 AM, Nikolay Borisov wrote: >> >> >> On 09/14/2016 05:53 PM, Ilya Dryomov wrote: >>> On Wed, Sep 14, 2016 at 3:30 PM, Nikolay Borisov wrote: On 09/14/2016

Re: [ceph-users] Jewel ceph-mon : high memory usage after few days

2016-09-15 Thread Wido den Hollander
> Op 15 september 2016 om 10:34 schreef Florent B : > > > Hi everyone, > > I have a Ceph cluster on Jewel. > > Monitors are on 32GB ram hosts. > > After a few days, ceph-mon process uses 25 to 35% of 32GB (8 to 11 GB) : > > 1150 ceph 20 0 15.454g 7.983g 7852 S

[ceph-users] CephFS: Upper limit for number of files in a directory?

2016-09-15 Thread Burkhard Linke
Hi, does CephFS impose an upper limit on the number of files in a directory? We currently have one directory with a large number of subdirectories: $ ls | wc -l 158141 Creating a new subdirectory fails: $ touch foo touch: cannot touch 'foo': No space left on device Creating files in a

Re: [ceph-users] Replacing a failed OSD

2016-09-15 Thread Dennis Kramer (DBS)
Hi Jim, I'm using a location script for OSDs, so when I add an OSD this script will determine its place in the cluster and in which bucket it belongs. In your ceph.conf there is a setting you can use: osd_crush_location_hook = With regards, On 09/14/2016 09:30 PM, Jim Kilborn wrote: > Reed, >

[ceph-users] Suiciding and corrupted OSDs zero out Ceph cluster IO

2016-09-15 Thread Kostis Fardelas
Hello cephers, last week we survived a 3-day outage on our ceph cluster (Hammer 0.94.7, 162 OSDs, 27 "fat" nodes, 1000s of clients) due to 6 out of 162 OSDs crash in the SAME node. The outage was caused in the following timeline: time 0: OSDs living in the same node (rd0-19) start heavily

Re: [ceph-users] Jewel ceph-mon : high memory usage after few days

2016-09-15 Thread Wido den Hollander
> Op 15 september 2016 om 10:40 schreef Florent B : > > > On 09/15/2016 10:37 AM, Wido den Hollander wrote: > >> Op 15 september 2016 om 10:34 schreef Florent B : > >> > >> > >> Hi everyone, > >> > >> I have a Ceph cluster on Jewel. > >> > >> Monitors

[ceph-users] rgw: Swift API X-Storage-Url

2016-09-15 Thread Василий Ангапов
Hello, I have Ceph Jewel 10.2.1 cluster and RadosGW. Issue is that when authenticating against Swift API I receive different values for X-Storage-Url header: # curl -i -H "X-Auth-User: internal-it:swift" -H "X-Auth-Key: ***" https://ed-1-vip.cloud/auth/v1.0 | grep X-Storage-Url X-Storage-Url:

Re: [ceph-users] Consistency problems when taking RBD snapshot

2016-09-15 Thread Nikolay Borisov
On 09/15/2016 01:24 PM, Ilya Dryomov wrote: > On Thu, Sep 15, 2016 at 10:22 AM, Nikolay Borisov > wrote: >> >> >> On 09/15/2016 09:22 AM, Nikolay Borisov wrote: >>> >>> >>> On 09/14/2016 05:53 PM, Ilya Dryomov wrote: On Wed, Sep 14, 2016 at 3:30 PM, Nikolay

Re: [ceph-users] OSDs thread leak during degraded cluster state

2016-09-15 Thread Wido den Hollander
> Op 15 september 2016 om 13:27 schreef Kostis Fardelas : > > > Hello cephers, > being in a degraded cluster state with 6/162 OSDs down ((Hammer > 0.94.7, 162 OSDs, 27 "fat" nodes, 1000s of clients) ) like the below > ceph cluster log indicates: > > 2016-09-12

Re: [ceph-users] Consistency problems when taking RBD snapshot

2016-09-15 Thread Ilya Dryomov
On Thu, Sep 15, 2016 at 12:54 PM, Nikolay Borisov wrote: > > > On 09/15/2016 01:24 PM, Ilya Dryomov wrote: >> On Thu, Sep 15, 2016 at 10:22 AM, Nikolay Borisov >> wrote: >>> >>> >>> On 09/15/2016 09:22 AM, Nikolay Borisov wrote: On

Re: [ceph-users] CephFS: Upper limit for number of files in a directory?

2016-09-15 Thread John Spray
On Thu, Sep 15, 2016 at 2:20 PM, Burkhard Linke wrote: > Hi, > > does CephFS impose an upper limit on the number of files in a directory? > > > We currently have one directory with a large number of subdirectories: > > $ ls | wc -l > 158141 > >

Re: [ceph-users] Replacing a failed OSD

2016-09-15 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Jim > Kilborn > Sent: 14 September 2016 20:30 > To: Reed Dier > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Replacing a failed OSD > > Reed, > > > >

Re: [ceph-users] CephFS: Upper limit for number of files in adirectory?

2016-09-15 Thread Burkhard Linke
Hi, On 09/15/2016 12:00 PM, John Spray wrote: On Thu, Sep 15, 2016 at 2:20 PM, Burkhard Linke wrote: Hi, does CephFS impose an upper limit on the number of files in a directory? We currently have one directory with a large number of