Re: [ceph-users] using jemalloc in trusty

2016-05-23 Thread Luis Periquito
Thanks Somnath, I expected that much. But given the hint in the config files do you know if they are built to use jemalloc? it seems not... On Mon, May 23, 2016 at 3:34 PM, Somnath Roy wrote: > You need to build ceph code base to use jemalloc for OSDs..LD_PRELOAD won't

Re: [ceph-users] using jemalloc in trusty

2016-05-23 Thread Somnath Roy
Yes, if you are using do_autogen , use -J option. If you are using config files directly , use --with-jemalloc -Original Message- From: Luis Periquito [mailto:periqu...@gmail.com] Sent: Monday, May 23, 2016 7:44 AM To: Somnath Roy Cc: Ceph Users Subject: Re: [ceph-users] using jemalloc

Re: [ceph-users] mark out vs crush weight 0

2016-05-23 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 Check out the Weighted Priority Queue option in Jewel, this really helped reduce the impact of recovery and backfill on client traffic with my testing. I think it really addresses a lot of the pain points you mention. -BEGIN PGP SIGNATURE-

[ceph-users] using jemalloc in trusty

2016-05-23 Thread Luis Periquito
I've been running some tests with jewel, and wanted to enable jemalloc. I noticed that the new jewel release now loads properly /etc/default/ceph and has an option to use jemalloc. I've installed jemalloc, enabled the LD_PRELOAD option, however doing some tests it seems that it's still using

Re: [ceph-users] using jemalloc in trusty

2016-05-23 Thread Somnath Roy
You need to build ceph code base to use jemalloc for OSDs..LD_PRELOAD won't work.. Thanks & regards Somnath -Original Message- From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Luis Periquito Sent: Monday, May 23, 2016 7:30 AM To: Ceph Users Subject: [ceph-users]

Re: [ceph-users] radosgw hammer -> jewel upgrade (default zone & region config)

2016-05-23 Thread Ben Hines
I for one am terrified of upgrading due to these messages (and indications that the problem still may not be resolved even in 10.2.1) - holding off until a clean upgrade is possible without running any hacky scripts. -Ben On Mon, May 23, 2016 at 2:23 AM, nick wrote: > Hi, > we

[ceph-users] New user questions with radosgw with Jewel 10.2.1

2016-05-23 Thread Sylvain, Eric
Hello, I'm a fairly new user and I am trying to bring up radosgw. I am following this page: http://docs.ceph.com/docs/master/install/install-ceph-gateway/ I have Jewel 10.2.1 installed with a co-located admin/mon host and a separate osd host First a question: Can I run radosgw on a

Re: [ceph-users] Jewel ubuntu release is half cooked

2016-05-23 Thread Anthony D'Atri
Re: > 2. Inefficient chown documentation - The documentation states that one should > "chown -R ceph:ceph /var/lib/ceph" if one is looking to have ceph-osd ran as > user ceph and not as root. Now, this command would run a chown process one > osd at a time. I am considering my cluster to be a

Re: [ceph-users] Public and Private network over 1 interface

2016-05-23 Thread Somnath Roy
4MB block size EC-object use case scenario (Mostly for reads , not so much for writes) we saw some benefit separating public/cluster network for 40GbE. We didn’t use two NIC though. We configured two ports on a NIC. Both network can give up to 48Gb/s but with Mellanox card/Mellanox switch

[ceph-users] Public and Private network over 1 interface

2016-05-23 Thread Brady Deetz
TLDR; Has anybody deployed a Ceph cluster using a single 40 gig nic? This is discouraged in http://docs.ceph.com/docs/master/rados/configuration/network-config-ref/ "One NIC OSD in a Two Network Cluster: Generally, we do not recommend deploying an OSD host with a single NIC in a cluster with two

Re: [ceph-users] RadosGW performance s3 many objects

2016-05-23 Thread Wade Holler
We (my customer ) are trying to test at Jewell now but I can say that the above behavior was also observed by my customer at Infernalis. After 300 million or so objects in a single bucket the cluster basically fell down as described above. Few hundred osds in this cluster. We are concerned that

Re: [ceph-users] civetweb vs Apache for rgw

2016-05-23 Thread fridifree
What apache gives that civetweb not? Thank you On May 23, 2016 11:49 AM, "Anand Bhat" wrote: > For performance, civetweb is better as fastcgi module associated with > apache is single threaded. But Apache does have fancy features which > civetweb lacks. If you are looking

Re: [ceph-users] Public and Private network over 1 interface

2016-05-23 Thread Wido den Hollander
> Op 23 mei 2016 om 21:53 schreef Brady Deetz : > > > TLDR; > Has anybody deployed a Ceph cluster using a single 40 gig nic? This is > discouraged in > http://docs.ceph.com/docs/master/rados/configuration/network-config-ref/ > > "One NIC OSD in a Two Network Cluster: >

[ceph-users] Falls cluster then one node switch off

2016-05-23 Thread Никитенко Виталий
Hello! I have a cluster of 2 nodes with 3 OSD each. The cluster full about 80%. df -H /dev/sdc127G 24G 3.9G 86% /var/lib/ceph/osd/ceph-1 /dev/sdd127G 20G 6.9G 75% /var/lib/ceph/osd/ceph-2 /dev/sdb127G 24G 3.5G 88% /var/lib/ceph/osd/ceph-0 When I switch off one

Re: [ceph-users] Diagnosing slow requests

2016-05-23 Thread Christian Balzer
Hello, On Mon, 23 May 2016 10:45:41 +0200 Peter Kerdisle wrote: > Hey, > > Sadly I'm still battling this issue. I did notice one interesting thing. > > I changed the cache settings for my cache tier to add redundancy to the > pool which means a lot of recover activity on the cache. During all

Re: [ceph-users] Public and Private network over 1 interface

2016-05-23 Thread Christian Balzer
Hello, separate public/private networks make sense for clusters that: a) have vastly more storage bandwidth than a single link can handle or b) are extremely read heavy on the client side, so replication reads can be separated from the client ones. I posit that neither is the case in your

Re: [ceph-users] NVRAM cards as OSD journals

2016-05-23 Thread Christian Balzer
Hello, On Fri, 20 May 2016 15:52:45 + EP Komarla wrote: > Hi, > > I am contemplating using a NVRAM card for OSD journals in place of SSD > drives in our ceph cluster. > > Configuration: > > * 4 Ceph servers > > * Each server has 24 OSDs (each OSD is a 1TB SAS drive) > >

Re: [ceph-users] dense storage nodes

2016-05-23 Thread Christian Balzer
Hello, On Fri, 20 May 2016 10:57:10 -0700 Anthony D'Atri wrote: > [ too much to quote ] > > Dense nodes often work better for object-focused workloads than > block-focused, the impact of delayed operations is simply speed vs. a > tenant VM crashing. > Especially if they're don't have SSD

Re: [ceph-users] Public and Private network over 1 interface

2016-05-23 Thread Brady Deetz
To be clear for future responders, separate mds and mon servers are in the design. Everything is the same as the osd hardware except the chassis and there aren't 24 hdds in there. On May 23, 2016 4:27 PM, "Oliver Dzombic" wrote: > Hi, > > keep it simple, would be in my

Re: [ceph-users] RBD removal issue

2016-05-23 Thread Adrian Saul
Thanks - all sorted. > -Original Message- > From: Nick Fisk [mailto:n...@fisk.me.uk] > Sent: Monday, 23 May 2016 6:58 PM > To: Adrian Saul; ceph-users@lists.ceph.com > Subject: RE: RBD removal issue > > See here: > > http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image > >

Re: [ceph-users] Public and Private network over 1 interface

2016-05-23 Thread Oliver Dzombic
Hi, keep it simple, would be in my opinion devide different tasks to different servers and networks. The more stuff is running on one device, the higher is the chance that they will influence each other and this way make debugging harder. Our first setup's had all ( mon, osd, mds ) on one

Re: [ceph-users] RadosGW performance s3 many objects

2016-05-23 Thread Vickey Singh
Hello Guys Is several millions of object with Ceph ( for RGW use case ) still an issue ? Or is it fixed ? Thnx Vickey On Thu, Jan 28, 2016 at 12:55 AM, Krzysztof Księżyk wrote: > Stefan Rogge writes: > > > > > > > Hi, > > we are using the Ceph with

[ceph-users] CephFS quotas in kernel client

2016-05-23 Thread Mykola Dvornik
Any plans to support quotas in CephFS kernel client? -Mykola___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] RBD removal issue

2016-05-23 Thread Nick Fisk
See here: http://cephnotes.ksperis.com/blog/2014/07/04/remove-big-rbd-image > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Adrian Saul > Sent: 23 May 2016 09:37 > To: 'ceph-users@lists.ceph.com' > Subject:

Re: [ceph-users] radosgw hammer -> jewel upgrade (default zone & region config)

2016-05-23 Thread nick
Hi, we ran into the same rgw problem when updating from infernalis to jewel (version 10.2.1). Now I would like to run the script from Yehuda, but I am a bit scared by >I can create and get new buckets and objects but I've "lost" all my > old buckets. As I understood so far we do not need the

[ceph-users] RBD removal issue

2016-05-23 Thread Adrian Saul
A while back I attempted to create an RBD volume manually - intending it to be an exact size of another LUN around 100G. The command line instead took this to be the default MB argument for size and so I ended up with a 102400 TB volume. Deletion was painfully slow (I never used the volume,

Re: [ceph-users] Diagnosing slow requests

2016-05-23 Thread Peter Kerdisle
Hey, Sadly I'm still battling this issue. I did notice one interesting thing. I changed the cache settings for my cache tier to add redundancy to the pool which means a lot of recover activity on the cache. During all this there were absolutely no slow requests reported. Is there anything I can

Re: [ceph-users] ceph -s output

2016-05-23 Thread Anand Bhat
Check states of PGs using "ceph pg dump" and for every PG that is not "active+clean", issue "ceph pg map " and get mapping OSDs. Check the state of those OSDs by looking at their logs under /var/log/ceph/. Regards, Anand On Mon, May 23, 2016 at 6:53 AM, Ken Peng wrote: > Hi,

Re: [ceph-users] civetweb vs Apache for rgw

2016-05-23 Thread Anand Bhat
For performance, civetweb is better as fastcgi module associated with apache is single threaded. But Apache does have fancy features which civetweb lacks. If you are looking for just the performance, then go for civetweb. Regards, Anand On Mon, May 23, 2016 at 12:43 PM, fridifree

Re: [ceph-users] NVRAM cards as OSD journals

2016-05-23 Thread Nick Fisk
> -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Heath Albritton > Sent: 23 May 2016 01:24 > To: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] NVRAM cards as OSD journals > > I'm contemplating the same thing as well. Or rather,

[ceph-users] Ceph Status - Segmentation Fault

2016-05-23 Thread Mathias Buresch
Hi there, I was updating Ceph to 0.94.7 and now I am getting segmantation faults. When getting status via "ceph -s" or "ceph health detail" I am getting an error "Segmentation fault". I have only two Monitor Deamon.. but didn't had any problems yet with that.. maybe they maintenance time was too

Re: [ceph-users] Jewel CephFS quota (setfattr, getfattr)

2016-05-23 Thread Goncalo Borges
Hi Edgaras... Just quoting a previous statement frim Yang: " To use ACL, you need to add "--fuse_default_permissions=0 --client_acl_type=posix_acl" options to ceph-fuse. The '--fuse_default_permissions=0' option disables kernel file permission check and let ceph-fuse do the check." Cheers

[ceph-users] civetweb vs Apache for rgw

2016-05-23 Thread fridifree
Hi everyone, What would give the best performance in most of the cases, civetweb or apache? Thank you ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[ceph-users] Removing objects and bucket after issues with placement group

2016-05-23 Thread Romero Junior
Hi group, A couple of weeks ago I ran into a couple of issues with an incomplete placement group. Luckly the data inside that PG was not relevant at all, so we decided to simply remove it completely (using the flag osd_find_best_info_ignore_history_les). However, now I cannot remove the

[ceph-users] openSuse Leap 42.1, slow krbd, max_sectors_kb = 127

2016-05-23 Thread David
Hi All I'm doing some testing with OpenSUSE Leap 42.1, it ships with kernel 4.1.12 but I've also tested with 4.1.24 When I map an image with the kernel RBD client, max_sectors_kb = 127. I'm unable to increase: # echo 4096 > /sys/block/rbd0/queue/max_sectors_kb -bash: echo: write error: Invalid

Re: [ceph-users] Jewel CephFS quota (setfattr, getfattr)

2016-05-23 Thread Yan, Zheng
To enable quota, you need to pass "--client-quota" option to ceph-fuse Yan, Zheng On Mon, May 23, 2016 at 3:18 PM, Goncalo Borges wrote: > Hi Edgaras... > > > Just quoting a previous statement frim Yang: > > " To use ACL, you need to add

Re: [ceph-users] CephFS quotas in kernel client

2016-05-23 Thread Yan, Zheng
No plan so far. Current quota design requires client to do bottom-to-top path walk, which is unfriendly for kernel client (due to lock design of kernel). On Mon, May 23, 2016 at 4:55 PM, Mykola Dvornik wrote: > Any plans to support quotas in CephFS kernel client? > >

Re: [ceph-users] CephFS quotas in kernel client

2016-05-23 Thread Edgaras Lukoševičius
Parallel VFS lookups should improve this: https://lwn.net/Articles/685108/ On 23/05/16 15:08, Yan, Zheng wrote: No plan so far. Current quota design requires client to do bottom-to-top path walk, which is unfriendly for kernel client (due to lock design of kernel). On Mon, May 23, 2016 at

Re: [ceph-users] CephFS quotas in kernel client

2016-05-23 Thread Mykola Dvornik
Thanks for a quick reply. On Mon, 2016-05-23 at 20:08 +0800, Yan, Zheng wrote: > No plan so far.  Current quota design requires client to do > bottom-to-top path walk, which is unfriendly for kernel client (due > to > lock design of kernel). > > On Mon, May 23, 2016 at 4:55 PM, Mykola Dvornik >

[ceph-users] rbd mapping error on ubuntu 16.04

2016-05-23 Thread Albert Archer
Hello All, There is a problem to mapping RBD images to ubuntu 16.04(Kernel 4.4.0-22-generic). All of the ceph solution is based on ubunut 16.04(deploy, monitors, OSDs and Clients). # # there are some output of my config : *$ ceph

Re: [ceph-users] Jewel ubuntu release is half cooked

2016-05-23 Thread Max A. Krasilnikov
Hello! On Mon, May 23, 2016 at 11:26:38AM +0100, andrei wrote: > 1. Ceph journals - After performing the upgrade the ceph-osd processes are > not starting. I've followed the instructions and chowned /var/lib/ceph (also > see point 2 below). The issue relates to the journal partitions, which

Re: [ceph-users] Recommended OSD size

2016-05-23 Thread gjprabu
Hi Christian, Please share your suggestion. Regards Prabu GJ On Sat, 21 May 2016 17:33:33 +0530 gjprabu gjpr...@zohocorp.comwrote HI Christian, Typo in my previous mail. Thanks for your reply, It will be very helpful if we get the details on osd per

[ceph-users] Jewel ubuntu release is half cooked

2016-05-23 Thread Andrei Mikhailovsky
Hello I've recently updated my Hammer ceph cluster running on Ubuntu 14.04 LTS servers and noticed a few issues during the upgrade. Just wanted to share my experience. I've installed the latest Jewel release. In my opinion, some of the issues I came across relate to poor upgrade

Re: [ceph-users] rbd mapping error on ubuntu 16.04

2016-05-23 Thread Ilya Dryomov
On Mon, May 23, 2016 at 12:27 PM, Albert Archer wrote: > Hello All, > There is a problem to mapping RBD images to ubuntu 16.04(Kernel > 4.4.0-22-generic). > All of the ceph solution is based on ubunut 16.04(deploy, monitors, OSDs and > Clients). >

Re: [ceph-users] rbd mapping error on ubuntu 16.04

2016-05-23 Thread Ilya Dryomov
On Mon, May 23, 2016 at 2:59 PM, Albert Archer wrote: > > Thanks. > but how to use these features ??? > so,there is no way to implement them on ubuntu 16.04 kernel (4.4.0) ??? > it's strange !!! What is your use case? If you are using the kernel client, create your

Re: [ceph-users] openSuse Leap 42.1, slow krbd, max_sectors_kb = 127

2016-05-23 Thread Ilya Dryomov
On Mon, May 23, 2016 at 10:12 AM, David wrote: > Hi All > > I'm doing some testing with OpenSUSE Leap 42.1, it ships with kernel 4.1.12 > but I've also tested with 4.1.24 > > When I map an image with the kernel RBD client, max_sectors_kb = 127. I'm > unable to increase: >

Re: [ceph-users] rbd mapping error on ubuntu 16.04

2016-05-23 Thread Albert Archer
Thanks. but how to use these features ??? so,there is no way to implement them on ubuntu 16.04 kernel (4.4.0) ??? it's strange !!! 樂 On Mon, May 23, 2016 at 5:28 PM, Albert Archer wrote: > Thanks. > but how to use these features ??? > so,there is no way to implement

Re: [ceph-users] Falls cluster then one node switch off

2016-05-23 Thread Christian Balzer
Hello, On Tue, 24 May 2016 10:28:02 +0700 Никитенко Виталий wrote: > Hello! > I have a cluster of 2 nodes with 3 OSD each. The cluster full about 80%. > According to your CRUSH map that's not quite true, namely ceph1-node2 entry. And while that again according to your CRUSH map isn't in the