Re: [ceph-users] .New Ceph cluster - cannot add additional monitor

2015-06-17 Thread Mike Carlson
Just to follow up, I started from scratch, and I think the key was to run ceph-deploy purge (nodes) , ceph-deploy purgdata (nodes) and finally ceph-deploy forgetkeys Thanks for the replies Alex and Alex! Mike C ___ ceph-users mailing list

[ceph-users] Interesting postmortem on SSDs from Algolia

2015-06-17 Thread Steve Anthony
There's often a great deal of discussion about which SSDs to use for journals, and why some of the cheaper SSDs end up being more expensive in the long run. The recent blog post at Algoria, though not Ceph specific, provides a good illustration of exactly how insidious kernel/SSD interactions can

Re: [ceph-users] Very chatty MON logs: Is this normal?

2015-06-17 Thread Somnath Roy
However, I'd rather not set the level to 0/0, as that would disable all logging from the MONs I don't think so. All the error scenarios and stack trace (in case of crash) are supposed to be logged with log level 0. But, generally, we need the highest log level (say 20) to get all the

Re: [ceph-users] Ceph OSD with OCFS2

2015-06-17 Thread Somnath Roy
Sorry Prabu, I forgot to mention the bold settings in the conf file you need to tweak based on your HW configuration (cpu, disk etc.) and number of OSDs otherwise it may hit you back badly. Thanks Regards Somnath From: Somnath Roy Sent: Wednesday, June 17, 2015 11:25 AM To: 'gjprabu' Cc:

[ceph-users] best Linux distro for Ceph

2015-06-17 Thread Shane Gibson
Ok - I know this post has the potential to spread to unsavory corners of discussion about the best linux distro ... blah blah blah ... please, don't let it go there ... ! I'm seeking some input from people that have been running larger Ceph clusters ... on the order of 100s of physical

Re: [ceph-users] Ceph OSD with OCFS2

2015-06-17 Thread Somnath Roy
Okay..You didn’t mention anything about your rbd client host config and the cpu cores of OSD/rbd system..Some thoughts what you can do… 1. Considering pretty lean cpu config you have , I would say check for cpu usage of both OSD and rbd nodes first. If it is saturated already, you are out of

[ceph-users] Explanation for ceph osd set nodown and ceph osd cluster_snap

2015-06-17 Thread Jan Schermer
1) Flags available in ceph osd set are pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub|nodeep-scrub|notieragent I know or can guess most of them (the docs are a “bit” lacking) But with ceph osd set nodown” I have no idea what it should be used for - to keep hammering a faulty OSD?

Re: [ceph-users] Very chatty MON logs: Is this normal?

2015-06-17 Thread Daniel Schneller
On 2015-06-17 18:52:51 +, Somnath Roy said: This is presently written from log level 1 onwards :-) So, only log level 0 will not log this.. Try, 'debug_mon = 0/0' in the conf file.. Yeah, once I had sent the mail I realized that 1 in the log line was the level. Had overlooked that

[ceph-users] Expanding a ceph cluster with ansible

2015-06-17 Thread Stillwell, Bryan
I've been working on automating a lot of our ceph admin tasks lately and am pretty pleased with how the puppet-ceph module has worked for installing packages, managing ceph.conf, and creating the mon nodes. However, I don't like the idea of puppet managing the OSDs. Since we also use ansible in

[ceph-users] radosgw did not create auth url for swift

2015-06-17 Thread Vickie ch
Hi all, I want to use swift-client to connect ceph cluster. I have done s3 test on this cluster before. So I follow the guide to create a subuser and use swift client to test it. But always got an error 404 Not Found How can I create the auth page?Any help will be appreciated. - 1

Re: [ceph-users] Hardware cache settings recomendation

2015-06-17 Thread Mateusz Skała
Thanks for answer, I made some test, first leave dwc=enabled and caching on journal drive disabled. Latency grows from 20ms to 90ms on this drive. Next I enabled cache on journal drive and disabled all cache on data drives. Latency on data drives grows from 30 – 50ms to 1500 – 2000ms. Test

[ceph-users] osd_scrub_chunk_min/max scrub_sleep?

2015-06-17 Thread Tu Holmes
Hey gang, Some options are just not documented well… What’s up with: osd_scrub_chunk_min osd_scrub_chunk_max osd_scrub_sleep? === Tu Holmes tu.hol...@gmail.com ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] Rename pool by id

2015-06-17 Thread Georgios Dimitrakakis
Pavel, unfortunately there isn't a way to rename a pool usign its ID as I have learned myself the hard way since I 've faced a few months ago the exact same issue. It would be a good idea for developers to also include a way to manipulate (rename, delete, etc.) pools using the ID which is

Re: [ceph-users] xattrs vs. omap with radosgw

2015-06-17 Thread Nathan Cutler
We've since merged something that stripes over several small xattrs so that we can keep things inline, but it hasn't been backported to hammer yet. See c6cdb4081e366f471b372102905a1192910ab2da. Hi Sage: You wrote yet - should we earmark it for hammer backport? Nathan

Re: [ceph-users] Hardware cache settings recomendation

2015-06-17 Thread Mateusz Skała
Yes, all disk are in single drive raid 0. Now cache is enabled for all drives, should I disable cache for SSD drives? Regards, Mateusz From: Tyler Bishop [mailto:tyler.bis...@beyondhosting.net] Sent: Thursday, June 11, 2015 7:30 PM To: Mateusz Skała Cc: ceph-users@lists.ceph.com Subject:

Re: [ceph-users] v0.94.2 Hammer released

2015-06-17 Thread Dan van der Ster
On Thu, Jun 11, 2015 at 7:34 PM, Sage Weil sw...@redhat.com wrote: * ceph-objectstore-tool should be in the ceph server package (#11376, Ken Dreyer) We had a little trouble yum updating from 0.94.1 to 0.94.2: file /usr/bin/ceph-objectstore-tool from install of ceph-1:0.94.2-0.el6.x86_64

Re: [ceph-users] Hardware cache settings recomendation

2015-06-17 Thread Jan Schermer
Cache on top of the data drives (not journal) will not help in most cases, those writes are already buffered in the OS - so unless your OS is very light on memory and flushing constantly it will have no effect, it just adds overhead in case a flush comes. I haven’t tested this extensively with

Re: [ceph-users] CephFS: 'ls -alR' performance terrible unless Linux cache flushed

2015-06-17 Thread negillen negillen
I have done some quick tests with FUSE too: it seems to me that, both with the old and with the new kernel, FUSE is approx. five times slower than kernel driver for both reading files and getting stats. I don't know whether it is just me or if it is expected. On Wed, Jun 17, 2015 at 2:56 AM,

Re: [ceph-users] rbd_cache, limiting read on high iops around 40k

2015-06-17 Thread Irek Fasikhov
If necessary, there are RPM files for centos 7: ​ gperftools.spec https://drive.google.com/file/d/0BxoNLVWxzOJWaVVmWTA3Z18zbUE/edit?usp=drive_web ​​ pprof-2.4-1.el7.centos.noarch.rpm https://drive.google.com/file/d/0BxoNLVWxzOJWRmQ2ZEt6a1pnSVk/edit?usp=drive_web ​​

[ceph-users] 10d

2015-06-17 Thread Dan van der Ster
Hi, After upgrading to 0.94.2 yesterday on our test cluster, we've had 3 PGs go inconsistent. First, immediately after we updated the OSDs PG 34.10d went inconsistent: 2015-06-16 13:42:19.086170 osd.52 137.138.39.211:6806/926964 2 : cluster [ERR] 34.10d scrub stat mismatch, got 4/5 objects, 0/0

Re: [ceph-users] SSD LifeTime for Monitors

2015-06-17 Thread Gregory Farnum
On Wed, Jun 17, 2015 at 10:18 AM, Stefan Priebe - Profihost AG s.pri...@profihost.ag wrote: Hi, Does anybody know how many data gets written from the monitors? I was using some cheaper ssds for monitors and was wondering why they had already written 80 TB after 8 month. 3.8MB/s? That's a

Re: [ceph-users] 10d

2015-06-17 Thread Gregory Farnum
On Wed, Jun 17, 2015 at 8:56 AM, Dan van der Ster d...@vanderster.com wrote: Hi, After upgrading to 0.94.2 yesterday on our test cluster, we've had 3 PGs go inconsistent. First, immediately after we updated the OSDs PG 34.10d went inconsistent: 2015-06-16 13:42:19.086170 osd.52

Re: [ceph-users] xattrs vs. omap with radosgw

2015-06-17 Thread Abhishek L
On Wed, Jun 17, 2015 at 1:02 PM, Nathan Cutler ncut...@suse.cz wrote: We've since merged something that stripes over several small xattrs so that we can keep things inline, but it hasn't been backported to hammer yet. See c6cdb4081e366f471b372102905a1192910ab2da. Hi Sage: You wrote yet -

[ceph-users] SSD LifeTime for Monitors

2015-06-17 Thread Stefan Priebe - Profihost AG
Hi, Does anybody know how many data gets written from the monitors? I was using some cheaper ssds for monitors and was wondering why they had already written 80 TB after 8 month. Stefan ___ ceph-users mailing list ceph-users@lists.ceph.com

[ceph-users] rbd performance issue - can't find bottleneck

2015-06-17 Thread Jacek Jarosiewicz
Hi, We've been doing some testing of ceph hammer (0.94.2), but the performance is very slow and we can't find what's causing the problem. Initially we've started with four nodes with 10 osd's total. The drives we've used were SATA enterprise drives and on top of that we've used SSD drives as

Re: [ceph-users] 10d

2015-06-17 Thread Dan van der Ster
On Wed, Jun 17, 2015 at 10:52 AM, Gregory Farnum g...@gregs42.com wrote: On Wed, Jun 17, 2015 at 8:56 AM, Dan van der Ster d...@vanderster.com wrote: Hi, After upgrading to 0.94.2 yesterday on our test cluster, we've had 3 PGs go inconsistent. First, immediately after we updated the OSDs PG

Re: [ceph-users] rbd performance issue - can't find bottleneck

2015-06-17 Thread Mark Nelson
On 06/17/2015 04:10 AM, Jacek Jarosiewicz wrote: Hi, We've been doing some testing of ceph hammer (0.94.2), but the performance is very slow and we can't find what's causing the problem. Initially we've started with four nodes with 10 osd's total. The drives we've used were SATA enterprise

Re: [ceph-users] rbd performance issue - can't find bottleneck

2015-06-17 Thread Alexandre DERUMIER
Hi, can you post your ceph.conf ? Which tools do you use for benchmark ? which block size, iodepth, number of client/rbd volume do you use ? Is it with krbd kernel driver ? (I have seen some bad performance with kernel 3.16, but at much higher rate (100k iops) Is it with ethernet switches ?

Re: [ceph-users] rbd performance issue - can't find bottleneck

2015-06-17 Thread Jacek Jarosiewicz
On 06/17/2015 03:38 PM, Alexandre DERUMIER wrote: Hi, can you post your ceph.conf ? sure: [global] fsid = e96fdc70-4f9c-4c12-aae8-63dd7c64c876 mon initial members = cf01,cf02 mon host = 10.4.10.211,10.4.10.212 auth cluster required = cephx auth service required = cephx auth client required

Re: [ceph-users] rbd performance issue - can't find bottleneck

2015-06-17 Thread Christian Balzer
On Wed, 17 Jun 2015 16:03:17 +0200 Jacek Jarosiewicz wrote: On 06/17/2015 03:34 PM, Mark Nelson wrote: On 06/17/2015 04:10 AM, Jacek Jarosiewicz wrote: Hi, [ cut ] ~60MB/s seq writes ~100MB/s seq reads ~2-3k iops random reads Is this per SSD or aggregate? aggregate (if

[ceph-users] Accessing Ceph from Spark

2015-06-17 Thread Milan Sladky
Is it possible to access Ceph from Spark as it is mentioned here for Openstack Swift? https://spark.apache.org/docs/latest/storage-openstack-swift.html Thanks for help. Milan Sladky ___ ceph-users mailing list

Re: [ceph-users] rbd performance issue - can't find bottleneck

2015-06-17 Thread Jacek Jarosiewicz
On 06/17/2015 03:34 PM, Mark Nelson wrote: On 06/17/2015 04:10 AM, Jacek Jarosiewicz wrote: Hi, [ cut ] ~60MB/s seq writes ~100MB/s seq reads ~2-3k iops random reads Is this per SSD or aggregate? aggregate (if I understand You correctly). This is what I see when I run tests on client

Re: [ceph-users] xattrs vs. omap with radosgw

2015-06-17 Thread Sage Weil
On Wed, 17 Jun 2015, Nathan Cutler wrote: We've since merged something that stripes over several small xattrs so that we can keep things inline, but it hasn't been backported to hammer yet. See c6cdb4081e366f471b372102905a1192910ab2da. Hi Sage: You wrote yet - should we earmark it

Re: [ceph-users] Accessing Ceph from Spark

2015-06-17 Thread Gregory Farnum
On Wed, Jun 17, 2015 at 2:58 PM, Milan Sladky milan.sla...@outlook.com wrote: Is it possible to access Ceph from Spark as it is mentioned here for Openstack Swift? https://spark.apache.org/docs/latest/storage-openstack-swift.html Depends on what you're trying to do. It's possible that the

Re: [ceph-users] rbd performance issue - can't find bottleneck

2015-06-17 Thread Mark Nelson
On 06/17/2015 09:03 AM, Jacek Jarosiewicz wrote: On 06/17/2015 03:34 PM, Mark Nelson wrote: On 06/17/2015 04:10 AM, Jacek Jarosiewicz wrote: Hi, [ cut ] ~60MB/s seq writes ~100MB/s seq reads ~2-3k iops random reads Is this per SSD or aggregate? aggregate (if I understand You

Re: [ceph-users] Erasure Coded Pools and PGs

2015-06-17 Thread Loic Dachary
Hi, On 17/06/2015 18:04, Garg, Pankaj wrote: Hi, I have 5 OSD servers, with total of 45 OSDS in my clusters. I am trying out Erasure Coding with different K and m values. I seem to always get Warnings about : Degraded and Undersized PGs, whenever I create a profile and create a

Re: [ceph-users] Accessing Ceph from Spark

2015-06-17 Thread ZHOU Yuan
Hi Milan, We've done some tests here and our hadoop can talk to RGW successfully with this SwiftFS plugin. But we haven't tried Spark yet. One thing is the data locality feature, it actually requires some special configuration of Swift proxy-server, so RGW is not able to archive the data locality

[ceph-users] Erasure Coded Pools and PGs

2015-06-17 Thread Garg, Pankaj
Hi, I have 5 OSD servers, with total of 45 OSDS in my clusters. I am trying out Erasure Coding with different K and m values. I seem to always get Warnings about : Degraded and Undersized PGs, whenever I create a profile and create a Pool based on that profile. I have profiles with K and M