Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-07-27 Thread Wido den Hollander
On 27-07-15 14:21, Jan Schermer wrote: Hi! The /cgroup/* mount point is probably a RHEL6 thing, recent distributions seem to use /sys/fs/cgroup like in your case (maybe because of systemd?). On RHEL 6 the mount points are configured in /etc/cgconfig.conf and /cgroup is the default. I

Re: [ceph-users] osd daemons stuck in D state

2015-07-27 Thread Jan Schermer
When those processes become blocked are the drives busy or idle? Can you post the output from ps -awexo pid,tt,user,fname,tmout,f,wchan” on those processes when that happens? My guess would be they really are waiting for the disk array for some reason - can you check if you can read/write to

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-07-27 Thread Dan van der Ster
On Mon, Jul 27, 2015 at 2:51 PM, Wido den Hollander w...@42on.com wrote: I'm testing with it on 48-core, 256GB machines with 90 OSDs each. This is a +/- 20PB Ceph cluster and I'm trying to see how much we would benefit from it. Cool. How many OSDs total? Cheers, Dan

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-07-27 Thread Saverio Proto
Hello Jan, I am testing your scripts, because we want also to test OSDs and VMs on the same server. I am new to cgroups, so this might be a very newbie question. In your script you always reference to the file /cgroup/cpuset/libvirt/cpuset.cpus but I have the file in

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-07-27 Thread Wido den Hollander
On 27-07-15 14:56, Dan van der Ster wrote: On Mon, Jul 27, 2015 at 2:51 PM, Wido den Hollander w...@42on.com wrote: I'm testing with it on 48-core, 256GB machines with 90 OSDs each. This is a +/- 20PB Ceph cluster and I'm trying to see how much we would benefit from it. Cool. How many

Re: [ceph-users] Enclosure power failure pausing client IO till all connected hosts up

2015-07-27 Thread Max A. Krasilnikov
Здравствуйте! On Tue, Jul 07, 2015 at 02:21:56PM +0530, mallikarjuna.biradar wrote: Hi all, Setup details: Two storage enclosures each connected to 4 OSD nodes (Shared storage). Failure domain is Chassis (enclosure) level. Replication count is 2. Each host has allotted with 4 drives. I

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-07-27 Thread Jan Schermer
Hi! The /cgroup/* mount point is probably a RHEL6 thing, recent distributions seem to use /sys/fs/cgroup like in your case (maybe because of systemd?). On RHEL 6 the mount points are configured in /etc/cgconfig.conf and /cgroup is the default. I also saw the pull request from you on github and

[ceph-users] osd daemons stuck in D state

2015-07-27 Thread Simion Rad
Hello all, When I try to add more than one osds to a host and the backfilling process starts , all the osd daemons except one of them become stuck in D state. When this happends they are shown as out and down (when running ceph osd tree). The only way I can kill the processes is to remove the

Re: [ceph-users] How to use cgroup to bind ceph-osd to a specific cpu core?

2015-07-27 Thread Wido den Hollander
On 27-07-15 15:28, Jan Schermer wrote: Cool! Any immediate effect you noticed? Did you partition it into 2 cpusets corresponding to NUMA nodes or more? Not yet. Cluster is still in build state. Will run benchmarks with and without pinning set. Currently the setup is to 2 cpusets with 2

[ceph-users] State of nfs-ganesha CEPH fsal

2015-07-27 Thread Burkhard Linke
Hi, the nfs-ganesha documentation states: ... This FSAL links to a modified version of the CEPH library that has been extended to expose its distributed cluster and replication facilities to the pNFS operations in the FSAL. ... The CEPH library modifications have not been merged into the

Re: [ceph-users] State of nfs-ganesha CEPH fsal

2015-07-27 Thread Gregory Farnum
On Mon, Jul 27, 2015 at 4:33 PM, Burkhard Linke burkhard.li...@computational.bio.uni-giessen.de wrote: Hi, the nfs-ganesha documentation states: ... This FSAL links to a modified version of the CEPH library that has been extended to expose its distributed cluster and replication facilities

Re: [ceph-users] Weird behaviour of cephfs with samba

2015-07-27 Thread Eric Eastman
I don't have any answers but I am also seeing some strange results exporting a Ceph file system using the Samba VFS interface on Ceph version 9.0.2. If I mount a Linux client with vers=1, I see the file system the same as I see it on a ceph file system mount. If I use vers=2.0 or vers=3.0 on the

Re: [ceph-users] Weird behaviour of cephfs with samba

2015-07-27 Thread Gregory Farnum
What's the full stack you're using to run this with? If you're using the kernel client, try updating it or switching to the userspace (ceph-fuse, or Samba built-in) client. If using userspace, please make sure you've got the latest one. -Greg On Mon, Jul 27, 2015 at 3:16 PM, Jörg Henne

[ceph-users] Performance Issues

2015-07-27 Thread Ray Sun
Dear Cephers, I did a simple test to understand the performance loss of ceph. Here's my environment: CPU: 2 * Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz Memory: 4 * 8G 1067 MHz NIC: 2 * Intel Corporation 10-Gigabit X540-AT2 HDD: 1 * WDC WD1003FZEX ATA Disk 1TB 4 * Seagate ST2000NM0011 ATA Disk 2TB

[ceph-users] Weird behaviour of cephfs with samba

2015-07-27 Thread Jörg Henne
Hi all, the faq at http://ceph.com/docs/cuttlefish/faq/ mentions the possibility to run export a mounted cephfs via samba. This combination exhibits a very weird behaviour, though. We have a directory on cephfs with many small xml snippets. If I repeadtedly ls the directory on Unix, I get the

[ceph-users] Algorithm for default pg_count calculation

2015-07-27 Thread Konstantin Danilov
Hi all, I'm working on algorithm to estimate PG count for set of pools with minimal input from user. The main target is openstack deployments. I know about ceph.com/pgcalc/, but would like to write down a rules and get a python code. Can you comment following, please? Input: * pg_count has no

Re: [ceph-users] osd daemons stuck in D state

2015-07-27 Thread Simion Rad
The server has 128 GB RAM (it also runs KVM virtual machines and they use ~ 95 GB ). The HBA is LSI Logic SAS1068 PCI-X Fusion-MPT SAS (kernel module is mptsas version 3.04.20) I have two HBAs ,but I didn't want to use multipath, so there is only one path / LUN (the array's controllers cannot

Re: [ceph-users] Algorithm for default pg_count calculation

2015-07-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 Some more info: Small PG count: * More varied data distribution in cluster Large PG count: * More even data distribution in cluster * A very high number of PG can starve CPU/RAM causing performance to decrease We are targeting 50 PGs per OSD to

Re: [ceph-users] Trying to remove osd

2015-07-27 Thread Paul Schaleger
Thanks, stopping the osd daemon seemed to do the trick. -Original Message- From: Robert LeBlanc [mailto:rob...@leblancnet.us] Sent: Monday, July 27, 2015 11:48 AM To: Paul Schaleger Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Trying to remove osd -BEGIN PGP SIGNED

Re: [ceph-users] Weird behaviour of cephfs with samba

2015-07-27 Thread Jörg Henne
Gregory Farnum greg@... writes: Yeah, I think there were some directory listing bugs in that version that Samba is probably running into. They're fixed in a newer kernel release (I'm not sure which one exactly, sorry). Ok, thanks, good to know! and then detaches itself but the mountpoint

Re: [ceph-users] Weird behaviour of cephfs with samba

2015-07-27 Thread Gregory Farnum
On Mon, Jul 27, 2015 at 5:46 PM, Jörg Henne henn...@gmail.com wrote: Gregory Farnum greg@... writes: What's the full stack you're using to run this with? If you're using the kernel client, try updating it or switching to the userspace (ceph-fuse, or Samba built-in) client. If using userspace,

Re: [ceph-users] hadoop on ceph

2015-07-27 Thread Patrick McGarry
Moving this to the ceph-user list where it has a better chance of being answered. On Mon, Jul 27, 2015 at 5:35 AM, jingxia@baifendian.com jingxia@baifendian.com wrote: Dear , I have questions to ask. The doc says hadoop on ceph but requires Hadoop 1.1.X stable series I want to know

[ceph-users] Trying to remove osd

2015-07-27 Thread Paul Schaleger
I added an OSD to a device that I did not really want to and now I am unable to remove it. Any suggestions as to what I am missing? Thanks in advance [rdo@n001 c140-ceph]$ ceph osd out 21 osd.21 is already out. [rdo@n001 c140-ceph]$ ceph osd down 21 marked down osd.21. [rdo@n001 c140-ceph]$

Re: [ceph-users] Trying to remove osd

2015-07-27 Thread Robert LeBlanc
-BEGIN PGP SIGNED MESSAGE- Hash: SHA256 Did you kill the OSD process, you are still showing 28 OSD up. I'm not sure that should stop you from removing it though. You can also try ceph osd crush rm osd.21 - Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E

Re: [ceph-users] Weird behaviour of cephfs with samba

2015-07-27 Thread Jörg Henne
Gregory Farnum greg@... writes: What's the full stack you're using to run this with? If you're using the kernel client, try updating it or switching to the userspace (ceph-fuse, or Samba built-in) client. If using userspace, please make sure you've got the latest one. -Greg The system is:

Re: [ceph-users] why are there degraded PGs when adding OSDs?

2015-07-27 Thread Chad William Seys
Hi Sam, The pg might also be degraded right after a map change which changes the up/acting sets since the few objects updated right before the map change might be new on some replicas and old on the other replicas. While in that state, those specific objects are degraded, and the pg would

Re: [ceph-users] why are there degraded PGs when adding OSDs?

2015-07-27 Thread Samuel Just
Hmm, that's odd. Can you attach the osdmap and ceph pg dump prior to the addition (with all pgs active+clean), then the osdmap and ceph pg dump afterwards? -Sam - Original Message - From: Chad William Seys cws...@physics.wisc.edu To: Samuel Just sj...@redhat.com, ceph-users

[ceph-users] why are there degraded PGs when adding OSDs?

2015-07-27 Thread Chad William Seys
Hi All, I recently added some OSDs to the Ceph cluster (0.94.2). I noticed that 'ceph -s' reported both misplaced AND degraded PGs. Why should any PGs become degraded? Seems as though Ceph should only be reporting misplaced PGs? From the Giant release notes: Degraded vs misplaced: the Ceph

Re: [ceph-users] why are there degraded PGs when adding OSDs?

2015-07-27 Thread Chad William Seys
Hi Sam, I'll need help getting the osdmap and pg dump prior to addition. I can remove the OSDs and add again if the osdmap (etc.) is not logged somewhere. Chad. Hmm, that's odd. Can you attach the osdmap and ceph pg dump prior to the addition (with all pgs active+clean), then

Re: [ceph-users] why are there degraded PGs when adding OSDs?

2015-07-27 Thread Chad William Seys
Hi Sam, I think I may have the problem: I noticed that the new host was created with straw2 instead of straw. Would this account for 50% of PGs being degraded? (I'm removing the OSDs on that host and will recreate with 'firefly' tunables.) Thanks! Chad. On Monday, July 27, 2015 15:09:21 Chad

Re: [ceph-users] Can a cephfs volume get errors and how are they fixed?

2015-07-27 Thread Roland Giesler
On 15 July 2015 at 17:34, John Spray john.sp...@redhat.com wrote: On 15/07/15 16:11, Roland Giesler wrote: I mount cephfs in /etc/fstab and all seemed well for quite a few months. Now however, I start seeing strange things like directories with corrupted files names in the file system.

Re: [ceph-users] Best method to limit snapshot/clone space overhead

2015-07-27 Thread Jason Dillaman
If I understand correctly you want to look at how many “guest filesystem block size” blocks there are that are empty? This might not be that precise because we do not discard blocks inside the guests, but if you tell me how to gather this - I can certainly try that. I’m not sure if my bash-fu

Re: [ceph-users] State of nfs-ganesha CEPH fsal

2015-07-27 Thread Eric Eastman
We are looking at using Ganesha NFS with the Ceph file system. Currently I am testing the FSAL interface on Ganesha NFS Release = V2.2.0-2 running on Ceph 9.0.2. This is all early work, as Ceph FS is still not considered production ready, and Ceph 9.0.2 is a development release. Currently I am

[ceph-users] How is Storage Object managed in Ceph Object Storage

2015-07-27 Thread Jiwan Ninglekhu
Hi: I am defining Objects broken down from a file to be stored in Object Storage Cloud, as Storage Objects. What I know: I have read documents and papers about object storage cloud. Most of the time, the documents assume that storage objects from a file (to be stored) have been already created

Re: [ceph-users] Trying to remove osd

2015-07-27 Thread Nikhil Mitra (nikmitra)
Hi Paul, Did you try to stop the osd first before marking it down and out ? stop ceph-osd id=21 or /etc/init.d/ceph stop osd.21 Ceph osd crush remove osd.21 Ceph auth del osd.21 Ceph osd rm osd.310 Regards, Nikhil Mitra From: ceph-users

[ceph-users] wrong documentation in add or rm mons

2015-07-27 Thread Makkelie, R (ITCDCC) - KLM
i followed the following documentation to add monitors to my already existing cluster with 1 mon http://ceph.com/docs/master/rados/operations/add-or-rm-mons/ when i follow this documentation. the monitor assimilates the old monitor so my monitor status is gone. but when i skip the ceph mon add