Re: [ceph-users] CephFS: delayed objects deletion ?

2015-03-16 Thread Yan, Zheng
On Mon, Mar 16, 2015 at 5:08 PM, Florent B flor...@coppint.com wrote: Since then I deleted the pool. But I now have another problem, in fact the opposite of the previous : now I never deleted files in clients, data objects and metadata are still in pools, but directory is empty for clients

[ceph-users] ceph.conf

2015-03-16 Thread Jesus Chavez (jeschave)
Hi all I have seen that new versions of CEPH with new OS like RHEL7 and Cento7 doesn’t need information like mon.node1 and osd.0 etc.. anymore, can anybody tell me if is that for real? or do I need still need to write config like this: [osd.0] host = sagitario addr = 192.168.1.67

Re: [ceph-users] PHP Rados failed in read operation if object size is large (say more than 10 MB )

2015-03-16 Thread Wido den Hollander
On 03/16/2015 01:55 PM, Gaurang Vyas wrote: running on ubuntu with nginx + php-fpm ?php $rados = rados_create('admin'); rados_conf_read_file($rados, '/etc/ceph/ceph.conf'); rados_conf_set($rados, 'keyring','/etc/ceph/ceph.client.admin.keyring'); $temp = rados_conf_get($rados,

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Steffen W Sørensen
On 16/03/2015, at 12.23, Alexandre DERUMIER aderum...@odiso.com wrote: We use Proxmox, so I think it uses librbd ? As It's me that I made the proxmox rbd plugin, I can confirm that yes, it's librbd ;) Is the ceph cluster on dedicated nodes ? or vms are running on same nodes than osd

[ceph-users] PHP Rados failed in read operation if object size is large (say more than 10 MB )

2015-03-16 Thread Gaurang Vyas
running on ubuntu with nginx + php-fpm ?php $rados = rados_create('admin'); rados_conf_read_file($rados, '/etc/ceph/ceph.conf'); rados_conf_set($rados, 'keyring','/etc/ceph/ceph.client.admin.keyring'); $temp = rados_conf_get($rados, rados_osd_op_timeout); echo osd ; echo $temp; $temp =

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Alexandre DERUMIER
That full system slows down, OK, but brutal stop... This is strange, that could be: - qemu crash, maybe a bug in rbd block storage (if you use librbd) - oom-killer on you host (any logs ?) what is your qemu version ? - Mail original - De: Florent Bautista flor...@coppint.com À:

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Alexandre DERUMIER
We use Proxmox, so I think it uses librbd ? As It's me that I made the proxmox rbd plugin, I can confirm that yes, it's librbd ;) Is the ceph cluster on dedicated nodes ? or vms are running on same nodes than osd daemons ? And I precise that not all VMs on that pool crashed, only some of

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Steffen W Sørensen
On 16/03/2015, at 11.14, Florent B flor...@coppint.com wrote: On 03/16/2015 11:03 AM, Alexandre DERUMIER wrote: This is strange, that could be: - qemu crash, maybe a bug in rbd block storage (if you use librbd) - oom-killer on you host (any logs ?) what is your qemu version ? Now,

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Azad Aliyar
May I know your ceph version.?. The latest version of firefly 80.9 has patches to avoid excessive data migrations during rewighting osds. You may need set a tunable inorder make this patch active. This is a bugfix release for firefly. It fixes a performance regression in librbd, an important

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Michael Kuriger
I always keep my pg number a power of 2. So I’d go from 2048 to 4096. I’m not sure if this is the safest way, but it’s worked for me. [yp] Michael Kuriger Sr. Unix Systems Engineer • mk7...@yp.commailto:mk7...@yp.com |• 818-649-7235 From: Chu Duc Minh

Re: [ceph-users] Calamari - Data

2015-03-16 Thread John Spray
Sumit, You may have better luck on the ceph-calamari mailing list. Anyway - calamari uses graphite to handle metrics, and graphite does indeed write them to files. John On 11/03/2015 05:09, Sumit Gaur wrote: Hi I have a basic architecture related question. I know Calamari collect system

Re: [ceph-users] CephFS: authorizations ?

2015-03-16 Thread John Spray
On 13/03/2015 11:51, Florent B wrote: Hi all, My question is about user management in CephFS. Is it possible to restrict a CephX user to access some subdirectories ? Not yet. The syntax for setting a path= part in the authorization caps for a cephx user exists, but the code for enforcing it

Re: [ceph-users] CephFS: delayed objects deletion ?

2015-03-16 Thread John Spray
On 16/03/2015 16:30, Florent B wrote: Thank you John :) Hammer is not released yet, is it ? Is it 'safe' to upgrade a production cluster to 0.93 ? I keep forgetting that -- yes, I should have added ...when it's released :-) John ___ ceph-users

Re: [ceph-users] CephFS: delayed objects deletion ?

2015-03-16 Thread John Spray
On 14/03/2015 09:22, Florent B wrote: Hi, What do you call old MDS ? I'm on Giant release, it is not very old... With CephFS we have a special definition of old that is anything that doesn't have the very latest bug fixes ;-) There have definitely been fixes to stray file handling[1] between

[ceph-users] OS file Cache, Ceph RBD cache and Network files systems

2015-03-16 Thread Stéphane DUGRAVOT
Hi Cephers, Our university could deploy ceph. The goal is to store datas for research laboratories (non-HPC) . To do this, we plan to use Ceph with RBD (mount block device) from a NFS ( or CIFS ) server (ceph client) to workstations in laboratories. According to our tests, the OS (ubuntu or

Re: [ceph-users] Ceph release timeline

2015-03-16 Thread David Moreau Simard
Great work ! David Moreau Simard On 2015-03-15 06:29 PM, Loic Dachary wrote: Hi Ceph, In an attempt to clarify what Ceph release is stable, LTS or development. a new page was added to the documentation: http://ceph.com/docs/master/releases/ It is a matrix where each cell is a release

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Chu Duc Minh
@Michael Kuriger: when ceph/librbd operate normally, i know that double the pg_num is the safe way. But when it has problem, i think double it can make many many VMs die (maybe = 50%?) On Mon, Mar 16, 2015 at 9:53 PM, Michael Kuriger mk7...@yp.com wrote: I always keep my pg number a power of

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Chu Duc Minh
I'm using the latest Giant and have the same issue. When i increase PG_num of a pool from 2048 to 2148, my VMs is still ok. When i increase from 2148 to 2400, some VMs die (Qemu-kvm process die). My physical servers (host VMs) running kernel 3.13 and use librbd. I think it's a bug in librbd with

Re: [ceph-users] Mapping users to different rgw pools

2015-03-16 Thread Craig Lewis
Yes, the placement target feature is logically separate from multi-zone setups. Placement targets are configured in the region though, which somewhat muddies the issue. Placement targets are useful feature for multi-zone, so different zones in a cluster don't share the same disks. Federation

Re: [ceph-users] osd laggy algorithm

2015-03-16 Thread Gregory Farnum
On Wed, Mar 11, 2015 at 8:40 AM, Artem Savinov asavi...@asdco.ru wrote: hello. ceph transfers osd node in the down status by default , after receiving 3 reports about disabled nodes. Reports are sent per osd heartbeat grace seconds, but the settings of mon_osd_adjust_heartbeat_gratse = true,

Re: [ceph-users] Cache Tier Flush = immediate base tier journal sync?

2015-03-16 Thread Gregory Farnum
On Wed, Mar 11, 2015 at 2:25 PM, Nick Fisk n...@fisk.me.uk wrote: I’m not sure if it’s something I’m doing wrong or just experiencing an oddity, but when my cache tier flushes dirty blocks out to the base tier, the writes seem to hit the OSD’s straight away instead of coalescing in the

Re: [ceph-users] PGs stuck unclean active+remapped after an osd marked out

2015-03-16 Thread Gregory Farnum
On Wed, Mar 11, 2015 at 3:49 PM, Francois Lafont flafdiv...@free.fr wrote: Hi, I was always in the same situation: I couldn't remove an OSD without have some PGs definitely stuck to the active+remapped state. But I remembered I read on IRC that, before to mark out an OSD, it could be

Re: [ceph-users] client-ceph [can not connect from client][connect protocol feature mismatch]

2015-03-16 Thread Sonal Dubey
Thanks a lot Stephane and Kamil, Your reply was really helpful. I needed a different version of ceph client on my client machine. Initially my java application using librados was throwing connection time out. Then I considered querying ceph from command line (ceph --id ...), which was giving the

Re: [ceph-users] Cache Tier Flush = immediate base tier journal sync?

2015-03-16 Thread Christian Balzer
On Mon, 16 Mar 2015 16:09:12 -0700 Gregory Farnum wrote: Nothing here particularly surprises me. I don't remember all the details of the filestore's rate limiting off the top of my head, but it goes to great lengths to try and avoid letting the journal get too far ahead of the backing store.

Re: [ceph-users] RadosGW Direct Upload Limitation

2015-03-16 Thread Yehuda Sadeh-Weinraub
- Original Message - From: Craig Lewis cle...@centraldesktop.com To: Gregory Farnum g...@gregs42.com Cc: ceph-users@lists.ceph.com Sent: Monday, March 16, 2015 11:48:15 AM Subject: Re: [ceph-users] RadosGW Direct Upload Limitation Maybe, but I'm not sure if Yehuda would

Re: [ceph-users] CephFS unexplained writes

2015-03-16 Thread Gregory Farnum
The information you're giving sounds a little contradictory, but my guess is that you're seeing the impacts of object promotion and flushing. You can sample the operations the OSDs are doing at any given time by running ops_in_progress (or similar, I forget exact phrasing) command on the OSD admin

Re: [ceph-users] Cache Tier Flush = immediate base tier journal sync?

2015-03-16 Thread Nick Fisk
-Original Message- From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Gregory Farnum Sent: 16 March 2015 17:33 To: Nick Fisk Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Cache Tier Flush = immediate base tier journal sync? On Wed, Mar 11,

[ceph-users] CephFS unexplained writes

2015-03-16 Thread Erik Logtenberg
Hi, I am getting relatively bad performance from cephfs. I use a replicated cache pool on ssd in front of an erasure coded pool on rotating media. When reading big files (streaming video), I see a lot of disk i/o, especially writes. I have no clue what could cause these writes. The writes are

Re: [ceph-users] CephFS unexplained writes

2015-03-16 Thread Erik Logtenberg
Hi, I forgot to mention: while I am seeing these writes in iotop and /proc/diskstats for the hdd's, I am -not- seeing any writes in rados df for the pool residing on these disks. There is only one pool active on the hdd's and according to rados df it is getting zero writes when I'm just reading

Re: [ceph-users] Cache Tier Flush = immediate base tier journal sync?

2015-03-16 Thread Gregory Farnum
Nothing here particularly surprises me. I don't remember all the details of the filestore's rate limiting off the top of my head, but it goes to great lengths to try and avoid letting the journal get too far ahead of the backing store. Disabling the filestore flusher and increasing the sync

Re: [ceph-users] Mapping users to different rgw pools

2015-03-16 Thread Sreenath BH
Thanks. Is this possible outside of multi-zone setup. (With only one Zone)? For example, I want to have pools with different replication factors(or erasure codings) and map users to these pools. -Sreenath On 3/13/15, Craig Lewis cle...@centraldesktop.com wrote: Yes, RadosGW has the concept

Re: [ceph-users] RadosGW Direct Upload Limitation

2015-03-16 Thread Craig Lewis
Maybe, but I'm not sure if Yehuda would want to take it upstream or not. This limit is present because it's part of the S3 spec. For larger objects you should use multi-part upload, which can get much bigger. -Greg Note that the multi-part upload has a lower limit of 4MiB per part, and

Re: [ceph-users] PGs stuck unclean active+remapped after an osd marked out

2015-03-16 Thread Craig Lewis
If I remember/guess correctly, if you mark an OSD out it won't necessarily change the weight of the bucket above it (ie, the host), whereas if you change the weight of the OSD then the host bucket's weight changes. -Greg That sounds right. Marking an OSD out is a ceph osd reweight, not

Re: [ceph-users] query about mapping of Swift/S3 APIs to Ceph cluster APIs

2015-03-16 Thread Craig Lewis
On Sat, Mar 14, 2015 at 3:04 AM, pragya jain prag_2...@yahoo.co.in wrote: Hello all! I am working on Ceph object storage architecture from last few months. I am unable to search a document which can describe how Ceph object storage APIs (Swift/S3 APIs) are mappedd with Ceph storage cluster

Re: [ceph-users] PGs stuck unclean active+remapped after an osd marked out

2015-03-16 Thread Francois Lafont
Hi, Gregory Farnum a wrote : If I remember/guess correctly, if you mark an OSD out it won't necessarily change the weight of the bucket above it (ie, the host), whereas if you change the weight of the OSD then the host bucket's weight changes. I can just say that, indeed, I have noticed

Re: [ceph-users] RadosGW Direct Upload Limitation

2015-03-16 Thread Gregory Farnum
On Mon, Mar 16, 2015 at 11:14 AM, Georgios Dimitrakakis gior...@acmac.uoc.gr wrote: Hi all! I have recently updated to CEPH version 0.80.9 (latest Firefly release) which presumably supports direct upload. I 've tried to upload a file using this functionality and it seems that is working

Re: [ceph-users] Shadow files

2015-03-16 Thread Craig Lewis
Out of curiousity, what's the frequency of the peaks and troughs? RadosGW has configs on how long it should wait after deleting before garbage collecting, how long between GC runs, and how many objects it can GC in per run. The defaults are 2 hours, 1 hour, and 32 respectively. Search

Re: [ceph-users] [SPAM] Changing pg_num = RBD VM down !

2015-03-16 Thread Alexandre DERUMIER
VMs are running on the same nodes than OSD Are you sure that you didn't some kind of out of memory. pg rebalance can be memory hungry. (depend how many osd you have). do you see oom-killer in your host logs ? - Mail original - De: Florent Bautista flor...@coppint.com À: aderumier

Re: [ceph-users] Shadow files

2015-03-16 Thread Gregory Farnum
On Mon, Mar 16, 2015 at 12:12 PM, Craig Lewis cle...@centraldesktop.com wrote: Out of curiousity, what's the frequency of the peaks and troughs? RadosGW has configs on how long it should wait after deleting before garbage collecting, how long between GC runs, and how many objects it can GC in

[ceph-users] Fw: query about mapping of Swift/S3 APIs to Ceph cluster APIs

2015-03-16 Thread pragya jain
please somebody answer my queries. -RegardsPragya JainDepartment of Computer ScienceUniversity of DelhiDelhi, India On Saturday, 14 March 2015 3:34 PM, pragya jain prag_2...@yahoo.co.in wrote: Hello all! I am working on Ceph object storage architecture from last few months. I

[ceph-users] query about region and zone creation while configuring RADOSGW

2015-03-16 Thread pragya jain
hello all! I am working on Ceph object storage architecture.I have some queries: In case of configuring federated system, we need to create regions containing one or more zones and the cluster must have a master region and each region must have a master zone. but in case of simple gateway