On Mon, Mar 16, 2015 at 5:08 PM, Florent B flor...@coppint.com wrote:
Since then I deleted the pool.
But I now have another problem, in fact the opposite of the previous :
now I never deleted files in clients, data objects and metadata are
still in pools, but directory is empty for clients
Hi all I have seen that new versions of CEPH with new OS like RHEL7 and Cento7
doesn’t need information like mon.node1 and osd.0 etc.. anymore, can anybody
tell me if is that for real? or do I need still need to write config like this:
[osd.0]
host = sagitario
addr = 192.168.1.67
On 03/16/2015 01:55 PM, Gaurang Vyas wrote:
running on ubuntu with nginx + php-fpm
?php
$rados = rados_create('admin');
rados_conf_read_file($rados, '/etc/ceph/ceph.conf');
rados_conf_set($rados, 'keyring','/etc/ceph/ceph.client.admin.keyring');
$temp = rados_conf_get($rados,
On 16/03/2015, at 12.23, Alexandre DERUMIER aderum...@odiso.com wrote:
We use Proxmox, so I think it uses librbd ?
As It's me that I made the proxmox rbd plugin, I can confirm that yes, it's
librbd ;)
Is the ceph cluster on dedicated nodes ? or vms are running on same nodes
than osd
running on ubuntu with nginx + php-fpm
?php
$rados = rados_create('admin');
rados_conf_read_file($rados, '/etc/ceph/ceph.conf');
rados_conf_set($rados, 'keyring','/etc/ceph/ceph.client.admin.keyring');
$temp = rados_conf_get($rados, rados_osd_op_timeout);
echo osd ;
echo $temp;
$temp =
That full system slows down, OK, but brutal stop...
This is strange, that could be:
- qemu crash, maybe a bug in rbd block storage (if you use librbd)
- oom-killer on you host (any logs ?)
what is your qemu version ?
- Mail original -
De: Florent Bautista flor...@coppint.com
À:
We use Proxmox, so I think it uses librbd ?
As It's me that I made the proxmox rbd plugin, I can confirm that yes, it's
librbd ;)
Is the ceph cluster on dedicated nodes ? or vms are running on same nodes than
osd daemons ?
And I precise that not all VMs on that pool crashed, only some of
On 16/03/2015, at 11.14, Florent B flor...@coppint.com wrote:
On 03/16/2015 11:03 AM, Alexandre DERUMIER wrote:
This is strange, that could be:
- qemu crash, maybe a bug in rbd block storage (if you use librbd)
- oom-killer on you host (any logs ?)
what is your qemu version ?
Now,
May I know your ceph version.?. The latest version of firefly 80.9 has
patches to avoid excessive data migrations during rewighting osds. You may
need set a tunable inorder make this patch active.
This is a bugfix release for firefly. It fixes a performance regression
in librbd, an important
I always keep my pg number a power of 2. So I’d go from 2048 to 4096. I’m not
sure if this is the safest way, but it’s worked for me.
[yp]
Michael Kuriger
Sr. Unix Systems Engineer
• mk7...@yp.commailto:mk7...@yp.com |• 818-649-7235
From: Chu Duc Minh
Sumit,
You may have better luck on the ceph-calamari mailing list. Anyway -
calamari uses graphite to handle metrics, and graphite does indeed write
them to files.
John
On 11/03/2015 05:09, Sumit Gaur wrote:
Hi
I have a basic architecture related question. I know Calamari collect
system
On 13/03/2015 11:51, Florent B wrote:
Hi all,
My question is about user management in CephFS.
Is it possible to restrict a CephX user to access some subdirectories ?
Not yet. The syntax for setting a path= part in the authorization
caps for a cephx user exists, but the code for enforcing it
On 16/03/2015 16:30, Florent B wrote:
Thank you John :) Hammer is not released yet, is it ? Is it 'safe' to
upgrade a production cluster to 0.93 ?
I keep forgetting that -- yes, I should have added ...when it's
released :-)
John
___
ceph-users
On 14/03/2015 09:22, Florent B wrote:
Hi,
What do you call old MDS ? I'm on Giant release, it is not very old...
With CephFS we have a special definition of old that is anything that
doesn't have the very latest bug fixes ;-)
There have definitely been fixes to stray file handling[1] between
Hi Cephers,
Our university could deploy ceph. The goal is to store datas for research
laboratories (non-HPC) . To do this, we plan to use Ceph with RBD (mount block
device) from a NFS ( or CIFS ) server (ceph client) to workstations in
laboratories. According to our tests, the OS (ubuntu or
Great work !
David Moreau Simard
On 2015-03-15 06:29 PM, Loic Dachary wrote:
Hi Ceph,
In an attempt to clarify what Ceph release is stable, LTS or development. a
new page was added to the documentation:
http://ceph.com/docs/master/releases/ It is a matrix where each cell is a
release
@Michael Kuriger: when ceph/librbd operate normally, i know that double the
pg_num is the safe way. But when it has problem, i think double it can make
many many VMs die (maybe = 50%?)
On Mon, Mar 16, 2015 at 9:53 PM, Michael Kuriger mk7...@yp.com wrote:
I always keep my pg number a power of
I'm using the latest Giant and have the same issue. When i increase PG_num
of a pool from 2048 to 2148, my VMs is still ok. When i increase from 2148
to 2400, some VMs die (Qemu-kvm process die).
My physical servers (host VMs) running kernel 3.13 and use librbd.
I think it's a bug in librbd with
Yes, the placement target feature is logically separate from multi-zone
setups. Placement targets are configured in the region though, which
somewhat muddies the issue.
Placement targets are useful feature for multi-zone, so different zones in
a cluster don't share the same disks. Federation
On Wed, Mar 11, 2015 at 8:40 AM, Artem Savinov asavi...@asdco.ru wrote:
hello.
ceph transfers osd node in the down status by default , after receiving 3
reports about disabled nodes. Reports are sent per osd heartbeat grace
seconds, but the settings of mon_osd_adjust_heartbeat_gratse = true,
On Wed, Mar 11, 2015 at 2:25 PM, Nick Fisk n...@fisk.me.uk wrote:
I’m not sure if it’s something I’m doing wrong or just experiencing an
oddity, but when my cache tier flushes dirty blocks out to the base tier, the
writes seem to hit the OSD’s straight away instead of coalescing in the
On Wed, Mar 11, 2015 at 3:49 PM, Francois Lafont flafdiv...@free.fr wrote:
Hi,
I was always in the same situation: I couldn't remove an OSD without
have some PGs definitely stuck to the active+remapped state.
But I remembered I read on IRC that, before to mark out an OSD, it
could be
Thanks a lot Stephane and Kamil,
Your reply was really helpful. I needed a different version of ceph client
on my client machine. Initially my java application using librados was
throwing connection time out. Then I considered querying ceph from command
line (ceph --id ...), which was giving the
On Mon, 16 Mar 2015 16:09:12 -0700 Gregory Farnum wrote:
Nothing here particularly surprises me. I don't remember all the
details of the filestore's rate limiting off the top of my head, but
it goes to great lengths to try and avoid letting the journal get too
far ahead of the backing store.
- Original Message -
From: Craig Lewis cle...@centraldesktop.com
To: Gregory Farnum g...@gregs42.com
Cc: ceph-users@lists.ceph.com
Sent: Monday, March 16, 2015 11:48:15 AM
Subject: Re: [ceph-users] RadosGW Direct Upload Limitation
Maybe, but I'm not sure if Yehuda would
The information you're giving sounds a little contradictory, but my
guess is that you're seeing the impacts of object promotion and
flushing. You can sample the operations the OSDs are doing at any
given time by running ops_in_progress (or similar, I forget exact
phrasing) command on the OSD admin
-Original Message-
From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
Gregory Farnum
Sent: 16 March 2015 17:33
To: Nick Fisk
Cc: ceph-users@lists.ceph.com
Subject: Re: [ceph-users] Cache Tier Flush = immediate base tier journal
sync?
On Wed, Mar 11,
Hi,
I am getting relatively bad performance from cephfs. I use a replicated
cache pool on ssd in front of an erasure coded pool on rotating media.
When reading big files (streaming video), I see a lot of disk i/o,
especially writes. I have no clue what could cause these writes. The
writes are
Hi,
I forgot to mention: while I am seeing these writes in iotop and
/proc/diskstats for the hdd's, I am -not- seeing any writes in rados
df for the pool residing on these disks. There is only one pool active
on the hdd's and according to rados df it is getting zero writes when
I'm just reading
Nothing here particularly surprises me. I don't remember all the
details of the filestore's rate limiting off the top of my head, but
it goes to great lengths to try and avoid letting the journal get too
far ahead of the backing store. Disabling the filestore flusher and
increasing the sync
Thanks.
Is this possible outside of multi-zone setup. (With only one Zone)?
For example, I want to have pools with different replication
factors(or erasure codings) and map users to these pools.
-Sreenath
On 3/13/15, Craig Lewis cle...@centraldesktop.com wrote:
Yes, RadosGW has the concept
Maybe, but I'm not sure if Yehuda would want to take it upstream or
not. This limit is present because it's part of the S3 spec. For
larger objects you should use multi-part upload, which can get much
bigger.
-Greg
Note that the multi-part upload has a lower limit of 4MiB per part, and
If I remember/guess correctly, if you mark an OSD out it won't
necessarily change the weight of the bucket above it (ie, the host),
whereas if you change the weight of the OSD then the host bucket's
weight changes.
-Greg
That sounds right. Marking an OSD out is a ceph osd reweight, not
On Sat, Mar 14, 2015 at 3:04 AM, pragya jain prag_2...@yahoo.co.in wrote:
Hello all!
I am working on Ceph object storage architecture from last few months.
I am unable to search a document which can describe how Ceph object
storage APIs (Swift/S3 APIs) are mappedd with Ceph storage cluster
Hi,
Gregory Farnum a wrote :
If I remember/guess correctly, if you mark an OSD out it won't
necessarily change the weight of the bucket above it (ie, the host),
whereas if you change the weight of the OSD then the host bucket's
weight changes.
I can just say that, indeed, I have noticed
On Mon, Mar 16, 2015 at 11:14 AM, Georgios Dimitrakakis
gior...@acmac.uoc.gr wrote:
Hi all!
I have recently updated to CEPH version 0.80.9 (latest Firefly release)
which presumably
supports direct upload.
I 've tried to upload a file using this functionality and it seems that is
working
Out of curiousity, what's the frequency of the peaks and troughs?
RadosGW has configs on how long it should wait after deleting before
garbage collecting, how long between GC runs, and how many objects it can
GC in per run.
The defaults are 2 hours, 1 hour, and 32 respectively. Search
VMs are running on the same nodes than OSD
Are you sure that you didn't some kind of out of memory.
pg rebalance can be memory hungry. (depend how many osd you have).
do you see oom-killer in your host logs ?
- Mail original -
De: Florent Bautista flor...@coppint.com
À: aderumier
On Mon, Mar 16, 2015 at 12:12 PM, Craig Lewis cle...@centraldesktop.com wrote:
Out of curiousity, what's the frequency of the peaks and troughs?
RadosGW has configs on how long it should wait after deleting before garbage
collecting, how long between GC runs, and how many objects it can GC in
please somebody answer my queries. -RegardsPragya JainDepartment of
Computer ScienceUniversity of DelhiDelhi, India
On Saturday, 14 March 2015 3:34 PM, pragya jain prag_2...@yahoo.co.in
wrote:
Hello all!
I am working on Ceph object storage architecture from last few months.
I
hello all!
I am working on Ceph object storage architecture.I have some queries:
In case of configuring federated system, we need to create regions containing
one or more zones and the cluster must have a master region and each region
must have a master zone.
but in case of simple gateway
41 matches
Mail list logo