On Mon, Apr 22, 2013 at 9:42 AM, John Nielsen <[email protected]> wrote:
> On Mar 15, 2013, at 8:15 PM, Yehuda Sadeh <[email protected]> wrote:
>
>> On Fri, Mar 15, 2013 at 5:06 PM, Mandell Degerness
>> <[email protected]> wrote:
>>> How are the pools used by rgw defined?
>>>
>>> Specifically, if I want to ensure that all of the data stored by rgw
>>> uses pools which are replicated 3 times and have a pgnum and a pgpnum
>>> greater than 8, what do I need to set?
>>
>> There are a bunch of pools that are create automatically. Currently
>> the best way to avoid them being created with a very low pg number is
>> to pre-create them before starting the gateways. Also, there's the
>> actual data pool (the pool that holds the user data and bucket
>> indexes), which is the .rgw.buckets pool, and which you can modify by
>> using the 'radosgw-admin pool pool add/rm. The following is
>> (currently) the default pools that are being used. The ability to set
>> and modify these will be part of the disaster-recovery/georeplication
>> feature.
>
> Is there an ETA for this work? I suppose seeing it in Cuttlefish isn't 
> likely, is it?
>
> I'd like the ability to run two (or more) completely distinct gateways on the 
> same rados cluster. Each gateway should have its own (ceph/authx) user and 
> its own set of pools. It sounds like that is not currently possible since the 
> pool names are read-only and apparently global. Is having separate pools per 
> gateway user/instance on the roadmap at all?
>
>> Current version only allows to view this list.
>>
>> { "domain_root": ".rgw",
>>  "control_pool": ".rgw.control",
>>  "gc_pool": ".rgw.gc",
>>  "log_pool": ".log",
>>  "intent_log_pool": ".intent-log",
>>  "usage_log_pool": ".usage",
>>  "user_keys_pool": ".users",
>>  "user_email_pool": ".users.email",
>>  "user_swift_pool": ".users.swift",
>>  "user_uid_pool ": ".users.uid"}


The ability to configure the pools that are being used will make it to
Cuttlefish. The 'radosgw-admin zone info' will replace 'radosgw-admin
cluster info', and there will be a new 'radosgw-admin zone set'
command that will allow setting that info.
In practice you'll also be able to have more than one setup on a
single cluster, leveraging the 'rgw root zone pool' configurable.

Basically you'd need to do something like this:

# radosgw-admin --rgw-root-zone-pool=.zone2.root zone info > zone.info

<now edit zone.info, replace the default pool names with new pools>

#radosgw-admin --rgw-root-zone-pool=.zone2.root zone set < zone.info

and for the second instance run radosgw with
--rgw-root-zone-pool=.zone2.root (or put it in your ceph.conf as 'rgw
root zone pool = .zone2.root).

>
> I'm also trying to fine-tune the pools used by our gateway. Obviously pools 
> that store lots of objects and/or data (like .rgw.buckets) should have more 
> than the low default number of placement groups. Is the same true for the 
> rest of the pools? Or in other words, if a pool is only going to contain a 
> few small objects, does it make sense to inflate the number of placement 
> groups in the cluster as a whole?
>
> If such a distinction makes sense, which pools are typically 'large'?
>

Well, pools other than the data pool can have smaller number of pgs.
Though you'd definitely want to have more than the default 8.

Yehuda
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to