well, nice one :)

*step chooseleaf firstn 0 type host* - it is the part of default crush map
(3 hosts, 2 OSDs per host)

It means: write 3 replicas (in my case) to 3 hosts...and randomly select
OSD from each host ?

I already read all the docs...and still not sure how to proceed...


On 16 October 2013 23:27, Mike Dawson <mike.daw...@cloudapt.com> wrote:

> Andrija,
>
> You can use a single pool and the proper CRUSH rule
>
>
> step chooseleaf firstn 0 type host
>
>
> to accomplish your goal.
>
> http://ceph.com/docs/master/**rados/operations/crush-map/<http://ceph.com/docs/master/rados/operations/crush-map/>
>
>
> Cheers,
> Mike Dawson
>
>
>
> On 10/16/2013 5:16 PM, Andrija Panic wrote:
>
>> Hi,
>>
>> I have 2 x  2TB disks, in 3 servers, so total of 6 disks... I have
>> deployed total of 6 OSDs.
>> ie:
>> host1 = osd.0 and osd.1
>> host2 = osd.2 and osd.3
>> host4 = osd.4 and osd.5
>>
>> Now, since I will have total of 3 replica (original + 2 replicas), I
>> want my replica placement to be such, that I don't end up having 2
>> replicas on 1 host (replica on osd0, osd1 (both on host1) and replica on
>> osd2. I want all 3 replicas spread on different hosts...
>>
>> I know this is to be done via crush maps, but I'm not sure if it would
>> be better to have 2 pools, 1 pool on osd0,2,4 and and another pool on
>> osd1,3,5.
>>
>> If possible, I would want only 1 pool, spread across all 6 OSDs, but
>> with data placement such, that I don't end up having 2 replicas on 1
>> host...not sure if this is possible at all...
>>
>> Is that possible, or maybe I should go for RAID0 in each server (2 x 2Tb
>> = 4TB for osd0) or maybe JBOD  (1 volume, so 1 OSD per host) ?
>>
>> Any suggesting about best practice ?
>>
>> Regards,
>>
>> --
>>
>> Andrija Panić
>>
>>
>> ______________________________**_________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>>
>>


-- 

Andrija Panić
--------------------------------------
  http://admintweets.com
--------------------------------------
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to