On 17/12/2014 19:46, Italo Santos wrote:> Understood.
> Thanks for your help, the cluster is healthy now :D
> 
> Also, using for example k=6,m=1 and failure domain by host I’ll be able lose 
> all OSD on the same host, but if a lose 2 disks on different hosts I can lose 
> data right? So, it is possible been a failure domain which allow me to lose 
> an OSD or a host?

That's actually a good way to put it :-)

> 
> Regards.
> 
> *Italo Santos*
> http://italosantos.com.br/
> 
> On Wednesday, December 17, 2014 at 4:27 PM, Loic Dachary wrote:
> 
>>
>>
>> On 17/12/2014 19:22, Italo Santos wrote:
>>> Loic,
>>>
>>> So, if want have a failure domain by host, I’ll need set up a erasure 
>>> profile which k+m = total number of hosts I have, right?
>>
>> Yes, k+m has to be <= number of hosts.
>>
>>>
>>> Regards.
>>>
>>> *Italo Santos*
>>> http://italosantos.com.br/
>>>
>>> On Wednesday, December 17, 2014 at 3:24 PM, Loic Dachary wrote:
>>>
>>>>
>>>>
>>>> On 17/12/2014 18:18, Italo Santos wrote:
>>>>> Hello,
>>>>>
>>>>> I’ve take a look to this documentation (which help a lot) and if I 
>>>>> understand right, when I set a profile like:
>>>>>
>>>>> ===
>>>>> ceph osd erasure-code-profile set isilon k=8 m=2 
>>>>> ruleset-failure-domain=host
>>>>> ===
>>>>>
>>>>> And create a pool following the recommendations on doc, I’ll need 
>>>>> (100*16)/2 = 800 PGs, I’ll need the sufficient number of hosts to support 
>>>>> create total PGs?
>>>>
>>>> You will need k+m = 10 host per OSD. If you only have 10 hosts that should 
>>>> be ok and the 800 PGs will use these 10 OSD in various orders. It also 
>>>> means that you will end up having 800 PG per OSD which is a bit too mche. 
>>>> If you have 20 OSDs that will be better : each PG will get 10 OSD out of 
>>>> 20 and each OSD will have 400 PGs. Ideally you want the number of PG per 
>>>> OSD to be in the range (approximately) [20,300].
>>>>
>>>> Cheers
>>>>
>>>>>
>>>>> Regards.
>>>>>
>>>>> *Italo Santos*
>>>>> http://italosantos.com.br/
>>>>>
>>>>> On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> Thanks for the update : good news are much appreciated :-) Would you 
>>>>>> have time to review the documentation at 
>>>>>> https://github.com/ceph/ceph/pull/3194/files ? It was partly motivated 
>>>>>> by the problem you had.
>>>>>>
>>>>>> Cheers
>>>>>>
>>>>>> On 17/12/2014 14:03, Italo Santos wrote:
>>>>>>> Hello Loic,
>>>>>>>
>>>>>>> Thanks for you help, I’ve take a look to my crush map and I replace 
>>>>>>> "step chooseleaf indep 0 type osd” by "step choose indep 0 type osd” 
>>>>>>> and all PGs was created successfully.
>>>>>>>
>>>>>>> At.
>>>>>>>
>>>>>>> *Italo Santos*
>>>>>>> http://italosantos.com.br/
>>>>>>>
>>>>>>> On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> The 2147483647 means that CRUSH did not find enough OSD for a given 
>>>>>>>> PG. If you check the crush rule associated with the erasure coded 
>>>>>>>> pool, you will most probably find why.
>>>>>>>>
>>>>>>>> Cheers
>>>>>>>>
>>>>>>>> On 16/12/2014 23:32, Italo Santos wrote:
>>>>>>>>> Hello,
>>>>>>>>>
>>>>>>>>> I'm trying to create an erasure pool following 
>>>>>>>>> http://docs.ceph.com/docs/master/rados/operations/erasure-code/, but 
>>>>>>>>> when I try create a pool with a specifc erasure-code-profile 
>>>>>>>>> ("myprofile") the PGs became on incomplete state.
>>>>>>>>>
>>>>>>>>> Anyone can help me?
>>>>>>>>>
>>>>>>>>> Below the profile I created:
>>>>>>>>> root@ceph0001:~# ceph osd erasure-code-profile get myprofile
>>>>>>>>> directory=/usr/lib/ceph/erasure-code
>>>>>>>>> k=6
>>>>>>>>> m=2
>>>>>>>>> plugin=jerasure
>>>>>>>>> technique=reed_sol_van
>>>>>>>>>
>>>>>>>>> The status of cluster:
>>>>>>>>> root@ceph0001:~# ceph health
>>>>>>>>> HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
>>>>>>>>> unclean
>>>>>>>>>
>>>>>>>>> health detail:
>>>>>>>>> root@ceph0001:~# ceph health detail
>>>>>>>>> HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs stuck 
>>>>>>>>> unclean
>>>>>>>>> pg 2.9 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.8 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
>>>>>>>>> pg 2.b is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.a is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
>>>>>>>>> pg 2.5 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.4 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
>>>>>>>>> pg 2.7 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.6 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.1 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
>>>>>>>>> pg 2.0 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.3 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
>>>>>>>>> pg 2.2 is stuck inactive since forever, current state incomplete, 
>>>>>>>>> last acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
>>>>>>>>> pg 2.9 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.8 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
>>>>>>>>> pg 2.b is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.a is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
>>>>>>>>> pg 2.5 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.4 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
>>>>>>>>> pg 2.7 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.6 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.1 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
>>>>>>>>> pg 2.0 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
>>>>>>>>> pg 2.3 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
>>>>>>>>> pg 2.2 is stuck unclean since forever, current state incomplete, last 
>>>>>>>>> acting [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
>>>>>>>>> pg 2.9 is incomplete, acting 
>>>>>>>>> [4,10,15,2147483647,3,2147483647,2147483647,2147483647] (reducing 
>>>>>>>>> pool ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.8 is incomplete, acting 
>>>>>>>>> [0,2147483647,4,2147483647,10,2147483647,15,2147483647] (reducing 
>>>>>>>>> pool ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.b is incomplete, acting 
>>>>>>>>> [8,3,14,2147483647,5,2147483647,2147483647,2147483647] (reducing pool 
>>>>>>>>> ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.a is incomplete, acting 
>>>>>>>>> [11,7,2,2147483647,2147483647,2147483647,15,2147483647] (reducing 
>>>>>>>>> pool ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.5 is incomplete, acting 
>>>>>>>>> [12,8,5,1,2147483647,2147483647,2147483647,2147483647] (reducing pool 
>>>>>>>>> ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.4 is incomplete, acting 
>>>>>>>>> [5,2147483647,13,1,2147483647,2147483647,8,2147483647] (reducing pool 
>>>>>>>>> ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.7 is incomplete, acting 
>>>>>>>>> [12,2,10,7,2147483647,2147483647,2147483647,2147483647] (reducing 
>>>>>>>>> pool ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.6 is incomplete, acting 
>>>>>>>>> [9,15,2147483647,4,2,2147483647,2147483647,2147483647] (reducing pool 
>>>>>>>>> ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.1 is incomplete, acting 
>>>>>>>>> [2,4,2147483647,13,2147483647,10,2147483647,2147483647] (reducing 
>>>>>>>>> pool ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.0 is incomplete, acting 
>>>>>>>>> [14,1,2147483647,4,10,2147483647,2147483647,2147483647] (reducing 
>>>>>>>>> pool ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.3 is incomplete, acting 
>>>>>>>>> [14,11,6,2147483647,2147483647,2147483647,2,2147483647] (reducing 
>>>>>>>>> pool ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>> pg 2.2 is incomplete, acting 
>>>>>>>>> [13,5,11,2147483647,2147483647,3,2147483647,2147483647] (reducing 
>>>>>>>>> pool ecpool min_size from 6 may help; search ceph.com/docs 
>>>>>>>>> <http://ceph.com/docs> for 'incomplete')
>>>>>>>>>
>>>>>>>>> At.
>>>>>>>>>
>>>>>>>>> *Italo Santos*
>>>>>>>>> http://italosantos.com.br/
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> ceph-users mailing list
>>>>>>>>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>>
>>>>>>>> -- 
>>>>>>>> Loïc Dachary, Artisan Logiciel Libre
>>>>>>
>>>>>> -- 
>>>>>> Loïc Dachary, Artisan Logiciel Libre
>>>>
>>>> -- 
>>>> Loïc Dachary, Artisan Logiciel Libre
>>
>> -- 
>> Loïc Dachary, Artisan Logiciel Libre
> 

-- 
Loïc Dachary, Artisan Logiciel Libre

Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to