Understood.
Thanks for your help, the cluster is healthy now :D

Also, using for example k=6,m=1 and failure domain by host I’ll be able lose 
all OSD on the same host, but if a lose 2 disks on different hosts I can lose 
data right? So, it is possible been a failure domain which allow me to lose an 
OSD or a host?  

Regards.

Italo Santos
http://italosantos.com.br/


On Wednesday, December 17, 2014 at 4:27 PM, Loic Dachary wrote:

>  
>  
> On 17/12/2014 19:22, Italo Santos wrote:
> > Loic,
> >  
> > So, if want have a failure domain by host, I’ll need set up a erasure 
> > profile which k+m = total number of hosts I have, right?
>  
> Yes, k+m has to be <= number of hosts.
>  
> >  
> > Regards.
> >  
> > *Italo Santos*
> > http://italosantos.com.br/
> >  
> > On Wednesday, December 17, 2014 at 3:24 PM, Loic Dachary wrote:
> >  
> > >  
> > >  
> > > On 17/12/2014 18:18, Italo Santos wrote:
> > > > Hello,
> > > >  
> > > > I’ve take a look to this documentation (which help a lot) and if I 
> > > > understand right, when I set a profile like:
> > > >  
> > > > ===
> > > > ceph osd erasure-code-profile set isilon k=8 m=2 
> > > > ruleset-failure-domain=host
> > > > ===
> > > >  
> > > > And create a pool following the recommendations on doc, I’ll need 
> > > > (100*16)/2 = 800 PGs, I’ll need the sufficient number of hosts to 
> > > > support create total PGs?
> > >  
> > > You will need k+m = 10 host per OSD. If you only have 10 hosts that 
> > > should be ok and the 800 PGs will use these 10 OSD in various orders. It 
> > > also means that you will end up having 800 PG per OSD which is a bit too 
> > > mche. If you have 20 OSDs that will be better : each PG will get 10 OSD 
> > > out of 20 and each OSD will have 400 PGs. Ideally you want the number of 
> > > PG per OSD to be in the range (approximately) [20,300].
> > >  
> > > Cheers
> > >  
> > > >  
> > > > Regards.
> > > >  
> > > > *Italo Santos*
> > > > http://italosantos.com.br/
> > > >  
> > > > On Wednesday, December 17, 2014 at 2:42 PM, Loic Dachary wrote:
> > > >  
> > > > > Hi,
> > > > >  
> > > > > Thanks for the update : good news are much appreciated :-) Would you 
> > > > > have time to review the documentation at 
> > > > > https://github.com/ceph/ceph/pull/3194/files ? It was partly 
> > > > > motivated by the problem you had.
> > > > >  
> > > > > Cheers
> > > > >  
> > > > > On 17/12/2014 14:03, Italo Santos wrote:
> > > > > > Hello Loic,
> > > > > >  
> > > > > > Thanks for you help, I’ve take a look to my crush map and I replace 
> > > > > > "step chooseleaf indep 0 type osd” by "step choose indep 0 type 
> > > > > > osd” and all PGs was created successfully.
> > > > > >  
> > > > > > At.
> > > > > >  
> > > > > > *Italo Santos*
> > > > > > http://italosantos.com.br/
> > > > > >  
> > > > > > On Tuesday, December 16, 2014 at 8:39 PM, Loic Dachary wrote:
> > > > > >  
> > > > > > > Hi,
> > > > > > >  
> > > > > > > The 2147483647 means that CRUSH did not find enough OSD for a 
> > > > > > > given PG. If you check the crush rule associated with the erasure 
> > > > > > > coded pool, you will most probably find why.
> > > > > > >  
> > > > > > > Cheers
> > > > > > >  
> > > > > > > On 16/12/2014 23:32, Italo Santos wrote:
> > > > > > > > Hello,
> > > > > > > >  
> > > > > > > > I'm trying to create an erasure pool following 
> > > > > > > > http://docs.ceph.com/docs/master/rados/operations/erasure-code/,
> > > > > > > >  but when I try create a pool with a specifc 
> > > > > > > > erasure-code-profile ("myprofile") the PGs became on incomplete 
> > > > > > > > state.
> > > > > > > >  
> > > > > > > > Anyone can help me?
> > > > > > > >  
> > > > > > > > Below the profile I created:
> > > > > > > > root@ceph0001:~# ceph osd erasure-code-profile get myprofile
> > > > > > > > directory=/usr/lib/ceph/erasure-code
> > > > > > > > k=6
> > > > > > > > m=2
> > > > > > > > plugin=jerasure
> > > > > > > > technique=reed_sol_van
> > > > > > > >  
> > > > > > > > The status of cluster:
> > > > > > > > root@ceph0001:~# ceph health
> > > > > > > > HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs 
> > > > > > > > stuck unclean
> > > > > > > >  
> > > > > > > > health detail:
> > > > > > > > root@ceph0001:~# ceph health detail
> > > > > > > > HEALTH_WARN 12 pgs incomplete; 12 pgs stuck inactive; 12 pgs 
> > > > > > > > stuck unclean
> > > > > > > > pg 2.9 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.8 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
> > > > > > > > pg 2.b is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.a is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
> > > > > > > > pg 2.5 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.4 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
> > > > > > > > pg 2.7 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.6 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.1 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
> > > > > > > > pg 2.0 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.3 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
> > > > > > > > pg 2.2 is stuck inactive since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
> > > > > > > > pg 2.9 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [4,10,15,2147483647,3,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.8 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [0,2147483647,4,2147483647,10,2147483647,15,2147483647]
> > > > > > > > pg 2.b is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [8,3,14,2147483647,5,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.a is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [11,7,2,2147483647,2147483647,2147483647,15,2147483647]
> > > > > > > > pg 2.5 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [12,8,5,1,2147483647,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.4 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [5,2147483647,13,1,2147483647,2147483647,8,2147483647]
> > > > > > > > pg 2.7 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [12,2,10,7,2147483647,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.6 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [9,15,2147483647,4,2,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.1 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [2,4,2147483647,13,2147483647,10,2147483647,2147483647]
> > > > > > > > pg 2.0 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [14,1,2147483647,4,10,2147483647,2147483647,2147483647]
> > > > > > > > pg 2.3 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [14,11,6,2147483647,2147483647,2147483647,2,2147483647]
> > > > > > > > pg 2.2 is stuck unclean since forever, current state 
> > > > > > > > incomplete, last acting 
> > > > > > > > [13,5,11,2147483647,2147483647,3,2147483647,2147483647]
> > > > > > > > pg 2.9 is incomplete, acting 
> > > > > > > > [4,10,15,2147483647,3,2147483647,2147483647,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.8 is incomplete, acting 
> > > > > > > > [0,2147483647,4,2147483647,10,2147483647,15,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.b is incomplete, acting 
> > > > > > > > [8,3,14,2147483647,5,2147483647,2147483647,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.a is incomplete, acting 
> > > > > > > > [11,7,2,2147483647,2147483647,2147483647,15,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.5 is incomplete, acting 
> > > > > > > > [12,8,5,1,2147483647,2147483647,2147483647,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.4 is incomplete, acting 
> > > > > > > > [5,2147483647,13,1,2147483647,2147483647,8,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.7 is incomplete, acting 
> > > > > > > > [12,2,10,7,2147483647,2147483647,2147483647,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.6 is incomplete, acting 
> > > > > > > > [9,15,2147483647,4,2,2147483647,2147483647,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.1 is incomplete, acting 
> > > > > > > > [2,4,2147483647,13,2147483647,10,2147483647,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.0 is incomplete, acting 
> > > > > > > > [14,1,2147483647,4,10,2147483647,2147483647,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.3 is incomplete, acting 
> > > > > > > > [14,11,6,2147483647,2147483647,2147483647,2,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > > pg 2.2 is incomplete, acting 
> > > > > > > > [13,5,11,2147483647,2147483647,3,2147483647,2147483647] 
> > > > > > > > (reducing pool ecpool min_size from 6 may help; search 
> > > > > > > > ceph.com/docs <http://ceph.com/docs> for 'incomplete')
> > > > > > > >  
> > > > > > > > At.
> > > > > > > >  
> > > > > > > > *Italo Santos*
> > > > > > > > http://italosantos.com.br/
> > > > > > > >  
> > > > > > > >  
> > > > > > > >  
> > > > > > > > _______________________________________________
> > > > > > > > ceph-users mailing list
> > > > > > > > ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> > > > > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > > > > > >  
> > > > > > >  
> > > > > > >  
> > > > > > > --  
> > > > > > > Loïc Dachary, Artisan Logiciel Libre
> > > > > > >  
> > > > > >  
> > > > > >  
> > > > >  
> > > > >  
> > > > > --  
> > > > > Loïc Dachary, Artisan Logiciel Libre
> > > > >  
> > > >  
> > > >  
> > >  
> > >  
> > > --  
> > > Loïc Dachary, Artisan Logiciel Libre
> > >  
> >  
> >  
>  
>  
> --  
> Loïc Dachary, Artisan Logiciel Libre
>  
>  


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to