Try " osd crush chooseleaf type = 0" in /etc/ceph/<clustername>.conf


Regards,
CY.

-----Original Message-----
From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Wah 
Peng
Sent: 2015年10月29日 9:14
To: Robert LeBlanc
Cc: Lindsay Mathieson; Gurjar, Unmesh; ceph-users@lists.ceph.com
Subject: Re: [ceph-users] creating+incomplete issues

wow this sounds hard to me. can you show the details?
thanks a lot.


On 2015/10/29 星期四 9:01, Robert LeBlanc wrote:
> You need to change the CRUSH map to select osd instead of host.
>
> Robert LeBlanc
>
> Sent from a mobile device please excuse any typos.
>
> On Oct 28, 2015 7:00 PM, "Wah Peng" <wah_p...@yahoo.com.sg 
> <mailto:wah_p...@yahoo.com.sg>> wrote:
>
>     $ ceph osd tree
>     # id    weight  type name       up/down reweight
>     -1      0.24    root default
>     -2      0.24            host ceph2
>     0       0.07999                 osd.0   up      1
>     1       0.07999                 osd.1   up      1
>     2       0.07999                 osd.2   up      1
>
>
>     On 2015/10/29 星期四 8:55, Robert LeBlanc wrote:
>
>         Please paste 'ceph osd tree'.
>
>         Robert LeBlanc
>
>         Sent from a mobile device please excuse any typos.
>
>         On Oct 28, 2015 6:54 PM, "Wah Peng" <wah_p...@yahoo.com.sg
>         <mailto:wah_p...@yahoo.com.sg>
>         <mailto:wah_p...@yahoo.com.sg <mailto:wah_p...@yahoo.com.sg>>>
>         wrote:
>
>              Hello,
>
>              Just did it, but still no good health. can you help? thanks.
>
>              ceph@ceph:~/my-cluster$ ceph osd stat
>                    osdmap e24: 3 osds: 3 up, 3 in
>
>              ceph@ceph:~/my-cluster$ ceph health
>              HEALTH_WARN 89 pgs degraded; 67 pgs incomplete; 67 pgs stuck
>              inactive; 192 pgs stuck unclean
>
>
>              On 2015/10/29 星期四 8:38, Lindsay Mathieson wrote:
>
>
>                  On 29 October 2015 at 10:29, Wah Peng
>         <wah_p...@yahoo.com.sg <mailto:wah_p...@yahoo.com.sg>
>                  <mailto:wah_p...@yahoo.com.sg
>         <mailto:wah_p...@yahoo.com.sg>>
>                  <mailto:wah_p...@yahoo.com.sg
>         <mailto:wah_p...@yahoo.com.sg> <mailto:wah_p...@yahoo.com.sg
>         <mailto:wah_p...@yahoo.com.sg>>>>
>                  wrote:
>
>                       $ ceph osd stat
>                             osdmap e18: 2 osds: 2 up, 2 in
>
>                       this is what it shows.
>                       does it mean I need to add up to 3 osds? I just
>         use  the
>                  default setup.
>
>
>                  If you went with the defaults then your pool size will
>         be 3,
>                  meaning it
>                  needs 3 copies of the data (replica 3) to be valid - as
>         you only
>                  have
>                  two nodes/osd's that can never happen :)
>
>                  Your options are:
>                  - Add another node and osd.
>                  or
>                  - reduce the size to 2.(ceph osd set <poolname> size 
> 2)
>
>
>
>                  --
>                  Lindsay
>
>              _______________________________________________
>              ceph-users mailing list
>         ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>         <mailto:ceph-users@lists.ceph.com
>         <mailto:ceph-users@lists.ceph.com>>
>         http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to