Hi,
I've removed the rbd pool and created it again. It picked up my default
settings but there are still some problems.
After running "sudo ceph -s", the output is as follow:
cluster 0b9b05db-98fe-49e6-b12b-1cce0645c015
health HEALTH_WARN
512 pgs stuck inactive
512 pgs stuck unclean
monmap e1: 1 mons at {monitorOne=192.168.1.153:6789/0}
election epoch 1, quorum 0 monitorOne
osdmap e62: 8 osds: 8 up, 8 in
pgmap v219: 512 pgs, 1 pools, 0 bytes data, 0 objects
8460 MB used, 4162 GB / 4171 GB avail
512 creating
Ceph stucks in creating the pgs forever. Those pgs are stuck in inactive and
unclean. And the Ceph pg query hangs forever.
I googled this problem and didn't get a clue.
Is there anything I missed?
Any idea to help me?
--
Zhen Wang
At 2015-10-07 13:05:51, "Christian Balzer" <[email protected]> wrote:
>
>Hello,
>On Wed, 7 Oct 2015 12:57:58 +0800 (CST) wikison wrote:
>
>This is a very old bug, misfeature.
>And creeps up every week or so here, google is your friend.
>
>> Hi,
>> I have a cluster of one monitor and eight OSDs. These OSDs are running
>> on four hosts(each host has two OSDs). When I set up everything and
>> started Ceph, I got this: esta@monitorOne:~$ sudo ceph -s [sudo]
>> password for esta: cluster 0b9b05db-98fe-49e6-b12b-1cce0645c015
>> health HEALTH_WARN
>> 64 pgs stuck inactive
>> 64 pgs stuck unclean
>> too few PGs per OSD (8 < min 30)
>
>Those 3 lines tell you pretty much all there is wrong.
>You did (correctly) set the defaul pg and pgp nums to something sensible
>(512) in your ceph.conf.
>Unfortunately when creating the initial pool (rbd) it still ignores those
>settings.
>
>You could try to increase those for your pool, which may or may not work.
>
>The easier and faster way is to remove the rbd pool and create it again.
>This should pick up your default settings.
>
>Christian
>
>> monmap e1: 1 mons at {monitorOne=192.168.1.153:6789/0}
>> election epoch 1, quorum 0 monitorOne
>> osdmap e58: 8 osds: 8 up, 8 in
>> pgmap v191: 64 pgs, 1 pools, 0 bytes data, 0 objects
>> 8460 MB used, 4162 GB / 4171 GB avail
>> 64 creating
>>
>>
>> How to deal with this HEALTH_WARN status?
>> This is my ceph.conf:
>> [global]
>>
>>
>> fsid = 0b9b05db-98fe-49e6-b12b-1cce0645c015
>>
>>
>> mon initial members = monitorOne
>> mon host = 192.168.1.153
>> filestore_xattr_use_omap = true
>>
>>
>> public network = 192.168.1.0/24
>> cluster network = 10.0.0.0/24
>> pid file = /var/run/ceph/$name.pid
>>
>>
>> auth cluster required = cephx
>> auth service required = cephx
>> auth client required = cephx
>>
>>
>> osd pool default size = 3
>> osd pool default min size = 2
>> osd pool default pg num = 512
>> osd pool default pgp num = 512
>> osd crush chooseleaf type = 1
>> osd journal size = 1024
>>
>>
>> [mon]
>>
>>
>> [mon.0]
>> host = monitorOne
>> mon addr = 192.168.1.153:6789
>>
>>
>> [osd]
>>
>>
>> [osd.0]
>> host = storageOne
>>
>>
>> [osd.1]
>> host = storageTwo
>>
>>
>> [osd.2]
>> host = storageFour
>>
>>
>> [osd.3]
>> host = storageLast
>>
>>
>> Could anybody help me?
>>
>> best regards,
>>
>> --
>>
>> Zhen Wang
>
>--
>Christian Balzer Network/Systems Engineer
>[email protected] Global OnLine Japan/Fusion Communications
>http://www.gol.com/
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com