Hi,
I am deploying firefly version on CentOs 6.4. I am following quick
installation instructions available at ceph.com.
Kernel version in CentOs 6.4 is 2.6.32-358.
I am using virtual machines for all the nodes. As per the setup, there
are one admin-node, one monitor node and two OSD nodes.
I have added four OSDs created from four scsi disks of 10 GB on both OSD
nodes, instead of just creating directory on OSD nodes as shown in that
quick installation section.
At the end when I run ceph health command, the website says it should
achieve active+clean state. But my cluster is not able to achieve it as
shown below:
[ceph@node1 ~]$ ceph health
HEALTH_WARN 192 pgs incomplete; 192 pgs stuck inactive; 192 pgs stuck
unclean
[ceph@node1 ~]$ ceph status
cluster 08c77eb5-4fa9-4d4a-938c-af812137cb2c
health HEALTH_WARN 192 pgs incomplete; 192 pgs stuck inactive; 192
pgs stuck unclean
monmap e1: 1 mons at {node1=172.17.35.17:6789/0}, election epoch
1, quorum 0 node1
osdmap e36: 8 osds: 8 up, 8 in
pgmap v95: 192 pgs, 3 pools, 0 bytes data, 0 objects
271 MB used, 40600 MB / 40871 MB avail
192 creating+incomplete
[ceph@node1 ~]$
[ceph@node1 ~]$ ceph osd tree
# id weight type name up/down reweight
-1 0 root default
-2 0 host node2
0 0 osd.0 up 1
1 0 osd.1 up 1
2 0 osd.2 up 1
3 0 osd.3 up 1
-3 0 host node3
4 0 osd.4 up 1
5 0 osd.5 up 1
6 0 osd.6 up 1
7 0 osd.7 up 1
[ceph@node1 ~]$
Please let me know if I am missing anything. Do I still need to do to
bring my Ceph cluster in HEALTH OK state.
Regards,
Pratik Rupala
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com