All:
I have a 16-OSD cluster running 0.48 (Argonaut), built from source.
I rebuilt the entire cluster on Sunday Evening 8-19-2012, and started some
rados testing.
I have a custom CRUSH map, that calls for the "rbd", "metadata" pools and a
custom pool called "SCSI" to be pulled from osd.0-11, while the "data" pool is
pulled from osd.12-15. While testing, I find that the cluster is putting data
where I want it to, with one exception: the SCSI pool is not storing data
evenly thoughout the osd.0-11. Through "df", I find that about every other osd
is seeing space utilization.
So, whether good or bad, I did a "ceph osd reweight-by-utilization", which did
improve the situation.
Now, after doing some more research in the mailing lists, I find that I should
have just let the cluster figure it out on its own.
All that to lead to the problem I'm having now, and, I wish to use this mistake
as a learning tool. My ceph status is this:
ceph -s
##
health HEALTH_WARN 377 pgs stale; 4 pgs stuck inactive; 377 pgs stuck stale;
948 pgs stuck unclean
monmap e1: 3 mons at
{a=10.9.181.10:6789/0,b=10.9.181.11:6789/0,c=10.9.181.12:6789/0}, election
epoch 2, quorum 0,1,2 a,b,c
osdmap e90: 16 osds: 16 up, 16 in
pgmap v5085: 3080 pgs: 4 creating, 1755 active+clean, 377
stale+active+clean, 944 active+remapped; 10175 MB data, 52057 MB used, 12244 GB
/ 12815 GB avail
mdsmap e16: 1/1/1 up {0=b=up:replay}, 2 up:standby
##
Side-affects: I can create and map any Rados pools. I cannot for the life of
me, write to them, format them, anything them. Making my entire cluster offline
to clients.
While I've parse and poured over the documentation, I really need experienced
help, just to know how to get Ceph to recover, and then allow for operation
again.
I've restarted each daemon individually several times, after which I've also
tried a complete stop and start of the cluster. After things settle, this
reveals the same ceph -s status as I've posted above.
Thanks for your time!
Ryan Nicholson
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html