Hello, 

Problem1:

Ceph health says HEALTH_WARN clock skew detected on mon.b, mon.c

But:

root@hcmonko1:~# date && ssh hcmonko3 date && ssh hcmonko2 date
Thu May 23 18:19:38 CEST 2013
Thu May 23 18:19:38 CEST 2013
Thu May 23 18:19:38 CEST 2013

Seems to be ok, ntp configured. Restarted all the servers, still the same.


Problem 2:

Service ceph -a works seriell, not parallel.
Situation: every osd and mon is offline via service ceph -a stop. When I want 
to bring them online with "service ceph -a start", it begins with mon a, then 
mon b, then osd0, osd1 etc.
But the procedure hangs at the point ceph-create keys for the first mon in the 
row.

root@hcosdko1:~# service ceph -a start
=== mon.a ===
Starting Ceph mon.a on hcmonko1...
Starting ceph-create-keys on hcmonko1...

I get to know via your irc-channel that ceph-create-keys needs the mon quorum 
to proceed. I think the ceph -a start script is designed to not make it serial, 
rather it has to do it parallel for all mons+osd.
But in my scenario it hangs at this point: ceph is started on mon1, hangs at 
the keys because there is no quorum => ceph will not start on mon2,mon3, osd0 
and so on. (when I type "service ceph -a start on mon1, it will begin on the 
host itself, create the keys, proceed with starting ceph on mon2 and can create 
the keys there, because there is a quorum because ceph is started already on 
mon1).


Question1
For testing we have only 1 osd node at this moment, there will follow 2 other 
nodes, perhaps we will add certain osd hdds in one server/node.
when i want to add a node or only an osd.. is it right, that i change the 
ceph.conf, copy it via ssh to each node(mon,osd,..) and then make ceph -a start 
to reload the new config, so the new servers,osds,mons etc are activated?
When I wants to add an osd (hdd), I have to type makecephfs --mkfs.. will it 
destroy the data on the already existing osd-hdds?

Question2
when i have to restart 1 server, i think the ceph mechanism will start 
replicating the data to the other hosts directly. is it possible to define a 
server/osd for maintenance mode, so rados does not begin to start replicating 
the data, only because 1 osd is restarting?


Thank you very much!

Regards
Philipp



--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to