On 01/09/2013 08:59 AM, Wido den Hollander wrote:
> Hi,
> 
> On 01/09/2013 01:53 AM, Chen, Xiaoxi wrote:
>> Hi,
>>      Setting rep size to 3 only make the data triple-replication, that means 
>> when you "fail" all OSDs in 2 out of 3 DCs, the data still accessable.
>>      But Monitor is another story, for monitor clusters with 2N+1 nodes, it 
>> require at least N+1 nodes alive, and indeed this is why you Ceph failed.
>>      It looks to me this discipline make it hard to design a proper 
>> deployment which is robust in DC outage. But hoping for inputs from 
>> community,how to make Monitor cluster reliable.
>>
> 
>  From what I understand he didn't kill the second mon, still leaving 2
> out of 3 mons running.

Indeed. A good hint that this is the case is this bit of Shawn's message:

>> When I fail a datacenter (including 1 of 3 mon's) I eventually get:
>> 2013-01-08 13:58:54.020477 mon.0 [INF] pgmap v2712139: 7104 pgs: 7104 
>> active+degraded; 60264 MB data, 137 GB used, 13570 GB / 14146 GB avail; 
>> 16362/49086 degraded (33.333%)
>>
>> At this point everything is still ok.  But when I fail the 2nd datacenter 
>> (still leaving 2 out of 3 mons running) I get:
>> 2013-01-08 14:01:25.600056 mon.0 [INF] pgmap v2712189: 7104 pgs: 7104 
>> incomplete; 60264 MB data, 137 GB used, 13570 GB / 14146 GB avail

If you still manage to get these messages, it means your monitors are
still handling and answering requests, and that only happens when you
have a quorum :)

  -Joao
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to