Hi! Maybe you should check your network.Does the network of your mon to each
other is ok. Then do you only run ceph in the mon host.Is there any others
program run and waste the cpu of your host. If that is all ok , you can try
expend the expire time for each mon to lease timeout.This may reduce the new
elelction for
each mon.
原始邮件
发件人: <[email protected]>
收件人: <[email protected]>
日 期 :2017年03月02日 08:32
主 题 :Re: [ceph-users] ceph-users Digest, Vol 50, Issue 1
Thank you for your response. :)
Version was Jewel - 10.2.2. And, yes I did restart the monitors with no
change in results.
For the record, here's the problem. It was a multi-pool cluster, and
the Crush rules had an inappropriately large number for the step
chooseleaf line. I won't get into details because it would raise more
questions than it would answer. But enough OSDs went down to result in
there being no solution for some PG groups. Neverthess the monitors
continued searching for a solution (based on the chooseleaf parameter
applied to each placement group).
The monitors were spending all their time running crush and then calling
an election when a timeout expired.
That's the cause of the problem. Maybe I'll post a solution if and when
we get out of this state.
Sadly, my fault. It might be nice to get a warning when you try to do
something really stupid like that.
On 03/01/2017 03:02 PM, [email protected] wrote:
Date: Tue, 28 Feb 2017 23:52:26 +0000 From: Joao Eduardo Luis <[email protected]>
To: [email protected] Subject: Re: [ceph-users] monitors at 100%
cluster out of service Message-ID:
<[email protected]> Content-Type: text/plain
charset=windows-1252 format=flowed On 02/28/2017 09:53 PM, WRIGHT, JON R (JON
R) wrote: > I currently have a situation where the monitors are running at
100% CPU, > and can't run any commands because authentication times out after
300 > seconds. > > I stopped the leader, and the resulting election picked a
new leader, > but that monitor shows exactly the same behavor. > > Now both
monitors *think* they are the leader and call new elections > against the third
monitor, both winning each time. Essentially they > alternate between calling
an election (which they win) and then pegging > one of the CPUs at 100%. > >
strace suggests that the monitor daemons are spending the "pegged" time > in
user space, and attaching a debugger to the running process suggests > that the
monitor is spending its time doing crushmap calculations in > fn_monstore. > >
Setting paxos_debug to 10 produces this log message: > > 2017-02-28
16:50:49.503712 7f218ccd4700 7 >
mon.hlxkvm001-storage@0(leader).paxosservice(osdmap 1252..1873) _active >
creating new pending > > during the time when the monitor process is pegged at
100%. > > The problem started when one of the hosts running a peon was
rebooted, > but didn't have the correct mtu setting in /etc/network/interfaces.
> The problem showed up after correcting the mtu value. > > Also, we are using
a hyperconverged architecture where the same host > runs a monitor and multiple
OSDs. > > Any thoughts on recovery would be greatly appreciated. What
version is this? How many monitors are you running? Are the monitors
consuming an unusual amount of memory? What about the OSDs in the same nodes?
Is the size of the monitor stores abnormally high? Have you tried restarting
all monitors and see if they hit the same issue? -Joao_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com