Re: [ceph-users] Testing CephFS
> On Sep 1, 2015, at 16:13, Simon Hallam <s...@pml.ac.uk> wrote: > > Hi Greg, Zheng, > > Is this fixed in a later version of the kernel client? Or would it be wise > for us to start using the fuse client? > > Cheers, I just wrote a fix https://github.com/ceph/ceph-client/commit/33b68dde7f27927a7cb1a7691e3c5b6f847ffd14 <https://github.com/ceph/ceph-client/commit/33b68dde7f27927a7cb1a7691e3c5b6f847ffd14>. Yes, you should try ceps-fuse if this bug causes problems for you. Regards Yan, Zheng > > Simon > >> -Original Message- >> From: Gregory Farnum [mailto:gfar...@redhat.com] >> Sent: 31 August 2015 13:02 >> To: Yan, Zheng >> Cc: Simon Hallam; Zheng Yan; ceph-users@lists.ceph.com >> Subject: Re: [ceph-users] Testing CephFS >> >> On Mon, Aug 31, 2015 at 12:16 PM, Yan, Zheng <uker...@gmail.com> wrote: >>> On Mon, Aug 24, 2015 at 6:38 PM, Gregory Farnum >> <gfar...@redhat.com> wrote: >>>> On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam <s...@pml.ac.uk> >> wrote: >>>>> Hi Greg, >>>>> >>>>> The MDS' detect that the other one went down and started the replay. >>>>> >>>>> I did some further testing with 20 client machines. Of the 20 client >> machines, 5 hung with the following error: >>>>> >>>>> [Aug24 10:53] ceph: mds0 caps stale >>>>> [Aug24 10:54] ceph: mds0 caps stale >>>>> [Aug24 10:58] ceph: mds0 hung >>>>> [Aug24 11:03] ceph: mds0 came back >>>>> [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state >> OPEN) >>>>> [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new >> mon >>>>> [Aug24 11:04] ceph: mds0 reconnect start >>>>> [ +0.084938] libceph: mon2 10.15.0.3:6789 session established >>>>> [ +0.008475] ceph: mds0 reconnect denied >>>> >>>> Oh, this might be a kernel bug, failing to ask for mdsmap updates when >>>> the connection goes away. Zheng, does that sound familiar? >>>> -Greg >>>> >>> >>> I reproduced this locally (use SIGSTOP to stop the monitor) . I think >>> the root cause is that kernel client does not implement >>> CEPH_FEATURE_MSGR_KEEPALIVE2. So the kernel client couldn't reliably >>> detect the event that network cable got unplugged. It kept waiting for >>> new events from the disconnected connection. >> >> Yeah, the userspace client maintains an ongoing MDSMap subscription >> from the monitors in order to hear about this. It puts more load on >> the monitors but right now that's the solution we're going with: the >> monitor times out the MDS, publishes a series of new maps (pushed to >> the clients) in order to activate a standby, and the clients see that >> they need to connect to the new MDS instance. >> -Greg > > > Please visit our new website at www.pml.ac.uk and follow us on Twitter > @PlymouthMarine > > Winner of the Environment & Conservation category, the Charity Awards 2014. > > Plymouth Marine Laboratory (PML) is a company limited by guarantee registered > in England & Wales, company number 4178503. Registered Charity No. 1091222. > Registered Office: Prospect Place, The Hoe, Plymouth PL1 3DH, UK. > > This message is private and confidential. If you have received this message > in error, please notify the sender and remove it from your system. You are > reminded that e-mail communications are not secure and may contain viruses; > PML accepts no liability for any loss or damage which may be caused by > viruses. > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Testing CephFS
Hi Greg, Zheng, Is this fixed in a later version of the kernel client? Or would it be wise for us to start using the fuse client? Cheers, Simon > -Original Message- > From: Gregory Farnum [mailto:gfar...@redhat.com] > Sent: 31 August 2015 13:02 > To: Yan, Zheng > Cc: Simon Hallam; Zheng Yan; ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Testing CephFS > > On Mon, Aug 31, 2015 at 12:16 PM, Yan, Zheng <uker...@gmail.com> wrote: > > On Mon, Aug 24, 2015 at 6:38 PM, Gregory Farnum > <gfar...@redhat.com> wrote: > >> On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam <s...@pml.ac.uk> > wrote: > >>> Hi Greg, > >>> > >>> The MDS' detect that the other one went down and started the replay. > >>> > >>> I did some further testing with 20 client machines. Of the 20 client > machines, 5 hung with the following error: > >>> > >>> [Aug24 10:53] ceph: mds0 caps stale > >>> [Aug24 10:54] ceph: mds0 caps stale > >>> [Aug24 10:58] ceph: mds0 hung > >>> [Aug24 11:03] ceph: mds0 came back > >>> [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state > OPEN) > >>> [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new > mon > >>> [Aug24 11:04] ceph: mds0 reconnect start > >>> [ +0.084938] libceph: mon2 10.15.0.3:6789 session established > >>> [ +0.008475] ceph: mds0 reconnect denied > >> > >> Oh, this might be a kernel bug, failing to ask for mdsmap updates when > >> the connection goes away. Zheng, does that sound familiar? > >> -Greg > >> > > > > I reproduced this locally (use SIGSTOP to stop the monitor) . I think > > the root cause is that kernel client does not implement > > CEPH_FEATURE_MSGR_KEEPALIVE2. So the kernel client couldn't reliably > > detect the event that network cable got unplugged. It kept waiting for > > new events from the disconnected connection. > > Yeah, the userspace client maintains an ongoing MDSMap subscription > from the monitors in order to hear about this. It puts more load on > the monitors but right now that's the solution we're going with: the > monitor times out the MDS, publishes a series of new maps (pushed to > the clients) in order to activate a standby, and the clients see that > they need to connect to the new MDS instance. > -Greg Please visit our new website at www.pml.ac.uk and follow us on Twitter @PlymouthMarine Winner of the Environment & Conservation category, the Charity Awards 2014. Plymouth Marine Laboratory (PML) is a company limited by guarantee registered in England & Wales, company number 4178503. Registered Charity No. 1091222. Registered Office: Prospect Place, The Hoe, Plymouth PL1 3DH, UK. This message is private and confidential. If you have received this message in error, please notify the sender and remove it from your system. You are reminded that e-mail communications are not secure and may contain viruses; PML accepts no liability for any loss or damage which may be caused by viruses. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Testing CephFS
On Mon, Aug 31, 2015 at 12:16 PM, Yan, Zhengwrote: > On Mon, Aug 24, 2015 at 6:38 PM, Gregory Farnum wrote: >> On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam wrote: >>> Hi Greg, >>> >>> The MDS' detect that the other one went down and started the replay. >>> >>> I did some further testing with 20 client machines. Of the 20 client >>> machines, 5 hung with the following error: >>> >>> [Aug24 10:53] ceph: mds0 caps stale >>> [Aug24 10:54] ceph: mds0 caps stale >>> [Aug24 10:58] ceph: mds0 hung >>> [Aug24 11:03] ceph: mds0 came back >>> [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state OPEN) >>> [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new mon >>> [Aug24 11:04] ceph: mds0 reconnect start >>> [ +0.084938] libceph: mon2 10.15.0.3:6789 session established >>> [ +0.008475] ceph: mds0 reconnect denied >> >> Oh, this might be a kernel bug, failing to ask for mdsmap updates when >> the connection goes away. Zheng, does that sound familiar? >> -Greg >> > > I reproduced this locally (use SIGSTOP to stop the monitor) . I think > the root cause is that kernel client does not implement > CEPH_FEATURE_MSGR_KEEPALIVE2. So the kernel client couldn't reliably > detect the event that network cable got unplugged. It kept waiting for > new events from the disconnected connection. Yeah, the userspace client maintains an ongoing MDSMap subscription from the monitors in order to hear about this. It puts more load on the monitors but right now that's the solution we're going with: the monitor times out the MDS, publishes a series of new maps (pushed to the clients) in order to activate a standby, and the clients see that they need to connect to the new MDS instance. -Greg ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Testing CephFS
After further testing, it does appear to be completely random (eg, not the same clients are affected). I have one machine running the latest kernel/client which hasn't hung yet. [root@f23-alpha ~]# uname -a Linux f23-alpha.npm.ac.uk 4.2.0-0.rc5.git0.2.fc23.x86_64 #1 SMP Tue Aug 4 01:37:40 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux [root@f23-alpha ~]# ceph -v ceph version 0.94.2 (5fb85614ca8f354284c713a2f9c610860720bbf3) Cheers, Simon -Original Message- From: Gregory Farnum [mailto:gfar...@redhat.com] Sent: 27 August 2015 14:24 To: Yan, Zheng Cc: Simon Hallam; ceph-users@lists.ceph.com Subject: Re: [ceph-users] Testing CephFS On Mon, Aug 24, 2015 at 12:53 PM, Simon Hallam s...@pml.ac.uk wrote: The clients are: [root@gridnode50 ~]# uname -a Linux gridnode50 4.0.8-200.fc21.x86_64 #1 SMP Fri Jul 10 21:09:54 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux [root@gridnode50 ~]# ceph -v ceph version 0.80.10 (ea6c958c38df1216bf95c927f143d8b13c4a9e70) I don't think it is a reconnect timeout, as they don't even attempt to reconnect until I plug the Ethernet cable back into the original MDS? Right, the default timeout is going to be 15 seconds. But they clients should be getting MDSMap updates from the monitor that tell them the MDS has failed over. It looks like they're not timing out, then when the MDS *does* come back it tells them about its own death. Is that possible, Zheng? -Greg Cheers, Simon -Original Message- From: Yan, Zheng [mailto:z...@redhat.com] Sent: 24 August 2015 12:28 To: Simon Hallam Cc: ceph-users@lists.ceph.com; Gregory Farnum Subject: Re: [ceph-users] Testing CephFS On Aug 24, 2015, at 18:38, Gregory Farnum gfar...@redhat.com wrote: On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam s...@pml.ac.uk wrote: Hi Greg, The MDS' detect that the other one went down and started the replay. I did some further testing with 20 client machines. Of the 20 client machines, 5 hung with the following error: [Aug24 10:53] ceph: mds0 caps stale [Aug24 10:54] ceph: mds0 caps stale [Aug24 10:58] ceph: mds0 hung [Aug24 11:03] ceph: mds0 came back [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state OPEN) [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new mon [Aug24 11:04] ceph: mds0 reconnect start [ +0.084938] libceph: mon2 10.15.0.3:6789 session established [ +0.008475] ceph: mds0 reconnect denied Oh, this might be a kernel bug, failing to ask for mdsmap updates when the connection goes away. Zheng, does that sound familiar? -Greg This seems like reconnect timeout. you can try enlarging mds_reconnect_timeout config option. Which version of kernel are you using? Yan, Zheng 10.15.0.3 was the active MDS at the time I unplugged the Ethernet cable. This was the output of ceph -w as I ran the test (I've removed a lot of the pg remapping): 2015-08-24 11:02:39.547529 mon.1 [INF] mon.ceph2 calling new monitor election 2015-08-24 11:02:40.011995 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:02:45.245869 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1 2015-08-24 11:02:45.257440 mon.0 [INF] HEALTH_WARN; 1 mons down, quorum 0,1 ceph1,ceph2 2015-08-24 11:02:45.535369 mon.0 [INF] monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} 2015-08-24 11:02:45.535444 mon.0 [INF] pgmap v15803: 8256 pgs: 8256 active+clean; 1248 GB data, 2503 GB used, 193 TB / 196 TB avail; 47 B/s wr, 0 op/s 2015-08-24 11:02:45.535541 mon.0 [INF] mdsmap e38: 1/1/1 up {0=ceph3=up:active}, 2 up:standby 2015-08-24 11:02:45.535629 mon.0 [INF] osdmap e197: 36 osds: 36 up, 36 in 2015-08-24 11:03:01.946397 mon.0 [INF] mdsmap e39: 1/1/1 up {0=ceph2=up:replay}, 1 up:standby 2015-08-24 11:03:02.993880 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:reconnect 2015-08-24 11:03:02.993930 mon.0 [INF] mdsmap e40: 1/1/1 up {0=ceph2=up:reconnect}, 1 up:standby 2015-08-24 11:03:51.461248 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:rejoin 2015-08-24 11:03:55.807131 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:03:55.807195 mon.0 [INF] mdsmap e42: 1/1/1 up {0=ceph2=up:active}, 1 up:standby 2015-08-24 11:06:48.036736 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:06:48.036799 mon.0 [INF] mdsmap e43: 1/1/1 up {0=ceph2=up:active}, 1 up:standby *cable plugged back in* 2015-08-24 11:13:13.230714 mon.0 [INF] osd.32 10.15.0.3:6832/11565 boot 2015-08-24 11:13:13.230765 mon.0 [INF] osdmap e212: 36 osds: 25 up, 25 in 2015-08-24 11:13:13.230809 mon.0 [INF] mds.? 10.15.0.3:6833/16993 up:boot 2015-08-24 11:13:13.230837 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:30.799429 mon.2 [INF] mon.ceph3 calling new monitor election
Re: [ceph-users] Testing CephFS
On Mon, Aug 24, 2015 at 12:53 PM, Simon Hallam s...@pml.ac.uk wrote: The clients are: [root@gridnode50 ~]# uname -a Linux gridnode50 4.0.8-200.fc21.x86_64 #1 SMP Fri Jul 10 21:09:54 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux [root@gridnode50 ~]# ceph -v ceph version 0.80.10 (ea6c958c38df1216bf95c927f143d8b13c4a9e70) I don't think it is a reconnect timeout, as they don't even attempt to reconnect until I plug the Ethernet cable back into the original MDS? Right, the default timeout is going to be 15 seconds. But they clients should be getting MDSMap updates from the monitor that tell them the MDS has failed over. It looks like they're not timing out, then when the MDS *does* come back it tells them about its own death. Is that possible, Zheng? -Greg Cheers, Simon -Original Message- From: Yan, Zheng [mailto:z...@redhat.com] Sent: 24 August 2015 12:28 To: Simon Hallam Cc: ceph-users@lists.ceph.com; Gregory Farnum Subject: Re: [ceph-users] Testing CephFS On Aug 24, 2015, at 18:38, Gregory Farnum gfar...@redhat.com wrote: On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam s...@pml.ac.uk wrote: Hi Greg, The MDS' detect that the other one went down and started the replay. I did some further testing with 20 client machines. Of the 20 client machines, 5 hung with the following error: [Aug24 10:53] ceph: mds0 caps stale [Aug24 10:54] ceph: mds0 caps stale [Aug24 10:58] ceph: mds0 hung [Aug24 11:03] ceph: mds0 came back [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state OPEN) [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new mon [Aug24 11:04] ceph: mds0 reconnect start [ +0.084938] libceph: mon2 10.15.0.3:6789 session established [ +0.008475] ceph: mds0 reconnect denied Oh, this might be a kernel bug, failing to ask for mdsmap updates when the connection goes away. Zheng, does that sound familiar? -Greg This seems like reconnect timeout. you can try enlarging mds_reconnect_timeout config option. Which version of kernel are you using? Yan, Zheng 10.15.0.3 was the active MDS at the time I unplugged the Ethernet cable. This was the output of ceph -w as I ran the test (I've removed a lot of the pg remapping): 2015-08-24 11:02:39.547529 mon.1 [INF] mon.ceph2 calling new monitor election 2015-08-24 11:02:40.011995 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:02:45.245869 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1 2015-08-24 11:02:45.257440 mon.0 [INF] HEALTH_WARN; 1 mons down, quorum 0,1 ceph1,ceph2 2015-08-24 11:02:45.535369 mon.0 [INF] monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} 2015-08-24 11:02:45.535444 mon.0 [INF] pgmap v15803: 8256 pgs: 8256 active+clean; 1248 GB data, 2503 GB used, 193 TB / 196 TB avail; 47 B/s wr, 0 op/s 2015-08-24 11:02:45.535541 mon.0 [INF] mdsmap e38: 1/1/1 up {0=ceph3=up:active}, 2 up:standby 2015-08-24 11:02:45.535629 mon.0 [INF] osdmap e197: 36 osds: 36 up, 36 in 2015-08-24 11:03:01.946397 mon.0 [INF] mdsmap e39: 1/1/1 up {0=ceph2=up:replay}, 1 up:standby 2015-08-24 11:03:02.993880 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:reconnect 2015-08-24 11:03:02.993930 mon.0 [INF] mdsmap e40: 1/1/1 up {0=ceph2=up:reconnect}, 1 up:standby 2015-08-24 11:03:51.461248 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:rejoin 2015-08-24 11:03:55.807131 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:03:55.807195 mon.0 [INF] mdsmap e42: 1/1/1 up {0=ceph2=up:active}, 1 up:standby 2015-08-24 11:06:48.036736 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:06:48.036799 mon.0 [INF] mdsmap e43: 1/1/1 up {0=ceph2=up:active}, 1 up:standby *cable plugged back in* 2015-08-24 11:13:13.230714 mon.0 [INF] osd.32 10.15.0.3:6832/11565 boot 2015-08-24 11:13:13.230765 mon.0 [INF] osdmap e212: 36 osds: 25 up, 25 in 2015-08-24 11:13:13.230809 mon.0 [INF] mds.? 10.15.0.3:6833/16993 up:boot 2015-08-24 11:13:13.230837 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:30.799429 mon.2 [INF] mon.ceph3 calling new monitor election 2015-08-24 11:13:30.826158 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:13:30.926331 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1,2 2015-08-24 11:13:30.968739 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:28.383203 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24155 10.10.10.95:0/3238635414 after 625.375507 (allowed interval 45) 2015-08-24 11:13:29.721653 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24146 10.10.10.99:0/3454703638 after 626.713952 (allowed interval 45) 2015-08-24 11:13:31.113004 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24140 10.10.10.60:0/359606080 after 628.105302 (allowed interval 45
Re: [ceph-users] Testing CephFS
Need to be more careful but probably you're right -; ./net/ceph/messenger.c Shinobu On Mon, Aug 24, 2015 at 8:53 PM, Simon Hallam s...@pml.ac.uk wrote: The clients are: [root@gridnode50 ~]# uname -a Linux gridnode50 4.0.8-200.fc21.x86_64 #1 SMP Fri Jul 10 21:09:54 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux [root@gridnode50 ~]# ceph -v ceph version 0.80.10 (ea6c958c38df1216bf95c927f143d8b13c4a9e70) I don't think it is a reconnect timeout, as they don't even attempt to reconnect until I plug the Ethernet cable back into the original MDS? Cheers, Simon -Original Message- From: Yan, Zheng [mailto:z...@redhat.com] Sent: 24 August 2015 12:28 To: Simon Hallam Cc: ceph-users@lists.ceph.com; Gregory Farnum Subject: Re: [ceph-users] Testing CephFS On Aug 24, 2015, at 18:38, Gregory Farnum gfar...@redhat.com wrote: On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam s...@pml.ac.uk wrote: Hi Greg, The MDS' detect that the other one went down and started the replay. I did some further testing with 20 client machines. Of the 20 client machines, 5 hung with the following error: [Aug24 10:53] ceph: mds0 caps stale [Aug24 10:54] ceph: mds0 caps stale [Aug24 10:58] ceph: mds0 hung [Aug24 11:03] ceph: mds0 came back [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state OPEN) [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new mon [Aug24 11:04] ceph: mds0 reconnect start [ +0.084938] libceph: mon2 10.15.0.3:6789 session established [ +0.008475] ceph: mds0 reconnect denied Oh, this might be a kernel bug, failing to ask for mdsmap updates when the connection goes away. Zheng, does that sound familiar? -Greg This seems like reconnect timeout. you can try enlarging mds_reconnect_timeout config option. Which version of kernel are you using? Yan, Zheng 10.15.0.3 was the active MDS at the time I unplugged the Ethernet cable. This was the output of ceph -w as I ran the test (I've removed a lot of the pg remapping): 2015-08-24 11:02:39.547529 mon.1 [INF] mon.ceph2 calling new monitor election 2015-08-24 11:02:40.011995 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:02:45.245869 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1 2015-08-24 11:02:45.257440 mon.0 [INF] HEALTH_WARN; 1 mons down, quorum 0,1 ceph1,ceph2 2015-08-24 11:02:45.535369 mon.0 [INF] monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} 2015-08-24 11:02:45.535444 mon.0 [INF] pgmap v15803: 8256 pgs: 8256 active+clean; 1248 GB data, 2503 GB used, 193 TB / 196 TB avail; 47 B/s wr, 0 op/s 2015-08-24 11:02:45.535541 mon.0 [INF] mdsmap e38: 1/1/1 up {0=ceph3=up:active}, 2 up:standby 2015-08-24 11:02:45.535629 mon.0 [INF] osdmap e197: 36 osds: 36 up, 36 in 2015-08-24 11:03:01.946397 mon.0 [INF] mdsmap e39: 1/1/1 up {0=ceph2=up:replay}, 1 up:standby 2015-08-24 11:03:02.993880 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:reconnect 2015-08-24 11:03:02.993930 mon.0 [INF] mdsmap e40: 1/1/1 up {0=ceph2=up:reconnect}, 1 up:standby 2015-08-24 11:03:51.461248 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:rejoin 2015-08-24 11:03:55.807131 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:03:55.807195 mon.0 [INF] mdsmap e42: 1/1/1 up {0=ceph2=up:active}, 1 up:standby 2015-08-24 11:06:48.036736 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:06:48.036799 mon.0 [INF] mdsmap e43: 1/1/1 up {0=ceph2=up:active}, 1 up:standby *cable plugged back in* 2015-08-24 11:13:13.230714 mon.0 [INF] osd.32 10.15.0.3:6832/11565 boot 2015-08-24 11:13:13.230765 mon.0 [INF] osdmap e212: 36 osds: 25 up, 25 in 2015-08-24 11:13:13.230809 mon.0 [INF] mds.? 10.15.0.3:6833/16993 up:boot 2015-08-24 11:13:13.230837 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:30.799429 mon.2 [INF] mon.ceph3 calling new monitor election 2015-08-24 11:13:30.826158 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:13:30.926331 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1,2 2015-08-24 11:13:30.968739 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:28.383203 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24155 10.10.10.95:0/3238635414 after 625.375507 (allowed interval 45) 2015-08-24 11:13:29.721653 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24146 10.10.10.99:0/3454703638 after 626.713952 (allowed interval 45) 2015-08-24 11:13:31.113004 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24140 10.10.10.60:0/359606080 after 628.105302 (allowed interval 45) 2015-08-24 11:13:50.933020 mds.0 [INF] denied reconnect attempt (mds is up:active) from client
Re: [ceph-users] Testing CephFS
Hi Greg, The MDS' detect that the other one went down and started the replay. I did some further testing with 20 client machines. Of the 20 client machines, 5 hung with the following error: [Aug24 10:53] ceph: mds0 caps stale [Aug24 10:54] ceph: mds0 caps stale [Aug24 10:58] ceph: mds0 hung [Aug24 11:03] ceph: mds0 came back [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state OPEN) [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new mon [Aug24 11:04] ceph: mds0 reconnect start [ +0.084938] libceph: mon2 10.15.0.3:6789 session established [ +0.008475] ceph: mds0 reconnect denied 10.15.0.3 was the active MDS at the time I unplugged the Ethernet cable. This was the output of ceph -w as I ran the test (I've removed a lot of the pg remapping): 2015-08-24 11:02:39.547529 mon.1 [INF] mon.ceph2 calling new monitor election 2015-08-24 11:02:40.011995 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:02:45.245869 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1 2015-08-24 11:02:45.257440 mon.0 [INF] HEALTH_WARN; 1 mons down, quorum 0,1 ceph1,ceph2 2015-08-24 11:02:45.535369 mon.0 [INF] monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} 2015-08-24 11:02:45.535444 mon.0 [INF] pgmap v15803: 8256 pgs: 8256 active+clean; 1248 GB data, 2503 GB used, 193 TB / 196 TB avail; 47 B/s wr, 0 op/s 2015-08-24 11:02:45.535541 mon.0 [INF] mdsmap e38: 1/1/1 up {0=ceph3=up:active}, 2 up:standby 2015-08-24 11:02:45.535629 mon.0 [INF] osdmap e197: 36 osds: 36 up, 36 in 2015-08-24 11:03:01.946397 mon.0 [INF] mdsmap e39: 1/1/1 up {0=ceph2=up:replay}, 1 up:standby 2015-08-24 11:03:02.993880 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:reconnect 2015-08-24 11:03:02.993930 mon.0 [INF] mdsmap e40: 1/1/1 up {0=ceph2=up:reconnect}, 1 up:standby 2015-08-24 11:03:51.461248 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:rejoin 2015-08-24 11:03:55.807131 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:03:55.807195 mon.0 [INF] mdsmap e42: 1/1/1 up {0=ceph2=up:active}, 1 up:standby 2015-08-24 11:06:48.036736 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:06:48.036799 mon.0 [INF] mdsmap e43: 1/1/1 up {0=ceph2=up:active}, 1 up:standby *cable plugged back in* 2015-08-24 11:13:13.230714 mon.0 [INF] osd.32 10.15.0.3:6832/11565 boot 2015-08-24 11:13:13.230765 mon.0 [INF] osdmap e212: 36 osds: 25 up, 25 in 2015-08-24 11:13:13.230809 mon.0 [INF] mds.? 10.15.0.3:6833/16993 up:boot 2015-08-24 11:13:13.230837 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:30.799429 mon.2 [INF] mon.ceph3 calling new monitor election 2015-08-24 11:13:30.826158 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:13:30.926331 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1,2 2015-08-24 11:13:30.968739 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:28.383203 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24155 10.10.10.95:0/3238635414 after 625.375507 (allowed interval 45) 2015-08-24 11:13:29.721653 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24146 10.10.10.99:0/3454703638 after 626.713952 (allowed interval 45) 2015-08-24 11:13:31.113004 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24140 10.10.10.60:0/359606080 after 628.105302 (allowed interval 45) 2015-08-24 11:13:50.933020 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24152 10.10.10.67:0/3475305031 after 647.925323 (allowed interval 45) 2015-08-24 11:13:51.037681 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24149 10.10.10.68:0/22416725 after 648.029988 (allowed interval 45) I did just notice that none of the times match up. So may try again once I fix ntp/chrony and see if that makes a difference. Cheers, Simon -Original Message- From: Gregory Farnum [mailto:gfar...@redhat.com] Sent: 21 August 2015 12:16 To: Simon Hallam Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Testing CephFS On Thu, Aug 20, 2015 at 11:07 AM, Simon Hallam s...@pml.ac.uk wrote: Hey all, We are currently testing CephFS on a small (3 node) cluster. The setup is currently: Each server has 12 OSDs, 1 Monitor and 1 MDS running on it: The servers are running: 0.94.2-0.el7 The clients are running: Ceph: 0.80.10-1.fc21, Kernel: 4.0.6-200.fc21.x86_64 ceph -s cluster 4ed5ecdd-0c5b-4422-9d99-c9e42c6bd4cd health HEALTH_OK monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} election epoch 20, quorum 0,1,2 ceph1,ceph2,ceph3 mdsmap e12: 1/1/1 up {0=ceph3=up:active}, 2 up:standby osdmap e389: 36 osds: 36 up, 36 in pgmap v19370: 8256 pgs, 3 pools, 51217 MB data, 14035 objects 95526 MB used, 196 TB / 196 TB
Re: [ceph-users] Testing CephFS
On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam s...@pml.ac.uk wrote: Hi Greg, The MDS' detect that the other one went down and started the replay. I did some further testing with 20 client machines. Of the 20 client machines, 5 hung with the following error: [Aug24 10:53] ceph: mds0 caps stale [Aug24 10:54] ceph: mds0 caps stale [Aug24 10:58] ceph: mds0 hung [Aug24 11:03] ceph: mds0 came back [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state OPEN) [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new mon [Aug24 11:04] ceph: mds0 reconnect start [ +0.084938] libceph: mon2 10.15.0.3:6789 session established [ +0.008475] ceph: mds0 reconnect denied Oh, this might be a kernel bug, failing to ask for mdsmap updates when the connection goes away. Zheng, does that sound familiar? -Greg 10.15.0.3 was the active MDS at the time I unplugged the Ethernet cable. This was the output of ceph -w as I ran the test (I've removed a lot of the pg remapping): 2015-08-24 11:02:39.547529 mon.1 [INF] mon.ceph2 calling new monitor election 2015-08-24 11:02:40.011995 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:02:45.245869 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1 2015-08-24 11:02:45.257440 mon.0 [INF] HEALTH_WARN; 1 mons down, quorum 0,1 ceph1,ceph2 2015-08-24 11:02:45.535369 mon.0 [INF] monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} 2015-08-24 11:02:45.535444 mon.0 [INF] pgmap v15803: 8256 pgs: 8256 active+clean; 1248 GB data, 2503 GB used, 193 TB / 196 TB avail; 47 B/s wr, 0 op/s 2015-08-24 11:02:45.535541 mon.0 [INF] mdsmap e38: 1/1/1 up {0=ceph3=up:active}, 2 up:standby 2015-08-24 11:02:45.535629 mon.0 [INF] osdmap e197: 36 osds: 36 up, 36 in 2015-08-24 11:03:01.946397 mon.0 [INF] mdsmap e39: 1/1/1 up {0=ceph2=up:replay}, 1 up:standby 2015-08-24 11:03:02.993880 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:reconnect 2015-08-24 11:03:02.993930 mon.0 [INF] mdsmap e40: 1/1/1 up {0=ceph2=up:reconnect}, 1 up:standby 2015-08-24 11:03:51.461248 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:rejoin 2015-08-24 11:03:55.807131 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:03:55.807195 mon.0 [INF] mdsmap e42: 1/1/1 up {0=ceph2=up:active}, 1 up:standby 2015-08-24 11:06:48.036736 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:06:48.036799 mon.0 [INF] mdsmap e43: 1/1/1 up {0=ceph2=up:active}, 1 up:standby *cable plugged back in* 2015-08-24 11:13:13.230714 mon.0 [INF] osd.32 10.15.0.3:6832/11565 boot 2015-08-24 11:13:13.230765 mon.0 [INF] osdmap e212: 36 osds: 25 up, 25 in 2015-08-24 11:13:13.230809 mon.0 [INF] mds.? 10.15.0.3:6833/16993 up:boot 2015-08-24 11:13:13.230837 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:30.799429 mon.2 [INF] mon.ceph3 calling new monitor election 2015-08-24 11:13:30.826158 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:13:30.926331 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1,2 2015-08-24 11:13:30.968739 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:28.383203 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24155 10.10.10.95:0/3238635414 after 625.375507 (allowed interval 45) 2015-08-24 11:13:29.721653 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24146 10.10.10.99:0/3454703638 after 626.713952 (allowed interval 45) 2015-08-24 11:13:31.113004 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24140 10.10.10.60:0/359606080 after 628.105302 (allowed interval 45) 2015-08-24 11:13:50.933020 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24152 10.10.10.67:0/3475305031 after 647.925323 (allowed interval 45) 2015-08-24 11:13:51.037681 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24149 10.10.10.68:0/22416725 after 648.029988 (allowed interval 45) I did just notice that none of the times match up. So may try again once I fix ntp/chrony and see if that makes a difference. Cheers, Simon -Original Message- From: Gregory Farnum [mailto:gfar...@redhat.com] Sent: 21 August 2015 12:16 To: Simon Hallam Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Testing CephFS On Thu, Aug 20, 2015 at 11:07 AM, Simon Hallam s...@pml.ac.uk wrote: Hey all, We are currently testing CephFS on a small (3 node) cluster. The setup is currently: Each server has 12 OSDs, 1 Monitor and 1 MDS running on it: The servers are running: 0.94.2-0.el7 The clients are running: Ceph: 0.80.10-1.fc21, Kernel: 4.0.6-200.fc21.x86_64 ceph -s cluster 4ed5ecdd-0c5b-4422-9d99-c9e42c6bd4cd health HEALTH_OK monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0
Re: [ceph-users] Testing CephFS
On Aug 24, 2015, at 18:38, Gregory Farnum gfar...@redhat.com wrote: On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam s...@pml.ac.uk wrote: Hi Greg, The MDS' detect that the other one went down and started the replay. I did some further testing with 20 client machines. Of the 20 client machines, 5 hung with the following error: [Aug24 10:53] ceph: mds0 caps stale [Aug24 10:54] ceph: mds0 caps stale [Aug24 10:58] ceph: mds0 hung [Aug24 11:03] ceph: mds0 came back [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state OPEN) [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new mon [Aug24 11:04] ceph: mds0 reconnect start [ +0.084938] libceph: mon2 10.15.0.3:6789 session established [ +0.008475] ceph: mds0 reconnect denied Oh, this might be a kernel bug, failing to ask for mdsmap updates when the connection goes away. Zheng, does that sound familiar? -Greg This seems like reconnect timeout. you can try enlarging mds_reconnect_timeout config option. Which version of kernel are you using? Yan, Zheng 10.15.0.3 was the active MDS at the time I unplugged the Ethernet cable. This was the output of ceph -w as I ran the test (I've removed a lot of the pg remapping): 2015-08-24 11:02:39.547529 mon.1 [INF] mon.ceph2 calling new monitor election 2015-08-24 11:02:40.011995 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:02:45.245869 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1 2015-08-24 11:02:45.257440 mon.0 [INF] HEALTH_WARN; 1 mons down, quorum 0,1 ceph1,ceph2 2015-08-24 11:02:45.535369 mon.0 [INF] monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} 2015-08-24 11:02:45.535444 mon.0 [INF] pgmap v15803: 8256 pgs: 8256 active+clean; 1248 GB data, 2503 GB used, 193 TB / 196 TB avail; 47 B/s wr, 0 op/s 2015-08-24 11:02:45.535541 mon.0 [INF] mdsmap e38: 1/1/1 up {0=ceph3=up:active}, 2 up:standby 2015-08-24 11:02:45.535629 mon.0 [INF] osdmap e197: 36 osds: 36 up, 36 in 2015-08-24 11:03:01.946397 mon.0 [INF] mdsmap e39: 1/1/1 up {0=ceph2=up:replay}, 1 up:standby 2015-08-24 11:03:02.993880 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:reconnect 2015-08-24 11:03:02.993930 mon.0 [INF] mdsmap e40: 1/1/1 up {0=ceph2=up:reconnect}, 1 up:standby 2015-08-24 11:03:51.461248 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:rejoin 2015-08-24 11:03:55.807131 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:03:55.807195 mon.0 [INF] mdsmap e42: 1/1/1 up {0=ceph2=up:active}, 1 up:standby 2015-08-24 11:06:48.036736 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:06:48.036799 mon.0 [INF] mdsmap e43: 1/1/1 up {0=ceph2=up:active}, 1 up:standby *cable plugged back in* 2015-08-24 11:13:13.230714 mon.0 [INF] osd.32 10.15.0.3:6832/11565 boot 2015-08-24 11:13:13.230765 mon.0 [INF] osdmap e212: 36 osds: 25 up, 25 in 2015-08-24 11:13:13.230809 mon.0 [INF] mds.? 10.15.0.3:6833/16993 up:boot 2015-08-24 11:13:13.230837 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:30.799429 mon.2 [INF] mon.ceph3 calling new monitor election 2015-08-24 11:13:30.826158 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:13:30.926331 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1,2 2015-08-24 11:13:30.968739 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:28.383203 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24155 10.10.10.95:0/3238635414 after 625.375507 (allowed interval 45) 2015-08-24 11:13:29.721653 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24146 10.10.10.99:0/3454703638 after 626.713952 (allowed interval 45) 2015-08-24 11:13:31.113004 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24140 10.10.10.60:0/359606080 after 628.105302 (allowed interval 45) 2015-08-24 11:13:50.933020 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24152 10.10.10.67:0/3475305031 after 647.925323 (allowed interval 45) 2015-08-24 11:13:51.037681 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24149 10.10.10.68:0/22416725 after 648.029988 (allowed interval 45) I did just notice that none of the times match up. So may try again once I fix ntp/chrony and see if that makes a difference. Cheers, Simon -Original Message- From: Gregory Farnum [mailto:gfar...@redhat.com] Sent: 21 August 2015 12:16 To: Simon Hallam Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] Testing CephFS On Thu, Aug 20, 2015 at 11:07 AM, Simon Hallam s...@pml.ac.uk wrote: Hey all, We are currently testing CephFS on a small (3 node) cluster. The setup is currently: Each server has 12 OSDs, 1 Monitor and 1 MDS running on it: The servers are running: 0.94.2-0.el7 The clients are running: Ceph: 0.80.10-1.fc21, Kernel
Re: [ceph-users] Testing CephFS
The clients are: [root@gridnode50 ~]# uname -a Linux gridnode50 4.0.8-200.fc21.x86_64 #1 SMP Fri Jul 10 21:09:54 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux [root@gridnode50 ~]# ceph -v ceph version 0.80.10 (ea6c958c38df1216bf95c927f143d8b13c4a9e70) I don't think it is a reconnect timeout, as they don't even attempt to reconnect until I plug the Ethernet cable back into the original MDS? Cheers, Simon -Original Message- From: Yan, Zheng [mailto:z...@redhat.com] Sent: 24 August 2015 12:28 To: Simon Hallam Cc: ceph-users@lists.ceph.com; Gregory Farnum Subject: Re: [ceph-users] Testing CephFS On Aug 24, 2015, at 18:38, Gregory Farnum gfar...@redhat.com wrote: On Mon, Aug 24, 2015 at 11:35 AM, Simon Hallam s...@pml.ac.uk wrote: Hi Greg, The MDS' detect that the other one went down and started the replay. I did some further testing with 20 client machines. Of the 20 client machines, 5 hung with the following error: [Aug24 10:53] ceph: mds0 caps stale [Aug24 10:54] ceph: mds0 caps stale [Aug24 10:58] ceph: mds0 hung [Aug24 11:03] ceph: mds0 came back [ +8.803334] libceph: mon2 10.15.0.3:6789 socket closed (con state OPEN) [ +0.18] libceph: mon2 10.15.0.3:6789 session lost, hunting for new mon [Aug24 11:04] ceph: mds0 reconnect start [ +0.084938] libceph: mon2 10.15.0.3:6789 session established [ +0.008475] ceph: mds0 reconnect denied Oh, this might be a kernel bug, failing to ask for mdsmap updates when the connection goes away. Zheng, does that sound familiar? -Greg This seems like reconnect timeout. you can try enlarging mds_reconnect_timeout config option. Which version of kernel are you using? Yan, Zheng 10.15.0.3 was the active MDS at the time I unplugged the Ethernet cable. This was the output of ceph -w as I ran the test (I've removed a lot of the pg remapping): 2015-08-24 11:02:39.547529 mon.1 [INF] mon.ceph2 calling new monitor election 2015-08-24 11:02:40.011995 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:02:45.245869 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1 2015-08-24 11:02:45.257440 mon.0 [INF] HEALTH_WARN; 1 mons down, quorum 0,1 ceph1,ceph2 2015-08-24 11:02:45.535369 mon.0 [INF] monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} 2015-08-24 11:02:45.535444 mon.0 [INF] pgmap v15803: 8256 pgs: 8256 active+clean; 1248 GB data, 2503 GB used, 193 TB / 196 TB avail; 47 B/s wr, 0 op/s 2015-08-24 11:02:45.535541 mon.0 [INF] mdsmap e38: 1/1/1 up {0=ceph3=up:active}, 2 up:standby 2015-08-24 11:02:45.535629 mon.0 [INF] osdmap e197: 36 osds: 36 up, 36 in 2015-08-24 11:03:01.946397 mon.0 [INF] mdsmap e39: 1/1/1 up {0=ceph2=up:replay}, 1 up:standby 2015-08-24 11:03:02.993880 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:reconnect 2015-08-24 11:03:02.993930 mon.0 [INF] mdsmap e40: 1/1/1 up {0=ceph2=up:reconnect}, 1 up:standby 2015-08-24 11:03:51.461248 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:rejoin 2015-08-24 11:03:55.807131 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:03:55.807195 mon.0 [INF] mdsmap e42: 1/1/1 up {0=ceph2=up:active}, 1 up:standby 2015-08-24 11:06:48.036736 mon.0 [INF] mds.0 10.15.0.2:6849/17644 up:active 2015-08-24 11:06:48.036799 mon.0 [INF] mdsmap e43: 1/1/1 up {0=ceph2=up:active}, 1 up:standby *cable plugged back in* 2015-08-24 11:13:13.230714 mon.0 [INF] osd.32 10.15.0.3:6832/11565 boot 2015-08-24 11:13:13.230765 mon.0 [INF] osdmap e212: 36 osds: 25 up, 25 in 2015-08-24 11:13:13.230809 mon.0 [INF] mds.? 10.15.0.3:6833/16993 up:boot 2015-08-24 11:13:13.230837 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:30.799429 mon.2 [INF] mon.ceph3 calling new monitor election 2015-08-24 11:13:30.826158 mon.0 [INF] mon.ceph1 calling new monitor election 2015-08-24 11:13:30.926331 mon.0 [INF] mon.ceph1@0 won leader election with quorum 0,1,2 2015-08-24 11:13:30.968739 mon.0 [INF] mdsmap e47: 1/1/1 up {0=ceph2=up:active}, 2 up:standby 2015-08-24 11:13:28.383203 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24155 10.10.10.95:0/3238635414 after 625.375507 (allowed interval 45) 2015-08-24 11:13:29.721653 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24146 10.10.10.99:0/3454703638 after 626.713952 (allowed interval 45) 2015-08-24 11:13:31.113004 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24140 10.10.10.60:0/359606080 after 628.105302 (allowed interval 45) 2015-08-24 11:13:50.933020 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24152 10.10.10.67:0/3475305031 after 647.925323 (allowed interval 45) 2015-08-24 11:13:51.037681 mds.0 [INF] denied reconnect attempt (mds is up:active) from client.24149 10.10.10.68:0/22416725 after 648.029988 (allowed interval 45) I did just notice that none
Re: [ceph-users] Testing CephFS
On Thu, Aug 20, 2015 at 11:07 AM, Simon Hallam s...@pml.ac.uk wrote: Hey all, We are currently testing CephFS on a small (3 node) cluster. The setup is currently: Each server has 12 OSDs, 1 Monitor and 1 MDS running on it: The servers are running: 0.94.2-0.el7 The clients are running: Ceph: 0.80.10-1.fc21, Kernel: 4.0.6-200.fc21.x86_64 ceph -s cluster 4ed5ecdd-0c5b-4422-9d99-c9e42c6bd4cd health HEALTH_OK monmap e1: 3 mons at {ceph1=10.15.0.1:6789/0,ceph2=10.15.0.2:6789/0,ceph3=10.15.0.3:6789/0} election epoch 20, quorum 0,1,2 ceph1,ceph2,ceph3 mdsmap e12: 1/1/1 up {0=ceph3=up:active}, 2 up:standby osdmap e389: 36 osds: 36 up, 36 in pgmap v19370: 8256 pgs, 3 pools, 51217 MB data, 14035 objects 95526 MB used, 196 TB / 196 TB avail 8256 active+clean Our Ceph.conf is relatively simple at the moment: cat /etc/ceph/ceph.conf [global] fsid = 4ed5ecdd-0c5b-4422-9d99-c9e42c6bd4cd mon_initial_members = ceph1, ceph2, ceph3 mon_host = 10.15.0.1,10.15.0.2,10.15.0.3 mon_pg_warn_max_per_osd = 1000 auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx filestore_xattr_use_omap = true osd_pool_default_size = 2 When I pulled the plug on the master MDS last time (ceph1), it stopped all IO until I plugged it back in. I was under the assumption that the MDS would fail over the other 2 MDS’s and IO would continue? Is there something I need to do to allow the MDS’s to failover from each other without too much interruption? Or is this because the clients ceph version? That's quite strange. How long did you wait for it to fail over? Did the output of ceph -s (or ceph -w, whichever) change during that time? By default the monitors should have detected the MDS was dead after 30 seconds and put one of the other MDS nodes into replay and active. ...I wonder if this is because you lost a monitor at the same time as the MDS. What kind of logging do you have available from during your test? -Greg Cheers, Simon Hallam Linux Support Development Officer ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com