Sage,

Thanks for the quick response.

That seems to fix the problem. Both of my mds are up and active.

mds e1838: 2/2/2 up {0=up:active,1=up:active}

The only thing that doesn't seem right to me (I am not a developer so my 
understanding of git is limited) is that my version is of the following....

ceph version 0.26-303-g36f0068 (commit:36f00685633a6f953b046106f5dd31a9169c82d4)

I don't think this is correct?

I do have a few questions for the group as well.

1. What is the proper way to upgrade from version to version. Has anyone 
documentation the proper procedure?

2. If I just want to test out the file system capabilities, do I need to 
install all the packages or is there some I can do without. I have been 
currently installing the following packages (Ubuntu v10.10) in this order.

sudo dpkg -i libcrush1_0.26-1_amd64.deb
sudo dpkg -i libceph1_0.26-1_amd64.deb
sudo dpkg -i ceph-fuse_0.26-1_amd64.deb
sudo dpkg -i librados2_0.26-1_amd64.deb
sudo dpkg -i librbd1_0.26-1_amd64.deb
sudo dpkg -i ceph_0.26-1_amd64.deb

Thanks for everyone's help.

Mark Nigh
Systems Architect
[email protected]
 (p) 314.392.6926


-----Original Message-----
From: Sage Weil [mailto:[email protected]]
Sent: Tuesday, April 19, 2011 11:18 AM
To: Mark Nigh
Cc: [email protected]
Subject: Re: mds crash

Hi Mark,

This should be fixed by d55399ffec224206ea324e83bb8ead1e9ca1eddc in the
'next' branch of ceph.git.  Can you test it out and see if that allows
journal replay to complete?

Thanks!
sage

http://tracker.newdream.net/issues/1019



On Tue, 19 Apr 2011, Mark Nigh wrote:

> I recently have been working with exporting ceph to NFS. I have had stability 
> problems with NFS (ceph is working but NFS crashes). But most recently, my 
> mds0 will not start after one of these instances with NFS.
>
> My setup. 2 mds, 1 mon (located on mds0), 5 osds. All running Ubuntu v10.10.
>
> Here is the output when I try to start the mds0. Is there other debugging I 
> can turn on?
>
> /etc/init.d/ceph start mds0
>
> 2011-04-19 10:06:58.602640 7fb202fe4700 mds0.11 ms_handle_connect on 
> 10.6.1.93:6800/945
> ./include/elist.h: In function 'elist<T>::item::~item() [with T = 
> MDSlaveUpdate*]', in thread '0x7fb2004d5700'
> ./include/elist.h: 39: FAILED assert(!is_on_list())
>  ceph version 0.26 (commit:9981ff90968398da43c63106694d661f5e3d07d5)
>  1: (MDSlaveUpdate::~MDSlaveUpdate()+0x59) [0x4d9fe9]
>  2: (ESlaveUpdate::replay(MDS*)+0x422) [0x4d2772]
>  3: (MDLog::_replay_thread()+0xb90) [0x67f850]
>  4: (MDLog::ReplayThread::entry()+0xd) [0x4b89ed]
>  5: (()+0x7971) [0x7fb20564a971]
>  6: (clone()+0x6d) [0x7fb2042e692d]
>  ceph version 0.26 (commit:9981ff90968398da43c63106694d661f5e3d07d5)
>  1: (MDSlaveUpdate::~MDSlaveUpdate()+0x59) [0x4d9fe9]
>  2: (ESlaveUpdate::replay(MDS*)+0x422) [0x4d2772]
>  3: (MDLog::_replay_thread()+0xb90) [0x67f850]
>  4: (MDLog::ReplayThread::entry()+0xd) [0x4b89ed]
>  5: (()+0x7971) [0x7fb20564a971]
>  6: (clone()+0x6d) [0x7fb2042e692d]
> *** Caught signal (Aborted) **
>  in thread 0x7fb2004d5700
>  ceph version 0.26 (commit:9981ff90968398da43c63106694d661f5e3d07d5)
>  1: /usr/bin/cmds() [0x70fc38]
>  2: (()+0xfb40) [0x7fb205652b40]
>  3: (gsignal()+0x35) [0x7fb204233ba5]
>  4: (abort()+0x180) [0x7fb2042376b0]
>  5: (__gnu_cxx::__verbose_terminate_handler()+0x11d) [0x7fb204ad76bd]
>  6: (()+0xb9906) [0x7fb204ad5906]
>  7: (()+0xb9933) [0x7fb204ad5933]
>  8: (()+0xb9a3e) [0x7fb204ad5a3e]
>  9: (ceph::__ceph_assert_fail(char const*, char const*, int, char 
> const*)+0x36a) [0x6f5eaa]
>  10: (MDSlaveUpdate::~MDSlaveUpdate()+0x59) [0x4d9fe9]
>  11: (ESlaveUpdate::replay(MDS*)+0x422) [0x4d2772]
>  12: (MDLog::_replay_thread()+0xb90) [0x67f850]
>  13: (MDLog::ReplayThread::entry()+0xd) [0x4b89ed]
>  14: (()+0x7971) [0x7fb20564a971]
>  15: (clone()+0x6d) [0x7fb2042e692d]
>
> I am not sure why the IP address of 0.0.0.0 shows up with starting the mds0.
>
> root@mds0:/var/log/ceph# /etc/init.d/ceph start mds0
> === mds.0 ===
> Starting Ceph mds.0 on mds0...
>  ** WARNING: Ceph is still under heavy development, and is only suitable for 
> **
>  **          testing and review.  Do not trust it with important data.       
> **
> starting mds.0 at 0.0.0.0:6800/2994
>
> Thanks for your assistance.
>
> Mark Nigh
> Systems Architect
> [email protected]
>  (p) 314.392.6926
>
>
>
>
> This transmission and any attached files are privileged, confidential or 
> otherwise the exclusive property of the intended recipient or Netelligent 
> Corporation. If you are not the intended recipient, any disclosure, copying, 
> distribution or use of any of the information contained in or attached to 
> this transmission is strictly prohibited. If you have received this 
> transmission in error, please contact us immediately by responding to this 
> message or by telephone (314-392-6900) and promptly destroy the original 
> transmission and its attachments.
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to [email protected]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
>

This transmission and any attached files are privileged, confidential or 
otherwise the exclusive property of the intended recipient or Netelligent 
Corporation. If you are not the intended recipient, any disclosure, copying, 
distribution or use of any of the information contained in or attached to this 
transmission is strictly prohibited. If you have received this transmission in 
error, please contact us immediately by responding to this message or by 
telephone (314-392-6900) and promptly destroy the original transmission and its 
attachments.
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to