That is a good question, and I'm not sure how to answer. The journal is on its own volume, and is not a symlink. Also how does one flush the journal? That seems like an important step when bringing down a cluster safely.
-Chris > On Mar 1, 2017, at 8:37 AM, Peter Maloney > <peter.malo...@brockmann-consult.de> wrote: > > On 02/28/17 18:55, Heller, Chris wrote: >> Quick update. So I'm trying out the procedure as documented here. >> >> So far I've: >> >> 1. Stopped ceph-mds >> 2. set noout, norecover, norebalance, nobackfill >> 3. Stopped all ceph-osd >> 4. Stopped ceph-mon >> 5. Installed new OS >> 6. Started ceph-mon >> 7. Started all ceph-osd >> >> This is where I've stopped. All but one OSD came back online. One has this >> backtrace: >> >> 2017-02-28 17:44:54.884235 7fb2ba3187c0 -1 journal FileJournal::_open: >> disabling aio for non-block journal. Use journal_force_aio to force use of >> aio anyway > Are the journals inline? or separate? If they're separate, the above means > the journal symlink/config is missing, so it would possibly make a new > journal, which would be bad if you didn't flush the old journal before. > > And also just one osd is easy enough to replace (which I wouldn't do until > the cluster settled down and recovered). So it's lame for it to be broken, > but it's still recoverable if that's the only issue.
smime.p7s
Description: S/MIME cryptographic signature
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com