On Fri, 2 Feb 2018, Frank Li wrote:
> Hi, I ran the ceph osd force-create-pg command in luminious 12.2.2 to recover
> a failed pg, and it
> Instantly caused all of the monitor to crash, is there anyway to revert back
> to an earlier state of the cluster ?
> Right now, the monitors refuse to come up, the error message is as follows:
> I’ve filed a ceph ticket for the crash, but just wonder if there is a way to
> get the cluster back up ?
Can you includ the bit of the log a few lines up that includes the
assertion and file line number that failed?
Also, "during the course of trouble-shooting an osd issue" makes me
nervous: force-create-pg creates a new, *empty* PG when all copies of the
old one have been lost. Is that what you meant to do? It is essentially
telling the system to give up and accepting that there is data loss. Is
that what you meant?
> --- begin dump of recent events ---
> 0> 2018-01-31 22:47:22.959665 7fc64350e700 -1 *** Caught signal
> (Aborted) **
> in thread 7fc64350e700 thread_name:cpu_tp
> ceph version 12.2.2 (cf0baeeeeba3b47f9427c6c97e2144b094b7e5ba) luminous
> 1: (()+0x8eae11) [0x55f1113fae11]
> 2: (()+0xf5e0) [0x7fc64aafa5e0]
> 3: (gsignal()+0x37) [0x7fc647fca1f7]
> 4: (abort()+0x148) [0x7fc647fcb8e8]
> 5: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> const*)+0x284) [0x55f1110fa4a4]
> 6: (()+0x2ccc4e) [0x55f110ddcc4e]
> 7: (OSDMonitor::update_creating_pgs()+0x98b) [0x55f11102232b]
> 8: (C_UpdateCreatingPGs::finish(int)+0x79) [0x55f1110777b9]
> 9: (Context::complete(int)+0x9) [0x55f110ed30c9]
> 10: (ParallelPGMapper::WQ::_process(ParallelPGMapper::Item*,
> ThreadPool::TPHandle&)+0x7f) [0x55f111204e1f]
> 11: (ThreadPool::worker(ThreadPool::WorkThread*)+0xa8e) [0x55f111100f1e]
> 12: (ThreadPool::WorkThread::entry()+0x10) [0x55f111101e00]
> 13: (()+0x7e25) [0x7fc64aaf2e25]
> 14: (clone()+0x6d) [0x7fc64808d34d]
> NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to
> interpret this.
> Efficiency is Intelligent Laziness
ceph-users mailing list