On Tue, Apr 21, 2020 at 12:44 PM Brad Hubbard <bhubb...@redhat.com> wrote:
>
> On Tue, Apr 21, 2020 at 6:35 PM Paul Emmerich <paul.emmer...@croit.io> wrote:
> >
> > On Tue, Apr 21, 2020 at 3:20 AM Brad Hubbard <bhubb...@redhat.com> wrote:
> > >
> > > Wait for recovery to finish so you know whether any data from the down
> > > OSDs is required. If not just reprovision them.
> >
> > Recovery will not finish from this state as several PGs are down and/or 
> > stale.
>
> What I meant was let recovery get as far as it can.

Which doesn't solve anything, you can already see that you need to get
at least some of these OSDs back in order to fix it.
No point in waiting for the recovery.

I agree that it looks like https://tracker.ceph.com/issues/36337
I happen to know Jonas who opened that issue and wrote the script;
I'll poke him maybe he has an idea or additional input


Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90

>
> >
> >
> > Paul
> >
> > >
> > > If data is required from the down OSDs you will need to run a query on
> > > the pg(s) to find out what OSDs have the required copies of the
> > > pg/object required. you can then export the pg from the down osd using
> > > the ceph-objectstore-tool, back it up, then import it back into the
> > > cluster.
> > >
> > > On Tue, Apr 21, 2020 at 1:05 AM Robert Sander
> > > <r.san...@heinlein-support.de> wrote:
> > > >
> > > > Hi,
> > > >
> > > > one of our customers had his Ceph cluster crashed due to a power or 
> > > > network outage (they still try to figure out what happened).
> > > >
> > > > The cluster is very unhealthy but recovering:
> > > >
> > > > # ceph -s
> > > >   cluster:
> > > >     id:     1c95ca5d-948b-4113-9246-14761cb9a82a
> > > >     health: HEALTH_ERR
> > > >             1 filesystem is degraded
> > > >             1 mds daemon damaged
> > > >             1 osds down
> > > >             1 pools have many more objects per pg than average
> > > >             1/115117480 objects unfound (0.000%)
> > > >             Reduced data availability: 71 pgs inactive, 53 pgs down, 18 
> > > > pgs peering, 27 pgs stale
> > > >             Possible data damage: 1 pg recovery_unfound
> > > >             Degraded data redundancy: 7303464/230234960 objects 
> > > > degraded (3.172%), 693 pgs degraded, 945 pgs undersized
> > > >             14 daemons have recently crashed
> > > >
> > > >   services:
> > > >     mon: 3 daemons, quorum 
> > > > maslxlabstore01,maslxlabstore02,maslxlabstore04 (age 64m)
> > > >     mgr: maslxlabstore01(active, since 69m), standbys: maslxlabstore03, 
> > > > maslxlabstore02, maslxlabstore04
> > > >     mds: cephfs:2/3 
> > > > {0=maslxlabstore03=up:resolve,1=maslxlabstore01=up:resolve} 2 
> > > > up:standby, 1 damaged
> > > >     osd: 140 osds: 130 up (since 4m), 131 in (since 4m); 847 remapped 
> > > > pgs
> > > >     rgw: 4 daemons active (maslxlabstore01.rgw0, maslxlabstore02.rgw0, 
> > > > maslxlabstore03.rgw0, maslxlabstore04.rgw0)
> > > >
> > > >   data:
> > > >     pools:   6 pools, 8328 pgs
> > > >     objects: 115.12M objects, 218 TiB
> > > >     usage:   425 TiB used, 290 TiB / 715 TiB avail
> > > >     pgs:     0.853% pgs not active
> > > >              7303464/230234960 objects degraded (3.172%)
> > > >              13486/230234960 objects misplaced (0.006%)
> > > >              1/115117480 objects unfound (0.000%)
> > > >              7311 active+clean
> > > >              338  active+undersized+degraded+remapped+backfill_wait
> > > >              255  active+undersized+degraded+remapped+backfilling
> > > >              215  active+undersized+remapped+backfilling
> > > >              99   active+undersized+degraded
> > > >              44   down
> > > >              37   active+undersized+remapped+backfill_wait
> > > >              13   stale+peering
> > > >              9    stale+down
> > > >              5    stale+remapped+peering
> > > >              1    active+recovery_unfound+undersized+degraded+remapped
> > > >              1    active+clean+remapped
> > > >
> > > >   io:
> > > >     client:   168 B/s rd, 0 B/s wr, 0 op/s rd, 0 op/s wr
> > > >     recovery: 1.9 GiB/s, 15 keys/s, 948 objects/s
> > > >
> > > >
> > > > The MDS cluster is unable to start because one of them is damaged.
> > > >
> > > > 10 of the OSDs do not start. They crash very early in the boot process:
> > > >
> > > > 2020-04-20 16:26:14.935 7f818ec8cc00  0 set uid:gid to 64045:64045 
> > > > (ceph:ceph)
> > > > 2020-04-20 16:26:14.935 7f818ec8cc00  0 ceph version 14.2.9 
> > > > (581f22da52345dba46ee232b73b990f06029a2a0) nautilus (stable), process 
> > > > ceph-osd, pid 69463
> > > > 2020-04-20 16:26:14.935 7f818ec8cc00  0 pidfile_write: ignore empty 
> > > > --pid-file
> > > > 2020-04-20 16:26:15.503 7f818ec8cc00  0 starting osd.42 osd_data 
> > > > /var/lib/ceph/osd/ceph-42 /var/lib/ceph/osd/ceph-42/journal
> > > > 2020-04-20 16:26:15.523 7f818ec8cc00  0 load: jerasure load: lrc load: 
> > > > isa
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_readahead_size = 2MB
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_style = kCompactionStyleLevel
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_threads = 32
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option compression 
> > > > = kNoCompression
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > flusher_threads = 8
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_file_num_compaction_trigger = 8
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_slowdown_writes_trigger = 32
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_stop_writes_trigger = 64
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_background_compactions = 31
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_base = 536870912
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_multiplier = 8
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_write_buffer_number = 32
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > min_write_buffer_number_to_merge = 2
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > recycle_log_file_num = 32
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > target_file_size_base = 67108864
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > write_buffer_size = 67108864
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_readahead_size = 2MB
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_style = kCompactionStyleLevel
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_threads = 32
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option compression 
> > > > = kNoCompression
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > flusher_threads = 8
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_file_num_compaction_trigger = 8
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_slowdown_writes_trigger = 32
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_stop_writes_trigger = 64
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_background_compactions = 31
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_base = 536870912
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_multiplier = 8
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_write_buffer_number = 32
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > min_write_buffer_number_to_merge = 2
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > recycle_log_file_num = 32
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > target_file_size_base = 67108864
> > > > 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > write_buffer_size = 67108864
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_readahead_size = 2MB
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_style = kCompactionStyleLevel
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_threads = 32
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option compression 
> > > > = kNoCompression
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > flusher_threads = 8
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_file_num_compaction_trigger = 8
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_slowdown_writes_trigger = 32
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_stop_writes_trigger = 64
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_background_compactions = 31
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_base = 536870912
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_multiplier = 8
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_write_buffer_number = 32
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > min_write_buffer_number_to_merge = 2
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > recycle_log_file_num = 32
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > target_file_size_base = 67108864
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > write_buffer_size = 67108864
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_readahead_size = 2MB
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_style = kCompactionStyleLevel
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_threads = 32
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option compression 
> > > > = kNoCompression
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > flusher_threads = 8
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_file_num_compaction_trigger = 8
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_slowdown_writes_trigger = 32
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_stop_writes_trigger = 64
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_background_compactions = 31
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_base = 536870912
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_multiplier = 8
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_write_buffer_number = 32
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > min_write_buffer_number_to_merge = 2
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > recycle_log_file_num = 32
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > target_file_size_base = 67108864
> > > > 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > write_buffer_size = 67108864
> > > > 2020-04-20 16:26:18.131 7f818ec8cc00  0 _get_class not permitted to 
> > > > load lua
> > > > 2020-04-20 16:26:18.131 7f818ec8cc00  0 _get_class not permitted to 
> > > > load kvs
> > > > 2020-04-20 16:26:18.131 7f818ec8cc00  0 _get_class not permitted to 
> > > > load sdk
> > > > 2020-04-20 16:26:18.131 7f818ec8cc00  0 <cls> 
> > > > /build/ceph-14.2.9/src/cls/cephfs/cls_cephfs.cc:197: loading cephfs
> > > > 2020-04-20 16:26:18.131 7f818ec8cc00  0 <cls> 
> > > > /build/ceph-14.2.9/src/cls/hello/cls_hello.cc:296: loading cls_hello
> > > > 2020-04-20 16:26:18.131 7f818ec8cc00  0 osd.42 6008 crush map has 
> > > > features 288514051259236352, adjusting msgr requires for clients
> > > > 2020-04-20 16:26:18.131 7f818ec8cc00  0 osd.42 6008 crush map has 
> > > > features 288514051259236352 was 8705, adjusting msgr requires for mons
> > > > 2020-04-20 16:26:18.131 7f818ec8cc00  0 osd.42 6008 crush map has 
> > > > features 3314933000852226048, adjusting msgr requires for osds
> > > > 2020-04-20 16:26:22.023 7f818ec8cc00  0 osd.42 6008 load_pgs
> > > > 2020-04-20 16:26:22.499 7f818ec8cc00  0 osd.42 6008 load_pgs opened 109 
> > > > pgs
> > > > 2020-04-20 16:26:22.499 7f818ec8cc00  0 osd.42 6008 using 
> > > > weightedpriority op queue with priority op cut off at 64.
> > > > 2020-04-20 16:26:22.499 7f818ec8cc00 -1 osd.42 6008 log_to_monitors 
> > > > {default=true}
> > > > 2020-04-20 16:26:22.511 7f818ec8cc00  0 osd.42 6008 done with init, 
> > > > starting boot process
> > > > 2020-04-20 16:26:23.883 7f815331c700 -1 
> > > > /build/ceph-14.2.9/src/osd/PGLog.cc: In function 'void 
> > > > PGLog::merge_log(pg_info_t&, pg_log_t&, pg_shard_t, pg_info_t&, 
> > > > PGLog::LogEntryHandler*, bool&, bool&)' thread 7f815331c700 time 
> > > > 2020-04-20 16:26:23.884183
> > > > /build/ceph-14.2.9/src/osd/PGLog.cc: 368: FAILED ceph_assert(log.head 
> > > > >= olog.tail && olog.head >= log.tail)
> > > >
> > > >  ceph version 14.2.9 (581f22da52345dba46ee232b73b990f06029a2a0) 
> > > > nautilus (stable)
> > > >  1: (ceph::__ceph_assert_fail(char const*, char const*, int, char 
> > > > const*)+0x152) [0x564fbd9349d2]
> > > >  2: (ceph::__ceph_assertf_fail(char const*, char const*, int, char 
> > > > const*, char const*, ...)+0) [0x564fbd934bad]
> > > >  3: (PGLog::merge_log(pg_info_t&, pg_log_t&, pg_shard_t, pg_info_t&, 
> > > > PGLog::LogEntryHandler*, bool&, bool&)+0x1cc0) [0x564fbdaff930]
> > > >  4: (PG::merge_log(ObjectStore::Transaction&, pg_info_t&, pg_log_t&, 
> > > > pg_shard_t)+0x64) [0x564fbda4eca4]
> > > >  5: (PG::proc_master_log(ObjectStore::Transaction&, pg_info_t&, 
> > > > pg_log_t&, pg_missing_set<false>&, pg_shard_t)+0x97) [0x564fbda7fe47]
> > > >  6: (PG::RecoveryState::GetLog::react(PG::RecoveryState::GotLog 
> > > > const&)+0xa6) [0x564fbda9d4f6]
> > > >  7: (boost::statechart::simple_state<PG::RecoveryState::GetLog, 
> > > > PG::RecoveryState::Peering, boost::mpl::list<mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na>, 
> > > > (boost::statechart::history_mode)0>::react_impl(boost::statechart::event_base
> > > >  const&, void const*)+0x191) [0x564fbdaf0e21]
> > > >  8: 
> > > > (boost::statechart::state_machine<PG::RecoveryState::RecoveryMachine, 
> > > > PG::RecoveryState::Initial, std::allocator<boost::statechart::none>, 
> > > > boost::statechart::null_exception_translator>::process_queued_events()+0xb3)
> > > >  [0x564fbdabdfc3]
> > > >  9: 
> > > > (boost::statechart::state_machine<PG::RecoveryState::RecoveryMachine, 
> > > > PG::RecoveryState::Initial, std::allocator<boost::statechart::none>, 
> > > > boost::statechart::null_exception_translator>::process_event(boost::statechart::event_base
> > > >  const&)+0x87) [0x564fbdabe227]
> > > >  10: (PG::do_peering_event(std::shared_ptr<PGPeeringEvent>, 
> > > > PG::RecoveryCtx*)+0x122) [0x564fbdaada12]
> > > >  11: (OSD::dequeue_peering_evt(OSDShard*, PG*, 
> > > > std::shared_ptr<PGPeeringEvent>, ThreadPool::TPHandle&)+0x1b4) 
> > > > [0x564fbd9e1f54]
> > > >  12: (PGPeeringItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, 
> > > > ThreadPool::TPHandle&)+0x50) [0x564fbdc710c0]
> > > >  13: (OSD::ShardedOpWQ::_process(unsigned int, 
> > > > ceph::heartbeat_handle_d*)+0xbf5) [0x564fbd9d5995]
> > > >  14: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x4ac) 
> > > > [0x564fbdfdb8cc]
> > > >  15: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) 
> > > > [0x564fbdfdea90]
> > > >  16: (()+0x76db) [0x7f818c8666db]
> > > >  17: (clone()+0x3f) [0x7f818b60688f]
> > > >
> > > > 2020-04-20 16:26:23.887 7f815331c700 -1 *** Caught signal (Aborted) **
> > > >  in thread 7f815331c700 thread_name:tp_osd_tp
> > > >
> > > >  ceph version 14.2.9 (581f22da52345dba46ee232b73b990f06029a2a0) 
> > > > nautilus (stable)
> > > >  1: (()+0x12890) [0x7f818c871890]
> > > >  2: (gsignal()+0xc7) [0x7f818b523e97]
> > > >  3: (abort()+0x141) [0x7f818b525801]
> > > >  4: (ceph::__ceph_assert_fail(char const*, char const*, int, char 
> > > > const*)+0x1a3) [0x564fbd934a23]
> > > >  5: (ceph::__ceph_assertf_fail(char const*, char const*, int, char 
> > > > const*, char const*, ...)+0) [0x564fbd934bad]
> > > >  6: (PGLog::merge_log(pg_info_t&, pg_log_t&, pg_shard_t, pg_info_t&, 
> > > > PGLog::LogEntryHandler*, bool&, bool&)+0x1cc0) [0x564fbdaff930]
> > > >  7: (PG::merge_log(ObjectStore::Transaction&, pg_info_t&, pg_log_t&, 
> > > > pg_shard_t)+0x64) [0x564fbda4eca4]
> > > >  8: (PG::proc_master_log(ObjectStore::Transaction&, pg_info_t&, 
> > > > pg_log_t&, pg_missing_set<false>&, pg_shard_t)+0x97) [0x564fbda7fe47]
> > > >  9: (PG::RecoveryState::GetLog::react(PG::RecoveryState::GotLog 
> > > > const&)+0xa6) [0x564fbda9d4f6]
> > > >  10: (boost::statechart::simple_state<PG::RecoveryState::GetLog, 
> > > > PG::RecoveryState::Peering, boost::mpl::list<mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na>, 
> > > > (boost::statechart::history_mode)0>::react_impl(boost::statechart::event_base
> > > >  const&, void const*)+0x191) [0x564fbdaf0e21]
> > > >  11: 
> > > > (boost::statechart::state_machine<PG::RecoveryState::RecoveryMachine, 
> > > > PG::RecoveryState::Initial, std::allocator<boost::statechart::none>, 
> > > > boost::statechart::null_exception_translator>::process_queued_events()+0xb3)
> > > >  [0x564fbdabdfc3]
> > > >  12: 
> > > > (boost::statechart::state_machine<PG::RecoveryState::RecoveryMachine, 
> > > > PG::RecoveryState::Initial, std::allocator<boost::statechart::none>, 
> > > > boost::statechart::null_exception_translator>::process_event(boost::statechart::event_base
> > > >  const&)+0x87) [0x564fbdabe227]
> > > >  13: (PG::do_peering_event(std::shared_ptr<PGPeeringEvent>, 
> > > > PG::RecoveryCtx*)+0x122) [0x564fbdaada12]
> > > >  14: (OSD::dequeue_peering_evt(OSDShard*, PG*, 
> > > > std::shared_ptr<PGPeeringEvent>, ThreadPool::TPHandle&)+0x1b4) 
> > > > [0x564fbd9e1f54]
> > > >  15: (PGPeeringItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, 
> > > > ThreadPool::TPHandle&)+0x50) [0x564fbdc710c0]
> > > >  16: (OSD::ShardedOpWQ::_process(unsigned int, 
> > > > ceph::heartbeat_handle_d*)+0xbf5) [0x564fbd9d5995]
> > > >  17: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x4ac) 
> > > > [0x564fbdfdb8cc]
> > > >  18: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) 
> > > > [0x564fbdfdea90]
> > > >  19: (()+0x76db) [0x7f818c8666db]
> > > >  20: (clone()+0x3f) [0x7f818b60688f]
> > > >  NOTE: a copy of the executable, or `objdump -rdS <executable>` is 
> > > > needed to interpret this.
> > > >
> > > > --- begin dump of recent events ---
> > > >   -105> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command assert hook 0x564fc83ae510
> > > >   -104> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command abort hook 0x564fc83ae510
> > > >   -103> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command perfcounters_dump hook 0x564fc83ae510
> > > >   -102> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command 1 hook 0x564fc83ae510
> > > >   -101> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command perf dump hook 0x564fc83ae510
> > > >   -100> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command perfcounters_schema hook 0x564fc83ae510
> > > >    -99> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command perf histogram dump hook 0x564fc83ae510
> > > >    -98> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command 2 hook 0x564fc83ae510
> > > >    -97> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command perf schema hook 0x564fc83ae510
> > > >    -96> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command perf histogram schema hook 0x564fc83ae510
> > > >    -95> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command perf reset hook 0x564fc83ae510
> > > >    -94> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command config show hook 0x564fc83ae510
> > > >    -93> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command config help hook 0x564fc83ae510
> > > >    -92> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command config set hook 0x564fc83ae510
> > > >    -91> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command config unset hook 0x564fc83ae510
> > > >    -90> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command config get hook 0x564fc83ae510
> > > >    -89> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command config diff hook 0x564fc83ae510
> > > >    -88> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command config diff get hook 0x564fc83ae510
> > > >    -87> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command log flush hook 0x564fc83ae510
> > > >    -86> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command log dump hook 0x564fc83ae510
> > > >    -85> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command log reopen hook 0x564fc83ae510
> > > >    -84> 2020-04-20 16:26:14.923 7f818ec8cc00  5 asok(0x564fc8472000) 
> > > > register_command dump_mempools hook 0x564fc9076068
> > > >    -83> 2020-04-20 16:26:14.935 7f818ec8cc00  0 set uid:gid to 
> > > > 64045:64045 (ceph:ceph)
> > > >    -82> 2020-04-20 16:26:14.935 7f818ec8cc00  0 ceph version 14.2.9 
> > > > (581f22da52345dba46ee232b73b990f06029a2a0) nautilus (stable), process 
> > > > ceph-osd, pid 69463
> > > >    -81> 2020-04-20 16:26:14.935 7f818ec8cc00  0 pidfile_write: ignore 
> > > > empty --pid-file
> > > >    -80> 2020-04-20 16:26:15.503 7f818ec8cc00  0 starting osd.42 
> > > > osd_data /var/lib/ceph/osd/ceph-42 /var/lib/ceph/osd/ceph-42/journal
> > > >    -79> 2020-04-20 16:26:15.523 7f818ec8cc00  0 load: jerasure load: 
> > > > lrc load: isa
> > > >    -78> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_readahead_size = 2MB
> > > >    -77> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_style = kCompactionStyleLevel
> > > >    -76> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_threads = 32
> > > >    -75> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compression = kNoCompression
> > > >    -74> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > flusher_threads = 8
> > > >    -73> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_file_num_compaction_trigger = 8
> > > >    -72> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_slowdown_writes_trigger = 32
> > > >    -71> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_stop_writes_trigger = 64
> > > >    -70> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_background_compactions = 31
> > > >    -69> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_base = 536870912
> > > >    -68> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_multiplier = 8
> > > >    -67> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_write_buffer_number = 32
> > > >    -66> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > min_write_buffer_number_to_merge = 2
> > > >    -65> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > recycle_log_file_num = 32
> > > >    -64> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > target_file_size_base = 67108864
> > > >    -63> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > write_buffer_size = 67108864
> > > >    -62> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_readahead_size = 2MB
> > > >    -61> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_style = kCompactionStyleLevel
> > > >    -60> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_threads = 32
> > > >    -59> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > compression = kNoCompression
> > > >    -58> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > flusher_threads = 8
> > > >    -57> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_file_num_compaction_trigger = 8
> > > >    -56> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_slowdown_writes_trigger = 32
> > > >    -55> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > level0_stop_writes_trigger = 64
> > > >    -54> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_background_compactions = 31
> > > >    -53> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_base = 536870912
> > > >    -52> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_multiplier = 8
> > > >    -51> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > max_write_buffer_number = 32
> > > >    -50> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > min_write_buffer_number_to_merge = 2
> > > >    -49> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > recycle_log_file_num = 32
> > > >    -48> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > target_file_size_base = 67108864
> > > >    -47> 2020-04-20 16:26:16.339 7f818ec8cc00  0  set rocksdb option 
> > > > write_buffer_size = 67108864
> > > >    -46> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_readahead_size = 2MB
> > > >    -45> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_style = kCompactionStyleLevel
> > > >    -44> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_threads = 32
> > > >    -43> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compression = kNoCompression
> > > >    -42> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > flusher_threads = 8
> > > >    -41> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_file_num_compaction_trigger = 8
> > > >    -40> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_slowdown_writes_trigger = 32
> > > >    -39> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_stop_writes_trigger = 64
> > > >    -38> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_background_compactions = 31
> > > >    -37> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_base = 536870912
> > > >    -36> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_multiplier = 8
> > > >    -35> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_write_buffer_number = 32
> > > >    -34> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > min_write_buffer_number_to_merge = 2
> > > >    -33> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > recycle_log_file_num = 32
> > > >    -32> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > target_file_size_base = 67108864
> > > >    -31> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > write_buffer_size = 67108864
> > > >    -30> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_readahead_size = 2MB
> > > >    -29> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_style = kCompactionStyleLevel
> > > >    -28> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compaction_threads = 32
> > > >    -27> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > compression = kNoCompression
> > > >    -26> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > flusher_threads = 8
> > > >    -25> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_file_num_compaction_trigger = 8
> > > >    -24> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_slowdown_writes_trigger = 32
> > > >    -23> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > level0_stop_writes_trigger = 64
> > > >    -22> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_background_compactions = 31
> > > >    -21> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_base = 536870912
> > > >    -20> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_bytes_for_level_multiplier = 8
> > > >    -19> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > max_write_buffer_number = 32
> > > >    -18> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > min_write_buffer_number_to_merge = 2
> > > >    -17> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > recycle_log_file_num = 32
> > > >    -16> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > target_file_size_base = 67108864
> > > >    -15> 2020-04-20 16:26:17.731 7f818ec8cc00  0  set rocksdb option 
> > > > write_buffer_size = 67108864
> > > >    -14> 2020-04-20 16:26:18.131 7f818ec8cc00  0 _get_class not 
> > > > permitted to load lua
> > > >    -13> 2020-04-20 16:26:18.131 7f818ec8cc00  0 _get_class not 
> > > > permitted to load kvs
> > > >    -12> 2020-04-20 16:26:18.131 7f818ec8cc00  0 _get_class not 
> > > > permitted to load sdk
> > > >    -11> 2020-04-20 16:26:18.131 7f818ec8cc00  0 <cls> 
> > > > /build/ceph-14.2.9/src/cls/cephfs/cls_cephfs.cc:197: loading cephfs
> > > >    -10> 2020-04-20 16:26:18.131 7f818ec8cc00  0 <cls> 
> > > > /build/ceph-14.2.9/src/cls/hello/cls_hello.cc:296: loading cls_hello
> > > >     -9> 2020-04-20 16:26:18.131 7f818ec8cc00  0 osd.42 6008 crush map 
> > > > has features 288514051259236352, adjusting msgr requires for clients
> > > >     -8> 2020-04-20 16:26:18.131 7f818ec8cc00  0 osd.42 6008 crush map 
> > > > has features 288514051259236352 was 8705, adjusting msgr requires for 
> > > > mons
> > > >     -7> 2020-04-20 16:26:18.131 7f818ec8cc00  0 osd.42 6008 crush map 
> > > > has features 3314933000852226048, adjusting msgr requires for osds
> > > >     -6> 2020-04-20 16:26:22.023 7f818ec8cc00  0 osd.42 6008 load_pgs
> > > >     -5> 2020-04-20 16:26:22.499 7f818ec8cc00  0 osd.42 6008 load_pgs 
> > > > opened 109 pgs
> > > >     -4> 2020-04-20 16:26:22.499 7f818ec8cc00  0 osd.42 6008 using 
> > > > weightedpriority op queue with priority op cut off at 64.
> > > >     -3> 2020-04-20 16:26:22.499 7f818ec8cc00 -1 osd.42 6008 
> > > > log_to_monitors {default=true}
> > > >     -2> 2020-04-20 16:26:22.511 7f818ec8cc00  0 osd.42 6008 done with 
> > > > init, starting boot process
> > > >     -1> 2020-04-20 16:26:23.883 7f815331c700 -1 
> > > > /build/ceph-14.2.9/src/osd/PGLog.cc: In function 'void 
> > > > PGLog::merge_log(pg_info_t&, pg_log_t&, pg_shard_t, pg_info_t&, 
> > > > PGLog::LogEntryHandler*, bool&, bool&)' thread 7f815331c700 time 
> > > > 2020-04-20 16:26:23.884183
> > > > /build/ceph-14.2.9/src/osd/PGLog.cc: 368: FAILED ceph_assert(log.head 
> > > > >= olog.tail && olog.head >= log.tail)
> > > >
> > > >  ceph version 14.2.9 (581f22da52345dba46ee232b73b990f06029a2a0) 
> > > > nautilus (stable)
> > > >  1: (ceph::__ceph_assert_fail(char const*, char const*, int, char 
> > > > const*)+0x152) [0x564fbd9349d2]
> > > >  2: (ceph::__ceph_assertf_fail(char const*, char const*, int, char 
> > > > const*, char const*, ...)+0) [0x564fbd934bad]
> > > >  3: (PGLog::merge_log(pg_info_t&, pg_log_t&, pg_shard_t, pg_info_t&, 
> > > > PGLog::LogEntryHandler*, bool&, bool&)+0x1cc0) [0x564fbdaff930]
> > > >  4: (PG::merge_log(ObjectStore::Transaction&, pg_info_t&, pg_log_t&, 
> > > > pg_shard_t)+0x64) [0x564fbda4eca4]
> > > >  5: (PG::proc_master_log(ObjectStore::Transaction&, pg_info_t&, 
> > > > pg_log_t&, pg_missing_set<false>&, pg_shard_t)+0x97) [0x564fbda7fe47]
> > > >  6: (PG::RecoveryState::GetLog::react(PG::RecoveryState::GotLog 
> > > > const&)+0xa6) [0x564fbda9d4f6]
> > > >  7: (boost::statechart::simple_state<PG::RecoveryState::GetLog, 
> > > > PG::RecoveryState::Peering, boost::mpl::list<mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na>, 
> > > > (boost::statechart::history_mode)0>::react_impl(boost::statechart::event_base
> > > >  const&, void const*)+0x191) [0x564fbdaf0e21]
> > > >  8: 
> > > > (boost::statechart::state_machine<PG::RecoveryState::RecoveryMachine, 
> > > > PG::RecoveryState::Initial, std::allocator<boost::statechart::none>, 
> > > > boost::statechart::null_exception_translator>::process_queued_events()+0xb3)
> > > >  [0x564fbdabdfc3]
> > > >  9: 
> > > > (boost::statechart::state_machine<PG::RecoveryState::RecoveryMachine, 
> > > > PG::RecoveryState::Initial, std::allocator<boost::statechart::none>, 
> > > > boost::statechart::null_exception_translator>::process_event(boost::statechart::event_base
> > > >  const&)+0x87) [0x564fbdabe227]
> > > >  10: (PG::do_peering_event(std::shared_ptr<PGPeeringEvent>, 
> > > > PG::RecoveryCtx*)+0x122) [0x564fbdaada12]
> > > >  11: (OSD::dequeue_peering_evt(OSDShard*, PG*, 
> > > > std::shared_ptr<PGPeeringEvent>, ThreadPool::TPHandle&)+0x1b4) 
> > > > [0x564fbd9e1f54]
> > > >  12: (PGPeeringItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, 
> > > > ThreadPool::TPHandle&)+0x50) [0x564fbdc710c0]
> > > >  13: (OSD::ShardedOpWQ::_process(unsigned int, 
> > > > ceph::heartbeat_handle_d*)+0xbf5) [0x564fbd9d5995]
> > > >  14: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x4ac) 
> > > > [0x564fbdfdb8cc]
> > > >  15: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) 
> > > > [0x564fbdfdea90]
> > > >  16: (()+0x76db) [0x7f818c8666db]
> > > >  17: (clone()+0x3f) [0x7f818b60688f]
> > > >
> > > >      0> 2020-04-20 16:26:23.887 7f815331c700 -1 *** Caught signal 
> > > > (Aborted) **
> > > >  in thread 7f815331c700 thread_name:tp_osd_tp
> > > >
> > > >  ceph version 14.2.9 (581f22da52345dba46ee232b73b990f06029a2a0) 
> > > > nautilus (stable)
> > > >  1: (()+0x12890) [0x7f818c871890]
> > > >  2: (gsignal()+0xc7) [0x7f818b523e97]
> > > >  3: (abort()+0x141) [0x7f818b525801]
> > > >  4: (ceph::__ceph_assert_fail(char const*, char const*, int, char 
> > > > const*)+0x1a3) [0x564fbd934a23]
> > > >  5: (ceph::__ceph_assertf_fail(char const*, char const*, int, char 
> > > > const*, char const*, ...)+0) [0x564fbd934bad]
> > > >  6: (PGLog::merge_log(pg_info_t&, pg_log_t&, pg_shard_t, pg_info_t&, 
> > > > PGLog::LogEntryHandler*, bool&, bool&)+0x1cc0) [0x564fbdaff930]
> > > >  7: (PG::merge_log(ObjectStore::Transaction&, pg_info_t&, pg_log_t&, 
> > > > pg_shard_t)+0x64) [0x564fbda4eca4]
> > > >  8: (PG::proc_master_log(ObjectStore::Transaction&, pg_info_t&, 
> > > > pg_log_t&, pg_missing_set<false>&, pg_shard_t)+0x97) [0x564fbda7fe47]
> > > >  9: (PG::RecoveryState::GetLog::react(PG::RecoveryState::GotLog 
> > > > const&)+0xa6) [0x564fbda9d4f6]
> > > >  10: (boost::statechart::simple_state<PG::RecoveryState::GetLog, 
> > > > PG::RecoveryState::Peering, boost::mpl::list<mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, mpl_::na, 
> > > > mpl_::na, mpl_::na, mpl_::na, mpl_::na>, 
> > > > (boost::statechart::history_mode)0>::react_impl(boost::statechart::event_base
> > > >  const&, void const*)+0x191) [0x564fbdaf0e21]
> > > >  11: 
> > > > (boost::statechart::state_machine<PG::RecoveryState::RecoveryMachine, 
> > > > PG::RecoveryState::Initial, std::allocator<boost::statechart::none>, 
> > > > boost::statechart::null_exception_translator>::process_queued_events()+0xb3)
> > > >  [0x564fbdabdfc3]
> > > >  12: 
> > > > (boost::statechart::state_machine<PG::RecoveryState::RecoveryMachine, 
> > > > PG::RecoveryState::Initial, std::allocator<boost::statechart::none>, 
> > > > boost::statechart::null_exception_translator>::process_event(boost::statechart::event_base
> > > >  const&)+0x87) [0x564fbdabe227]
> > > >  13: (PG::do_peering_event(std::shared_ptr<PGPeeringEvent>, 
> > > > PG::RecoveryCtx*)+0x122) [0x564fbdaada12]
> > > >  14: (OSD::dequeue_peering_evt(OSDShard*, PG*, 
> > > > std::shared_ptr<PGPeeringEvent>, ThreadPool::TPHandle&)+0x1b4) 
> > > > [0x564fbd9e1f54]
> > > >  15: (PGPeeringItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, 
> > > > ThreadPool::TPHandle&)+0x50) [0x564fbdc710c0]
> > > >  16: (OSD::ShardedOpWQ::_process(unsigned int, 
> > > > ceph::heartbeat_handle_d*)+0xbf5) [0x564fbd9d5995]
> > > >  17: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x4ac) 
> > > > [0x564fbdfdb8cc]
> > > >  18: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) 
> > > > [0x564fbdfdea90]
> > > >  19: (()+0x76db) [0x7f818c8666db]
> > > >  20: (clone()+0x3f) [0x7f818b60688f]
> > > >  NOTE: a copy of the executable, or `objdump -rdS <executable>` is 
> > > > needed to interpret this.
> > > >
> > > > --- logging levels ---
> > > >    0/ 0 none
> > > >    0/ 0 lockdep
> > > >    0/ 0 context
> > > >    0/ 0 crush
> > > >    0/ 0 mds
> > > >    0/ 0 mds_balancer
> > > >    0/ 0 mds_locker
> > > >    0/ 0 mds_log
> > > >    0/ 0 mds_log_expire
> > > >    0/ 0 mds_migrator
> > > >    0/ 0 buffer
> > > >    0/ 0 timer
> > > >    0/ 0 filer
> > > >    0/ 0 striper
> > > >    0/ 0 objecter
> > > >    0/ 0 rados
> > > >    0/ 0 rbd
> > > >    0/ 0 rbd_mirror
> > > >    0/ 0 rbd_replay
> > > >    0/ 0 journaler
> > > >    0/ 0 objectcacher
> > > >    0/ 0 client
> > > >    0/ 0 osd
> > > >    0/ 0 optracker
> > > >    0/ 0 objclass
> > > >    0/ 0 filestore
> > > >    0/ 0 journal
> > > >    0/ 0 ms
> > > >    0/ 0 mon
> > > >    0/ 0 monc
> > > >    0/ 0 paxos
> > > >    0/ 0 tp
> > > >    0/ 0 auth
> > > >    0/ 0 crypto
> > > >    0/ 0 finisher
> > > >    0/ 0 reserver
> > > >    0/ 0 heartbeatmap
> > > >    0/ 0 perfcounter
> > > >    0/ 0 rgw
> > > >    1/ 5 rgw_sync
> > > >    0/ 0 civetweb
> > > >    0/ 0 javaclient
> > > >    0/ 0 asok
> > > >    0/ 0 throttle
> > > >    0/ 0 refs
> > > >    0/ 0 xio
> > > >    0/ 0 compressor
> > > >    0/ 0 bluestore
> > > >    0/ 0 bluefs
> > > >    0/ 0 bdev
> > > >    0/ 0 kstore
> > > >    0/ 0 rocksdb
> > > >    0/ 0 leveldb
> > > >    0/ 0 memdb
> > > >    0/ 0 kinetic
> > > >    0/ 0 fuse
> > > >    0/ 0 mgr
> > > >    0/ 0 mgrc
> > > >    0/ 0 dpdk
> > > >    0/ 0 eventtrace
> > > >    1/ 5 prioritycache
> > > >   -2/-2 (syslog threshold)
> > > >   -1/-1 (stderr threshold)
> > > >   max_recent     10000
> > > >   max_new         1000
> > > >   log_file /var/log/ceph/ceph-osd.42.log
> > > > --- end dump of recent events ---
> > > >
> > > > It would be nice if anybody could give me a hint on where to look 
> > > > further.
> > > >
> > > > Regards
> > > > --
> > > > Robert Sander
> > > > Heinlein Support GmbH
> > > > Schwedter Str. 8/9b, 10119 Berlin
> > > >
> > > > http://www.heinlein-support.de
> > > >
> > > > Tel: 030 / 405051-43
> > > > Fax: 030 / 405051-19
> > > >
> > > > Zwangsangaben lt. §35a GmbHG:
> > > > HRB 93818 B / Amtsgericht Berlin-Charlottenburg,
> > > > Geschäftsführer: Peer Heinlein -- Sitz: Berlin
> > > >
> > > > _______________________________________________
> > > > ceph-users mailing list -- ceph-users@ceph.io
> > > > To unsubscribe send an email to ceph-users-le...@ceph.io
> > >
> > >
> > >
> > > --
> > > Cheers,
> > > Brad
> > > _______________________________________________
> > > ceph-users mailing list -- ceph-users@ceph.io
> > > To unsubscribe send an email to ceph-users-le...@ceph.io
> >
>
>
> --
> Cheers,
> Brad
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to