I just did this and backfilling started.  Let's see where this takes me. ceph 
osd lost 0 --yes-i-really-mean-it
Regards,Hong 

    On Friday, September 15, 2017 12:44 AM, hjcho616 <hjcho...@yahoo.com> wrote:
 

 Ronny,
Working with all of the pgs shown in the "ceph health detail", I ran below for 
each PG to export.ceph-objectstore-tool --op export --pgid 0.1c   --data-path 
/var/lib/ceph/osd/ceph-0 --journal-path /var/lib/ceph/osd/ceph-0/journal 
--skip-journal-replay --file 0.1c.export

I have all PGs exported, except 1... PG 1.28.  It is on ceph-4.  This error 
doesn't make much sense to me.  Looking at the source code from 
https://github.com/ceph/ceph/blob/master/src/osd/osd_types.cc, that message is 
telling me struct_v is 1... but not sure how it ended up in the default in the 
case statement when 1 case is defined...  I tried with --skip-journal-replay, 
fails with same error message.ceph-objectstore-tool --op export --pgid 1.28  
--data-path /var/lib/ceph/osd/ceph-4 --journal-path 
/var/lib/ceph/osd/ceph-4/journal --file 1.28.exportterminate called after 
throwing an instance of 'std::domain_error'  what():  coll_t::decode(): don't 
know how to decode version 1*** Caught signal (Aborted) ** in thread 
7fabc5ecc940 thread_name:ceph-objectstor ceph version 10.2.9 
(2ee413f77150c0f375ff6f10edd6c8f9c7d060d0) 1: (()+0x996a57) [0x55b2d3323a57] 2: 
(()+0x110c0) [0x7fabc46d50c0] 3: (gsignal()+0xcf) [0x7fabc2b08fcf] 4: 
(abort()+0x16a) [0x7fabc2b0a3fa] 5: 
(__gnu_cxx::__verbose_terminate_handler()+0x15d) [0x7fabc33efb3d] 6: 
(()+0x5ebb6) [0x7fabc33edbb6] 7: (()+0x5ec01) [0x7fabc33edc01] 8: (()+0x5ee19) 
[0x7fabc33ede19] 9: (coll_t::decode(ceph::buffer::list::iterator&)+0x21e) 
[0x55b2d2ff401e] 10: 
(DBObjectMap::_Header::decode(ceph::buffer::list::iterator&)+0x125) 
[0x55b2d31315f5] 11: (DBObjectMap::check(std::ostream&, bool)+0x279) 
[0x55b2d3126bb9] 12: (DBObjectMap::init(bool)+0x288) [0x55b2d3125eb8] 13: 
(FileStore::mount()+0x2525) [0x55b2d305ceb5] 14: (main()+0x28c0) 
[0x55b2d2c8d400] 15: (__libc_start_main()+0xf1) [0x7fabc2af62b1] 16: 
(()+0x34f747) [0x55b2d2cdc747]Aborted
Then wrote a simple script to run import process... just created an OSD per PG. 
 Basically ran below for each PG.mkdir 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/ceph-disk prepare 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/chown -R ceph.ceph 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/ceph-disk activate 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/ceph osd crush reweight osd.$(cat 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/whoami) 0systemctl stop ceph-osd@$(cat 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/whoami)ceph-objectstore-tool --op import 
--pgid 0.1c   --data-path /var/lib/ceph/osd/ceph-$(cat 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/whoami) --journal-path 
/var/lib/ceph/osd/ceph-$(cat 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/whoami)/journal --file 
./export/0.1c.export   chown -R ceph.ceph 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/systemctl start ceph-osd@$(cat 
/var/lib/ceph/osd/ceph-5/tmposd_0.1c/whoami)
Sometimes import didn't work.. but stopping OSD and rerunning 
ceph-objectstore-tool again seems to help or when some PG didn't really want to 
import .
Unfound messages are gone!   But I still have down+peering, or 
down+remapped+peering. # ceph health detailHEALTH_ERR 22 pgs are stuck inactive 
for more than 300 seconds; 22 pgs down; 1 pgs inconsistent; 22 pgs peering; 22 
pgs stuck inactive; 22 pgs stuck unclean; 1 requests are blocked > 32 sec; 1 
osds have slow requests; 2 scrub errors; mds cluster is degraded; noout flag(s) 
set; no legacy OSD present but 'sortbitwise' flag is not setpg 1.d is stuck 
inactive since forever, current state down+peering, last acting [11,2]pg 0.a is 
stuck inactive since forever, current state down+remapped+peering, last acting 
[11,7]pg 2.8 is stuck inactive since forever, current state 
down+remapped+peering, last acting [11,7]pg 2.b is stuck inactive since 
forever, current state down+remapped+peering, last acting [7,11]pg 1.9 is stuck 
inactive since forever, current state down+remapped+peering, last acting 
[11,7]pg 0.e is stuck inactive since forever, current state down+peering, last 
acting [11,2]pg 1.3d is stuck inactive since forever, current state 
down+remapped+peering, last acting [10,6]pg 0.2c is stuck inactive since 
forever, current state down+peering, last acting [1,11]pg 0.0 is stuck inactive 
since forever, current state down+remapped+peering, last acting [10,7]pg 1.2b 
is stuck inactive since forever, current state down+peering, last acting 
[1,11]pg 0.29 is stuck inactive since forever, current state down+peering, last 
acting [11,6]pg 1.28 is stuck inactive since forever, current state 
down+peering, last acting [11,6]pg 2.3 is stuck inactive since forever, current 
state down+peering, last acting [11,7]pg 1.1b is stuck inactive since forever, 
current state down+remapped+peering, last acting [11,6]pg 0.d is stuck inactive 
since forever, current state down+remapped+peering, last acting [7,11]pg 1.c is 
stuck inactive since forever, current state down+remapped+peering, last acting 
[7,11]pg 0.3b is stuck inactive since forever, current state 
down+remapped+peering, last acting [10,7]pg 2.39 is stuck inactive since 
forever, current state down+remapped+peering, last acting [10,7]pg 1.3a is 
stuck inactive since forever, current state down+remapped+peering, last acting 
[10,7]pg 0.5 is stuck inactive since forever, current state down+peering, last 
acting [11,7]pg 1.4 is stuck inactive since forever, current state 
down+peering, last acting [11,7]pg 0.1c is stuck inactive since forever, 
current state down+peering, last acting [11,6]pg 1.d is stuck unclean since 
forever, current state down+peering, last acting [11,2]pg 0.a is stuck unclean 
since forever, current state down+remapped+peering, last acting [11,7]pg 2.8 is 
stuck unclean since forever, current state down+remapped+peering, last acting 
[11,7]pg 2.b is stuck unclean since forever, current state 
down+remapped+peering, last acting [7,11]pg 1.9 is stuck unclean since forever, 
current state down+remapped+peering, last acting [11,7]pg 0.e is stuck unclean 
since forever, current state down+peering, last acting [11,2]pg 1.3d is stuck 
unclean since forever, current state down+remapped+peering, last acting 
[10,6]pg 0.d is stuck unclean since forever, current state 
down+remapped+peering, last acting [7,11]pg 1.c is stuck unclean since forever, 
current state down+remapped+peering, last acting [7,11]pg 0.3b is stuck unclean 
since forever, current state down+remapped+peering, last acting [10,7]pg 1.3a 
is stuck unclean since forever, current state down+remapped+peering, last 
acting [10,7]pg 2.39 is stuck unclean since forever, current state 
down+remapped+peering, last acting [10,7]pg 0.5 is stuck unclean since forever, 
current state down+peering, last acting [11,7]pg 1.4 is stuck unclean since 
forever, current state down+peering, last acting [11,7]pg 0.1c is stuck unclean 
since forever, current state down+peering, last acting [11,6]pg 1.1b is stuck 
unclean since forever, current state down+remapped+peering, last acting 
[11,6]pg 2.3 is stuck unclean since forever, current state down+peering, last 
acting [11,7]pg 0.0 is stuck unclean since forever, current state 
down+remapped+peering, last acting [10,7]pg 1.28 is stuck unclean since 
forever, current state down+peering, last acting [11,6]pg 0.29 is stuck unclean 
since forever, current state down+peering, last acting [11,6]pg 1.2b is stuck 
unclean since forever, current state down+peering, last acting [1,11]pg 0.2c is 
stuck unclean since forever, current state down+peering, last acting [1,11]pg 
0.2c is down+peering, acting [1,11]pg 1.2b is down+peering, acting [1,11]pg 
0.29 is down+peering, acting [11,6]pg 1.28 is down+peering, acting [11,6]pg 0.0 
is down+remapped+peering, acting [10,7]pg 2.3 is down+peering, acting [11,7]pg 
1.1b is down+remapped+peering, acting [11,6]pg 0.1c is down+peering, acting 
[11,6]pg 2.39 is down+remapped+peering, acting [10,7]pg 1.3a is 
down+remapped+peering, acting [10,7]pg 0.3b is down+remapped+peering, acting 
[10,7]pg 1.3d is down+remapped+peering, acting [10,6]pg 2.7 is 
active+clean+inconsistent, acting [2,11]pg 1.4 is down+peering, acting [11,7]pg 
0.5 is down+peering, acting [11,7]pg 1.9 is down+remapped+peering, acting 
[11,7]pg 2.b is down+remapped+peering, acting [7,11]pg 2.8 is 
down+remapped+peering, acting [11,7]pg 0.a is down+remapped+peering, acting 
[11,7]pg 1.d is down+peering, acting [11,2]pg 1.c is down+remapped+peering, 
acting [7,11]pg 0.d is down+remapped+peering, acting [7,11]pg 0.e is 
down+peering, acting [11,2]1 ops are blocked > 8388.61 sec on osd.101 osds have 
slow requests2 scrub errorsmds cluster is degradedmds.MDS1.2 at 
192.168.1.20:6801/3142084617 rank 0 is replaying journalnoout flag(s) set
What would be the next step?
Regards,Hong

   
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to