I just upgraded a cephadm cluster from 15.2.1 to 15.2.2.
Everything went fine on the upgrade, however after restarting one node that has
3 OSD's for ecmeta two of the 3 ODS's now wont boot with the following error:
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.598+0000
7fbcc46f7ec0 4 rocksdb: [db/version_set.cc:3757] Recovered from manifest
file:db/MANIFEST-002768 succeeded,manifest_file_number is 2768,
next_file_number is 2775, last_sequence is 188026749, log_number is
2767,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.598+0000
7fbcc46f7ec0 4 rocksdb: [db/version_set.cc:3766] Column family [default] (ID
0), log number is 2767
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.598+0000
7fbcc46f7ec0 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1589963382599157, "job":
1, "event": "recovery_started", "log_files": [2769]}
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.598+0000
7fbcc46f7ec0 4 rocksdb: [db/db_impl_open.cc:583] Recovering log #2769 mode 0
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.598+0000
7fbcc46f7ec0 3 rocksdb: [db/db_impl_open.cc:518] db/002769.log: dropping
537526 bytes; Corruption: error in middle of record
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.598+0000
7fbcc46f7ec0 3 rocksdb: [db/db_impl_open.cc:518] db/002769.log: dropping 32757
bytes; Corruption: missing start of fragmented record(1)
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 3 rocksdb: [db/db_impl_open.cc:518] db/002769.log: dropping 32757
bytes; Corruption: missing start of fragmented record(1)
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 3 rocksdb: [db/db_impl_open.cc:518] db/002769.log: dropping 32757
bytes; Corruption: missing start of fragmented record(1)
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 3 rocksdb: [db/db_impl_open.cc:518] db/002769.log: dropping 32757
bytes; Corruption: missing start of fragmented record(1)
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 3 rocksdb: [db/db_impl_open.cc:518] db/002769.log: dropping 32757
bytes; Corruption: missing start of fragmented record(1)
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 3 rocksdb: [db/db_impl_open.cc:518] db/002769.log: dropping 32757
bytes; Corruption: missing start of fragmented record(1)
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 3 rocksdb: [db/db_impl_open.cc:518] db/002769.log: dropping 23263
bytes; Corruption: missing start of fragmented record(2)
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 4 rocksdb: [db/db_impl.cc:390] Shutdown: canceling all background
work
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 4 rocksdb: [db/db_impl.cc:563] Shutdown complete
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 -1 rocksdb: Corruption: error in middle of record
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 -1 bluestore(/var/lib/ceph/osd/ceph-0) _open_db erroring opening
db:
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.602+0000
7fbcc46f7ec0 1 bdev(0x558a28dd0700 /var/lib/ceph/osd/ceph-0/block) close
May 20 08:29:42 sn-m01 bash[6833]: debug 2020-05-20T08:29:42.870+0000
7fbcc46f7ec0 1 bdev(0x558a28dd0000 /var/lib/ceph/osd/ceph-0/block) close
May 20 08:29:43 sn-m01 bash[6833]: debug 2020-05-20T08:29:43.118+0000
7fbcc46f7ec0 -1 osd.0 0 OSD:init: unable to mount object store
May 20 08:29:43 sn-m01 bash[6833]: debug 2020-05-20T08:29:43.118+0000
7fbcc46f7ec0 -1 ** ERROR: osd init failed: (5) Input/output error
Have I hit a bug, or is there something I can do to try and fix these OSD's?
Thanks
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]