Hi there,

I changed the SSD on the problematic node with the new one and reconfigure
OSDs and MON service on it.
but the problem occurred again with:

"rocksdb: submit_transaction error: Corruption: block checksum mismatch
code = 2"

I get fully confused now.



On Tue, Feb 20, 2018 at 5:16 PM, Behnam Loghmani <behnam.loghm...@gmail.com>
wrote:

> Hi Caspar,
>
> I checked the filesystem and there isn't any error on filesystem.
> The disk is SSD and it doesn't any attribute related to Wear level in
> smartctl and filesystem is mounted with default options and no discard.
>
> my ceph structure on this node is like this:
>
> it has osd,mon,rgw services
> 1 SSD for OS and WAL/DB
> 2 HDD
>
> OSDs are created by ceph-volume lvm.
>
> the whole SSD is on 1 vg.
> OS is on root lv
> OSD.1 DB is on db-a
> OSD.1 WAL is on wal-a
> OSD.2 DB is on db-b
> OSD.2 WAL is on wal-b
>
> output of lvs:
>
>   data-a data-a -wi-a-----
>
>   data-b data-b -wi-a-----
>   db-a   vg0    -wi-a-----
>
>   db-b   vg0    -wi-a-----
>
>   root   vg0    -wi-ao----
>
>   wal-a  vg0    -wi-a-----
>
>   wal-b  vg0    -wi-a-----
>
> after making a heavy write on the radosgw, OSD.1 and OSD.2 has stopped
> with "block checksum mismatch" error.
> Now on this node MON and OSDs services has stopped working with this error
>
> I think my issue is related to this bug: http://tracker.ceph.com/
> issues/22102
>
> I ran
> #ceph-bluestore-tool fsck --path /var/lib/ceph/osd/ceph-1 --deep 1
> but it returns the same error:
>
> *** Caught signal (Aborted) **
>  in thread 7fbf6c923d00 thread_name:ceph-bluestore-
> 2018-02-20 16:44:30.128787 7fbf6c923d00 -1 abort: Corruption: block
> checksum mismatch
>  ceph version 12.2.2 (cf0baeeeeba3b47f9427c6c97e2144b094b7e5ba) luminous
> (stable)
>  1: (()+0x3eb0b1) [0x55f779e6e0b1]
>  2: (()+0xf5e0) [0x7fbf61ae15e0]
>  3: (gsignal()+0x37) [0x7fbf604d31f7]
>  4: (abort()+0x148) [0x7fbf604d48e8]
>  5: (RocksDBStore::get(std::string const&, char const*, unsigned long,
> ceph::buffer::list*)+0x1ce) [0x55f779d2b5ce]
>  6: (BlueStore::Collection::get_onode(ghobject_t const&, bool)+0x545)
> [0x55f779cd8f75]
>  7: (BlueStore::_fsck(bool, bool)+0x1bb5) [0x55f779cf1a75]
>  8: (main()+0xde0) [0x55f779baab90]
>  9: (__libc_start_main()+0xf5) [0x7fbf604bfc05]
>  10: (()+0x1bc59f) [0x55f779c3f59f]
> 2018-02-20 16:44:30.131334 7fbf6c923d00 -1 *** Caught signal (Aborted) **
>  in thread 7fbf6c923d00 thread_name:ceph-bluestore-
>
>  ceph version 12.2.2 (cf0baeeeeba3b47f9427c6c97e2144b094b7e5ba) luminous
> (stable)
>  1: (()+0x3eb0b1) [0x55f779e6e0b1]
>  2: (()+0xf5e0) [0x7fbf61ae15e0]
>  3: (gsignal()+0x37) [0x7fbf604d31f7]
>  4: (abort()+0x148) [0x7fbf604d48e8]
>  5: (RocksDBStore::get(std::string const&, char const*, unsigned long,
> ceph::buffer::list*)+0x1ce) [0x55f779d2b5ce]
>  6: (BlueStore::Collection::get_onode(ghobject_t const&, bool)+0x545)
> [0x55f779cd8f75]
>  7: (BlueStore::_fsck(bool, bool)+0x1bb5) [0x55f779cf1a75]
>  8: (main()+0xde0) [0x55f779baab90]
>  9: (__libc_start_main()+0xf5) [0x7fbf604bfc05]
>  10: (()+0x1bc59f) [0x55f779c3f59f]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
>     -1> 2018-02-20 16:44:30.128787 7fbf6c923d00 -1 abort: Corruption:
> block checksum mismatch
>      0> 2018-02-20 16:44:30.131334 7fbf6c923d00 -1 *** Caught signal
> (Aborted) **
>  in thread 7fbf6c923d00 thread_name:ceph-bluestore-
>
>  ceph version 12.2.2 (cf0baeeeeba3b47f9427c6c97e2144b094b7e5ba) luminous
> (stable)
>  1: (()+0x3eb0b1) [0x55f779e6e0b1]
>  2: (()+0xf5e0) [0x7fbf61ae15e0]
>  3: (gsignal()+0x37) [0x7fbf604d31f7]
>  4: (abort()+0x148) [0x7fbf604d48e8]
>  5: (RocksDBStore::get(std::string const&, char const*, unsigned long,
> ceph::buffer::list*)+0x1ce) [0x55f779d2b5ce]
>  6: (BlueStore::Collection::get_onode(ghobject_t const&, bool)+0x545)
> [0x55f779cd8f75]
>  7: (BlueStore::_fsck(bool, bool)+0x1bb5) [0x55f779cf1a75]
>  8: (main()+0xde0) [0x55f779baab90]
>  9: (__libc_start_main()+0xf5) [0x7fbf604bfc05]
>  10: (()+0x1bc59f) [0x55f779c3f59f]
>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
> to interpret this.
>
>
>
> Could you please help me to recover this node or find a way to prove SSD
> disk problem.
>
> Best regards,
> Behnam Loghmani
>
>
>
>
> On Mon, Feb 19, 2018 at 1:35 PM, Caspar Smit <caspars...@supernas.eu>
> wrote:
>
>> Hi Behnam,
>>
>> I would firstly recommend running a filesystem check on the monitor disk
>> first to see if there are any inconsistencies.
>>
>> Is the disk where the monitor is running on a spinning disk or SSD?
>>
>> If SSD you should check the Wear level stats through smartctl.
>> Maybe trim (discard) enabled on the filesystem mount? (discard could
>> cause problems/corruption in combination with certain SSD firmwares)
>>
>> Caspar
>>
>> 2018-02-16 23:03 GMT+01:00 Behnam Loghmani <behnam.loghm...@gmail.com>:
>>
>>> I checked the disk that monitor is on it with smartctl and it didn't
>>> return any error and it doesn't have any Current_Pending_Sector.
>>> Do you recommend any disk checks to make sure that this disk has problem
>>> and then I can send the report to the provider for replacing the disk
>>>
>>> On Sat, Feb 17, 2018 at 1:09 AM, Gregory Farnum <gfar...@redhat.com>
>>> wrote:
>>>
>>>> The disk that the monitor is on...there isn't anything for you to
>>>> configure about a monitor WAL though so I'm not sure how that enters into
>>>> it?
>>>>
>>>> On Fri, Feb 16, 2018 at 12:46 PM Behnam Loghmani <
>>>> behnam.loghm...@gmail.com> wrote:
>>>>
>>>>> Thanks for your reply
>>>>>
>>>>> Do you mean, that's the problem with the disk I use for WAL and DB?
>>>>>
>>>>> On Fri, Feb 16, 2018 at 11:33 PM, Gregory Farnum <gfar...@redhat.com>
>>>>> wrote:
>>>>>
>>>>>>
>>>>>> On Fri, Feb 16, 2018 at 7:37 AM Behnam Loghmani <
>>>>>> behnam.loghm...@gmail.com> wrote:
>>>>>>
>>>>>>> Hi there,
>>>>>>>
>>>>>>> I have a Ceph cluster version 12.2.2 on CentOS 7.
>>>>>>>
>>>>>>> It is a testing cluster and I have set it up 2 weeks ago.
>>>>>>> after some days, I see that one of the three mons has stopped(out of
>>>>>>> quorum) and I can't start it anymore.
>>>>>>> I checked the mon service log and the output shows this error:
>>>>>>>
>>>>>>> """
>>>>>>> mon.XXXXXX@-1(probing) e4 preinit clean up potentially inconsistent
>>>>>>> store state
>>>>>>> rocksdb: submit_transaction_sync error: Corruption: block checksum
>>>>>>> mismatch
>>>>>>>
>>>>>>
>>>>>> This bit is the important one. Your disk is bad and it’s feeding back
>>>>>> corrupted data.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>> code = 2 Rocksdb transaction:
>>>>>>>      0> 2018-02-16 17:37:07.041812 7f45a1e52e40 -1
>>>>>>> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/A
>>>>>>> VAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MAC
>>>>>>> HINE_SIZE/huge/release/12.2.2/rpm/el7/BUI
>>>>>>> LD/ceph-12.2.2/src/mon/MonitorDBStore.h: In function 'void
>>>>>>> MonitorDBStore::clear(std::set<std::basic_string<char> >&)' thread
>>>>>>> 7f45a1e52e40 time 2018-02-16 17:37:07.040846
>>>>>>> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/A
>>>>>>> VAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MAC
>>>>>>> HINE_SIZE/huge/release/12.2.2/rpm/el7/BUILD/ceph-12.2.2/src/mon/MonitorDBStore.h:
>>>>>>> 581: FAILE
>>>>>>> D assert(r >= 0)
>>>>>>> """
>>>>>>>
>>>>>>> the only solution I found is to remove this mon from quorum and
>>>>>>> remove all mon data and re-add this mon to quorum again.
>>>>>>> and ceph goes to the healthy status again.
>>>>>>>
>>>>>>> but now after some days this mon has stopped and I face the same
>>>>>>> problem again.
>>>>>>>
>>>>>>> My cluster setup is:
>>>>>>> 4 osd hosts
>>>>>>> total 8 osds
>>>>>>> 3 mons
>>>>>>> 1 rgw
>>>>>>>
>>>>>>> this cluster has setup with ceph-volume lvm and wal/db separation on
>>>>>>> logical volumes.
>>>>>>>
>>>>>>> Best regards,
>>>>>>> Behnam Loghmani
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> ceph-users mailing list
>>>>>>> ceph-users@lists.ceph.com
>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>
>>>>>>
>>>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to