Hi
We have a 4 physical nodes cluster running Jewel, our app talks S3 to
the cluster and uses S3 index heavily no-doubt. We've had several big
outages in the past that seem caused by a deep-scrub on one of the PGs
in S3 index pool. Generally it starts from a deep scrub on one such PG
then follows with lots of slow requests blocking/accumulating which
eventually makes the whole cluster down. In the event like this, we have
to set OSD to noup/nodown/noout to let OSD not suicide during such
deep-scrub.
In a recent outage, the deep-scrub of one PG took 2 hours to finish,
after finished, I happened to try listing all omap keys of the objects
in that PG and found that listing keys of one particular object can
cause same outage described above, it indicates to me that the index
object was corrupted, but I can't find anything in the logs.
Interestingly (to me), 2 days later that index object seems have fixed
itself: listing omap keys quick and easy, deep-scrubbing same PG only
takes 3 seconds.
The deep-scrub that took 2 hours to finish:
xxxx.log-20170730.gz:2017-07-29 12:14:10.476325 osd.2 x.x.x.x:6800/78482
217 : cluster [INF] 11.11 deep-scrub starts
xxxx.log-20170730.gz:2017-07-29 14:05:12.108523 osd.2
x.x.x.203:6800/78482 1795 : cluster [INF] 11.11 deep-scrub ok
The command I used to list all omap keys:
rados -p .rgw.buckets.index listomapkeys
.dir.c82cdc62-7926-440d-8085-4e7879ef8155.26048.647 | wc -l
Most recent deep-scrub kicked off manually:
2017-07-31 09:54:37.997911 7f78bc333700 0 log_channel(cluster) log
[INF] : 11.11 deep-scrub starts
2017-07-31 09:54:40.539494 7f78bc333700 0 log_channel(cluster) log
[INF] : 11.11 deep-scrub ok
Setting debug_leveldb to 20/5 didn't log any useful information for the
event, sorry, but a perf record shows most (83%) of the time was used on
LevelDB operations (screenshot or perf file can be supplied if anybody
interested since it's over 150KB size limit.).
I wonder if anybody came across similar issue before or can explain what
happened to the index object to make it not-usable before but usable 2
days later? One thing that might fix the index object is leveldb
compactions I guess. By the way the above problematic index object has
~30k keys, the biggest index object in our cluster holds about 300k keys.
Regards
Stanley
--
*Stanley Zhang | * Senior Operations Engineer
*Telephone:* +64 9 302 0515 *Fax:* +64 9 302 0518
*Mobile:* +64 22 318 3664 *Freephone:* 0800 SMX SMX (769 769)
*SMX Limited:* Level 15, 19 Victoria Street West, Auckland, New Zealand
*Web:* http://smxemail.com
SMX | Cloud Email Hosting & Security
_____________________________________________________________________________
This email has been filtered by SMX. For more info visit http://smxemail.com
_____________________________________________________________________________
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com