Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
Thanks Wido - I will do that. On 13 March 2015 at 09:46, Wido den Hollander w...@42on.com wrote: On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
On 13-03-15 12:00, Andrija Panic wrote: Nice - so I just realized I need to manually scrub 1216 placements groups :) With manual I meant using a script. Loop through 'ceph pg dump', get the PGid, issue a scrub, sleep for X seconds and issue the next scrub. Wido On 13 March 2015 at 10:16, Andrija Panic andrija.pa...@gmail.com mailto:andrija.pa...@gmail.com wrote: Thanks Wido - I will do that. On 13 March 2015 at 09:46, Wido den Hollander w...@42on.com mailto:w...@42on.com wrote: On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com mailto:ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com mailto:ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Andrija Panić -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
Nice - so I just realized I need to manually scrub 1216 placements groups :) On 13 March 2015 at 10:16, Andrija Panic andrija.pa...@gmail.com wrote: Thanks Wido - I will do that. On 13 March 2015 at 09:46, Wido den Hollander w...@42on.com wrote: On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Andrija Panić -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
Interestingthx for that Henrik. BTW, my placements groups are arround 1800 objects (ceph pg dump) - meainng max of 7GB od data at the moment, regular scrub just took 5-10sec to finish. Deep scrub would I guess take some minutes for sure What about deepscrub - timestamp is still some months ago, but regular scrub is fine now with fresh timestamp...? I don't see max deep scrub setings - or are these settings applied in general for both kind on scrubs ? Thanks On 13 March 2015 at 12:22, Henrik Korkuc li...@kirneh.eu wrote: I think that there will be no big scrub, as there are limits of maximum scrubs at a time. http://ceph.com/docs/master/rados/configuration/osd-config-ref/#scrubbing If we take osd max scrubs which is 1 by default, then you will not get more than 1 scrub per OSD. I couldn't quickly find if there are cluster wide limits. On 3/13/15 10:46, Wido den Hollander wrote: On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
Hmnice Thx guys On 13 March 2015 at 12:33, Henrik Korkuc li...@kirneh.eu wrote: I think settings apply to both kinds of scrubs On 3/13/15 13:31, Andrija Panic wrote: Interestingthx for that Henrik. BTW, my placements groups are arround 1800 objects (ceph pg dump) - meainng max of 7GB od data at the moment, regular scrub just took 5-10sec to finish. Deep scrub would I guess take some minutes for sure What about deepscrub - timestamp is still some months ago, but regular scrub is fine now with fresh timestamp...? I don't see max deep scrub setings - or are these settings applied in general for both kind on scrubs ? Thanks On 13 March 2015 at 12:22, Henrik Korkuc li...@kirneh.eu wrote: I think that there will be no big scrub, as there are limits of maximum scrubs at a time. http://ceph.com/docs/master/rados/configuration/osd-config-ref/#scrubbing If we take osd max scrubs which is 1 by default, then you will not get more than 1 scrub per OSD. I couldn't quickly find if there are cluster wide limits. On 3/13/15 10:46, Wido den Hollander wrote: On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Andrija Panić -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
Will do, of course :) THx Wido for quick help, as always ! On 13 March 2015 at 12:04, Wido den Hollander w...@42on.com wrote: On 13-03-15 12:00, Andrija Panic wrote: Nice - so I just realized I need to manually scrub 1216 placements groups :) With manual I meant using a script. Loop through 'ceph pg dump', get the PGid, issue a scrub, sleep for X seconds and issue the next scrub. Wido On 13 March 2015 at 10:16, Andrija Panic andrija.pa...@gmail.com mailto:andrija.pa...@gmail.com wrote: Thanks Wido - I will do that. On 13 March 2015 at 09:46, Wido den Hollander w...@42on.com mailto:w...@42on.com wrote: On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com mailto:ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com mailto:ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Andrija Panić -- Andrija Panić -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
I think that there will be no big scrub, as there are limits of maximum scrubs at a time. http://ceph.com/docs/master/rados/configuration/osd-config-ref/#scrubbing If we take osd max scrubs which is 1 by default, then you will not get more than 1 scrub per OSD. I couldn't quickly find if there are cluster wide limits. On 3/13/15 10:46, Wido den Hollander wrote: On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Turning on SCRUB back on - any suggestion ?
I think settings apply to both kinds of scrubs On 3/13/15 13:31, Andrija Panic wrote: Interestingthx for that Henrik. BTW, my placements groups are arround 1800 objects (ceph pg dump) - meainng max of 7GB od data at the moment, regular scrub just took 5-10sec to finish. Deep scrub would I guess take some minutes for sure What about deepscrub - timestamp is still some months ago, but regular scrub is fine now with fresh timestamp...? I don't see max deep scrub setings - or are these settings applied in general for both kind on scrubs ? Thanks On 13 March 2015 at 12:22, Henrik Korkuc li...@kirneh.eu mailto:li...@kirneh.eu wrote: I think that there will be no big scrub, as there are limits of maximum scrubs at a time. http://ceph.com/docs/master/rados/configuration/osd-config-ref/#scrubbing If we take osd max scrubs which is 1 by default, then you will not get more than 1 scrub per OSD. I couldn't quickly find if there are cluster wide limits. On 3/13/15 10:46, Wido den Hollander wrote: On 13-03-15 09:42, Andrija Panic wrote: Hi all, I have set nodeep-scrub and noscrub while I had small/slow hardware for the cluster. It has been off for a while now. Now we are upgraded with hardware/networking/SSDs and I would like to activate - or unset these flags. Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I was wondering what is the best way to unset flags - meaning if I just unset the flags, should I expect that the SCRUB will start all of the sudden on all disks - or is there way to let the SCRUB do drives one by one... So, I *think* that unsetting these flags will trigger a big scrub, since all PGs have a very old last_scrub_stamp and last_deepscrub_stamp You can verify this with: $ ceph pg pgid query A solution would be to scrub each PG manually first in a timely fashion. $ ceph pg scrub pgid That way you set the timestamps and slowly scrub each PG. When that's done, unset the flags. Wido In other words - should I expect BIG performance impact ornot ? Any experience is very appreciated... Thanks, -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com mailto:ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com mailto:ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com mailto:ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Andrija Panić ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com