Interesting....thx for that Henrik.

BTW, my placements groups are arround 1800 objects (ceph pg dump) - meainng
max of 7GB od data at the moment,

regular scrub just took 5-10sec to finish. Deep scrub would I guess take
some minutes for sure

What about deepscrub - timestamp is still some months ago, but regular
scrub is fine now with fresh timestamp...?

I don't see max deep scrub setings - or are these settings applied in
general for both kind on scrubs ?

Thanks



On 13 March 2015 at 12:22, Henrik Korkuc <li...@kirneh.eu> wrote:

>  I think that there will be no big scrub, as there are limits of maximum
> scrubs at a time.
> http://ceph.com/docs/master/rados/configuration/osd-config-ref/#scrubbing
>
> If we take "osd max scrubs" which is 1 by default, then you will not get
> more than 1 scrub per OSD.
>
> I couldn't quickly find if there are cluster wide limits.
>
>
> On 3/13/15 10:46, Wido den Hollander wrote:
>
>
> On 13-03-15 09:42, Andrija Panic wrote:
>
>  Hi all,
>
> I have set nodeep-scrub and noscrub while I had small/slow hardware for
> the cluster.
> It has been off for a while now.
>
> Now we are upgraded with hardware/networking/SSDs and I would like to
> activate - or unset these flags.
>
> Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I
> was wondering what is the best way to unset flags - meaning if I just
> unset the flags, should I expect that the SCRUB will start all of the
> sudden on all disks - or is there way to let the SCRUB do drives one by
> one...
>
>
>  So, I *think* that unsetting these flags will trigger a big scrub, since
> all PGs have a very old last_scrub_stamp and last_deepscrub_stamp
>
> You can verify this with:
>
> $ ceph pg <pgid> query
>
> A solution would be to scrub each PG manually first in a timely fashion.
>
> $ ceph pg scrub <pgid>
>
> That way you set the timestamps and slowly scrub each PG.
>
> When that's done, unset the flags.
>
> Wido
>
>
>  In other words - should I expect BIG performance impact or....not ?
>
> Any experience is very appreciated...
>
> Thanks,
>
> --
>
> Andrija Panić
>
>
> _______________________________________________
> ceph-users mailing 
> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>  _______________________________________________
> ceph-users mailing 
> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>


-- 

Andrija Panić
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to