Nice - so I just realized I need to manually scrub 1216 placements groups :)


On 13 March 2015 at 10:16, Andrija Panic <andrija.pa...@gmail.com> wrote:

> Thanks Wido - I will do that.
>
> On 13 March 2015 at 09:46, Wido den Hollander <w...@42on.com> wrote:
>
>>
>>
>> On 13-03-15 09:42, Andrija Panic wrote:
>> > Hi all,
>> >
>> > I have set nodeep-scrub and noscrub while I had small/slow hardware for
>> > the cluster.
>> > It has been off for a while now.
>> >
>> > Now we are upgraded with hardware/networking/SSDs and I would like to
>> > activate - or unset these flags.
>> >
>> > Since I now have 3 servers with 12 OSDs each (SSD based Journals) - I
>> > was wondering what is the best way to unset flags - meaning if I just
>> > unset the flags, should I expect that the SCRUB will start all of the
>> > sudden on all disks - or is there way to let the SCRUB do drives one by
>> > one...
>> >
>>
>> So, I *think* that unsetting these flags will trigger a big scrub, since
>> all PGs have a very old last_scrub_stamp and last_deepscrub_stamp
>>
>> You can verify this with:
>>
>> $ ceph pg <pgid> query
>>
>> A solution would be to scrub each PG manually first in a timely fashion.
>>
>> $ ceph pg scrub <pgid>
>>
>> That way you set the timestamps and slowly scrub each PG.
>>
>> When that's done, unset the flags.
>>
>> Wido
>>
>> > In other words - should I expect BIG performance impact or....not ?
>> >
>> > Any experience is very appreciated...
>> >
>> > Thanks,
>> >
>> > --
>> >
>> > Andrija Panić
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>
> --
>
> Andrija Panić
>



-- 

Andrija Panić
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to