> -----Original Message-----
> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of 
> Eugen Block
> Sent: 22 November 2016 10:11
> To: Nick Fisk <n...@fisk.me.uk>
> Cc: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] deep-scrubbing has large impact on performance
> 
> Thanks for the very quick answer!
> 
> > If you are using Jewel
> 
> We are still using Hammer (0.94.7), we wanted to upgrade to Jewel in a couple 
> of weeks, would you recommend to do it now?

It's been fairly solid for me, but you might want to wait for the scrubbing 
hang bug to be fixed before upgrading. I think this
might be fixed in the upcoming 10.2.4 release.

> 
> 
> Zitat von Nick Fisk <n...@fisk.me.uk>:
> 
> >> -----Original Message-----
> >> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf
> >> Of Eugen Block
> >> Sent: 22 November 2016 09:55
> >> To: ceph-users@lists.ceph.com
> >> Subject: [ceph-users] deep-scrubbing has large impact on performance
> >>
> >> Hi list,
> >>
> >> I've been searching the mail archive and the web for some help. I
> >> tried the things I found, but I can't see the effects. We use
> > Ceph for
> >> our Openstack environment.
> >>
> >> When our cluster (2 pools, each 4092 PGs, in 20 OSDs on 4 nodes, 3
> >> MONs) starts deep-scrubbing, it's impossible to work with the VMs.
> >> Currently, the deep-scrubs happen to start on Monday, which is
> >> unfortunate. I already plan to start the next deep-scrub on
> > Saturday,
> >> so it has no impact on our work days. But if I imagine we had a large
> >> multi-datacenter, such performance breaks are not
> > reasonable. So
> >> I'm wondering how do you guys manage that?
> >>
> >> What I've tried so far:
> >>
> >> ceph tell osd.* injectargs '--osd_scrub_sleep 0.1'
> >> ceph tell osd.* injectargs '--osd_disk_thread_ioprio_priority 7'
> >> ceph tell osd.* injectargs '--osd_disk_thread_ioprio_class idle'
> >> ceph tell osd.* injectargs '--osd_scrub_begin_hour 0'
> >> ceph tell osd.* injectargs '--osd_scrub_end_hour 7'
> >>
> >> And I also added these options to the ceph.conf.
> >> To be able to work again, I had to set the nodeep-scrub option and
> >> unset it when I left the office. Today, I see the cluster deep-
> >> scrubbing again, but only one PG at a time, it seems that now the
> >> default for osd_max_scrubs is working now and I don't see major
> >> impacts yet.
> >>
> >> But is there something else I can do to reduce the performance impact?
> >
> > If you are using Jewel, the scrubing is now done in the client IO
> > thread, so those disk thread options won't do anything. Instead there
> > is a new priority setting, which seems to work for me, along with a
> > few other settings.
> >
> > osd_scrub_priority = 1
> > osd_scrub_sleep = .1
> > osd_scrub_chunk_min = 1
> > osd_scrub_chunk_max = 5
> > osd_scrub_load_threshold = 5
> >
> > Also enabling the weighted priority queue can assist the new priority
> > options
> >
> > osd_op_queue = wpq
> >
> >
> >> I just found [1] and will have a look into it.
> >>
> >> [1] http://prob6.com/en/ceph-pg-deep-scrub-cron/
> >>
> >> Thanks!
> >> Eugen
> >>
> >> --
> >> Eugen Block                             voice   : +49-40-559 51 75
> >> NDE Netzdesign und -entwicklung AG      fax     : +49-40-559 51 77
> >> Postfach 61 03 15
> >> D-22423 Hamburg                         e-mail  : ebl...@nde.ag
> >>
> >>          Vorsitzende des Aufsichtsrates: Angelika Mozdzen
> >>            Sitz und Registergericht: Hamburg, HRB 90934
> >>                    Vorstand: Jens-U. Mozdzen
> >>                     USt-IdNr. DE 814 013 983
> >>
> >> _______________________________________________
> >> ceph-users mailing list
> >> ceph-users@lists.ceph.com
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
> 
> --
> Eugen Block                             voice   : +49-40-559 51 75
> NDE Netzdesign und -entwicklung AG      fax     : +49-40-559 51 77
> Postfach 61 03 15
> D-22423 Hamburg                         e-mail  : ebl...@nde.ag
> 
>          Vorsitzende des Aufsichtsrates: Angelika Mozdzen
>            Sitz und Registergericht: Hamburg, HRB 90934
>                    Vorstand: Jens-U. Mozdzen
>                     USt-IdNr. DE 814 013 983
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to