Re: [ceph-users] RBD: How many snapshots is too many?

2017-09-11 Thread Mclean, Patrick
On 2017-09-08 06:06 PM, Gregory Farnum wrote: > On Fri, Sep 8, 2017 at 5:47 PM, Mclean, Patrick > wrote: > >> On a related note, we are very curious why the snapshot id is >> incremented when a snapshot is deleted, this creates lots >> phantom entries in the deleted

Re: [ceph-users] RBD: How many snapshots is too many?

2017-09-08 Thread Mclean, Patrick
On 2017-09-08 01:36 PM, Gregory Farnum wrote: > On Thu, Sep 7, 2017 at 1:46 PM, Mclean, Patrick > wrote: >> On 2017-09-05 02:41 PM, Gregory Farnum wrote: >>> On Tue, Sep 5, 2017 at 1:44 PM, Florian Haas > wrote: >>> >> Hi everyone, >> >> with

Re: [ceph-users] RBD: How many snapshots is too many?

2017-09-08 Thread Mclean, Patrick
On 2017-09-08 01:59 PM, Gregory Farnum wrote: > On Fri, Sep 8, 2017 at 1:45 AM, Florian Haas wrote: >>> In our use case, we are severly hampered by the size of removed_snaps >>> (50k+) in the OSDMap to the point were ~80% of ALL cpu time is spent in >>> PGPool::update and its interval calculation

Re: [ceph-users] RBD: How many snapshots is too many?

2017-09-07 Thread Mclean, Patrick
On 2017-09-05 02:41 PM, Gregory Farnum wrote: > On Tue, Sep 5, 2017 at 1:44 PM, Florian Haas > wrote: > >> Hi everyone, >> >> with the Luminous release out the door and the Labor Day weekend >> over, I hope I can kick off a discussion on another issue that has >> irked me a bit for quite a while.

Re: [ceph-users] ceph cluster experiencing major performance issues

2017-08-08 Thread Mclean, Patrick
hot count down below 510), we are currently testing the possibility of using the nbd-rbd driver as an alternative. > On Mon, Aug 7, 2017 at 11:49 PM Mclean, Patrick <mailto:patrick.mcl...@sony.com>> wrote: > > High CPU utilization and inexplicably slow I/O requests > > We

[ceph-users] ceph cluster experiencing major performance issues

2017-08-07 Thread Mclean, Patrick
High CPU utilization and inexplicably slow I/O requests We have been having similar performance issues across several ceph clusters. When all the OSDs are up in the cluster, it can stay HEALTH_OK for a while, but eventually performance worsens and becomes (at first intermittently, but eventually c