We have not seen this issue, but we don't run EC pools yet (we are waiting
for multiple layers to be available). We are not running 0.94.6 in
production yet either. We have adopted the policy to only run released
versions in production unless there is a really pressing need to have a
patch. We are running 0.94.6 through our alpha and staging clusters and
hoping to do the upgrade in the next couple of weeks. We won't know how
much the recency fix will help until then because we have not been able to
replicate our workload with fio accurately enough to get good test results.
Unfortunately we will probably be swapping out our M600s with S3610s. We've
burned through 30% of the life in 2 months and they have 8x the op latency.
Due to the 10 Minutes of Terror, we are going to have to do both at the
same time to reduce the impact. Luckily, when you have weighted out OSDs or
empty ones, it is much less impactful. If you get your upgrade done before
ours, I'd like to know how it went. I'll be posting the results from ours
when it is done.

Sent from a mobile device, please excuse any typos.
On Feb 24, 2016 5:43 PM, "Christian Balzer" <ch...@gol.com> wrote:

>
> Hello Jason (Ceph devs et al),
>
> On Wed, 24 Feb 2016 13:15:34 -0500 (EST) Jason Dillaman wrote:
>
> > If you run "rados -p <cache pool> ls | grep "rbd_id.<yyy-disk1>" and
> > don't see that object, you are experiencing that issue [1].
> >
> > You can attempt to work around this issue by running "rados -p irfu-virt
> > setomapval rbd_id.<yyy-disk1> dummy value" to force-promote the object
> > to the cache pool.  I haven't tested / verified that will alleviate the
> > issue, though.
> >
> > [1] http://tracker.ceph.com/issues/14762
> >
>
> This concerns me greatly, as I'm about to phase in a cache tier this
> weekend into a very busy, VERY mission critical Ceph cluster.
> That is on top of a replicated pool, Hammer.
>
> That issue and the related git blurb are less than crystal clear, so for
> my and everybody else's benefit could you elaborate a bit more on this?
>
> 1. Does this only affect EC base pools?
> 2. Is this a regressions of sorts and when came it about?
>    I have a hard time imagining people not running into this earlier,
>    unless that problem is very hard to trigger.
> 3. One assumes that this isn't fixed in any released version of Ceph,
>    correct?
>
> Robert, sorry for CC'ing you, but AFAICT your cluster is about the closest
> approximation in terms of busyness to mine here.
> And I a assume that you're neither using EC pools (since you need
> performance, not space) and haven't experienced this bug all?
>
> Also, would you consider the benefits of the recency fix (thanks for
> that) being worth risk of being an early adopter of 0.94.6?
> In other words, are you eating your own dog food already and 0.94.6 hasn't
> eaten your data babies yet? ^o^
>
> Regards,
>
> Christian
> --
> Christian Balzer        Network/Systems Engineer
> ch...@gol.com           Global OnLine Japan/Rakuten Communications
> http://www.gol.com/
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to