OK...  When I hit the wall, Ceph got pretty unusable; I haven't figured out how 
to restore it to health.  So to do as you suggest, I am going to have to scrape 
everything into the trash and try again (3rd time's the charm, right?) - so let 
me get started on that.  I will pause before I run the big test that can 
overflow the cache and consult with you on what specific steps you might 
recommend.

-don-


-----Original Message-----
From: Nick Fisk [mailto:n...@fisk.me.uk] 
Sent: 30 April, 2015 10:58
To: Don Doerner; ceph-users@lists.ceph.com
Subject: RE: RHEL7/HAMMER cache tier doesn't flush or evict?
Sensitivity: Personal

I'm using Inkscope to monitor my cluster and looking at the pool details I saw 
that mode was set to none. I'm pretty sure there must be a ceph cmd line to get 
the option state but I couldn't find anything obvious when I was looking for it.

> -----Original Message-----
> From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf 
> Of Don Doerner
> Sent: 30 April 2015 18:47
> To: Nick Fisk; ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] RHEL7/HAMMER cache tier doesn't flush or evict?
> Sensitivity: Personal
> 
> Hi Nick,
> 
> For brevity, I didn't detail all of the commands I issued.  Looking 
> back
through
> my command history, I can confirm that I did explicitly set cache-mode 
> to writeback (and later reset it to forward to try flush-and-evict).
Question:
> how did you determine that your cache-mode was not writeback?  I'll do 
> that, just to  confirm that this is the problem, then reestablish the
cache-
> mode.
> 
> Thank you very much for your assistance!
> 
> -don-
> 
> -----Original Message-----
> From: Nick Fisk [mailto:n...@fisk.me.uk]
> Sent: 30 April, 2015 10:38
> To: Don Doerner; ceph-users@lists.ceph.com
> Subject: RE: RHEL7/HAMMER cache tier doesn't flush or evict?
> Sensitivity: Personal
> 
> Hi Don,
> 
> I experienced the same thing a couple of days ago on Hammer. On 
> investigation the cache mode wasn't set to writeback even though I'm 
> sure
it
> accepted the command successfully when I set the pool up.
> 
> Could you reapply the cache mode writeback command and see if that 
> makes a difference?
> 
> Nick
> 
> > -----Original Message-----
> > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On 
> > Behalf Of Don Doerner
> > Sent: 30 April 2015 17:57
> > To: ceph-users@lists.ceph.com
> > Subject: [ceph-users] RHEL7/HAMMER cache tier doesn't flush or evict?
> > Sensitivity: Personal
> >
> > All,
> >
> > Synopsis: I can't get cache tiering to work in HAMMER on RHEL7.
> >
> > Process:
> > 1. Fresh install of HAMMER on RHEL7 went well.
> > 2. Crush map adapted to provide two "root" level resources a.
> > "ctstorage", to use as a cache tier based on very high-performance,
> high
> > IOPS SSD (intrinsic journal).  2 OSDs.
> > b. "ecstorage", to use as an erasure-coded poolbased on 
> > low-performance, cost effective storage (extrinsic journal).  12 OSDs.
> > 3. Established a pool "ctpool", 32 PGs on ctstorage (pool size = 
> > min_size
> = 1).
> > Ran a quick RADOS bench test, all worked as expected.  Cleaned up.
> > 4. Established a pool "ecpool", 256 PGs on ecstorage (5+3 profile).
> > Ran a quick RADOS bench test, all worked as expected.  Cleaned up.
> > 5. Ensured that both pools were empty (i.e., "rados ls" shows no
> > objects) 6. Put the cache tier on the erasure coded storage (one 
> > Bloom hit set, interval 900 seconds), set up the overlay.  Used 
> > defaults for flushing and eviction.  No errors.
> > 7. Started a 3600-second write test to ecpool.
> >
> > Objects piled up in ctpool (as expected) - went past the 40% mark 
> > (as expected), then past the 80% mark (unexpected), then ran into 
> > the wall (95% full - very unexpected).  Using "rados df", I can see 
> > that the cache
> tier is
> > full (duh!) but not one single object lives in the ecpool.  Nothing 
> > was
> ever
> > flushed, nothing was ever evicted.  Thought I might be misreading 
> > that, so
> I
> > went back to SAR data that I captured during the test: the SSDs were 
> > the
> only
> > [ceph] devices that sustained I/O.
> >
> > I based this experiment on another (much more successful) experiment 
> > that I performed using GIANT (.1) on RHEL7 a couple of weeks ago 
> > (wherein I used RAM as a cache tier); that all worked.  It seems 
> > there are at least
> three
> > possibilities.
> > . I forgot a critical step this time around.
> > . The steps needed for a cache tier in HAMMER are different than the 
> > steps needed in GIANT (and different than the documentation online).
> > . There is a problem with HAMMER in the area of cache tier.
> >
> > Has anyone successfully deployed cache-tiering in HAMMER?  Did you 
> > have to do anything unusual?  Do you see any steps that I missed?
> >
> > Regards,
> >
> > -don-
> >
> > ________________________________________
> > The information contained in this transmission may be confidential.
> > Any disclosure, copying, or further distribution of confidential 
> > information
> is not
> > permitted unless such privilege is explicitly granted in writing by
> Quantum.
> > Quantum reserves the right to have electronic communications, 
> > including email and attachments, sent across its networks filtered 
> > through anti
> virus
> > and spam software programs and retain such messages in order to 
> > comply with applicable data security and retention requirements. 
> > Quantum is not responsible for the proper and complete transmission 
> > of the substance of this communication or for any delay in its receipt.
> 
> 
> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> https://urldefense.proofpoint.com/v2/url?u=http-3A__lists.ceph.com_lis
> tinfo.cgi_ceph-2Dusers-2Dceph.com&d=BQICAg&c=8S5idjlO_n28Ko3lg6lskTMwn
> eSC-WqZ5EBTEEvDlkg&r=DAW8QzIBpV_iBddECxqMq8sRPZPQOBqikPmeCBg26bM&m=aNoZ-EWdwG72G_Y6BINvMTf6VjXHuuxJRqXN2Dqn1k0&s=WGWzgEUz8djDKJe0c62GmYz6OqoATjVpYeMHeLiXT_c&e=




_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to