On Fri, Feb 5, 2016 at 6:39 AM, Stephen Lord <steve.l...@quantum.com> wrote:
>
> I looked at this system this morning, and the it actually finished what it was
> doing. The erasure coded pool still contains all the data and the cache
> pool has about a million zero sized objects:
>
>
> GLOBAL:
>     SIZE       AVAIL     RAW USED     %RAW USED     OBJECTS
>     15090G     9001G        6080G         40.29       2127k
> POOLS:
>     NAME                ID     CATEGORY     USED       %USED     MAX AVAIL    
>  OBJECTS     DIRTY     READ       WRITE
>     cache-data          21     -                 0         0         7962G    
>  1162258     1057k      22969     3220k
>     cephfs-data         22     -             3964G     26.27         5308G    
>  1014840      991k       891k     1143k
>
> Definitely seems like a bug since I removed all references to these from the 
> filesystem
> which created them.
>
> I originally wrote 4.5 Tbytes of data into the file system, the erasure coded
> pool is setup as 4+2, and the cache has a size limit of 1 Tbyte. Looks like 
> not
> all the data made it out of the cache tier before I removed content, it 
> removed the
> content which was only present in the cache tier and created a zero sized 
> object
> in the cache for all the content. The used capacity is somewhat consistent 
> with
> this.
>
> I tried to look at the extended attributes on one of the zero size object 
> with ceph-dencoder,
> but it failed:
>
> error: buffer::malformed_input: void 
> object_info_t::decode(ceph::buffer::list::iterator&) unknown encoding version 
> > 15
>
> Same error on one of the objects in the erasure coded pool.
>
> Looks like I am a little too bleeding edge for this, or the contents of the 
> .ceph_ attribute are not an object_info_t

ghobject_info_t

You can get the EC stuff actually deleted by getting the cache pool to
flush everything. That's discussed in the docs and in various mailing
list archives.
-Greg

>
>
>
> Steve
>
>> On Feb 4, 2016, at 7:10 PM, Gregory Farnum <gfar...@redhat.com> wrote:
>>
>> On Thu, Feb 4, 2016 at 5:07 PM, Stephen Lord <steve.l...@quantum.com> wrote:
>>>
>>>> On Feb 4, 2016, at 6:51 PM, Gregory Farnum <gfar...@redhat.com> wrote:
>>>>
>>>> I presume we're doing reads in order to gather some object metadata
>>>> from the cephfs-data pool; and the (small) newly-created objects in
>>>> cache-data are definitely whiteout objects indicating the object no
>>>> longer exists logically.
>>>>
>>>> What kinds of reads are you actually seeing? Does it appear to be
>>>> transferring data, or merely doing a bunch of seeks? I thought we were
>>>> trying to avoid doing reads-to-delete, but perhaps the way we're
>>>> handling snapshots or something is invoking behavior that isn't
>>>> amicable to a full-FS delete.
>>>>
>>>> I presume you're trying to characterize the system's behavior, but of
>>>> course if you just want to empty it out entirely you're better off
>>>> deleting the pools and the CephFS instance entirely and then starting
>>>> it over again from scratch.
>>>> -Greg
>>>
>>> I believe it is reading all the data, just from the volume of traffic and
>>> the cpu load on the OSDs maybe suggests it is doing more than
>>> just that.
>>>
>>> iostat is showing a lot of data moving, I am seeing about the same volume
>>> of read and write activity here. Because the OSDs underneath both pools
>>> are the same ones, I know that’s not exactly optimal, it is hard to tell 
>>> what
>>> which pool is responsible for which I/O. Large reads and small writes 
>>> suggest
>>> it is reading up all the data from the objects,  the write traffic is I 
>>> presume all
>>> journal activity relating to deleting objects and creating the empty ones.
>>>
>>> The 9:1 ratio between things being deleted and created seems odd though.
>>>
>>> A previous version of this exercise with just a regular replicated data pool
>>> did not read anything, just a lot of write activity and eventually the 
>>> content
>>> disappeared. So definitely related to the pool configuration here and 
>>> probably
>>> not to the filesystem layer.
>>
>> Sam, does this make any sense to you in terms of how RADOS handles deletes?
>> -Greg
>
>
> ----------------------------------------------------------------------
> The information contained in this transmission may be confidential. Any 
> disclosure, copying, or further distribution of confidential information is 
> not permitted unless such privilege is explicitly granted in writing by 
> Quantum. Quantum reserves the right to have electronic communications, 
> including email and attachments, sent across its networks filtered through 
> anti virus and spam software programs and retain such messages in order to 
> comply with applicable data security and retention requirements. Quantum is 
> not responsible for the proper and complete transmission of the substance of 
> this communication or for any delay in its receipt.
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to