If I understand correctly you have one sas disk as a journal for multiple OSDs.
If you do small synchronous writes it will become a IO bottleneck pretty 
quickly:
Due to multiple journals on the same disk it will no longer be sequential 
writes writes to one journal but  4k writes to x journals making it fully 
random.
I would expect a performance of 100 to 200 IOPS max.
Doing an iostat -x or atop should show this bottleneck immediately.
This is also the reason to go with SSDs: they have reasonable random IO 
performance.

Cheers,
Robert van Leeuwen

Sent from my iPad

> On 6 dec. 2013, at 17:05, "nicolasc" <nicolas.cance...@surfsara.nl> wrote:
> 
> Hi James,
> 
> Thank you for this clarification. I am quite aware of that, which is why the 
> journals are on SAS disks in RAID0 (SSDs out of scope).
> 
> I still have trouble believing that fast-but-not-super-fast journals is the 
> main reason for the poor performances observed. Maybe I am mistaken?
> 
> Best regards,
> 
> Nicolas Canceill
> Scalable Storage Systems
> SURFsara (Amsterdam, NL)
> 
> 
> 
> On 12/03/2013 03:01 PM, James Pearce wrote:
>>> I would really appreciate it if someone could:
>>> - explain why the journal setup is way more important than striping 
>>> settings;
>> 
>> I'm not sure if it's what you're asking, but any write must be physically 
>> written to the journal before the operation is acknowledged.  So the overall 
>> cluster performance (or rather write latency) is always governed by the 
>> speed of those journals.  Data is then gathered up into (hopefully) larger 
>> blocks and committed to OSDs later.
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to