Am 22.05.2012 23:11, schrieb Greg Farnum: > On Tuesday, May 22, 2012 at 2:00 PM, Stefan Priebe wrote: >> Am 22.05.2012 22:49, schrieb Greg Farnum: >>> Anyway, it looks like you're just paying a synchronous write penalty >> >> >> What does that exactly mean? Shouldn't one threaded write to four >> 260MB/s devices gives at least 100Mb/s? > > Well, with dd you've got a single thread issuing synchronous IO requests to > the kernel. We could have it set up so that those synchronous requests get > split up, but they aren't, and between the kernel and KVM it looks like when > it needs to make a write out to disk it sends one request at a time to the > Ceph backend. So you aren't writing to four 260MB/s devices; you are writing > to one 260MB/s device without any pipelining — meaning you send off a 4MB > write, then wait until it's done, then send off a second 4MB write, then wait > until it's done, etc. > Frankly I'm surprised you aren't getting a bit more throughput than you're > seeing (I remember other people getting much more out of less beefy boxes), > but it doesn't much matter because what you really want to do is enable the > client-side writeback cache in RBD, which will dispatch multiple requests at > once and not force writes to be committed before reporting back to the > kernel. Then you should indeed be writing to four 260MB/s devices at once. :)
OK i understand that but still the question where is the bottlenek in this case. I mean i see not more than 40% network load, not more than 10% cpu load and only 40MB/s to the SSD. I would still expect a network load of 70-90%. Greets and thanks, Stefan -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html