Hello Minchan, excellent analysis! On (01/30/15 23:41), Minchan Kim wrote: > Yes, __srcu_read_lock is a little bit heavier but the number of instruction > are not too much difference to make difference 10%. A culprit is > __cond_resched but I don't think, either because our test was CPU intensive > soS I don't think schedule latency affects total bandwidth. > > More cuprit is your data pattern. > It seems you didn't use scramble_buffers=0, zero_buffers in fio so that > fio fills random data pattern so zram bandwidth could be different by > compression/decompression ratio.
Completely agree. Shame on me. gotten so used to iozone (iozone uses same data pattern 0xA5, this is +Z option what for), so I didn't even think about data pattern in fio. sorry. > 1) randread > srcu is worse as 0.63% but the difference is really marginal. > > 2) randwrite > srcu is better as 1.24% is better. > > 3) randrw > srcu is better as 2.3% hm, interesting. I'll re-check. > Okay, if you concerns on the data still, how about this? I'm not so upset to lose 0.6234187%. my concerns were about iozone's 10% different (which looks a bit worse). I'll review your patch. Thanks for your effort. > > > > by "data pattern" you mean usage scenario? well, I usually use zram for > > `make -jX', where X=[4..N]. so N concurrent read-write ops scenario. > > What I meant is what data fills I/O buffer, which is really important > to evaluate zram because the compression/decompression speeds relys on it. > I see. I never test it with `make' anyway, only iozone +Z. -ss -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/