> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Erik Trimble
> 
> (1) Unless you are using Zvols for "raw" disk partitions (for use with
> something like a database), the recordsize value is a MAXIMUM value, NOT
> an absolute value.  Thus, if you have a ZFS filesystem with a record size
of
> 128k, it will break up I/O into 128k chunks for writing, but it will also
write
> smaller chunks.  I forget what the minimum size is (512b or 1k, IIRC), but
what
> ZFS does is use a Variable block size, up to the
> maximum size specified in the "recordsize" property.   So, if
> recordsize=128k and you have a 190k write I/O op, it will write a 128k
chunk,
> and a 64k chunk (64 being the smallest multiple of 2 greater than the
> remaining 62 bits of info).  It WON'T write two 128k chunks.

So ... Suppose there is a raidz2 with 8+2 disks.  You write a 128K chunk,
which gets divided up into 8 parts, and each disk writes a 16K block, right?


It seems to me, limiting the max size of data that a disk can write will
ultimately result in more random scattering of information about in the
drives, and degrade performance.  

We previously calculated (in some other thread) that in order for a drive to
be "efficient" which we defined as 99% useful and 1% wasted time seeking,
then each disk would need to be read/writing 40M blocks consistently.  (Of
course, depending on the specs of the drive, but typical consumer &
enterprise disks were consistently around 40M.)  

So wouldn't it be useful to set the recordsize to something huge?  Then if
you've got a large chunk of data to be written, it's actually *permitted* to
be written as a large chunk instead of forcibly breaking it up?

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to