On Mon, 2010-05-17 at 12:54 -0400, Dan Pritts wrote:
> On Mon, May 17, 2010 at 06:25:18PM +0200, Tomas Ögren wrote:
> > Resilver does a whole lot of random io itself, not bulk reads.. It reads
> > the filesystem tree, not "block 0, block 1, block 2..". You won't get
> > 60MB/s sustained, not even close.
> 
> Even with large, unfragmented files?  
> 
> danno
> --
> Dan Pritts, Sr. Systems Engineer
> Internet2
> office: +1-734-352-4953 | mobile: +1-734-834-7224

Having large, unfragmented files will certainly help keep sustained
throughput.  But, also, you have to consider the amount of deletions
done on the pool.

For instance, let's say you wrote files A, B, and C one right after
another, and they're all big files.  Doing a re-silver, you'd be pretty
well off on getting reasonable throughput reading A, then B, then C,
since they're going to be contiguous on the drive (both internally, and
across the three files).  However, if you have deleted B at some time,
and say wrote a file D (where D < B in size) into B's old space, then,
well, you seek to A, read A, seek forward to C, read C, seek back to D,
etc.

Thus, you'll get good throughput for resilver on these drives pretty
much in just ONE case:  large files with NO deletions.  If you're using
them for write-once/read-many/no-delete archives, then you're OK.
Anything else is going to suck.

:-)



-- 
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to