On 9/8/2012 1:10 PM, Martin Steigerwald wrote: > Am Freitag, 7. September 2012 schrieb Stan Hoeppner: >> On 9/7/2012 12:42 PM, Dan Ritter wrote: > […] >>> Now, the next thing: I know it's tempting to make a single >>> filesystem over all these disks. Don't. The fsck times will be >>> horrendous. Make filesystems which are the size you need, plus a >>> little extra. It's rare to actually need a single gigantic fs. >> >> Whjat? Are you talking crash recovery boot time "fsck"? With any >> modern journaled FS log recovery is instantaneous. If you're talking >> about an actual structure check, XFS is pretty quick regardless of >> inode count as the check is done in parallel. I can't speak to EXTx >> as I don't use them. For a multi terabyte backup server, XFS is the >> only way to go anyway. Using XFS also allows infinite growth without >> requiring array reshapes nor LVM, while maintaining striped write >> alignment and thus maintaining performance. >> >> There are hundreds of 30TB+ and dozens of 100TB+ XFS filesystems in >> production today, and I know of one over 300TB and one over 500TB, >> attached to NASA's two archival storage servers. >> >> When using correctly architected reliable hardware there's no reason >> one can't use a single 500TB XFS filesystem. > > I assume that such correctly architected hardware contains a lot of RAM in > order to be able to xfs_repair the filesystem in case of any filesystem > corruption. > > I know RAM usage of xfs_repair has been lowered, but still such a 500 TiB > XFS filesystem can contain a lot of inodes.
The system I've been referring to with the ~500TB XFS is an IA64 SGI Altix with 64P and 128GB RAM. I'm pretty sure 128GB is plenty for xfs_repair on filesystems much larger than 500TB. > But for upto 10 TiB XFS filesystem I wouldn´t care too much about those > issues. Yeah, an 8GB machine typically allows for much larger than 10TB xfs_repair. -- Stan -- To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org Archive: http://lists.debian.org/504c587b.90...@hardwarefreak.com