Cc'ing supervisor to loop him in as well. On Wed, Dec 6, 2023, 9:18 AM Edward Zuniga <eazun...@vcu.edu> wrote:
> Thanks everyone for the feedback! I've learned so much from reading the > discussions. > > For our application, we will have a LAN with a single server (1TB RAID1 > array for OS, 200TB RAID5 array for data) and up to 16 workstations (1TB > RAID1 array for OS). Our IT department is more familiar with Rocky Linux 8, > which I assume will perform the same as AlmaLinux 8. Some of our MRI > processing can take weeks to finish, so we need a system that is very > reliable. We also work with individual files in the hundreds of gigabytes. > > While reading the Red Hat 8 manual > <https://urldefense.proofpoint.com/v2/url?u=https-3A__access.redhat.com_documentation_en-2Dus_red-5Fhat-5Fenterprise-5Flinux_8_html_managing-5Ffile-5Fsystems_overview-2Dof-2Davailable-2Dfile-2Dsystems-5Fmanaging-2Dfile-2Dsystems&d=DwIFaQ&c=gRgGjJ3BkIsb5y6s49QqsA&r=gd8BzeSQcySVxr0gDWSEbN-P-pgDXkdyCtaMqdCgPPdW1cyL5RIpaIYrCn8C5x2A&m=NWrdCkO_Rv2xr06ZFmX0tmfqqeYNrrwhynckTqel03PtwMXxPfTvgwA0pa8NEDQP&s=PuCPAQ38-YIaby8e4N7dh0ORNT6UbvsXS04mQ0wfKnw&e= > >, > I found a few possible issues regarding XFS. I'm curious to see if anyone > has experienced these as well. > > 1. Metadata error behaviorIn ext4, you can configure the behavior when > the file system encounters metadata errors. The default behavior is to > simply continue the operation. When XFS encounters an unrecoverable > metadata error, it shuts down the file system and returns the EFSCORRUPTED > error.*This could be problematic for processing that takes several > weeks.*2. Inode numbers > > The ext4 file system does not support more than 232 inodes. > > XFS dynamically allocates inodes. An XFS file system cannot run out of > inodes as long as there is free space on the file system. > > Certain applications cannot properly handle inode numbers larger than 232 on > an XFS file system. These applications might cause the failure of 32-bit > stat calls with the EOVERFLOW return value. Inode number exceed 232 under > the following conditions: > > - The file system is larger than 1 TiB with 256-byte inodes. > - The file system is larger than 2 TiB with 512-byte inodes. > > If your application fails with large inode numbers, mount the XFS file > system with the -o inode32 option to enforce inode numbers below 232. > Note that using inode32 does not affect inodes that are already allocated > with 64-bit numbers. > *Has anyone encountered this issue? *3. The Red Hat 8 manual also warns > that using xfs_repair -L might cause significant file system damage and > data loss and should only be used as a last resort. The manual does not > mention a similar warning about using e2fsck to repair an ext4 file system. > Has anyone experienced issues repairing a corrupt XFS file system? > Thanks,Eddie > > On Tue, Dec 5, 2023 at 8:46 PM Konstantin Olchanski <olcha...@triumf.ca> > wrote: > >> On Mon, Dec 04, 2023 at 03:03:46PM -0500, Edward Zuniga wrote: >> > >> > We are upgrading our MRI Lab servers and workstations to AlmaLinux 8. We >> > have used ext4 for the past 10 years, however we are considering using >> XFS >> > for its better performance with larger files. Which file system do you >> use >> > for your lab? >> > >> >> Historical background. >> >> XFS filesystem with the companion XLV logical volume manager (aka >> "partitioning tool") >> came to Linux from SGI IRIX, where it was developed circa late-1990-ies. >> XFS was copied >> to Linux verbatim (initially with shims and kludges, later, fully >> integrated). >> XLV was reimplemented as LVM. >> >> The EXT series of filesystems were developed together with the linux >> kernel (first ext >> filesystem may have originated with MINIX, look it up). As improvements >> were made, >> journaling, no need to fsck after crash, online grow/shrink, etc, they >> were >> renamed ext2/ext3/ext4 and they are still largely compatible between >> themselves. >> >> For many purposes, both filesystems are obsoleted by ZFS, which added: >> >> - added metadata and data checksums - to detect silent bit rot on >> current-generation HDDs and SSDs >> - added online filesystem check - for broken data, gives you list of >> filenames instead of inode numbers >> - added "built-in" mirroring - together with checksums, online fsck (zfs >> scrub) and monthly zfs scrub cron job, allows automatic healing of bit rot. >> - added "built-in" raid-5 and raid-6 - again, together with checksums and >> online fsck, allows automatic healing and robust operation in presence of >> disk bad sectors, I/O errors, corruption and single-disk failure. >> - other goodies like snapshots, large ram cache, dedup, online >> compression, etc are taken for granted for current generation filesystems. >> >> On current generation HDDs and SSds, use of bare XFS and ext4 is >> dangerous, SSD failure or "HDD grows bad sectors" will destroy your data >> completely. >> >> On current generation HDDs, use of mirrored XFS and ext4 is dangerous >> (using mdadm or LVM mirroring), (a) bit rot inevitably causes differences >> between data between the two disks. Lacking checksums, mdadm and LVM >> mirroring cannot decide which of the two copies is the correct one. (b) >> after a crash, mirror rebuild fill fail if both disks happen to have bad >> sectors (or throw random I/O errors). >> >> Ditto for RAID5 and RAID6, probability of RAID rebuild failing because >> multiple disks have have sectors and I/O errors goes up with the number of >> disks. >> >> ZFS was invented to resolve all these problems. (BTRFS was invented as a >> NIH erzatz ZFS, is still incomplete wrt RAID5/RAID6). >> >> Bottom line, if you can, use ZFS. Current Ubuntu installer has a button >> "install on ZFS", use it! >> >> -- >> Konstantin Olchanski >> Data Acquisition Systems: The Bytes Must Flow! >> Email: olchansk-at-triumf-dot-ca >> Snail mail: 4004 Wesbrook Mall, TRIUMF, Vancouver, B.C., V6T 2A3, Canada >> > > > -- > Edward A. Zuniga > Senior Research Analyst/CARI MRI Lab Manager > Virginia Commonwealth University > C. Kenneth and Dianne Wright Center for Clinical and Translational Research > Institute for Drug and Alcohol Studies > Collaborative Advanced Research Imaging (CARI) Facility > 203 East Cary Street, Suite 202 > Richmond, Virginia 23219 > Phone: (804) 828-4184 > Fax: (804) 827-2565 > eazun...@vcu.edu >