Thanx for the details on the Inode number, but I'm still having an issue where I'm not getting the number I expected from the MDS creation, but I suspect its not a reporting error from lfs.
When I create the MDS, I specified '-i 1024' and I can see (locally) 800M inodes, but only part of the available space is allocated. Also, when the client mounts the filesystem, the MDS only has 400M blocks available: gulfwork-MDT0000_UUID 430781784 500264 387274084 0% /gulfwork[MDT:0] As we were creating files for testing, I saw that each inode allocation on the MDS was consuming 4k of space, so even though I have 800M inodes available on actual mds partition, it appears that the actual space available was only allowing 100M inodes in the lustre fs. Am I understanding that correctly? I tried to force the MDS creation to use a smaller size per inode but that produced an error: mkfs.lustre --fsname gulfwork --mdt --mgs --mkfsoptions='-i 1024 -I 1024' --reformat --failnode=10.18.12.1 /dev/sda ... mke2fs: inode_size (1024) * inodes_count (860148736) too big for a filesystem with 215037184 blocks, specify higher inode_ratio (-i) or lower inode count (-N). ... yet the actual drive has many more blocks available: SCSI device sda: 1720297472 512-byte hdwr sectors (880792 MB) Is this ext4 setting the block size limit? FYI, I am using: lustre-1.8.2-2.6.18_164.11.1.el5-ext4_lustre.1.8.2.x86_64.rpm lustre-ldiskfs-3.0.9-2.6.18_164.11.1.el5-ext4_lustre.1.8.2.x86_64.rpm e2fsprogs-1.41.6.sun1-0redhat.rhel5.x86_64.rpm -- Gary Molenkamp SHARCNET Systems Administrator University of Western Ontario g...@sharcnet.ca http://www.sharcnet.ca (519) 661-2111 x88429 (519) 661-4000 _______________________________________________ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss