On 2010-11-11, at 19:53, Christopher Walker wrote: > Thanks very much for your reply. I've tried remaking the mdsdb and all > of the ostdb's, but I still get the same error -- it checks the first 34 > osts without a problem, but can't find the ostdb file for the 35th > (which has ost_idx 42): > > with the filesystem up I can see files on this OST: > > [cwal...@iliadaccess04 P-Gadget3.3.1]$ lfs getstripe predict.c > OBDS: > 0: aegalfs-OST0000_UUID ACTIVE > ... > 33: aegalfs-OST0021_UUID ACTIVE > 42: aegalfs-OST002a_UUID ACTIVE > predict.c > obdidx objid objid group > 42 10 0xa 0 > > > lfsck identifies several hundred GB of orphan data that we'd like to > recover, so we'd really like to run lfsck on this array. We're willing > to forgo the recovery on the 35th ost, but I want to make sure that > running lfsck -l with the current configuration won't make things worse.
I'm not sure that what lfsck is reporting in this case is correct. Is the orphan data all on the same OST, or spread around separate OSTs? My concern is that if lfsck thinks the in-use objects on your estranged OST is actually orphan data it will destroy that data. If there are a small number of very large objects on other OSTs that are making up the bulk of the orphan space usage, you could mount those OSTs as type ldiskfs and delete the objects by hand to free up the space. Cheers, Andreas -- Andreas Dilger Lustre Technical Lead Oracle Corporation Canada Inc. _______________________________________________ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss