Our problem seems to correlate with an unintentional creation of a tree of 
>500M files.  Some of the crashes we've had since then appeared
to be related to vm.zone_reclaim_mode=1.  We also enabled quotas right after 
the 500M file thing, and were thinking that inconsistent
quota records might cause this sort of crash.

Have you set vm.zone_reclaim_mode=0 yet?  I had an issue with this on my
file system a while back when it was set to 1.

all our existing Lustre MDSes run happily with vm.zone_reclaim_mode=0,
and making this one consistent appears to have resolved a problem
(in which one family of lustre kernel threads would appear to spin,
"perf top" showing nearly all time spent in spinlock_irq.  iirc.)

might your system have had a *lot* of memory? ours tend to be fairly modest (32-64G, dual-socket intel.)

thanks,
Mark Hahn | SHARCnet Sysadmin | h...@sharcnet.ca | http://www.sharcnet.ca
          | McMaster RHPCS    | h...@mcmaster.ca | 905 525 9140 x24687
          | Compute/Calcul Canada                | http://www.computecanada.ca
_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org

Reply via email to