On Thu, Feb 08, 2024 at 02:59:15PM +0000, Michal Hruška wrote: > @Uwe > Using iohist we found out that gpfs is overloading one dm-device (it took > about 500ms to finish IOs). We replaced the "problematic" dm-device (as we > have enough drives to play with) for new one but the overloading issue just > jumped to another dm-device. > We believe that this behaviour is caused by the gpfs but we are unable to > locate the root cause of it.
Hello, this behaviour could be caused by an assymmetry in data paths of your storage, relatively small imbalance can make request queue of a slightly slower disk grow seemingly unproportionally. In general, I think you need to scale your GPFS parameters down, not up, in order to force better write clustering and achieve top speed of rotational disks unless array controllers use huge cache memory. If you can change your benchmark workload, try synchronous writes (dd oflag=dsync ...). Best regards, Zdenek Salvet sal...@ics.muni.cz Institute of Computer Science of Masaryk University, Brno, Czech Republic and CESNET, z.s.p.o., Prague, Czech Republic Phone: ++420-549 49 6534 Fax: ++420-541 212 747 ---------------------------------------------------------------------------- Teamwork is essential -- it allows you to blame someone else. _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org