FYI Scale 4.2.3 went out of support in September 2020.  DDN may still support it but there are no updates/enhancements being made to that code stream.  It is quite old.  Scale 5.0.x goes end of support at the end of April 2022.  Scale 5.1.2 was just released and if possible I suggest you upgrade your entire cluster to the Scale 5.1.x release stream.
 
The change to the prefetchAggressivenessRead variable should help, but performance may still not be as you require.

Fred
_______________________________________________________
Fred Stock | Spectrum Scale Development Advocacy | 720-430-8821
sto...@us.ibm.com
 
 
----- Original message -----
From: "Stuart Barkley" <stua...@4gh.net>
Sent by: gpfsug-discuss-boun...@spectrumscale.org
To: gpfsug-discuss@spectrumscale.org
Cc:
Subject: [EXTERNAL] [gpfsug-discuss] alphafold and mmap performance
Date: Tue, Oct 19, 2021 1:25 PM
 
Over the years there have been several discussions about performance
problems with mmap() on GPFS/Spectrum Scale.

We are currently having problems with mmap() performance on our
systems with new alphafold <https://github.com/deepmind/alphafold >
protein folding software.  Things look similar to previous times we
have had mmap() problems.

The software component "hhblits" appears to mmap a large file with
genomic data and then does random reads throughout the file.  GPFS
appears to be doing 4K reads for each block limiting the performance.

The first run takes 20+ hours to run.  Subsequent identical runs
complete in just 1-2 hours.  After clearing the linux system cache
(echo 3 > /proc/sys/vm/drop_caches) the slow performance returns for
the next run.

GPFS Server is 4.2.3-5 running on DDN hardware.  CentOS 7.3
Default GPFS Client is 4.2.3-22. CentOS 7.9

We have tried a number of things including Spectrum Scale client
version 5.0.5-9 which should have Sven's recent mmap performance
improvements. Are the recent mmap performance improvements in the
client code or the server code?

Only now do I notice a suggestion:
    mmchconfig prefetchAggressivenessRead=0 -i
I did not use this.  Would a performance change be expected?

Would the pagepool size be involved in this?

Stuart Barkley
--
I've never been lost; I was once bewildered for three days, but never lost!
                                        --  Daniel Boone
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss 
 


_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to