The filesystem looks like a test system (only 200GB OSTs, the whole filesystem 
is smaller than a single HDD), so it doesn't make sense to be testing Lustre 
2.9 at this point.  At a very minimum it should be 2.12.8, but for a new test 
filesystem it makes sense to try out master (2.15.0-RC2) since it will be the 
new LTS release very soon.

Cheers, Andreas.

On Mar 29, 2022, at 11:08, Patrick Farrell via lustre-discuss 
<lustre-discuss@lists.lustre.org<mailto:lustre-discuss@lists.lustre.org>> wrote:

Hasan,

That's quite old - 2.9 was released nearly 6 years ago and wasn't a maintenance 
release, so had no updates after release.  I expect some of those bugs are 
present there.  A newer version should help.

-Patrick
________________________________
From: Hasan Rashid <mrash...@uncc.edu<mailto:mrash...@uncc.edu>>
Sent: Tuesday, March 29, 2022 11:50 AM
To: Patrick Farrell <pfarr...@ddn.com<mailto:pfarr...@ddn.com>>
Cc: lustre-discuss@lists.lustre.org<mailto:lustre-discuss@lists.lustre.org> 
<lustre-discuss@lists.lustre.org<mailto:lustre-discuss@lists.lustre.org>>
Subject: Re: [EXTERNAL] Re: [lustre-discuss] Write Performance is Abnormal for 
max_dirty_mb Value of 2047

Hi Patrick,

Thank you so much for getting back to me. The Lustre installation was 2.9.0. I 
am working to run the workloads on Lustre 2.12 release.

I have 1 machine for MGS and MDS and 4 machines for OSS. In each OSS, I have 2 
OSTs and each of them is 200 GiB in size. All the OSTs are mounted on SSD. The 
striping layout is the default one. Since for the workload, we have one large 
file of 5 GiB in size, the whole resides in a single OST. We also verified that 
by looking at the rpc_stats.

<image.png>

Following is the hardware configuration for all the machines.

<image.png>

Please let me know if you need further information.

Thanks,
Hasan

On Sun, Mar 27, 2022 at 1:54 PM Patrick Farrell 
<pfarr...@ddn.com<mailto:pfarr...@ddn.com>> wrote:
[Caution: Email from External Sender. Do not click or open links or attachments 
unless you know this sender.]

Hasan,

Historically, there have been several bugs related to write grant when 
max_dirty_mb is set to large values (depending on a few other details of system 
setup).

Write grant allows the client to write data in to memory and write it out 
asynchronously.  When write grant is not available to the client, the client is 
forced to do sync writes at small sizes.  The result looks exactly like this, 
write performance drops severely.

Depending on what version you're running, you may not have fixes for these 
bugs.  You could either try a newer Lustre version (you didn't mention what 
you're running) or just use a smaller value of max_dirty_mb.

I am surprised to see you're still seeing a speedup from max_dirty_mb values 
over 1 GiB in size.

Can you describe your system a bit more?  How many OSTs do you have and how 
many stripes are you using?  max_dirty_mb is a per OST value on the client, not 
a global one.

-Patrick
________________________________
From: lustre-discuss 
<lustre-discuss-boun...@lists.lustre.org<mailto:lustre-discuss-boun...@lists.lustre.org>>
 on behalf of Hasan Rashid via lustre-discuss 
<lustre-discuss@lists.lustre.org<mailto:lustre-discuss@lists.lustre.org>>
Sent: Friday, March 25, 2022 11:45 AM
To: lustre-discuss@lists.lustre.org<mailto:lustre-discuss@lists.lustre.org> 
<lustre-discuss@lists.lustre.org<mailto:lustre-discuss@lists.lustre.org>>
Subject: [lustre-discuss] Write Performance is Abnormal for max_dirty_mb Value 
of 2047

Hi Everyone,

As the manual suggests, the valid value range for max_dirty_mb is the values 
larger than 0 and smaller than the lesser of 2048 MiB or 1/4 of client RAM. In 
my system, the client's RAM is 196 GiB. So, the maximum valid value for 
max_dirty_mb(mdm) is 2047 MiB.

However, when we set the max_dirty_mb value to 2047, we see very low write 
throughput for multiple Filebench workloads that we have tested so far. I am 
providing details for one example of the tested workload below.

Workload Detail: We are doing only random write operation of 1MiB size from one 
process and one thread to a single large file of 5GiB size.

Observed Result: As you can see from the below diagram, as we increase the mdm 
value from 768 to 1792 by an amount of 256 in each step, the write throughput 
has increased gradually. However, for the mdm value of 2047, the result dropped 
very significantly. The observation holds true for all the workloads we tested 
so far.


[https://lh3.googleusercontent.com/iEqpGNZhI9r9jJCLq0rWPvFADJRXkKKKZnyCV_8m3nhiHggNqWU9d_7WTUU0yeb011nxjULF4_iLkI7TIc0qe5el11PJI3i9Jot9KveXUil98A_UEnBojFqAHfK94ve1foQT39m2]

I am unable to figure out why we would have such low performance at the mdm 
value of 2047. Please share any insights you have that would be helpful for me 
to understand the aforementioned scenario.

Best Wishes,
Md Hasanur Rashid
_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org<mailto:lustre-discuss@lists.lustre.org>
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org

Cheers, Andreas
--
Andreas Dilger
Lustre Principal Architect
Whamcloud







_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org

Reply via email to