Peter,
This looks suspiciously similar to an issue you and I discussed a couple
of years ago, titled "Lustre caching and NUMA nodes" December 6,2023.
http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/2023-December/018956.html
The application program, in my case dd, is simply doing a streaming
write, 8000 @ 1MB, reading from /dev/zero. For reasons that escape me,
the cached memory of all OSCs, for all file systems, intermittently gets
dropped, causing pauses is the dd write. The other non-Lustre cached
data does not get dropped. An image that depicts this is at:
https://www.dropbox.com/scl/fi/augs88r7lcdfd6wb7nwrf/pfe27_allOSC_cached.png?rlkey=ynaw60yknwmfjavfy5gxsuk76&e=1&dl=0
I would be curious to see what behavior the OSC caching is exhibiting on
your compute node when you experience these write pauses.
John
On 10/28/2025 3:03 PM, [email protected] wrote:
Send lustre-discuss mailing list submissions to
[email protected]
To subscribe or unsubscribe via the World Wide Web, visit
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
or, via email, send a message with subject or body 'help' to
[email protected]
You can reach the person managing the list at
[email protected]
When replying, please edit your Subject line so it is more specific
than "Re: Contents of lustre-discuss digest..."
Today's Topics:
1. issue: strange pauses between writes, but not everywhere
(Peter Grandi)
----------------------------------------------------------------------
Message: 1
Date: Tue, 28 Oct 2025 20:01:41 +0000
From:[email protected] (Peter Grandi)
To: list Linux fs Lustre<[email protected]>
Subject: [lustre-discuss] issue: strange pauses between writes, but
not everywhere
Message-ID:<[email protected]>
Content-Type: text/plain
So I have 3 Lustre storage clusters which recently have develope a
strange issue:
* Each cluster has 1 MDT with 3 "enterprise" SSDs and 8 OSTs each with 3
"entreprise" SSDs, MDT and OSTs all done with ZFS, on top of Alma
8.10. Lustre version is 2.15.7. Each server is pretty overspecified
(28 cores, 128GiB), 100Gb/s cards and switch, and the clients are the
same as the servers except they run the client version of the Lustre
2.16.1 drivers.
* For an example I will use the Lustre 'temp01' where the servers have
addresses 192.168.102.40-48 where .40 is the MDT and some clients with
addresses 192.168.102.13-36.
* Reading is quite good for all clients. But since yesterday early
afternoon inexplicably the clients .13-36 have a maximum average write
speed of around 35-40MB/s; but if I mount 'temp01' on any of the
Lustre servers (and I usually have it mounted on the MDT .40) write
rates are as good as before. Mysteriously today for a while one of the
clients (.14) wrote at previous good speeds for a while and then
reverted to slow. I was tweaking the some '/proc/sys/net/ipv4/tcp_*'
parameters at the time but the same parameters on .13 did not improve
the situation.
* I have collected 'tcpdump' traces on all the 'temp01' servers and a
client while writing and examined with WireShark's "TCP Stream Graphs"
(etc.) and what is happening is that the clients send at full speed
for a little while and then pause for around 2-3 seconds and then
resume. The servers when accessing 'temp01' as clients do not pauses.
* If I use NFS Ganesha with NFSv4-over-TCP on the MDT exporting 'temp01'
I can write to that at high rates (not as high as with native Lustre
of course).
* I have used 'iperf3' to check basic network rates and for "reasons"
they are around 25-30Gb/s, but still much higher than observed
*average* write speeds.
* The issues persists after rebooting the clients (have not reebooted
all the servers of at least one cluster, but I recently rebooted
one of the MDTs).
* I have checked the relevant switch logs and ports and there are no
obvious errors or significant rates of packet issues.
My current guesses are some issue with IP flow control or TCP window
size but bare TCP with 'iperf3' and NFSv4-over-TCP both give good rates.
So perhaps it is something weird with the LNET drivers with receive
pacing in the Lustre driver.
Please let me know if you have seen something similar or other
suggestions.
------------------------------
Subject: Digest Footer
_______________________________________________
lustre-discuss mailing list
[email protected]
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
------------------------------
End of lustre-discuss Digest, Vol 235, Issue 29
***********************************************
_______________________________________________
lustre-discuss mailing list
[email protected]
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org