On 04.09.24 15:59, Henrik Cednert wrote:
Thanks Uwe
Just saw: the app seems to have issued IO requests not sequentially but in bunches.
what is apparent (but I have not recognized it before):after such a bunch of IOs, the next bunch is typically issued when the longest-taking IO from the prious bunch was completed. For example:
13:24:59.341629 R data 17:805453824 16384 3,993 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.341629 R data 6:1720532992 16384 27,471 cli C0A82DD5:63877BCE 192.168.45.214 13:24:59.341629 R data 14:1720532992 16384 44,953 cli C0A82DD5:63877BD6 192.168.45.214
3 read IOs issued at 13:24:59.341629, longest taking 44.953ms. 59.341629+0.044953=59.386582, which is 1ms before the next IO req is issued at 13:24:59.387584
13:24:59.387584 R data 17:805453824 16384 5,990 cli C0A82DD5:63877BDA 192.168.45.213
that one takes just 5.990ms , the next bunch of IO reqs is seen 6.988 ms later (i.e. again the service time plus about 1 ms):
13:24:59.394572 R data 17:805453824 16384 7,993 cli C0A82DD5:63877BDA 192.168.45.213
now a triplet of IOs again, but one of them taking very long:13:24:59.402565 R data 25:805453824 16384 4,991 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.402565 R data 22:1720532992 16384 26,309 cli C0A82DD5:63877BDF 192.168.45.214 13:24:59.402565 R data 7:1720532992 16384 142,856 cli C0A82DD5:63877BCF 192.168.45.213
That incurs the expected delay of 143.854ms :13:24:59.546419 R data 25:805453824 16384 7,992 cli C0A82DD5:63877BE2 192.168.45.213
etc. pp.13:24:59.556408 R data 25:805453824 16384 7,987 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.564395 R data 10:805453824 16384 5,505 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.564395 R data 23:1720532992 16384 28,053 cli C0A82DD5:63877BE0 192.168.45.213 13:24:59.564395 R data 15:1720532992 16384 33,044 cli C0A82DD5:63877BD8 192.168.45.213 13:24:59.598437 R data 10:805453824 16384 5,504 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.604939 R data 18:805453824 16384 4,993 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.604939 R data 8:1720532992 16384 36,015 cli C0A82DD5:63877BD0 192.168.45.214
...so there is some parallelization in IO, but using multiple threads should improve things as the app apparently still waits for IOs being performed. and of course with an otherwise unloaded storage and just one client, service times >25ms should not be seen and most should be 10...20ms.
Uwe -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email:uwe.fa...@kit.edu www.scc.kit.edu Registered office: Kaiserstraße 12, 76131 Karlsruhe, Germany KIT – The Research University in the Helmholtz Association
smime.p7s
Description: S/MIME Cryptographic Signature
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org