This has nothing to do with keepalivetimeout – that is to do with keeping a connection open to send subsequent requests without re-negotiating the connection.
It is TimeOut which is the gap between sending packets of the response. If your response is taking more than 1 minute to generate then you are hitting this problem which gives you the 503. How can you get round this: * Look for bottlenecks and speed up response; * Look at how you serve the data? * do you collect it altogether and return it as one big blob or can you stream the data as you generate it; * We have a script that generates many MB of data and can take upwards of an hour to generate the data – we simply stream that output 1 line at a time, memory usage is extremely small and there are no timeout issues; * Can you look at a ticketing solution * The page generates a “ticket” which kicks of the data export job and returns saying data is being produced * You then create a unique URL which will fetch the data/or say comeback later * You then have a ticker in the page which retrieves the data via AJAX or just waits till ready and redirects From: alchemist vk <alchemist...@gmail.com> Sent: 07 October 2020 11:53 To: users@httpd.apache.org Subject: [users@httpd] Configuring KeepAliveTimeout to individual URIs [EXT] Hi All, I have a requirement where serving GET on few URIs whose payload is large takes more than 1min compared to our configured "KeepAliveTimeout 60" directive. And this is resulting in 503 error to clients. Is there a way where I can group few URIs and increase KeepAliveTimeout to 300 secs ? Tried for locationmatch but cant configure KeepAliveTimeout directive inside the locationmatch. Any help is appreciated . WR A -- The Wellcome Sanger Institute is operated by Genome Research Limited, a charity registered in England with number 1021457 and a company registered in England with number 2742969, whose registered office is 215 Euston Road, London, NW1 2BE.