Re: I think I found a fix for the memory leak issue on gRPC module

2024-05-27 Thread Sergey Kandaurov
On Mon, May 27, 2024 at 02:06:58PM +0400, Roman Arutyunyan wrote:
> Hi,

> # HG changeset patch
> # User Roman Arutyunyan 
> # Date 1716477338 -14400
> #  Thu May 23 19:15:38 2024 +0400
> # Node ID f7d53c7f70140b1cd1eaf51ce4346a873692f879
> # Parent  f58b6f6362387eeace46043a6fc0bceb56a6786a
> Optimized chain link usage (ticket #2614).
> 
> Previously chain links could sometimes be dropped instead of being reused,
> which could result in increased memory consumption during long requests.
> 
> A similar chain link issue in ngx_http_gzip_filter_module was fixed in
> da46bfc484ef (1.11.10).
> 
> Based on a patch by Sangmin Lee.
> 
> diff --git a/src/core/ngx_output_chain.c b/src/core/ngx_output_chain.c
> --- a/src/core/ngx_output_chain.c
> +++ b/src/core/ngx_output_chain.c
> @@ -117,7 +117,10 @@ ngx_output_chain(ngx_output_chain_ctx_t 
>  
>  ngx_debug_point();
>  
> -ctx->in = ctx->in->next;
> +cl = ctx->in;
> +ctx->in = cl->next;
> +
> +ngx_free_chain(ctx->pool, cl);
>  
>  continue;
>  }
> @@ -203,7 +206,10 @@ ngx_output_chain(ngx_output_chain_ctx_t 
>  /* delete the completed buf from the ctx->in chain */
>  
>  if (ngx_buf_size(ctx->in->buf) == 0) {
> -ctx->in = ctx->in->next;
> +cl = ctx->in;
> +ctx->in = cl->next;
> +
> +ngx_free_chain(ctx->pool, cl);
>  }
>  
>  cl = ngx_alloc_chain_link(ctx->pool);
> diff --git a/src/http/modules/ngx_http_grpc_module.c 
> b/src/http/modules/ngx_http_grpc_module.c
> --- a/src/http/modules/ngx_http_grpc_module.c
> +++ b/src/http/modules/ngx_http_grpc_module.c
> @@ -1231,7 +1231,7 @@ ngx_http_grpc_body_output_filter(void *d
>  ngx_buf_t  *b;
>  ngx_int_t   rc;
>  ngx_uint_t  next, last;
> -ngx_chain_t*cl, *out, **ll;
> +ngx_chain_t*cl, *out, *ln, **ll;
>  ngx_http_upstream_t*u;
>  ngx_http_grpc_ctx_t*ctx;
>  ngx_http_grpc_frame_t  *f;
> @@ -1459,7 +1459,10 @@ ngx_http_grpc_body_output_filter(void *d
>  last = 1;
>  }
>  
> +ln = in;
>  in = in->next;
> +
> +ngx_free_chain(r->pool, ln);
>  }

Looks good now.

>  
>  ctx->in = in;
> diff --git a/src/http/modules/ngx_http_gunzip_filter_module.c 
> b/src/http/modules/ngx_http_gunzip_filter_module.c
> --- a/src/http/modules/ngx_http_gunzip_filter_module.c
> +++ b/src/http/modules/ngx_http_gunzip_filter_module.c
> @@ -333,6 +333,8 @@ static ngx_int_t
>  ngx_http_gunzip_filter_add_data(ngx_http_request_t *r,
>  ngx_http_gunzip_ctx_t *ctx)
>  {
> +ngx_chain_t  *cl;
> +
>  if (ctx->zstream.avail_in || ctx->flush != Z_NO_FLUSH || ctx->redo) {
>  return NGX_OK;
>  }
> @@ -344,8 +346,11 @@ ngx_http_gunzip_filter_add_data(ngx_http
>  return NGX_DECLINED;
>  }
>  
> -ctx->in_buf = ctx->in->buf;
> -ctx->in = ctx->in->next;
> +cl = ctx->in;
> +ctx->in_buf = cl->buf;
> +ctx->in = cl->next;
> +
> +ngx_free_chain(r->pool, cl);
>  
>  ctx->zstream.next_in = ctx->in_buf->pos;
>  ctx->zstream.avail_in = ctx->in_buf->last - ctx->in_buf->pos;
> @@ -374,6 +379,7 @@ static ngx_int_t
>  ngx_http_gunzip_filter_get_buf(ngx_http_request_t *r,
>  ngx_http_gunzip_ctx_t *ctx)
>  {
> +ngx_chain_t *cl;
>  ngx_http_gunzip_conf_t  *conf;
>  
>  if (ctx->zstream.avail_out) {
> @@ -383,8 +389,12 @@ ngx_http_gunzip_filter_get_buf(ngx_http_
>  conf = ngx_http_get_module_loc_conf(r, ngx_http_gunzip_filter_module);
>  
>  if (ctx->free) {
> -ctx->out_buf = ctx->free->buf;
> -ctx->free = ctx->free->next;
> +
> +cl = ctx->free;
> +ctx->out_buf = cl->buf;
> +ctx->free = cl->next;
> +
> +ngx_free_chain(r->pool, cl);
>  
>  ctx->out_buf->flush = 0;
>  
> diff --git a/src/http/modules/ngx_http_gzip_filter_module.c 
> b/src/http/modules/ngx_http_gzip_filter_module.c
> --- a/src/http/modules/ngx_http_gzip_filter_module.c
> +++ b/src/http/modules/ngx_http_gzip_filter_module.c
> @@ -985,10 +985,14 @@ static void
>  ngx_http_gzip_filter_free_copy_buf(ngx_http_request_t *r,
>  ngx_http_gzip_ctx_t *ctx)
>  {
> -ngx_chain_t  *cl;
> +ngx_chain_t  *cl, *ln;
>  
> -for (cl = ctx->copied; cl; cl = cl->next) {
> -ngx_pfree(r->pool, cl->buf->start);
> +for (cl = ctx->copied; cl; /* void */) {
> +ln = cl;
> +cl = cl->next;
> +
> +ngx_pfree(r->pool, ln->buf->start);
> +ngx_free_chain(r->pool, ln);
>  }
>  
>  ctx->copied = NULL;
> diff --git a/src/http/modules/ngx_http_ssi_filter_module.c 
> b/src/http/modules/ngx_http_ssi_filter_module.c
> --- a/src/http/modules/ngx_http_ssi_filter_module.c
> +++ b/src/http/modules/ngx_http_ssi_filter_module.c
> @@ -482,9 +482,13 @@ 

Re: I think I found a fix for the memory leak issue on gRPC module

2024-05-27 Thread Sergey Kandaurov
On Wed, May 22, 2024 at 06:14:26PM +0400, Roman Arutyunyan wrote:
> Hi,
> 
> Indeed there's a problem there.  We have similar problems in other places as
> well.  Attached is a patch that fixes all I could find.
> 
> I did some testing for the sub_filter with the following config.  Small 
> buffers
> exaggerate the problem.
> 
> http {
> server {
> listen 8000;
> 
> location / {
> root html;
> 
> output_buffers 2 128;
> 
> sub_filter 1 2;
> sub_filter_types *;
> sub_filter_once off;
> }
> }
> }
> 
> Retrieving a 10m file resulted in 7304 (patched) vs 1317704 (unpatched) bytes
> allocated from the request pool.  With the default output_buffers (2 32768),
> it's 79880 vs 84040.

Note that unlike in sub_filter and gzip modules,
gRPC logic is more flawed as the buffer size in user controlled.

> 
> 
> On Thu, May 02, 2024 at 07:59:44AM +, Pavel Pautov via nginx-devel wrote:
> > Hi Sangmin,
> > 
> > Your analysis looks correct. Client streaming RPCs can lead to unbound 
> > accumulation of ngx_chain_t objects (although, at very slow rate). Gzip 
> > module had a similar issue (https://trac.nginx.org/nginx/ticket/1046).
> > Attaching somewhat simplified patch based on yours. I was able to reproduce 
> > the issue locally and the patch fixes it.
> > 
> > From: nginx-devel  On Behalf Of Sangmin Lee
> > Sent: Thursday, April 4, 2024 19:29
> > To: nginx-devel@nginx.org
> > Subject: I think I found a fix for the memory leak issue on gRPC module
> > 
> > CAUTION: This email has been sent from an external source. Do not click 
> > links, open attachments, or provide sensitive business information unless 
> > you can verify the sender's legitimacy.
> > 
> > I am sending this mail again because I did a mistake while I was writing a 
> > mail. I didn't know, in gmail, "Ctrl - Enter" would send a mail immediately 
> > even without a title!
> > I am sorry for that, so I am sending again.
> > 
> > Hello,
> > 
> > I think I found the main cause of the memory leak issue when using gRPC 
> > stream so I made a patch for it.
> > Please find the test scenario and details here -- This is what I wrote.:
> > https://trac.nginx.org/nginx/ticket/2614
> > After I changed the memory pool totally on nginx and tested our workload -- 
> > long-lived gRPC streams with many connections -- using Valgrind and massif, 
> > I was able to find what brought up the memory leak issue.
> > like the picture below.
> > 
> > [cid:image001.png@01DA9C29.C792CD90]
> > ( I am not sure whether this picture will be sent properly )
> > 
> > After I patched one part, it seems okay now I have tested it for 1 week 
> > with out workload.
> > 
> > [cid:image002.png@01DA9C29.C792CD90]
> > ( I am not sure whether this picture will be sent properly )
> > 
> > But because I am not familiar with Mercurial, I couldn't find a way to 
> > create PR like on github. I guess this mailing list is for this patch.
> > From my point of view, it is more like a workaround and I think the way of 
> > using ngx_chain_add_copy() or itself needs to be changed because it 
> > allocates a ngx_chain_t structure using ngx_alloc_chain_link() but inside 
> > of that, it just copies pointer, like cl->buf = in->buf;
> > so this ngx_chain_t instance should be dealt with differently unlike other 
> > ngx_chain_t instances.
> > But I am quite new to nginx codes so my view might be wrong.
> > Anyhow, please go over this patch and I would like to further talk here.
> > 
> > 
> > 
> > diff --git a/src/http/modules/ngx_http_grpc_module.c 
> > b/src/http/modules/ngx_http_grpc_module.c
> > index dfe49c586..1db67bd0a 100644
> > --- a/src/http/modules/ngx_http_grpc_module.c
> > +++ b/src/http/modules/ngx_http_grpc_module.c
> > @@ -1462,6 +1462,12 @@ ngx_http_grpc_body_output_filter(void *data, 
> > ngx_chain_t *in)
> >  in = in->next;
> >  }
> > 
> > +   ngx_chain_t *nl;
> > +   for (ngx_chain_t *dl = ctx->in; dl != in; dl = nl ) {
> > +   nl = dl->next;
> > +   ngx_free_chain(r->pool, dl);
> > +   }
> > +
> >  ctx->in = in;
> > 
> >  if (last) {
> > 
> > ---

Re: I think I found a fix for the memory leak issue on gRPC module

2024-05-27 Thread Roman Arutyunyan
Hi,

Following an internal discussion with Sergey, here's an updated version of
the patch.

On Thu, May 23, 2024 at 01:42:24PM +0400, Roman Arutyunyan wrote:
> Hi,
> 
> On Wed, May 22, 2024 at 06:14:26PM +0400, Roman Arutyunyan wrote:
> > Hi,
> > 
> > Indeed there's a problem there.  We have similar problems in other places as
> > well.  Attached is a patch that fixes all I could find.
> > 
> > I did some testing for the sub_filter with the following config.  Small 
> > buffers
> > exaggerate the problem.
> > 
> > http {
> > server {
> > listen 8000;
> > 
> > location / {
> > root html;
> > 
> > output_buffers 2 128;
> > 
> > sub_filter 1 2;
> > sub_filter_types *;
> > sub_filter_once off;
> > }
> > }
> > }
> > 
> > Retrieving a 10m file resulted in 7304 (patched) vs 1317704 (unpatched) 
> > bytes
> > allocated from the request pool.  With the default output_buffers (2 32768),
> > it's 79880 vs 84040.
> 
> I tested ssi with the following config.
> 
> server {
> listen 8000;
> 
> location / {
> root html;
> 
> output_buffers 2 128;
> 
> ssi on;
> ssi_types *;
> }
> }
> 
> The result is similar:
> 
> 6224 vs 1316912 with small buffers
> 38864 vs 43952 with the default buffers
> 
> > On Thu, May 02, 2024 at 07:59:44AM +, Pavel Pautov via nginx-devel 
> > wrote:
> > > Hi Sangmin,
> > > 
> > > Your analysis looks correct. Client streaming RPCs can lead to unbound 
> > > accumulation of ngx_chain_t objects (although, at very slow rate). Gzip 
> > > module had a similar issue (https://trac.nginx.org/nginx/ticket/1046).
> > > Attaching somewhat simplified patch based on yours. I was able to 
> > > reproduce the issue locally and the patch fixes it.
> > > 
> > > From: nginx-devel  On Behalf Of Sangmin Lee
> > > Sent: Thursday, April 4, 2024 19:29
> > > To: nginx-devel@nginx.org
> > > Subject: I think I found a fix for the memory leak issue on gRPC module
> > > 
> > > CAUTION: This email has been sent from an external source. Do not click 
> > > links, open attachments, or provide sensitive business information unless 
> > > you can verify the sender's legitimacy.
> > > 
> > > I am sending this mail again because I did a mistake while I was writing 
> > > a mail. I didn't know, in gmail, "Ctrl - Enter" would send a mail 
> > > immediately even without a title!
> > > I am sorry for that, so I am sending again.
> > > 
> > > Hello,
> > > 
> > > I think I found the main cause of the memory leak issue when using gRPC 
> > > stream so I made a patch for it.
> > > Please find the test scenario and details here -- This is what I wrote.:
> > > https://trac.nginx.org/nginx/ticket/2614
> > > After I changed the memory pool totally on nginx and tested our workload 
> > > -- long-lived gRPC streams with many connections -- using Valgrind and 
> > > massif, I was able to find what brought up the memory leak issue.
> > > like the picture below.
> > > 
> > > [cid:image001.png@01DA9C29.C792CD90]
> > > ( I am not sure whether this picture will be sent properly )
> > > 
> > > After I patched one part, it seems okay now I have tested it for 1 week 
> > > with out workload.
> > > 
> > > [cid:image002.png@01DA9C29.C792CD90]
> > > ( I am not sure whether this picture will be sent properly )
> > > 
> > > But because I am not familiar with Mercurial, I couldn't find a way to 
> > > create PR like on github. I guess this mailing list is for this patch.
> > > From my point of view, it is more like a workaround and I think the way 
> > > of using ngx_chain_add_copy() or itself needs to be changed because it 
> > > allocates a ngx_chain_t structure using ngx_alloc_chain_link() but inside 
> > > of that, it just copies pointer, like cl->buf = in->buf;
> > > so this ngx_chain_t instance should be dealt with differently unlike 
> > > other ngx_chain_t instances.
> > > But I am quite new to nginx codes so my view might be wrong.
> > > Anyhow, please go over this patch and I would like to further talk here.
> > > 
> > > --

Re: I think I found a fix for the memory leak issue on gRPC module

2024-05-23 Thread Roman Arutyunyan
Hi,

On Wed, May 22, 2024 at 06:14:26PM +0400, Roman Arutyunyan wrote:
> Hi,
> 
> Indeed there's a problem there.  We have similar problems in other places as
> well.  Attached is a patch that fixes all I could find.
> 
> I did some testing for the sub_filter with the following config.  Small 
> buffers
> exaggerate the problem.
> 
> http {
> server {
> listen 8000;
> 
> location / {
> root html;
> 
> output_buffers 2 128;
> 
> sub_filter 1 2;
> sub_filter_types *;
> sub_filter_once off;
> }
> }
> }
> 
> Retrieving a 10m file resulted in 7304 (patched) vs 1317704 (unpatched) bytes
> allocated from the request pool.  With the default output_buffers (2 32768),
> it's 79880 vs 84040.

I tested ssi with the following config.

server {
listen 8000;

location / {
root html;

output_buffers 2 128;

ssi on;
ssi_types *;
}
}

The result is similar:

6224 vs 1316912 with small buffers
38864 vs 43952 with the default buffers

> On Thu, May 02, 2024 at 07:59:44AM +, Pavel Pautov via nginx-devel wrote:
> > Hi Sangmin,
> > 
> > Your analysis looks correct. Client streaming RPCs can lead to unbound 
> > accumulation of ngx_chain_t objects (although, at very slow rate). Gzip 
> > module had a similar issue (https://trac.nginx.org/nginx/ticket/1046).
> > Attaching somewhat simplified patch based on yours. I was able to reproduce 
> > the issue locally and the patch fixes it.
> > 
> > From: nginx-devel  On Behalf Of Sangmin Lee
> > Sent: Thursday, April 4, 2024 19:29
> > To: nginx-devel@nginx.org
> > Subject: I think I found a fix for the memory leak issue on gRPC module
> > 
> > CAUTION: This email has been sent from an external source. Do not click 
> > links, open attachments, or provide sensitive business information unless 
> > you can verify the sender's legitimacy.
> > 
> > I am sending this mail again because I did a mistake while I was writing a 
> > mail. I didn't know, in gmail, "Ctrl - Enter" would send a mail immediately 
> > even without a title!
> > I am sorry for that, so I am sending again.
> > 
> > Hello,
> > 
> > I think I found the main cause of the memory leak issue when using gRPC 
> > stream so I made a patch for it.
> > Please find the test scenario and details here -- This is what I wrote.:
> > https://trac.nginx.org/nginx/ticket/2614
> > After I changed the memory pool totally on nginx and tested our workload -- 
> > long-lived gRPC streams with many connections -- using Valgrind and massif, 
> > I was able to find what brought up the memory leak issue.
> > like the picture below.
> > 
> > [cid:image001.png@01DA9C29.C792CD90]
> > ( I am not sure whether this picture will be sent properly )
> > 
> > After I patched one part, it seems okay now I have tested it for 1 week 
> > with out workload.
> > 
> > [cid:image002.png@01DA9C29.C792CD90]
> > ( I am not sure whether this picture will be sent properly )
> > 
> > But because I am not familiar with Mercurial, I couldn't find a way to 
> > create PR like on github. I guess this mailing list is for this patch.
> > From my point of view, it is more like a workaround and I think the way of 
> > using ngx_chain_add_copy() or itself needs to be changed because it 
> > allocates a ngx_chain_t structure using ngx_alloc_chain_link() but inside 
> > of that, it just copies pointer, like cl->buf = in->buf;
> > so this ngx_chain_t instance should be dealt with differently unlike other 
> > ngx_chain_t instances.
> > But I am quite new to nginx codes so my view might be wrong.
> > Anyhow, please go over this patch and I would like to further talk here.
> > 
> > 
> > 
> > diff --git a/src/http/modules/ngx_http_grpc_module.c 
> > b/src/http/modules/ngx_http_grpc_module.c
> > index dfe49c586..1db67bd0a 100644
> > --- a/src/http/modules/ngx_http_grpc_module.c
> > +++ b/src/http/modules/ngx_http_grpc_module.c
> > @@ -1462,6 +1462,12 @@ ngx_http_grpc_body_output_filter(void *data, 
> > ngx_chain_t *in)
> >  in = in->next;
> >  }
> > 
> > +   ngx_chain_t *nl;
> > +   for (ngx_chain_t *dl = ctx->in; dl != in; dl = nl ) {
> > +   nl = dl->next;
> > +   ngx_free_chain(r->pool, dl);
> > +   }
> > +
> >  ctx-&g

Re: I think I found a fix for the memory leak issue on gRPC module

2024-05-22 Thread Roman Arutyunyan
Hi,

Indeed there's a problem there.  We have similar problems in other places as
well.  Attached is a patch that fixes all I could find.

I did some testing for the sub_filter with the following config.  Small buffers
exaggerate the problem.

http {
server {
listen 8000;

location / {
root html;

output_buffers 2 128;

sub_filter 1 2;
sub_filter_types *;
sub_filter_once off;
}
}
}

Retrieving a 10m file resulted in 7304 (patched) vs 1317704 (unpatched) bytes
allocated from the request pool.  With the default output_buffers (2 32768),
it's 79880 vs 84040.


On Thu, May 02, 2024 at 07:59:44AM +, Pavel Pautov via nginx-devel wrote:
> Hi Sangmin,
> 
> Your analysis looks correct. Client streaming RPCs can lead to unbound 
> accumulation of ngx_chain_t objects (although, at very slow rate). Gzip 
> module had a similar issue (https://trac.nginx.org/nginx/ticket/1046).
> Attaching somewhat simplified patch based on yours. I was able to reproduce 
> the issue locally and the patch fixes it.
> 
> From: nginx-devel  On Behalf Of Sangmin Lee
> Sent: Thursday, April 4, 2024 19:29
> To: nginx-devel@nginx.org
> Subject: I think I found a fix for the memory leak issue on gRPC module
> 
> CAUTION: This email has been sent from an external source. Do not click 
> links, open attachments, or provide sensitive business information unless you 
> can verify the sender's legitimacy.
> 
> I am sending this mail again because I did a mistake while I was writing a 
> mail. I didn't know, in gmail, "Ctrl - Enter" would send a mail immediately 
> even without a title!
> I am sorry for that, so I am sending again.
> 
> Hello,
> 
> I think I found the main cause of the memory leak issue when using gRPC 
> stream so I made a patch for it.
> Please find the test scenario and details here -- This is what I wrote.:
> https://trac.nginx.org/nginx/ticket/2614
> After I changed the memory pool totally on nginx and tested our workload -- 
> long-lived gRPC streams with many connections -- using Valgrind and massif, I 
> was able to find what brought up the memory leak issue.
> like the picture below.
> 
> [cid:image001.png@01DA9C29.C792CD90]
> ( I am not sure whether this picture will be sent properly )
> 
> After I patched one part, it seems okay now I have tested it for 1 week with 
> out workload.
> 
> [cid:image002.png@01DA9C29.C792CD90]
> ( I am not sure whether this picture will be sent properly )
> 
> But because I am not familiar with Mercurial, I couldn't find a way to create 
> PR like on github. I guess this mailing list is for this patch.
> From my point of view, it is more like a workaround and I think the way of 
> using ngx_chain_add_copy() or itself needs to be changed because it allocates 
> a ngx_chain_t structure using ngx_alloc_chain_link() but inside of that, it 
> just copies pointer, like cl->buf = in->buf;
> so this ngx_chain_t instance should be dealt with differently unlike other 
> ngx_chain_t instances.
> But I am quite new to nginx codes so my view might be wrong.
> Anyhow, please go over this patch and I would like to further talk here.
> 
> 
> 
> diff --git a/src/http/modules/ngx_http_grpc_module.c 
> b/src/http/modules/ngx_http_grpc_module.c
> index dfe49c586..1db67bd0a 100644
> --- a/src/http/modules/ngx_http_grpc_module.c
> +++ b/src/http/modules/ngx_http_grpc_module.c
> @@ -1462,6 +1462,12 @@ ngx_http_grpc_body_output_filter(void *data, 
> ngx_chain_t *in)
>  in = in->next;
>  }
> 
> +   ngx_chain_t *nl;
> +   for (ngx_chain_t *dl = ctx->in; dl != in; dl = nl ) {
> +   nl = dl->next;
> +   ngx_free_chain(r->pool, dl);
> +   }
> +
>  ctx->in = in;
> 
>  if (last) {
> 
> 
> 
> Best regards,
> Sangmin





> ___
> nginx-devel mailing list
> nginx-devel@nginx.org
> https://mailman.nginx.org/mailman/listinfo/nginx-devel


--
Roman Arutyunyan
# HG changeset patch
# User Roman Arutyunyan 
# Date 1716386385 -14400
#  Wed May 22 17:59:45 2024 +0400
# Node ID 95af5fe4921294b48e634defc03b6b0f0201d6f7
# Parent  e60377bdee3d0f8dbd237b5ae8a5deab2af785ef
Optimized chain link usage (ticket #2614).

Previously chain links could sometimes be dropped instead of being reused,
which could result in increased memory consumption during long requests.

A similar chain link issue