Re: [PATCH] xfs: remove kmem_zalloc_greedy

2017-03-07 Thread Michal Hocko
On Mon 06-03-17 10:41:09, Darrick J. Wong wrote:
> The sole remaining caller of kmem_zalloc_greedy is bulkstat, which uses
> it to grab 1-4 pages for staging of inobt records.  The infinite loop in
> the greedy allocation function is causing hangs[1] in generic/269, so
> just get rid of the greedy allocator in favor of kmem_zalloc_large.
> This makes bulkstat somewhat more likely to ENOMEM if there's really no
> pages to spare, but eliminates a source of hangs.
> 
> [1] https://lkml.org/lkml/2017/2/28/832

I cannot really comment on the patch but I would suggest not using
lkml.org reference in the changelog because I've seen those links being
broken many times. Could you use
http://lkml.kernel.org/r/20170301044634.rgidgdqqiiwsmfpj%40XZHOUW.usersys.redhat.com

instead please? Thanks for taking care of this!

> 
> Signed-off-by: Darrick J. Wong 
> ---
>  fs/xfs/kmem.c   |   18 --
>  fs/xfs/kmem.h   |2 --
>  fs/xfs/xfs_itable.c |   14 --
>  3 files changed, 8 insertions(+), 26 deletions(-)
> 
> diff --git a/fs/xfs/kmem.c b/fs/xfs/kmem.c
> index 339c696..bb2beae 100644
> --- a/fs/xfs/kmem.c
> +++ b/fs/xfs/kmem.c
> @@ -24,24 +24,6 @@
>  #include "kmem.h"
>  #include "xfs_message.h"
>  
> -/*
> - * Greedy allocation.  May fail and may return vmalloced memory.
> - */
> -void *
> -kmem_zalloc_greedy(size_t *size, size_t minsize, size_t maxsize)
> -{
> - void*ptr;
> - size_t  kmsize = maxsize;
> -
> - while (!(ptr = vzalloc(kmsize))) {
> - if ((kmsize >>= 1) <= minsize)
> - kmsize = minsize;
> - }
> - if (ptr)
> - *size = kmsize;
> - return ptr;
> -}
> -
>  void *
>  kmem_alloc(size_t size, xfs_km_flags_t flags)
>  {
> diff --git a/fs/xfs/kmem.h b/fs/xfs/kmem.h
> index 689f746..f0fc84f 100644
> --- a/fs/xfs/kmem.h
> +++ b/fs/xfs/kmem.h
> @@ -69,8 +69,6 @@ static inline void  kmem_free(const void *ptr)
>  }
>  
>  
> -extern void *kmem_zalloc_greedy(size_t *, size_t, size_t);
> -
>  static inline void *
>  kmem_zalloc(size_t size, xfs_km_flags_t flags)
>  {
> diff --git a/fs/xfs/xfs_itable.c b/fs/xfs/xfs_itable.c
> index 8b2150d..283e76c 100644
> --- a/fs/xfs/xfs_itable.c
> +++ b/fs/xfs/xfs_itable.c
> @@ -362,7 +362,6 @@ xfs_bulkstat(
>   xfs_agino_t agino;  /* inode # in allocation group */
>   xfs_agnumber_t  agno;   /* allocation group number */
>   xfs_btree_cur_t *cur;   /* btree cursor for ialloc btree */
> - size_t  irbsize; /* size of irec buffer in bytes */
>   xfs_inobt_rec_incore_t  *irbuf; /* start of irec buffer */
>   int nirbuf; /* size of irbuf */
>   int ubcount; /* size of user's buffer */
> @@ -389,11 +388,14 @@ xfs_bulkstat(
>   *ubcountp = 0;
>   *done = 0;
>  
> - irbuf = kmem_zalloc_greedy(&irbsize, PAGE_SIZE, PAGE_SIZE * 4);
> - if (!irbuf)
> - return -ENOMEM;
> -
> - nirbuf = irbsize / sizeof(*irbuf);
> + nirbuf = (PAGE_SIZE * 4) / sizeof(*irbuf);
> + irbuf = kmem_zalloc_large(PAGE_SIZE * 4, KM_SLEEP);
> + if (!irbuf) {
> + irbuf = kmem_zalloc_large(PAGE_SIZE, KM_SLEEP);
> + if (!irbuf)
> + return -ENOMEM;
> + nirbuf /= 4;
> + }
>  
>   /*
>* Loop over the allocation groups, starting from the last

-- 
Michal Hocko
SUSE Labs


Re: [PATCH] xfs: remove kmem_zalloc_greedy

2017-03-06 Thread Christoph Hellwig
On Tue, Mar 07, 2017 at 11:54:20AM +1100, Dave Chinner wrote:
> > Or maybe I've misunderstood, and you're asking if we should try
> > kmem_zalloc(4 pages), then kmem_zalloc(1 page), and only then switch to
> > the __vmalloc calls?
> 
> Just call kmem_zalloc_large() for 4 pages without a fallback on
> failure - that's exactly how we handle allocations for things like
> the 64k xattr buffers

Yeah, that sounds fine.  I didn't remember that we actually tried
kmalloc before vmalloc for kmem_zalloc_large.


Re: [PATCH] xfs: remove kmem_zalloc_greedy

2017-03-06 Thread Darrick J. Wong
On Tue, Mar 07, 2017 at 01:07:54AM +0100, Christoph Hellwig wrote:
> I like killing it, but shouldn't we just try a normal kmem_zalloc?
> At least for the fallback it's the right thing, and even for an
> order 2 allocation it seems like a useful first try.

I'm confused -- kmem_zalloc_large tries kmem_zalloc with KM_MAYFAIL and
only falls back to __vmalloc if it doesn't get anything.

Or maybe I've misunderstood, and you're asking if we should try
kmem_zalloc(4 pages), then kmem_zalloc(1 page), and only then switch to
the __vmalloc calls?

--D


Re: [PATCH] xfs: remove kmem_zalloc_greedy

2017-03-06 Thread Dave Chinner
On Mon, Mar 06, 2017 at 04:13:28PM -0800, Darrick J. Wong wrote:
> On Tue, Mar 07, 2017 at 01:07:54AM +0100, Christoph Hellwig wrote:
> > I like killing it, but shouldn't we just try a normal kmem_zalloc?
> > At least for the fallback it's the right thing, and even for an
> > order 2 allocation it seems like a useful first try.
> 
> I'm confused -- kmem_zalloc_large tries kmem_zalloc with KM_MAYFAIL and
> only falls back to __vmalloc if it doesn't get anything.

Yup, that's right.

> Or maybe I've misunderstood, and you're asking if we should try
> kmem_zalloc(4 pages), then kmem_zalloc(1 page), and only then switch to
> the __vmalloc calls?

Just call kmem_zalloc_large() for 4 pages without a fallback on
failure - that's exactly how we handle allocations for things like
the 64k xattr buffers

Cheers,

Dave.
-- 
Dave Chinner
da...@fromorbit.com


Re: [PATCH] xfs: remove kmem_zalloc_greedy

2017-03-06 Thread Christoph Hellwig
I like killing it, but shouldn't we just try a normal kmem_zalloc?
At least for the fallback it's the right thing, and even for an
order 2 allocation it seems like a useful first try.


[PATCH] xfs: remove kmem_zalloc_greedy

2017-03-06 Thread Darrick J. Wong
The sole remaining caller of kmem_zalloc_greedy is bulkstat, which uses
it to grab 1-4 pages for staging of inobt records.  The infinite loop in
the greedy allocation function is causing hangs[1] in generic/269, so
just get rid of the greedy allocator in favor of kmem_zalloc_large.
This makes bulkstat somewhat more likely to ENOMEM if there's really no
pages to spare, but eliminates a source of hangs.

[1] https://lkml.org/lkml/2017/2/28/832

Signed-off-by: Darrick J. Wong 
---
 fs/xfs/kmem.c   |   18 --
 fs/xfs/kmem.h   |2 --
 fs/xfs/xfs_itable.c |   14 --
 3 files changed, 8 insertions(+), 26 deletions(-)

diff --git a/fs/xfs/kmem.c b/fs/xfs/kmem.c
index 339c696..bb2beae 100644
--- a/fs/xfs/kmem.c
+++ b/fs/xfs/kmem.c
@@ -24,24 +24,6 @@
 #include "kmem.h"
 #include "xfs_message.h"
 
-/*
- * Greedy allocation.  May fail and may return vmalloced memory.
- */
-void *
-kmem_zalloc_greedy(size_t *size, size_t minsize, size_t maxsize)
-{
-   void*ptr;
-   size_t  kmsize = maxsize;
-
-   while (!(ptr = vzalloc(kmsize))) {
-   if ((kmsize >>= 1) <= minsize)
-   kmsize = minsize;
-   }
-   if (ptr)
-   *size = kmsize;
-   return ptr;
-}
-
 void *
 kmem_alloc(size_t size, xfs_km_flags_t flags)
 {
diff --git a/fs/xfs/kmem.h b/fs/xfs/kmem.h
index 689f746..f0fc84f 100644
--- a/fs/xfs/kmem.h
+++ b/fs/xfs/kmem.h
@@ -69,8 +69,6 @@ static inline void  kmem_free(const void *ptr)
 }
 
 
-extern void *kmem_zalloc_greedy(size_t *, size_t, size_t);
-
 static inline void *
 kmem_zalloc(size_t size, xfs_km_flags_t flags)
 {
diff --git a/fs/xfs/xfs_itable.c b/fs/xfs/xfs_itable.c
index 8b2150d..283e76c 100644
--- a/fs/xfs/xfs_itable.c
+++ b/fs/xfs/xfs_itable.c
@@ -362,7 +362,6 @@ xfs_bulkstat(
xfs_agino_t agino;  /* inode # in allocation group */
xfs_agnumber_t  agno;   /* allocation group number */
xfs_btree_cur_t *cur;   /* btree cursor for ialloc btree */
-   size_t  irbsize; /* size of irec buffer in bytes */
xfs_inobt_rec_incore_t  *irbuf; /* start of irec buffer */
int nirbuf; /* size of irbuf */
int ubcount; /* size of user's buffer */
@@ -389,11 +388,14 @@ xfs_bulkstat(
*ubcountp = 0;
*done = 0;
 
-   irbuf = kmem_zalloc_greedy(&irbsize, PAGE_SIZE, PAGE_SIZE * 4);
-   if (!irbuf)
-   return -ENOMEM;
-
-   nirbuf = irbsize / sizeof(*irbuf);
+   nirbuf = (PAGE_SIZE * 4) / sizeof(*irbuf);
+   irbuf = kmem_zalloc_large(PAGE_SIZE * 4, KM_SLEEP);
+   if (!irbuf) {
+   irbuf = kmem_zalloc_large(PAGE_SIZE, KM_SLEEP);
+   if (!irbuf)
+   return -ENOMEM;
+   nirbuf /= 4;
+   }
 
/*
 * Loop over the allocation groups, starting from the last