Re: [PATCH] mm/slub: fix accumulate per cpu partial cache objects

2013-12-30 Thread Pekka Enberg

On 12/30/2013 03:08 AM, Wanpeng Li wrote:

Zefan's patch is good enough, mine doesn't need any more.


OK, thanks guys!

Pekka
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: [PATCH] mm/slub: fix accumulate per cpu partial cache objects

2013-12-30 Thread Pekka Enberg

On 12/30/2013 03:08 AM, Wanpeng Li wrote:

Zefan's patch is good enough, mine doesn't need any more.


OK, thanks guys!

Pekka
--
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: [PATCH] mm/slub: fix accumulate per cpu partial cache objects

2013-12-29 Thread Pekka Enberg
On Sat, Dec 28, 2013 at 3:50 AM, Li Zefan  wrote:
> On 2013/12/27 17:46, Wanpeng Li wrote:
>> SLUB per cpu partial cache is a list of slab caches to accelerate objects
>> allocation. However, current codes just accumulate the objects number of
>> the first slab cache of per cpu partial cache instead of traverse the whole
>> list.
>>
>> Signed-off-by: Wanpeng Li 
>> ---
>>  mm/slub.c |   32 +++-
>>  1 files changed, 23 insertions(+), 9 deletions(-)
>>
>> diff --git a/mm/slub.c b/mm/slub.c
>> index 545a170..799bfdc 100644
>> --- a/mm/slub.c
>> +++ b/mm/slub.c
>> @@ -4280,7 +4280,7 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
>>   struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab,
>>  cpu);
>>   int node;
>> - struct page *page;
>> + struct page *page, *p;
>>
>>   page = ACCESS_ONCE(c->page);
>>   if (!page)
>> @@ -4298,8 +4298,9 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
>>   nodes[node] += x;
>>
>>   page = ACCESS_ONCE(c->partial);
>> - if (page) {
>> - x = page->pobjects;
>> + while ((p = page)) {
>> + page = p->next;
>> + x = p->pobjects;
>>   total += x;
>>   nodes[node] += x;
>>   }
>
> Can we apply this patch first? It was sent month ago, but Pekka was not 
> responsive.

Applied. Wanpeng, care to resend your patch?
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: [PATCH] mm/slub: fix accumulate per cpu partial cache objects

2013-12-29 Thread Pekka Enberg
On Sat, Dec 28, 2013 at 3:50 AM, Li Zefan lize...@huawei.com wrote:
 On 2013/12/27 17:46, Wanpeng Li wrote:
 SLUB per cpu partial cache is a list of slab caches to accelerate objects
 allocation. However, current codes just accumulate the objects number of
 the first slab cache of per cpu partial cache instead of traverse the whole
 list.

 Signed-off-by: Wanpeng Li liw...@linux.vnet.ibm.com
 ---
  mm/slub.c |   32 +++-
  1 files changed, 23 insertions(+), 9 deletions(-)

 diff --git a/mm/slub.c b/mm/slub.c
 index 545a170..799bfdc 100644
 --- a/mm/slub.c
 +++ b/mm/slub.c
 @@ -4280,7 +4280,7 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
   struct kmem_cache_cpu *c = per_cpu_ptr(s-cpu_slab,
  cpu);
   int node;
 - struct page *page;
 + struct page *page, *p;

   page = ACCESS_ONCE(c-page);
   if (!page)
 @@ -4298,8 +4298,9 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
   nodes[node] += x;

   page = ACCESS_ONCE(c-partial);
 - if (page) {
 - x = page-pobjects;
 + while ((p = page)) {
 + page = p-next;
 + x = p-pobjects;
   total += x;
   nodes[node] += x;
   }

 Can we apply this patch first? It was sent month ago, but Pekka was not 
 responsive.

Applied. Wanpeng, care to resend your patch?
--
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: [PATCH] mm/slub: fix accumulate per cpu partial cache objects

2013-12-27 Thread Li Zefan
On 2013/12/27 17:46, Wanpeng Li wrote:
> SLUB per cpu partial cache is a list of slab caches to accelerate objects 
> allocation. However, current codes just accumulate the objects number of 
> the first slab cache of per cpu partial cache instead of traverse the whole 
> list.
> 
> Signed-off-by: Wanpeng Li 
> ---
>  mm/slub.c |   32 +++-
>  1 files changed, 23 insertions(+), 9 deletions(-)
> 
> diff --git a/mm/slub.c b/mm/slub.c
> index 545a170..799bfdc 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -4280,7 +4280,7 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
>   struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab,
>  cpu);
>   int node;
> - struct page *page;
> + struct page *page, *p;
>  
>   page = ACCESS_ONCE(c->page);
>   if (!page)
> @@ -4298,8 +4298,9 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
>   nodes[node] += x;
>  
>   page = ACCESS_ONCE(c->partial);
> - if (page) {
> - x = page->pobjects;
> + while ((p = page)) {
> + page = p->next;
> + x = p->pobjects;
>   total += x;
>   nodes[node] += x;
>   }

Can we apply this patch first? It was sent month ago, but Pekka was not 
responsive.

=

[PATCH] slub: Fix calculation of cpu slabs

  /sys/kernel/slab/:t-048 # cat cpu_slabs
  231 N0=16 N1=215
  /sys/kernel/slab/:t-048 # cat slabs
  145 N0=36 N1=109

See, the number of slabs is smaller than that of cpu slabs.

The bug was introduced by commit 49e2258586b423684f03c278149ab46d8f8b6700
("slub: per cpu cache for partial pages").

We should use page->pages instead of page->pobjects when calculating
the number of cpu partial slabs. This also fixes the mapping of slabs
and nodes.

As there's no variable storing the number of total/active objects in
cpu partial slabs, and we don't have user interfaces requiring those
statistics, I just add WARN_ON for those cases.

Cc:  # 3.2+
Signed-off-by: Li Zefan 
Acked-by: Christoph Lameter 
Reviewed-by: Wanpeng Li 
---
 mm/slub.c | 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)

diff --git a/mm/slub.c b/mm/slub.c
index e3ba1f2..6ea461d 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -4300,7 +4300,13 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
 
page = ACCESS_ONCE(c->partial);
if (page) {
-   x = page->pobjects;
+   node = page_to_nid(page);
+   if (flags & SO_TOTAL)
+   WARN_ON_ONCE(1);
+   else if (flags & SO_OBJECTS)
+   WARN_ON_ONCE(1);
+   else
+   x = page->pages;
total += x;
nodes[node] += x;
}
-- 1.8.0.2 

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


[PATCH] mm/slub: fix accumulate per cpu partial cache objects

2013-12-27 Thread Wanpeng Li
SLUB per cpu partial cache is a list of slab caches to accelerate objects 
allocation. However, current codes just accumulate the objects number of 
the first slab cache of per cpu partial cache instead of traverse the whole 
list.

Signed-off-by: Wanpeng Li 
---
 mm/slub.c |   32 +++-
 1 files changed, 23 insertions(+), 9 deletions(-)

diff --git a/mm/slub.c b/mm/slub.c
index 545a170..799bfdc 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -4280,7 +4280,7 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab,
   cpu);
int node;
-   struct page *page;
+   struct page *page, *p;
 
page = ACCESS_ONCE(c->page);
if (!page)
@@ -4298,8 +4298,9 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
nodes[node] += x;
 
page = ACCESS_ONCE(c->partial);
-   if (page) {
-   x = page->pobjects;
+   while ((p = page)) {
+   page = p->next;
+   x = p->pobjects;
total += x;
nodes[node] += x;
}
@@ -4520,13 +4521,15 @@ static ssize_t slabs_cpu_partial_show(struct kmem_cache 
*s, char *buf)
int pages = 0;
int cpu;
int len;
+   struct page *p;
 
for_each_online_cpu(cpu) {
struct page *page = per_cpu_ptr(s->cpu_slab, cpu)->partial;
 
-   if (page) {
-   pages += page->pages;
-   objects += page->pobjects;
+   while ((p = page)) {
+   page = p->next;
+   pages += p->pages;
+   objects += p->pobjects;
}
}
 
@@ -4535,10 +4538,21 @@ static ssize_t slabs_cpu_partial_show(struct kmem_cache 
*s, char *buf)
 #ifdef CONFIG_SMP
for_each_online_cpu(cpu) {
struct page *page = per_cpu_ptr(s->cpu_slab, cpu) ->partial;
+   objects = 0;
+   pages = 0;
+
+   if (!page)
+   continue;
+
+   while ((p = page)) {
+   page = p->next;
+   pages += p->pages;
+   objects += p->pobjects;
+   }
 
-   if (page && len < PAGE_SIZE - 20)
-   len += sprintf(buf + len, " C%d=%d(%d)", cpu,
-   page->pobjects, page->pages);
+   if (len < PAGE_SIZE - 20)
+   len += sprintf(buf + len, " C%d=%d(%d)", cpu,
+   objects, pages);
}
 #endif
return len + sprintf(buf + len, "\n");
-- 
1.7.7.6

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


[PATCH] mm/slub: fix accumulate per cpu partial cache objects

2013-12-27 Thread Wanpeng Li
SLUB per cpu partial cache is a list of slab caches to accelerate objects 
allocation. However, current codes just accumulate the objects number of 
the first slab cache of per cpu partial cache instead of traverse the whole 
list.

Signed-off-by: Wanpeng Li liw...@linux.vnet.ibm.com
---
 mm/slub.c |   32 +++-
 1 files changed, 23 insertions(+), 9 deletions(-)

diff --git a/mm/slub.c b/mm/slub.c
index 545a170..799bfdc 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -4280,7 +4280,7 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
struct kmem_cache_cpu *c = per_cpu_ptr(s-cpu_slab,
   cpu);
int node;
-   struct page *page;
+   struct page *page, *p;
 
page = ACCESS_ONCE(c-page);
if (!page)
@@ -4298,8 +4298,9 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
nodes[node] += x;
 
page = ACCESS_ONCE(c-partial);
-   if (page) {
-   x = page-pobjects;
+   while ((p = page)) {
+   page = p-next;
+   x = p-pobjects;
total += x;
nodes[node] += x;
}
@@ -4520,13 +4521,15 @@ static ssize_t slabs_cpu_partial_show(struct kmem_cache 
*s, char *buf)
int pages = 0;
int cpu;
int len;
+   struct page *p;
 
for_each_online_cpu(cpu) {
struct page *page = per_cpu_ptr(s-cpu_slab, cpu)-partial;
 
-   if (page) {
-   pages += page-pages;
-   objects += page-pobjects;
+   while ((p = page)) {
+   page = p-next;
+   pages += p-pages;
+   objects += p-pobjects;
}
}
 
@@ -4535,10 +4538,21 @@ static ssize_t slabs_cpu_partial_show(struct kmem_cache 
*s, char *buf)
 #ifdef CONFIG_SMP
for_each_online_cpu(cpu) {
struct page *page = per_cpu_ptr(s-cpu_slab, cpu) -partial;
+   objects = 0;
+   pages = 0;
+
+   if (!page)
+   continue;
+
+   while ((p = page)) {
+   page = p-next;
+   pages += p-pages;
+   objects += p-pobjects;
+   }
 
-   if (page  len  PAGE_SIZE - 20)
-   len += sprintf(buf + len,  C%d=%d(%d), cpu,
-   page-pobjects, page-pages);
+   if (len  PAGE_SIZE - 20)
+   len += sprintf(buf + len,  C%d=%d(%d), cpu,
+   objects, pages);
}
 #endif
return len + sprintf(buf + len, \n);
-- 
1.7.7.6

--
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: [PATCH] mm/slub: fix accumulate per cpu partial cache objects

2013-12-27 Thread Li Zefan
On 2013/12/27 17:46, Wanpeng Li wrote:
 SLUB per cpu partial cache is a list of slab caches to accelerate objects 
 allocation. However, current codes just accumulate the objects number of 
 the first slab cache of per cpu partial cache instead of traverse the whole 
 list.
 
 Signed-off-by: Wanpeng Li liw...@linux.vnet.ibm.com
 ---
  mm/slub.c |   32 +++-
  1 files changed, 23 insertions(+), 9 deletions(-)
 
 diff --git a/mm/slub.c b/mm/slub.c
 index 545a170..799bfdc 100644
 --- a/mm/slub.c
 +++ b/mm/slub.c
 @@ -4280,7 +4280,7 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
   struct kmem_cache_cpu *c = per_cpu_ptr(s-cpu_slab,
  cpu);
   int node;
 - struct page *page;
 + struct page *page, *p;
  
   page = ACCESS_ONCE(c-page);
   if (!page)
 @@ -4298,8 +4298,9 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
   nodes[node] += x;
  
   page = ACCESS_ONCE(c-partial);
 - if (page) {
 - x = page-pobjects;
 + while ((p = page)) {
 + page = p-next;
 + x = p-pobjects;
   total += x;
   nodes[node] += x;
   }

Can we apply this patch first? It was sent month ago, but Pekka was not 
responsive.

=

[PATCH] slub: Fix calculation of cpu slabs

  /sys/kernel/slab/:t-048 # cat cpu_slabs
  231 N0=16 N1=215
  /sys/kernel/slab/:t-048 # cat slabs
  145 N0=36 N1=109

See, the number of slabs is smaller than that of cpu slabs.

The bug was introduced by commit 49e2258586b423684f03c278149ab46d8f8b6700
(slub: per cpu cache for partial pages).

We should use page-pages instead of page-pobjects when calculating
the number of cpu partial slabs. This also fixes the mapping of slabs
and nodes.

As there's no variable storing the number of total/active objects in
cpu partial slabs, and we don't have user interfaces requiring those
statistics, I just add WARN_ON for those cases.

Cc: sta...@vger.kernel.org # 3.2+
Signed-off-by: Li Zefan lize...@huawei.com
Acked-by: Christoph Lameter c...@linux.com
Reviewed-by: Wanpeng Li liw...@linux.vnet.ibm.com
---
 mm/slub.c | 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)

diff --git a/mm/slub.c b/mm/slub.c
index e3ba1f2..6ea461d 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -4300,7 +4300,13 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
 
page = ACCESS_ONCE(c-partial);
if (page) {
-   x = page-pobjects;
+   node = page_to_nid(page);
+   if (flags  SO_TOTAL)
+   WARN_ON_ONCE(1);
+   else if (flags  SO_OBJECTS)
+   WARN_ON_ONCE(1);
+   else
+   x = page-pages;
total += x;
nodes[node] += x;
}
-- 1.8.0.2 

--
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/