In large memory(32TB) powerpc servers, we watched several soft lockup under
stress tests.
The call trace are as follows:
1.
get_page_from_freelist+0x2d8/0xd50  
__alloc_pages_nodemask+0x180/0xc20  
alloc_fresh_huge_page+0xb0/0x190    
set_max_huge_pages+0x164/0x3b0      

2.
prep_new_huge_page+0x5c/0x100             
alloc_fresh_huge_page+0xc8/0x190          
set_max_huge_pages+0x164/0x3b0

This patch is to fix such soft lockup. I thouhgt it is safe to call 
cond_resched() because alloc_fresh_gigantic_page and alloc_fresh_huge_page 
are out of spin_lock/unlock section.

Signed-off-by: Jia He <hejia...@gmail.com>
Cc: Andrew Morton <a...@linux-foundation.org>
Cc: Naoya Horiguchi <n-horigu...@ah.jp.nec.com>
Cc: Mike Kravetz <mike.krav...@oracle.com>
Cc: "Kirill A. Shutemov" <kirill.shute...@linux.intel.com>
Cc: Michal Hocko <mho...@suse.com>
Cc: Dave Hansen <dave.han...@linux.intel.com>
Cc: Paul Gortmaker <paul.gortma...@windriver.com>

---
 mm/hugetlb.c | 8 ++++++++
 1 file changed, 8 insertions(+)

diff --git a/mm/hugetlb.c b/mm/hugetlb.c
index addfe4ac..d51759d 100644
--- a/mm/hugetlb.c
+++ b/mm/hugetlb.c
@@ -1146,6 +1146,10 @@ static int alloc_fresh_gigantic_page(struct hstate *h,
 
        for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) {
                page = alloc_fresh_gigantic_page_node(h, node);
+
+               /* yield cpu */
+               cond_resched();
+
                if (page)
                        return 1;
        }
@@ -1381,6 +1385,10 @@ static int alloc_fresh_huge_page(struct hstate *h, 
nodemask_t *nodes_allowed)
 
        for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) {
                page = alloc_fresh_huge_page_node(h, node);
+
+               /* yield cpu */
+               cond_resched();
+
                if (page) {
                        ret = 1;
                        break;
-- 
2.5.0

Reply via email to