Hi,

We are running into lockups during the memory pressure tests on our
boards, which essentially NMI panic them. In short the test case is

- THP shmem
    echo advise > /sys/kernel/mm/transparent_hugepage/shmem_enabled

- And a user-space process doing madvise(MADV_HUGEPAGE) on new mappings,
  and madvise(MADV_REMOVE) when it wants to remove the page range

The problem boils down to the reverse locking chain:
        kswapd does

                lock_page(page) -> down_read(page->mapping->i_mmap_rwsem)

        madvise() process does

                down_write(page->mapping->i_mmap_rwsem) -> lock_page(page)



CPU0                                                       CPU1

kswapd                                                     vfs_fallocate()
 shrink_node()                                              shmem_fallocate()
  shrink_active_list()                                       
unmap_mapping_range()
   page_referenced() << lock page:PG_locked >>                
unmap_mapping_pages()  << down_write(mapping->i_mmap_rwsem) >>
    rmap_walk_file()                                           
zap_page_range_single()
     down_read(mapping->i_mmap_rwsem) << W-locked on CPU1>>     
unmap_page_range()
      rwsem_down_read_failed()                                   
__split_huge_pmd()
       __rwsem_down_read_failed_common()                          __lock_page() 
 << PG_locked on CPU0 >>
        schedule()                                                 
wait_on_page_bit_common()
                                                                    
io_schedule()

        -ss

Reply via email to