From: Matthew Brost <[email protected]>

Introduce migrate_device_split_page() to split a device page into
lower-order pages. Used when a folio allocated as higher-order is freed
and later reallocated at a smaller order by the driver memory manager.

Cc: Andrew Morton <[email protected]>
Cc: Balbir Singh <[email protected]>
Cc: [email protected]
Cc: [email protected]
Signed-off-by: Matthew Brost <[email protected]>
Signed-off-by: Francois Dugast <[email protected]>
---
 include/linux/huge_mm.h |  3 +++
 include/linux/migrate.h |  1 +
 mm/huge_memory.c        |  6 ++---
 mm/migrate_device.c     | 49 +++++++++++++++++++++++++++++++++++++++++
 4 files changed, 56 insertions(+), 3 deletions(-)

diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h
index a4d9f964dfde..6ad8f359bc0d 100644
--- a/include/linux/huge_mm.h
+++ b/include/linux/huge_mm.h
@@ -374,6 +374,9 @@ int __split_huge_page_to_list_to_order(struct page *page, 
struct list_head *list
 int folio_split_unmapped(struct folio *folio, unsigned int new_order);
 unsigned int min_order_for_split(struct folio *folio);
 int split_folio_to_list(struct folio *folio, struct list_head *list);
+int __split_unmapped_folio(struct folio *folio, int new_order,
+                          struct page *split_at, struct xa_state *xas,
+                          struct address_space *mapping, enum split_type 
split_type);
 int folio_check_splittable(struct folio *folio, unsigned int new_order,
                           enum split_type split_type);
 int folio_split(struct folio *folio, unsigned int new_order, struct page *page,
diff --git a/include/linux/migrate.h b/include/linux/migrate.h
index 26ca00c325d9..ec65e4fd5f88 100644
--- a/include/linux/migrate.h
+++ b/include/linux/migrate.h
@@ -192,6 +192,7 @@ void migrate_device_pages(unsigned long *src_pfns, unsigned 
long *dst_pfns,
                        unsigned long npages);
 void migrate_device_finalize(unsigned long *src_pfns,
                        unsigned long *dst_pfns, unsigned long npages);
+int migrate_device_split_page(struct page *page);
 
 #endif /* CONFIG_MIGRATION */
 
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
index 40cf59301c21..7ded35a3ecec 100644
--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -3621,9 +3621,9 @@ static void __split_folio_to_order(struct folio *folio, 
int old_order,
  * Return: 0 - successful, <0 - failed (if -ENOMEM is returned, @folio might be
  * split but not to @new_order, the caller needs to check)
  */
-static int __split_unmapped_folio(struct folio *folio, int new_order,
-               struct page *split_at, struct xa_state *xas,
-               struct address_space *mapping, enum split_type split_type)
+int __split_unmapped_folio(struct folio *folio, int new_order,
+                          struct page *split_at, struct xa_state *xas,
+                          struct address_space *mapping, enum split_type 
split_type)
 {
        const bool is_anon = folio_test_anon(folio);
        int old_order = folio_order(folio);
diff --git a/mm/migrate_device.c b/mm/migrate_device.c
index 23379663b1e1..eb0f0e938947 100644
--- a/mm/migrate_device.c
+++ b/mm/migrate_device.c
@@ -775,6 +775,49 @@ int migrate_vma_setup(struct migrate_vma *args)
 EXPORT_SYMBOL(migrate_vma_setup);
 
 #ifdef CONFIG_ARCH_ENABLE_THP_MIGRATION
+/**
+ * migrate_device_split_page() - Split device page
+ * @page: Device page to split
+ *
+ * Splits a device page into smaller pages. Typically called when reallocating 
a
+ * folio to a smaller size. Inherently racy—only safe if the caller ensures
+ * mutual exclusion within the page's folio (i.e., no other threads are using
+ * pages within the folio). Expected to be called a free device page and
+ * restores all split out pages to a free state.
+ */
+int migrate_device_split_page(struct page *page)
+{
+       struct folio *folio = page_folio(page);
+       struct dev_pagemap *pgmap = folio->pgmap;
+       struct page *unlock_page = folio_page(folio, 0);
+       unsigned int order = folio_order(folio), i;
+       int ret = 0;
+
+       VM_BUG_ON_FOLIO(!order, folio);
+       VM_BUG_ON_FOLIO(!folio_is_device_private(folio), folio);
+       VM_BUG_ON_FOLIO(folio_ref_count(folio), folio);
+
+       folio_lock(folio);
+
+       ret = __split_unmapped_folio(folio, 0, page, NULL, NULL, 
SPLIT_TYPE_UNIFORM);
+       if (ret) {
+              /*
+               * We can't fail here unless the caller doesn't know what they
+               * are doing.
+               */
+               VM_BUG_ON_FOLIO(ret, folio);
+
+               return ret;
+       }
+
+       for (i = 0; i < 0x1 << order; ++i, ++unlock_page) {
+               page_folio(unlock_page)->pgmap = pgmap;
+               folio_unlock(page_folio(unlock_page));
+       }
+
+       return 0;
+}
+
 /**
  * migrate_vma_insert_huge_pmd_page: Insert a huge folio into 
@migrate->vma->vm_mm
  * at @addr. folio is already allocated as a part of the migration process with
@@ -927,6 +970,11 @@ static int migrate_vma_split_unmapped_folio(struct 
migrate_vma *migrate,
        return ret;
 }
 #else /* !CONFIG_ARCH_ENABLE_THP_MIGRATION */
+int migrate_device_split_page(struct page *page)
+{
+       return 0;
+}
+
 static int migrate_vma_insert_huge_pmd_page(struct migrate_vma *migrate,
                                         unsigned long addr,
                                         struct page *page,
@@ -943,6 +991,7 @@ static int migrate_vma_split_unmapped_folio(struct 
migrate_vma *migrate,
        return 0;
 }
 #endif
+EXPORT_SYMBOL(migrate_device_split_page);
 
 static unsigned long migrate_vma_nr_pages(unsigned long *src)
 {
-- 
2.43.0

Reply via email to