This refactors ttm_pool_free_page(), and by adding extra entries
to ttm_pool_page_dat, we then use it for all allocations, which
allows us to simplify the arguments needed to be passed to
ttm_pool_free_page().

This is critical for allowing the free function to be called
by the sharable drm_page_pool logic.

Cc: Daniel Vetter <dan...@ffwll.ch>
Cc: Christian Koenig <christian.koe...@amd.com>
Cc: Sumit Semwal <sumit.sem...@linaro.org>
Cc: Liam Mark <lm...@codeaurora.org>
Cc: Chris Goldsworthy <cgold...@codeaurora.org>
Cc: Laura Abbott <labb...@kernel.org>
Cc: Brian Starkey <brian.star...@arm.com>
Cc: Hridya Valsaraju <hri...@google.com>
Cc: Suren Baghdasaryan <sur...@google.com>
Cc: Sandeep Patil <sspa...@google.com>
Cc: Daniel Mentz <danielme...@google.com>
Cc: Ørjan Eide <orjan.e...@arm.com>
Cc: Robin Murphy <robin.mur...@arm.com>
Cc: Ezequiel Garcia <ezequ...@collabora.com>
Cc: Simon Ser <cont...@emersion.fr>
Cc: James Jones <jajo...@nvidia.com>
Cc: linux-me...@vger.kernel.org
Cc: dri-de...@lists.freedesktop.org
Signed-off-by: John Stultz <john.stu...@linaro.org>
---
 drivers/gpu/drm/ttm/ttm_pool.c | 60 ++++++++++++++++++----------------
 1 file changed, 32 insertions(+), 28 deletions(-)

diff --git a/drivers/gpu/drm/ttm/ttm_pool.c b/drivers/gpu/drm/ttm/ttm_pool.c
index c0274e256be3..eca36678f967 100644
--- a/drivers/gpu/drm/ttm/ttm_pool.c
+++ b/drivers/gpu/drm/ttm/ttm_pool.c
@@ -44,10 +44,14 @@
 /**
  * struct ttm_pool_page_dat - Helper object for coherent DMA mappings
  *
+ * @pool: ttm_pool pointer the page was allocated by
+ * @caching: the caching value the allocated page was configured for
  * @addr: original DMA address returned for the mapping
  * @vaddr: original vaddr return for the mapping and order in the lower bits
  */
 struct ttm_pool_page_dat {
+       struct ttm_pool *pool;
+       enum ttm_caching caching;
        dma_addr_t addr;
        unsigned long vaddr;
 };
@@ -71,13 +75,20 @@ static struct shrinker mm_shrinker;
 
 /* Allocate pages of size 1 << order with the given gfp_flags */
 static struct page *ttm_pool_alloc_page(struct ttm_pool *pool, gfp_t gfp_flags,
-                                       unsigned int order)
+                                       unsigned int order, enum ttm_caching 
caching)
 {
        unsigned long attr = DMA_ATTR_FORCE_CONTIGUOUS;
        struct ttm_pool_page_dat *dat;
        struct page *p;
        void *vaddr;
 
+       dat = kmalloc(sizeof(*dat), GFP_KERNEL);
+       if (!dat)
+               return NULL;
+
+       dat->pool = pool;
+       dat->caching = caching;
+
        /* Don't set the __GFP_COMP flag for higher order allocations.
         * Mapping pages directly into an userspace process and calling
         * put_page() on a TTM allocated page is illegal.
@@ -88,15 +99,13 @@ static struct page *ttm_pool_alloc_page(struct ttm_pool 
*pool, gfp_t gfp_flags,
 
        if (!pool->use_dma_alloc) {
                p = alloc_pages(gfp_flags, order);
-               if (p)
-                       p->private = order;
+               if (!p)
+                       goto error_free;
+               dat->vaddr = order;
+               p->private = (unsigned long)dat;
                return p;
        }
 
-       dat = kmalloc(sizeof(*dat), GFP_KERNEL);
-       if (!dat)
-               return NULL;
-
        if (order)
                attr |= DMA_ATTR_NO_WARN;
 
@@ -123,34 +132,34 @@ static struct page *ttm_pool_alloc_page(struct ttm_pool 
*pool, gfp_t gfp_flags,
 }
 
 /* Reset the caching and pages of size 1 << order */
-static void ttm_pool_free_page(struct ttm_pool *pool, enum ttm_caching caching,
-                              unsigned int order, struct page *p)
+static int ttm_pool_free_page(struct page *p, unsigned int order)
 {
        unsigned long attr = DMA_ATTR_FORCE_CONTIGUOUS;
-       struct ttm_pool_page_dat *dat;
+       struct ttm_pool_page_dat *dat = (void *)p->private;
        void *vaddr;
 
 #ifdef CONFIG_X86
        /* We don't care that set_pages_wb is inefficient here. This is only
         * used when we have to shrink and CPU overhead is irrelevant then.
         */
-       if (caching != ttm_cached && !PageHighMem(p))
+       if (dat->caching != ttm_cached && !PageHighMem(p))
                set_pages_wb(p, 1 << order);
 #endif
 
-       if (!pool || !pool->use_dma_alloc) {
+       if (!dat->pool || !dat->pool->use_dma_alloc) {
                __free_pages(p, order);
-               return;
+               goto out;
        }
 
        if (order)
                attr |= DMA_ATTR_NO_WARN;
 
-       dat = (void *)p->private;
        vaddr = (void *)(dat->vaddr & PAGE_MASK);
-       dma_free_attrs(pool->dev, (1UL << order) * PAGE_SIZE, vaddr, dat->addr,
+       dma_free_attrs(dat->pool->dev, (1UL << order) * PAGE_SIZE, vaddr, 
dat->addr,
                       attr);
+out:
        kfree(dat);
+       return 1 << order;
 }
 
 /* Apply a new caching to an array of pages */
@@ -264,7 +273,7 @@ static void ttm_pool_type_fini(struct ttm_pool_type *pt)
        mutex_unlock(&shrinker_lock);
 
        list_for_each_entry_safe(p, tmp, &pt->pages, lru)
-               ttm_pool_free_page(pt->pool, pt->caching, pt->order, p);
+               ttm_pool_free_page(p, pt->order);
 }
 
 /* Return the pool_type to use for the given caching and order */
@@ -307,7 +316,7 @@ static unsigned int ttm_pool_shrink(void)
 
        p = ttm_pool_type_take(pt);
        if (p) {
-               ttm_pool_free_page(pt->pool, pt->caching, pt->order, p);
+               ttm_pool_free_page(p, pt->order);
                num_freed = 1 << pt->order;
        } else {
                num_freed = 0;
@@ -322,13 +331,9 @@ static unsigned int ttm_pool_shrink(void)
 /* Return the allocation order based for a page */
 static unsigned int ttm_pool_page_order(struct ttm_pool *pool, struct page *p)
 {
-       if (pool->use_dma_alloc) {
-               struct ttm_pool_page_dat *dat = (void *)p->private;
-
-               return dat->vaddr & ~PAGE_MASK;
-       }
+       struct ttm_pool_page_dat *dat = (void *)p->private;
 
-       return p->private;
+       return dat->vaddr & ~PAGE_MASK;
 }
 
 /**
@@ -379,7 +384,7 @@ int ttm_pool_alloc(struct ttm_pool *pool, struct ttm_tt *tt,
                if (p) {
                        apply_caching = true;
                } else {
-                       p = ttm_pool_alloc_page(pool, gfp_flags, order);
+                       p = ttm_pool_alloc_page(pool, gfp_flags, order, 
tt->caching);
                        if (p && PageHighMem(p))
                                apply_caching = true;
                }
@@ -428,13 +433,13 @@ int ttm_pool_alloc(struct ttm_pool *pool, struct ttm_tt 
*tt,
        ttm_mem_global_free_page(&ttm_mem_glob, p, (1 << order) * PAGE_SIZE);
 
 error_free_page:
-       ttm_pool_free_page(pool, tt->caching, order, p);
+       ttm_pool_free_page(p, order);
 
 error_free_all:
        num_pages = tt->num_pages - num_pages;
        for (i = 0; i < num_pages; ) {
                order = ttm_pool_page_order(pool, tt->pages[i]);
-               ttm_pool_free_page(pool, tt->caching, order, tt->pages[i]);
+               ttm_pool_free_page(tt->pages[i], order);
                i += 1 << order;
        }
 
@@ -470,8 +475,7 @@ void ttm_pool_free(struct ttm_pool *pool, struct ttm_tt *tt)
                if (pt)
                        ttm_pool_type_give(pt, tt->pages[i]);
                else
-                       ttm_pool_free_page(pool, tt->caching, order,
-                                          tt->pages[i]);
+                       ttm_pool_free_page(tt->pages[i], order);
 
                i += num_pages;
        }
-- 
2.25.1

Reply via email to