On 08/05/2010 09:09 AM, Ben Skeggs wrote:
> From: Ben Skeggs <bske...@redhat.com>
> 
> Nouveau will need this on GeForce 8 and up to account for the GPU
> reordering physical VRAM for some memory types.
> 
> Signed-off-by: Ben Skeggs <bske...@redhat.com>
Reviewed-by: Jerome Glisse <jgli...@redhat.com>
> ---
>  drivers/gpu/drm/nouveau/nouveau_bo.c       |   12 ++-
>  drivers/gpu/drm/nouveau/nouveau_channel.c  |    6 +-
>  drivers/gpu/drm/nouveau/nouveau_notifier.c |    2 +-
>  drivers/gpu/drm/nouveau/nouveau_sgdma.c    |    4 +-
>  drivers/gpu/drm/nouveau/nv50_crtc.c        |    3 +-
>  drivers/gpu/drm/nouveau/nv50_display.c     |    2 +-
>  drivers/gpu/drm/nouveau/nv50_instmem.c     |    2 +-
>  drivers/gpu/drm/radeon/radeon_object.c     |    6 +-
>  drivers/gpu/drm/radeon/radeon_ttm.c        |   16 ++--
>  drivers/gpu/drm/ttm/Makefile               |    3 +-
>  drivers/gpu/drm/ttm/ttm_agp_backend.c      |    3 +-
>  drivers/gpu/drm/ttm/ttm_bo.c               |  100 ++++---------------
>  drivers/gpu/drm/ttm/ttm_bo_manager.c       |  148 
> ++++++++++++++++++++++++++++
>  drivers/gpu/drm/ttm/ttm_bo_util.c          |    3 +-
>  drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c     |    3 +-
>  include/drm/ttm/ttm_bo_api.h               |    3 +-
>  include/drm/ttm/ttm_bo_driver.h            |   21 ++++-
>  17 files changed, 225 insertions(+), 112 deletions(-)
>  create mode 100644 drivers/gpu/drm/ttm/ttm_bo_manager.c
> 
> diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c 
> b/drivers/gpu/drm/nouveau/nouveau_bo.c
> index 92290fa..f403737 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_bo.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_bo.c
> @@ -390,6 +390,7 @@ nouveau_bo_init_mem_type(struct ttm_bo_device *bdev, 
> uint32_t type,
>               man->default_caching = TTM_PL_FLAG_CACHED;
>               break;
>       case TTM_PL_VRAM:
> +             man->func = &ttm_bo_manager_func;
>               man->flags = TTM_MEMTYPE_FLAG_FIXED |
>                            TTM_MEMTYPE_FLAG_MAPPABLE;
>               man->available_caching = TTM_PL_FLAG_UNCACHED |
> @@ -398,6 +399,7 @@ nouveau_bo_init_mem_type(struct ttm_bo_device *bdev, 
> uint32_t type,
>               man->gpu_offset = dev_priv->vm_vram_base;
>               break;
>       case TTM_PL_TT:
> +             man->func = &ttm_bo_manager_func;
>               switch (dev_priv->gart_info.type) {
>               case NOUVEAU_GART_AGP:
>                       man->flags = TTM_MEMTYPE_FLAG_MAPPABLE;
> @@ -500,8 +502,8 @@ nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int 
> evict, bool intr,
>       if (!chan || nvbo->tile_flags || nvbo->no_vm)
>               chan = dev_priv->channel;
>  
> -     src_offset = old_mem->mm_node->start << PAGE_SHIFT;
> -     dst_offset = new_mem->mm_node->start << PAGE_SHIFT;
> +     src_offset = old_mem->start << PAGE_SHIFT;
> +     dst_offset = new_mem->start << PAGE_SHIFT;
>       if (chan != dev_priv->channel) {
>               if (old_mem->mem_type == TTM_PL_TT)
>                       src_offset += dev_priv->vm_gart_base;
> @@ -650,7 +652,7 @@ nouveau_bo_vm_bind(struct ttm_buffer_object *bo, struct 
> ttm_mem_reg *new_mem,
>               return 0;
>       }
>  
> -     offset = new_mem->mm_node->start << PAGE_SHIFT;
> +     offset = new_mem->start << PAGE_SHIFT;
>  
>       if (dev_priv->card_type == NV_50) {
>               ret = nv50_mem_vm_bind_linear(dev,
> @@ -764,14 +766,14 @@ nouveau_ttm_io_mem_reserve(struct ttm_bo_device *bdev, 
> struct ttm_mem_reg *mem)
>       case TTM_PL_TT:
>  #if __OS_HAS_AGP
>               if (dev_priv->gart_info.type == NOUVEAU_GART_AGP) {
> -                     mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> +                     mem->bus.offset = mem->start << PAGE_SHIFT;
>                       mem->bus.base = dev_priv->gart_info.aper_base;
>                       mem->bus.is_iomem = true;
>               }
>  #endif
>               break;
>       case TTM_PL_VRAM:
> -             mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> +             mem->bus.offset = mem->start << PAGE_SHIFT;
>               mem->bus.base = pci_resource_start(dev->pdev, 1);
>               mem->bus.is_iomem = true;
>               break;
> diff --git a/drivers/gpu/drm/nouveau/nouveau_channel.c 
> b/drivers/gpu/drm/nouveau/nouveau_channel.c
> index 90fdcda..90d3450 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_channel.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_channel.c
> @@ -48,14 +48,14 @@ nouveau_channel_pushbuf_ctxdma_init(struct 
> nouveau_channel *chan)
>                                                 dev_priv->gart_info.aper_size,
>                                                 NV_DMA_ACCESS_RO, &pushbuf,
>                                                 NULL);
> -             chan->pushbuf_base = pb->bo.mem.mm_node->start << PAGE_SHIFT;
> +             chan->pushbuf_base = pb->bo.mem.start << PAGE_SHIFT;
>       } else
>       if (dev_priv->card_type != NV_04) {
>               ret = nouveau_gpuobj_dma_new(chan, NV_CLASS_DMA_IN_MEMORY, 0,
>                                            dev_priv->fb_available_size,
>                                            NV_DMA_ACCESS_RO,
>                                            NV_DMA_TARGET_VIDMEM, &pushbuf);
> -             chan->pushbuf_base = pb->bo.mem.mm_node->start << PAGE_SHIFT;
> +             chan->pushbuf_base = pb->bo.mem.start << PAGE_SHIFT;
>       } else {
>               /* NV04 cmdbuf hack, from original ddx.. not sure of it's
>                * exact reason for existing :)  PCI access to cmdbuf in
> @@ -67,7 +67,7 @@ nouveau_channel_pushbuf_ctxdma_init(struct nouveau_channel 
> *chan)
>                                            dev_priv->fb_available_size,
>                                            NV_DMA_ACCESS_RO,
>                                            NV_DMA_TARGET_PCI, &pushbuf);
> -             chan->pushbuf_base = pb->bo.mem.mm_node->start << PAGE_SHIFT;
> +             chan->pushbuf_base = pb->bo.mem.start << PAGE_SHIFT;
>       }
>  
>       ret = nouveau_gpuobj_ref_add(dev, chan, 0, pushbuf, &chan->pushbuf);
> diff --git a/drivers/gpu/drm/nouveau/nouveau_notifier.c 
> b/drivers/gpu/drm/nouveau/nouveau_notifier.c
> index 3ec181f..d218427 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_notifier.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_notifier.c
> @@ -112,7 +112,7 @@ nouveau_notifier_alloc(struct nouveau_channel *chan, 
> uint32_t handle,
>               return -ENOMEM;
>       }
>  
> -     offset = chan->notifier_bo->bo.mem.mm_node->start << PAGE_SHIFT;
> +     offset = chan->notifier_bo->bo.mem.start << PAGE_SHIFT;
>       if (chan->notifier_bo->bo.mem.mem_type == TTM_PL_VRAM) {
>               target = NV_DMA_TARGET_VIDMEM;
>       } else
> diff --git a/drivers/gpu/drm/nouveau/nouveau_sgdma.c 
> b/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> index 491767f..4374366 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> @@ -95,9 +95,9 @@ nouveau_sgdma_bind(struct ttm_backend *be, struct 
> ttm_mem_reg *mem)
>       struct nouveau_gpuobj *gpuobj = dev_priv->gart_info.sg_ctxdma;
>       unsigned i, j, pte;
>  
> -     NV_DEBUG(dev, "pg=0x%lx\n", mem->mm_node->start);
> +     NV_DEBUG(dev, "pg=0x%lx\n", mem->start);
>  
> -     pte = nouveau_sgdma_pte(nvbe->dev, mem->mm_node->start << PAGE_SHIFT);
> +     pte = nouveau_sgdma_pte(nvbe->dev, mem->start << PAGE_SHIFT);
>       nvbe->pte_start = pte;
>       for (i = 0; i < nvbe->nr_pages; i++) {
>               dma_addr_t dma_offset = nvbe->pages[i];
> diff --git a/drivers/gpu/drm/nouveau/nv50_crtc.c 
> b/drivers/gpu/drm/nouveau/nv50_crtc.c
> index 5d11ea1..a6a0f95 100644
> --- a/drivers/gpu/drm/nouveau/nv50_crtc.c
> +++ b/drivers/gpu/drm/nouveau/nv50_crtc.c
> @@ -104,8 +104,7 @@ nv50_crtc_blank(struct nouveau_crtc *nv_crtc, bool 
> blanked)
>               OUT_RING(evo, nv_crtc->lut.depth == 8 ?
>                               NV50_EVO_CRTC_CLUT_MODE_OFF :
>                               NV50_EVO_CRTC_CLUT_MODE_ON);
> -             OUT_RING(evo, (nv_crtc->lut.nvbo->bo.mem.mm_node->start <<
> -                              PAGE_SHIFT) >> 8);
> +             OUT_RING(evo, (nv_crtc->lut.nvbo->bo.mem.start << PAGE_SHIFT) 
> >> 8);
>               if (dev_priv->chipset != 0x50) {
>                       BEGIN_RING(evo, 0, NV84_EVO_CRTC(index, CLUT_DMA), 1);
>                       OUT_RING(evo, NvEvoVRAM);
> diff --git a/drivers/gpu/drm/nouveau/nv50_display.c 
> b/drivers/gpu/drm/nouveau/nv50_display.c
> index f13ad0d..5f88e93 100644
> --- a/drivers/gpu/drm/nouveau/nv50_display.c
> +++ b/drivers/gpu/drm/nouveau/nv50_display.c
> @@ -322,7 +322,7 @@ nv50_display_init(struct drm_device *dev)
>  
>       /* initialise fifo */
>       nv_wr32(dev, NV50_PDISPLAY_CHANNEL_DMA_CB(0),
> -             ((evo->pushbuf_bo->bo.mem.mm_node->start << PAGE_SHIFT) >> 8) |
> +             ((evo->pushbuf_bo->bo.mem.start << PAGE_SHIFT) >> 8) |
>               NV50_PDISPLAY_CHANNEL_DMA_CB_LOCATION_VRAM |
>               NV50_PDISPLAY_CHANNEL_DMA_CB_VALID);
>       nv_wr32(dev, NV50_PDISPLAY_CHANNEL_UNK2(0), 0x00010000);
> diff --git a/drivers/gpu/drm/nouveau/nv50_instmem.c 
> b/drivers/gpu/drm/nouveau/nv50_instmem.c
> index 37c7b48..a35e04c 100644
> --- a/drivers/gpu/drm/nouveau/nv50_instmem.c
> +++ b/drivers/gpu/drm/nouveau/nv50_instmem.c
> @@ -397,7 +397,7 @@ nv50_instmem_populate(struct drm_device *dev, struct 
> nouveau_gpuobj *gpuobj,
>               return ret;
>       }
>  
> -     gpuobj->im_backing_start = gpuobj->im_backing->bo.mem.mm_node->start;
> +     gpuobj->im_backing_start = gpuobj->im_backing->bo.mem.start;
>       gpuobj->im_backing_start <<= PAGE_SHIFT;
>  
>       return 0;
> diff --git a/drivers/gpu/drm/radeon/radeon_object.c 
> b/drivers/gpu/drm/radeon/radeon_object.c
> index 0afd1e6..c261060 100644
> --- a/drivers/gpu/drm/radeon/radeon_object.c
> +++ b/drivers/gpu/drm/radeon/radeon_object.c
> @@ -435,7 +435,7 @@ int radeon_bo_get_surface_reg(struct radeon_bo *bo)
>  
>  out:
>       radeon_set_surface_reg(rdev, i, bo->tiling_flags, bo->pitch,
> -                            bo->tbo.mem.mm_node->start << PAGE_SHIFT,
> +                            bo->tbo.mem.start << PAGE_SHIFT,
>                              bo->tbo.num_pages << PAGE_SHIFT);
>       return 0;
>  }
> @@ -532,7 +532,7 @@ int radeon_bo_fault_reserve_notify(struct 
> ttm_buffer_object *bo)
>       rdev = rbo->rdev;
>       if (bo->mem.mem_type == TTM_PL_VRAM) {
>               size = bo->mem.num_pages << PAGE_SHIFT;
> -             offset = bo->mem.mm_node->start << PAGE_SHIFT;
> +             offset = bo->mem.start << PAGE_SHIFT;
>               if ((offset + size) > rdev->mc.visible_vram_size) {
>                       /* hurrah the memory is not visible ! */
>                       radeon_ttm_placement_from_domain(rbo, 
> RADEON_GEM_DOMAIN_VRAM);
> @@ -540,7 +540,7 @@ int radeon_bo_fault_reserve_notify(struct 
> ttm_buffer_object *bo)
>                       r = ttm_bo_validate(bo, &rbo->placement, false, true, 
> false);
>                       if (unlikely(r != 0))
>                               return r;
> -                     offset = bo->mem.mm_node->start << PAGE_SHIFT;
> +                     offset = bo->mem.start << PAGE_SHIFT;
>                       /* this should not happen */
>                       if ((offset + size) > rdev->mc.visible_vram_size)
>                               return -EINVAL;
> diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c 
> b/drivers/gpu/drm/radeon/radeon_ttm.c
> index cc19aba..0921910 100644
> --- a/drivers/gpu/drm/radeon/radeon_ttm.c
> +++ b/drivers/gpu/drm/radeon/radeon_ttm.c
> @@ -152,6 +152,7 @@ static int radeon_init_mem_type(struct ttm_bo_device 
> *bdev, uint32_t type,
>               man->default_caching = TTM_PL_FLAG_CACHED;
>               break;
>       case TTM_PL_TT:
> +             man->func = &ttm_bo_manager_func;
>               man->gpu_offset = rdev->mc.gtt_start;
>               man->available_caching = TTM_PL_MASK_CACHING;
>               man->default_caching = TTM_PL_FLAG_CACHED;
> @@ -173,6 +174,7 @@ static int radeon_init_mem_type(struct ttm_bo_device 
> *bdev, uint32_t type,
>               break;
>       case TTM_PL_VRAM:
>               /* "On-card" video ram */
> +             man->func = &ttm_bo_manager_func;
>               man->gpu_offset = rdev->mc.vram_start;
>               man->flags = TTM_MEMTYPE_FLAG_FIXED |
>                            TTM_MEMTYPE_FLAG_MAPPABLE;
> @@ -246,8 +248,8 @@ static int radeon_move_blit(struct ttm_buffer_object *bo,
>       if (unlikely(r)) {
>               return r;
>       }
> -     old_start = old_mem->mm_node->start << PAGE_SHIFT;
> -     new_start = new_mem->mm_node->start << PAGE_SHIFT;
> +     old_start = old_mem->start << PAGE_SHIFT;
> +     new_start = new_mem->start << PAGE_SHIFT;
>  
>       switch (old_mem->mem_type) {
>       case TTM_PL_VRAM:
> @@ -435,14 +437,14 @@ static int radeon_ttm_io_mem_reserve(struct 
> ttm_bo_device *bdev, struct ttm_mem_
>  #if __OS_HAS_AGP
>               if (rdev->flags & RADEON_IS_AGP) {
>                       /* RADEON_IS_AGP is set only if AGP is active */
> -                     mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> +                     mem->bus.offset = mem->start << PAGE_SHIFT;
>                       mem->bus.base = rdev->mc.agp_base;
>                       mem->bus.is_iomem = !rdev->ddev->agp->cant_use_aperture;
>               }
>  #endif
>               break;
>       case TTM_PL_VRAM:
> -             mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> +             mem->bus.offset = mem->start << PAGE_SHIFT;
>               /* check if it's visible */
>               if ((mem->bus.offset + mem->bus.size) > 
> rdev->mc.visible_vram_size)
>                       return -EINVAL;
> @@ -685,7 +687,7 @@ static int radeon_ttm_backend_bind(struct ttm_backend 
> *backend,
>       int r;
>  
>       gtt = container_of(backend, struct radeon_ttm_backend, backend);
> -     gtt->offset = bo_mem->mm_node->start << PAGE_SHIFT;
> +     gtt->offset = bo_mem->start << PAGE_SHIFT;
>       if (!gtt->num_pages) {
>               WARN(1, "nothing to bind %lu pages for mreg %p back %p!\n", 
> gtt->num_pages, bo_mem, backend);
>       }
> @@ -784,9 +786,9 @@ static int radeon_ttm_debugfs_init(struct radeon_device 
> *rdev)
>               radeon_mem_types_list[i].show = &radeon_mm_dump_table;
>               radeon_mem_types_list[i].driver_features = 0;
>               if (i == 0)
> -                     radeon_mem_types_list[i].data = 
> &rdev->mman.bdev.man[TTM_PL_VRAM].manager;
> +                     radeon_mem_types_list[i].data = 
> &rdev->mman.bdev.man[TTM_PL_VRAM].priv;
>               else
> -                     radeon_mem_types_list[i].data = 
> &rdev->mman.bdev.man[TTM_PL_TT].manager;
> +                     radeon_mem_types_list[i].data = 
> &rdev->mman.bdev.man[TTM_PL_TT].priv;
>  
>       }
>       /* Add ttm page pool to debugfs */
> diff --git a/drivers/gpu/drm/ttm/Makefile b/drivers/gpu/drm/ttm/Makefile
> index b256d4a..f3cf6f0 100644
> --- a/drivers/gpu/drm/ttm/Makefile
> +++ b/drivers/gpu/drm/ttm/Makefile
> @@ -4,6 +4,7 @@
>  ccflags-y := -Iinclude/drm
>  ttm-y := ttm_agp_backend.o ttm_memory.o ttm_tt.o ttm_bo.o \
>       ttm_bo_util.o ttm_bo_vm.o ttm_module.o \
> -     ttm_object.o ttm_lock.o ttm_execbuf_util.o ttm_page_alloc.o
> +     ttm_object.o ttm_lock.o ttm_execbuf_util.o ttm_page_alloc.o \
> +     ttm_bo_manager.o
>  
>  obj-$(CONFIG_DRM_TTM) += ttm.o
> diff --git a/drivers/gpu/drm/ttm/ttm_agp_backend.c 
> b/drivers/gpu/drm/ttm/ttm_agp_backend.c
> index 4bf69c4..f999e36 100644
> --- a/drivers/gpu/drm/ttm/ttm_agp_backend.c
> +++ b/drivers/gpu/drm/ttm/ttm_agp_backend.c
> @@ -74,6 +74,7 @@ static int ttm_agp_bind(struct ttm_backend *backend, struct 
> ttm_mem_reg *bo_mem)
>  {
>       struct ttm_agp_backend *agp_be =
>           container_of(backend, struct ttm_agp_backend, backend);
> +     struct drm_mm_node *node = bo_mem->mm_node;
>       struct agp_memory *mem = agp_be->mem;
>       int cached = (bo_mem->placement & TTM_PL_FLAG_CACHED);
>       int ret;
> @@ -81,7 +82,7 @@ static int ttm_agp_bind(struct ttm_backend *backend, struct 
> ttm_mem_reg *bo_mem)
>       mem->is_flushed = 1;
>       mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY;
>  
> -     ret = agp_bind_memory(mem, bo_mem->mm_node->start);
> +     ret = agp_bind_memory(mem, node->start);
>       if (ret)
>               printk(KERN_ERR TTM_PFX "AGP Bind memory failed.\n");
>  
> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
> index 80d37b4..af7b57a 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo.c
> @@ -84,11 +84,8 @@ static void ttm_mem_type_debug(struct ttm_bo_device *bdev, 
> int mem_type)
>               man->available_caching);
>       printk(KERN_ERR TTM_PFX "    default_caching: 0x%08X\n",
>               man->default_caching);
> -     if (mem_type != TTM_PL_SYSTEM) {
> -             spin_lock(&bdev->glob->lru_lock);
> -             drm_mm_debug_table(&man->manager, TTM_PFX);
> -             spin_unlock(&bdev->glob->lru_lock);
> -     }
> +     if (mem_type != TTM_PL_SYSTEM)
> +             (*man->func->debug)(man, TTM_PFX);
>  }
>  
>  static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo,
> @@ -421,7 +418,7 @@ moved:
>  
>       if (bo->mem.mm_node) {
>               spin_lock(&bo->lock);
> -             bo->offset = (bo->mem.mm_node->start << PAGE_SHIFT) +
> +             bo->offset = (bo->mem.start << PAGE_SHIFT) +
>                   bdev->man[bo->mem.mem_type].gpu_offset;
>               bo->cur_placement = bo->mem.placement;
>               spin_unlock(&bo->lock);
> @@ -724,52 +721,12 @@ retry:
>       return ret;
>  }
>  
> -static int ttm_bo_man_get_node(struct ttm_buffer_object *bo,
> -                             struct ttm_mem_type_manager *man,
> -                             struct ttm_placement *placement,
> -                             struct ttm_mem_reg *mem,
> -                             struct drm_mm_node **node)
> -{
> -     struct ttm_bo_global *glob = bo->glob;
> -     unsigned long lpfn;
> -     int ret;
> -
> -     lpfn = placement->lpfn;
> -     if (!lpfn)
> -             lpfn = man->size;
> -     *node = NULL;
> -     do {
> -             ret = drm_mm_pre_get(&man->manager);
> -             if (unlikely(ret))
> -                     return ret;
> -
> -             spin_lock(&glob->lru_lock);
> -             *node = drm_mm_search_free_in_range(&man->manager,
> -                                     mem->num_pages, mem->page_alignment,
> -                                     placement->fpfn, lpfn, 1);
> -             if (unlikely(*node == NULL)) {
> -                     spin_unlock(&glob->lru_lock);
> -                     return 0;
> -             }
> -             *node = drm_mm_get_block_atomic_range(*node, mem->num_pages,
> -                                                     mem->page_alignment,
> -                                                     placement->fpfn,
> -                                                     lpfn);
> -             spin_unlock(&glob->lru_lock);
> -     } while (*node == NULL);
> -     return 0;
> -}
> -
>  void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem)
>  {
> -     struct ttm_bo_global *glob = bo->glob;
> +     struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type];
>  
> -     if (mem->mm_node) {
> -             spin_lock(&glob->lru_lock);
> -             drm_mm_put_block(mem->mm_node);
> -             spin_unlock(&glob->lru_lock);
> -             mem->mm_node = NULL;
> -     }
> +     if (mem->mm_node)
> +             (*man->func->put_node)(man, mem);
>  }
>  EXPORT_SYMBOL(ttm_bo_mem_put);
>  
> @@ -788,14 +745,13 @@ static int ttm_bo_mem_force_space(struct 
> ttm_buffer_object *bo,
>       struct ttm_bo_device *bdev = bo->bdev;
>       struct ttm_bo_global *glob = bdev->glob;
>       struct ttm_mem_type_manager *man = &bdev->man[mem_type];
> -     struct drm_mm_node *node;
>       int ret;
>  
>       do {
> -             ret = ttm_bo_man_get_node(bo, man, placement, mem, &node);
> +             ret = (*man->func->get_node)(man, bo, placement, mem);
>               if (unlikely(ret != 0))
>                       return ret;
> -             if (node)
> +             if (mem->mm_node)
>                       break;
>               spin_lock(&glob->lru_lock);
>               if (list_empty(&man->lru)) {
> @@ -808,9 +764,8 @@ static int ttm_bo_mem_force_space(struct 
> ttm_buffer_object *bo,
>               if (unlikely(ret != 0))
>                       return ret;
>       } while (1);
> -     if (node == NULL)
> +     if (mem->mm_node == NULL)
>               return -ENOMEM;
> -     mem->mm_node = node;
>       mem->mem_type = mem_type;
>       return 0;
>  }
> @@ -884,7 +839,6 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo,
>       bool type_found = false;
>       bool type_ok = false;
>       bool has_erestartsys = false;
> -     struct drm_mm_node *node = NULL;
>       int i, ret;
>  
>       mem->mm_node = NULL;
> @@ -918,17 +872,15 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo,
>  
>               if (man->has_type && man->use_type) {
>                       type_found = true;
> -                     ret = ttm_bo_man_get_node(bo, man, placement, mem,
> -                                                     &node);
> +                     ret = (*man->func->get_node)(man, bo, placement, mem);
>                       if (unlikely(ret))
>                               return ret;
>               }
> -             if (node)
> +             if (mem->mm_node)
>                       break;
>       }
>  
> -     if ((type_ok && (mem_type == TTM_PL_SYSTEM)) || node) {
> -             mem->mm_node = node;
> +     if ((type_ok && (mem_type == TTM_PL_SYSTEM)) || mem->mm_node) {
>               mem->mem_type = mem_type;
>               mem->placement = cur_flags;
>               return 0;
> @@ -998,7 +950,6 @@ int ttm_bo_move_buffer(struct ttm_buffer_object *bo,
>                       bool interruptible, bool no_wait_reserve,
>                       bool no_wait_gpu)
>  {
> -     struct ttm_bo_global *glob = bo->glob;
>       int ret = 0;
>       struct ttm_mem_reg mem;
>  
> @@ -1026,11 +977,8 @@ int ttm_bo_move_buffer(struct ttm_buffer_object *bo,
>               goto out_unlock;
>       ret = ttm_bo_handle_move_mem(bo, &mem, false, interruptible, 
> no_wait_reserve, no_wait_gpu);
>  out_unlock:
> -     if (ret && mem.mm_node) {
> -             spin_lock(&glob->lru_lock);
> -             drm_mm_put_block(mem.mm_node);
> -             spin_unlock(&glob->lru_lock);
> -     }
> +     if (ret && mem.mm_node)
> +             ttm_bo_mem_put(bo, &mem);
>       return ret;
>  }
>  
> @@ -1038,11 +986,10 @@ static int ttm_bo_mem_compat(struct ttm_placement 
> *placement,
>                            struct ttm_mem_reg *mem)
>  {
>       int i;
> -     struct drm_mm_node *node = mem->mm_node;
>  
> -     if (node && placement->lpfn != 0 &&
> -         (node->start < placement->fpfn ||
> -          node->start + node->size > placement->lpfn))
> +     if (mem->mm_node && placement->lpfn != 0 &&
> +         (mem->start < placement->fpfn ||
> +          mem->start + mem->num_pages > placement->lpfn))
>               return -1;
>  
>       for (i = 0; i < placement->num_placement; i++) {
> @@ -1286,7 +1233,6 @@ static int ttm_bo_force_list_clean(struct ttm_bo_device 
> *bdev,
>  
>  int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type)
>  {
> -     struct ttm_bo_global *glob = bdev->glob;
>       struct ttm_mem_type_manager *man;
>       int ret = -EINVAL;
>  
> @@ -1309,13 +1255,7 @@ int ttm_bo_clean_mm(struct ttm_bo_device *bdev, 
> unsigned mem_type)
>       if (mem_type > 0) {
>               ttm_bo_force_list_clean(bdev, mem_type, false);
>  
> -             spin_lock(&glob->lru_lock);
> -             if (drm_mm_clean(&man->manager))
> -                     drm_mm_takedown(&man->manager);
> -             else
> -                     ret = -EBUSY;
> -
> -             spin_unlock(&glob->lru_lock);
> +             ret = (*man->func->takedown)(man);
>       }
>  
>       return ret;
> @@ -1366,6 +1306,7 @@ int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned 
> type,
>       ret = bdev->driver->init_mem_type(bdev, type, man);
>       if (ret)
>               return ret;
> +     man->bdev = bdev;
>  
>       ret = 0;
>       if (type != TTM_PL_SYSTEM) {
> @@ -1375,7 +1316,8 @@ int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned 
> type,
>                              type);
>                       return ret;
>               }
> -             ret = drm_mm_init(&man->manager, 0, p_size);
> +
> +             ret = (*man->func->init)(man, p_size);
>               if (ret)
>                       return ret;
>       }
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_manager.c 
> b/drivers/gpu/drm/ttm/ttm_bo_manager.c
> new file mode 100644
> index 0000000..7410c19
> --- /dev/null
> +++ b/drivers/gpu/drm/ttm/ttm_bo_manager.c
> @@ -0,0 +1,148 @@
> +/**************************************************************************
> + *
> + * Copyright (c) 2007-2009 VMware, Inc., Palo Alto, CA., USA
> + * All Rights Reserved.
> + *
> + * Permission is hereby granted, free of charge, to any person obtaining a
> + * copy of this software and associated documentation files (the
> + * "Software"), to deal in the Software without restriction, including
> + * without limitation the rights to use, copy, modify, merge, publish,
> + * distribute, sub license, and/or sell copies of the Software, and to
> + * permit persons to whom the Software is furnished to do so, subject to
> + * the following conditions:
> + *
> + * The above copyright notice and this permission notice (including the
> + * next paragraph) shall be included in all copies or substantial portions
> + * of the Software.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
> + * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
> + * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY 
> CLAIM,
> + * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
> + * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
> + * USE OR OTHER DEALINGS IN THE SOFTWARE.
> + *
> + **************************************************************************/
> +/*
> + * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com>
> + */
> +
> +#include "ttm/ttm_module.h"
> +#include "ttm/ttm_bo_driver.h"
> +#include "ttm/ttm_placement.h"
> +#include <linux/jiffies.h>
> +#include <linux/slab.h>
> +#include <linux/sched.h>
> +#include <linux/mm.h>
> +#include <linux/file.h>
> +#include <linux/module.h>
> +
> +static int ttm_bo_man_get_node(struct ttm_mem_type_manager *man,
> +                            struct ttm_buffer_object *bo,
> +                            struct ttm_placement *placement,
> +                            struct ttm_mem_reg *mem)
> +{
> +     struct ttm_bo_global *glob = man->bdev->glob;
> +     struct drm_mm *mm = man->priv;
> +     struct drm_mm_node *node = NULL;
> +     unsigned long lpfn;
> +     int ret;
> +
> +     lpfn = placement->lpfn;
> +     if (!lpfn)
> +             lpfn = man->size;
> +     do {
> +             ret = drm_mm_pre_get(mm);
> +             if (unlikely(ret))
> +                     return ret;
> +
> +             spin_lock(&glob->lru_lock);
> +             node = drm_mm_search_free_in_range(mm,
> +                                     mem->num_pages, mem->page_alignment,
> +                                     placement->fpfn, lpfn, 1);
> +             if (unlikely(node == NULL)) {
> +                     spin_unlock(&glob->lru_lock);
> +                     return 0;
> +             }
> +             node = drm_mm_get_block_atomic_range(node, mem->num_pages,
> +                                                     mem->page_alignment,
> +                                                     placement->fpfn,
> +                                                     lpfn);
> +             spin_unlock(&glob->lru_lock);
> +     } while (node == NULL);
> +
> +     mem->mm_node = node;
> +     mem->start = node->start;
> +     return 0;
> +}
> +
> +static void ttm_bo_man_put_node(struct ttm_mem_type_manager *man,
> +                             struct ttm_mem_reg *mem)
> +{
> +     struct ttm_bo_global *glob = man->bdev->glob;
> +
> +     if (mem->mm_node) {
> +             spin_lock(&glob->lru_lock);
> +             drm_mm_put_block(mem->mm_node);
> +             spin_unlock(&glob->lru_lock);
> +             mem->mm_node = NULL;
> +     }
> +}
> +
> +static int ttm_bo_man_init(struct ttm_mem_type_manager *man,
> +                        unsigned long p_size)
> +{
> +     struct drm_mm *mm;
> +     int ret;
> +
> +     mm = kzalloc(sizeof(*mm), GFP_KERNEL);
> +     if (!mm)
> +             return -ENOMEM;
> +
> +     ret = drm_mm_init(mm, 0, p_size);
> +     if (ret) {
> +             kfree(mm);
> +             return ret;
> +     }
> +
> +     man->priv = mm;
> +     return 0;
> +}
> +
> +static int ttm_bo_man_takedown(struct ttm_mem_type_manager *man)
> +{
> +     struct ttm_bo_global *glob = man->bdev->glob;
> +     struct drm_mm *mm = man->priv;
> +     int ret = 0;
> +
> +     spin_lock(&glob->lru_lock);
> +     if (drm_mm_clean(mm)) {
> +             drm_mm_takedown(mm);
> +             kfree(mm);
> +             man->priv = NULL;
> +     } else
> +             ret = -EBUSY;
> +     spin_unlock(&glob->lru_lock);
> +     return ret;
> +}
> +
> +static void ttm_bo_man_debug(struct ttm_mem_type_manager *man,
> +                          const char *prefix)
> +{
> +     struct ttm_bo_global *glob = man->bdev->glob;
> +     struct drm_mm *mm = man->priv;
> +
> +     spin_lock(&glob->lru_lock);
> +     drm_mm_debug_table(mm, prefix);
> +     spin_unlock(&glob->lru_lock);
> +}
> +
> +const struct ttm_mem_type_manager_func ttm_bo_manager_func = {
> +     ttm_bo_man_init,
> +     ttm_bo_man_takedown,
> +     ttm_bo_man_get_node,
> +     ttm_bo_man_put_node,
> +     ttm_bo_man_debug
> +};
> +EXPORT_SYMBOL(ttm_bo_manager_func);
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c 
> b/drivers/gpu/drm/ttm/ttm_bo_util.c
> index 0ebfe0d..c9d2d4d 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
> @@ -256,8 +256,7 @@ int ttm_bo_move_memcpy(struct ttm_buffer_object *bo,
>       dir = 1;
>  
>       if ((old_mem->mem_type == new_mem->mem_type) &&
> -         (new_mem->mm_node->start <
> -          old_mem->mm_node->start + old_mem->mm_node->size)) {
> +         (new_mem->start < old_mem->start + old_mem->size)) {
>               dir = -1;
>               add = new_mem->num_pages - 1;
>       }
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c 
> b/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c
> index c4f5114..1b3bd8c 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c
> @@ -147,6 +147,7 @@ int vmw_init_mem_type(struct ttm_bo_device *bdev, 
> uint32_t type,
>               break;
>       case TTM_PL_VRAM:
>               /* "On-card" video ram */
> +             man->func = &ttm_bo_manager_func;
>               man->gpu_offset = 0;
>               man->flags = TTM_MEMTYPE_FLAG_FIXED | TTM_MEMTYPE_FLAG_MAPPABLE;
>               man->available_caching = TTM_PL_MASK_CACHING;
> @@ -203,7 +204,7 @@ static int vmw_ttm_io_mem_reserve(struct ttm_bo_device 
> *bdev, struct ttm_mem_reg
>               /* System memory */
>               return 0;
>       case TTM_PL_VRAM:
> -             mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> +             mem->bus.offset = mem->start << PAGE_SHIFT;
>               mem->bus.base = dev_priv->vram_start;
>               mem->bus.is_iomem = true;
>               break;
> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h
> index 267a86c..49b43c2 100644
> --- a/include/drm/ttm/ttm_bo_api.h
> +++ b/include/drm/ttm/ttm_bo_api.h
> @@ -102,7 +102,8 @@ struct ttm_bus_placement {
>   */
>  
>  struct ttm_mem_reg {
> -     struct drm_mm_node *mm_node;
> +     void *mm_node;
> +     unsigned long start;
>       unsigned long size;
>       unsigned long num_pages;
>       uint32_t page_alignment;
> diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h
> index 6c694d8..e3371db 100644
> --- a/include/drm/ttm/ttm_bo_driver.h
> +++ b/include/drm/ttm/ttm_bo_driver.h
> @@ -203,7 +203,22 @@ struct ttm_tt {
>   * It's set up by the ttm_bo_driver::init_mem_type method.
>   */
>  
> +struct ttm_mem_type_manager;
> +
> +struct ttm_mem_type_manager_func {
> +     int  (*init)(struct ttm_mem_type_manager *man, unsigned long p_size);
> +     int  (*takedown)(struct ttm_mem_type_manager *man);
> +     int  (*get_node)(struct ttm_mem_type_manager *man,
> +                      struct ttm_buffer_object *bo,
> +                      struct ttm_placement *placement,
> +                      struct ttm_mem_reg *mem);
> +     void (*put_node)(struct ttm_mem_type_manager *man,
> +                      struct ttm_mem_reg *mem);
> +     void (*debug)(struct ttm_mem_type_manager *man, const char *prefix);
> +};
> +
>  struct ttm_mem_type_manager {
> +     struct ttm_bo_device *bdev;
>  
>       /*
>        * No protection. Constant from start.
> @@ -222,8 +237,8 @@ struct ttm_mem_type_manager {
>        * TODO: Consider one lru_lock per ttm_mem_type_manager.
>        * Plays ill with list removal, though.
>        */
> -
> -     struct drm_mm manager;
> +     const struct ttm_mem_type_manager_func *func;
> +     void *priv;
>       struct list_head lru;
>  };
>  
> @@ -895,6 +910,8 @@ extern int ttm_bo_move_accel_cleanup(struct 
> ttm_buffer_object *bo,
>   */
>  extern pgprot_t ttm_io_prot(uint32_t caching_flags, pgprot_t tmp);
>  
> +extern const struct ttm_mem_type_manager_func ttm_bo_manager_func;
> +
>  #if (defined(CONFIG_AGP) || (defined(CONFIG_AGP_MODULE) && defined(MODULE)))
>  #define TTM_HAS_AGP
>  #include <linux/agp_backend.h>

_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
http://lists.freedesktop.org/mailman/listinfo/dri-devel

Reply via email to