[PATCH 2/2] drm/ttm: restructure to allow driver to plug in alternate memory manager
Jerome Glisse
glisse at freedesktop.org
Mon Aug 9 12:24:39 PDT 2010
On 08/05/2010 09:09 AM, Ben Skeggs wrote:
> From: Ben Skeggs <bskeggs at redhat.com>
>
> Nouveau will need this on GeForce 8 and up to account for the GPU
> reordering physical VRAM for some memory types.
>
> Signed-off-by: Ben Skeggs <bskeggs at redhat.com>
Reviewed-by: Jerome Glisse <jglisse at redhat.com>
> ---
> drivers/gpu/drm/nouveau/nouveau_bo.c | 12 ++-
> drivers/gpu/drm/nouveau/nouveau_channel.c | 6 +-
> drivers/gpu/drm/nouveau/nouveau_notifier.c | 2 +-
> drivers/gpu/drm/nouveau/nouveau_sgdma.c | 4 +-
> drivers/gpu/drm/nouveau/nv50_crtc.c | 3 +-
> drivers/gpu/drm/nouveau/nv50_display.c | 2 +-
> drivers/gpu/drm/nouveau/nv50_instmem.c | 2 +-
> drivers/gpu/drm/radeon/radeon_object.c | 6 +-
> drivers/gpu/drm/radeon/radeon_ttm.c | 16 ++--
> drivers/gpu/drm/ttm/Makefile | 3 +-
> drivers/gpu/drm/ttm/ttm_agp_backend.c | 3 +-
> drivers/gpu/drm/ttm/ttm_bo.c | 100 ++++---------------
> drivers/gpu/drm/ttm/ttm_bo_manager.c | 148 ++++++++++++++++++++++++++++
> drivers/gpu/drm/ttm/ttm_bo_util.c | 3 +-
> drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c | 3 +-
> include/drm/ttm/ttm_bo_api.h | 3 +-
> include/drm/ttm/ttm_bo_driver.h | 21 ++++-
> 17 files changed, 225 insertions(+), 112 deletions(-)
> create mode 100644 drivers/gpu/drm/ttm/ttm_bo_manager.c
>
> diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c b/drivers/gpu/drm/nouveau/nouveau_bo.c
> index 92290fa..f403737 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_bo.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_bo.c
> @@ -390,6 +390,7 @@ nouveau_bo_init_mem_type(struct ttm_bo_device *bdev, uint32_t type,
> man->default_caching = TTM_PL_FLAG_CACHED;
> break;
> case TTM_PL_VRAM:
> + man->func = &ttm_bo_manager_func;
> man->flags = TTM_MEMTYPE_FLAG_FIXED |
> TTM_MEMTYPE_FLAG_MAPPABLE;
> man->available_caching = TTM_PL_FLAG_UNCACHED |
> @@ -398,6 +399,7 @@ nouveau_bo_init_mem_type(struct ttm_bo_device *bdev, uint32_t type,
> man->gpu_offset = dev_priv->vm_vram_base;
> break;
> case TTM_PL_TT:
> + man->func = &ttm_bo_manager_func;
> switch (dev_priv->gart_info.type) {
> case NOUVEAU_GART_AGP:
> man->flags = TTM_MEMTYPE_FLAG_MAPPABLE;
> @@ -500,8 +502,8 @@ nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, bool intr,
> if (!chan || nvbo->tile_flags || nvbo->no_vm)
> chan = dev_priv->channel;
>
> - src_offset = old_mem->mm_node->start << PAGE_SHIFT;
> - dst_offset = new_mem->mm_node->start << PAGE_SHIFT;
> + src_offset = old_mem->start << PAGE_SHIFT;
> + dst_offset = new_mem->start << PAGE_SHIFT;
> if (chan != dev_priv->channel) {
> if (old_mem->mem_type == TTM_PL_TT)
> src_offset += dev_priv->vm_gart_base;
> @@ -650,7 +652,7 @@ nouveau_bo_vm_bind(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem,
> return 0;
> }
>
> - offset = new_mem->mm_node->start << PAGE_SHIFT;
> + offset = new_mem->start << PAGE_SHIFT;
>
> if (dev_priv->card_type == NV_50) {
> ret = nv50_mem_vm_bind_linear(dev,
> @@ -764,14 +766,14 @@ nouveau_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem)
> case TTM_PL_TT:
> #if __OS_HAS_AGP
> if (dev_priv->gart_info.type == NOUVEAU_GART_AGP) {
> - mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> + mem->bus.offset = mem->start << PAGE_SHIFT;
> mem->bus.base = dev_priv->gart_info.aper_base;
> mem->bus.is_iomem = true;
> }
> #endif
> break;
> case TTM_PL_VRAM:
> - mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> + mem->bus.offset = mem->start << PAGE_SHIFT;
> mem->bus.base = pci_resource_start(dev->pdev, 1);
> mem->bus.is_iomem = true;
> break;
> diff --git a/drivers/gpu/drm/nouveau/nouveau_channel.c b/drivers/gpu/drm/nouveau/nouveau_channel.c
> index 90fdcda..90d3450 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_channel.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_channel.c
> @@ -48,14 +48,14 @@ nouveau_channel_pushbuf_ctxdma_init(struct nouveau_channel *chan)
> dev_priv->gart_info.aper_size,
> NV_DMA_ACCESS_RO, &pushbuf,
> NULL);
> - chan->pushbuf_base = pb->bo.mem.mm_node->start << PAGE_SHIFT;
> + chan->pushbuf_base = pb->bo.mem.start << PAGE_SHIFT;
> } else
> if (dev_priv->card_type != NV_04) {
> ret = nouveau_gpuobj_dma_new(chan, NV_CLASS_DMA_IN_MEMORY, 0,
> dev_priv->fb_available_size,
> NV_DMA_ACCESS_RO,
> NV_DMA_TARGET_VIDMEM, &pushbuf);
> - chan->pushbuf_base = pb->bo.mem.mm_node->start << PAGE_SHIFT;
> + chan->pushbuf_base = pb->bo.mem.start << PAGE_SHIFT;
> } else {
> /* NV04 cmdbuf hack, from original ddx.. not sure of it's
> * exact reason for existing :) PCI access to cmdbuf in
> @@ -67,7 +67,7 @@ nouveau_channel_pushbuf_ctxdma_init(struct nouveau_channel *chan)
> dev_priv->fb_available_size,
> NV_DMA_ACCESS_RO,
> NV_DMA_TARGET_PCI, &pushbuf);
> - chan->pushbuf_base = pb->bo.mem.mm_node->start << PAGE_SHIFT;
> + chan->pushbuf_base = pb->bo.mem.start << PAGE_SHIFT;
> }
>
> ret = nouveau_gpuobj_ref_add(dev, chan, 0, pushbuf, &chan->pushbuf);
> diff --git a/drivers/gpu/drm/nouveau/nouveau_notifier.c b/drivers/gpu/drm/nouveau/nouveau_notifier.c
> index 3ec181f..d218427 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_notifier.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_notifier.c
> @@ -112,7 +112,7 @@ nouveau_notifier_alloc(struct nouveau_channel *chan, uint32_t handle,
> return -ENOMEM;
> }
>
> - offset = chan->notifier_bo->bo.mem.mm_node->start << PAGE_SHIFT;
> + offset = chan->notifier_bo->bo.mem.start << PAGE_SHIFT;
> if (chan->notifier_bo->bo.mem.mem_type == TTM_PL_VRAM) {
> target = NV_DMA_TARGET_VIDMEM;
> } else
> diff --git a/drivers/gpu/drm/nouveau/nouveau_sgdma.c b/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> index 491767f..4374366 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> @@ -95,9 +95,9 @@ nouveau_sgdma_bind(struct ttm_backend *be, struct ttm_mem_reg *mem)
> struct nouveau_gpuobj *gpuobj = dev_priv->gart_info.sg_ctxdma;
> unsigned i, j, pte;
>
> - NV_DEBUG(dev, "pg=0x%lx\n", mem->mm_node->start);
> + NV_DEBUG(dev, "pg=0x%lx\n", mem->start);
>
> - pte = nouveau_sgdma_pte(nvbe->dev, mem->mm_node->start << PAGE_SHIFT);
> + pte = nouveau_sgdma_pte(nvbe->dev, mem->start << PAGE_SHIFT);
> nvbe->pte_start = pte;
> for (i = 0; i < nvbe->nr_pages; i++) {
> dma_addr_t dma_offset = nvbe->pages[i];
> diff --git a/drivers/gpu/drm/nouveau/nv50_crtc.c b/drivers/gpu/drm/nouveau/nv50_crtc.c
> index 5d11ea1..a6a0f95 100644
> --- a/drivers/gpu/drm/nouveau/nv50_crtc.c
> +++ b/drivers/gpu/drm/nouveau/nv50_crtc.c
> @@ -104,8 +104,7 @@ nv50_crtc_blank(struct nouveau_crtc *nv_crtc, bool blanked)
> OUT_RING(evo, nv_crtc->lut.depth == 8 ?
> NV50_EVO_CRTC_CLUT_MODE_OFF :
> NV50_EVO_CRTC_CLUT_MODE_ON);
> - OUT_RING(evo, (nv_crtc->lut.nvbo->bo.mem.mm_node->start <<
> - PAGE_SHIFT) >> 8);
> + OUT_RING(evo, (nv_crtc->lut.nvbo->bo.mem.start << PAGE_SHIFT) >> 8);
> if (dev_priv->chipset != 0x50) {
> BEGIN_RING(evo, 0, NV84_EVO_CRTC(index, CLUT_DMA), 1);
> OUT_RING(evo, NvEvoVRAM);
> diff --git a/drivers/gpu/drm/nouveau/nv50_display.c b/drivers/gpu/drm/nouveau/nv50_display.c
> index f13ad0d..5f88e93 100644
> --- a/drivers/gpu/drm/nouveau/nv50_display.c
> +++ b/drivers/gpu/drm/nouveau/nv50_display.c
> @@ -322,7 +322,7 @@ nv50_display_init(struct drm_device *dev)
>
> /* initialise fifo */
> nv_wr32(dev, NV50_PDISPLAY_CHANNEL_DMA_CB(0),
> - ((evo->pushbuf_bo->bo.mem.mm_node->start << PAGE_SHIFT) >> 8) |
> + ((evo->pushbuf_bo->bo.mem.start << PAGE_SHIFT) >> 8) |
> NV50_PDISPLAY_CHANNEL_DMA_CB_LOCATION_VRAM |
> NV50_PDISPLAY_CHANNEL_DMA_CB_VALID);
> nv_wr32(dev, NV50_PDISPLAY_CHANNEL_UNK2(0), 0x00010000);
> diff --git a/drivers/gpu/drm/nouveau/nv50_instmem.c b/drivers/gpu/drm/nouveau/nv50_instmem.c
> index 37c7b48..a35e04c 100644
> --- a/drivers/gpu/drm/nouveau/nv50_instmem.c
> +++ b/drivers/gpu/drm/nouveau/nv50_instmem.c
> @@ -397,7 +397,7 @@ nv50_instmem_populate(struct drm_device *dev, struct nouveau_gpuobj *gpuobj,
> return ret;
> }
>
> - gpuobj->im_backing_start = gpuobj->im_backing->bo.mem.mm_node->start;
> + gpuobj->im_backing_start = gpuobj->im_backing->bo.mem.start;
> gpuobj->im_backing_start <<= PAGE_SHIFT;
>
> return 0;
> diff --git a/drivers/gpu/drm/radeon/radeon_object.c b/drivers/gpu/drm/radeon/radeon_object.c
> index 0afd1e6..c261060 100644
> --- a/drivers/gpu/drm/radeon/radeon_object.c
> +++ b/drivers/gpu/drm/radeon/radeon_object.c
> @@ -435,7 +435,7 @@ int radeon_bo_get_surface_reg(struct radeon_bo *bo)
>
> out:
> radeon_set_surface_reg(rdev, i, bo->tiling_flags, bo->pitch,
> - bo->tbo.mem.mm_node->start << PAGE_SHIFT,
> + bo->tbo.mem.start << PAGE_SHIFT,
> bo->tbo.num_pages << PAGE_SHIFT);
> return 0;
> }
> @@ -532,7 +532,7 @@ int radeon_bo_fault_reserve_notify(struct ttm_buffer_object *bo)
> rdev = rbo->rdev;
> if (bo->mem.mem_type == TTM_PL_VRAM) {
> size = bo->mem.num_pages << PAGE_SHIFT;
> - offset = bo->mem.mm_node->start << PAGE_SHIFT;
> + offset = bo->mem.start << PAGE_SHIFT;
> if ((offset + size) > rdev->mc.visible_vram_size) {
> /* hurrah the memory is not visible ! */
> radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_VRAM);
> @@ -540,7 +540,7 @@ int radeon_bo_fault_reserve_notify(struct ttm_buffer_object *bo)
> r = ttm_bo_validate(bo, &rbo->placement, false, true, false);
> if (unlikely(r != 0))
> return r;
> - offset = bo->mem.mm_node->start << PAGE_SHIFT;
> + offset = bo->mem.start << PAGE_SHIFT;
> /* this should not happen */
> if ((offset + size) > rdev->mc.visible_vram_size)
> return -EINVAL;
> diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c b/drivers/gpu/drm/radeon/radeon_ttm.c
> index cc19aba..0921910 100644
> --- a/drivers/gpu/drm/radeon/radeon_ttm.c
> +++ b/drivers/gpu/drm/radeon/radeon_ttm.c
> @@ -152,6 +152,7 @@ static int radeon_init_mem_type(struct ttm_bo_device *bdev, uint32_t type,
> man->default_caching = TTM_PL_FLAG_CACHED;
> break;
> case TTM_PL_TT:
> + man->func = &ttm_bo_manager_func;
> man->gpu_offset = rdev->mc.gtt_start;
> man->available_caching = TTM_PL_MASK_CACHING;
> man->default_caching = TTM_PL_FLAG_CACHED;
> @@ -173,6 +174,7 @@ static int radeon_init_mem_type(struct ttm_bo_device *bdev, uint32_t type,
> break;
> case TTM_PL_VRAM:
> /* "On-card" video ram */
> + man->func = &ttm_bo_manager_func;
> man->gpu_offset = rdev->mc.vram_start;
> man->flags = TTM_MEMTYPE_FLAG_FIXED |
> TTM_MEMTYPE_FLAG_MAPPABLE;
> @@ -246,8 +248,8 @@ static int radeon_move_blit(struct ttm_buffer_object *bo,
> if (unlikely(r)) {
> return r;
> }
> - old_start = old_mem->mm_node->start << PAGE_SHIFT;
> - new_start = new_mem->mm_node->start << PAGE_SHIFT;
> + old_start = old_mem->start << PAGE_SHIFT;
> + new_start = new_mem->start << PAGE_SHIFT;
>
> switch (old_mem->mem_type) {
> case TTM_PL_VRAM:
> @@ -435,14 +437,14 @@ static int radeon_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_
> #if __OS_HAS_AGP
> if (rdev->flags & RADEON_IS_AGP) {
> /* RADEON_IS_AGP is set only if AGP is active */
> - mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> + mem->bus.offset = mem->start << PAGE_SHIFT;
> mem->bus.base = rdev->mc.agp_base;
> mem->bus.is_iomem = !rdev->ddev->agp->cant_use_aperture;
> }
> #endif
> break;
> case TTM_PL_VRAM:
> - mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> + mem->bus.offset = mem->start << PAGE_SHIFT;
> /* check if it's visible */
> if ((mem->bus.offset + mem->bus.size) > rdev->mc.visible_vram_size)
> return -EINVAL;
> @@ -685,7 +687,7 @@ static int radeon_ttm_backend_bind(struct ttm_backend *backend,
> int r;
>
> gtt = container_of(backend, struct radeon_ttm_backend, backend);
> - gtt->offset = bo_mem->mm_node->start << PAGE_SHIFT;
> + gtt->offset = bo_mem->start << PAGE_SHIFT;
> if (!gtt->num_pages) {
> WARN(1, "nothing to bind %lu pages for mreg %p back %p!\n", gtt->num_pages, bo_mem, backend);
> }
> @@ -784,9 +786,9 @@ static int radeon_ttm_debugfs_init(struct radeon_device *rdev)
> radeon_mem_types_list[i].show = &radeon_mm_dump_table;
> radeon_mem_types_list[i].driver_features = 0;
> if (i == 0)
> - radeon_mem_types_list[i].data = &rdev->mman.bdev.man[TTM_PL_VRAM].manager;
> + radeon_mem_types_list[i].data = &rdev->mman.bdev.man[TTM_PL_VRAM].priv;
> else
> - radeon_mem_types_list[i].data = &rdev->mman.bdev.man[TTM_PL_TT].manager;
> + radeon_mem_types_list[i].data = &rdev->mman.bdev.man[TTM_PL_TT].priv;
>
> }
> /* Add ttm page pool to debugfs */
> diff --git a/drivers/gpu/drm/ttm/Makefile b/drivers/gpu/drm/ttm/Makefile
> index b256d4a..f3cf6f0 100644
> --- a/drivers/gpu/drm/ttm/Makefile
> +++ b/drivers/gpu/drm/ttm/Makefile
> @@ -4,6 +4,7 @@
> ccflags-y := -Iinclude/drm
> ttm-y := ttm_agp_backend.o ttm_memory.o ttm_tt.o ttm_bo.o \
> ttm_bo_util.o ttm_bo_vm.o ttm_module.o \
> - ttm_object.o ttm_lock.o ttm_execbuf_util.o ttm_page_alloc.o
> + ttm_object.o ttm_lock.o ttm_execbuf_util.o ttm_page_alloc.o \
> + ttm_bo_manager.o
>
> obj-$(CONFIG_DRM_TTM) += ttm.o
> diff --git a/drivers/gpu/drm/ttm/ttm_agp_backend.c b/drivers/gpu/drm/ttm/ttm_agp_backend.c
> index 4bf69c4..f999e36 100644
> --- a/drivers/gpu/drm/ttm/ttm_agp_backend.c
> +++ b/drivers/gpu/drm/ttm/ttm_agp_backend.c
> @@ -74,6 +74,7 @@ static int ttm_agp_bind(struct ttm_backend *backend, struct ttm_mem_reg *bo_mem)
> {
> struct ttm_agp_backend *agp_be =
> container_of(backend, struct ttm_agp_backend, backend);
> + struct drm_mm_node *node = bo_mem->mm_node;
> struct agp_memory *mem = agp_be->mem;
> int cached = (bo_mem->placement & TTM_PL_FLAG_CACHED);
> int ret;
> @@ -81,7 +82,7 @@ static int ttm_agp_bind(struct ttm_backend *backend, struct ttm_mem_reg *bo_mem)
> mem->is_flushed = 1;
> mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY;
>
> - ret = agp_bind_memory(mem, bo_mem->mm_node->start);
> + ret = agp_bind_memory(mem, node->start);
> if (ret)
> printk(KERN_ERR TTM_PFX "AGP Bind memory failed.\n");
>
> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
> index 80d37b4..af7b57a 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo.c
> @@ -84,11 +84,8 @@ static void ttm_mem_type_debug(struct ttm_bo_device *bdev, int mem_type)
> man->available_caching);
> printk(KERN_ERR TTM_PFX " default_caching: 0x%08X\n",
> man->default_caching);
> - if (mem_type != TTM_PL_SYSTEM) {
> - spin_lock(&bdev->glob->lru_lock);
> - drm_mm_debug_table(&man->manager, TTM_PFX);
> - spin_unlock(&bdev->glob->lru_lock);
> - }
> + if (mem_type != TTM_PL_SYSTEM)
> + (*man->func->debug)(man, TTM_PFX);
> }
>
> static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo,
> @@ -421,7 +418,7 @@ moved:
>
> if (bo->mem.mm_node) {
> spin_lock(&bo->lock);
> - bo->offset = (bo->mem.mm_node->start << PAGE_SHIFT) +
> + bo->offset = (bo->mem.start << PAGE_SHIFT) +
> bdev->man[bo->mem.mem_type].gpu_offset;
> bo->cur_placement = bo->mem.placement;
> spin_unlock(&bo->lock);
> @@ -724,52 +721,12 @@ retry:
> return ret;
> }
>
> -static int ttm_bo_man_get_node(struct ttm_buffer_object *bo,
> - struct ttm_mem_type_manager *man,
> - struct ttm_placement *placement,
> - struct ttm_mem_reg *mem,
> - struct drm_mm_node **node)
> -{
> - struct ttm_bo_global *glob = bo->glob;
> - unsigned long lpfn;
> - int ret;
> -
> - lpfn = placement->lpfn;
> - if (!lpfn)
> - lpfn = man->size;
> - *node = NULL;
> - do {
> - ret = drm_mm_pre_get(&man->manager);
> - if (unlikely(ret))
> - return ret;
> -
> - spin_lock(&glob->lru_lock);
> - *node = drm_mm_search_free_in_range(&man->manager,
> - mem->num_pages, mem->page_alignment,
> - placement->fpfn, lpfn, 1);
> - if (unlikely(*node == NULL)) {
> - spin_unlock(&glob->lru_lock);
> - return 0;
> - }
> - *node = drm_mm_get_block_atomic_range(*node, mem->num_pages,
> - mem->page_alignment,
> - placement->fpfn,
> - lpfn);
> - spin_unlock(&glob->lru_lock);
> - } while (*node == NULL);
> - return 0;
> -}
> -
> void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem)
> {
> - struct ttm_bo_global *glob = bo->glob;
> + struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type];
>
> - if (mem->mm_node) {
> - spin_lock(&glob->lru_lock);
> - drm_mm_put_block(mem->mm_node);
> - spin_unlock(&glob->lru_lock);
> - mem->mm_node = NULL;
> - }
> + if (mem->mm_node)
> + (*man->func->put_node)(man, mem);
> }
> EXPORT_SYMBOL(ttm_bo_mem_put);
>
> @@ -788,14 +745,13 @@ static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo,
> struct ttm_bo_device *bdev = bo->bdev;
> struct ttm_bo_global *glob = bdev->glob;
> struct ttm_mem_type_manager *man = &bdev->man[mem_type];
> - struct drm_mm_node *node;
> int ret;
>
> do {
> - ret = ttm_bo_man_get_node(bo, man, placement, mem, &node);
> + ret = (*man->func->get_node)(man, bo, placement, mem);
> if (unlikely(ret != 0))
> return ret;
> - if (node)
> + if (mem->mm_node)
> break;
> spin_lock(&glob->lru_lock);
> if (list_empty(&man->lru)) {
> @@ -808,9 +764,8 @@ static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo,
> if (unlikely(ret != 0))
> return ret;
> } while (1);
> - if (node == NULL)
> + if (mem->mm_node == NULL)
> return -ENOMEM;
> - mem->mm_node = node;
> mem->mem_type = mem_type;
> return 0;
> }
> @@ -884,7 +839,6 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo,
> bool type_found = false;
> bool type_ok = false;
> bool has_erestartsys = false;
> - struct drm_mm_node *node = NULL;
> int i, ret;
>
> mem->mm_node = NULL;
> @@ -918,17 +872,15 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo,
>
> if (man->has_type && man->use_type) {
> type_found = true;
> - ret = ttm_bo_man_get_node(bo, man, placement, mem,
> - &node);
> + ret = (*man->func->get_node)(man, bo, placement, mem);
> if (unlikely(ret))
> return ret;
> }
> - if (node)
> + if (mem->mm_node)
> break;
> }
>
> - if ((type_ok && (mem_type == TTM_PL_SYSTEM)) || node) {
> - mem->mm_node = node;
> + if ((type_ok && (mem_type == TTM_PL_SYSTEM)) || mem->mm_node) {
> mem->mem_type = mem_type;
> mem->placement = cur_flags;
> return 0;
> @@ -998,7 +950,6 @@ int ttm_bo_move_buffer(struct ttm_buffer_object *bo,
> bool interruptible, bool no_wait_reserve,
> bool no_wait_gpu)
> {
> - struct ttm_bo_global *glob = bo->glob;
> int ret = 0;
> struct ttm_mem_reg mem;
>
> @@ -1026,11 +977,8 @@ int ttm_bo_move_buffer(struct ttm_buffer_object *bo,
> goto out_unlock;
> ret = ttm_bo_handle_move_mem(bo, &mem, false, interruptible, no_wait_reserve, no_wait_gpu);
> out_unlock:
> - if (ret && mem.mm_node) {
> - spin_lock(&glob->lru_lock);
> - drm_mm_put_block(mem.mm_node);
> - spin_unlock(&glob->lru_lock);
> - }
> + if (ret && mem.mm_node)
> + ttm_bo_mem_put(bo, &mem);
> return ret;
> }
>
> @@ -1038,11 +986,10 @@ static int ttm_bo_mem_compat(struct ttm_placement *placement,
> struct ttm_mem_reg *mem)
> {
> int i;
> - struct drm_mm_node *node = mem->mm_node;
>
> - if (node && placement->lpfn != 0 &&
> - (node->start < placement->fpfn ||
> - node->start + node->size > placement->lpfn))
> + if (mem->mm_node && placement->lpfn != 0 &&
> + (mem->start < placement->fpfn ||
> + mem->start + mem->num_pages > placement->lpfn))
> return -1;
>
> for (i = 0; i < placement->num_placement; i++) {
> @@ -1286,7 +1233,6 @@ static int ttm_bo_force_list_clean(struct ttm_bo_device *bdev,
>
> int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type)
> {
> - struct ttm_bo_global *glob = bdev->glob;
> struct ttm_mem_type_manager *man;
> int ret = -EINVAL;
>
> @@ -1309,13 +1255,7 @@ int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type)
> if (mem_type > 0) {
> ttm_bo_force_list_clean(bdev, mem_type, false);
>
> - spin_lock(&glob->lru_lock);
> - if (drm_mm_clean(&man->manager))
> - drm_mm_takedown(&man->manager);
> - else
> - ret = -EBUSY;
> -
> - spin_unlock(&glob->lru_lock);
> + ret = (*man->func->takedown)(man);
> }
>
> return ret;
> @@ -1366,6 +1306,7 @@ int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned type,
> ret = bdev->driver->init_mem_type(bdev, type, man);
> if (ret)
> return ret;
> + man->bdev = bdev;
>
> ret = 0;
> if (type != TTM_PL_SYSTEM) {
> @@ -1375,7 +1316,8 @@ int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned type,
> type);
> return ret;
> }
> - ret = drm_mm_init(&man->manager, 0, p_size);
> +
> + ret = (*man->func->init)(man, p_size);
> if (ret)
> return ret;
> }
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_manager.c b/drivers/gpu/drm/ttm/ttm_bo_manager.c
> new file mode 100644
> index 0000000..7410c19
> --- /dev/null
> +++ b/drivers/gpu/drm/ttm/ttm_bo_manager.c
> @@ -0,0 +1,148 @@
> +/**************************************************************************
> + *
> + * Copyright (c) 2007-2009 VMware, Inc., Palo Alto, CA., USA
> + * All Rights Reserved.
> + *
> + * Permission is hereby granted, free of charge, to any person obtaining a
> + * copy of this software and associated documentation files (the
> + * "Software"), to deal in the Software without restriction, including
> + * without limitation the rights to use, copy, modify, merge, publish,
> + * distribute, sub license, and/or sell copies of the Software, and to
> + * permit persons to whom the Software is furnished to do so, subject to
> + * the following conditions:
> + *
> + * The above copyright notice and this permission notice (including the
> + * next paragraph) shall be included in all copies or substantial portions
> + * of the Software.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
> + * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
> + * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
> + * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
> + * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
> + * USE OR OTHER DEALINGS IN THE SOFTWARE.
> + *
> + **************************************************************************/
> +/*
> + * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com>
> + */
> +
> +#include "ttm/ttm_module.h"
> +#include "ttm/ttm_bo_driver.h"
> +#include "ttm/ttm_placement.h"
> +#include <linux/jiffies.h>
> +#include <linux/slab.h>
> +#include <linux/sched.h>
> +#include <linux/mm.h>
> +#include <linux/file.h>
> +#include <linux/module.h>
> +
> +static int ttm_bo_man_get_node(struct ttm_mem_type_manager *man,
> + struct ttm_buffer_object *bo,
> + struct ttm_placement *placement,
> + struct ttm_mem_reg *mem)
> +{
> + struct ttm_bo_global *glob = man->bdev->glob;
> + struct drm_mm *mm = man->priv;
> + struct drm_mm_node *node = NULL;
> + unsigned long lpfn;
> + int ret;
> +
> + lpfn = placement->lpfn;
> + if (!lpfn)
> + lpfn = man->size;
> + do {
> + ret = drm_mm_pre_get(mm);
> + if (unlikely(ret))
> + return ret;
> +
> + spin_lock(&glob->lru_lock);
> + node = drm_mm_search_free_in_range(mm,
> + mem->num_pages, mem->page_alignment,
> + placement->fpfn, lpfn, 1);
> + if (unlikely(node == NULL)) {
> + spin_unlock(&glob->lru_lock);
> + return 0;
> + }
> + node = drm_mm_get_block_atomic_range(node, mem->num_pages,
> + mem->page_alignment,
> + placement->fpfn,
> + lpfn);
> + spin_unlock(&glob->lru_lock);
> + } while (node == NULL);
> +
> + mem->mm_node = node;
> + mem->start = node->start;
> + return 0;
> +}
> +
> +static void ttm_bo_man_put_node(struct ttm_mem_type_manager *man,
> + struct ttm_mem_reg *mem)
> +{
> + struct ttm_bo_global *glob = man->bdev->glob;
> +
> + if (mem->mm_node) {
> + spin_lock(&glob->lru_lock);
> + drm_mm_put_block(mem->mm_node);
> + spin_unlock(&glob->lru_lock);
> + mem->mm_node = NULL;
> + }
> +}
> +
> +static int ttm_bo_man_init(struct ttm_mem_type_manager *man,
> + unsigned long p_size)
> +{
> + struct drm_mm *mm;
> + int ret;
> +
> + mm = kzalloc(sizeof(*mm), GFP_KERNEL);
> + if (!mm)
> + return -ENOMEM;
> +
> + ret = drm_mm_init(mm, 0, p_size);
> + if (ret) {
> + kfree(mm);
> + return ret;
> + }
> +
> + man->priv = mm;
> + return 0;
> +}
> +
> +static int ttm_bo_man_takedown(struct ttm_mem_type_manager *man)
> +{
> + struct ttm_bo_global *glob = man->bdev->glob;
> + struct drm_mm *mm = man->priv;
> + int ret = 0;
> +
> + spin_lock(&glob->lru_lock);
> + if (drm_mm_clean(mm)) {
> + drm_mm_takedown(mm);
> + kfree(mm);
> + man->priv = NULL;
> + } else
> + ret = -EBUSY;
> + spin_unlock(&glob->lru_lock);
> + return ret;
> +}
> +
> +static void ttm_bo_man_debug(struct ttm_mem_type_manager *man,
> + const char *prefix)
> +{
> + struct ttm_bo_global *glob = man->bdev->glob;
> + struct drm_mm *mm = man->priv;
> +
> + spin_lock(&glob->lru_lock);
> + drm_mm_debug_table(mm, prefix);
> + spin_unlock(&glob->lru_lock);
> +}
> +
> +const struct ttm_mem_type_manager_func ttm_bo_manager_func = {
> + ttm_bo_man_init,
> + ttm_bo_man_takedown,
> + ttm_bo_man_get_node,
> + ttm_bo_man_put_node,
> + ttm_bo_man_debug
> +};
> +EXPORT_SYMBOL(ttm_bo_manager_func);
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c
> index 0ebfe0d..c9d2d4d 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
> @@ -256,8 +256,7 @@ int ttm_bo_move_memcpy(struct ttm_buffer_object *bo,
> dir = 1;
>
> if ((old_mem->mem_type == new_mem->mem_type) &&
> - (new_mem->mm_node->start <
> - old_mem->mm_node->start + old_mem->mm_node->size)) {
> + (new_mem->start < old_mem->start + old_mem->size)) {
> dir = -1;
> add = new_mem->num_pages - 1;
> }
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c b/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c
> index c4f5114..1b3bd8c 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_buffer.c
> @@ -147,6 +147,7 @@ int vmw_init_mem_type(struct ttm_bo_device *bdev, uint32_t type,
> break;
> case TTM_PL_VRAM:
> /* "On-card" video ram */
> + man->func = &ttm_bo_manager_func;
> man->gpu_offset = 0;
> man->flags = TTM_MEMTYPE_FLAG_FIXED | TTM_MEMTYPE_FLAG_MAPPABLE;
> man->available_caching = TTM_PL_MASK_CACHING;
> @@ -203,7 +204,7 @@ static int vmw_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg
> /* System memory */
> return 0;
> case TTM_PL_VRAM:
> - mem->bus.offset = mem->mm_node->start << PAGE_SHIFT;
> + mem->bus.offset = mem->start << PAGE_SHIFT;
> mem->bus.base = dev_priv->vram_start;
> mem->bus.is_iomem = true;
> break;
> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h
> index 267a86c..49b43c2 100644
> --- a/include/drm/ttm/ttm_bo_api.h
> +++ b/include/drm/ttm/ttm_bo_api.h
> @@ -102,7 +102,8 @@ struct ttm_bus_placement {
> */
>
> struct ttm_mem_reg {
> - struct drm_mm_node *mm_node;
> + void *mm_node;
> + unsigned long start;
> unsigned long size;
> unsigned long num_pages;
> uint32_t page_alignment;
> diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h
> index 6c694d8..e3371db 100644
> --- a/include/drm/ttm/ttm_bo_driver.h
> +++ b/include/drm/ttm/ttm_bo_driver.h
> @@ -203,7 +203,22 @@ struct ttm_tt {
> * It's set up by the ttm_bo_driver::init_mem_type method.
> */
>
> +struct ttm_mem_type_manager;
> +
> +struct ttm_mem_type_manager_func {
> + int (*init)(struct ttm_mem_type_manager *man, unsigned long p_size);
> + int (*takedown)(struct ttm_mem_type_manager *man);
> + int (*get_node)(struct ttm_mem_type_manager *man,
> + struct ttm_buffer_object *bo,
> + struct ttm_placement *placement,
> + struct ttm_mem_reg *mem);
> + void (*put_node)(struct ttm_mem_type_manager *man,
> + struct ttm_mem_reg *mem);
> + void (*debug)(struct ttm_mem_type_manager *man, const char *prefix);
> +};
> +
> struct ttm_mem_type_manager {
> + struct ttm_bo_device *bdev;
>
> /*
> * No protection. Constant from start.
> @@ -222,8 +237,8 @@ struct ttm_mem_type_manager {
> * TODO: Consider one lru_lock per ttm_mem_type_manager.
> * Plays ill with list removal, though.
> */
> -
> - struct drm_mm manager;
> + const struct ttm_mem_type_manager_func *func;
> + void *priv;
> struct list_head lru;
> };
>
> @@ -895,6 +910,8 @@ extern int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
> */
> extern pgprot_t ttm_io_prot(uint32_t caching_flags, pgprot_t tmp);
>
> +extern const struct ttm_mem_type_manager_func ttm_bo_manager_func;
> +
> #if (defined(CONFIG_AGP) || (defined(CONFIG_AGP_MODULE) && defined(MODULE)))
> #define TTM_HAS_AGP
> #include <linux/agp_backend.h>
More information about the dri-devel
mailing list