[PATCH v6 5/5] drm/amdgpu: track bo memory stats at runtime
Li, Yunxiang (Teddy)
Yunxiang.Li at amd.com
Thu Oct 31 13:48:14 UTC 2024
[Public]
> From: Christian König <ckoenig.leichtzumerken at gmail.com>
> Sent: Thursday, October 31, 2024 8:54
> Am 25.10.24 um 19:41 schrieb Yunxiang Li:
> > Before, every time fdinfo is queried we try to lock all the BOs in the
> > VM and calculate memory usage from scratch. This works okay if the
> > fdinfo is rarely read and the VMs don't have a ton of BOs. If either
> > of these conditions is not true, we get a massive performance hit.
> >
> > In this new revision, we track the BOs as they change states. This way
> > when the fdinfo is queried we only need to take the status lock and
> > copy out the usage stats with minimal impact to the runtime performance.
> >
> > Signed-off-by: Yunxiang Li <Yunxiang.Li at amd.com>
> > ---
> > drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c | 14 +-
> > drivers/gpu/drm/amd/amdgpu/amdgpu_fdinfo.c | 10 +-
> > drivers/gpu/drm/amd/amdgpu/amdgpu_object.c | 107 +++--------
> > drivers/gpu/drm/amd/amdgpu/amdgpu_object.h | 5 +-
> > drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h | 2 +-
> > drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 189 +++++++++++++++-----
> > drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h | 12 +-
> > drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c | 1 +
> > 8 files changed, 199 insertions(+), 141 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> > b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> > index b144404902255..1d8a0ff3c8604 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c
> > @@ -36,6 +36,7 @@
> > #include "amdgpu_gem.h"
> > #include "amdgpu_dma_buf.h"
> > #include "amdgpu_xgmi.h"
> > +#include "amdgpu_vm.h"
> > #include <drm/amdgpu_drm.h>
> > #include <drm/ttm/ttm_tt.h>
> > #include <linux/dma-buf.h>
> > @@ -190,6 +191,13 @@ static void amdgpu_dma_buf_unmap(struct
> dma_buf_attachment *attach,
> > }
> > }
> >
> > +static void amdgpu_dma_buf_release(struct dma_buf *buf) {
> > + struct amdgpu_bo *bo = gem_to_amdgpu_bo(buf->priv);
> > + amdgpu_vm_bo_update_shared(bo, -1);
> > + drm_gem_dmabuf_release(buf);
>
> Please run checkpatch.pl on the patch. As far as I can see it would complain about
> the coding style here (empty line between declaration and code).
>
> Not much of an issue but we would like to prevent upstream from complaining about
> such things.
Will do
> > +}
> > +
> > /**
> > * amdgpu_dma_buf_begin_cpu_access - &dma_buf_ops.begin_cpu_access
> implementation
> > * @dma_buf: Shared DMA buffer
> > @@ -237,7 +245,7 @@ const struct dma_buf_ops amdgpu_dmabuf_ops = {
> > .unpin = amdgpu_dma_buf_unpin,
> > .map_dma_buf = amdgpu_dma_buf_map,
> > .unmap_dma_buf = amdgpu_dma_buf_unmap,
> > - .release = drm_gem_dmabuf_release,
> > + .release = amdgpu_dma_buf_release,
> > .begin_cpu_access = amdgpu_dma_buf_begin_cpu_access,
> > .mmap = drm_gem_dmabuf_mmap,
> > .vmap = drm_gem_dmabuf_vmap,
> > @@ -265,8 +273,10 @@ struct dma_buf *amdgpu_gem_prime_export(struct
> drm_gem_object *gobj,
> > return ERR_PTR(-EPERM);
> >
> > buf = drm_gem_prime_export(gobj, flags);
> > - if (!IS_ERR(buf))
> > + if (!IS_ERR(buf)) {
> > buf->ops = &amdgpu_dmabuf_ops;
> > + amdgpu_vm_bo_update_shared(bo, +1);
> > + }
> >
> > return buf;
> > }
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_fdinfo.c
> > b/drivers/gpu/drm/amd/amdgpu/amdgpu_fdinfo.c
> > index 7a9573958d87c..e0e09f7b39d10 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_fdinfo.c
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_fdinfo.c
> > @@ -60,7 +60,7 @@ void amdgpu_show_fdinfo(struct drm_printer *p, struct
> drm_file *file)
> > struct amdgpu_fpriv *fpriv = file->driver_priv;
> > struct amdgpu_vm *vm = &fpriv->vm;
> >
> > - struct amdgpu_mem_stats stats[__AMDGPU_PL_LAST + 1] = { };
> > + struct amdgpu_mem_stats stats[__AMDGPU_PL_LAST] = { };
> > ktime_t usage[AMDGPU_HW_IP_NUM];
> > const char *pl_name[] = {
> > [TTM_PL_VRAM] = "vram",
> > @@ -70,13 +70,7 @@ void amdgpu_show_fdinfo(struct drm_printer *p, struct
> drm_file *file)
> > unsigned int hw_ip, i;
> > int ret;
> >
> > - ret = amdgpu_bo_reserve(vm->root.bo, false);
> > - if (ret)
> > - return;
> > -
> > - amdgpu_vm_get_memory(vm, stats, ARRAY_SIZE(stats));
> > - amdgpu_bo_unreserve(vm->root.bo);
> > -
> > + amdgpu_vm_get_memory(vm, stats);
> > amdgpu_ctx_mgr_usage(&fpriv->ctx_mgr, usage);
> >
> > /*
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> > b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> > index 2436b7c9ad12b..98563124ff99c 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
> > @@ -1156,7 +1156,7 @@ void amdgpu_bo_move_notify(struct ttm_buffer_object
> *bo,
> > return;
> >
> > abo = ttm_to_amdgpu_bo(bo);
> > - amdgpu_vm_bo_invalidate(abo, evict);
> > + amdgpu_vm_bo_move(abo, new_mem, evict);
> >
> > amdgpu_bo_kunmap(abo);
> >
> > @@ -1169,86 +1169,6 @@ void amdgpu_bo_move_notify(struct
> ttm_buffer_object *bo,
> > old_mem ? old_mem->mem_type : -1);
> > }
> >
> > -void amdgpu_bo_get_memory(struct amdgpu_bo *bo,
> > - struct amdgpu_mem_stats *stats,
> > - unsigned int sz)
> > -{
> > - const unsigned int domain_to_pl[] = {
> > - [ilog2(AMDGPU_GEM_DOMAIN_CPU)] = TTM_PL_SYSTEM,
> > - [ilog2(AMDGPU_GEM_DOMAIN_GTT)] = TTM_PL_TT,
> > - [ilog2(AMDGPU_GEM_DOMAIN_VRAM)] = TTM_PL_VRAM,
> > - [ilog2(AMDGPU_GEM_DOMAIN_GDS)] =
> AMDGPU_PL_GDS,
> > - [ilog2(AMDGPU_GEM_DOMAIN_GWS)] =
> AMDGPU_PL_GWS,
> > - [ilog2(AMDGPU_GEM_DOMAIN_OA)] = AMDGPU_PL_OA,
> > - [ilog2(AMDGPU_GEM_DOMAIN_DOORBELL)] =
> AMDGPU_PL_DOORBELL,
> > - };
> > - struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
> > - struct ttm_resource *res = bo->tbo.resource;
> > - struct drm_gem_object *obj = &bo->tbo.base;
> > - uint64_t size = amdgpu_bo_size(bo);
> > - unsigned int type;
> > -
> > - if (!res) {
> > - /*
> > - * If no backing store use one of the preferred domain for basic
> > - * stats. We take the MSB since that should give a reasonable
> > - * view.
> > - */
> > - BUILD_BUG_ON(TTM_PL_VRAM < TTM_PL_TT ||
> > - TTM_PL_VRAM < TTM_PL_SYSTEM);
> > - type = fls(bo->preferred_domains &
> AMDGPU_GEM_DOMAIN_MASK);
> > - if (!type)
> > - return;
> > - type--;
> > - if (drm_WARN_ON_ONCE(&adev->ddev,
> > - type >= ARRAY_SIZE(domain_to_pl)))
> > - return;
> > - type = domain_to_pl[type];
> > - } else {
> > - type = res->mem_type;
> > - }
> > -
> > - /* Squash some into 'cpu' to keep the legacy userspace view. */
> > - switch (type) {
> > - case TTM_PL_VRAM:
> > - case TTM_PL_TT:
> > - case TTM_PL_SYSTEM:
> > - break;
> > - default:
> > - type = TTM_PL_SYSTEM;
> > - break;
> > - }
> > -
> > - if (drm_WARN_ON_ONCE(&adev->ddev, type >= sz))
> > - return;
> > -
> > - /* DRM stats common fields: */
> > -
> > - if (drm_gem_object_is_shared_for_memory_stats(obj))
> > - stats[type].drm.shared += size;
> > - else
> > - stats[type].drm.private += size;
> > -
> > - if (res) {
> > - stats[type].drm.resident += size;
> > -
> > - if (!dma_resv_test_signaled(obj->resv,
> DMA_RESV_USAGE_BOOKKEEP))
> > - stats[type].drm.active += size;
> > - else if (bo->flags & AMDGPU_GEM_CREATE_DISCARDABLE)
> > - stats[type].drm.purgeable += size;
> > - }
> > -
> > - /* amdgpu specific stats: */
> > -
> > - if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM) {
> > - stats[TTM_PL_VRAM].requested += size;
> > - if (type != TTM_PL_VRAM)
> > - stats[TTM_PL_VRAM].evicted += size;
> > - } else if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_GTT) {
> > - stats[TTM_PL_TT].requested += size;
> > - }
> > -}
> > -
> > /**
> > * amdgpu_bo_release_notify - notification about a BO being released
> > * @bo: pointer to a buffer object
> > @@ -1463,6 +1383,31 @@ u64 amdgpu_bo_gpu_offset_no_check(struct
> amdgpu_bo *bo)
> > return amdgpu_gmc_sign_extend(offset);
> > }
> >
> > +uint32_t amdgpu_bo_get_preferred_placement(struct amdgpu_bo *bo) {
> > + uint32_t domain = bo->preferred_domains &
> AMDGPU_GEM_DOMAIN_MASK;
> > + if (!domain)
> > + return TTM_PL_SYSTEM;
> > +
> > + switch (ilog2(domain)) {
> > + case AMDGPU_GEM_DOMAIN_CPU:
> > + return TTM_PL_SYSTEM;
> > + case AMDGPU_GEM_DOMAIN_GTT:
> > + return TTM_PL_TT;
> > + case AMDGPU_GEM_DOMAIN_VRAM:
> > + return TTM_PL_VRAM;
> > + case AMDGPU_GEM_DOMAIN_GDS:
> > + return AMDGPU_PL_GDS;
> > + case AMDGPU_GEM_DOMAIN_GWS:
> > + return AMDGPU_PL_GWS;
> > + case AMDGPU_GEM_DOMAIN_OA:
> > + return AMDGPU_PL_OA;
> > + case AMDGPU_GEM_DOMAIN_DOORBELL:
> > + return AMDGPU_PL_DOORBELL;
> > + default:
> > + return TTM_PL_SYSTEM;
>
> If I'm not completely mistaken that won't work like that.
>
> The AMDGPU_GEM_DOMAIN_* defines are masks and not shifts.
Yeah you are right, should have been rounddown_pow_of_two
> > + }
> > +}
> > +
> > /**
> > * amdgpu_bo_get_preferred_domain - get preferred domain
> > * @adev: amdgpu device object
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
> > b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
> > index be6769852ece4..bd58a8b0ece66 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
> > @@ -30,6 +30,7 @@
> >
> > #include <drm/amdgpu_drm.h>
> > #include "amdgpu.h"
> > +#include "amdgpu_ttm.h"
> > #include "amdgpu_res_cursor.h"
>
> Why is that necessary?
I got a compile error otherwise for those AMDGPU_PL_*
> >
> > #ifdef CONFIG_MMU_NOTIFIER
> > @@ -300,9 +301,7 @@ int amdgpu_bo_sync_wait_resv(struct amdgpu_device
> *adev, struct dma_resv *resv,
> > int amdgpu_bo_sync_wait(struct amdgpu_bo *bo, void *owner, bool intr);
> > u64 amdgpu_bo_gpu_offset(struct amdgpu_bo *bo);
> > u64 amdgpu_bo_gpu_offset_no_check(struct amdgpu_bo *bo); -void
> > amdgpu_bo_get_memory(struct amdgpu_bo *bo,
> > - struct amdgpu_mem_stats *stats,
> > - unsigned int size);
> > +uint32_t amdgpu_bo_get_preferred_placement(struct amdgpu_bo *bo);
> > uint32_t amdgpu_bo_get_preferred_domain(struct amdgpu_device *adev,
> > uint32_t domain);
> >
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
> > b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
> > index 2852a6064c9ac..a9088e864fde4 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
> > @@ -26,8 +26,8 @@
> >
> > #include <linux/dma-direction.h>
> > #include <drm/gpu_scheduler.h>
> > +#include <drm/ttm/ttm_placement.h>
> > #include "amdgpu_vram_mgr.h"
> > -#include "amdgpu.h"
>
> Looks like a valuable cleanup, but should probably a separate patch.
Without this there's a circular include that breaks compilation combined with the above
> >
> > #define AMDGPU_PL_GDS (TTM_PL_PRIV + 0)
> > #define AMDGPU_PL_GWS (TTM_PL_PRIV + 1)
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> > b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> > index 9fab64edd0530..abd35c18ddaa8 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> > @@ -36,6 +36,7 @@
> > #include <drm/ttm/ttm_tt.h>
> > #include <drm/drm_exec.h>
> > #include "amdgpu.h"
> > +#include "amdgpu_vm.h"
> > #include "amdgpu_trace.h"
> > #include "amdgpu_amdkfd.h"
> > #include "amdgpu_gmc.h"
> > @@ -310,6 +311,94 @@ static void amdgpu_vm_bo_reset_state_machine(struct
> amdgpu_vm *vm)
> > spin_unlock(&vm->status_lock);
> > }
> >
> > +/**
> > + * amdgpu_vm_update_shared - helper to update shared memory stat
> > + * @base: base structure for tracking BO usage in a VM
> > + * @sign: if we should add (+1) or subtract (-1) from the shared stat
> > + *
> > + * Takes the vm status_lock and updates the shared memory stat. If
> > +the basic
> > + * stat changed (e.g. buffer was moved) amdgpu_vm_update_stats need
> > +to be called
> > + * as well.
> > + */
> > +static void amdgpu_vm_update_shared(struct amdgpu_vm_bo_base *base,
> > +int sign) {
> > + struct amdgpu_vm *vm = base->vm;
> > + struct amdgpu_bo *bo = base->bo;
> > + struct ttm_resource *res;
> > + int64_t size;
> > + uint32_t type;
> > +
> > + if (!vm || !bo)
> > + return;
> > +
> > + size = sign * amdgpu_bo_size(bo);
> > + if ((res = bo->tbo.resource))
> > + type = res->mem_type;
> > + else
> > + type = amdgpu_bo_get_preferred_placement(bo);
>
> As discussed with Tvrtko that won't work like this.
>
> Either use the preferred placement or the actual backing store, but don't use a
> fallback here.
I had a follow up discussion with Tvrtko here https://lists.freedesktop.org/archives/amd-gfx/2024-October/116024.html it seems like this is the intended semantics for the drm-total-* stats. I can see it going both ways, I guess it's just up to which design is most useful for whom ever is reading the stats.
Current design is for it to mean "all the buffer currently at X" + "all the buffer that wants to be at X but currently don't have a backing"
The alternative I guess is for it to mean "all the buffer that wants to be at X"
Btw, I'm having trouble figuring out where I should account for drm-active-* it's for buffers that are currently being used (e.g. have a fence not signaled) it seems like the work scheduling part is quite far removed from the individual BOs...
> > + if (type >= __AMDGPU_PL_LAST)
> > + return;
> > +
> > + spin_lock(&vm->status_lock);
> > + vm->stats[type].drm.shared += size;
> > + vm->stats[type].drm.private -= size;
> > + spin_unlock(&vm->status_lock);
> > +}
> > +
> > +/**
> > + * amdgpu_vm_update_stats - helper to update normal memory stat
> > + * @base: base structure for tracking BO usage in a VM
> > + * @new_res: if not NULL, the ttm_resource to use for the purpose of
> > +accounting
> > + * (i.e. ignore the one in the BO)
> > + * @sign: if we should add (+1) or subtract (-1) from the stat
> > + *
> > + * Takes the vm status_lock and updates the basic memory stat. If the
> > +shared
> > + * stat changed (e.g. buffer was exported) amdgpu_vm_update_shared
> > +need to be
> > + * called as well.
> > + */
> > +void amdgpu_vm_update_stats(struct amdgpu_vm_bo_base *base,
> > + struct ttm_resource *new_res, int sign) {
> > + struct amdgpu_vm *vm = base->vm;
> > + struct amdgpu_bo *bo = base->bo;
> > + struct ttm_resource *res;
> > + int64_t size;
> > + uint32_t type;
> > + bool shared;
> > +
> > + if (!vm || !bo)
> > + return;
> > +
> > + size = sign * amdgpu_bo_size(bo);
> > + res = new_res ? new_res : bo->tbo.resource;
> > + type = res ? res->mem_type : amdgpu_bo_get_preferred_placement(bo);
>
> Same here. Don't use the preferred domain as fallback.
>
> > + shared = drm_gem_object_is_shared_for_memory_stats(&bo->tbo.base);
> > +
> > + if (type >= __AMDGPU_PL_LAST)
> > + return;
> > +
> > + spin_lock(&vm->status_lock);
> > +
> > + if (shared)
> > + vm->stats[type].drm.shared += size;
> > + else
> > + vm->stats[type].drm.private += size;
> > + if (res)
> > + vm->stats[type].drm.resident += size;
> > + if (bo->flags & AMDGPU_GEM_CREATE_DISCARDABLE)
> > + vm->stats[type].drm.purgeable += size;
> > +
> > + if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM) {
> > + vm->stats[TTM_PL_VRAM].requested += size;
> > + if (type != TTM_PL_VRAM)
> > + vm->stats[TTM_PL_VRAM].evicted += size;
>
> That check is probably not correct. We have BOs which can be placed in both
> VRAM and GTT.
That is true, but does it make sense to count it towards evicted if say our picking order prefers VRAM over GTT?
> > + } else if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_GTT) {
> > + vm->stats[TTM_PL_TT].requested += size;
> > + }
> > +
> > + spin_unlock(&vm->status_lock);
> > +}
> > +
> > /**
> > * amdgpu_vm_bo_base_init - Adds bo to the list of bos associated with the vm
> > *
> > @@ -332,6 +421,7 @@ void amdgpu_vm_bo_base_init(struct
> amdgpu_vm_bo_base *base,
> > return;
> > base->next = bo->vm_bo;
> > bo->vm_bo = base;
> > + amdgpu_vm_update_stats(base, NULL, +1);
> >
> > if (!amdgpu_vm_is_bo_always_valid(vm, bo))
> > return;
> > @@ -1082,53 +1172,11 @@ int amdgpu_vm_update_range(struct
> amdgpu_device *adev, struct amdgpu_vm *vm,
> > return r;
> > }
> >
> > -static void amdgpu_vm_bo_get_memory(struct amdgpu_bo_va *bo_va,
> > - struct amdgpu_mem_stats *stats,
> > - unsigned int size)
> > -{
> > - struct amdgpu_vm *vm = bo_va->base.vm;
> > - struct amdgpu_bo *bo = bo_va->base.bo;
> > -
> > - if (!bo)
> > - return;
> > -
> > - /*
> > - * For now ignore BOs which are currently locked and potentially
> > - * changing their location.
> > - */
> > - if (!amdgpu_vm_is_bo_always_valid(vm, bo) &&
> > - !dma_resv_trylock(bo->tbo.base.resv))
> > - return;
> > -
> > - amdgpu_bo_get_memory(bo, stats, size);
> > - if (!amdgpu_vm_is_bo_always_valid(vm, bo))
> > - dma_resv_unlock(bo->tbo.base.resv);
> > -}
> > -
> > void amdgpu_vm_get_memory(struct amdgpu_vm *vm,
> > - struct amdgpu_mem_stats *stats,
> > - unsigned int size)
> > + struct amdgpu_mem_stats stats[__AMDGPU_PL_LAST])
> > {
> > - struct amdgpu_bo_va *bo_va, *tmp;
> > -
> > spin_lock(&vm->status_lock);
> > - list_for_each_entry_safe(bo_va, tmp, &vm->idle, base.vm_status)
> > - amdgpu_vm_bo_get_memory(bo_va, stats, size);
> > -
> > - list_for_each_entry_safe(bo_va, tmp, &vm->evicted, base.vm_status)
> > - amdgpu_vm_bo_get_memory(bo_va, stats, size);
> > -
> > - list_for_each_entry_safe(bo_va, tmp, &vm->relocated, base.vm_status)
> > - amdgpu_vm_bo_get_memory(bo_va, stats, size);
> > -
> > - list_for_each_entry_safe(bo_va, tmp, &vm->moved, base.vm_status)
> > - amdgpu_vm_bo_get_memory(bo_va, stats, size);
> > -
> > - list_for_each_entry_safe(bo_va, tmp, &vm->invalidated, base.vm_status)
> > - amdgpu_vm_bo_get_memory(bo_va, stats, size);
> > -
> > - list_for_each_entry_safe(bo_va, tmp, &vm->done, base.vm_status)
> > - amdgpu_vm_bo_get_memory(bo_va, stats, size);
> > + memcpy(stats, vm->stats, sizeof(*stats) * __AMDGPU_PL_LAST);
> > spin_unlock(&vm->status_lock);
> > }
> >
> > @@ -2071,6 +2119,7 @@ void amdgpu_vm_bo_del(struct amdgpu_device
> *adev,
> > if (*base != &bo_va->base)
> > continue;
> >
> > + amdgpu_vm_update_stats(*base, NULL, -1);
> > *base = bo_va->base.next;
> > break;
> > }
> > @@ -2136,6 +2185,22 @@ bool amdgpu_vm_evictable(struct amdgpu_bo *bo)
> > return true;
> > }
> >
> > +/**
> > + * amdgpu_vm_bo_update_shared - called when bo gets shared/unshared
> > + *
> > + * @bo: amdgpu buffer object
> > + * @sign: if we should add (+1) or subtract (-1) the memory stat
> > + *
> > + * Update the per VM stats for all the vm */ void
> > +amdgpu_vm_bo_update_shared(struct amdgpu_bo *bo, int sign) {
> > + struct amdgpu_vm_bo_base *bo_base;
> > +
> > + for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next)
> > + amdgpu_vm_update_shared(bo_base, sign); }
> > +
> > /**
> > * amdgpu_vm_bo_invalidate - mark the bo as invalid
> > *
> > @@ -2169,6 +2234,28 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_bo
> *bo, bool evicted)
> > }
> > }
> >
> > +/**
> > + * amdgpu_vm_bo_move - handle BO move
> > + *
> > + * @bo: amdgpu buffer object
> > + * @new_mem: the new placement of the BO move
> > + * @evicted: is the BO evicted
> > + *
> > + * Update the memory stats for the new placement and mark @bo as invalid.
> > + */
> > +void amdgpu_vm_bo_move(struct amdgpu_bo *bo, struct ttm_resource
> *new_mem,
> > + bool evicted)
> > +{
> > + struct amdgpu_vm_bo_base *bo_base;
> > +
> > + for (bo_base = bo->vm_bo; bo_base; bo_base = bo_base->next) {
> > + amdgpu_vm_update_stats(bo_base, bo->tbo.resource, -1);
> > + amdgpu_vm_update_stats(bo_base, new_mem, +1);
> > + }
> > +
> > + amdgpu_vm_bo_invalidate(bo, evicted); }
> > +
> > /**
> > * amdgpu_vm_get_block_size - calculate VM page table size as power of two
> > *
> > @@ -2585,6 +2672,18 @@ void amdgpu_vm_release_compute(struct
> amdgpu_device *adev, struct amdgpu_vm *vm)
> > vm->is_compute_context = false;
> > }
> >
> > +static int amdgpu_vm_stats_is_zero(struct amdgpu_vm *vm) {
> > + int is_zero = 1;
> > + for (int i = 0; i < __AMDGPU_PL_LAST; ++i) {
> > + is_zero = drm_memory_stats_is_zero(&vm->stats[i].drm) &&
> > + vm->stats->evicted == 0 && vm->stats->requested == 0;
> > + if (!is_zero)
> > + break;
>
> Just make that an "if (...) return false", no need for the local variable.
>
> Regards,
> Christian.
>
D'oh!
Teddy
More information about the dri-devel
mailing list