[Intel-xe] [PATCH 2/2] fixup! drm/xe: Introduce a new DRM driver for Intel GPUs
Rodrigo Vivi
rodrigo.vivi at intel.com
Mon May 1 19:03:14 UTC 2023
Signed-off-by: Rodrigo Vivi <rodrigo.vivi at intel.com>
---
drivers/gpu/drm/xe/Makefile | 1 +
drivers/gpu/drm/xe/xe_pt.c | 110 +++++++++++++++----------------
drivers/gpu/drm/xe/xe_pt_types.h | 4 +-
3 files changed, 56 insertions(+), 59 deletions(-)
diff --git a/drivers/gpu/drm/xe/Makefile b/drivers/gpu/drm/xe/Makefile
index 74a84080f242..b84e191ba14f 100644
--- a/drivers/gpu/drm/xe/Makefile
+++ b/drivers/gpu/drm/xe/Makefile
@@ -73,6 +73,7 @@ xe-y += xe_bb.o \
xe_pm.o \
xe_preempt_fence.o \
xe_pt.o \
+ xe_pt_walk.o \
xe_query.o \
xe_reg_sr.o \
xe_reg_whitelist.o \
diff --git a/drivers/gpu/drm/xe/xe_pt.c b/drivers/gpu/drm/xe/xe_pt.c
index 4ee5ea2cabc9..f15282996c3b 100644
--- a/drivers/gpu/drm/xe/xe_pt.c
+++ b/drivers/gpu/drm/xe/xe_pt.c
@@ -5,14 +5,13 @@
#include "xe_pt.h"
-#include <drm/drm_pt_walk.h>
-
#include "xe_bo.h"
#include "xe_device.h"
#include "xe_gt.h"
#include "xe_gt_tlb_invalidation.h"
#include "xe_migrate.h"
#include "xe_pt_types.h"
+#include "xe_pt_walk.h"
#include "xe_res_cursor.h"
#include "xe_trace.h"
#include "xe_ttm_stolen_mgr.h"
@@ -20,8 +19,8 @@
struct xe_pt_dir {
struct xe_pt pt;
- /** @dir: Directory structure for the drm_pt_walk functionality */
- struct drm_pt_dir dir;
+ /** @dir: Directory structure for the xe_pt_walk functionality */
+ struct xe_ptw_dir dir;
};
#if IS_ENABLED(CONFIG_DRM_XE_DEBUG_VM)
@@ -44,7 +43,7 @@ static struct xe_pt_dir *as_xe_pt_dir(struct xe_pt *pt)
static struct xe_pt *xe_pt_entry(struct xe_pt_dir *pt_dir, unsigned int index)
{
- return container_of(pt_dir->dir.entries[index], struct xe_pt, drm);
+ return container_of(pt_dir->dir.entries[index], struct xe_pt, base);
}
/**
@@ -211,7 +210,7 @@ struct xe_pt *xe_pt_create(struct xe_vm *vm, struct xe_gt *gt,
int err;
size = !level ? sizeof(struct xe_pt) : sizeof(struct xe_pt_dir) +
- XE_PDES * sizeof(struct drm_pt *);
+ XE_PDES * sizeof(struct xe_ptw *);
pt = kzalloc(size, GFP_KERNEL);
if (!pt)
return ERR_PTR(-ENOMEM);
@@ -227,7 +226,7 @@ struct xe_pt *xe_pt_create(struct xe_vm *vm, struct xe_gt *gt,
}
pt->bo = bo;
pt->level = level;
- pt->drm.dir = level ? &as_xe_pt_dir(pt)->dir : NULL;
+ pt->base.dir = level ? &as_xe_pt_dir(pt)->dir : NULL;
XE_BUG_ON(level > XE_VM_MAX_LEVEL);
@@ -404,8 +403,8 @@ struct xe_pt_update {
};
struct xe_pt_stage_bind_walk {
- /** drm: The base class. */
- struct drm_pt_walk drm;
+ /** base: The base class. */
+ struct xe_pt_walk base;
/* Input parameters for the walk */
/** @vm: The vm we're building for. */
@@ -532,7 +531,7 @@ xe_pt_insert_entry(struct xe_pt_stage_bind_walk *xe_walk, struct xe_pt *parent,
struct iosys_map *map = &parent->bo->vmap;
if (unlikely(xe_child))
- parent->drm.dir->entries[offset] = &xe_child->drm;
+ parent->base.dir->entries[offset] = &xe_child->base;
xe_pt_write(xe_walk->vm->xe, map, offset, pte);
parent->num_live++;
@@ -556,7 +555,7 @@ static bool xe_pt_hugepte_possible(u64 addr, u64 next, unsigned int level,
u64 size, dma;
/* Does the virtual range requested cover a huge pte? */
- if (!drm_pt_covers(addr, next, level, &xe_walk->drm))
+ if (!xe_pt_covers(addr, next, level, &xe_walk->base))
return false;
/* Does the DMA segment cover the whole pte? */
@@ -618,15 +617,15 @@ xe_pt_is_pte_ps64K(u64 addr, u64 next, struct xe_pt_stage_bind_walk *xe_walk)
}
static int
-xe_pt_stage_bind_entry(struct drm_pt *parent, pgoff_t offset,
+xe_pt_stage_bind_entry(struct xe_ptw *parent, pgoff_t offset,
unsigned int level, u64 addr, u64 next,
- struct drm_pt **child,
+ struct xe_ptw **child,
enum page_walk_action *action,
- struct drm_pt_walk *walk)
+ struct xe_pt_walk *walk)
{
struct xe_pt_stage_bind_walk *xe_walk =
- container_of(walk, typeof(*xe_walk), drm);
- struct xe_pt *xe_parent = container_of(parent, typeof(*xe_parent), drm);
+ container_of(walk, typeof(*xe_walk), base);
+ struct xe_pt *xe_parent = container_of(parent, typeof(*xe_parent), base);
struct xe_pt *xe_child;
bool covers;
int ret = 0;
@@ -675,7 +674,7 @@ xe_pt_stage_bind_entry(struct drm_pt *parent, pgoff_t offset,
xe_walk->l0_end_addr = next;
}
- covers = drm_pt_covers(addr, next, level, &xe_walk->drm);
+ covers = xe_pt_covers(addr, next, level, &xe_walk->base);
if (covers || !*child) {
u64 flags = 0;
@@ -689,7 +688,7 @@ xe_pt_stage_bind_entry(struct drm_pt *parent, pgoff_t offset,
if (!covers)
xe_pt_populate_empty(xe_walk->gt, xe_walk->vm, xe_child);
- *child = &xe_child->drm;
+ *child = &xe_child->base;
/*
* Prefer the compact pagetable layout for L0 if possible.
@@ -712,7 +711,7 @@ xe_pt_stage_bind_entry(struct drm_pt *parent, pgoff_t offset,
return ret;
}
-static const struct drm_pt_walk_ops xe_pt_stage_bind_ops = {
+static const struct xe_pt_walk_ops xe_pt_stage_bind_ops = {
.pt_entry = xe_pt_stage_bind_entry,
};
@@ -742,7 +741,7 @@ xe_pt_stage_bind(struct xe_gt *gt, struct xe_vma *vma,
bool is_vram = !xe_vma_is_userptr(vma) && bo && xe_bo_is_vram(bo);
struct xe_res_cursor curs;
struct xe_pt_stage_bind_walk xe_walk = {
- .drm = {
+ .base = {
.ops = &xe_pt_stage_bind_ops,
.shifts = xe_normal_pt_shifts,
.max_level = XE_PT_HIGHEST_LEVEL,
@@ -787,8 +786,8 @@ xe_pt_stage_bind(struct xe_gt *gt, struct xe_vma *vma,
xe_res_first_sg(xe_bo_get_sg(bo), vma->bo_offset,
vma->end - vma->start + 1, &curs);
- ret = drm_pt_walk_range(&pt->drm, pt->level, vma->start, vma->end + 1,
- &xe_walk.drm);
+ ret = xe_pt_walk_range(&pt->base, pt->level, vma->start, vma->end + 1,
+ &xe_walk.base);
*num_entries = xe_walk.wupd.num_used_entries;
return ret;
@@ -814,20 +813,17 @@ xe_pt_stage_bind(struct xe_gt *gt, struct xe_vma *vma,
* be shared page tables also at lower levels, so it adjusts the walk action
* accordingly.
*
- * Note that the function is not device-specific so could be made a drm
- * pagewalk helper.
- *
* Return: true if there were non-shared entries, false otherwise.
*/
static bool xe_pt_nonshared_offsets(u64 addr, u64 end, unsigned int level,
- struct drm_pt_walk *walk,
+ struct xe_pt_walk *walk,
enum page_walk_action *action,
pgoff_t *offset, pgoff_t *end_offset)
{
u64 size = 1ull << walk->shifts[level];
- *offset = drm_pt_offset(addr, level, walk);
- *end_offset = drm_pt_num_entries(addr, end, level, walk) + *offset;
+ *offset = xe_pt_offset(addr, level, walk);
+ *end_offset = xe_pt_num_entries(addr, end, level, walk) + *offset;
if (!level)
return true;
@@ -851,8 +847,8 @@ static bool xe_pt_nonshared_offsets(u64 addr, u64 end, unsigned int level,
}
struct xe_pt_zap_ptes_walk {
- /** @drm: The walk base-class */
- struct drm_pt_walk drm;
+ /** @base: The walk base-class */
+ struct xe_pt_walk base;
/* Input parameters for the walk */
/** @gt: The gt we're building for */
@@ -863,15 +859,15 @@ struct xe_pt_zap_ptes_walk {
bool needs_invalidate;
};
-static int xe_pt_zap_ptes_entry(struct drm_pt *parent, pgoff_t offset,
+static int xe_pt_zap_ptes_entry(struct xe_ptw *parent, pgoff_t offset,
unsigned int level, u64 addr, u64 next,
- struct drm_pt **child,
+ struct xe_ptw **child,
enum page_walk_action *action,
- struct drm_pt_walk *walk)
+ struct xe_pt_walk *walk)
{
struct xe_pt_zap_ptes_walk *xe_walk =
- container_of(walk, typeof(*xe_walk), drm);
- struct xe_pt *xe_child = container_of(*child, typeof(*xe_child), drm);
+ container_of(walk, typeof(*xe_walk), base);
+ struct xe_pt *xe_child = container_of(*child, typeof(*xe_child), base);
pgoff_t end_offset;
XE_BUG_ON(!*child);
@@ -893,7 +889,7 @@ static int xe_pt_zap_ptes_entry(struct drm_pt *parent, pgoff_t offset,
return 0;
}
-static const struct drm_pt_walk_ops xe_pt_zap_ptes_ops = {
+static const struct xe_pt_walk_ops xe_pt_zap_ptes_ops = {
.pt_entry = xe_pt_zap_ptes_entry,
};
@@ -916,7 +912,7 @@ static const struct drm_pt_walk_ops xe_pt_zap_ptes_ops = {
bool xe_pt_zap_ptes(struct xe_gt *gt, struct xe_vma *vma)
{
struct xe_pt_zap_ptes_walk xe_walk = {
- .drm = {
+ .base = {
.ops = &xe_pt_zap_ptes_ops,
.shifts = xe_normal_pt_shifts,
.max_level = XE_PT_HIGHEST_LEVEL,
@@ -928,8 +924,8 @@ bool xe_pt_zap_ptes(struct xe_gt *gt, struct xe_vma *vma)
if (!(vma->gt_present & BIT(gt->info.id)))
return false;
- (void)drm_pt_walk_shared(&pt->drm, pt->level, vma->start, vma->end + 1,
- &xe_walk.drm);
+ (void)xe_pt_walk_shared(&pt->base, pt->level, vma->start, vma->end + 1,
+ &xe_walk.base);
return xe_walk.needs_invalidate;
}
@@ -1015,7 +1011,7 @@ static void xe_pt_commit_bind(struct xe_vma *vma,
xe_pt_destroy(xe_pt_entry(pt_dir, j_),
vma->vm->flags, deferred);
- pt_dir->dir.entries[j_] = &newpte->drm;
+ pt_dir->dir.entries[j_] = &newpte->base;
}
kfree(entries[i].pt_entries);
}
@@ -1375,8 +1371,8 @@ __xe_pt_bind_vma(struct xe_gt *gt, struct xe_vma *vma, struct xe_engine *e,
}
struct xe_pt_stage_unbind_walk {
- /** @drm: The pagewalk base-class. */
- struct drm_pt_walk drm;
+ /** @base: The pagewalk base-class. */
+ struct xe_pt_walk base;
/* Input parameters for the walk */
/** @gt: The gt we're unbinding from. */
@@ -1404,10 +1400,10 @@ struct xe_pt_stage_unbind_walk {
static bool xe_pt_check_kill(u64 addr, u64 next, unsigned int level,
const struct xe_pt *child,
enum page_walk_action *action,
- struct drm_pt_walk *walk)
+ struct xe_pt_walk *walk)
{
struct xe_pt_stage_unbind_walk *xe_walk =
- container_of(walk, typeof(*xe_walk), drm);
+ container_of(walk, typeof(*xe_walk), base);
unsigned int shift = walk->shifts[level];
u64 size = 1ull << shift;
@@ -1428,13 +1424,13 @@ static bool xe_pt_check_kill(u64 addr, u64 next, unsigned int level,
return false;
}
-static int xe_pt_stage_unbind_entry(struct drm_pt *parent, pgoff_t offset,
+static int xe_pt_stage_unbind_entry(struct xe_ptw *parent, pgoff_t offset,
unsigned int level, u64 addr, u64 next,
- struct drm_pt **child,
+ struct xe_ptw **child,
enum page_walk_action *action,
- struct drm_pt_walk *walk)
+ struct xe_pt_walk *walk)
{
- struct xe_pt *xe_child = container_of(*child, typeof(*xe_child), drm);
+ struct xe_pt *xe_child = container_of(*child, typeof(*xe_child), base);
XE_BUG_ON(!*child);
XE_BUG_ON(!level && xe_child->is_compact);
@@ -1445,15 +1441,15 @@ static int xe_pt_stage_unbind_entry(struct drm_pt *parent, pgoff_t offset,
}
static int
-xe_pt_stage_unbind_post_descend(struct drm_pt *parent, pgoff_t offset,
+xe_pt_stage_unbind_post_descend(struct xe_ptw *parent, pgoff_t offset,
unsigned int level, u64 addr, u64 next,
- struct drm_pt **child,
+ struct xe_ptw **child,
enum page_walk_action *action,
- struct drm_pt_walk *walk)
+ struct xe_pt_walk *walk)
{
struct xe_pt_stage_unbind_walk *xe_walk =
- container_of(walk, typeof(*xe_walk), drm);
- struct xe_pt *xe_child = container_of(*child, typeof(*xe_child), drm);
+ container_of(walk, typeof(*xe_walk), base);
+ struct xe_pt *xe_child = container_of(*child, typeof(*xe_child), base);
pgoff_t end_offset;
u64 size = 1ull << walk->shifts[--level];
@@ -1477,7 +1473,7 @@ xe_pt_stage_unbind_post_descend(struct drm_pt *parent, pgoff_t offset,
return 0;
}
-static const struct drm_pt_walk_ops xe_pt_stage_unbind_ops = {
+static const struct xe_pt_walk_ops xe_pt_stage_unbind_ops = {
.pt_entry = xe_pt_stage_unbind_entry,
.pt_post_descend = xe_pt_stage_unbind_post_descend,
};
@@ -1500,7 +1496,7 @@ static unsigned int xe_pt_stage_unbind(struct xe_gt *gt, struct xe_vma *vma,
struct xe_vm_pgtable_update *entries)
{
struct xe_pt_stage_unbind_walk xe_walk = {
- .drm = {
+ .base = {
.ops = &xe_pt_stage_unbind_ops,
.shifts = xe_normal_pt_shifts,
.max_level = XE_PT_HIGHEST_LEVEL,
@@ -1512,8 +1508,8 @@ static unsigned int xe_pt_stage_unbind(struct xe_gt *gt, struct xe_vma *vma,
};
struct xe_pt *pt = vma->vm->pt_root[gt->info.id];
- (void)drm_pt_walk_shared(&pt->drm, pt->level, vma->start, vma->end + 1,
- &xe_walk.drm);
+ (void)xe_pt_walk_shared(&pt->base, pt->level, vma->start, vma->end + 1,
+ &xe_walk.base);
return xe_walk.wupd.num_used_entries;
}
diff --git a/drivers/gpu/drm/xe/xe_pt_types.h b/drivers/gpu/drm/xe/xe_pt_types.h
index 2bb5d0e319b7..2ed64c0a4485 100644
--- a/drivers/gpu/drm/xe/xe_pt_types.h
+++ b/drivers/gpu/drm/xe/xe_pt_types.h
@@ -6,7 +6,7 @@
#ifndef _XE_PT_TYPES_H_
#define _XE_PT_TYPES_H_
-#include <drm/drm_pt_walk.h>
+#include "xe_pt_walk.h"
enum xe_cache_level {
XE_CACHE_NONE,
@@ -17,7 +17,7 @@ enum xe_cache_level {
#define XE_VM_MAX_LEVEL 4
struct xe_pt {
- struct drm_pt drm;
+ struct xe_ptw base;
struct xe_bo *bo;
unsigned int level;
unsigned int num_live;
--
2.39.2
More information about the Intel-xe
mailing list