[Intel-xe] [PATCH 1/3] drm/xe: Simplify rebar sizing

Lucas De Marchi lucas.demarchi at intel.com
Thu Apr 27 20:59:35 UTC 2023


On Thu, Apr 27, 2023 at 03:52:11PM -0400, Michael J. Ruhl wrote:
>"Right sizing" the PCI BAR is not necessary.  If rebar is needed
>size to the maximum available.
>
>Allow for specific sizing as well.

Instead of bullets, maybe convert to freeform text? Here it took me a
while to figure out this info doens't contradict the first one about
right sizing. Suggestion: "the behavior for force_vram_bar_size is still
preserved though, allowing to test with arbitrary BAR values".

>
>Update associated code for consistency.
>
>Signed-off-by: Michael J. Ruhl <michael.j.ruhl at intel.com>
>---
> drivers/gpu/drm/xe/xe_device_types.h   |  12 +--
> drivers/gpu/drm/xe/xe_migrate.c        |   2 +-
> drivers/gpu/drm/xe/xe_mmio.c           | 130 +++++++++++--------------
> drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c |  16 +--
> 4 files changed, 69 insertions(+), 91 deletions(-)
>
>diff --git a/drivers/gpu/drm/xe/xe_device_types.h b/drivers/gpu/drm/xe/xe_device_types.h
>index 1cb404e48aaa..24dc5d2e9e05 100644
>--- a/drivers/gpu/drm/xe/xe_device_types.h
>+++ b/drivers/gpu/drm/xe/xe_device_types.h
>@@ -194,14 +194,14 @@ struct xe_device {
> 			 * @io_size: IO size of VRAM.
> 			 *
> 			 * This represents how much of VRAM we can access via
>-			 * the CPU through the VRAM BAR. This can be smaller
>-			 * than @size, in which case only part of VRAM is CPU
>-			 * accessible (typically the first 256M). This
>-			 * configuration is known as small-bar.
>+			 * the CPU through the VRAM BAR.

the via + through here read odd. Maybe something like:

			* This represents how much of VRAM the CPU can access via
			* the VRAM BAR. ....

>+			 * On systems that do not support large BAR IO space,
>+			 * this can be smaller than the actual memory size, in
>+			 * which case only part of VRAM is CPU accessible
>+			 * (typically the first 256M).  This configuration is
>+			 * known as small-bar.
> 			 */
> 			resource_size_t io_size;
>-			/** @size: Total size of VRAM */
>-			resource_size_t size;
> 			/** @mapping: pointer to VRAM mappable space */
> 			void *__iomem mapping;
> 		} vram;
>diff --git a/drivers/gpu/drm/xe/xe_migrate.c b/drivers/gpu/drm/xe/xe_migrate.c
>index f40f47ccb76f..a1d507db0098 100644
>--- a/drivers/gpu/drm/xe/xe_migrate.c
>+++ b/drivers/gpu/drm/xe/xe_migrate.c
>@@ -270,7 +270,7 @@ static int xe_migrate_prepare_vm(struct xe_gt *gt, struct xe_migrate *m,
> 		 * Use 1GB pages, it shouldn't matter the physical amount of
> 		 * vram is less, when we don't access it.
> 		 */
>-		for (pos = 0; pos < xe->mem.vram.size; pos += SZ_1G, ofs += 8)
>+		for (pos = 0; pos < xe->mem.vram.io_size; pos += SZ_1G, ofs += 8)
> 			xe_map_wr(xe, &bo->vmap, ofs, u64, pos | flags);
> 	}
>
>diff --git a/drivers/gpu/drm/xe/xe_mmio.c b/drivers/gpu/drm/xe/xe_mmio.c
>index 5536f84682c0..6c1591a4f43e 100644
>--- a/drivers/gpu/drm/xe/xe_mmio.c
>+++ b/drivers/gpu/drm/xe/xe_mmio.c
>@@ -1,6 +1,6 @@
> // SPDX-License-Identifier: MIT
> /*
>- * Copyright © 2021 Intel Corporation
>+ * Copyright © 2021-2023 Intel Corporation
>  */
>
> #include "xe_mmio.h"
>@@ -21,6 +21,8 @@
> #define TILE_COUNT		REG_GENMASK(15, 8)
> #define GEN12_LMEM_BAR		2
>
>+#define BAR_SIZE_SHIFT 20


I had to look at include/linux/pci.h:pci_rebar_bytes_to_size()
to understand what this is about. Not a requirement for this patch, but
since it's something from the resizable BAR specification, maybe it
could deserve a define in include/linux/pci.h rather than hardcoding
in the drivers?

>+
> static int xe_set_dma_info(struct xe_device *xe)
> {
> 	unsigned int mask_size = xe->info.dma_mask_size;
>@@ -61,50 +63,65 @@ _resize_bar(struct xe_device *xe, int resno, resource_size_t size)
> 	if (ret) {
> 		drm_info(&xe->drm, "Failed to resize BAR%d to %dM (%pe). Consider enabling 'Resizable BAR' support in your BIOS\n",
> 			 resno, 1 << bar_size, ERR_PTR(ret));
>-		return -1;
>+		return ret;
> 	}
>
> 	drm_info(&xe->drm, "BAR%d resized to %dM\n", resno, 1 << bar_size);
>-	return 1;
>+	return ret;
> }
>
>-static int xe_resize_vram_bar(struct xe_device *xe, resource_size_t vram_size)
>+/*
>+ * if force_vram_bar_size is set, attempt to set to the requested size
>+ * else set to maximum possible size
>+ */
>+static int xe_resize_vram_bar(struct xe_device *xe)
> {
>+	u64 force_vram_bar_size = xe_force_vram_bar_size;
> 	struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
> 	struct pci_bus *root = pdev->bus;
>-	struct resource *root_res;
>-	resource_size_t rebar_size;
> 	resource_size_t current_size;
>+	resource_size_t rebar_size;
>+	struct resource *root_res;
>+	u32 bar_sizes;
> 	u32 pci_cmd;
> 	int i;
> 	int ret;
>-	u64 force_vram_bar_size = xe_force_vram_bar_size;
>
>-	current_size = roundup_pow_of_two(pci_resource_len(pdev, GEN12_LMEM_BAR));
>+	/* gather some relevant info */
>+	current_size = pci_resource_len(pdev, GEN12_LMEM_BAR);
>+	bar_sizes = pci_rebar_get_possible_sizes(pdev, GEN12_LMEM_BAR);


presumably this already has the >> 20 to return it in "BAR size"
quantity.

>+
>+	if (!bar_sizes)
>+		return 0;
>
>+	/* set to a specific size? */
> 	if (force_vram_bar_size) {
>-		u32 bar_sizes;
>+		u32 bar_size;
>
> 		rebar_size = force_vram_bar_size * (resource_size_t)SZ_1M;
>-		bar_sizes = pci_rebar_get_possible_sizes(pdev, GEN12_LMEM_BAR);
>
>-		if (rebar_size == current_size)
>-			return 0;
>+		bar_size = bar_sizes & BIT(pci_rebar_bytes_to_size(rebar_size));

but then why you add BIT() here?

Supposing bar_sizes is 0x3 (i.e. valid bar sizes in bytes are 2097152
and 4194304) and rebar_size is 2097152, shouldn't this be

		bar_size = bar_sizes & pci_rebar_bytes_to_size(rebar_size);
?

/me confused. Maybe rename the variables that are measurable in bytes to
avoid some of the confusion?


Lucas De Marchi

>
>-		if (!(bar_sizes & BIT(pci_rebar_bytes_to_size(rebar_size))) ||
>-		    rebar_size >= roundup_pow_of_two(vram_size)) {
>-			rebar_size = vram_size;
>+		if (!bar_size) {
> 			drm_info(&xe->drm,
>-				 "Given bar size is not within supported size, setting it to default: %lluMiB\n",
>-				 (u64)vram_size >> 20);
>+				 "Requested size: 0x%llx is not supported by rebar sizes: 0x%x. Leaving default: 0x%llx\n",
>+				 (u64)rebar_size >> 2, bar_sizes, (u64)current_size >> 20);
>+			return 0;
> 		}
>+
>+		rebar_size = 1ULL << (bar_size + BAR_SIZE_SHIFT);
>+
>+		if (rebar_size == current_size)
>+			return 0;
> 	} else {
>-		rebar_size = current_size;
>+		rebar_size = 1ULL << (__fls(bar_sizes) + BAR_SIZE_SHIFT);
>
>-		if (rebar_size != roundup_pow_of_two(vram_size))
>-			rebar_size = vram_size;
>-		else
>+		/* only resize if larger than current */
>+		if (rebar_size <= current_size) {
>+			drm_info(&xe->drm, "Rebar size: 0x%llx vs. actual size: 0x%llx\n",
>+				 rebar_size, current_size);
> 			return 0;
>+		}
> 	}
>
> 	drm_info(&xe->drm, "Resizing bar from %lluMiB -> %lluMiB\n",
>@@ -148,49 +165,39 @@ static bool xe_pci_resource_valid(struct pci_dev *pdev, int bar)
> 	return true;
> }
>
>-int xe_mmio_total_vram_size(struct xe_device *xe, u64 *vram_size, u64 *usable_size)
>+int xe_determine_lmem_bar_size(struct xe_device *xe)
> {
>-	struct xe_gt *gt = xe_device_get_gt(xe, 0);
> 	struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
> 	int err;
>-	u32 reg;
>
>-	if (!xe->info.has_flat_ccs)  {
>-		*vram_size = pci_resource_len(pdev, GEN12_LMEM_BAR);
>-		if (usable_size)
>-			*usable_size = min(*vram_size, xe_mmio_read64(gt, GEN12_GSMBASE.reg));
>-		return 0;
>+	if (!xe_pci_resource_valid(pdev, GEN12_LMEM_BAR)) {
>+		drm_err(&xe->drm, "pci resource is not valid\n");
>+		return -ENXIO;
> 	}
>
>-	err = xe_force_wake_get(gt_to_fw(gt), XE_FW_GT);
>+	err = xe_resize_vram_bar(xe);
> 	if (err)
> 		return err;
>
>-	reg = xe_gt_mcr_unicast_read_any(gt, XEHP_TILE0_ADDR_RANGE);
>-	*vram_size = (u64)REG_FIELD_GET(GENMASK(14, 8), reg) * SZ_1G;
>-	if (usable_size) {
>-		reg = xe_gt_mcr_unicast_read_any(gt, XEHP_FLAT_CCS_BASE_ADDR);
>-		*usable_size = (u64)REG_FIELD_GET(GENMASK(31, 8), reg) * SZ_64K;
>-		drm_info(&xe->drm, "vram_size: 0x%llx usable_size: 0x%llx\n",
>-			 *vram_size, *usable_size);
>-	}
>+	xe->mem.vram.io_start = pci_resource_start(pdev, GEN12_LMEM_BAR);
>+	xe->mem.vram.io_size = pci_resource_len(pdev, GEN12_LMEM_BAR);
>+	if (!xe->mem.vram.io_size)
>+		return -EIO;
>+
>+	/* set up a map to the total memory area. */
>+	xe->mem.vram.mapping = ioremap_wc(xe->mem.vram.io_start, xe->mem.vram.io_size);
>
>-	return xe_force_wake_put(gt_to_fw(gt), XE_FW_GT);
>+	return 0;
> }
>
> int xe_mmio_probe_vram(struct xe_device *xe)
> {
>-	struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
> 	struct xe_gt *gt;
>-	u8 id;
>-	u64 vram_size;
>-	u64 original_size;
>-	u64 usable_size;
> 	int err;
>+	u8 id;
>
> 	if (!IS_DGFX(xe)) {
> 		xe->mem.vram.mapping = 0;
>-		xe->mem.vram.size = 0;
> 		xe->mem.vram.io_start = 0;
> 		xe->mem.vram.io_size = 0;
>
>@@ -203,35 +210,12 @@ int xe_mmio_probe_vram(struct xe_device *xe)
> 		return 0;
> 	}
>
>-	if (!xe_pci_resource_valid(pdev, GEN12_LMEM_BAR)) {
>-		drm_err(&xe->drm, "pci resource is not valid\n");
>-		return -ENXIO;
>-	}
>-
>-	gt = xe_device_get_gt(xe, 0);
>-	original_size = pci_resource_len(pdev, GEN12_LMEM_BAR);
>-
>-	err = xe_mmio_total_vram_size(xe, &vram_size, &usable_size);
>+	err = xe_determine_lmem_bar_size(xe);
> 	if (err)
> 		return err;
>
>-	xe_resize_vram_bar(xe, vram_size);
>-	xe->mem.vram.io_start = pci_resource_start(pdev, GEN12_LMEM_BAR);
>-	xe->mem.vram.io_size = min(usable_size,
>-				   pci_resource_len(pdev, GEN12_LMEM_BAR));
>-	xe->mem.vram.size = xe->mem.vram.io_size;
>-
>-	if (!xe->mem.vram.size)
>-		return -EIO;
>-
>-	if (usable_size > xe->mem.vram.io_size)
>-		drm_warn(&xe->drm, "Restricting VRAM size to PCI resource size (%lluMiB->%lluMiB)\n",
>-			 (u64)usable_size >> 20, (u64)xe->mem.vram.io_size >> 20);
>-
>-	xe->mem.vram.mapping = ioremap_wc(xe->mem.vram.io_start, xe->mem.vram.io_size);
>-	xe->mem.vram.size = min_t(u64, xe->mem.vram.size, usable_size);
>-
>-	drm_info(&xe->drm, "TOTAL VRAM: %pa, %pa\n", &xe->mem.vram.io_start, &xe->mem.vram.size);
>+	drm_info(&xe->drm, "VISIBLE VRAM: %pa, %pa\n", &xe->mem.vram.io_start,
>+		 &xe->mem.vram.io_size);
>
> 	/* FIXME: Assuming equally partitioned VRAM, incorrect */
> 	if (xe->info.tile_count > 1) {
>@@ -244,7 +228,7 @@ int xe_mmio_probe_vram(struct xe_device *xe)
>
> 		XE_BUG_ON(!adj_tile_count);
>
>-		size = xe->mem.vram.size / adj_tile_count;
>+		size = xe->mem.vram.io_size / adj_tile_count;
> 		io_start = xe->mem.vram.io_start;
> 		io_size = xe->mem.vram.io_size;
>
>@@ -277,7 +261,7 @@ int xe_mmio_probe_vram(struct xe_device *xe)
> 				 &gt->mem.vram.size);
> 		}
> 	} else {
>-		gt->mem.vram.size = xe->mem.vram.size;
>+		gt->mem.vram.size = xe->mem.vram.io_size;
> 		gt->mem.vram.io_start = xe->mem.vram.io_start;
> 		gt->mem.vram.io_size = xe->mem.vram.io_size;
> 		gt->mem.vram.mapping = xe->mem.vram.mapping;
>diff --git a/drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c b/drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
>index 31887fec1073..9136c035db0e 100644
>--- a/drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
>+++ b/drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
>@@ -53,25 +53,19 @@ bool xe_ttm_stolen_cpu_access_needs_ggtt(struct xe_device *xe)
>
> static s64 detect_bar2_dgfx(struct xe_device *xe, struct xe_ttm_stolen_mgr *mgr)
> {
>-	struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
>-	struct xe_gt *gt = to_gt(xe);
> 	u64 vram_size, stolen_size;
>-	int err;
>
>-	err = xe_mmio_total_vram_size(xe, &vram_size, NULL);
>-	if (err) {
>-		drm_info(&xe->drm, "Querying total vram size failed\n");
>-		return 0;
>-	}
>+	vram_size = xe->mem.vram.io_size;
>
> 	/* Use DSM base address instead for stolen memory */
>-	mgr->stolen_base = xe_mmio_read64(gt, GEN12_DSMBASE.reg) & GEN12_BDSM_MASK;
>+	mgr->stolen_base = xe_mmio_read64(to_gt(xe), GEN12_DSMBASE.reg) & GEN12_BDSM_MASK;
> 	if (drm_WARN_ON(&xe->drm, vram_size < mgr->stolen_base))
> 		return 0;
>
> 	stolen_size = vram_size - mgr->stolen_base;
>-	if (mgr->stolen_base + stolen_size <= pci_resource_len(pdev, 2))
>-		mgr->io_base = pci_resource_start(pdev, 2) + mgr->stolen_base;
>+
>+	if (mgr->stolen_base + stolen_size <= vram_size)
>+		mgr->io_base = xe->mem.vram.io_start + mgr->stolen_base;
>
> 	/*
> 	 * There may be few KB of platform dependent reserved memory at the end
>-- 
>2.39.2
>


More information about the Intel-xe mailing list