[Intel-gfx] [PATCH v5 1/2] drm/i915/gt: Split intel-gtt functions by arch

Lucas De Marchi lucas.demarchi at intel.com
Wed Apr 6 20:19:19 UTC 2022


On Tue, Apr 05, 2022 at 11:05:55AM -0700, Casey Bowman wrote:
>@Jani/Lucas, any other thoughts here?

This is now applied. Thanks.

Tvrtko / Jani / Rodrigo,

There was a small conflict due to s/intel_vtd_active/i915_vtd_active/
in drm-intel-next. I added a fixup to drm-rerere.

Lucas De Marchi


>
>Regards,
>Casey
>
>On 3/31/22 07:36, Tvrtko Ursulin wrote:
>>
>>On 31/03/2022 00:48, Casey Bowman wrote:
>>>Some functions defined in the intel-gtt module are used in several
>>>areas, but is only supported on x86 platforms.
>>>
>>>By separating these calls and their static underlying functions to
>>>another area, we are able to compile out these functions for
>>>non-x86 builds and provide stubs for the non-x86 implementations.
>>>
>>>In addition to the problematic calls, we are moving the gmch-related
>>>functions to the new area.
>>
>>Glanced over the latest version - looks acceptable to me.
>>
>>Acked-by: Tvrtko Ursulin <tvrtko.ursulin at intel.com>
>>
>>Regards,
>>
>>Tvrtko
>>
>>>Signed-off-by: Casey Bowman <casey.g.bowman at intel.com>
>>>---
>>>  drivers/gpu/drm/i915/Makefile           |   2 +
>>>  drivers/gpu/drm/i915/gt/intel_ggtt.c    | 664 +-----------------------
>>>  drivers/gpu/drm/i915/gt/intel_gt.c      |   4 +-
>>>  drivers/gpu/drm/i915/gt/intel_gt.h      |   9 +
>>>  drivers/gpu/drm/i915/gt/intel_gt_gmch.c | 654 +++++++++++++++++++++++
>>>  drivers/gpu/drm/i915/gt/intel_gt_gmch.h |  46 ++
>>>  drivers/gpu/drm/i915/gt/intel_gtt.h     |   9 +
>>>  7 files changed, 736 insertions(+), 652 deletions(-)
>>>  create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_gmch.c
>>>  create mode 100644 drivers/gpu/drm/i915/gt/intel_gt_gmch.h
>>>
>>>diff --git a/drivers/gpu/drm/i915/Makefile 
>>>b/drivers/gpu/drm/i915/Makefile
>>>index c1d5540f6052..30b75b5895ae 100644
>>>--- a/drivers/gpu/drm/i915/Makefile
>>>+++ b/drivers/gpu/drm/i915/Makefile
>>>@@ -126,6 +126,8 @@ gt-y += \
>>>      gt/intel_workarounds.o \
>>>      gt/shmem_utils.o \
>>>      gt/sysfs_engines.o
>>>+# x86 intel-gtt module support
>>>+gt-$(CONFIG_X86) += gt/intel_gt_gmch.o
>>>  # autogenerated null render state
>>>  gt-y += \
>>>      gt/gen6_renderstate.o \
>>>diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c 
>>>b/drivers/gpu/drm/i915/gt/intel_ggtt.c
>>>index 86b2cd2a9f34..be25375ac6f1 100644
>>>--- a/drivers/gpu/drm/i915/gt/intel_ggtt.c
>>>+++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c
>>>@@ -3,22 +3,19 @@
>>>   * Copyright © 2020 Intel Corporation
>>>   */
>>>  -#include <linux/agp_backend.h>
>>>-#include <linux/stop_machine.h>
>>>-
>>>+#include <linux/types.h>
>>>  #include <asm/set_memory.h>
>>>  #include <asm/smp.h>
>>>    #include <drm/i915_drm.h>
>>>-#include <drm/intel-gtt.h>
>>>    #include "gem/i915_gem_lmem.h"
>>>    #include "intel_gt.h"
>>>+#include "intel_gt_gmch.h"
>>>  #include "intel_gt_regs.h"
>>>  #include "i915_drv.h"
>>>  #include "i915_scatterlist.h"
>>>-#include "i915_utils.h"
>>>  #include "i915_vgpu.h"
>>>    #include "intel_gtt.h"
>>>@@ -95,28 +92,6 @@ int i915_ggtt_init_hw(struct drm_i915_private *i915)
>>>      return 0;
>>>  }
>>>  -/*
>>>- * Certain Gen5 chipsets require idling the GPU before
>>>- * unmapping anything from the GTT when VT-d is enabled.
>>>- */
>>>-static bool needs_idle_maps(struct drm_i915_private *i915)
>>>-{
>>>-    /*
>>>-     * Query intel_iommu to see if we need the workaround. 
>>>Presumably that
>>>-     * was loaded first.
>>>-     */
>>>-    if (!i915_vtd_active(i915))
>>>-        return false;
>>>-
>>>-    if (GRAPHICS_VER(i915) == 5 && IS_MOBILE(i915))
>>>-        return true;
>>>-
>>>-    if (GRAPHICS_VER(i915) == 12)
>>>-        return true; /* XXX DMAR fault reason 7 */
>>>-
>>>-    return false;
>>>-}
>>>-
>>>  /**
>>>   * i915_ggtt_suspend_vm - Suspend the memory mappings for a GGTT 
>>>or DPT VM
>>>   * @vm: The VM to suspend the mappings for
>>>@@ -205,7 +180,7 @@ void gen6_ggtt_invalidate(struct i915_ggtt *ggtt)
>>>      spin_unlock_irq(&uncore->lock);
>>>  }
>>>  -static void gen8_ggtt_invalidate(struct i915_ggtt *ggtt)
>>>+void gen8_ggtt_invalidate(struct i915_ggtt *ggtt)
>>>  {
>>>      struct intel_uncore *uncore = ggtt->vm.gt->uncore;
>>>  @@ -230,11 +205,6 @@ static void guc_ggtt_invalidate(struct 
>>>i915_ggtt *ggtt)
>>>          intel_uncore_write_fw(uncore, GEN8_GTCR, 
>>>GEN8_GTCR_INVALIDATE);
>>>  }
>>>  -static void gmch_ggtt_invalidate(struct i915_ggtt *ggtt)
>>>-{
>>>-    intel_gtt_chipset_flush();
>>>-}
>>>-
>>>  u64 gen8_ggtt_pte_encode(dma_addr_t addr,
>>>               enum i915_cache_level level,
>>>               u32 flags)
>>>@@ -247,258 +217,7 @@ u64 gen8_ggtt_pte_encode(dma_addr_t addr,
>>>      return pte;
>>>  }
>>>  -static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte)
>>>-{
>>>-    writeq(pte, addr);
>>>-}
>>>-
>>>-static void gen8_ggtt_insert_page(struct i915_address_space *vm,
>>>-                  dma_addr_t addr,
>>>-                  u64 offset,
>>>-                  enum i915_cache_level level,
>>>-                  u32 flags)
>>>-{
>>>-    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>-    gen8_pte_t __iomem *pte =
>>>-        (gen8_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE;
>>>-
>>>-    gen8_set_pte(pte, gen8_ggtt_pte_encode(addr, level, flags));
>>>-
>>>-    ggtt->invalidate(ggtt);
>>>-}
>>>-
>>>-static void gen8_ggtt_insert_entries(struct i915_address_space *vm,
>>>-                     struct i915_vma_resource *vma_res,
>>>-                     enum i915_cache_level level,
>>>-                     u32 flags)
>>>-{
>>>-    const gen8_pte_t pte_encode = gen8_ggtt_pte_encode(0, level, 
>>>flags);
>>>-    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>-    gen8_pte_t __iomem *gte;
>>>-    gen8_pte_t __iomem *end;
>>>-    struct sgt_iter iter;
>>>-    dma_addr_t addr;
>>>-
>>>-    /*
>>>-     * Note that we ignore PTE_READ_ONLY here. The caller must be 
>>>careful
>>>-     * not to allow the user to override access to a read only page.
>>>-     */
>>>-
>>>-    gte = (gen8_pte_t __iomem *)ggtt->gsm;
>>>-    gte += vma_res->start / I915_GTT_PAGE_SIZE;
>>>-    end = gte + vma_res->node_size / I915_GTT_PAGE_SIZE;
>>>-
>>>-    for_each_sgt_daddr(addr, iter, vma_res->bi.pages)
>>>-        gen8_set_pte(gte++, pte_encode | addr);
>>>-    GEM_BUG_ON(gte > end);
>>>-
>>>-    /* Fill the allocated but "unused" space beyond the end of 
>>>the buffer */
>>>-    while (gte < end)
>>>-        gen8_set_pte(gte++, vm->scratch[0]->encode);
>>>-
>>>-    /*
>>>-     * We want to flush the TLBs only after we're certain all the PTE
>>>-     * updates have finished.
>>>-     */
>>>-    ggtt->invalidate(ggtt);
>>>-}
>>>-
>>>-static void gen6_ggtt_insert_page(struct i915_address_space *vm,
>>>-                  dma_addr_t addr,
>>>-                  u64 offset,
>>>-                  enum i915_cache_level level,
>>>-                  u32 flags)
>>>-{
>>>-    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>-    gen6_pte_t __iomem *pte =
>>>-        (gen6_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE;
>>>-
>>>-    iowrite32(vm->pte_encode(addr, level, flags), pte);
>>>-
>>>-    ggtt->invalidate(ggtt);
>>>-}
>>>-
>>>-/*
>>>- * Binds an object into the global gtt with the specified cache level.
>>>- * The object will be accessible to the GPU via commands whose operands
>>>- * reference offsets within the global GTT as well as accessible 
>>>by the GPU
>>>- * through the GMADR mapped BAR (i915->mm.gtt->gtt).
>>>- */
>>>-static void gen6_ggtt_insert_entries(struct i915_address_space *vm,
>>>-                     struct i915_vma_resource *vma_res,
>>>-                     enum i915_cache_level level,
>>>-                     u32 flags)
>>>-{
>>>-    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>-    gen6_pte_t __iomem *gte;
>>>-    gen6_pte_t __iomem *end;
>>>-    struct sgt_iter iter;
>>>-    dma_addr_t addr;
>>>-
>>>-    gte = (gen6_pte_t __iomem *)ggtt->gsm;
>>>-    gte += vma_res->start / I915_GTT_PAGE_SIZE;
>>>-    end = gte + vma_res->node_size / I915_GTT_PAGE_SIZE;
>>>-
>>>-    for_each_sgt_daddr(addr, iter, vma_res->bi.pages)
>>>-        iowrite32(vm->pte_encode(addr, level, flags), gte++);
>>>-    GEM_BUG_ON(gte > end);
>>>-
>>>-    /* Fill the allocated but "unused" space beyond the end of 
>>>the buffer */
>>>-    while (gte < end)
>>>-        iowrite32(vm->scratch[0]->encode, gte++);
>>>-
>>>-    /*
>>>-     * We want to flush the TLBs only after we're certain all the PTE
>>>-     * updates have finished.
>>>-     */
>>>-    ggtt->invalidate(ggtt);
>>>-}
>>>-
>>>-static void nop_clear_range(struct i915_address_space *vm,
>>>-                u64 start, u64 length)
>>>-{
>>>-}
>>>-
>>>-static void gen8_ggtt_clear_range(struct i915_address_space *vm,
>>>-                  u64 start, u64 length)
>>>-{
>>>-    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>-    unsigned int first_entry = start / I915_GTT_PAGE_SIZE;
>>>-    unsigned int num_entries = length / I915_GTT_PAGE_SIZE;
>>>-    const gen8_pte_t scratch_pte = vm->scratch[0]->encode;
>>>-    gen8_pte_t __iomem *gtt_base =
>>>-        (gen8_pte_t __iomem *)ggtt->gsm + first_entry;
>>>-    const int max_entries = ggtt_total_entries(ggtt) - first_entry;
>>>-    int i;
>>>-
>>>-    if (WARN(num_entries > max_entries,
>>>-         "First entry = %d; Num entries = %d (max=%d)\n",
>>>-         first_entry, num_entries, max_entries))
>>>-        num_entries = max_entries;
>>>-
>>>-    for (i = 0; i < num_entries; i++)
>>>-        gen8_set_pte(&gtt_base[i], scratch_pte);
>>>-}
>>>-
>>>-static void bxt_vtd_ggtt_wa(struct i915_address_space *vm)
>>>-{
>>>-    /*
>>>-     * Make sure the internal GAM fifo has been cleared of all GTT
>>>-     * writes before exiting stop_machine(). This guarantees that
>>>-     * any aperture accesses waiting to start in another process
>>>-     * cannot back up behind the GTT writes causing a hang.
>>>-     * The register can be any arbitrary GAM register.
>>>-     */
>>>-    intel_uncore_posting_read_fw(vm->gt->uncore, GFX_FLSH_CNTL_GEN6);
>>>-}
>>>-
>>>-struct insert_page {
>>>-    struct i915_address_space *vm;
>>>-    dma_addr_t addr;
>>>-    u64 offset;
>>>-    enum i915_cache_level level;
>>>-};
>>>-
>>>-static int bxt_vtd_ggtt_insert_page__cb(void *_arg)
>>>-{
>>>-    struct insert_page *arg = _arg;
>>>-
>>>-    gen8_ggtt_insert_page(arg->vm, arg->addr, arg->offset, 
>>>arg->level, 0);
>>>-    bxt_vtd_ggtt_wa(arg->vm);
>>>-
>>>-    return 0;
>>>-}
>>>-
>>>-static void bxt_vtd_ggtt_insert_page__BKL(struct 
>>>i915_address_space *vm,
>>>-                      dma_addr_t addr,
>>>-                      u64 offset,
>>>-                      enum i915_cache_level level,
>>>-                      u32 unused)
>>>-{
>>>-    struct insert_page arg = { vm, addr, offset, level };
>>>-
>>>-    stop_machine(bxt_vtd_ggtt_insert_page__cb, &arg, NULL);
>>>-}
>>>-
>>>-struct insert_entries {
>>>-    struct i915_address_space *vm;
>>>-    struct i915_vma_resource *vma_res;
>>>-    enum i915_cache_level level;
>>>-    u32 flags;
>>>-};
>>>-
>>>-static int bxt_vtd_ggtt_insert_entries__cb(void *_arg)
>>>-{
>>>-    struct insert_entries *arg = _arg;
>>>-
>>>-    gen8_ggtt_insert_entries(arg->vm, arg->vma_res, arg->level, 
>>>arg->flags);
>>>-    bxt_vtd_ggtt_wa(arg->vm);
>>>-
>>>-    return 0;
>>>-}
>>>-
>>>-static void bxt_vtd_ggtt_insert_entries__BKL(struct 
>>>i915_address_space *vm,
>>>-                         struct i915_vma_resource *vma_res,
>>>-                         enum i915_cache_level level,
>>>-                         u32 flags)
>>>-{
>>>-    struct insert_entries arg = { vm, vma_res, level, flags };
>>>-
>>>-    stop_machine(bxt_vtd_ggtt_insert_entries__cb, &arg, NULL);
>>>-}
>>>-
>>>-static void gen6_ggtt_clear_range(struct i915_address_space *vm,
>>>-                  u64 start, u64 length)
>>>-{
>>>-    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>-    unsigned int first_entry = start / I915_GTT_PAGE_SIZE;
>>>-    unsigned int num_entries = length / I915_GTT_PAGE_SIZE;
>>>-    gen6_pte_t scratch_pte, __iomem *gtt_base =
>>>-        (gen6_pte_t __iomem *)ggtt->gsm + first_entry;
>>>-    const int max_entries = ggtt_total_entries(ggtt) - first_entry;
>>>-    int i;
>>>-
>>>-    if (WARN(num_entries > max_entries,
>>>-         "First entry = %d; Num entries = %d (max=%d)\n",
>>>-         first_entry, num_entries, max_entries))
>>>-        num_entries = max_entries;
>>>-
>>>-    scratch_pte = vm->scratch[0]->encode;
>>>-    for (i = 0; i < num_entries; i++)
>>>-        iowrite32(scratch_pte, &gtt_base[i]);
>>>-}
>>>-
>>>-static void i915_ggtt_insert_page(struct i915_address_space *vm,
>>>-                  dma_addr_t addr,
>>>-                  u64 offset,
>>>-                  enum i915_cache_level cache_level,
>>>-                  u32 unused)
>>>-{
>>>-    unsigned int flags = (cache_level == I915_CACHE_NONE) ?
>>>-        AGP_USER_MEMORY : AGP_USER_CACHED_MEMORY;
>>>-
>>>-    intel_gtt_insert_page(addr, offset >> PAGE_SHIFT, flags);
>>>-}
>>>-
>>>-static void i915_ggtt_insert_entries(struct i915_address_space *vm,
>>>-                     struct i915_vma_resource *vma_res,
>>>-                     enum i915_cache_level cache_level,
>>>-                     u32 unused)
>>>-{
>>>-    unsigned int flags = (cache_level == I915_CACHE_NONE) ?
>>>-        AGP_USER_MEMORY : AGP_USER_CACHED_MEMORY;
>>>-
>>>-    intel_gtt_insert_sg_entries(vma_res->bi.pages, vma_res->start 
>>>>> PAGE_SHIFT,
>>>-                    flags);
>>>-}
>>>-
>>>-static void i915_ggtt_clear_range(struct i915_address_space *vm,
>>>-                  u64 start, u64 length)
>>>-{
>>>-    intel_gtt_clear_range(start >> PAGE_SHIFT, length >> PAGE_SHIFT);
>>>-}
>>>-
>>>-static void ggtt_bind_vma(struct i915_address_space *vm,
>>>+void intel_ggtt_bind_vma(struct i915_address_space *vm,
>>>                struct i915_vm_pt_stash *stash,
>>>                struct i915_vma_resource *vma_res,
>>>                enum i915_cache_level cache_level,
>>>@@ -522,7 +241,7 @@ static void ggtt_bind_vma(struct 
>>>i915_address_space *vm,
>>>      vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE;
>>>  }
>>>  -static void ggtt_unbind_vma(struct i915_address_space *vm,
>>>+void intel_ggtt_unbind_vma(struct i915_address_space *vm,
>>>                  struct i915_vma_resource *vma_res)
>>>  {
>>>      vm->clear_range(vm, vma_res->start, vma_res->vma_size);
>>>@@ -725,10 +444,10 @@ static int init_aliasing_ppgtt(struct 
>>>i915_ggtt *ggtt)
>>>      ggtt->alias = ppgtt;
>>>      ggtt->vm.bind_async_flags |= ppgtt->vm.bind_async_flags;
>>>  -    GEM_BUG_ON(ggtt->vm.vma_ops.bind_vma != ggtt_bind_vma);
>>>+    GEM_BUG_ON(ggtt->vm.vma_ops.bind_vma != intel_ggtt_bind_vma);
>>>      ggtt->vm.vma_ops.bind_vma = aliasing_gtt_bind_vma;
>>>  -    GEM_BUG_ON(ggtt->vm.vma_ops.unbind_vma != ggtt_unbind_vma);
>>>+    GEM_BUG_ON(ggtt->vm.vma_ops.unbind_vma != intel_ggtt_unbind_vma);
>>>      ggtt->vm.vma_ops.unbind_vma = aliasing_gtt_unbind_vma;
>>>        i915_vm_free_pt_stash(&ppgtt->vm, &stash);
>>>@@ -751,8 +470,8 @@ static void fini_aliasing_ppgtt(struct 
>>>i915_ggtt *ggtt)
>>>        i915_vm_put(&ppgtt->vm);
>>>  -    ggtt->vm.vma_ops.bind_vma   = ggtt_bind_vma;
>>>-    ggtt->vm.vma_ops.unbind_vma = ggtt_unbind_vma;
>>>+    ggtt->vm.vma_ops.bind_vma   = intel_ggtt_bind_vma;
>>>+    ggtt->vm.vma_ops.unbind_vma = intel_ggtt_unbind_vma;
>>>  }
>>>    int i915_init_ggtt(struct drm_i915_private *i915)
>>>@@ -840,364 +559,12 @@ void i915_ggtt_driver_late_release(struct 
>>>drm_i915_private *i915)
>>>      dma_resv_fini(&ggtt->vm._resv);
>>>  }
>>>  -static unsigned int gen6_get_total_gtt_size(u16 snb_gmch_ctl)
>>>-{
>>>-    snb_gmch_ctl >>= SNB_GMCH_GGMS_SHIFT;
>>>-    snb_gmch_ctl &= SNB_GMCH_GGMS_MASK;
>>>-    return snb_gmch_ctl << 20;
>>>-}
>>>-
>>>-static unsigned int gen8_get_total_gtt_size(u16 bdw_gmch_ctl)
>>>-{
>>>-    bdw_gmch_ctl >>= BDW_GMCH_GGMS_SHIFT;
>>>-    bdw_gmch_ctl &= BDW_GMCH_GGMS_MASK;
>>>-    if (bdw_gmch_ctl)
>>>-        bdw_gmch_ctl = 1 << bdw_gmch_ctl;
>>>-
>>>-#ifdef CONFIG_X86_32
>>>-    /* Limit 32b platforms to a 2GB GGTT: 4 << 20 / pte size * 
>>>I915_GTT_PAGE_SIZE */
>>>-    if (bdw_gmch_ctl > 4)
>>>-        bdw_gmch_ctl = 4;
>>>-#endif
>>>-
>>>-    return bdw_gmch_ctl << 20;
>>>-}
>>>-
>>>-static unsigned int chv_get_total_gtt_size(u16 gmch_ctrl)
>>>-{
>>>-    gmch_ctrl >>= SNB_GMCH_GGMS_SHIFT;
>>>-    gmch_ctrl &= SNB_GMCH_GGMS_MASK;
>>>-
>>>-    if (gmch_ctrl)
>>>-        return 1 << (20 + gmch_ctrl);
>>>-
>>>-    return 0;
>>>-}
>>>-
>>>-static unsigned int gen6_gttmmadr_size(struct drm_i915_private *i915)
>>>-{
>>>-    /*
>>>-     * GEN6: GTTMMADR size is 4MB and GTTADR starts at 2MB offset
>>>-     * GEN8: GTTMMADR size is 16MB and GTTADR starts at 8MB offset
>>>-     */
>>>-    GEM_BUG_ON(GRAPHICS_VER(i915) < 6);
>>>-    return (GRAPHICS_VER(i915) < 8) ? SZ_4M : SZ_16M;
>>>-}
>>>-
>>>-static unsigned int gen6_gttadr_offset(struct drm_i915_private *i915)
>>>-{
>>>-    return gen6_gttmmadr_size(i915) / 2;
>>>-}
>>>-
>>>-static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
>>>-{
>>>-    struct drm_i915_private *i915 = ggtt->vm.i915;
>>>-    struct pci_dev *pdev = to_pci_dev(i915->drm.dev);
>>>-    phys_addr_t phys_addr;
>>>-    u32 pte_flags;
>>>-    int ret;
>>>-
>>>-    GEM_WARN_ON(pci_resource_len(pdev, 0) != gen6_gttmmadr_size(i915));
>>>-    phys_addr = pci_resource_start(pdev, 0) + gen6_gttadr_offset(i915);
>>>-
>>>-    /*
>>>-     * On BXT+/ICL+ writes larger than 64 bit to the GTT 
>>>pagetable range
>>>-     * will be dropped. For WC mappings in general we have 64 
>>>byte burst
>>>-     * writes when the WC buffer is flushed, so we can't use it, 
>>>but have to
>>>-     * resort to an uncached mapping. The WC issue is easily 
>>>caught by the
>>>-     * readback check when writing GTT PTE entries.
>>>-     */
>>>-    if (IS_GEN9_LP(i915) || GRAPHICS_VER(i915) >= 11)
>>>-        ggtt->gsm = ioremap(phys_addr, size);
>>>-    else
>>>-        ggtt->gsm = ioremap_wc(phys_addr, size);
>>>-    if (!ggtt->gsm) {
>>>-        drm_err(&i915->drm, "Failed to map the ggtt page table\n");
>>>-        return -ENOMEM;
>>>-    }
>>>-
>>>-    kref_init(&ggtt->vm.resv_ref);
>>>-    ret = setup_scratch_page(&ggtt->vm);
>>>-    if (ret) {
>>>-        drm_err(&i915->drm, "Scratch setup failed\n");
>>>-        /* iounmap will also get called at remove, but meh */
>>>-        iounmap(ggtt->gsm);
>>>-        return ret;
>>>-    }
>>>-
>>>-    pte_flags = 0;
>>>-    if (i915_gem_object_is_lmem(ggtt->vm.scratch[0]))
>>>-        pte_flags |= PTE_LM;
>>>-
>>>-    ggtt->vm.scratch[0]->encode =
>>>-        ggtt->vm.pte_encode(px_dma(ggtt->vm.scratch[0]),
>>>-                    I915_CACHE_NONE, pte_flags);
>>>-
>>>-    return 0;
>>>-}
>>>-
>>>-static void gen6_gmch_remove(struct i915_address_space *vm)
>>>-{
>>>-    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>-
>>>-    iounmap(ggtt->gsm);
>>>-    free_scratch(vm);
>>>-}
>>>-
>>>-static struct resource pci_resource(struct pci_dev *pdev, int bar)
>>>+struct resource intel_pci_resource(struct pci_dev *pdev, int bar)
>>>  {
>>>      return (struct 
>>>resource)DEFINE_RES_MEM(pci_resource_start(pdev, bar),
>>>                             pci_resource_len(pdev, bar));
>>>  }
>>>  -static int gen8_gmch_probe(struct i915_ggtt *ggtt)
>>>-{
>>>-    struct drm_i915_private *i915 = ggtt->vm.i915;
>>>-    struct pci_dev *pdev = to_pci_dev(i915->drm.dev);
>>>-    unsigned int size;
>>>-    u16 snb_gmch_ctl;
>>>-
>>>-    /* TODO: We're not aware of mappable constraints on gen8 yet */
>>>-    if (!HAS_LMEM(i915)) {
>>>-        ggtt->gmadr = pci_resource(pdev, 2);
>>>-        ggtt->mappable_end = resource_size(&ggtt->gmadr);
>>>-    }
>>>-
>>>-    pci_read_config_word(pdev, SNB_GMCH_CTRL, &snb_gmch_ctl);
>>>-    if (IS_CHERRYVIEW(i915))
>>>-        size = chv_get_total_gtt_size(snb_gmch_ctl);
>>>-    else
>>>-        size = gen8_get_total_gtt_size(snb_gmch_ctl);
>>>-
>>>-    ggtt->vm.alloc_pt_dma = alloc_pt_dma;
>>>-    ggtt->vm.alloc_scratch_dma = alloc_pt_dma;
>>>-    ggtt->vm.lmem_pt_obj_flags = I915_BO_ALLOC_PM_EARLY;
>>>-
>>>-    ggtt->vm.total = (size / sizeof(gen8_pte_t)) * I915_GTT_PAGE_SIZE;
>>>-    ggtt->vm.cleanup = gen6_gmch_remove;
>>>-    ggtt->vm.insert_page = gen8_ggtt_insert_page;
>>>-    ggtt->vm.clear_range = nop_clear_range;
>>>-    if (intel_scanout_needs_vtd_wa(i915))
>>>-        ggtt->vm.clear_range = gen8_ggtt_clear_range;
>>>-
>>>-    ggtt->vm.insert_entries = gen8_ggtt_insert_entries;
>>>-
>>>-    /*
>>>-     * Serialize GTT updates with aperture access on BXT if VT-d is on,
>>>-     * and always on CHV.
>>>-     */
>>>-    if (intel_vm_no_concurrent_access_wa(i915)) {
>>>-        ggtt->vm.insert_entries = bxt_vtd_ggtt_insert_entries__BKL;
>>>-        ggtt->vm.insert_page    = bxt_vtd_ggtt_insert_page__BKL;
>>>-        ggtt->vm.bind_async_flags =
>>>-            I915_VMA_GLOBAL_BIND | I915_VMA_LOCAL_BIND;
>>>-    }
>>>-
>>>-    ggtt->invalidate = gen8_ggtt_invalidate;
>>>-
>>>-    ggtt->vm.vma_ops.bind_vma    = ggtt_bind_vma;
>>>-    ggtt->vm.vma_ops.unbind_vma  = ggtt_unbind_vma;
>>>-
>>>-    ggtt->vm.pte_encode = gen8_ggtt_pte_encode;
>>>-
>>>-    setup_private_pat(ggtt->vm.gt->uncore);
>>>-
>>>-    return ggtt_probe_common(ggtt, size);
>>>-}
>>>-
>>>-static u64 snb_pte_encode(dma_addr_t addr,
>>>-              enum i915_cache_level level,
>>>-              u32 flags)
>>>-{
>>>-    gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>-
>>>-    switch (level) {
>>>-    case I915_CACHE_L3_LLC:
>>>-    case I915_CACHE_LLC:
>>>-        pte |= GEN6_PTE_CACHE_LLC;
>>>-        break;
>>>-    case I915_CACHE_NONE:
>>>-        pte |= GEN6_PTE_UNCACHED;
>>>-        break;
>>>-    default:
>>>-        MISSING_CASE(level);
>>>-    }
>>>-
>>>-    return pte;
>>>-}
>>>-
>>>-static u64 ivb_pte_encode(dma_addr_t addr,
>>>-              enum i915_cache_level level,
>>>-              u32 flags)
>>>-{
>>>-    gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>-
>>>-    switch (level) {
>>>-    case I915_CACHE_L3_LLC:
>>>-        pte |= GEN7_PTE_CACHE_L3_LLC;
>>>-        break;
>>>-    case I915_CACHE_LLC:
>>>-        pte |= GEN6_PTE_CACHE_LLC;
>>>-        break;
>>>-    case I915_CACHE_NONE:
>>>-        pte |= GEN6_PTE_UNCACHED;
>>>-        break;
>>>-    default:
>>>-        MISSING_CASE(level);
>>>-    }
>>>-
>>>-    return pte;
>>>-}
>>>-
>>>-static u64 byt_pte_encode(dma_addr_t addr,
>>>-              enum i915_cache_level level,
>>>-              u32 flags)
>>>-{
>>>-    gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>-
>>>-    if (!(flags & PTE_READ_ONLY))
>>>-        pte |= BYT_PTE_WRITEABLE;
>>>-
>>>-    if (level != I915_CACHE_NONE)
>>>-        pte |= BYT_PTE_SNOOPED_BY_CPU_CACHES;
>>>-
>>>-    return pte;
>>>-}
>>>-
>>>-static u64 hsw_pte_encode(dma_addr_t addr,
>>>-              enum i915_cache_level level,
>>>-              u32 flags)
>>>-{
>>>-    gen6_pte_t pte = HSW_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>-
>>>-    if (level != I915_CACHE_NONE)
>>>-        pte |= HSW_WB_LLC_AGE3;
>>>-
>>>-    return pte;
>>>-}
>>>-
>>>-static u64 iris_pte_encode(dma_addr_t addr,
>>>-               enum i915_cache_level level,
>>>-               u32 flags)
>>>-{
>>>-    gen6_pte_t pte = HSW_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>-
>>>-    switch (level) {
>>>-    case I915_CACHE_NONE:
>>>-        break;
>>>-    case I915_CACHE_WT:
>>>-        pte |= HSW_WT_ELLC_LLC_AGE3;
>>>-        break;
>>>-    default:
>>>-        pte |= HSW_WB_ELLC_LLC_AGE3;
>>>-        break;
>>>-    }
>>>-
>>>-    return pte;
>>>-}
>>>-
>>>-static int gen6_gmch_probe(struct i915_ggtt *ggtt)
>>>-{
>>>-    struct drm_i915_private *i915 = ggtt->vm.i915;
>>>-    struct pci_dev *pdev = to_pci_dev(i915->drm.dev);
>>>-    unsigned int size;
>>>-    u16 snb_gmch_ctl;
>>>-
>>>-    ggtt->gmadr = pci_resource(pdev, 2);
>>>-    ggtt->mappable_end = resource_size(&ggtt->gmadr);
>>>-
>>>-    /*
>>>-     * 64/512MB is the current min/max we actually know of, but this is
>>>-     * just a coarse sanity check.
>>>-     */
>>>-    if (ggtt->mappable_end < (64<<20) || ggtt->mappable_end > 
>>>(512<<20)) {
>>>-        drm_err(&i915->drm, "Unknown GMADR size (%pa)\n",
>>>-            &ggtt->mappable_end);
>>>-        return -ENXIO;
>>>-    }
>>>-
>>>-    pci_read_config_word(pdev, SNB_GMCH_CTRL, &snb_gmch_ctl);
>>>-
>>>-    size = gen6_get_total_gtt_size(snb_gmch_ctl);
>>>-    ggtt->vm.total = (size / sizeof(gen6_pte_t)) * I915_GTT_PAGE_SIZE;
>>>-
>>>-    ggtt->vm.alloc_pt_dma = alloc_pt_dma;
>>>-    ggtt->vm.alloc_scratch_dma = alloc_pt_dma;
>>>-
>>>-    ggtt->vm.clear_range = nop_clear_range;
>>>-    if (!HAS_FULL_PPGTT(i915) || intel_scanout_needs_vtd_wa(i915))
>>>-        ggtt->vm.clear_range = gen6_ggtt_clear_range;
>>>-    ggtt->vm.insert_page = gen6_ggtt_insert_page;
>>>-    ggtt->vm.insert_entries = gen6_ggtt_insert_entries;
>>>-    ggtt->vm.cleanup = gen6_gmch_remove;
>>>-
>>>-    ggtt->invalidate = gen6_ggtt_invalidate;
>>>-
>>>-    if (HAS_EDRAM(i915))
>>>-        ggtt->vm.pte_encode = iris_pte_encode;
>>>-    else if (IS_HASWELL(i915))
>>>-        ggtt->vm.pte_encode = hsw_pte_encode;
>>>-    else if (IS_VALLEYVIEW(i915))
>>>-        ggtt->vm.pte_encode = byt_pte_encode;
>>>-    else if (GRAPHICS_VER(i915) >= 7)
>>>-        ggtt->vm.pte_encode = ivb_pte_encode;
>>>-    else
>>>-        ggtt->vm.pte_encode = snb_pte_encode;
>>>-
>>>-    ggtt->vm.vma_ops.bind_vma    = ggtt_bind_vma;
>>>-    ggtt->vm.vma_ops.unbind_vma  = ggtt_unbind_vma;
>>>-
>>>-    return ggtt_probe_common(ggtt, size);
>>>-}
>>>-
>>>-static void i915_gmch_remove(struct i915_address_space *vm)
>>>-{
>>>-    intel_gmch_remove();
>>>-}
>>>-
>>>-static int i915_gmch_probe(struct i915_ggtt *ggtt)
>>>-{
>>>-    struct drm_i915_private *i915 = ggtt->vm.i915;
>>>-    phys_addr_t gmadr_base;
>>>-    int ret;
>>>-
>>>-    ret = intel_gmch_probe(i915->bridge_dev, 
>>>to_pci_dev(i915->drm.dev), NULL);
>>>-    if (!ret) {
>>>-        drm_err(&i915->drm, "failed to set up gmch\n");
>>>-        return -EIO;
>>>-    }
>>>-
>>>-    intel_gtt_get(&ggtt->vm.total, &gmadr_base, &ggtt->mappable_end);
>>>-
>>>-    ggtt->gmadr =
>>>-        (struct resource)DEFINE_RES_MEM(gmadr_base, 
>>>ggtt->mappable_end);
>>>-
>>>-    ggtt->vm.alloc_pt_dma = alloc_pt_dma;
>>>-    ggtt->vm.alloc_scratch_dma = alloc_pt_dma;
>>>-
>>>-    if (needs_idle_maps(i915)) {
>>>-        drm_notice(&i915->drm,
>>>-               "Flushing DMA requests before IOMMU unmaps; 
>>>performance may be degraded\n");
>>>-        ggtt->do_idle_maps = true;
>>>-    }
>>>-
>>>-    ggtt->vm.insert_page = i915_ggtt_insert_page;
>>>-    ggtt->vm.insert_entries = i915_ggtt_insert_entries;
>>>-    ggtt->vm.clear_range = i915_ggtt_clear_range;
>>>-    ggtt->vm.cleanup = i915_gmch_remove;
>>>-
>>>-    ggtt->invalidate = gmch_ggtt_invalidate;
>>>-
>>>-    ggtt->vm.vma_ops.bind_vma    = ggtt_bind_vma;
>>>-    ggtt->vm.vma_ops.unbind_vma  = ggtt_unbind_vma;
>>>-
>>>-    if (unlikely(ggtt->do_idle_maps))
>>>-        drm_notice(&i915->drm,
>>>-               "Applying Ironlake quirks for intel_iommu\n");
>>>-
>>>-    return 0;
>>>-}
>>>-
>>>  static int ggtt_probe_hw(struct i915_ggtt *ggtt, struct intel_gt *gt)
>>>  {
>>>      struct drm_i915_private *i915 = gt->i915;
>>>@@ -1209,11 +576,11 @@ static int ggtt_probe_hw(struct i915_ggtt 
>>>*ggtt, struct intel_gt *gt)
>>>      dma_resv_init(&ggtt->vm._resv);
>>>        if (GRAPHICS_VER(i915) <= 5)
>>>-        ret = i915_gmch_probe(ggtt);
>>>+        ret = intel_gt_gmch_gen5_probe(ggtt);
>>>      else if (GRAPHICS_VER(i915) < 8)
>>>-        ret = gen6_gmch_probe(ggtt);
>>>+        ret = intel_gt_gmch_gen6_probe(ggtt);
>>>      else
>>>-        ret = gen8_gmch_probe(ggtt);
>>>+        ret = intel_gt_gmch_gen8_probe(ggtt);
>>>      if (ret) {
>>>          dma_resv_fini(&ggtt->vm._resv);
>>>          return ret;
>>>@@ -1267,10 +634,7 @@ int i915_ggtt_probe_hw(struct 
>>>drm_i915_private *i915)
>>>    int i915_ggtt_enable_hw(struct drm_i915_private *i915)
>>>  {
>>>-    if (GRAPHICS_VER(i915) < 6 && !intel_enable_gtt())
>>>-        return -EIO;
>>>-
>>>-    return 0;
>>>+    return intel_gt_gmch_gen5_enable_hw(i915);
>>>  }
>>>    void i915_ggtt_enable_guc(struct i915_ggtt *ggtt)
>>>diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c 
>>>b/drivers/gpu/drm/i915/gt/intel_gt.c
>>>index 5001a6168d56..f0014c5072c9 100644
>>>--- a/drivers/gpu/drm/i915/gt/intel_gt.c
>>>+++ b/drivers/gpu/drm/i915/gt/intel_gt.c
>>>@@ -4,7 +4,6 @@
>>>   */
>>>    #include <drm/drm_managed.h>
>>>-#include <drm/intel-gtt.h>
>>>    #include "gem/i915_gem_internal.h"
>>>  #include "gem/i915_gem_lmem.h"
>>>@@ -17,6 +16,7 @@
>>>  #include "intel_gt_buffer_pool.h"
>>>  #include "intel_gt_clock_utils.h"
>>>  #include "intel_gt_debugfs.h"
>>>+#include "intel_gt_gmch.h"
>>>  #include "intel_gt_pm.h"
>>>  #include "intel_gt_regs.h"
>>>  #include "intel_gt_requests.h"
>>>@@ -451,7 +451,7 @@ void intel_gt_chipset_flush(struct intel_gt *gt)
>>>  {
>>>      wmb();
>>>      if (GRAPHICS_VER(gt->i915) < 6)
>>>-        intel_gtt_chipset_flush();
>>>+        intel_gt_gmch_gen5_chipset_flush(gt);
>>>  }
>>>    void intel_gt_driver_register(struct intel_gt *gt)
>>>diff --git a/drivers/gpu/drm/i915/gt/intel_gt.h 
>>>b/drivers/gpu/drm/i915/gt/intel_gt.h
>>>index e76168e10a21..0163bba0959e 100644
>>>--- a/drivers/gpu/drm/i915/gt/intel_gt.h
>>>+++ b/drivers/gpu/drm/i915/gt/intel_gt.h
>>>@@ -13,6 +13,13 @@
>>>  struct drm_i915_private;
>>>  struct drm_printer;
>>>  +struct insert_entries {
>>>+    struct i915_address_space *vm;
>>>+    struct i915_vma_resource *vma_res;
>>>+    enum i915_cache_level level;
>>>+    u32 flags;
>>>+};
>>>+
>>>  #define GT_TRACE(gt, fmt, ...) do {                    \
>>>      const struct intel_gt *gt__ __maybe_unused = (gt); \
>>>      GEM_TRACE("%s " fmt, dev_name(gt__->i915->drm.dev),        \
>>>@@ -113,4 +120,6 @@ void intel_gt_watchdog_work(struct work_struct 
>>>*work);
>>>    void intel_gt_invalidate_tlbs(struct intel_gt *gt);
>>>  +struct resource intel_pci_resource(struct pci_dev *pdev, int bar);
>>>+
>>>  #endif /* __INTEL_GT_H__ */
>>>diff --git a/drivers/gpu/drm/i915/gt/intel_gt_gmch.c 
>>>b/drivers/gpu/drm/i915/gt/intel_gt_gmch.c
>>>new file mode 100644
>>>index 000000000000..18e488672d1b
>>>--- /dev/null
>>>+++ b/drivers/gpu/drm/i915/gt/intel_gt_gmch.c
>>>@@ -0,0 +1,654 @@
>>>+// SPDX-License-Identifier: MIT
>>>+/*
>>>+ * Copyright © 2022 Intel Corporation
>>>+ */
>>>+
>>>+#include <drm/intel-gtt.h>
>>>+#include <drm/i915_drm.h>
>>>+
>>>+#include <linux/agp_backend.h>
>>>+#include <linux/stop_machine.h>
>>>+
>>>+#include "i915_drv.h"
>>>+#include "intel_gt_gmch.h"
>>>+#include "intel_gt_regs.h"
>>>+#include "intel_gt.h"
>>>+#include "i915_utils.h"
>>>+
>>>+#include "gen8_ppgtt.h"
>>>+
>>>+struct insert_page {
>>>+    struct i915_address_space *vm;
>>>+    dma_addr_t addr;
>>>+    u64 offset;
>>>+    enum i915_cache_level level;
>>>+};
>>>+
>>>+static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte)
>>>+{
>>>+    writeq(pte, addr);
>>>+}
>>>+
>>>+static void nop_clear_range(struct i915_address_space *vm,
>>>+                u64 start, u64 length)
>>>+{
>>>+}
>>>+
>>>+static u64 snb_pte_encode(dma_addr_t addr,
>>>+              enum i915_cache_level level,
>>>+              u32 flags)
>>>+{
>>>+    gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>+
>>>+    switch (level) {
>>>+    case I915_CACHE_L3_LLC:
>>>+    case I915_CACHE_LLC:
>>>+        pte |= GEN6_PTE_CACHE_LLC;
>>>+        break;
>>>+    case I915_CACHE_NONE:
>>>+        pte |= GEN6_PTE_UNCACHED;
>>>+        break;
>>>+    default:
>>>+        MISSING_CASE(level);
>>>+    }
>>>+
>>>+    return pte;
>>>+}
>>>+
>>>+static u64 ivb_pte_encode(dma_addr_t addr,
>>>+              enum i915_cache_level level,
>>>+              u32 flags)
>>>+{
>>>+    gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>+
>>>+    switch (level) {
>>>+    case I915_CACHE_L3_LLC:
>>>+        pte |= GEN7_PTE_CACHE_L3_LLC;
>>>+        break;
>>>+    case I915_CACHE_LLC:
>>>+        pte |= GEN6_PTE_CACHE_LLC;
>>>+        break;
>>>+    case I915_CACHE_NONE:
>>>+        pte |= GEN6_PTE_UNCACHED;
>>>+        break;
>>>+    default:
>>>+        MISSING_CASE(level);
>>>+    }
>>>+
>>>+    return pte;
>>>+}
>>>+
>>>+static u64 byt_pte_encode(dma_addr_t addr,
>>>+              enum i915_cache_level level,
>>>+              u32 flags)
>>>+{
>>>+    gen6_pte_t pte = GEN6_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>+
>>>+    if (!(flags & PTE_READ_ONLY))
>>>+        pte |= BYT_PTE_WRITEABLE;
>>>+
>>>+    if (level != I915_CACHE_NONE)
>>>+        pte |= BYT_PTE_SNOOPED_BY_CPU_CACHES;
>>>+
>>>+    return pte;
>>>+}
>>>+
>>>+static u64 hsw_pte_encode(dma_addr_t addr,
>>>+              enum i915_cache_level level,
>>>+              u32 flags)
>>>+{
>>>+    gen6_pte_t pte = HSW_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>+
>>>+    if (level != I915_CACHE_NONE)
>>>+        pte |= HSW_WB_LLC_AGE3;
>>>+
>>>+    return pte;
>>>+}
>>>+
>>>+static u64 iris_pte_encode(dma_addr_t addr,
>>>+               enum i915_cache_level level,
>>>+               u32 flags)
>>>+{
>>>+    gen6_pte_t pte = HSW_PTE_ADDR_ENCODE(addr) | GEN6_PTE_VALID;
>>>+
>>>+    switch (level) {
>>>+    case I915_CACHE_NONE:
>>>+        break;
>>>+    case I915_CACHE_WT:
>>>+        pte |= HSW_WT_ELLC_LLC_AGE3;
>>>+        break;
>>>+    default:
>>>+        pte |= HSW_WB_ELLC_LLC_AGE3;
>>>+        break;
>>>+    }
>>>+
>>>+    return pte;
>>>+}
>>>+
>>>+static void gen5_ggtt_insert_page(struct i915_address_space *vm,
>>>+                  dma_addr_t addr,
>>>+                  u64 offset,
>>>+                  enum i915_cache_level cache_level,
>>>+                  u32 unused)
>>>+{
>>>+    unsigned int flags = (cache_level == I915_CACHE_NONE) ?
>>>+        AGP_USER_MEMORY : AGP_USER_CACHED_MEMORY;
>>>+
>>>+    intel_gtt_insert_page(addr, offset >> PAGE_SHIFT, flags);
>>>+}
>>>+
>>>+static void gen6_ggtt_insert_page(struct i915_address_space *vm,
>>>+                  dma_addr_t addr,
>>>+                  u64 offset,
>>>+                  enum i915_cache_level level,
>>>+                  u32 flags)
>>>+{
>>>+    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>+    gen6_pte_t __iomem *pte =
>>>+        (gen6_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE;
>>>+
>>>+    iowrite32(vm->pte_encode(addr, level, flags), pte);
>>>+
>>>+    ggtt->invalidate(ggtt);
>>>+}
>>>+
>>>+static void gen8_ggtt_insert_page(struct i915_address_space *vm,
>>>+                  dma_addr_t addr,
>>>+                  u64 offset,
>>>+                  enum i915_cache_level level,
>>>+                  u32 flags)
>>>+{
>>>+    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>+    gen8_pte_t __iomem *pte =
>>>+        (gen8_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE;
>>>+
>>>+    gen8_set_pte(pte, gen8_ggtt_pte_encode(addr, level, flags));
>>>+
>>>+    ggtt->invalidate(ggtt);
>>>+}
>>>+
>>>+static void gen5_ggtt_insert_entries(struct i915_address_space *vm,
>>>+                     struct i915_vma_resource *vma_res,
>>>+                     enum i915_cache_level cache_level,
>>>+                     u32 unused)
>>>+{
>>>+    unsigned int flags = (cache_level == I915_CACHE_NONE) ?
>>>+        AGP_USER_MEMORY : AGP_USER_CACHED_MEMORY;
>>>+
>>>+    intel_gtt_insert_sg_entries(vma_res->bi.pages, vma_res->start 
>>>>> PAGE_SHIFT,
>>>+                    flags);
>>>+}
>>>+
>>>+/*
>>>+ * Binds an object into the global gtt with the specified cache level.
>>>+ * The object will be accessible to the GPU via commands whose operands
>>>+ * reference offsets within the global GTT as well as accessible 
>>>by the GPU
>>>+ * through the GMADR mapped BAR (i915->mm.gtt->gtt).
>>>+ */
>>>+static void gen6_ggtt_insert_entries(struct i915_address_space *vm,
>>>+                     struct i915_vma_resource *vma_res,
>>>+                     enum i915_cache_level level,
>>>+                     u32 flags)
>>>+{
>>>+    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>+    gen6_pte_t __iomem *gte;
>>>+    gen6_pte_t __iomem *end;
>>>+    struct sgt_iter iter;
>>>+    dma_addr_t addr;
>>>+
>>>+    gte = (gen6_pte_t __iomem *)ggtt->gsm;
>>>+    gte += vma_res->start / I915_GTT_PAGE_SIZE;
>>>+    end = gte + vma_res->node_size / I915_GTT_PAGE_SIZE;
>>>+
>>>+    for_each_sgt_daddr(addr, iter, vma_res->bi.pages)
>>>+        iowrite32(vm->pte_encode(addr, level, flags), gte++);
>>>+    GEM_BUG_ON(gte > end);
>>>+
>>>+    /* Fill the allocated but "unused" space beyond the end of 
>>>the buffer */
>>>+    while (gte < end)
>>>+        iowrite32(vm->scratch[0]->encode, gte++);
>>>+
>>>+    /*
>>>+     * We want to flush the TLBs only after we're certain all the PTE
>>>+     * updates have finished.
>>>+     */
>>>+    ggtt->invalidate(ggtt);
>>>+}
>>>+
>>>+static void gen8_ggtt_insert_entries(struct i915_address_space *vm,
>>>+                     struct i915_vma_resource *vma_res,
>>>+                     enum i915_cache_level level,
>>>+                     u32 flags)
>>>+{
>>>+    const gen8_pte_t pte_encode = gen8_ggtt_pte_encode(0, level, 
>>>flags);
>>>+    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>+    gen8_pte_t __iomem *gte;
>>>+    gen8_pte_t __iomem *end;
>>>+    struct sgt_iter iter;
>>>+    dma_addr_t addr;
>>>+
>>>+    /*
>>>+     * Note that we ignore PTE_READ_ONLY here. The caller must be 
>>>careful
>>>+     * not to allow the user to override access to a read only page.
>>>+     */
>>>+
>>>+    gte = (gen8_pte_t __iomem *)ggtt->gsm;
>>>+    gte += vma_res->start / I915_GTT_PAGE_SIZE;
>>>+    end = gte + vma_res->node_size / I915_GTT_PAGE_SIZE;
>>>+
>>>+    for_each_sgt_daddr(addr, iter, vma_res->bi.pages)
>>>+        gen8_set_pte(gte++, pte_encode | addr);
>>>+    GEM_BUG_ON(gte > end);
>>>+
>>>+    /* Fill the allocated but "unused" space beyond the end of 
>>>the buffer */
>>>+    while (gte < end)
>>>+        gen8_set_pte(gte++, vm->scratch[0]->encode);
>>>+
>>>+    /*
>>>+     * We want to flush the TLBs only after we're certain all the PTE
>>>+     * updates have finished.
>>>+     */
>>>+    ggtt->invalidate(ggtt);
>>>+}
>>>+
>>>+static void bxt_vtd_ggtt_wa(struct i915_address_space *vm)
>>>+{
>>>+    /*
>>>+     * Make sure the internal GAM fifo has been cleared of all GTT
>>>+     * writes before exiting stop_machine(). This guarantees that
>>>+     * any aperture accesses waiting to start in another process
>>>+     * cannot back up behind the GTT writes causing a hang.
>>>+     * The register can be any arbitrary GAM register.
>>>+     */
>>>+    intel_uncore_posting_read_fw(vm->gt->uncore, GFX_FLSH_CNTL_GEN6);
>>>+}
>>>+
>>>+static int bxt_vtd_ggtt_insert_page__cb(void *_arg)
>>>+{
>>>+    struct insert_page *arg = _arg;
>>>+
>>>+    gen8_ggtt_insert_page(arg->vm, arg->addr, arg->offset, 
>>>arg->level, 0);
>>>+    bxt_vtd_ggtt_wa(arg->vm);
>>>+
>>>+    return 0;
>>>+}
>>>+
>>>+static void bxt_vtd_ggtt_insert_page__BKL(struct 
>>>i915_address_space *vm,
>>>+                      dma_addr_t addr,
>>>+                      u64 offset,
>>>+                      enum i915_cache_level level,
>>>+                      u32 unused)
>>>+{
>>>+    struct insert_page arg = { vm, addr, offset, level };
>>>+
>>>+    stop_machine(bxt_vtd_ggtt_insert_page__cb, &arg, NULL);
>>>+}
>>>+
>>>+static int bxt_vtd_ggtt_insert_entries__cb(void *_arg)
>>>+{
>>>+    struct insert_entries *arg = _arg;
>>>+
>>>+    gen8_ggtt_insert_entries(arg->vm, arg->vma_res, arg->level, 
>>>arg->flags);
>>>+    bxt_vtd_ggtt_wa(arg->vm);
>>>+
>>>+    return 0;
>>>+}
>>>+
>>>+static void bxt_vtd_ggtt_insert_entries__BKL(struct 
>>>i915_address_space *vm,
>>>+                         struct i915_vma_resource *vma_res,
>>>+                         enum i915_cache_level level,
>>>+                         u32 flags)
>>>+{
>>>+    struct insert_entries arg = { vm, vma_res, level, flags };
>>>+
>>>+    stop_machine(bxt_vtd_ggtt_insert_entries__cb, &arg, NULL);
>>>+}
>>>+
>>>+void intel_gt_gmch_gen5_chipset_flush(struct intel_gt *gt)
>>>+{
>>>+    intel_gtt_chipset_flush();
>>>+}
>>>+
>>>+static void gmch_ggtt_invalidate(struct i915_ggtt *ggtt)
>>>+{
>>>+    intel_gtt_chipset_flush();
>>>+}
>>>+
>>>+static void gen5_ggtt_clear_range(struct i915_address_space *vm,
>>>+                     u64 start, u64 length)
>>>+{
>>>+    intel_gtt_clear_range(start >> PAGE_SHIFT, length >> PAGE_SHIFT);
>>>+}
>>>+
>>>+static void gen6_ggtt_clear_range(struct i915_address_space *vm,
>>>+                  u64 start, u64 length)
>>>+{
>>>+    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>+    unsigned int first_entry = start / I915_GTT_PAGE_SIZE;
>>>+    unsigned int num_entries = length / I915_GTT_PAGE_SIZE;
>>>+    gen6_pte_t scratch_pte, __iomem *gtt_base =
>>>+        (gen6_pte_t __iomem *)ggtt->gsm + first_entry;
>>>+    const int max_entries = ggtt_total_entries(ggtt) - first_entry;
>>>+    int i;
>>>+
>>>+    if (WARN(num_entries > max_entries,
>>>+         "First entry = %d; Num entries = %d (max=%d)\n",
>>>+         first_entry, num_entries, max_entries))
>>>+        num_entries = max_entries;
>>>+
>>>+    scratch_pte = vm->scratch[0]->encode;
>>>+    for (i = 0; i < num_entries; i++)
>>>+        iowrite32(scratch_pte, &gtt_base[i]);
>>>+}
>>>+
>>>+static void gen8_ggtt_clear_range(struct i915_address_space *vm,
>>>+                  u64 start, u64 length)
>>>+{
>>>+    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>+    unsigned int first_entry = start / I915_GTT_PAGE_SIZE;
>>>+    unsigned int num_entries = length / I915_GTT_PAGE_SIZE;
>>>+    const gen8_pte_t scratch_pte = vm->scratch[0]->encode;
>>>+    gen8_pte_t __iomem *gtt_base =
>>>+        (gen8_pte_t __iomem *)ggtt->gsm + first_entry;
>>>+    const int max_entries = ggtt_total_entries(ggtt) - first_entry;
>>>+    int i;
>>>+
>>>+    if (WARN(num_entries > max_entries,
>>>+         "First entry = %d; Num entries = %d (max=%d)\n",
>>>+         first_entry, num_entries, max_entries))
>>>+        num_entries = max_entries;
>>>+
>>>+    for (i = 0; i < num_entries; i++)
>>>+        gen8_set_pte(&gtt_base[i], scratch_pte);
>>>+}
>>>+
>>>+static void gen5_gmch_remove(struct i915_address_space *vm)
>>>+{
>>>+    intel_gmch_remove();
>>>+}
>>>+
>>>+static void gen6_gmch_remove(struct i915_address_space *vm)
>>>+{
>>>+    struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
>>>+
>>>+    iounmap(ggtt->gsm);
>>>+    free_scratch(vm);
>>>+}
>>>+
>>>+/*
>>>+ * Certain Gen5 chipsets require idling the GPU before
>>>+ * unmapping anything from the GTT when VT-d is enabled.
>>>+ */
>>>+static bool needs_idle_maps(struct drm_i915_private *i915)
>>>+{
>>>+    /*
>>>+     * Query intel_iommu to see if we need the workaround. 
>>>Presumably that
>>>+     * was loaded first.
>>>+     */
>>>+    if (!i915_vtd_active(i915))
>>>+        return false;
>>>+
>>>+    if (GRAPHICS_VER(i915) == 5 && IS_MOBILE(i915))
>>>+        return true;
>>>+
>>>+    if (GRAPHICS_VER(i915) == 12)
>>>+        return true; /* XXX DMAR fault reason 7 */
>>>+
>>>+    return false;
>>>+}
>>>+
>>>+static unsigned int gen6_gttmmadr_size(struct drm_i915_private *i915)
>>>+{
>>>+    /*
>>>+     * GEN6: GTTMMADR size is 4MB and GTTADR starts at 2MB offset
>>>+     * GEN8: GTTMMADR size is 16MB and GTTADR starts at 8MB offset
>>>+     */
>>>+    GEM_BUG_ON(GRAPHICS_VER(i915) < 6);
>>>+    return (GRAPHICS_VER(i915) < 8) ? SZ_4M : SZ_16M;
>>>+}
>>>+
>>>+static unsigned int gen6_get_total_gtt_size(u16 snb_gmch_ctl)
>>>+{
>>>+    snb_gmch_ctl >>= SNB_GMCH_GGMS_SHIFT;
>>>+    snb_gmch_ctl &= SNB_GMCH_GGMS_MASK;
>>>+    return snb_gmch_ctl << 20;
>>>+}
>>>+
>>>+static unsigned int gen8_get_total_gtt_size(u16 bdw_gmch_ctl)
>>>+{
>>>+    bdw_gmch_ctl >>= BDW_GMCH_GGMS_SHIFT;
>>>+    bdw_gmch_ctl &= BDW_GMCH_GGMS_MASK;
>>>+    if (bdw_gmch_ctl)
>>>+        bdw_gmch_ctl = 1 << bdw_gmch_ctl;
>>>+
>>>+#ifdef CONFIG_X86_32
>>>+    /* Limit 32b platforms to a 2GB GGTT: 4 << 20 / pte size * 
>>>I915_GTT_PAGE_SIZE */
>>>+    if (bdw_gmch_ctl > 4)
>>>+        bdw_gmch_ctl = 4;
>>>+#endif
>>>+
>>>+    return bdw_gmch_ctl << 20;
>>>+}
>>>+
>>>+static unsigned int gen6_gttadr_offset(struct drm_i915_private *i915)
>>>+{
>>>+    return gen6_gttmmadr_size(i915) / 2;
>>>+}
>>>+
>>>+static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
>>>+{
>>>+    struct drm_i915_private *i915 = ggtt->vm.i915;
>>>+    struct pci_dev *pdev = to_pci_dev(i915->drm.dev);
>>>+    phys_addr_t phys_addr;
>>>+    u32 pte_flags;
>>>+    int ret;
>>>+
>>>+    GEM_WARN_ON(pci_resource_len(pdev, 0) != gen6_gttmmadr_size(i915));
>>>+    phys_addr = pci_resource_start(pdev, 0) + gen6_gttadr_offset(i915);
>>>+
>>>+    /*
>>>+     * On BXT+/ICL+ writes larger than 64 bit to the GTT 
>>>pagetable range
>>>+     * will be dropped. For WC mappings in general we have 64 
>>>byte burst
>>>+     * writes when the WC buffer is flushed, so we can't use it, 
>>>but have to
>>>+     * resort to an uncached mapping. The WC issue is easily 
>>>caught by the
>>>+     * readback check when writing GTT PTE entries.
>>>+     */
>>>+    if (IS_GEN9_LP(i915) || GRAPHICS_VER(i915) >= 11)
>>>+        ggtt->gsm = ioremap(phys_addr, size);
>>>+    else
>>>+        ggtt->gsm = ioremap_wc(phys_addr, size);
>>>+    if (!ggtt->gsm) {
>>>+        drm_err(&i915->drm, "Failed to map the ggtt page table\n");
>>>+        return -ENOMEM;
>>>+    }
>>>+
>>>+    kref_init(&ggtt->vm.resv_ref);
>>>+    ret = setup_scratch_page(&ggtt->vm);
>>>+    if (ret) {
>>>+        drm_err(&i915->drm, "Scratch setup failed\n");
>>>+        /* iounmap will also get called at remove, but meh */
>>>+        iounmap(ggtt->gsm);
>>>+        return ret;
>>>+    }
>>>+
>>>+    pte_flags = 0;
>>>+    if (i915_gem_object_is_lmem(ggtt->vm.scratch[0]))
>>>+        pte_flags |= PTE_LM;
>>>+
>>>+    ggtt->vm.scratch[0]->encode =
>>>+        ggtt->vm.pte_encode(px_dma(ggtt->vm.scratch[0]),
>>>+                    I915_CACHE_NONE, pte_flags);
>>>+
>>>+    return 0;
>>>+}
>>>+
>>>+int intel_gt_gmch_gen5_probe(struct i915_ggtt *ggtt)
>>>+{
>>>+    struct drm_i915_private *i915 = ggtt->vm.i915;
>>>+    phys_addr_t gmadr_base;
>>>+    int ret;
>>>+
>>>+    ret = intel_gmch_probe(i915->bridge_dev, 
>>>to_pci_dev(i915->drm.dev), NULL);
>>>+    if (!ret) {
>>>+        drm_err(&i915->drm, "failed to set up gmch\n");
>>>+        return -EIO;
>>>+    }
>>>+
>>>+    intel_gtt_get(&ggtt->vm.total, &gmadr_base, &ggtt->mappable_end);
>>>+
>>>+    ggtt->gmadr =
>>>+        (struct resource)DEFINE_RES_MEM(gmadr_base, 
>>>ggtt->mappable_end);
>>>+
>>>+    ggtt->vm.alloc_pt_dma = alloc_pt_dma;
>>>+    ggtt->vm.alloc_scratch_dma = alloc_pt_dma;
>>>+
>>>+    if (needs_idle_maps(i915)) {
>>>+        drm_notice(&i915->drm,
>>>+               "Flushing DMA requests before IOMMU unmaps; 
>>>performance may be degraded\n");
>>>+        ggtt->do_idle_maps = true;
>>>+    }
>>>+
>>>+    ggtt->vm.insert_page = gen5_ggtt_insert_page;
>>>+    ggtt->vm.insert_entries = gen5_ggtt_insert_entries;
>>>+    ggtt->vm.clear_range = gen5_ggtt_clear_range;
>>>+    ggtt->vm.cleanup = gen5_gmch_remove;
>>>+
>>>+    ggtt->invalidate = gmch_ggtt_invalidate;
>>>+
>>>+    ggtt->vm.vma_ops.bind_vma    = intel_ggtt_bind_vma;
>>>+    ggtt->vm.vma_ops.unbind_vma  = intel_ggtt_unbind_vma;
>>>+
>>>+    if (unlikely(ggtt->do_idle_maps))
>>>+        drm_notice(&i915->drm,
>>>+               "Applying Ironlake quirks for intel_iommu\n");
>>>+
>>>+    return 0;
>>>+}
>>>+
>>>+int intel_gt_gmch_gen6_probe(struct i915_ggtt *ggtt)
>>>+{
>>>+    struct drm_i915_private *i915 = ggtt->vm.i915;
>>>+    struct pci_dev *pdev = to_pci_dev(i915->drm.dev);
>>>+    unsigned int size;
>>>+    u16 snb_gmch_ctl;
>>>+
>>>+    ggtt->gmadr = intel_pci_resource(pdev, 2);
>>>+    ggtt->mappable_end = resource_size(&ggtt->gmadr);
>>>+
>>>+    /*
>>>+     * 64/512MB is the current min/max we actually know of, but this is
>>>+     * just a coarse sanity check.
>>>+     */
>>>+    if (ggtt->mappable_end < (64<<20) || ggtt->mappable_end > 
>>>(512<<20)) {
>>>+        drm_err(&i915->drm, "Unknown GMADR size (%pa)\n",
>>>+            &ggtt->mappable_end);
>>>+        return -ENXIO;
>>>+    }
>>>+
>>>+    pci_read_config_word(pdev, SNB_GMCH_CTRL, &snb_gmch_ctl);
>>>+
>>>+    size = gen6_get_total_gtt_size(snb_gmch_ctl);
>>>+    ggtt->vm.total = (size / sizeof(gen6_pte_t)) * I915_GTT_PAGE_SIZE;
>>>+
>>>+    ggtt->vm.alloc_pt_dma = alloc_pt_dma;
>>>+    ggtt->vm.alloc_scratch_dma = alloc_pt_dma;
>>>+
>>>+    ggtt->vm.clear_range = nop_clear_range;
>>>+    if (!HAS_FULL_PPGTT(i915) || intel_scanout_needs_vtd_wa(i915))
>>>+        ggtt->vm.clear_range = gen6_ggtt_clear_range;
>>>+    ggtt->vm.insert_page = gen6_ggtt_insert_page;
>>>+    ggtt->vm.insert_entries = gen6_ggtt_insert_entries;
>>>+    ggtt->vm.cleanup = gen6_gmch_remove;
>>>+
>>>+    ggtt->invalidate = gen6_ggtt_invalidate;
>>>+
>>>+    if (HAS_EDRAM(i915))
>>>+        ggtt->vm.pte_encode = iris_pte_encode;
>>>+    else if (IS_HASWELL(i915))
>>>+        ggtt->vm.pte_encode = hsw_pte_encode;
>>>+    else if (IS_VALLEYVIEW(i915))
>>>+        ggtt->vm.pte_encode = byt_pte_encode;
>>>+    else if (GRAPHICS_VER(i915) >= 7)
>>>+        ggtt->vm.pte_encode = ivb_pte_encode;
>>>+    else
>>>+        ggtt->vm.pte_encode = snb_pte_encode;
>>>+
>>>+    ggtt->vm.vma_ops.bind_vma    = intel_ggtt_bind_vma;
>>>+    ggtt->vm.vma_ops.unbind_vma  = intel_ggtt_unbind_vma;
>>>+
>>>+    return ggtt_probe_common(ggtt, size);
>>>+}
>>>+
>>>+static unsigned int chv_get_total_gtt_size(u16 gmch_ctrl)
>>>+{
>>>+    gmch_ctrl >>= SNB_GMCH_GGMS_SHIFT;
>>>+    gmch_ctrl &= SNB_GMCH_GGMS_MASK;
>>>+
>>>+    if (gmch_ctrl)
>>>+        return 1 << (20 + gmch_ctrl);
>>>+
>>>+    return 0;
>>>+}
>>>+
>>>+int intel_gt_gmch_gen8_probe(struct i915_ggtt *ggtt)
>>>+{
>>>+    struct drm_i915_private *i915 = ggtt->vm.i915;
>>>+    struct pci_dev *pdev = to_pci_dev(i915->drm.dev);
>>>+    unsigned int size;
>>>+    u16 snb_gmch_ctl;
>>>+
>>>+    /* TODO: We're not aware of mappable constraints on gen8 yet */
>>>+    if (!HAS_LMEM(i915)) {
>>>+        ggtt->gmadr = intel_pci_resource(pdev, 2);
>>>+        ggtt->mappable_end = resource_size(&ggtt->gmadr);
>>>+    }
>>>+
>>>+    pci_read_config_word(pdev, SNB_GMCH_CTRL, &snb_gmch_ctl);
>>>+    if (IS_CHERRYVIEW(i915))
>>>+        size = chv_get_total_gtt_size(snb_gmch_ctl);
>>>+    else
>>>+        size = gen8_get_total_gtt_size(snb_gmch_ctl);
>>>+
>>>+    ggtt->vm.alloc_pt_dma = alloc_pt_dma;
>>>+    ggtt->vm.alloc_scratch_dma = alloc_pt_dma;
>>>+    ggtt->vm.lmem_pt_obj_flags = I915_BO_ALLOC_PM_EARLY;
>>>+
>>>+    ggtt->vm.total = (size / sizeof(gen8_pte_t)) * I915_GTT_PAGE_SIZE;
>>>+    ggtt->vm.cleanup = gen6_gmch_remove;
>>>+    ggtt->vm.insert_page = gen8_ggtt_insert_page;
>>>+    ggtt->vm.clear_range = nop_clear_range;
>>>+    if (intel_scanout_needs_vtd_wa(i915))
>>>+        ggtt->vm.clear_range = gen8_ggtt_clear_range;
>>>+
>>>+    ggtt->vm.insert_entries = gen8_ggtt_insert_entries;
>>>+
>>>+    /*
>>>+     * Serialize GTT updates with aperture access on BXT if VT-d is on,
>>>+     * and always on CHV.
>>>+     */
>>>+    if (intel_vm_no_concurrent_access_wa(i915)) {
>>>+        ggtt->vm.insert_entries = bxt_vtd_ggtt_insert_entries__BKL;
>>>+        ggtt->vm.insert_page    = bxt_vtd_ggtt_insert_page__BKL;
>>>+        ggtt->vm.bind_async_flags =
>>>+            I915_VMA_GLOBAL_BIND | I915_VMA_LOCAL_BIND;
>>>+    }
>>>+
>>>+    ggtt->invalidate = gen8_ggtt_invalidate;
>>>+
>>>+    ggtt->vm.vma_ops.bind_vma    = intel_ggtt_bind_vma;
>>>+    ggtt->vm.vma_ops.unbind_vma  = intel_ggtt_unbind_vma;
>>>+
>>>+    ggtt->vm.pte_encode = gen8_ggtt_pte_encode;
>>>+
>>>+    setup_private_pat(ggtt->vm.gt->uncore);
>>>+
>>>+    return ggtt_probe_common(ggtt, size);
>>>+}
>>>+
>>>+int intel_gt_gmch_gen5_enable_hw(struct drm_i915_private *i915)
>>>+{
>>>+    if (GRAPHICS_VER(i915) < 6 && !intel_enable_gtt())
>>>+        return -EIO;
>>>+
>>>+    return 0;
>>>+}
>>>diff --git a/drivers/gpu/drm/i915/gt/intel_gt_gmch.h 
>>>b/drivers/gpu/drm/i915/gt/intel_gt_gmch.h
>>>new file mode 100644
>>>index 000000000000..75ed55c1f30a
>>>--- /dev/null
>>>+++ b/drivers/gpu/drm/i915/gt/intel_gt_gmch.h
>>>@@ -0,0 +1,46 @@
>>>+/* SPDX-License-Identifier: MIT */
>>>+/*
>>>+ * Copyright © 2022 Intel Corporation
>>>+ */
>>>+
>>>+#ifndef __INTEL_GT_GMCH_H__
>>>+#define __INTEL_GT_GMCH_H__
>>>+
>>>+#include "intel_gtt.h"
>>>+
>>>+/* For x86 platforms */
>>>+#if IS_ENABLED(CONFIG_X86)
>>>+void intel_gt_gmch_gen5_chipset_flush(struct intel_gt *gt);
>>>+int intel_gt_gmch_gen6_probe(struct i915_ggtt *ggtt);
>>>+int intel_gt_gmch_gen8_probe(struct i915_ggtt *ggtt);
>>>+int intel_gt_gmch_gen5_probe(struct i915_ggtt *ggtt);
>>>+int intel_gt_gmch_gen5_enable_hw(struct drm_i915_private *i915);
>>>+
>>>+/* Stubs for non-x86 platforms */
>>>+#else
>>>+static inline void intel_gt_gmch_gen5_chipset_flush(struct 
>>>intel_gt *gt)
>>>+{
>>>+}
>>>+static inline int intel_gt_gmch_gen5_probe(struct i915_ggtt *ggtt)
>>>+{
>>>+    /* No HW should be probed for this case yet, return fail */
>>>+    return -ENODEV;
>>>+}
>>>+static inline int intel_gt_gmch_gen6_probe(struct i915_ggtt *ggtt)
>>>+{
>>>+    /* No HW should be probed for this case yet, return fail */
>>>+    return -ENODEV;
>>>+}
>>>+static inline int intel_gt_gmch_gen8_probe(struct i915_ggtt *ggtt)
>>>+{
>>>+    /* No HW should be probed for this case yet, return fail */
>>>+    return -ENODEV;
>>>+}
>>>+static inline int intel_gt_gmch_gen5_enable_hw(struct 
>>>drm_i915_private *i915)
>>>+{
>>>+    /* No HW should be enabled for this case yet, return fail */
>>>+    return -ENODEV;
>>>+}
>>>+#endif
>>>+
>>>+#endif /* __INTEL_GT_GMCH_H__ */
>>>diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.h 
>>>b/drivers/gpu/drm/i915/gt/intel_gtt.h
>>>index 5922e2cf4d8d..a40d928b3888 100644
>>>--- a/drivers/gpu/drm/i915/gt/intel_gtt.h
>>>+++ b/drivers/gpu/drm/i915/gt/intel_gtt.h
>>>@@ -549,6 +549,14 @@ i915_page_dir_dma_addr(const struct 
>>>i915_ppgtt *ppgtt, const unsigned int n)
>>>  void ppgtt_init(struct i915_ppgtt *ppgtt, struct intel_gt *gt,
>>>          unsigned long lmem_pt_obj_flags);
>>>  +void intel_ggtt_bind_vma(struct i915_address_space *vm,
>>>+              struct i915_vm_pt_stash *stash,
>>>+              struct i915_vma_resource *vma_res,
>>>+              enum i915_cache_level cache_level,
>>>+              u32 flags);
>>>+void intel_ggtt_unbind_vma(struct i915_address_space *vm,
>>>+                struct i915_vma_resource *vma_res);
>>>+
>>>  int i915_ggtt_probe_hw(struct drm_i915_private *i915);
>>>  int i915_ggtt_init_hw(struct drm_i915_private *i915);
>>>  int i915_ggtt_enable_hw(struct drm_i915_private *i915);
>>>@@ -619,6 +627,7 @@ release_pd_entry(struct i915_page_directory * 
>>>const pd,
>>>           struct i915_page_table * const pt,
>>>           const struct drm_i915_gem_object * const scratch);
>>>  void gen6_ggtt_invalidate(struct i915_ggtt *ggtt);
>>>+void gen8_ggtt_invalidate(struct i915_ggtt *ggtt);
>>>    void ppgtt_bind_vma(struct i915_address_space *vm,
>>>              struct i915_vm_pt_stash *stash,
>


More information about the Intel-gfx mailing list