From: Fei Yang <fei.y...@intel.com>

On MTL, GT can no longer allocate on LLC - only the CPU can.
This, along with addition of support for ADM/L4 cache calls a
MOCS/PAT table update.
Also add PTE encode functions for MTL as it has different PAT
index definition than previous platforms.

BSpec: 44509, 45101, 44235

Cc: Matt Roper <matthew.d.ro...@intel.com>
Cc: Lucas De Marchi <lucas.demar...@intel.com>
Signed-off-by: Madhumitha Tolakanahalli Pradeep 
<madhumitha.tolakanahalli.prad...@intel.com>
Signed-off-by: Aravind Iddamsetty <aravind.iddamse...@intel.com>
Signed-off-by: Fei Yang <fei.y...@intel.com>
---
 drivers/gpu/drm/i915/display/intel_dpt.c |  2 +-
 drivers/gpu/drm/i915/gt/gen8_ppgtt.c     | 43 ++++++++++++--
 drivers/gpu/drm/i915/gt/gen8_ppgtt.h     |  3 +
 drivers/gpu/drm/i915/gt/intel_ggtt.c     | 36 ++++++++++-
 drivers/gpu/drm/i915/gt/intel_gtt.c      | 23 ++++++-
 drivers/gpu/drm/i915/gt/intel_gtt.h      | 20 ++++++-
 drivers/gpu/drm/i915/gt/intel_mocs.c     | 76 ++++++++++++++++++++++--
 drivers/gpu/drm/i915/gt/selftest_mocs.c  |  2 +-
 drivers/gpu/drm/i915/i915_pci.c          |  1 +
 9 files changed, 189 insertions(+), 17 deletions(-)

diff --git a/drivers/gpu/drm/i915/display/intel_dpt.c 
b/drivers/gpu/drm/i915/display/intel_dpt.c
index 2bf5cce232d5..dc82a756c899 100644
--- a/drivers/gpu/drm/i915/display/intel_dpt.c
+++ b/drivers/gpu/drm/i915/display/intel_dpt.c
@@ -300,7 +300,7 @@ intel_dpt_create(struct intel_framebuffer *fb)
        vm->vma_ops.bind_vma    = dpt_bind_vma;
        vm->vma_ops.unbind_vma  = dpt_unbind_vma;
 
-       vm->pte_encode = gen8_ggtt_pte_encode;
+       vm->pte_encode = vm->gt->ggtt->vm.pte_encode;
 
        dpt->obj = dpt_obj;
        dpt->obj->is_dpt = true;
diff --git a/drivers/gpu/drm/i915/gt/gen8_ppgtt.c 
b/drivers/gpu/drm/i915/gt/gen8_ppgtt.c
index 4daaa6f55668..4197b43150cc 100644
--- a/drivers/gpu/drm/i915/gt/gen8_ppgtt.c
+++ b/drivers/gpu/drm/i915/gt/gen8_ppgtt.c
@@ -55,6 +55,34 @@ static u64 gen8_pte_encode(dma_addr_t addr,
        return pte;
 }
 
+static u64 mtl_pte_encode(dma_addr_t addr,
+                         enum i915_cache_level level,
+                         u32 flags)
+{
+       gen8_pte_t pte = addr | GEN8_PAGE_PRESENT | GEN8_PAGE_RW;
+
+       if (unlikely(flags & PTE_READ_ONLY))
+               pte &= ~GEN8_PAGE_RW;
+
+       if (flags & PTE_LM)
+               pte |= GEN12_PPGTT_PTE_LM | GEN12_PPGTT_PTE_NC;
+
+       switch (level) {
+       case I915_CACHE_NONE:
+               pte |= GEN12_PPGTT_PTE_PAT1;
+               break;
+       case I915_CACHE_LLC:
+       case I915_CACHE_L3_LLC:
+               pte |= GEN12_PPGTT_PTE_PAT0 | GEN12_PPGTT_PTE_PAT1;
+               break;
+       case I915_CACHE_WT:
+               pte |= GEN12_PPGTT_PTE_PAT0;
+               break;
+       }
+
+       return pte;
+}
+
 static void gen8_ppgtt_notify_vgt(struct i915_ppgtt *ppgtt, bool create)
 {
        struct drm_i915_private *i915 = ppgtt->vm.i915;
@@ -427,7 +455,7 @@ gen8_ppgtt_insert_pte(struct i915_ppgtt *ppgtt,
                      u32 flags)
 {
        struct i915_page_directory *pd;
-       const gen8_pte_t pte_encode = gen8_pte_encode(0, cache_level, flags);
+       const gen8_pte_t pte_encode = ppgtt->vm.pte_encode(0, cache_level, 
flags);
        gen8_pte_t *vaddr;
 
        pd = i915_pd_entry(pdp, gen8_pd_index(idx, 2));
@@ -580,7 +608,7 @@ static void gen8_ppgtt_insert_huge(struct 
i915_address_space *vm,
                                   enum i915_cache_level cache_level,
                                   u32 flags)
 {
-       const gen8_pte_t pte_encode = gen8_pte_encode(0, cache_level, flags);
+       const gen8_pte_t pte_encode = vm->pte_encode(0, cache_level, flags);
        unsigned int rem = sg_dma_len(iter->sg);
        u64 start = vma_res->start;
 
@@ -743,7 +771,7 @@ static void gen8_ppgtt_insert_entry(struct 
i915_address_space *vm,
        GEM_BUG_ON(pt->is_compact);
 
        vaddr = px_vaddr(pt);
-       vaddr[gen8_pd_index(idx, 0)] = gen8_pte_encode(addr, level, flags);
+       vaddr[gen8_pd_index(idx, 0)] = vm->pte_encode(addr, level, flags);
        drm_clflush_virt_range(&vaddr[gen8_pd_index(idx, 0)], sizeof(*vaddr));
 }
 
@@ -773,7 +801,7 @@ static void __xehpsdv_ppgtt_insert_entry_lm(struct 
i915_address_space *vm,
        }
 
        vaddr = px_vaddr(pt);
-       vaddr[gen8_pd_index(idx, 0) / 16] = gen8_pte_encode(addr, level, flags);
+       vaddr[gen8_pd_index(idx, 0) / 16] = vm->pte_encode(addr, level, flags);
 }
 
 static void xehpsdv_ppgtt_insert_entry(struct i915_address_space *vm,
@@ -820,7 +848,7 @@ static int gen8_init_scratch(struct i915_address_space *vm)
                pte_flags |= PTE_LM;
 
        vm->scratch[0]->encode =
-               gen8_pte_encode(px_dma(vm->scratch[0]),
+               vm->pte_encode(px_dma(vm->scratch[0]),
                                I915_CACHE_NONE, pte_flags);
 
        for (i = 1; i <= vm->top; i++) {
@@ -963,7 +991,10 @@ struct i915_ppgtt *gen8_ppgtt_create(struct intel_gt *gt,
         */
        ppgtt->vm.alloc_scratch_dma = alloc_pt_dma;
 
-       ppgtt->vm.pte_encode = gen8_pte_encode;
+       if (GRAPHICS_VER_FULL(gt->i915) >= IP_VER(12, 70))
+               ppgtt->vm.pte_encode = mtl_pte_encode;
+       else
+               ppgtt->vm.pte_encode = gen8_pte_encode;
 
        ppgtt->vm.bind_async_flags = I915_VMA_LOCAL_BIND;
        ppgtt->vm.insert_entries = gen8_ppgtt_insert;
diff --git a/drivers/gpu/drm/i915/gt/gen8_ppgtt.h 
b/drivers/gpu/drm/i915/gt/gen8_ppgtt.h
index f541d19264b4..6b8ce7f4d25a 100644
--- a/drivers/gpu/drm/i915/gt/gen8_ppgtt.h
+++ b/drivers/gpu/drm/i915/gt/gen8_ppgtt.h
@@ -18,5 +18,8 @@ struct i915_ppgtt *gen8_ppgtt_create(struct intel_gt *gt,
 u64 gen8_ggtt_pte_encode(dma_addr_t addr,
                         enum i915_cache_level level,
                         u32 flags);
+u64 mtl_ggtt_pte_encode(dma_addr_t addr,
+                       unsigned int pat_index,
+                       u32 flags);
 
 #endif
diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c 
b/drivers/gpu/drm/i915/gt/intel_ggtt.c
index 3c7f1ed92f5b..ba3109338aee 100644
--- a/drivers/gpu/drm/i915/gt/intel_ggtt.c
+++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c
@@ -220,6 +220,33 @@ static void guc_ggtt_invalidate(struct i915_ggtt *ggtt)
        }
 }
 
+u64 mtl_ggtt_pte_encode(dma_addr_t addr,
+                       enum i915_cache_level level,
+                       u32 flags)
+{
+       gen8_pte_t pte = addr | GEN8_PAGE_PRESENT;
+
+       GEM_BUG_ON(addr & ~GEN12_GGTT_PTE_ADDR_MASK);
+
+       if (flags & PTE_LM)
+               pte |= GEN12_GGTT_PTE_LM;
+
+       switch (level) {
+       case I915_CACHE_NONE:
+               pte |= MTL_GGTT_PTE_PAT1;
+               break;
+       case I915_CACHE_LLC:
+       case I915_CACHE_L3_LLC:
+               pte |= MTL_GGTT_PTE_PAT0 | MTL_GGTT_PTE_PAT1;
+               break;
+       case I915_CACHE_WT:
+               pte |= MTL_GGTT_PTE_PAT0;
+               break;
+       }
+
+       return pte;
+}
+
 u64 gen8_ggtt_pte_encode(dma_addr_t addr,
                         enum i915_cache_level level,
                         u32 flags)
@@ -247,7 +274,7 @@ static void gen8_ggtt_insert_page(struct i915_address_space 
*vm,
        gen8_pte_t __iomem *pte =
                (gen8_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE;
 
-       gen8_set_pte(pte, gen8_ggtt_pte_encode(addr, level, flags));
+       gen8_set_pte(pte, ggtt->vm.pte_encode(addr, level, flags));
 
        ggtt->invalidate(ggtt);
 }
@@ -257,8 +284,8 @@ static void gen8_ggtt_insert_entries(struct 
i915_address_space *vm,
                                     enum i915_cache_level level,
                                     u32 flags)
 {
-       const gen8_pte_t pte_encode = gen8_ggtt_pte_encode(0, level, flags);
        struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
+       const gen8_pte_t pte_encode = ggtt->vm.pte_encode(0, level, flags);
        gen8_pte_t __iomem *gte;
        gen8_pte_t __iomem *end;
        struct sgt_iter iter;
@@ -981,7 +1008,10 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt)
        ggtt->vm.vma_ops.bind_vma    = intel_ggtt_bind_vma;
        ggtt->vm.vma_ops.unbind_vma  = intel_ggtt_unbind_vma;
 
-       ggtt->vm.pte_encode = gen8_ggtt_pte_encode;
+       if (GRAPHICS_VER_FULL(i915) >= IP_VER(12, 70))
+               ggtt->vm.pte_encode = mtl_ggtt_pte_encode;
+       else
+               ggtt->vm.pte_encode = gen8_ggtt_pte_encode;
 
        return ggtt_probe_common(ggtt, size);
 }
diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.c 
b/drivers/gpu/drm/i915/gt/intel_gtt.c
index 4f436ba7a3c8..1e1b34e22cf5 100644
--- a/drivers/gpu/drm/i915/gt/intel_gtt.c
+++ b/drivers/gpu/drm/i915/gt/intel_gtt.c
@@ -468,6 +468,25 @@ void gtt_write_workarounds(struct intel_gt *gt)
        }
 }
 
+static void mtl_setup_private_ppat(struct intel_uncore *uncore)
+{
+       intel_uncore_write(uncore, GEN12_PAT_INDEX(0),
+                          MTL_PPAT_L4_0_WB);
+       intel_uncore_write(uncore, GEN12_PAT_INDEX(1),
+                          MTL_PPAT_L4_1_WT);
+       intel_uncore_write(uncore, GEN12_PAT_INDEX(2),
+                          MTL_PPAT_L4_3_UC);
+       intel_uncore_write(uncore, GEN12_PAT_INDEX(3),
+                          MTL_PPAT_L4_0_WB | MTL_2_COH_1W);
+       intel_uncore_write(uncore, GEN12_PAT_INDEX(4),
+                          MTL_PPAT_L4_0_WB | MTL_3_COH_2W);
+
+       /*
+        * Remaining PAT entries are left at the hardware-default
+        * fully-cached setting
+        */
+}
+
 static void tgl_setup_private_ppat(struct intel_uncore *uncore)
 {
        /* TGL doesn't support LLC or AGE settings */
@@ -603,7 +622,9 @@ void setup_private_pat(struct intel_gt *gt)
 
        GEM_BUG_ON(GRAPHICS_VER(i915) < 8);
 
-       if (GRAPHICS_VER_FULL(i915) >= IP_VER(12, 50))
+       if (IS_METEORLAKE(i915))
+               mtl_setup_private_ppat(uncore);
+       else if (GRAPHICS_VER_FULL(i915) >= IP_VER(12, 50))
                xehp_setup_private_ppat(gt);
        else if (GRAPHICS_VER(i915) >= 12)
                tgl_setup_private_ppat(uncore);
diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.h 
b/drivers/gpu/drm/i915/gt/intel_gtt.h
index 69ce55f517f5..b632167eaf2e 100644
--- a/drivers/gpu/drm/i915/gt/intel_gtt.h
+++ b/drivers/gpu/drm/i915/gt/intel_gtt.h
@@ -88,9 +88,18 @@ typedef u64 gen8_pte_t;
 #define BYT_PTE_SNOOPED_BY_CPU_CACHES  REG_BIT(2)
 #define BYT_PTE_WRITEABLE              REG_BIT(1)
 
+#define GEN12_PPGTT_PTE_PAT3    BIT_ULL(62)
 #define GEN12_PPGTT_PTE_LM     BIT_ULL(11)
+#define GEN12_PPGTT_PTE_PAT2    BIT_ULL(7)
+#define GEN12_PPGTT_PTE_NC      BIT_ULL(5)
+#define GEN12_PPGTT_PTE_PAT1    BIT_ULL(4)
+#define GEN12_PPGTT_PTE_PAT0    BIT_ULL(3)
 
-#define GEN12_GGTT_PTE_LM      BIT_ULL(1)
+#define GEN12_GGTT_PTE_LM              BIT_ULL(1)
+#define MTL_GGTT_PTE_PAT0              BIT_ULL(52)
+#define MTL_GGTT_PTE_PAT1              BIT_ULL(53)
+#define GEN12_GGTT_PTE_ADDR_MASK       GENMASK_ULL(45, 12)
+#define MTL_GGTT_PTE_PAT_MASK          GENMASK_ULL(53, 52)
 
 #define GEN12_PDE_64K BIT(6)
 #define GEN12_PTE_PS64 BIT(8)
@@ -147,6 +156,15 @@ typedef u64 gen8_pte_t;
 #define GEN8_PDE_IPS_64K BIT(11)
 #define GEN8_PDE_PS_2M   BIT(7)
 
+#define MTL_PPAT_L4_CACHE_POLICY_MASK  REG_GENMASK(3, 2)
+#define MTL_PAT_INDEX_COH_MODE_MASK    REG_GENMASK(1, 0)
+#define MTL_PPAT_L4_3_UC       REG_FIELD_PREP(MTL_PPAT_L4_CACHE_POLICY_MASK, 3)
+#define MTL_PPAT_L4_1_WT       REG_FIELD_PREP(MTL_PPAT_L4_CACHE_POLICY_MASK, 1)
+#define MTL_PPAT_L4_0_WB       REG_FIELD_PREP(MTL_PPAT_L4_CACHE_POLICY_MASK, 0)
+#define MTL_3_COH_2W   REG_FIELD_PREP(MTL_PAT_INDEX_COH_MODE_MASK, 3)
+#define MTL_2_COH_1W   REG_FIELD_PREP(MTL_PAT_INDEX_COH_MODE_MASK, 2)
+#define MTL_0_COH_NON  REG_FIELD_PREP(MTL_PAT_INDEX_COH_MODE_MASK, 0)
+
 enum i915_cache_level;
 
 struct drm_i915_gem_object;
diff --git a/drivers/gpu/drm/i915/gt/intel_mocs.c 
b/drivers/gpu/drm/i915/gt/intel_mocs.c
index 69b489e8dfed..89570f137b2c 100644
--- a/drivers/gpu/drm/i915/gt/intel_mocs.c
+++ b/drivers/gpu/drm/i915/gt/intel_mocs.c
@@ -40,6 +40,10 @@ struct drm_i915_mocs_table {
 #define LE_COS(value)          ((value) << 15)
 #define LE_SSE(value)          ((value) << 17)
 
+/* Defines for the tables (GLOB_MOCS_0 - GLOB_MOCS_16) */
+#define _L4_CACHEABILITY(value)        ((value) << 2)
+#define IG_PAT(value)          ((value) << 8)
+
 /* Defines for the tables (LNCFMOCS0 - LNCFMOCS31) - two entries per word */
 #define L3_ESC(value)          ((value) << 0)
 #define L3_SCC(value)          ((value) << 1)
@@ -50,6 +54,7 @@ struct drm_i915_mocs_table {
 /* Helper defines */
 #define GEN9_NUM_MOCS_ENTRIES  64  /* 63-64 are reserved, but configured. */
 #define PVC_NUM_MOCS_ENTRIES   3
+#define MTL_NUM_MOCS_ENTRIES   16
 
 /* (e)LLC caching options */
 /*
@@ -73,6 +78,12 @@ struct drm_i915_mocs_table {
 #define L3_2_RESERVED          _L3_CACHEABILITY(2)
 #define L3_3_WB                        _L3_CACHEABILITY(3)
 
+/* L4 caching options */
+#define L4_0_WB                        _L4_CACHEABILITY(0)
+#define L4_1_WT                        _L4_CACHEABILITY(1)
+#define L4_2_RESERVED          _L4_CACHEABILITY(2)
+#define L4_3_UC                        _L4_CACHEABILITY(3)
+
 #define MOCS_ENTRY(__idx, __control_value, __l3cc_value) \
        [__idx] = { \
                .control_value = __control_value, \
@@ -416,6 +427,57 @@ static const struct drm_i915_mocs_entry pvc_mocs_table[] = 
{
        MOCS_ENTRY(2, 0, L3_3_WB),
 };
 
+static const struct drm_i915_mocs_entry mtl_mocs_table[] = {
+       /* Error - Reserved for Non-Use */
+       MOCS_ENTRY(0,
+                  IG_PAT(0),
+                  L3_LKUP(1) | L3_3_WB),
+       /* Cached - L3 + L4 */
+       MOCS_ENTRY(1,
+                  IG_PAT(1),
+                  L3_LKUP(1) | L3_3_WB),
+       /* L4 - GO:L3 */
+       MOCS_ENTRY(2,
+                  IG_PAT(1),
+                  L3_LKUP(1) | L3_1_UC),
+       /* Uncached - GO:L3 */
+       MOCS_ENTRY(3,
+                  IG_PAT(1) | L4_3_UC,
+                  L3_LKUP(1) | L3_1_UC),
+       /* L4 - GO:Mem */
+       MOCS_ENTRY(4,
+                  IG_PAT(1),
+                  L3_LKUP(1) | L3_GLBGO(1) | L3_1_UC),
+       /* Uncached - GO:Mem */
+       MOCS_ENTRY(5,
+                  IG_PAT(1) | L4_3_UC,
+                  L3_LKUP(1) | L3_GLBGO(1) | L3_1_UC),
+       /* L4 - L3:NoLKUP; GO:L3 */
+       MOCS_ENTRY(6,
+                  IG_PAT(1),
+                  L3_1_UC),
+       /* Uncached - L3:NoLKUP; GO:L3 */
+       MOCS_ENTRY(7,
+                  IG_PAT(1) | L4_3_UC,
+                  L3_1_UC),
+       /* L4 - L3:NoLKUP; GO:Mem */
+       MOCS_ENTRY(8,
+                  IG_PAT(1),
+                  L3_GLBGO(1) | L3_1_UC),
+       /* Uncached - L3:NoLKUP; GO:Mem */
+       MOCS_ENTRY(9,
+                  IG_PAT(1) | L4_3_UC,
+                  L3_GLBGO(1) | L3_1_UC),
+       /* Display - L3; L4:WT */
+       MOCS_ENTRY(14,
+                  IG_PAT(1) | L4_1_WT,
+                  L3_LKUP(1) | L3_3_WB),
+       /* CCS - Non-Displayable */
+       MOCS_ENTRY(15,
+                  IG_PAT(1),
+                  L3_GLBGO(1) | L3_1_UC),
+};
+
 enum {
        HAS_GLOBAL_MOCS = BIT(0),
        HAS_ENGINE_MOCS = BIT(1),
@@ -445,7 +507,13 @@ static unsigned int get_mocs_settings(const struct 
drm_i915_private *i915,
        memset(table, 0, sizeof(struct drm_i915_mocs_table));
 
        table->unused_entries_index = I915_MOCS_PTE;
-       if (IS_PONTEVECCHIO(i915)) {
+       if (IS_METEORLAKE(i915)) {
+               table->size = ARRAY_SIZE(mtl_mocs_table);
+               table->table = mtl_mocs_table;
+               table->n_entries = MTL_NUM_MOCS_ENTRIES;
+               table->uc_index = 9;
+               table->unused_entries_index = 1;
+       } else if (IS_PONTEVECCHIO(i915)) {
                table->size = ARRAY_SIZE(pvc_mocs_table);
                table->table = pvc_mocs_table;
                table->n_entries = PVC_NUM_MOCS_ENTRIES;
@@ -646,9 +714,9 @@ void intel_mocs_init_engine(struct intel_engine_cs *engine)
                init_l3cc_table(engine->gt, &table);
 }
 
-static u32 global_mocs_offset(void)
+static u32 global_mocs_offset(struct intel_gt *gt)
 {
-       return i915_mmio_reg_offset(GEN12_GLOBAL_MOCS(0));
+       return i915_mmio_reg_offset(GEN12_GLOBAL_MOCS(0)) + 
gt->uncore->gsi_offset;
 }
 
 void intel_set_mocs_index(struct intel_gt *gt)
@@ -671,7 +739,7 @@ void intel_mocs_init(struct intel_gt *gt)
         */
        flags = get_mocs_settings(gt->i915, &table);
        if (flags & HAS_GLOBAL_MOCS)
-               __init_mocs_table(gt->uncore, &table, global_mocs_offset());
+               __init_mocs_table(gt->uncore, &table, global_mocs_offset(gt));
 
        /*
         * Initialize the L3CC table as part of mocs initalization to make
diff --git a/drivers/gpu/drm/i915/gt/selftest_mocs.c 
b/drivers/gpu/drm/i915/gt/selftest_mocs.c
index ca009a6a13bd..730796346514 100644
--- a/drivers/gpu/drm/i915/gt/selftest_mocs.c
+++ b/drivers/gpu/drm/i915/gt/selftest_mocs.c
@@ -137,7 +137,7 @@ static int read_mocs_table(struct i915_request *rq,
                return 0;
 
        if (HAS_GLOBAL_MOCS_REGISTERS(rq->engine->i915))
-               addr = global_mocs_offset();
+               addr = global_mocs_offset(rq->engine->gt);
        else
                addr = mocs_offset(rq->engine);
 
diff --git a/drivers/gpu/drm/i915/i915_pci.c b/drivers/gpu/drm/i915/i915_pci.c
index 621730b6551c..480b128499ae 100644
--- a/drivers/gpu/drm/i915/i915_pci.c
+++ b/drivers/gpu/drm/i915/i915_pci.c
@@ -1149,6 +1149,7 @@ static const struct intel_device_info mtl_info = {
        .has_flat_ccs = 0,
        .has_gmd_id = 1,
        .has_guc_deprivilege = 1,
+       .has_llc = 0,
        .has_mslice_steering = 0,
        .has_snoop = 1,
        .__runtime.memory_regions = REGION_SMEM | REGION_STOLEN_LMEM,
-- 
2.25.1

Reply via email to