]> git.ipfire.org Git - thirdparty/kernel/stable.git/commitdiff
drm/xe: Decouple TLB invalidations from GT
authorMatthew Brost <matthew.brost@intel.com>
Tue, 26 Aug 2025 18:29:08 +0000 (18:29 +0000)
committerMatthew Brost <matthew.brost@intel.com>
Wed, 27 Aug 2025 18:49:18 +0000 (11:49 -0700)
Decouple TLB invalidations from the GT by updating the TLB invalidation
layer to accept a `struct xe_tlb_inval` instead of a `struct xe_gt`.
Also, rename *gt_tlb* to *tlb*. The internals of the TLB invalidation
code still operate on a GT, but this is now hidden from the rest of the
driver.

Signed-off-by: Stuart Summers <stuart.summers@intel.com>
Reviewed-by: Stuart Summers <stuart.summers@intel.com>
Signed-off-by: Matthew Brost <matthew.brost@intel.com>
Link: https://lore.kernel.org/r/20250826182911.392550-7-stuart.summers@intel.com
18 files changed:
drivers/gpu/drm/xe/Makefile
drivers/gpu/drm/xe/xe_ggtt.c
drivers/gpu/drm/xe/xe_gt.c
drivers/gpu/drm/xe/xe_gt_tlb_inval.h [deleted file]
drivers/gpu/drm/xe/xe_gt_tlb_inval_job.h [deleted file]
drivers/gpu/drm/xe/xe_gt_types.h
drivers/gpu/drm/xe/xe_guc_ct.c
drivers/gpu/drm/xe/xe_lmtt.c
drivers/gpu/drm/xe/xe_migrate.h
drivers/gpu/drm/xe/xe_pt.c
drivers/gpu/drm/xe/xe_svm.c
drivers/gpu/drm/xe/xe_tlb_inval.c [moved from drivers/gpu/drm/xe/xe_gt_tlb_inval.c with 80% similarity]
drivers/gpu/drm/xe/xe_tlb_inval.h [new file with mode: 0644]
drivers/gpu/drm/xe/xe_tlb_inval_job.c [moved from drivers/gpu/drm/xe/xe_gt_tlb_inval_job.c with 51% similarity]
drivers/gpu/drm/xe/xe_tlb_inval_job.h [new file with mode: 0644]
drivers/gpu/drm/xe/xe_tlb_inval_types.h [moved from drivers/gpu/drm/xe/xe_gt_tlb_inval_types.h with 56% similarity]
drivers/gpu/drm/xe/xe_trace.h
drivers/gpu/drm/xe/xe_vm.c

index a1e4b3b1a1b64e3b65ea7001c985cbebf79deb25..2b67debb8bc474ca3f0a1acefab9d08c8302a662 100644 (file)
@@ -61,8 +61,6 @@ xe-y += xe_bb.o \
        xe_gt_pagefault.o \
        xe_gt_sysfs.o \
        xe_gt_throttle.o \
-       xe_gt_tlb_inval.o \
-       xe_gt_tlb_inval_job.o \
        xe_gt_topology.o \
        xe_guc.o \
        xe_guc_ads.o \
@@ -119,6 +117,8 @@ xe-y += xe_bb.o \
        xe_sync.o \
        xe_tile.o \
        xe_tile_sysfs.o \
+       xe_tlb_inval.o \
+       xe_tlb_inval_job.o \
        xe_trace.o \
        xe_trace_bo.o \
        xe_trace_guc.o \
index c3e46c27011736baf22a6f6b58c42812283d2e2a..71c7690a92b3b55a4c188544ad0e8085d9a8e9fa 100644 (file)
 #include "xe_device.h"
 #include "xe_gt.h"
 #include "xe_gt_printk.h"
-#include "xe_gt_tlb_inval.h"
 #include "xe_map.h"
 #include "xe_mmio.h"
 #include "xe_pm.h"
 #include "xe_res_cursor.h"
 #include "xe_sriov.h"
 #include "xe_tile_sriov_vf.h"
+#include "xe_tlb_inval.h"
 #include "xe_wa.h"
 #include "xe_wopcm.h"
 
@@ -438,7 +438,7 @@ static void ggtt_invalidate_gt_tlb(struct xe_gt *gt)
        if (!gt)
                return;
 
-       err = xe_gt_tlb_inval_ggtt(gt);
+       err = xe_tlb_inval_ggtt(&gt->tlb_inval);
        xe_gt_WARN(gt, err, "Failed to invalidate GGTT (%pe)", ERR_PTR(err));
 }
 
index 9a4639732bd7bed7b91a39ff0d08c0e0afa345d1..67ee7cdbd6ec36b15a8a8c38ad9d979d0de71669 100644 (file)
@@ -37,7 +37,6 @@
 #include "xe_gt_sriov_pf.h"
 #include "xe_gt_sriov_vf.h"
 #include "xe_gt_sysfs.h"
-#include "xe_gt_tlb_inval.h"
 #include "xe_gt_topology.h"
 #include "xe_guc_exec_queue_types.h"
 #include "xe_guc_pc.h"
@@ -58,6 +57,7 @@
 #include "xe_sa.h"
 #include "xe_sched_job.h"
 #include "xe_sriov.h"
+#include "xe_tlb_inval.h"
 #include "xe_tuning.h"
 #include "xe_uc.h"
 #include "xe_uc_fw.h"
@@ -852,7 +852,7 @@ static int gt_reset(struct xe_gt *gt)
 
        xe_uc_stop(&gt->uc);
 
-       xe_gt_tlb_inval_reset(gt);
+       xe_tlb_inval_reset(&gt->tlb_inval);
 
        err = do_gt_reset(gt);
        if (err)
@@ -1066,5 +1066,5 @@ void xe_gt_declare_wedged(struct xe_gt *gt)
        xe_gt_assert(gt, gt_to_xe(gt)->wedged.mode);
 
        xe_uc_declare_wedged(&gt->uc);
-       xe_gt_tlb_inval_reset(gt);
+       xe_tlb_inval_reset(&gt->tlb_inval);
 }
diff --git a/drivers/gpu/drm/xe/xe_gt_tlb_inval.h b/drivers/gpu/drm/xe/xe_gt_tlb_inval.h
deleted file mode 100644 (file)
index b1258ac..0000000
+++ /dev/null
@@ -1,41 +0,0 @@
-/* SPDX-License-Identifier: MIT */
-/*
- * Copyright © 2023 Intel Corporation
- */
-
-#ifndef _XE_GT_TLB_INVAL_H_
-#define _XE_GT_TLB_INVAL_H_
-
-#include <linux/types.h>
-
-#include "xe_gt_tlb_inval_types.h"
-
-struct xe_gt;
-struct xe_guc;
-struct xe_vm;
-struct xe_vma;
-
-int xe_gt_tlb_inval_init_early(struct xe_gt *gt);
-void xe_gt_tlb_inval_fini(struct xe_gt *gt);
-
-void xe_gt_tlb_inval_reset(struct xe_gt *gt);
-int xe_gt_tlb_inval_ggtt(struct xe_gt *gt);
-void xe_gt_tlb_inval_vm(struct xe_gt *gt, struct xe_vm *vm);
-int xe_gt_tlb_inval_all(struct xe_gt *gt, struct xe_gt_tlb_inval_fence *fence);
-int xe_gt_tlb_inval_range(struct xe_gt *gt,
-                         struct xe_gt_tlb_inval_fence *fence,
-                         u64 start, u64 end, u32 asid);
-int xe_guc_tlb_inval_done_handler(struct xe_guc *guc, u32 *msg, u32 len);
-
-void xe_gt_tlb_inval_fence_init(struct xe_gt *gt,
-                               struct xe_gt_tlb_inval_fence *fence,
-                               bool stack);
-void xe_gt_tlb_inval_fence_signal(struct xe_gt_tlb_inval_fence *fence);
-
-static inline void
-xe_gt_tlb_inval_fence_wait(struct xe_gt_tlb_inval_fence *fence)
-{
-       dma_fence_wait(&fence->base, false);
-}
-
-#endif /* _XE_GT_TLB_INVAL_ */
diff --git a/drivers/gpu/drm/xe/xe_gt_tlb_inval_job.h b/drivers/gpu/drm/xe/xe_gt_tlb_inval_job.h
deleted file mode 100644 (file)
index 8838961..0000000
+++ /dev/null
@@ -1,34 +0,0 @@
-/* SPDX-License-Identifier: MIT */
-/*
- * Copyright © 2025 Intel Corporation
- */
-
-#ifndef _XE_GT_TLB_INVAL_JOB_H_
-#define _XE_GT_TLB_INVAL_JOB_H_
-
-#include <linux/types.h>
-
-struct dma_fence;
-struct drm_sched_job;
-struct kref;
-struct xe_exec_queue;
-struct xe_gt;
-struct xe_gt_tlb_inval_job;
-struct xe_migrate;
-
-struct xe_gt_tlb_inval_job *xe_gt_tlb_inval_job_create(struct xe_exec_queue *q,
-                                                      struct xe_gt *gt,
-                                                      u64 start, u64 end,
-                                                      u32 asid);
-
-int xe_gt_tlb_inval_job_alloc_dep(struct xe_gt_tlb_inval_job *job);
-
-struct dma_fence *xe_gt_tlb_inval_job_push(struct xe_gt_tlb_inval_job *job,
-                                          struct xe_migrate *m,
-                                          struct dma_fence *fence);
-
-void xe_gt_tlb_inval_job_get(struct xe_gt_tlb_inval_job *job);
-
-void xe_gt_tlb_inval_job_put(struct xe_gt_tlb_inval_job *job);
-
-#endif
index 7dc5a3f310f155ae582e8b3b87c25e0f2491477f..66158105aca5ad7137b84ef7699100354d188b4d 100644 (file)
 #include "xe_gt_sriov_pf_types.h"
 #include "xe_gt_sriov_vf_types.h"
 #include "xe_gt_stats_types.h"
-#include "xe_gt_tlb_inval_types.h"
 #include "xe_hw_engine_types.h"
 #include "xe_hw_fence_types.h"
 #include "xe_oa_types.h"
 #include "xe_reg_sr_types.h"
 #include "xe_sa_types.h"
+#include "xe_tlb_inval_types.h"
 #include "xe_uc_types.h"
 
 struct xe_exec_queue_ops;
index 9131d121d941cf958cd018b68db487735f16d653..5f38041cff4c143bc82e67d0216708a5617262d9 100644 (file)
 #include "xe_gt_sriov_pf_control.h"
 #include "xe_gt_sriov_pf_monitor.h"
 #include "xe_gt_sriov_printk.h"
-#include "xe_gt_tlb_inval.h"
 #include "xe_guc.h"
 #include "xe_guc_log.h"
 #include "xe_guc_relay.h"
 #include "xe_guc_submit.h"
 #include "xe_map.h"
 #include "xe_pm.h"
+#include "xe_tlb_inval.h"
 #include "xe_trace_guc.h"
 
 static void receive_g2h(struct xe_guc_ct *ct);
index e5aba03ff8ac808341d75e6fe16f49299e7ce388..f2bfbfa3efa16b3c946bcb6673a9c55d87fe6ca3 100644 (file)
@@ -11,7 +11,7 @@
 
 #include "xe_assert.h"
 #include "xe_bo.h"
-#include "xe_gt_tlb_inval.h"
+#include "xe_tlb_inval.h"
 #include "xe_lmtt.h"
 #include "xe_map.h"
 #include "xe_mmio.h"
@@ -228,8 +228,8 @@ void xe_lmtt_init_hw(struct xe_lmtt *lmtt)
 
 static int lmtt_invalidate_hw(struct xe_lmtt *lmtt)
 {
-       struct xe_gt_tlb_inval_fence fences[XE_MAX_GT_PER_TILE];
-       struct xe_gt_tlb_inval_fence *fence = fences;
+       struct xe_tlb_inval_fence fences[XE_MAX_GT_PER_TILE];
+       struct xe_tlb_inval_fence *fence = fences;
        struct xe_tile *tile = lmtt_to_tile(lmtt);
        struct xe_gt *gt;
        int result = 0;
@@ -237,8 +237,8 @@ static int lmtt_invalidate_hw(struct xe_lmtt *lmtt)
        u8 id;
 
        for_each_gt_on_tile(gt, tile, id) {
-               xe_gt_tlb_inval_fence_init(gt, fence, true);
-               err = xe_gt_tlb_inval_all(gt, fence);
+               xe_tlb_inval_fence_init(&gt->tlb_inval, fence, true);
+               err = xe_tlb_inval_all(&gt->tlb_inval, fence);
                result = result ?: err;
                fence++;
        }
@@ -252,7 +252,7 @@ static int lmtt_invalidate_hw(struct xe_lmtt *lmtt)
         */
        fence = fences;
        for_each_gt_on_tile(gt, tile, id)
-               xe_gt_tlb_inval_fence_wait(fence++);
+               xe_tlb_inval_fence_wait(fence++);
 
        return result;
 }
index 8978d2cc1a75bdaac8d25216db3b5f2c6b39c72e..4fad324b625356c99d53da2e7fada4e996d06126 100644 (file)
@@ -15,7 +15,7 @@ struct ttm_resource;
 
 struct xe_bo;
 struct xe_gt;
-struct xe_gt_tlb_inval_job;
+struct xe_tlb_inval_job;
 struct xe_exec_queue;
 struct xe_migrate;
 struct xe_migrate_pt_update;
@@ -94,13 +94,13 @@ struct xe_migrate_pt_update {
        /** @job: The job if a GPU page-table update. NULL otherwise */
        struct xe_sched_job *job;
        /**
-        * @ijob: The GT TLB invalidation job for primary tile. NULL otherwise
+        * @ijob: The TLB invalidation job for primary GT. NULL otherwise
         */
-       struct xe_gt_tlb_inval_job *ijob;
+       struct xe_tlb_inval_job *ijob;
        /**
-        * @mjob: The GT TLB invalidation job for media tile. NULL otherwise
+        * @mjob: The TLB invalidation job for media GT. NULL otherwise
         */
-       struct xe_gt_tlb_inval_job *mjob;
+       struct xe_tlb_inval_job *mjob;
        /** @tile_id: Tile ID of the update */
        u8 tile_id;
 };
index c0a70c80dff9a8de6bd34305acdab18966673971..c129048a9a09e1beca7b4bb89f6d3d2bdb12c6c4 100644 (file)
@@ -13,7 +13,7 @@
 #include "xe_drm_client.h"
 #include "xe_exec_queue.h"
 #include "xe_gt.h"
-#include "xe_gt_tlb_inval_job.h"
+#include "xe_tlb_inval_job.h"
 #include "xe_migrate.h"
 #include "xe_pt_types.h"
 #include "xe_pt_walk.h"
@@ -21,6 +21,7 @@
 #include "xe_sched_job.h"
 #include "xe_sync.h"
 #include "xe_svm.h"
+#include "xe_tlb_inval_job.h"
 #include "xe_trace.h"
 #include "xe_ttm_stolen_mgr.h"
 #include "xe_vm.h"
@@ -1276,8 +1277,8 @@ static int op_add_deps(struct xe_vm *vm, struct xe_vma_op *op,
 }
 
 static int xe_pt_vm_dependencies(struct xe_sched_job *job,
-                                struct xe_gt_tlb_inval_job *ijob,
-                                struct xe_gt_tlb_inval_job *mjob,
+                                struct xe_tlb_inval_job *ijob,
+                                struct xe_tlb_inval_job *mjob,
                                 struct xe_vm *vm,
                                 struct xe_vma_ops *vops,
                                 struct xe_vm_pgtable_update_ops *pt_update_ops,
@@ -1347,13 +1348,13 @@ static int xe_pt_vm_dependencies(struct xe_sched_job *job,
 
        if (job) {
                if (ijob) {
-                       err = xe_gt_tlb_inval_job_alloc_dep(ijob);
+                       err = xe_tlb_inval_job_alloc_dep(ijob);
                        if (err)
                                return err;
                }
 
                if (mjob) {
-                       err = xe_gt_tlb_inval_job_alloc_dep(mjob);
+                       err = xe_tlb_inval_job_alloc_dep(mjob);
                        if (err)
                                return err;
                }
@@ -2353,6 +2354,15 @@ static const struct xe_migrate_pt_update_ops svm_migrate_ops = {
 static const struct xe_migrate_pt_update_ops svm_migrate_ops;
 #endif
 
+static struct xe_dep_scheduler *to_dep_scheduler(struct xe_exec_queue *q,
+                                                struct xe_gt *gt)
+{
+       if (xe_gt_is_media_type(gt))
+               return q->tlb_inval[XE_EXEC_QUEUE_TLB_INVAL_MEDIA_GT].dep_scheduler;
+
+       return q->tlb_inval[XE_EXEC_QUEUE_TLB_INVAL_PRIMARY_GT].dep_scheduler;
+}
+
 /**
  * xe_pt_update_ops_run() - Run PT update operations
  * @tile: Tile of PT update operations
@@ -2371,7 +2381,7 @@ xe_pt_update_ops_run(struct xe_tile *tile, struct xe_vma_ops *vops)
        struct xe_vm_pgtable_update_ops *pt_update_ops =
                &vops->pt_update_ops[tile->id];
        struct dma_fence *fence, *ifence, *mfence;
-       struct xe_gt_tlb_inval_job *ijob = NULL, *mjob = NULL;
+       struct xe_tlb_inval_job *ijob = NULL, *mjob = NULL;
        struct dma_fence **fences = NULL;
        struct dma_fence_array *cf = NULL;
        struct xe_range_fence *rfence;
@@ -2403,11 +2413,15 @@ xe_pt_update_ops_run(struct xe_tile *tile, struct xe_vma_ops *vops)
 #endif
 
        if (pt_update_ops->needs_invalidation) {
-               ijob = xe_gt_tlb_inval_job_create(pt_update_ops->q,
-                                                 tile->primary_gt,
-                                                 pt_update_ops->start,
-                                                 pt_update_ops->last,
-                                                 vm->usm.asid);
+               struct xe_exec_queue *q = pt_update_ops->q;
+               struct xe_dep_scheduler *dep_scheduler =
+                       to_dep_scheduler(q, tile->primary_gt);
+
+               ijob = xe_tlb_inval_job_create(q, &tile->primary_gt->tlb_inval,
+                                              dep_scheduler,
+                                              pt_update_ops->start,
+                                              pt_update_ops->last,
+                                              vm->usm.asid);
                if (IS_ERR(ijob)) {
                        err = PTR_ERR(ijob);
                        goto kill_vm_tile1;
@@ -2415,11 +2429,14 @@ xe_pt_update_ops_run(struct xe_tile *tile, struct xe_vma_ops *vops)
                update.ijob = ijob;
 
                if (tile->media_gt) {
-                       mjob = xe_gt_tlb_inval_job_create(pt_update_ops->q,
-                                                         tile->media_gt,
-                                                         pt_update_ops->start,
-                                                         pt_update_ops->last,
-                                                         vm->usm.asid);
+                       dep_scheduler = to_dep_scheduler(q, tile->media_gt);
+
+                       mjob = xe_tlb_inval_job_create(q,
+                                                      &tile->media_gt->tlb_inval,
+                                                      dep_scheduler,
+                                                      pt_update_ops->start,
+                                                      pt_update_ops->last,
+                                                      vm->usm.asid);
                        if (IS_ERR(mjob)) {
                                err = PTR_ERR(mjob);
                                goto free_ijob;
@@ -2470,13 +2487,13 @@ xe_pt_update_ops_run(struct xe_tile *tile, struct xe_vma_ops *vops)
        if (ijob) {
                struct dma_fence *__fence;
 
-               ifence = xe_gt_tlb_inval_job_push(ijob, tile->migrate, fence);
+               ifence = xe_tlb_inval_job_push(ijob, tile->migrate, fence);
                __fence = ifence;
 
                if (mjob) {
                        fences[0] = ifence;
-                       mfence = xe_gt_tlb_inval_job_push(mjob, tile->migrate,
-                                                         fence);
+                       mfence = xe_tlb_inval_job_push(mjob, tile->migrate,
+                                                      fence);
                        fences[1] = mfence;
 
                        dma_fence_array_init(cf, 2, fences,
@@ -2519,8 +2536,8 @@ xe_pt_update_ops_run(struct xe_tile *tile, struct xe_vma_ops *vops)
        if (pt_update_ops->needs_userptr_lock)
                up_read(&vm->userptr.notifier_lock);
 
-       xe_gt_tlb_inval_job_put(mjob);
-       xe_gt_tlb_inval_job_put(ijob);
+       xe_tlb_inval_job_put(mjob);
+       xe_tlb_inval_job_put(ijob);
 
        return fence;
 
@@ -2529,8 +2546,8 @@ free_rfence:
 free_ijob:
        kfree(cf);
        kfree(fences);
-       xe_gt_tlb_inval_job_put(mjob);
-       xe_gt_tlb_inval_job_put(ijob);
+       xe_tlb_inval_job_put(mjob);
+       xe_tlb_inval_job_put(ijob);
 kill_vm_tile1:
        if (err != -EAGAIN && err != -ENODATA && tile->id)
                xe_vm_kill(vops->vm, false);
index babc64485b52a0665eeabb7980458c847afb0791..76c6d74c1208cb1d2bd24b7c214d732a25bc54c9 100644 (file)
@@ -7,7 +7,6 @@
 
 #include "xe_bo.h"
 #include "xe_gt_stats.h"
-#include "xe_gt_tlb_inval.h"
 #include "xe_migrate.h"
 #include "xe_module.h"
 #include "xe_pm.h"
similarity index 80%
rename from drivers/gpu/drm/xe/xe_gt_tlb_inval.c
rename to drivers/gpu/drm/xe/xe_tlb_inval.c
index 37b3b45ec230e38688d4e5f8b475815a0ee03825..f4b7c0c7489475e9456e3d6d2ff802316dcdc8fb 100644 (file)
@@ -13,7 +13,7 @@
 #include "xe_guc.h"
 #include "xe_guc_ct.h"
 #include "xe_gt_stats.h"
-#include "xe_gt_tlb_inval.h"
+#include "xe_tlb_inval.h"
 #include "xe_mmio.h"
 #include "xe_pm.h"
 #include "xe_sriov.h"
@@ -38,40 +38,47 @@ static long tlb_timeout_jiffies(struct xe_gt *gt)
        return hw_tlb_timeout + 2 * delay;
 }
 
-static void xe_gt_tlb_inval_fence_fini(struct xe_gt_tlb_inval_fence *fence)
+static void xe_tlb_inval_fence_fini(struct xe_tlb_inval_fence *fence)
 {
-       if (WARN_ON_ONCE(!fence->gt))
+       struct xe_gt *gt;
+
+       if (WARN_ON_ONCE(!fence->tlb_inval))
                return;
 
-       xe_pm_runtime_put(gt_to_xe(fence->gt));
-       fence->gt = NULL; /* fini() should be called once */
+       gt = fence->tlb_inval->private;
+
+       xe_pm_runtime_put(gt_to_xe(gt));
+       fence->tlb_inval = NULL; /* fini() should be called once */
 }
 
 static void
-__inval_fence_signal(struct xe_device *xe, struct xe_gt_tlb_inval_fence *fence)
+__inval_fence_signal(struct xe_device *xe, struct xe_tlb_inval_fence *fence)
 {
        bool stack = test_bit(FENCE_STACK_BIT, &fence->base.flags);
 
-       trace_xe_gt_tlb_inval_fence_signal(xe, fence);
-       xe_gt_tlb_inval_fence_fini(fence);
+       trace_xe_tlb_inval_fence_signal(xe, fence);
+       xe_tlb_inval_fence_fini(fence);
        dma_fence_signal(&fence->base);
        if (!stack)
                dma_fence_put(&fence->base);
 }
 
 static void
-inval_fence_signal(struct xe_device *xe, struct xe_gt_tlb_inval_fence *fence)
+inval_fence_signal(struct xe_device *xe, struct xe_tlb_inval_fence *fence)
 {
        list_del(&fence->link);
        __inval_fence_signal(xe, fence);
 }
 
-void xe_gt_tlb_inval_fence_signal(struct xe_gt_tlb_inval_fence *fence)
+void xe_tlb_inval_fence_signal(struct xe_tlb_inval_fence *fence)
 {
-       if (WARN_ON_ONCE(!fence->gt))
+       struct xe_gt *gt;
+
+       if (WARN_ON_ONCE(!fence->tlb_inval))
                return;
 
-       __inval_fence_signal(gt_to_xe(fence->gt), fence);
+       gt = fence->tlb_inval->private;
+       __inval_fence_signal(gt_to_xe(gt), fence);
 }
 
 static void xe_gt_tlb_fence_timeout(struct work_struct *work)
@@ -79,7 +86,7 @@ static void xe_gt_tlb_fence_timeout(struct work_struct *work)
        struct xe_gt *gt = container_of(work, struct xe_gt,
                                        tlb_inval.fence_tdr.work);
        struct xe_device *xe = gt_to_xe(gt);
-       struct xe_gt_tlb_inval_fence *fence, *next;
+       struct xe_tlb_inval_fence *fence, *next;
 
        LNL_FLUSH_WORK(&gt->uc.guc.ct.g2h_worker);
 
@@ -92,7 +99,7 @@ static void xe_gt_tlb_fence_timeout(struct work_struct *work)
                if (msecs_to_jiffies(since_inval_ms) < tlb_timeout_jiffies(gt))
                        break;
 
-               trace_xe_gt_tlb_inval_fence_timeout(xe, fence);
+               trace_xe_tlb_inval_fence_timeout(xe, fence);
                xe_gt_err(gt, "TLB invalidation fence timeout, seqno=%d recv=%d",
                          fence->seqno, gt->tlb_inval.seqno_recv);
 
@@ -110,7 +117,7 @@ static void xe_gt_tlb_fence_timeout(struct work_struct *work)
  * xe_gt_tlb_inval_init_early - Initialize GT TLB invalidation state
  * @gt: GT structure
  *
- * Initialize GT TLB invalidation state, purely software initialization, should
+ * Initialize TLB invalidation state, purely software initialization, should
  * be called once during driver load.
  *
  * Return: 0 on success, negative error code on error.
@@ -120,6 +127,7 @@ int xe_gt_tlb_inval_init_early(struct xe_gt *gt)
        struct xe_device *xe = gt_to_xe(gt);
        int err;
 
+       gt->tlb_inval.private = gt;
        gt->tlb_inval.seqno = 1;
        INIT_LIST_HEAD(&gt->tlb_inval.pending_fences);
        spin_lock_init(&gt->tlb_inval.pending_lock);
@@ -141,14 +149,15 @@ int xe_gt_tlb_inval_init_early(struct xe_gt *gt)
 }
 
 /**
- * xe_gt_tlb_inval_reset - Initialize GT TLB invalidation reset
- * @gt: GT structure
+ * xe_tlb_inval_reset - Initialize TLB invalidation reset
+ * @tlb_inval: TLB invalidation client
  *
  * Signal any pending invalidation fences, should be called during a GT reset
  */
-void xe_gt_tlb_inval_reset(struct xe_gt *gt)
+void xe_tlb_inval_reset(struct xe_tlb_inval *tlb_inval)
 {
-       struct xe_gt_tlb_inval_fence *fence, *next;
+       struct xe_gt *gt = tlb_inval->private;
+       struct xe_tlb_inval_fence *fence, *next;
        int pending_seqno;
 
        /*
@@ -213,7 +222,7 @@ static bool tlb_inval_seqno_past(struct xe_gt *gt, int seqno)
 }
 
 static int send_tlb_inval(struct xe_guc *guc,
-                         struct xe_gt_tlb_inval_fence *fence,
+                         struct xe_tlb_inval_fence *fence,
                          u32 *action, int len)
 {
        struct xe_gt *gt = guc_to_gt(guc);
@@ -232,7 +241,7 @@ static int send_tlb_inval(struct xe_guc *guc,
        mutex_lock(&gt->tlb_inval.seqno_lock);
        seqno = gt->tlb_inval.seqno;
        fence->seqno = seqno;
-       trace_xe_gt_tlb_inval_fence_send(xe, fence);
+       trace_xe_tlb_inval_fence_send(xe, fence);
        action[1] = seqno;
        ret = xe_guc_ct_send(&guc->ct, action, len,
                             G2H_LEN_DW_TLB_INVALIDATE, 1);
@@ -277,7 +286,7 @@ static int send_tlb_inval(struct xe_guc *guc,
                XE_GUC_TLB_INVAL_FLUSH_CACHE)
 
 /**
- * xe_gt_tlb_inval_guc - Issue a TLB invalidation on this GT for the GuC
+ * xe_tlb_inval_guc - Issue a TLB invalidation on this GT for the GuC
  * @gt: GT structure
  * @fence: invalidation fence which will be signal on TLB invalidation
  * completion
@@ -287,8 +296,8 @@ static int send_tlb_inval(struct xe_guc *guc,
  *
  * Return: 0 on success, negative error code on error
  */
-static int xe_gt_tlb_inval_guc(struct xe_gt *gt,
-                              struct xe_gt_tlb_inval_fence *fence)
+static int xe_tlb_inval_guc(struct xe_gt *gt,
+                           struct xe_tlb_inval_fence *fence)
 {
        u32 action[] = {
                XE_GUC_ACTION_TLB_INVALIDATION,
@@ -309,30 +318,31 @@ static int xe_gt_tlb_inval_guc(struct xe_gt *gt,
 }
 
 /**
- * xe_gt_tlb_inval_ggtt - Issue a TLB invalidation on this GT for the GGTT
- * @gt: GT structure
+ * xe_tlb_inval_ggtt - Issue a TLB invalidation on this GT for the GGTT
+ * @tlb_inval: TLB invalidation client
  *
  * Issue a TLB invalidation for the GGTT. Completion of TLB invalidation is
  * synchronous.
  *
  * Return: 0 on success, negative error code on error
  */
-int xe_gt_tlb_inval_ggtt(struct xe_gt *gt)
+int xe_tlb_inval_ggtt(struct xe_tlb_inval *tlb_inval)
 {
+       struct xe_gt *gt = tlb_inval->private;
        struct xe_device *xe = gt_to_xe(gt);
        unsigned int fw_ref;
 
        if (xe_guc_ct_enabled(&gt->uc.guc.ct) &&
            gt->uc.guc.submission_state.enabled) {
-               struct xe_gt_tlb_inval_fence fence;
+               struct xe_tlb_inval_fence fence;
                int ret;
 
-               xe_gt_tlb_inval_fence_init(gt, &fence, true);
-               ret = xe_gt_tlb_inval_guc(gt, &fence);
+               xe_tlb_inval_fence_init(tlb_inval, &fence, true);
+               ret = xe_tlb_inval_guc(gt, &fence);
                if (ret)
                        return ret;
 
-               xe_gt_tlb_inval_fence_wait(&fence);
+               xe_tlb_inval_fence_wait(&fence);
        } else if (xe_device_uc_enabled(xe) && !xe_device_wedged(xe)) {
                struct xe_mmio *mmio = &gt->mmio;
 
@@ -355,14 +365,17 @@ int xe_gt_tlb_inval_ggtt(struct xe_gt *gt)
        return 0;
 }
 
-static int send_tlb_inval_all(struct xe_gt *gt,
-                             struct xe_gt_tlb_inval_fence *fence)
+static int send_tlb_inval_all(struct xe_tlb_inval *tlb_inval,
+                             struct xe_tlb_inval_fence *fence)
 {
        u32 action[] = {
                XE_GUC_ACTION_TLB_INVALIDATION_ALL,
                0,  /* seqno, replaced in send_tlb_inval */
                MAKE_INVAL_OP(XE_GUC_TLB_INVAL_FULL),
        };
+       struct xe_gt *gt = tlb_inval->private;
+
+       xe_gt_assert(gt, fence);
 
        return send_tlb_inval(&gt->uc.guc, fence, action, ARRAY_SIZE(action));
 }
@@ -370,19 +383,19 @@ static int send_tlb_inval_all(struct xe_gt *gt,
 /**
  * xe_gt_tlb_invalidation_all - Invalidate all TLBs across PF and all VFs.
  * @gt: the &xe_gt structure
- * @fence: the &xe_gt_tlb_inval_fence to be signaled on completion
+ * @fence: the &xe_tlb_inval_fence to be signaled on completion
  *
  * Send a request to invalidate all TLBs across PF and all VFs.
  *
  * Return: 0 on success, negative error code on error
  */
-int xe_gt_tlb_inval_all(struct xe_gt *gt, struct xe_gt_tlb_inval_fence *fence)
+int xe_tlb_inval_all(struct xe_tlb_inval *tlb_inval,
+                    struct xe_tlb_inval_fence *fence)
 {
+       struct xe_gt *gt = tlb_inval->private;
        int err;
 
-       xe_gt_assert(gt, gt == fence->gt);
-
-       err = send_tlb_inval_all(gt, fence);
+       err = send_tlb_inval_all(tlb_inval, fence);
        if (err)
                xe_gt_err(gt, "TLB invalidation request failed (%pe)", ERR_PTR(err));
 
@@ -397,9 +410,8 @@ int xe_gt_tlb_inval_all(struct xe_gt *gt, struct xe_gt_tlb_inval_fence *fence)
 #define MAX_RANGE_TLB_INVALIDATION_LENGTH (rounddown_pow_of_two(ULONG_MAX))
 
 /**
- * xe_gt_tlb_inval_range - Issue a TLB invalidation on this GT for an address range
- *
- * @gt: GT structure
+ * xe_tlb_inval_range - Issue a TLB invalidation on this GT for an address range
+ * @tlb_inval: TLB invalidation client
  * @fence: invalidation fence which will be signal on TLB invalidation
  * completion
  * @start: start address
@@ -412,9 +424,11 @@ int xe_gt_tlb_inval_all(struct xe_gt *gt, struct xe_gt_tlb_inval_fence *fence)
  *
  * Return: Negative error code on error, 0 on success
  */
-int xe_gt_tlb_inval_range(struct xe_gt *gt, struct xe_gt_tlb_inval_fence *fence,
-                         u64 start, u64 end, u32 asid)
+int xe_tlb_inval_range(struct xe_tlb_inval *tlb_inval,
+                      struct xe_tlb_inval_fence *fence, u64 start, u64 end,
+                      u32 asid)
 {
+       struct xe_gt *gt = tlb_inval->private;
        struct xe_device *xe = gt_to_xe(gt);
 #define MAX_TLB_INVALIDATION_LEN       7
        u32 action[MAX_TLB_INVALIDATION_LEN];
@@ -484,38 +498,38 @@ int xe_gt_tlb_inval_range(struct xe_gt *gt, struct xe_gt_tlb_inval_fence *fence,
 }
 
 /**
- * xe_gt_tlb_inval_vm - Issue a TLB invalidation on this GT for a VM
- * @gt: graphics tile
+ * xe_tlb_inval_vm - Issue a TLB invalidation on this GT for a VM
+ * @tlb_inval: TLB invalidation client
  * @vm: VM to invalidate
  *
  * Invalidate entire VM's address space
  */
-void xe_gt_tlb_inval_vm(struct xe_gt *gt, struct xe_vm *vm)
+void xe_tlb_inval_vm(struct xe_tlb_inval *tlb_inval, struct xe_vm *vm)
 {
-       struct xe_gt_tlb_inval_fence fence;
+       struct xe_tlb_inval_fence fence;
        u64 range = 1ull << vm->xe->info.va_bits;
        int ret;
 
-       xe_gt_tlb_inval_fence_init(gt, &fence, true);
+       xe_tlb_inval_fence_init(tlb_inval, &fence, true);
 
-       ret = xe_gt_tlb_inval_range(gt, &fence, 0, range, vm->usm.asid);
+       ret = xe_tlb_inval_range(tlb_inval, &fence, 0, range, vm->usm.asid);
        if (ret < 0)
                return;
 
-       xe_gt_tlb_inval_fence_wait(&fence);
+       xe_tlb_inval_fence_wait(&fence);
 }
 
 /**
- * xe_gt_tlb_inval_done_handler - GT TLB invalidation done handler
+ * xe_tlb_inval_done_handler - TLB invalidation done handler
  * @gt: gt
  * @seqno: seqno of invalidation that is done
  *
- * Update recv seqno, signal any GT TLB invalidation fences, and restart TDR
+ * Update recv seqno, signal any TLB invalidation fences, and restart TDR
  */
-static void xe_gt_tlb_inval_done_handler(struct xe_gt *gt, int seqno)
+static void xe_tlb_inval_done_handler(struct xe_gt *gt, int seqno)
 {
        struct xe_device *xe = gt_to_xe(gt);
-       struct xe_gt_tlb_inval_fence *fence, *next;
+       struct xe_tlb_inval_fence *fence, *next;
        unsigned long flags;
 
        /*
@@ -543,7 +557,7 @@ static void xe_gt_tlb_inval_done_handler(struct xe_gt *gt, int seqno)
 
        list_for_each_entry_safe(fence, next,
                                 &gt->tlb_inval.pending_fences, link) {
-               trace_xe_gt_tlb_inval_fence_recv(xe, fence);
+               trace_xe_tlb_inval_fence_recv(xe, fence);
 
                if (!tlb_inval_seqno_past(gt, fence->seqno))
                        break;
@@ -580,7 +594,7 @@ int xe_guc_tlb_inval_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
        if (unlikely(len != 1))
                return -EPROTO;
 
-       xe_gt_tlb_inval_done_handler(gt, msg[0]);
+       xe_tlb_inval_done_handler(gt, msg[0]);
 
        return 0;
 }
@@ -603,19 +617,21 @@ static const struct dma_fence_ops inval_fence_ops = {
 };
 
 /**
- * xe_gt_tlb_inval_fence_init - Initialize TLB invalidation fence
- * @gt: GT
+ * xe_tlb_inval_fence_init - Initialize TLB invalidation fence
+ * @tlb_inval: TLB invalidation client
  * @fence: TLB invalidation fence to initialize
  * @stack: fence is stack variable
  *
- * Initialize TLB invalidation fence for use. xe_gt_tlb_inval_fence_fini
+ * Initialize TLB invalidation fence for use. xe_tlb_inval_fence_fini
  * will be automatically called when fence is signalled (all fences must signal),
  * even on error.
  */
-void xe_gt_tlb_inval_fence_init(struct xe_gt *gt,
-                               struct xe_gt_tlb_inval_fence *fence,
-                               bool stack)
+void xe_tlb_inval_fence_init(struct xe_tlb_inval *tlb_inval,
+                            struct xe_tlb_inval_fence *fence,
+                            bool stack)
 {
+       struct xe_gt *gt = tlb_inval->private;
+
        xe_pm_runtime_get_noresume(gt_to_xe(gt));
 
        spin_lock_irq(&gt->tlb_inval.lock);
@@ -628,5 +644,5 @@ void xe_gt_tlb_inval_fence_init(struct xe_gt *gt,
                set_bit(FENCE_STACK_BIT, &fence->base.flags);
        else
                dma_fence_get(&fence->base);
-       fence->gt = gt;
+       fence->tlb_inval = tlb_inval;
 }
diff --git a/drivers/gpu/drm/xe/xe_tlb_inval.h b/drivers/gpu/drm/xe/xe_tlb_inval.h
new file mode 100644 (file)
index 0000000..ab6f769
--- /dev/null
@@ -0,0 +1,41 @@
+/* SPDX-License-Identifier: MIT */
+/*
+ * Copyright © 2025 Intel Corporation
+ */
+
+#ifndef _XE_TLB_INVAL_H_
+#define _XE_TLB_INVAL_H_
+
+#include <linux/types.h>
+
+#include "xe_tlb_inval_types.h"
+
+struct xe_gt;
+struct xe_guc;
+struct xe_vm;
+
+int xe_gt_tlb_inval_init_early(struct xe_gt *gt);
+void xe_gt_tlb_inval_fini(struct xe_gt *gt);
+
+void xe_tlb_inval_reset(struct xe_tlb_inval *tlb_inval);
+int xe_tlb_inval_ggtt(struct xe_tlb_inval *tlb_inval);
+void xe_tlb_inval_vm(struct xe_tlb_inval *tlb_inval, struct xe_vm *vm);
+int xe_tlb_inval_all(struct xe_tlb_inval *tlb_inval,
+                    struct xe_tlb_inval_fence *fence);
+int xe_tlb_inval_range(struct xe_tlb_inval *tlb_inval,
+                      struct xe_tlb_inval_fence *fence,
+                      u64 start, u64 end, u32 asid);
+int xe_guc_tlb_inval_done_handler(struct xe_guc *guc, u32 *msg, u32 len);
+
+void xe_tlb_inval_fence_init(struct xe_tlb_inval *tlb_inval,
+                            struct xe_tlb_inval_fence *fence,
+                            bool stack);
+void xe_tlb_inval_fence_signal(struct xe_tlb_inval_fence *fence);
+
+static inline void
+xe_tlb_inval_fence_wait(struct xe_tlb_inval_fence *fence)
+{
+       dma_fence_wait(&fence->base, false);
+}
+
+#endif /* _XE_TLB_INVAL_ */
similarity index 51%
rename from drivers/gpu/drm/xe/xe_gt_tlb_inval_job.c
rename to drivers/gpu/drm/xe/xe_tlb_inval_job.c
index 41e0ea92ea5a257f1ab15cd0e6ae24722cfd5207..492def04a55953390fb43b8cf083e5689d24121e 100644 (file)
@@ -3,21 +3,22 @@
  * Copyright © 2025 Intel Corporation
  */
 
+#include "xe_assert.h"
 #include "xe_dep_job_types.h"
 #include "xe_dep_scheduler.h"
 #include "xe_exec_queue.h"
-#include "xe_gt.h"
-#include "xe_gt_tlb_inval.h"
-#include "xe_gt_tlb_inval_job.h"
+#include "xe_gt_types.h"
+#include "xe_tlb_inval.h"
+#include "xe_tlb_inval_job.h"
 #include "xe_migrate.h"
 #include "xe_pm.h"
 
-/** struct xe_gt_tlb_inval_job - GT TLB invalidation job */
-struct xe_gt_tlb_inval_job {
+/** struct xe_tlb_inval_job - TLB invalidation job */
+struct xe_tlb_inval_job {
        /** @dep: base generic dependency Xe job */
        struct xe_dep_job dep;
-       /** @gt: GT to invalidate */
-       struct xe_gt *gt;
+       /** @tlb_inval: TLB invalidation client */
+       struct xe_tlb_inval *tlb_inval;
        /** @q: exec queue issuing the invalidate */
        struct xe_exec_queue *q;
        /** @refcount: ref count of this job */
@@ -37,63 +38,56 @@ struct xe_gt_tlb_inval_job {
        bool fence_armed;
 };
 
-static struct dma_fence *xe_gt_tlb_inval_job_run(struct xe_dep_job *dep_job)
+static struct dma_fence *xe_tlb_inval_job_run(struct xe_dep_job *dep_job)
 {
-       struct xe_gt_tlb_inval_job *job =
+       struct xe_tlb_inval_job *job =
                container_of(dep_job, typeof(*job), dep);
-       struct xe_gt_tlb_inval_fence *ifence =
+       struct xe_tlb_inval_fence *ifence =
                container_of(job->fence, typeof(*ifence), base);
 
-       xe_gt_tlb_inval_range(job->gt, ifence, job->start,
-                             job->end, job->asid);
+       xe_tlb_inval_range(job->tlb_inval, ifence, job->start,
+                          job->end, job->asid);
 
        return job->fence;
 }
 
-static void xe_gt_tlb_inval_job_free(struct xe_dep_job *dep_job)
+static void xe_tlb_inval_job_free(struct xe_dep_job *dep_job)
 {
-       struct xe_gt_tlb_inval_job *job =
+       struct xe_tlb_inval_job *job =
                container_of(dep_job, typeof(*job), dep);
 
-       /* Pairs with get in xe_gt_tlb_inval_job_push */
-       xe_gt_tlb_inval_job_put(job);
+       /* Pairs with get in xe_tlb_inval_job_push */
+       xe_tlb_inval_job_put(job);
 }
 
 static const struct xe_dep_job_ops dep_job_ops = {
-       .run_job = xe_gt_tlb_inval_job_run,
-       .free_job = xe_gt_tlb_inval_job_free,
+       .run_job = xe_tlb_inval_job_run,
+       .free_job = xe_tlb_inval_job_free,
 };
 
-static int xe_gt_tlb_inval_context(struct xe_gt *gt)
-{
-       return xe_gt_is_media_type(gt) ? XE_EXEC_QUEUE_TLB_INVAL_MEDIA_GT :
-               XE_EXEC_QUEUE_TLB_INVAL_PRIMARY_GT;
-}
-
 /**
- * xe_gt_tlb_inval_job_create() - GT TLB invalidation job create
- * @gt: GT to invalidate
+ * xe_tlb_inval_job_create() - TLB invalidation job create
  * @q: exec queue issuing the invalidate
+ * @tlb_inval: TLB invalidation client
+ * @dep_scheduler: Dependency scheduler for job
  * @start: Start address to invalidate
  * @end: End address to invalidate
  * @asid: Address space ID to invalidate
  *
- * Create a GT TLB invalidation job and initialize internal fields. The caller is
+ * Create a TLB invalidation job and initialize internal fields. The caller is
  * responsible for releasing the creation reference.
  *
- * Return: GT TLB invalidation job object on success, ERR_PTR failure
+ * Return: TLB invalidation job object on success, ERR_PTR failure
  */
-struct xe_gt_tlb_inval_job *xe_gt_tlb_inval_job_create(struct xe_exec_queue *q,
-                                                      struct xe_gt *gt,
-                                                      u64 start, u64 end,
-                                                      u32 asid)
+struct xe_tlb_inval_job *
+xe_tlb_inval_job_create(struct xe_exec_queue *q, struct xe_tlb_inval *tlb_inval,
+                       struct xe_dep_scheduler *dep_scheduler, u64 start,
+                       u64 end, u32 asid)
 {
-       struct xe_gt_tlb_inval_job *job;
-       struct xe_dep_scheduler *dep_scheduler =
-               q->tlb_inval[xe_gt_tlb_inval_context(gt)].dep_scheduler;
+       struct xe_tlb_inval_job *job;
        struct drm_sched_entity *entity =
                xe_dep_scheduler_entity(dep_scheduler);
-       struct xe_gt_tlb_inval_fence *ifence;
+       struct xe_tlb_inval_fence *ifence;
        int err;
 
        job = kmalloc(sizeof(*job), GFP_KERNEL);
@@ -101,14 +95,14 @@ struct xe_gt_tlb_inval_job *xe_gt_tlb_inval_job_create(struct xe_exec_queue *q,
                return ERR_PTR(-ENOMEM);
 
        job->q = q;
-       job->gt = gt;
+       job->tlb_inval = tlb_inval;
        job->start = start;
        job->end = end;
        job->asid = asid;
        job->fence_armed = false;
        job->dep.ops = &dep_job_ops;
        kref_init(&job->refcount);
-       xe_exec_queue_get(q);   /* Pairs with put in xe_gt_tlb_inval_job_destroy */
+       xe_exec_queue_get(q);   /* Pairs with put in xe_tlb_inval_job_destroy */
 
        ifence = kmalloc(sizeof(*ifence), GFP_KERNEL);
        if (!ifence) {
@@ -122,8 +116,8 @@ struct xe_gt_tlb_inval_job *xe_gt_tlb_inval_job_create(struct xe_exec_queue *q,
        if (err)
                goto err_fence;
 
-       /* Pairs with put in xe_gt_tlb_inval_job_destroy */
-       xe_pm_runtime_get_noresume(gt_to_xe(job->gt));
+       /* Pairs with put in xe_tlb_inval_job_destroy */
+       xe_pm_runtime_get_noresume(gt_to_xe(q->gt));
 
        return job;
 
@@ -136,40 +130,40 @@ err_job:
        return ERR_PTR(err);
 }
 
-static void xe_gt_tlb_inval_job_destroy(struct kref *ref)
+static void xe_tlb_inval_job_destroy(struct kref *ref)
 {
-       struct xe_gt_tlb_inval_job *job = container_of(ref, typeof(*job),
-                                                      refcount);
-       struct xe_gt_tlb_inval_fence *ifence =
+       struct xe_tlb_inval_job *job = container_of(ref, typeof(*job),
+                                                   refcount);
+       struct xe_tlb_inval_fence *ifence =
                container_of(job->fence, typeof(*ifence), base);
-       struct xe_device *xe = gt_to_xe(job->gt);
        struct xe_exec_queue *q = job->q;
+       struct xe_device *xe = gt_to_xe(q->gt);
 
        if (!job->fence_armed)
                kfree(ifence);
        else
-               /* Ref from xe_gt_tlb_inval_fence_init */
+               /* Ref from xe_tlb_inval_fence_init */
                dma_fence_put(job->fence);
 
        drm_sched_job_cleanup(&job->dep.drm);
        kfree(job);
-       xe_exec_queue_put(q);   /* Pairs with get from xe_gt_tlb_inval_job_create */
-       xe_pm_runtime_put(xe);  /* Pairs with get from xe_gt_tlb_inval_job_create */
+       xe_exec_queue_put(q);   /* Pairs with get from xe_tlb_inval_job_create */
+       xe_pm_runtime_put(xe);  /* Pairs with get from xe_tlb_inval_job_create */
 }
 
 /**
- * xe_gt_tlb_inval_alloc_dep() - GT TLB invalidation job alloc dependency
- * @job: GT TLB invalidation job to alloc dependency for
+ * xe_tlb_inval_alloc_dep() - TLB invalidation job alloc dependency
+ * @job: TLB invalidation job to alloc dependency for
  *
- * Allocate storage for a dependency in the GT TLB invalidation fence. This
+ * Allocate storage for a dependency in the TLB invalidation fence. This
  * function should be called at most once per job and must be paired with
- * xe_gt_tlb_inval_job_push being called with a real fence.
+ * xe_tlb_inval_job_push being called with a real fence.
  *
  * Return: 0 on success, -errno on failure
  */
-int xe_gt_tlb_inval_job_alloc_dep(struct xe_gt_tlb_inval_job *job)
+int xe_tlb_inval_job_alloc_dep(struct xe_tlb_inval_job *job)
 {
-       xe_assert(gt_to_xe(job->gt), !xa_load(&job->dep.drm.dependencies, 0));
+       xe_assert(gt_to_xe(job->q->gt), !xa_load(&job->dep.drm.dependencies, 0));
        might_alloc(GFP_KERNEL);
 
        return drm_sched_job_add_dependency(&job->dep.drm,
@@ -177,24 +171,24 @@ int xe_gt_tlb_inval_job_alloc_dep(struct xe_gt_tlb_inval_job *job)
 }
 
 /**
- * xe_gt_tlb_inval_job_push() - GT TLB invalidation job push
- * @job: GT TLB invalidation job to push
+ * xe_tlb_inval_job_push() - TLB invalidation job push
+ * @job: TLB invalidation job to push
  * @m: The migration object being used
- * @fence: Dependency for GT TLB invalidation job
+ * @fence: Dependency for TLB invalidation job
  *
- * Pushes a GT TLB invalidation job for execution, using @fence as a dependency.
- * Storage for @fence must be preallocated with xe_gt_tlb_inval_job_alloc_dep
+ * Pushes a TLB invalidation job for execution, using @fence as a dependency.
+ * Storage for @fence must be preallocated with xe_tlb_inval_job_alloc_dep
  * prior to this call if @fence is not signaled. Takes a reference to the job’s
  * finished fence, which the caller is responsible for releasing, and return it
  * to the caller. This function is safe to be called in the path of reclaim.
  *
  * Return: Job's finished fence on success, cannot fail
  */
-struct dma_fence *xe_gt_tlb_inval_job_push(struct xe_gt_tlb_inval_job *job,
-                                          struct xe_migrate *m,
-                                          struct dma_fence *fence)
+struct dma_fence *xe_tlb_inval_job_push(struct xe_tlb_inval_job *job,
+                                       struct xe_migrate *m,
+                                       struct dma_fence *fence)
 {
-       struct xe_gt_tlb_inval_fence *ifence =
+       struct xe_tlb_inval_fence *ifence =
                container_of(job->fence, typeof(*ifence), base);
 
        if (!dma_fence_is_signaled(fence)) {
@@ -202,20 +196,20 @@ struct dma_fence *xe_gt_tlb_inval_job_push(struct xe_gt_tlb_inval_job *job,
 
                /*
                 * Can be in path of reclaim, hence the preallocation of fence
-                * storage in xe_gt_tlb_inval_job_alloc_dep. Verify caller did
+                * storage in xe_tlb_inval_job_alloc_dep. Verify caller did
                 * this correctly.
                 */
-               xe_assert(gt_to_xe(job->gt),
+               xe_assert(gt_to_xe(job->q->gt),
                          xa_load(&job->dep.drm.dependencies, 0) ==
                          dma_fence_get_stub());
 
                dma_fence_get(fence);   /* ref released once dependency processed by scheduler */
                ptr = xa_store(&job->dep.drm.dependencies, 0, fence,
                               GFP_ATOMIC);
-               xe_assert(gt_to_xe(job->gt), !xa_is_err(ptr));
+               xe_assert(gt_to_xe(job->q->gt), !xa_is_err(ptr));
        }
 
-       xe_gt_tlb_inval_job_get(job);   /* Pairs with put in free_job */
+       xe_tlb_inval_job_get(job);      /* Pairs with put in free_job */
        job->fence_armed = true;
 
        /*
@@ -225,8 +219,8 @@ struct dma_fence *xe_gt_tlb_inval_job_push(struct xe_gt_tlb_inval_job *job,
         */
        xe_migrate_job_lock(m, job->q);
 
-       /* Creation ref pairs with put in xe_gt_tlb_inval_job_destroy */
-       xe_gt_tlb_inval_fence_init(job->gt, ifence, false);
+       /* Creation ref pairs with put in xe_tlb_inval_job_destroy */
+       xe_tlb_inval_fence_init(job->tlb_inval, ifence, false);
        dma_fence_get(job->fence);      /* Pairs with put in DRM scheduler */
 
        drm_sched_job_arm(&job->dep.drm);
@@ -241,7 +235,7 @@ struct dma_fence *xe_gt_tlb_inval_job_push(struct xe_gt_tlb_inval_job *job,
 
        /*
         * Not using job->fence, as it has its own dma-fence context, which does
-        * not allow GT TLB invalidation fences on the same queue, GT tuple to
+        * not allow TLB invalidation fences on the same queue, GT tuple to
         * be squashed in dma-resv/DRM scheduler. Instead, we use the DRM scheduler
         * context and job's finished fence, which enables squashing.
         */
@@ -249,26 +243,26 @@ struct dma_fence *xe_gt_tlb_inval_job_push(struct xe_gt_tlb_inval_job *job,
 }
 
 /**
- * xe_gt_tlb_inval_job_get() - Get a reference to GT TLB invalidation job
- * @job: GT TLB invalidation job object
+ * xe_tlb_inval_job_get() - Get a reference to TLB invalidation job
+ * @job: TLB invalidation job object
  *
- * Increment the GT TLB invalidation job's reference count
+ * Increment the TLB invalidation job's reference count
  */
-void xe_gt_tlb_inval_job_get(struct xe_gt_tlb_inval_job *job)
+void xe_tlb_inval_job_get(struct xe_tlb_inval_job *job)
 {
        kref_get(&job->refcount);
 }
 
 /**
- * xe_gt_tlb_inval_job_put() - Put a reference to GT TLB invalidation job
- * @job: GT TLB invalidation job object
+ * xe_tlb_inval_job_put() - Put a reference to TLB invalidation job
+ * @job: TLB invalidation job object
  *
- * Decrement the GT TLB invalidation job's reference count, call
- * xe_gt_tlb_inval_job_destroy when reference count == 0. Skips decrement if
+ * Decrement the TLB invalidation job's reference count, call
+ * xe_tlb_inval_job_destroy when reference count == 0. Skips decrement if
  * input @job is NULL or IS_ERR.
  */
-void xe_gt_tlb_inval_job_put(struct xe_gt_tlb_inval_job *job)
+void xe_tlb_inval_job_put(struct xe_tlb_inval_job *job)
 {
        if (!IS_ERR_OR_NULL(job))
-               kref_put(&job->refcount, xe_gt_tlb_inval_job_destroy);
+               kref_put(&job->refcount, xe_tlb_inval_job_destroy);
 }
diff --git a/drivers/gpu/drm/xe/xe_tlb_inval_job.h b/drivers/gpu/drm/xe/xe_tlb_inval_job.h
new file mode 100644 (file)
index 0000000..e63edcb
--- /dev/null
@@ -0,0 +1,33 @@
+/* SPDX-License-Identifier: MIT */
+/*
+ * Copyright © 2025 Intel Corporation
+ */
+
+#ifndef _XE_TLB_INVAL_JOB_H_
+#define _XE_TLB_INVAL_JOB_H_
+
+#include <linux/types.h>
+
+struct dma_fence;
+struct xe_dep_scheduler;
+struct xe_exec_queue;
+struct xe_tlb_inval;
+struct xe_tlb_inval_job;
+struct xe_migrate;
+
+struct xe_tlb_inval_job *
+xe_tlb_inval_job_create(struct xe_exec_queue *q, struct xe_tlb_inval *tlb_inval,
+                       struct xe_dep_scheduler *dep_scheduler,
+                       u64 start, u64 end, u32 asid);
+
+int xe_tlb_inval_job_alloc_dep(struct xe_tlb_inval_job *job);
+
+struct dma_fence *xe_tlb_inval_job_push(struct xe_tlb_inval_job *job,
+                                       struct xe_migrate *m,
+                                       struct dma_fence *fence);
+
+void xe_tlb_inval_job_get(struct xe_tlb_inval_job *job);
+
+void xe_tlb_inval_job_put(struct xe_tlb_inval_job *job);
+
+#endif
similarity index 56%
rename from drivers/gpu/drm/xe/xe_gt_tlb_inval_types.h
rename to drivers/gpu/drm/xe/xe_tlb_inval_types.h
index 442f72b78ccf31c8317c52ae3eb1c0d7bd609cf3..6d14b9f17b915d61e69dd23efffc1e6a3de5b809 100644 (file)
@@ -3,58 +3,57 @@
  * Copyright © 2023 Intel Corporation
  */
 
-#ifndef _XE_GT_TLB_INVAL_TYPES_H_
-#define _XE_GT_TLB_INVAL_TYPES_H_
+#ifndef _XE_TLB_INVAL_TYPES_H_
+#define _XE_TLB_INVAL_TYPES_H_
 
 #include <linux/workqueue.h>
 #include <linux/dma-fence.h>
 
-struct xe_gt;
-
 /** struct xe_tlb_inval - TLB invalidation client */
 struct xe_tlb_inval {
+       /** @private: Backend private pointer */
+       void *private;
        /** @tlb_inval.seqno: TLB invalidation seqno, protected by CT lock */
 #define TLB_INVALIDATION_SEQNO_MAX     0x100000
        int seqno;
        /** @tlb_invalidation.seqno_lock: protects @tlb_invalidation.seqno */
        struct mutex seqno_lock;
        /**
-        * @tlb_inval.seqno_recv: last received TLB invalidation seqno,
-        * protected by CT lock
+        * @seqno_recv: last received TLB invalidation seqno, protected by
+        * CT lock
         */
        int seqno_recv;
        /**
-        * @tlb_inval.pending_fences: list of pending fences waiting TLB
-        * invaliations, protected by CT lock
+        * @pending_fences: list of pending fences waiting TLB invaliations,
+        * protected CT lock
         */
        struct list_head pending_fences;
        /**
-        * @tlb_inval.pending_lock: protects @tlb_inval.pending_fences
-        * and updating @tlb_inval.seqno_recv.
+        * @pending_lock: protects @pending_fences and updating @seqno_recv.
         */
        spinlock_t pending_lock;
        /**
-        * @tlb_inval.fence_tdr: schedules a delayed call to
-        * xe_gt_tlb_fence_timeout after the timeut interval is over.
+        * @fence_tdr: schedules a delayed call to xe_tlb_fence_timeout after
+        * the timeout interval is over.
         */
        struct delayed_work fence_tdr;
-       /** @wtlb_invalidation.wq: schedules GT TLB invalidation jobs */
+       /** @job_wq: schedules TLB invalidation jobs */
        struct workqueue_struct *job_wq;
        /** @tlb_inval.lock: protects TLB invalidation fences */
        spinlock_t lock;
 };
 
 /**
- * struct xe_gt_tlb_inval_fence - XE GT TLB invalidation fence
+ * struct xe_tlb_inval_fence - TLB invalidation fence
  *
- * Optionally passed to xe_gt_tlb_inval and will be signaled upon TLB
+ * Optionally passed to xe_tlb_inval* functions and will be signaled upon TLB
  * invalidation completion.
  */
-struct xe_gt_tlb_inval_fence {
+struct xe_tlb_inval_fence {
        /** @base: dma fence base */
        struct dma_fence base;
-       /** @gt: GT which fence belong to */
-       struct xe_gt *gt;
+       /** @tlb_inval: TLB invalidation client which fence belong to */
+       struct xe_tlb_inval *tlb_inval;
        /** @link: link into list of pending tlb fences */
        struct list_head link;
        /** @seqno: seqno of TLB invalidation to signal fence one */
index 36538f50d06ff185f3bfd73f5ccbfd3630a3b732..314f42fcbcbd93a5f52a06e32bfd9b7047498bb0 100644 (file)
 
 #include "xe_exec_queue_types.h"
 #include "xe_gpu_scheduler_types.h"
-#include "xe_gt_tlb_inval_types.h"
 #include "xe_gt_types.h"
 #include "xe_guc_exec_queue_types.h"
 #include "xe_sched_job.h"
+#include "xe_tlb_inval_types.h"
 #include "xe_vm.h"
 
 #define __dev_name_xe(xe)      dev_name((xe)->drm.dev)
 #define __dev_name_gt(gt)      __dev_name_xe(gt_to_xe((gt)))
 #define __dev_name_eq(q)       __dev_name_gt((q)->gt)
 
-DECLARE_EVENT_CLASS(xe_gt_tlb_inval_fence,
-                   TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_inval_fence *fence),
+DECLARE_EVENT_CLASS(xe_tlb_inval_fence,
+                   TP_PROTO(struct xe_device *xe, struct xe_tlb_inval_fence *fence),
                    TP_ARGS(xe, fence),
 
                    TP_STRUCT__entry(
                             __string(dev, __dev_name_xe(xe))
-                            __field(struct xe_gt_tlb_inval_fence *, fence)
+                            __field(struct xe_tlb_inval_fence *, fence)
                             __field(int, seqno)
                             ),
 
@@ -45,23 +45,23 @@ DECLARE_EVENT_CLASS(xe_gt_tlb_inval_fence,
                              __get_str(dev), __entry->fence, __entry->seqno)
 );
 
-DEFINE_EVENT(xe_gt_tlb_inval_fence, xe_gt_tlb_inval_fence_send,
-            TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_inval_fence *fence),
+DEFINE_EVENT(xe_tlb_inval_fence, xe_tlb_inval_fence_send,
+            TP_PROTO(struct xe_device *xe, struct xe_tlb_inval_fence *fence),
             TP_ARGS(xe, fence)
 );
 
-DEFINE_EVENT(xe_gt_tlb_inval_fence, xe_gt_tlb_inval_fence_recv,
-            TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_inval_fence *fence),
+DEFINE_EVENT(xe_tlb_inval_fence, xe_tlb_inval_fence_recv,
+            TP_PROTO(struct xe_device *xe, struct xe_tlb_inval_fence *fence),
             TP_ARGS(xe, fence)
 );
 
-DEFINE_EVENT(xe_gt_tlb_inval_fence, xe_gt_tlb_inval_fence_signal,
-            TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_inval_fence *fence),
+DEFINE_EVENT(xe_tlb_inval_fence, xe_tlb_inval_fence_signal,
+            TP_PROTO(struct xe_device *xe, struct xe_tlb_inval_fence *fence),
             TP_ARGS(xe, fence)
 );
 
-DEFINE_EVENT(xe_gt_tlb_inval_fence, xe_gt_tlb_inval_fence_timeout,
-            TP_PROTO(struct xe_device *xe, struct xe_gt_tlb_inval_fence *fence),
+DEFINE_EVENT(xe_tlb_inval_fence, xe_tlb_inval_fence_timeout,
+            TP_PROTO(struct xe_device *xe, struct xe_tlb_inval_fence *fence),
             TP_ARGS(xe, fence)
 );
 
index b758e9a6837aa686a54eef69cbaf6c9fa7e6fc32..3ff3c67aa79d90d8ec2f3097dcd82445045dd6d6 100644 (file)
@@ -28,7 +28,6 @@
 #include "xe_drm_client.h"
 #include "xe_exec_queue.h"
 #include "xe_gt_pagefault.h"
-#include "xe_gt_tlb_inval.h"
 #include "xe_migrate.h"
 #include "xe_pat.h"
 #include "xe_pm.h"
@@ -39,6 +38,7 @@
 #include "xe_svm.h"
 #include "xe_sync.h"
 #include "xe_tile.h"
+#include "xe_tlb_inval.h"
 #include "xe_trace_bo.h"
 #include "xe_wa.h"
 #include "xe_hmm.h"
@@ -1898,7 +1898,7 @@ static void xe_vm_close(struct xe_vm *vm)
                                        xe_pt_clear(xe, vm->pt_root[id]);
 
                        for_each_gt(gt, xe, id)
-                               xe_gt_tlb_inval_vm(gt, vm);
+                               xe_tlb_inval_vm(&gt->tlb_inval, vm);
                }
        }
 
@@ -4046,7 +4046,7 @@ void xe_vm_unlock(struct xe_vm *vm)
 int xe_vm_range_tilemask_tlb_inval(struct xe_vm *vm, u64 start,
                                   u64 end, u8 tile_mask)
 {
-       struct xe_gt_tlb_inval_fence
+       struct xe_tlb_inval_fence
                fence[XE_MAX_TILES_PER_DEVICE * XE_MAX_GT_PER_TILE];
        struct xe_tile *tile;
        u32 fence_id = 0;
@@ -4060,11 +4060,12 @@ int xe_vm_range_tilemask_tlb_inval(struct xe_vm *vm, u64 start,
                if (!(tile_mask & BIT(id)))
                        continue;
 
-               xe_gt_tlb_inval_fence_init(tile->primary_gt,
-                                          &fence[fence_id], true);
+               xe_tlb_inval_fence_init(&tile->primary_gt->tlb_inval,
+                                       &fence[fence_id], true);
 
-               err = xe_gt_tlb_inval_range(tile->primary_gt, &fence[fence_id],
-                                           start, end, vm->usm.asid);
+               err = xe_tlb_inval_range(&tile->primary_gt->tlb_inval,
+                                        &fence[fence_id], start, end,
+                                        vm->usm.asid);
                if (err)
                        goto wait;
                ++fence_id;
@@ -4072,11 +4073,12 @@ int xe_vm_range_tilemask_tlb_inval(struct xe_vm *vm, u64 start,
                if (!tile->media_gt)
                        continue;
 
-               xe_gt_tlb_inval_fence_init(tile->media_gt,
-                                          &fence[fence_id], true);
+               xe_tlb_inval_fence_init(&tile->media_gt->tlb_inval,
+                                       &fence[fence_id], true);
 
-               err = xe_gt_tlb_inval_range(tile->media_gt, &fence[fence_id],
-                                           start, end, vm->usm.asid);
+               err = xe_tlb_inval_range(&tile->media_gt->tlb_inval,
+                                        &fence[fence_id], start, end,
+                                        vm->usm.asid);
                if (err)
                        goto wait;
                ++fence_id;
@@ -4084,7 +4086,7 @@ int xe_vm_range_tilemask_tlb_inval(struct xe_vm *vm, u64 start,
 
 wait:
        for (id = 0; id < fence_id; ++id)
-               xe_gt_tlb_inval_fence_wait(&fence[id]);
+               xe_tlb_inval_fence_wait(&fence[id]);
 
        return err;
 }