]> git.ipfire.org Git - thirdparty/linux.git/commitdiff
drm/xe/bo: Add a bo remove callback
authorThomas Hellström <thomas.hellstrom@linux.intel.com>
Wed, 26 Mar 2025 08:05:49 +0000 (09:05 +0100)
committerThomas Hellström <thomas.hellstrom@linux.intel.com>
Thu, 27 Mar 2025 11:07:57 +0000 (12:07 +0100)
On device unbind, migrate exported bos, including pagemap bos to
system. This allows importers to take proper action without
disruption. In particular, SVM clients on remote devices may
continue as if nothing happened, and can chose a different
placement.

The evict_flags() placement is chosen in such a way that bos that
aren't exported are purged.

For pinned bos, we unmap DMA, but their pages are not freed yet
since we can't be 100% sure they are not accessed.

All pinned external bos (not just the VRAM ones) are put on the
pinned.external list with this patch. But this only affects the
xe_bo_pci_dev_remove_pinned() function since !VRAM bos are
ignored by the suspend / resume functionality. As a follow-up we
could look at removing the suspend / resume iteration over
pinned external bos since we currently don't allow pinning
external bos in VRAM, and other external bos don't need any
special treatment at suspend / resume.

v2:
- Address review comments. (Matthew Auld).
v3:
- Don't introduce an external_evicted list (Matthew Auld)
- Add a discussion around suspend / resume behaviour to the
  commit message.
- Formatting fixes.
v4:
- Move dma-unmaps of pinned kernel bos to a dev managed
  callback to give subsystems using these bos a chance to
  clean them up. (Matthew Auld)

Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Reviewed-by: Matthew Auld <matthew.auld@intel.com>
Link: https://lore.kernel.org/r/20250326080551.40201-4-thomas.hellstrom@linux.intel.com
drivers/gpu/drm/xe/xe_bo.c
drivers/gpu/drm/xe/xe_bo.h
drivers/gpu/drm/xe/xe_bo_evict.c
drivers/gpu/drm/xe/xe_bo_evict.h
drivers/gpu/drm/xe/xe_device.c
drivers/gpu/drm/xe/xe_device_types.h

index 64f9c936eea06383eb282eb0f73ca4b92a0bb412..9d043d2c30fdd4d5c99e90de0b80842c1d3050b4 100644 (file)
@@ -55,6 +55,8 @@ static struct ttm_placement sys_placement = {
        .placement = &sys_placement_flags,
 };
 
+static struct ttm_placement purge_placement;
+
 static const struct ttm_place tt_placement_flags[] = {
        {
                .fpfn = 0,
@@ -281,6 +283,8 @@ int xe_bo_placement_for_flags(struct xe_device *xe, struct xe_bo *bo,
 static void xe_evict_flags(struct ttm_buffer_object *tbo,
                           struct ttm_placement *placement)
 {
+       struct xe_device *xe = container_of(tbo->bdev, typeof(*xe), ttm);
+       bool device_unplugged = drm_dev_is_unplugged(&xe->drm);
        struct xe_bo *bo;
 
        if (!xe_bo_is_xe_bo(tbo)) {
@@ -290,7 +294,7 @@ static void xe_evict_flags(struct ttm_buffer_object *tbo,
                        return;
                }
 
-               *placement = sys_placement;
+               *placement = device_unplugged ? purge_placement : sys_placement;
                return;
        }
 
@@ -300,6 +304,11 @@ static void xe_evict_flags(struct ttm_buffer_object *tbo,
                return;
        }
 
+       if (device_unplugged && !tbo->base.dma_buf) {
+               *placement = purge_placement;
+               return;
+       }
+
        /*
         * For xe, sg bos that are evicted to system just triggers a
         * rebind of the sg list upon subsequent validation to XE_PL_TT.
@@ -657,11 +666,20 @@ static int xe_bo_move_dmabuf(struct ttm_buffer_object *ttm_bo,
        struct xe_ttm_tt *xe_tt = container_of(ttm_bo->ttm, struct xe_ttm_tt,
                                               ttm);
        struct xe_device *xe = ttm_to_xe_device(ttm_bo->bdev);
+       bool device_unplugged = drm_dev_is_unplugged(&xe->drm);
        struct sg_table *sg;
 
        xe_assert(xe, attach);
        xe_assert(xe, ttm_bo->ttm);
 
+       if (device_unplugged && new_res->mem_type == XE_PL_SYSTEM &&
+           ttm_bo->sg) {
+               dma_resv_wait_timeout(ttm_bo->base.resv, DMA_RESV_USAGE_BOOKKEEP,
+                                     false, MAX_SCHEDULE_TIMEOUT);
+               dma_buf_unmap_attachment(attach, ttm_bo->sg, DMA_BIDIRECTIONAL);
+               ttm_bo->sg = NULL;
+       }
+
        if (new_res->mem_type == XE_PL_SYSTEM)
                goto out;
 
@@ -1224,6 +1242,31 @@ err_res_free:
        return ret;
 }
 
+int xe_bo_dma_unmap_pinned(struct xe_bo *bo)
+{
+       struct ttm_buffer_object *ttm_bo = &bo->ttm;
+       struct ttm_tt *tt = ttm_bo->ttm;
+
+       if (tt) {
+               struct xe_ttm_tt *xe_tt = container_of(tt, typeof(*xe_tt), ttm);
+
+               if (ttm_bo->type == ttm_bo_type_sg && ttm_bo->sg) {
+                       dma_buf_unmap_attachment(ttm_bo->base.import_attach,
+                                                ttm_bo->sg,
+                                                DMA_BIDIRECTIONAL);
+                       ttm_bo->sg = NULL;
+                       xe_tt->sg = NULL;
+               } else if (xe_tt->sg) {
+                       dma_unmap_sgtable(xe_tt->xe->drm.dev, xe_tt->sg,
+                                         DMA_BIDIRECTIONAL, 0);
+                       sg_free_table(xe_tt->sg);
+                       xe_tt->sg = NULL;
+               }
+       }
+
+       return 0;
+}
+
 static unsigned long xe_ttm_io_mem_pfn(struct ttm_buffer_object *ttm_bo,
                                       unsigned long page_offset)
 {
@@ -2102,12 +2145,9 @@ int xe_bo_pin_external(struct xe_bo *bo)
                if (err)
                        return err;
 
-               if (xe_bo_is_vram(bo)) {
-                       spin_lock(&xe->pinned.lock);
-                       list_add_tail(&bo->pinned_link,
-                                     &xe->pinned.external_vram);
-                       spin_unlock(&xe->pinned.lock);
-               }
+               spin_lock(&xe->pinned.lock);
+               list_add_tail(&bo->pinned_link, &xe->pinned.external);
+               spin_unlock(&xe->pinned.lock);
        }
 
        ttm_bo_pin(&bo->ttm);
index ec3e4446d0272966b7ccc0a4848f0d5730783273..05479240bf75deec216f44c580338ef624a7347d 100644 (file)
@@ -276,6 +276,8 @@ int xe_bo_evict(struct xe_bo *bo, bool force_alloc);
 int xe_bo_evict_pinned(struct xe_bo *bo);
 int xe_bo_restore_pinned(struct xe_bo *bo);
 
+int xe_bo_dma_unmap_pinned(struct xe_bo *bo);
+
 extern const struct ttm_device_funcs xe_ttm_funcs;
 extern const char *const xe_mem_type_to_name[];
 
index 1eeb3910450b9d080a857dc84021fc2eeb6b9dec..a1f0661e7b0c861801da8cef6e5c9e3142b8223e 100644 (file)
@@ -93,8 +93,8 @@ int xe_bo_evict_all(struct xe_device *xe)
                }
        }
 
-       ret = xe_bo_apply_to_pinned(xe, &xe->pinned.external_vram,
-                                   &xe->pinned.external_vram,
+       ret = xe_bo_apply_to_pinned(xe, &xe->pinned.external,
+                                   &xe->pinned.external,
                                    xe_bo_evict_pinned);
 
        /*
@@ -181,8 +181,8 @@ int xe_bo_restore_user(struct xe_device *xe)
                return 0;
 
        /* Pinned user memory in VRAM should be validated on resume */
-       ret = xe_bo_apply_to_pinned(xe, &xe->pinned.external_vram,
-                                   &xe->pinned.external_vram,
+       ret = xe_bo_apply_to_pinned(xe, &xe->pinned.external,
+                                   &xe->pinned.external,
                                    xe_bo_restore_pinned);
 
        /* Wait for restore to complete */
@@ -191,3 +191,79 @@ int xe_bo_restore_user(struct xe_device *xe)
 
        return ret;
 }
+
+static void xe_bo_pci_dev_remove_pinned(struct xe_device *xe)
+{
+       struct xe_tile *tile;
+       unsigned int id;
+
+       (void)xe_bo_apply_to_pinned(xe, &xe->pinned.external,
+                                   &xe->pinned.external,
+                                   xe_bo_dma_unmap_pinned);
+       for_each_tile(tile, xe, id)
+               xe_tile_migrate_wait(tile);
+}
+
+/**
+ * xe_bo_pci_dev_remove_all() - Handle bos when the pci_device is about to be removed
+ * @xe: The xe device.
+ *
+ * On pci_device removal we need to drop all dma mappings and move
+ * the data of exported bos out to system. This includes SVM bos and
+ * exported dma-buf bos. This is done by evicting all bos, but
+ * the evict placement in xe_evict_flags() is chosen such that all
+ * bos except those mentioned are purged, and thus their memory
+ * is released.
+ *
+ * For pinned bos, we're unmapping dma.
+ */
+void xe_bo_pci_dev_remove_all(struct xe_device *xe)
+{
+       unsigned int mem_type;
+
+       /*
+        * Move pagemap bos and exported dma-buf to system, and
+        * purge everything else.
+        */
+       for (mem_type = XE_PL_VRAM1; mem_type >= XE_PL_TT; --mem_type) {
+               struct ttm_resource_manager *man =
+                       ttm_manager_type(&xe->ttm, mem_type);
+
+               if (man) {
+                       int ret = ttm_resource_manager_evict_all(&xe->ttm, man);
+
+                       drm_WARN_ON(&xe->drm, ret);
+               }
+       }
+
+       xe_bo_pci_dev_remove_pinned(xe);
+}
+
+static void xe_bo_pinned_fini(void *arg)
+{
+       struct xe_device *xe = arg;
+
+       (void)xe_bo_apply_to_pinned(xe, &xe->pinned.kernel_bo_present,
+                                   &xe->pinned.kernel_bo_present,
+                                   xe_bo_dma_unmap_pinned);
+}
+
+/**
+ * xe_bo_pinned_init() - Initialize pinned bo tracking
+ * @xe: The xe device.
+ *
+ * Initializes the lists and locks required for pinned bo
+ * tracking and registers a callback to dma-unmap
+ * any remaining pinned bos on pci device removal.
+ *
+ * Return: %0 on success, negative error code on error.
+ */
+int xe_bo_pinned_init(struct xe_device *xe)
+{
+       spin_lock_init(&xe->pinned.lock);
+       INIT_LIST_HEAD(&xe->pinned.kernel_bo_present);
+       INIT_LIST_HEAD(&xe->pinned.external);
+       INIT_LIST_HEAD(&xe->pinned.evicted);
+
+       return devm_add_action_or_reset(xe->drm.dev, xe_bo_pinned_fini, xe);
+}
index 746894798852d212e8c95b8b1b62a32f27d1bdb2..0708d50ddfa8c6c3844434c09aa96ecfba7d8173 100644 (file)
@@ -12,4 +12,7 @@ int xe_bo_evict_all(struct xe_device *xe);
 int xe_bo_restore_kernel(struct xe_device *xe);
 int xe_bo_restore_user(struct xe_device *xe);
 
+void xe_bo_pci_dev_remove_all(struct xe_device *xe);
+
+int xe_bo_pinned_init(struct xe_device *xe);
 #endif
index 1ffb7d1f6be6652bd41b534f0691ff5aea04d572..d8e227ddf255648e4e1d414286289bcb02ee76b1 100644 (file)
@@ -23,6 +23,7 @@
 #include "regs/xe_gt_regs.h"
 #include "regs/xe_regs.h"
 #include "xe_bo.h"
+#include "xe_bo_evict.h"
 #include "xe_debugfs.h"
 #include "xe_devcoredump.h"
 #include "xe_dma_buf.h"
@@ -467,10 +468,9 @@ struct xe_device *xe_device_create(struct pci_dev *pdev,
                        xa_erase(&xe->usm.asid_to_vm, asid);
        }
 
-       spin_lock_init(&xe->pinned.lock);
-       INIT_LIST_HEAD(&xe->pinned.kernel_bo_present);
-       INIT_LIST_HEAD(&xe->pinned.external_vram);
-       INIT_LIST_HEAD(&xe->pinned.evicted);
+       err = xe_bo_pinned_init(xe);
+       if (err)
+               goto err;
 
        xe->preempt_fence_wq = alloc_ordered_workqueue("xe-preempt-fence-wq",
                                                       WQ_MEM_RECLAIM);
@@ -939,6 +939,8 @@ void xe_device_remove(struct xe_device *xe)
        xe_display_unregister(xe);
 
        drm_dev_unplug(&xe->drm);
+
+       xe_bo_pci_dev_remove_all(xe);
 }
 
 void xe_device_shutdown(struct xe_device *xe)
index 4472bf849f2b55780e11078ffff23345a8e12bc2..0cc14daf75d3760e8533a38092666bf9ad6f9fb4 100644 (file)
@@ -426,8 +426,8 @@ struct xe_device {
                struct list_head kernel_bo_present;
                /** @pinned.evicted: pinned BO that have been evicted */
                struct list_head evicted;
-               /** @pinned.external_vram: pinned external BO in vram*/
-               struct list_head external_vram;
+               /** @pinned.external: pinned external and dma-buf. */
+               struct list_head external;
        } pinned;
 
        /** @ufence_wq: user fence wait queue */