drm/i915/ttm: add unmap_virtual callback
authorMatthew Auld <matthew.auld@intel.com>
Thu, 6 Jan 2022 17:49:09 +0000 (17:49 +0000)
committerMatthew Auld <matthew.auld@intel.com>
Mon, 10 Jan 2022 11:01:44 +0000 (11:01 +0000)
Ensure we call ttm_bo_unmap_virtual when releasing the pages.
Importantly this should now handle the ttm swapping case, and all other
places that already call into i915_ttm_move_notify().

v2: fix up the selftest

Fixes: cf3e3e86d779 ("drm/i915: Use ttm mmap handling for ttm bo's.")
Signed-off-by: Matthew Auld <matthew.auld@intel.com>
Cc: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Reviewed-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20220106174910.280616-3-matthew.auld@intel.com
drivers/gpu/drm/i915/gem/i915_gem_mman.c
drivers/gpu/drm/i915/gem/i915_gem_object_types.h
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c

index fae458019172d015cb84d69edcaf674dde3c6bef..55f256298d9018584a281abc16f2f8e466f640d9 100644 (file)
@@ -537,6 +537,9 @@ void i915_gem_object_release_mmap_offset(struct drm_i915_gem_object *obj)
 {
        struct i915_mmap_offset *mmo, *mn;
 
+       if (obj->ops->unmap_virtual)
+               obj->ops->unmap_virtual(obj);
+
        spin_lock(&obj->mmo.lock);
        rbtree_postorder_for_each_entry_safe(mmo, mn,
                                             &obj->mmo.offsets, offset) {
index 6f446cca4322a2270f14a935d4e748a7476d7c78..1c903d813097e5a911a92594e5bffaceb7e17161 100644 (file)
@@ -83,6 +83,7 @@ struct drm_i915_gem_object_ops {
        int (*pwrite)(struct drm_i915_gem_object *obj,
                      const struct drm_i915_gem_pwrite *arg);
        u64 (*mmap_offset)(struct drm_i915_gem_object *obj);
+       void (*unmap_virtual)(struct drm_i915_gem_object *obj);
 
        int (*dmabuf_export)(struct drm_i915_gem_object *obj);
 
index f2af76577e224099620af160be86938b6cedf232..d6d6878d2414231d45f2d46ae08f8937b315645b 100644 (file)
@@ -948,6 +948,11 @@ static u64 i915_ttm_mmap_offset(struct drm_i915_gem_object *obj)
        return drm_vma_node_offset_addr(&obj->base.vma_node);
 }
 
+static void i915_ttm_unmap_virtual(struct drm_i915_gem_object *obj)
+{
+       ttm_bo_unmap_virtual(i915_gem_to_ttm(obj));
+}
+
 static const struct drm_i915_gem_object_ops i915_gem_ttm_obj_ops = {
        .name = "i915_gem_object_ttm",
        .flags = I915_GEM_OBJECT_IS_SHRINKABLE |
@@ -963,6 +968,7 @@ static const struct drm_i915_gem_object_ops i915_gem_ttm_obj_ops = {
        .migrate = i915_ttm_migrate,
 
        .mmap_offset = i915_ttm_mmap_offset,
+       .unmap_virtual = i915_ttm_unmap_virtual,
        .mmap_ops = &vm_ops_ttm,
 };
 
index 743a098facf264b92806b95fb178807461654927..f61356b72b1c7fbe2adf0b16926f87de97ef1a90 100644 (file)
@@ -1369,20 +1369,10 @@ static int __igt_mmap_revoke(struct drm_i915_private *i915,
                }
        }
 
-       if (!obj->ops->mmap_ops) {
-               err = check_absent(addr, obj->base.size);
-               if (err) {
-                       pr_err("%s: was not absent\n", obj->mm.region->name);
-                       goto out_unmap;
-               }
-       } else {
-               /* ttm allows access to evicted regions by design */
-
-               err = check_present(addr, obj->base.size);
-               if (err) {
-                       pr_err("%s: was not present\n", obj->mm.region->name);
-                       goto out_unmap;
-               }
+       err = check_absent(addr, obj->base.size);
+       if (err) {
+               pr_err("%s: was not absent\n", obj->mm.region->name);
+               goto out_unmap;
        }
 
 out_unmap: