OSDN Git Service

drm/amdgpu: remove last_entry_used from the VM code
authorChristian König <christian.koenig@amd.com>
Thu, 30 Nov 2017 14:28:03 +0000 (15:28 +0100)
committerAlex Deucher <alexander.deucher@amd.com>
Tue, 12 Dec 2017 19:46:00 +0000 (14:46 -0500)
Not needed any more.

Signed-off-by: Christian König <christian.koenig@amd.com>
Reviewed-by: Chunming Zhou <david1.zhou@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h

index bd6296a..25fdcba 100644 (file)
@@ -329,9 +329,6 @@ static int amdgpu_vm_alloc_levels(struct amdgpu_device *adev,
            to >= amdgpu_vm_num_entries(adev, level))
                return -EINVAL;
 
-       if (to > parent->last_entry_used)
-               parent->last_entry_used = to;
-
        ++level;
        saddr = saddr & ((1 << shift) - 1);
        eaddr = eaddr & ((1 << shift) - 1);
@@ -1184,16 +1181,19 @@ error_free:
  *
  * Mark all PD level as invalid after an error.
  */
-static void amdgpu_vm_invalidate_level(struct amdgpu_vm *vm,
-                                      struct amdgpu_vm_pt *parent)
+static void amdgpu_vm_invalidate_level(struct amdgpu_device *adev,
+                                      struct amdgpu_vm *vm,
+                                      struct amdgpu_vm_pt *parent,
+                                      unsigned level)
 {
-       unsigned pt_idx;
+       unsigned pt_idx, num_entries;
 
        /*
         * Recurse into the subdirectories. This recursion is harmless because
         * we only have a maximum of 5 layers.
         */
-       for (pt_idx = 0; pt_idx <= parent->last_entry_used; ++pt_idx) {
+       num_entries = amdgpu_vm_num_entries(adev, level);
+       for (pt_idx = 0; pt_idx < num_entries; ++pt_idx) {
                struct amdgpu_vm_pt *entry = &parent->entries[pt_idx];
 
                if (!entry->base.bo)
@@ -1204,7 +1204,7 @@ static void amdgpu_vm_invalidate_level(struct amdgpu_vm *vm,
                if (list_empty(&entry->base.vm_status))
                        list_add(&entry->base.vm_status, &vm->relocated);
                spin_unlock(&vm->status_lock);
-               amdgpu_vm_invalidate_level(vm, entry);
+               amdgpu_vm_invalidate_level(adev, vm, entry, level + 1);
        }
 }
 
@@ -1246,7 +1246,8 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
 
                        r = amdgpu_vm_update_pde(adev, vm, pt, entry);
                        if (r) {
-                               amdgpu_vm_invalidate_level(vm, &vm->root);
+                               amdgpu_vm_invalidate_level(adev, vm,
+                                                          &vm->root, 0);
                                return r;
                        }
                        spin_lock(&vm->status_lock);
@@ -1649,7 +1650,7 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
 
 error_free:
        amdgpu_job_free(job);
-       amdgpu_vm_invalidate_level(vm, &vm->root);
+       amdgpu_vm_invalidate_level(adev, vm, &vm->root, 0);
        return r;
 }
 
@@ -2713,26 +2714,31 @@ error_free_sched_entity:
 /**
  * amdgpu_vm_free_levels - free PD/PT levels
  *
- * @level: PD/PT starting level to free
+ * @adev: amdgpu device structure
+ * @parent: PD/PT starting level to free
+ * @level: level of parent structure
  *
  * Free the page directory or page table level and all sub levels.
  */
-static void amdgpu_vm_free_levels(struct amdgpu_vm_pt *level)
+static void amdgpu_vm_free_levels(struct amdgpu_device *adev,
+                                 struct amdgpu_vm_pt *parent,
+                                 unsigned level)
 {
-       unsigned i;
+       unsigned i, num_entries = amdgpu_vm_num_entries(adev, level);
 
-       if (level->base.bo) {
-               list_del(&level->base.bo_list);
-               list_del(&level->base.vm_status);
-               amdgpu_bo_unref(&level->base.bo->shadow);
-               amdgpu_bo_unref(&level->base.bo);
+       if (parent->base.bo) {
+               list_del(&parent->base.bo_list);
+               list_del(&parent->base.vm_status);
+               amdgpu_bo_unref(&parent->base.bo->shadow);
+               amdgpu_bo_unref(&parent->base.bo);
        }
 
-       if (level->entries)
-               for (i = 0; i <= level->last_entry_used; i++)
-                       amdgpu_vm_free_levels(&level->entries[i]);
+       if (parent->entries)
+               for (i = 0; i < num_entries; i++)
+                       amdgpu_vm_free_levels(adev, &parent->entries[i],
+                                             level + 1);
 
-       kvfree(level->entries);
+       kvfree(parent->entries);
 }
 
 /**
@@ -2790,7 +2796,7 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
        if (r) {
                dev_err(adev->dev, "Leaking page tables because BO reservation failed\n");
        } else {
-               amdgpu_vm_free_levels(&vm->root);
+               amdgpu_vm_free_levels(adev, &vm->root, 0);
                amdgpu_bo_unreserve(root);
        }
        amdgpu_bo_unref(&root);
index 1599804..2f1d0c8 100644 (file)
@@ -142,7 +142,6 @@ struct amdgpu_vm_pt {
 
        /* array of page tables, one for each directory entry */
        struct amdgpu_vm_pt             *entries;
-       unsigned                        last_entry_used;
 };
 
 #define AMDGPU_VM_FAULT(pasid, addr) (((u64)(pasid) << 48) | (addr))