drm/amdgpu: make set_prt callback optional and fix error handling
authorChristian König <christian.koenig@amd.com>
Mon, 13 Mar 2017 09:13:36 +0000 (10:13 +0100)
committerAlex Deucher <alexander.deucher@amd.com>
Thu, 30 Mar 2017 03:53:59 +0000 (23:53 -0400)
PRT support is completely implemented now and we left it
turned on accidentially in the error path.

Signed-off-by: Christian König <christian.koenig@amd.com>
Reviewed-by: Junwei Zhang <Jerry.Zhang@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c

index a45de6e6a0f73af16700687e9a42833a0dfb628c..f9f4e20550840381e83d20214270d35fce734850 100644 (file)
@@ -1194,10 +1194,13 @@ static void amdgpu_vm_update_prt_state(struct amdgpu_device *adev)
 }
 
 /**
- * amdgpu_vm_prt_put - add a PRT user
+ * amdgpu_vm_prt_get - add a PRT user
  */
 static void amdgpu_vm_prt_get(struct amdgpu_device *adev)
 {
+       if (!adev->gart.gart_funcs->set_prt)
+               return;
+
        if (atomic_inc_return(&adev->vm_manager.num_prt_users) == 1)
                amdgpu_vm_update_prt_state(adev);
 }
@@ -1228,9 +1231,12 @@ static void amdgpu_vm_prt_cb(struct dma_fence *fence, struct dma_fence_cb *_cb)
 static void amdgpu_vm_add_prt_cb(struct amdgpu_device *adev,
                                 struct dma_fence *fence)
 {
-       struct amdgpu_prt_cb *cb = kmalloc(sizeof(struct amdgpu_prt_cb),
-                                          GFP_KERNEL);
+       struct amdgpu_prt_cb *cb;
 
+       if (!adev->gart.gart_funcs->set_prt)
+               return;
+
+       cb = kmalloc(sizeof(struct amdgpu_prt_cb), GFP_KERNEL);
        if (!cb) {
                /* Last resort when we are OOM */
                if (fence)
@@ -1445,14 +1451,6 @@ int amdgpu_vm_bo_map(struct amdgpu_device *adev,
            size == 0 || size & AMDGPU_GPU_PAGE_MASK)
                return -EINVAL;
 
-       if (flags & AMDGPU_PTE_PRT) {
-               /* Check if we have PRT hardware support */
-               if (!adev->gart.gart_funcs->set_prt)
-                       return -EINVAL;
-
-               amdgpu_vm_prt_get(adev);
-       }
-
        /* make sure object fit at this offset */
        eaddr = saddr + size - 1;
        if (saddr >= eaddr ||
@@ -1533,6 +1531,9 @@ int amdgpu_vm_bo_map(struct amdgpu_device *adev,
                vm->page_tables[pt_idx].addr = 0;
        }
 
+       if (flags & AMDGPU_PTE_PRT)
+               amdgpu_vm_prt_get(adev);
+
        return 0;
 
 error_free:
@@ -1753,7 +1754,7 @@ err:
 void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
 {
        struct amdgpu_bo_va_mapping *mapping, *tmp;
-       bool prt_fini_called = false;
+       bool prt_fini_needed = !!adev->gart.gart_funcs->set_prt;
        int i;
 
        amd_sched_entity_fini(vm->entity.sched, &vm->entity);
@@ -1767,9 +1768,9 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
                kfree(mapping);
        }
        list_for_each_entry_safe(mapping, tmp, &vm->freed, list) {
-               if (mapping->flags & AMDGPU_PTE_PRT && !prt_fini_called) {
+               if (mapping->flags & AMDGPU_PTE_PRT && prt_fini_needed) {
                        amdgpu_vm_prt_fini(adev, vm);
-                       prt_fini_called = true;
+                       prt_fini_needed = false;
                }
 
                list_del(&mapping->list);