aboutsummaryrefslogtreecommitdiffstats
path: root/common/recipes-kernel/linux/linux-yocto-4.19.8/0319-drm-amdgpu-separate-per-VM-BOs-from-normal-in-the-mo.patch
diff options
context:
space:
mode:
Diffstat (limited to 'common/recipes-kernel/linux/linux-yocto-4.19.8/0319-drm-amdgpu-separate-per-VM-BOs-from-normal-in-the-mo.patch')
-rw-r--r--common/recipes-kernel/linux/linux-yocto-4.19.8/0319-drm-amdgpu-separate-per-VM-BOs-from-normal-in-the-mo.patch199
1 files changed, 199 insertions, 0 deletions
diff --git a/common/recipes-kernel/linux/linux-yocto-4.19.8/0319-drm-amdgpu-separate-per-VM-BOs-from-normal-in-the-mo.patch b/common/recipes-kernel/linux/linux-yocto-4.19.8/0319-drm-amdgpu-separate-per-VM-BOs-from-normal-in-the-mo.patch
new file mode 100644
index 00000000..a3e9db46
--- /dev/null
+++ b/common/recipes-kernel/linux/linux-yocto-4.19.8/0319-drm-amdgpu-separate-per-VM-BOs-from-normal-in-the-mo.patch
@@ -0,0 +1,199 @@
+From 78113fe6d925f0bbb2875d69813a6de48be56711 Mon Sep 17 00:00:00 2001
+From: =?UTF-8?q?Christian=20K=C3=B6nig?= <christian.koenig@amd.com>
+Date: Sat, 1 Sep 2018 13:25:31 +0200
+Subject: [PATCH 0319/2940] drm/amdgpu: separate per VM BOs from normal in the
+ moved state
+MIME-Version: 1.0
+Content-Type: text/plain; charset=UTF-8
+Content-Transfer-Encoding: 8bit
+
+Allows us to avoid taking the spinlock in more places.
+
+Change-Id: I54c3dcb21ce28b3c5dde39c8f77af8016fbbd8a4
+Signed-off-by: Christian König <christian.koenig@amd.com>
+Reviewed-by: Junwei Zhang <Jerry.Zhang@amd.com>
+Signed-off-by: Kalyan Alle <kalyan.alle@amd.com>
+---
+ drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 71 +++++++++++++-------------
+ drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h | 7 ++-
+ 2 files changed, 40 insertions(+), 38 deletions(-)
+
+diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
+index 0a334a04e8c9..cfe946f01b93 100644
+--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
+@@ -342,9 +342,7 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
+ break;
+
+ if (bo->tbo.type != ttm_bo_type_kernel) {
+- spin_lock(&vm->moved_lock);
+ list_move(&bo_base->vm_status, &vm->moved);
+- spin_unlock(&vm->moved_lock);
+ } else {
+ if (vm->use_cpu_for_update)
+ r = amdgpu_bo_kmap(bo, NULL);
+@@ -1758,10 +1756,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
+ amdgpu_asic_flush_hdp(adev, NULL);
+ }
+
+- spin_lock(&vm->moved_lock);
+- list_del_init(&bo_va->base.vm_status);
+- spin_unlock(&vm->moved_lock);
+-
+ /* If the BO is not in its preferred location add it back to
+ * the evicted list so that it gets validated again on the
+ * next command submission.
+@@ -1770,9 +1764,13 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
+ uint32_t mem_type = bo->tbo.mem.mem_type;
+
+ if (!(bo->preferred_domains & amdgpu_mem_type_to_domain(mem_type)))
+- list_add_tail(&bo_va->base.vm_status, &vm->evicted);
++ list_move_tail(&bo_va->base.vm_status, &vm->evicted);
+ else
+- list_add(&bo_va->base.vm_status, &vm->idle);
++ list_move(&bo_va->base.vm_status, &vm->idle);
++ } else {
++ spin_lock(&vm->invalidated_lock);
++ list_del_init(&bo_va->base.vm_status);
++ spin_unlock(&vm->invalidated_lock);
+ }
+
+ list_splice_init(&bo_va->invalids, &bo_va->valids);
+@@ -1998,41 +1996,41 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev,
+ struct amdgpu_vm *vm)
+ {
+ struct amdgpu_bo_va *bo_va, *tmp;
+- struct list_head moved;
++ struct reservation_object *resv;
+ bool clear;
+ int r;
+
+- INIT_LIST_HEAD(&moved);
+- spin_lock(&vm->moved_lock);
+- list_splice_init(&vm->moved, &moved);
+- spin_unlock(&vm->moved_lock);
+-
+- list_for_each_entry_safe(bo_va, tmp, &moved, base.vm_status) {
+- struct reservation_object *resv = bo_va->base.bo->tbo.resv;
+-
++ list_for_each_entry_safe(bo_va, tmp, &vm->moved, base.vm_status) {
+ /* Per VM BOs never need to bo cleared in the page tables */
+- if (resv == vm->root.base.bo->tbo.resv)
+- clear = false;
++ r = amdgpu_vm_bo_update(adev, bo_va, false);
++ if (r)
++ return r;
++ }
++
++ spin_lock(&vm->invalidated_lock);
++ while (!list_empty(&vm->invalidated)) {
++ bo_va = list_first_entry(&vm->invalidated, struct amdgpu_bo_va,
++ base.vm_status);
++ resv = bo_va->base.bo->tbo.resv;
++ spin_unlock(&vm->invalidated_lock);
++
+ /* Try to reserve the BO to avoid clearing its ptes */
+- else if (!amdgpu_vm_debug && reservation_object_trylock(resv))
++ if (!amdgpu_vm_debug && reservation_object_trylock(resv))
+ clear = false;
+ /* Somebody else is using the BO right now */
+ else
+ clear = true;
+
+ r = amdgpu_vm_bo_update(adev, bo_va, clear);
+- if (r) {
+- spin_lock(&vm->moved_lock);
+- list_splice(&moved, &vm->moved);
+- spin_unlock(&vm->moved_lock);
++ if (r)
+ return r;
+- }
+
+- if (!clear && resv != vm->root.base.bo->tbo.resv)
++ if (!clear)
+ reservation_object_unlock(resv);
+-
++ spin_lock(&vm->invalidated_lock);
+ }
+-
++ spin_unlock(&vm->invalidated_lock);
++
+ return 0;
+ }
+
+@@ -2096,9 +2094,7 @@ static void amdgpu_vm_bo_insert_map(struct amdgpu_device *adev,
+
+ if (bo && bo->tbo.resv == vm->root.base.bo->tbo.resv &&
+ !bo_va->base.moved) {
+- spin_lock(&vm->moved_lock);
+ list_move(&bo_va->base.vm_status, &vm->moved);
+- spin_unlock(&vm->moved_lock);
+ }
+ trace_amdgpu_vm_bo_map(bo_va, mapping);
+ }
+@@ -2458,9 +2454,9 @@ void amdgpu_vm_bo_rmv(struct amdgpu_device *adev,
+
+ list_del(&bo_va->base.bo_list);
+
+- spin_lock(&vm->moved_lock);
++ spin_lock(&vm->invalidated_lock);
+ list_del(&bo_va->base.vm_status);
+- spin_unlock(&vm->moved_lock);
++ spin_unlock(&vm->invalidated_lock);
+
+ list_for_each_entry_safe(mapping, next, &bo_va->valids, list) {
+ list_del(&mapping->list);
+@@ -2517,10 +2513,12 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev,
+
+ if (bo->tbo.type == ttm_bo_type_kernel) {
+ list_move(&bo_base->vm_status, &vm->relocated);
+- } else {
+- spin_lock(&bo_base->vm->moved_lock);
++ } else if (bo->tbo.resv == vm->root.base.bo->tbo.resv) {
+ list_move(&bo_base->vm_status, &vm->moved);
+- spin_unlock(&bo_base->vm->moved_lock);
++ } else {
++ spin_lock(&vm->invalidated_lock);
++ list_move(&bo_base->vm_status, &vm->invalidated);
++ spin_unlock(&vm->invalidated_lock);
+ }
+ }
+ }
+@@ -2665,9 +2663,10 @@ int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm,
+ vm->reserved_vmid[i] = NULL;
+ INIT_LIST_HEAD(&vm->evicted);
+ INIT_LIST_HEAD(&vm->relocated);
+- spin_lock_init(&vm->moved_lock);
+ INIT_LIST_HEAD(&vm->moved);
+ INIT_LIST_HEAD(&vm->idle);
++ INIT_LIST_HEAD(&vm->invalidated);
++ spin_lock_init(&vm->invalidated_lock);
+ INIT_LIST_HEAD(&vm->freed);
+
+ /* create scheduler entity for page table updates */
+diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
+index 94fe47890adf..3787c1973f91 100644
+--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
+@@ -201,13 +201,16 @@ struct amdgpu_vm {
+ /* PT BOs which relocated and their parent need an update */
+ struct list_head relocated;
+
+- /* BOs moved, but not yet updated in the PT */
++ /* per VM BOs moved, but not yet updated in the PT */
+ struct list_head moved;
+- spinlock_t moved_lock;
+
+ /* All BOs of this VM not currently in the state machine */
+ struct list_head idle;
+
++ /* regular invalidated BOs, but not yet updated in the PT */
++ struct list_head invalidated;
++ spinlock_t invalidated_lock;
++
+ /* BO mappings freed, but not yet updated in the PT */
+ struct list_head freed;
+
+--
+2.17.1
+