diff options
Diffstat (limited to 'common/recipes-kernel/linux/files/0467-drm-amdgpu-use-a-sync-object-for-VMID-fences-v2.patch')
-rw-r--r-- | common/recipes-kernel/linux/files/0467-drm-amdgpu-use-a-sync-object-for-VMID-fences-v2.patch | 286 |
1 files changed, 286 insertions, 0 deletions
diff --git a/common/recipes-kernel/linux/files/0467-drm-amdgpu-use-a-sync-object-for-VMID-fences-v2.patch b/common/recipes-kernel/linux/files/0467-drm-amdgpu-use-a-sync-object-for-VMID-fences-v2.patch new file mode 100644 index 00000000..7edb7656 --- /dev/null +++ b/common/recipes-kernel/linux/files/0467-drm-amdgpu-use-a-sync-object-for-VMID-fences-v2.patch @@ -0,0 +1,286 @@ +From ce9a1734fbccf5fa7fb28a314f25e1f6f244f988 Mon Sep 17 00:00:00 2001 +From: =?UTF-8?q?Christian=20K=C3=B6nig?= <christian.koenig@amd.com> +Date: Mon, 15 Feb 2016 12:33:02 +0100 +Subject: [PATCH 0467/1110] drm/amdgpu: use a sync object for VMID fences v2 +MIME-Version: 1.0 +Content-Type: text/plain; charset=UTF-8 +Content-Transfer-Encoding: 8bit + +v2: rebase & cleanup + +This way we can store more than one fence as user for each VMID. + +Signed-off-by: Christian König <christian.koenig@amd.com> +Acked-by: Alex Deucher <alexander.deucher@amd.com> (v1) +Reviewed-by: Chunming Zhou <david1.zhou@amd.com> (v1) +Signed-off-by: Alex Deucher <alexander.deucher@amd.com> +Signed-off-by: Kalyan Alle <kalyan.alle@amd.com> +--- + drivers/gpu/drm/amd/amdgpu/amdgpu.h | 7 ++- + drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c | 105 +++++++++++++++++++++++++++++-- + drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 54 ++++++++-------- + 3 files changed, 133 insertions(+), 33 deletions(-) + +diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu.h b/drivers/gpu/drm/amd/amdgpu/amdgpu.h +index 4b9fe35..915c2e3 100644 +--- a/drivers/gpu/drm/amd/amdgpu/amdgpu.h ++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu.h +@@ -589,6 +589,9 @@ int amdgpu_sync_resv(struct amdgpu_device *adev, + struct amdgpu_sync *sync, + struct reservation_object *resv, + void *owner); ++bool amdgpu_sync_is_idle(struct amdgpu_sync *sync); ++int amdgpu_sync_cycle_fences(struct amdgpu_sync *dst, struct amdgpu_sync *src, ++ struct fence *fence); + struct fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync); + int amdgpu_sync_wait(struct amdgpu_sync *sync); + void amdgpu_sync_free(struct amdgpu_sync *sync); +@@ -874,8 +877,10 @@ struct amdgpu_vm { + + struct amdgpu_vm_id { + struct list_head list; +- struct fence *active; ++ struct fence *first; ++ struct amdgpu_sync active; + atomic_long_t owner; ++ + uint64_t pd_gpu_addr; + /* last flushed PD/PT update */ + struct fence *flushed_updates; +diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c +index c48b4fc..34a9280 100644 +--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c ++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c +@@ -109,6 +109,29 @@ static void amdgpu_sync_keep_later(struct fence **keep, struct fence *fence) + } + + /** ++ * amdgpu_sync_add_later - add the fence to the hash ++ * ++ * @sync: sync object to add the fence to ++ * @f: fence to add ++ * ++ * Tries to add the fence to an existing hash entry. Returns true when an entry ++ * was found, false otherwise. ++ */ ++static bool amdgpu_sync_add_later(struct amdgpu_sync *sync, struct fence *f) ++{ ++ struct amdgpu_sync_entry *e; ++ ++ hash_for_each_possible(sync->fences, e, node, f->context) { ++ if (unlikely(e->fence->context != f->context)) ++ continue; ++ ++ amdgpu_sync_keep_later(&e->fence, f); ++ return true; ++ } ++ return false; ++} ++ ++/** + * amdgpu_sync_fence - remember to sync to this fence + * + * @sync: sync object to add fence to +@@ -127,13 +150,8 @@ int amdgpu_sync_fence(struct amdgpu_device *adev, struct amdgpu_sync *sync, + amdgpu_sync_get_owner(f) == AMDGPU_FENCE_OWNER_VM) + amdgpu_sync_keep_later(&sync->last_vm_update, f); + +- hash_for_each_possible(sync->fences, e, node, f->context) { +- if (unlikely(e->fence->context != f->context)) +- continue; +- +- amdgpu_sync_keep_later(&e->fence, f); ++ if (amdgpu_sync_add_later(sync, f)) + return 0; +- } + + e = kmem_cache_alloc(amdgpu_sync_slab, GFP_KERNEL); + if (!e) +@@ -204,6 +222,81 @@ int amdgpu_sync_resv(struct amdgpu_device *adev, + return r; + } + ++/** ++ * amdgpu_sync_is_idle - test if all fences are signaled ++ * ++ * @sync: the sync object ++ * ++ * Returns true if all fences in the sync object are signaled. ++ */ ++bool amdgpu_sync_is_idle(struct amdgpu_sync *sync) ++{ ++ struct amdgpu_sync_entry *e; ++ struct hlist_node *tmp; ++ int i; ++ ++ hash_for_each_safe(sync->fences, i, tmp, e, node) { ++ struct fence *f = e->fence; ++ ++ if (fence_is_signaled(f)) { ++ hash_del(&e->node); ++ fence_put(f); ++ kmem_cache_free(amdgpu_sync_slab, e); ++ continue; ++ } ++ ++ return false; ++ } ++ ++ return true; ++} ++ ++/** ++ * amdgpu_sync_cycle_fences - move fences from one sync object into another ++ * ++ * @dst: the destination sync object ++ * @src: the source sync object ++ * @fence: fence to add to source ++ * ++ * Remove all fences from source and put them into destination and add ++ * fence as new one into source. ++ */ ++int amdgpu_sync_cycle_fences(struct amdgpu_sync *dst, struct amdgpu_sync *src, ++ struct fence *fence) ++{ ++ struct amdgpu_sync_entry *e, *newone; ++ struct hlist_node *tmp; ++ int i; ++ ++ /* Allocate the new entry before moving the old ones */ ++ newone = kmem_cache_alloc(amdgpu_sync_slab, GFP_KERNEL); ++ if (!newone) ++ return -ENOMEM; ++ ++ hash_for_each_safe(src->fences, i, tmp, e, node) { ++ struct fence *f = e->fence; ++ ++ hash_del(&e->node); ++ if (fence_is_signaled(f)) { ++ fence_put(f); ++ kmem_cache_free(amdgpu_sync_slab, e); ++ continue; ++ } ++ ++ if (amdgpu_sync_add_later(dst, f)) { ++ kmem_cache_free(amdgpu_sync_slab, e); ++ continue; ++ } ++ ++ hash_add(dst->fences, &e->node, f->context); ++ } ++ ++ hash_add(src->fences, &newone->node, fence->context); ++ newone->fence = fence_get(fence); ++ ++ return 0; ++} ++ + struct fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync) + { + struct amdgpu_sync_entry *e; +diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c +index 1c87193..8382b7f 100644 +--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c ++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c +@@ -196,14 +196,13 @@ int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring, + + if (usable) { + +- r = amdgpu_sync_fence(ring->adev, sync, id->active); +- if (r) { +- mutex_unlock(&adev->vm_manager.lock); +- return r; +- } ++ r = amdgpu_sync_fence(ring->adev, sync, id->first); ++ if (r) ++ goto error; + +- fence_put(id->active); +- id->active = fence_get(fence); ++ r = amdgpu_sync_fence(ring->adev, &id->active, fence); ++ if (r) ++ goto error; + + list_move_tail(&id->list, &adev->vm_manager.ids_lru); + +@@ -221,13 +220,13 @@ int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring, + struct amdgpu_vm_id, + list); + +- if (id->active && !fence_is_signaled(id->active)) { +- struct amdgpu_vm_id *tmp; ++ if (!amdgpu_sync_is_idle(&id->active)) { + struct list_head *head = &adev->vm_manager.ids_lru; ++ struct amdgpu_vm_id *tmp; + + list_for_each_entry_safe(id, tmp, &adev->vm_manager.ids_lru, + list) { +- if (id->active && fence_is_signaled(id->active)) { ++ if (amdgpu_sync_is_idle(&id->active)) { + list_move(&id->list, head); + head = &id->list; + } +@@ -237,25 +236,27 @@ int amdgpu_vm_grab_id(struct amdgpu_vm *vm, struct amdgpu_ring *ring, + list); + } + +- r = amdgpu_sync_fence(ring->adev, sync, id->active); +- if (!r) { +- fence_put(id->active); +- id->active = fence_get(fence); ++ r = amdgpu_sync_cycle_fences(sync, &id->active, fence); ++ if (r) ++ goto error; + +- fence_put(id->flushed_updates); +- id->flushed_updates = fence_get(updates); ++ fence_put(id->first); ++ id->first = fence_get(fence); + +- id->pd_gpu_addr = pd_addr; ++ fence_put(id->flushed_updates); ++ id->flushed_updates = fence_get(updates); + +- list_move_tail(&id->list, &adev->vm_manager.ids_lru); +- atomic_long_set(&id->owner, (long)&vm->ids[ring->idx]); +- vm->ids[ring->idx] = id; ++ id->pd_gpu_addr = pd_addr; + +- *vm_id = id - adev->vm_manager.ids; +- *vm_pd_addr = pd_addr; +- trace_amdgpu_vm_grab_id(vm, ring->idx, *vm_id, *vm_pd_addr); +- } ++ list_move_tail(&id->list, &adev->vm_manager.ids_lru); ++ atomic_long_set(&id->owner, (long)id); ++ vm->ids[ring->idx] = id; + ++ *vm_id = id - adev->vm_manager.ids; ++ *vm_pd_addr = pd_addr; ++ trace_amdgpu_vm_grab_id(vm, ring->idx, *vm_id, *vm_pd_addr); ++ ++error: + mutex_unlock(&adev->vm_manager.lock); + return r; + } +@@ -1470,7 +1471,7 @@ void amdgpu_vm_manager_init(struct amdgpu_device *adev) + + /* skip over VMID 0, since it is the system VM */ + for (i = 1; i < adev->vm_manager.num_ids; ++i) { +- amdgpu_vm_reset_id(adev, i); ++ amdgpu_sync_create(&adev->vm_manager.ids[i].active); + list_add_tail(&adev->vm_manager.ids[i].list, + &adev->vm_manager.ids_lru); + } +@@ -1493,7 +1494,8 @@ void amdgpu_vm_manager_fini(struct amdgpu_device *adev) + for (i = 0; i < AMDGPU_NUM_VM; ++i) { + struct amdgpu_vm_id *id = &adev->vm_manager.ids[i]; + +- fence_put(id->active); ++ fence_put(adev->vm_manager.ids[i].first); ++ amdgpu_sync_free(&adev->vm_manager.ids[i].active); + fence_put(id->flushed_updates); + } + } +-- +2.7.4 + |