diff options
Diffstat (limited to 'common/recipes-kernel/linux/files/0392-drm-amdgpu-dispatch-job-for-vm.patch')
-rw-r--r-- | common/recipes-kernel/linux/files/0392-drm-amdgpu-dispatch-job-for-vm.patch | 419 |
1 files changed, 0 insertions, 419 deletions
diff --git a/common/recipes-kernel/linux/files/0392-drm-amdgpu-dispatch-job-for-vm.patch b/common/recipes-kernel/linux/files/0392-drm-amdgpu-dispatch-job-for-vm.patch deleted file mode 100644 index 09953536..00000000 --- a/common/recipes-kernel/linux/files/0392-drm-amdgpu-dispatch-job-for-vm.patch +++ /dev/null @@ -1,419 +0,0 @@ -From d5fc5e82a3127b8839c4a4457f3b950a009737a7 Mon Sep 17 00:00:00 2001 -From: Chunming Zhou <david1.zhou@amd.com> -Date: Tue, 21 Jul 2015 16:52:10 +0800 -Subject: [PATCH 0392/1050] drm/amdgpu: dispatch job for vm - -use kernel context to submit command for vm - -Signed-off-by: Chunming Zhou <david1.zhou@amd.com> -Acked-by: Christian K?nig <christian.koenig@amd.com> -Reviewed-by: Jammy Zhou <Jammy.Zhou@amd.com> ---- - drivers/gpu/drm/amd/amdgpu/amdgpu.h | 20 +++ - drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 245 ++++++++++++++++++++++++++------- - 2 files changed, 217 insertions(+), 48 deletions(-) - -diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu.h b/drivers/gpu/drm/amd/amdgpu/amdgpu.h -index 47e4809..1b8d05f 100644 ---- a/drivers/gpu/drm/amd/amdgpu/amdgpu.h -+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu.h -@@ -1221,6 +1221,19 @@ struct amdgpu_cs_chunk { - void __user *user_ptr; - }; - -+union amdgpu_sched_job_param { -+ struct { -+ struct amdgpu_vm *vm; -+ uint64_t start; -+ uint64_t last; -+ struct amdgpu_fence **fence; -+ -+ } vm_mapping; -+ struct { -+ struct amdgpu_bo *bo; -+ } vm; -+}; -+ - struct amdgpu_cs_parser { - struct amdgpu_device *adev; - struct drm_file *filp; -@@ -1245,6 +1258,7 @@ struct amdgpu_cs_parser { - struct mutex job_lock; - struct work_struct job_work; - int (*prepare_job)(struct amdgpu_cs_parser *sched_job); -+ union amdgpu_sched_job_param job_param; - int (*run_job)(struct amdgpu_cs_parser *sched_job); - int (*free_job)(struct amdgpu_cs_parser *sched_job); - }; -@@ -2255,6 +2269,12 @@ void amdgpu_pci_config_reset(struct amdgpu_device *adev); - bool amdgpu_card_posted(struct amdgpu_device *adev); - void amdgpu_update_display_priority(struct amdgpu_device *adev); - bool amdgpu_boot_test_post_card(struct amdgpu_device *adev); -+struct amdgpu_cs_parser *amdgpu_cs_parser_create(struct amdgpu_device *adev, -+ struct drm_file *filp, -+ struct amdgpu_ctx *ctx, -+ struct amdgpu_ib *ibs, -+ uint32_t num_ibs); -+ - int amdgpu_cs_parser_init(struct amdgpu_cs_parser *p, void *data); - int amdgpu_cs_get_ring(struct amdgpu_device *adev, u32 ip_type, - u32 ip_instance, u32 ring, -diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c -index fd8395f..34938d2 100644 ---- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c -+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c -@@ -306,6 +306,24 @@ static void amdgpu_vm_update_pages(struct amdgpu_device *adev, - } - } - -+static int amdgpu_vm_free_job( -+ struct amdgpu_cs_parser *sched_job) -+{ -+ int i; -+ for (i = 0; i < sched_job->num_ibs; i++) -+ amdgpu_ib_free(sched_job->adev, &sched_job->ibs[i]); -+ kfree(sched_job->ibs); -+ return 0; -+} -+ -+static int amdgpu_vm_run_job( -+ struct amdgpu_cs_parser *sched_job) -+{ -+ amdgpu_bo_fence(sched_job->job_param.vm.bo, -+ sched_job->ibs[sched_job->num_ibs -1].fence, true); -+ return 0; -+} -+ - /** - * amdgpu_vm_clear_bo - initially clear the page dir/table - * -@@ -316,7 +334,8 @@ static int amdgpu_vm_clear_bo(struct amdgpu_device *adev, - struct amdgpu_bo *bo) - { - struct amdgpu_ring *ring = adev->vm_manager.vm_pte_funcs_ring; -- struct amdgpu_ib ib; -+ struct amdgpu_cs_parser *sched_job = NULL; -+ struct amdgpu_ib *ib; - unsigned entries; - uint64_t addr; - int r; -@@ -336,24 +355,54 @@ static int amdgpu_vm_clear_bo(struct amdgpu_device *adev, - addr = amdgpu_bo_gpu_offset(bo); - entries = amdgpu_bo_size(bo) / 8; - -- r = amdgpu_ib_get(ring, NULL, entries * 2 + 64, &ib); -- if (r) -+ ib = kzalloc(sizeof(struct amdgpu_ib), GFP_KERNEL); -+ if (!ib) - goto error_unreserve; - -- ib.length_dw = 0; -- -- amdgpu_vm_update_pages(adev, &ib, addr, 0, entries, 0, 0, 0); -- amdgpu_vm_pad_ib(adev, &ib); -- WARN_ON(ib.length_dw > 64); -- -- r = amdgpu_ib_schedule(adev, 1, &ib, AMDGPU_FENCE_OWNER_VM); -+ r = amdgpu_ib_get(ring, NULL, entries * 2 + 64, ib); - if (r) - goto error_free; - -- amdgpu_bo_fence(bo, ib.fence, true); -+ ib->length_dw = 0; -+ -+ amdgpu_vm_update_pages(adev, ib, addr, 0, entries, 0, 0, 0); -+ amdgpu_vm_pad_ib(adev, ib); -+ WARN_ON(ib->length_dw > 64); -+ -+ if (amdgpu_enable_scheduler) { -+ int r; -+ uint64_t v_seq; -+ sched_job = amdgpu_cs_parser_create(adev, AMDGPU_FENCE_OWNER_VM, -+ adev->kernel_ctx, ib, 1); -+ if(!sched_job) -+ goto error_free; -+ sched_job->job_param.vm.bo = bo; -+ sched_job->run_job = amdgpu_vm_run_job; -+ sched_job->free_job = amdgpu_vm_free_job; -+ v_seq = atomic64_inc_return(&adev->kernel_ctx->rings[ring->idx].c_entity.last_queued_v_seq); -+ sched_job->uf.sequence = v_seq; -+ amd_sched_push_job(ring->scheduler, -+ &adev->kernel_ctx->rings[ring->idx].c_entity, -+ sched_job); -+ r = amd_sched_wait_emit(&adev->kernel_ctx->rings[ring->idx].c_entity, -+ v_seq, -+ true, -+ -1); -+ if (r) -+ DRM_ERROR("emit timeout\n"); -+ -+ amdgpu_bo_unreserve(bo); -+ return 0; -+ } else { -+ r = amdgpu_ib_schedule(adev, 1, ib, AMDGPU_FENCE_OWNER_VM); -+ if (r) -+ goto error_free; -+ amdgpu_bo_fence(bo, ib->fence, true); -+ } - - error_free: -- amdgpu_ib_free(adev, &ib); -+ amdgpu_ib_free(adev, ib); -+ kfree(ib); - - error_unreserve: - amdgpu_bo_unreserve(bo); -@@ -406,7 +455,9 @@ int amdgpu_vm_update_page_directory(struct amdgpu_device *adev, - uint32_t incr = AMDGPU_VM_PTE_COUNT * 8; - uint64_t last_pde = ~0, last_pt = ~0; - unsigned count = 0, pt_idx, ndw; -- struct amdgpu_ib ib; -+ struct amdgpu_ib *ib; -+ struct amdgpu_cs_parser *sched_job = NULL; -+ - int r; - - /* padding, etc. */ -@@ -419,10 +470,14 @@ int amdgpu_vm_update_page_directory(struct amdgpu_device *adev, - if (ndw > 0xfffff) - return -ENOMEM; - -- r = amdgpu_ib_get(ring, NULL, ndw * 4, &ib); -+ ib = kzalloc(sizeof(struct amdgpu_ib), GFP_KERNEL); -+ if (!ib) -+ return -ENOMEM; -+ -+ r = amdgpu_ib_get(ring, NULL, ndw * 4, ib); - if (r) - return r; -- ib.length_dw = 0; -+ ib->length_dw = 0; - - /* walk over the address space and update the page directory */ - for (pt_idx = 0; pt_idx <= vm->max_pde_used; ++pt_idx) { -@@ -442,7 +497,7 @@ int amdgpu_vm_update_page_directory(struct amdgpu_device *adev, - ((last_pt + incr * count) != pt)) { - - if (count) { -- amdgpu_vm_update_pages(adev, &ib, last_pde, -+ amdgpu_vm_update_pages(adev, ib, last_pde, - last_pt, count, incr, - AMDGPU_PTE_VALID, 0); - } -@@ -456,23 +511,59 @@ int amdgpu_vm_update_page_directory(struct amdgpu_device *adev, - } - - if (count) -- amdgpu_vm_update_pages(adev, &ib, last_pde, last_pt, count, -+ amdgpu_vm_update_pages(adev, ib, last_pde, last_pt, count, - incr, AMDGPU_PTE_VALID, 0); - -- if (ib.length_dw != 0) { -- amdgpu_vm_pad_ib(adev, &ib); -- amdgpu_sync_resv(adev, &ib.sync, pd->tbo.resv, AMDGPU_FENCE_OWNER_VM); -- WARN_ON(ib.length_dw > ndw); -- r = amdgpu_ib_schedule(adev, 1, &ib, AMDGPU_FENCE_OWNER_VM); -- if (r) { -- amdgpu_ib_free(adev, &ib); -- return r; -+ if (ib->length_dw != 0) { -+ amdgpu_vm_pad_ib(adev, ib); -+ amdgpu_sync_resv(adev, &ib->sync, pd->tbo.resv, AMDGPU_FENCE_OWNER_VM); -+ WARN_ON(ib->length_dw > ndw); -+ -+ if (amdgpu_enable_scheduler) { -+ int r; -+ uint64_t v_seq; -+ sched_job = amdgpu_cs_parser_create(adev, AMDGPU_FENCE_OWNER_VM, -+ adev->kernel_ctx, -+ ib, 1); -+ if(!sched_job) -+ goto error_free; -+ sched_job->job_param.vm.bo = pd; -+ sched_job->run_job = amdgpu_vm_run_job; -+ sched_job->free_job = amdgpu_vm_free_job; -+ v_seq = atomic64_inc_return(&adev->kernel_ctx->rings[ring->idx].c_entity.last_queued_v_seq); -+ sched_job->uf.sequence = v_seq; -+ amd_sched_push_job(ring->scheduler, -+ &adev->kernel_ctx->rings[ring->idx].c_entity, -+ sched_job); -+ r = amd_sched_wait_emit(&adev->kernel_ctx->rings[ring->idx].c_entity, -+ v_seq, -+ true, -+ -1); -+ if (r) -+ DRM_ERROR("emit timeout\n"); -+ } else { -+ r = amdgpu_ib_schedule(adev, 1, ib, AMDGPU_FENCE_OWNER_VM); -+ if (r) { -+ amdgpu_ib_free(adev, ib); -+ return r; -+ } -+ amdgpu_bo_fence(pd, ib->fence, true); - } -- amdgpu_bo_fence(pd, ib.fence, true); - } -- amdgpu_ib_free(adev, &ib); -+ -+ if (!amdgpu_enable_scheduler || ib->length_dw == 0) { -+ amdgpu_ib_free(adev, ib); -+ kfree(ib); -+ } - - return 0; -+ -+error_free: -+ if (sched_job) -+ kfree(sched_job); -+ amdgpu_ib_free(adev, ib); -+ kfree(ib); -+ return -ENOMEM; - } - - /** -@@ -657,6 +748,20 @@ static void amdgpu_vm_fence_pts(struct amdgpu_vm *vm, - amdgpu_bo_fence(vm->page_tables[i].bo, fence, true); - } - -+static int amdgpu_vm_bo_update_mapping_run_job( -+ struct amdgpu_cs_parser *sched_job) -+{ -+ struct amdgpu_fence **fence = sched_job->job_param.vm_mapping.fence; -+ amdgpu_vm_fence_pts(sched_job->job_param.vm_mapping.vm, -+ sched_job->job_param.vm_mapping.start, -+ sched_job->job_param.vm_mapping.last + 1, -+ sched_job->ibs[sched_job->num_ibs -1].fence); -+ if (fence) { -+ amdgpu_fence_unref(fence); -+ *fence = amdgpu_fence_ref(sched_job->ibs[sched_job->num_ibs -1].fence); -+ } -+ return 0; -+} - /** - * amdgpu_vm_bo_update_mapping - update a mapping in the vm page table - * -@@ -681,7 +786,8 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev, - struct amdgpu_ring *ring = adev->vm_manager.vm_pte_funcs_ring; - unsigned nptes, ncmds, ndw; - uint32_t flags = gtt_flags; -- struct amdgpu_ib ib; -+ struct amdgpu_ib *ib; -+ struct amdgpu_cs_parser *sched_job = NULL; - int r; - - /* normally,bo_va->flags only contians READABLE and WIRTEABLE bit go here -@@ -728,48 +834,91 @@ static int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev, - if (ndw > 0xfffff) - return -ENOMEM; - -- r = amdgpu_ib_get(ring, NULL, ndw * 4, &ib); -- if (r) -+ ib = kzalloc(sizeof(struct amdgpu_ib), GFP_KERNEL); -+ if (!ib) -+ return -ENOMEM; -+ -+ r = amdgpu_ib_get(ring, NULL, ndw * 4, ib); -+ if (r) { -+ kfree(ib); - return r; -- ib.length_dw = 0; -+ } -+ -+ ib->length_dw = 0; - - if (!(flags & AMDGPU_PTE_VALID)) { - unsigned i; - - for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { - struct amdgpu_fence *f = vm->ids[i].last_id_use; -- r = amdgpu_sync_fence(adev, &ib.sync, &f->base); -+ r = amdgpu_sync_fence(adev, &ib->sync, &f->base); - if (r) - return r; - } - } - -- r = amdgpu_vm_update_ptes(adev, vm, &ib, mapping->it.start, -+ r = amdgpu_vm_update_ptes(adev, vm, ib, mapping->it.start, - mapping->it.last + 1, addr + mapping->offset, - flags, gtt_flags); - - if (r) { -- amdgpu_ib_free(adev, &ib); -+ amdgpu_ib_free(adev, ib); -+ kfree(ib); - return r; - } - -- amdgpu_vm_pad_ib(adev, &ib); -- WARN_ON(ib.length_dw > ndw); -+ amdgpu_vm_pad_ib(adev, ib); -+ WARN_ON(ib->length_dw > ndw); - -- r = amdgpu_ib_schedule(adev, 1, &ib, AMDGPU_FENCE_OWNER_VM); -- if (r) { -- amdgpu_ib_free(adev, &ib); -- return r; -- } -- amdgpu_vm_fence_pts(vm, mapping->it.start, -- mapping->it.last + 1, ib.fence); -- if (fence) { -- amdgpu_fence_unref(fence); -- *fence = amdgpu_fence_ref(ib.fence); -- } -- amdgpu_ib_free(adev, &ib); -+ if (amdgpu_enable_scheduler) { -+ int r; -+ uint64_t v_seq; -+ sched_job = amdgpu_cs_parser_create(adev, AMDGPU_FENCE_OWNER_VM, -+ adev->kernel_ctx, ib, 1); -+ if(!sched_job) -+ goto error_free; -+ sched_job->job_param.vm_mapping.vm = vm; -+ sched_job->job_param.vm_mapping.start = mapping->it.start; -+ sched_job->job_param.vm_mapping.last = mapping->it.last; -+ sched_job->job_param.vm_mapping.fence = fence; -+ sched_job->run_job = amdgpu_vm_bo_update_mapping_run_job; -+ sched_job->free_job = amdgpu_vm_free_job; -+ v_seq = atomic64_inc_return(&adev->kernel_ctx->rings[ring->idx].c_entity.last_queued_v_seq); -+ sched_job->uf.sequence = v_seq; -+ amd_sched_push_job(ring->scheduler, -+ &adev->kernel_ctx->rings[ring->idx].c_entity, -+ sched_job); -+ r = amd_sched_wait_emit(&adev->kernel_ctx->rings[ring->idx].c_entity, -+ v_seq, -+ true, -+ -1); -+ if (r) -+ DRM_ERROR("emit timeout\n"); -+ } else { -+ r = amdgpu_ib_schedule(adev, 1, ib, AMDGPU_FENCE_OWNER_VM); -+ if (r) { -+ amdgpu_ib_free(adev, ib); -+ return r; -+ } -+ -+ amdgpu_vm_fence_pts(vm, mapping->it.start, -+ mapping->it.last + 1, ib->fence); -+ if (fence) { -+ amdgpu_fence_unref(fence); -+ *fence = amdgpu_fence_ref(ib->fence); -+ } - -+ amdgpu_ib_free(adev, ib); -+ kfree(ib); -+ } - return 0; -+ -+error_free: -+ if (sched_job) -+ kfree(sched_job); -+ amdgpu_ib_free(adev, ib); -+ kfree(ib); -+ return -ENOMEM; - } - - /** --- -1.9.1 - |