diff options
Diffstat (limited to 'meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.19.8/2924-drm-amdgpu-add-paging-queue-support-for-8-SDMA-insta.patch')
-rw-r--r-- | meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.19.8/2924-drm-amdgpu-add-paging-queue-support-for-8-SDMA-insta.patch | 118 |
1 files changed, 118 insertions, 0 deletions
diff --git a/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.19.8/2924-drm-amdgpu-add-paging-queue-support-for-8-SDMA-insta.patch b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.19.8/2924-drm-amdgpu-add-paging-queue-support-for-8-SDMA-insta.patch new file mode 100644 index 00000000..85c1cd48 --- /dev/null +++ b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.19.8/2924-drm-amdgpu-add-paging-queue-support-for-8-SDMA-insta.patch @@ -0,0 +1,118 @@ +From 7b6455c3b0e66a9ea08370f54e2b04198d8101a7 Mon Sep 17 00:00:00 2001 +From: Le Ma <le.ma@amd.com> +Date: Wed, 27 Feb 2019 14:32:51 +0800 +Subject: [PATCH 2924/2940] drm/amdgpu: add paging queue support for 8 SDMA + instances on Arcturus +MIME-Version: 1.0 +Content-Type: text/plain; charset=UTF-8 +Content-Transfer-Encoding: 8bit + +Signed-off-by: Le Ma <le.ma@amd.com> +Reviewed-by: Christian König <christian.koenig@amd.com> +Signed-off-by: Alex Deucher <alexander.deucher@amd.com> +--- + drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c | 62 +++++++++++++++++++++----- + 1 file changed, 51 insertions(+), 11 deletions(-) + +diff --git a/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c b/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c +index 467fcfbe8085..72183c0382d2 100644 +--- a/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c ++++ b/drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c +@@ -730,16 +730,20 @@ static void sdma_v4_0_rlc_stop(struct amdgpu_device *adev) + */ + static void sdma_v4_0_page_stop(struct amdgpu_device *adev) + { +- struct amdgpu_ring *sdma0 = &adev->sdma.instance[0].page; +- struct amdgpu_ring *sdma1 = &adev->sdma.instance[1].page; ++ struct amdgpu_ring *sdma[AMDGPU_MAX_SDMA_INSTANCES]; + u32 rb_cntl, ib_cntl; + int i; +- +- if ((adev->mman.buffer_funcs_ring == sdma0) || +- (adev->mman.buffer_funcs_ring == sdma1)) +- amdgpu_ttm_set_buffer_funcs_status(adev, false); ++ bool unset = false; + + for (i = 0; i < adev->sdma.num_instances; i++) { ++ sdma[i] = &adev->sdma.instance[i].page; ++ ++ if ((adev->mman.buffer_funcs_ring == sdma[i]) && ++ (unset == false)) { ++ amdgpu_ttm_set_buffer_funcs_status(adev, false); ++ unset = true; ++ } ++ + rb_cntl = RREG32_SDMA(i, mmSDMA0_PAGE_RB_CNTL); + rb_cntl = REG_SET_FIELD(rb_cntl, SDMA0_PAGE_RB_CNTL, + RB_ENABLE, 0); +@@ -748,10 +752,9 @@ static void sdma_v4_0_page_stop(struct amdgpu_device *adev) + ib_cntl = REG_SET_FIELD(ib_cntl, SDMA0_PAGE_IB_CNTL, + IB_ENABLE, 0); + WREG32_SDMA(i, mmSDMA0_PAGE_IB_CNTL, ib_cntl); +- } + +- sdma0->sched.ready = false; +- sdma1->sched.ready = false; ++ sdma[i]->sched.ready = false; ++ } + } + + /** +@@ -2258,7 +2261,39 @@ static const struct amdgpu_ring_funcs sdma_v4_0_page_ring_funcs = { + .align_mask = 0xf, + .nop = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP), + .support_64bit_ptrs = true, +- .vmhub = AMDGPU_MMHUB_0, ++ .vmhub = AMDGPU_MMHUB_0, ++ .get_rptr = sdma_v4_0_ring_get_rptr, ++ .get_wptr = sdma_v4_0_page_ring_get_wptr, ++ .set_wptr = sdma_v4_0_page_ring_set_wptr, ++ .emit_frame_size = ++ 6 + /* sdma_v4_0_ring_emit_hdp_flush */ ++ 3 + /* hdp invalidate */ ++ 6 + /* sdma_v4_0_ring_emit_pipeline_sync */ ++ /* sdma_v4_0_ring_emit_vm_flush */ ++ SOC15_FLUSH_GPU_TLB_NUM_WREG * 3 + ++ SOC15_FLUSH_GPU_TLB_NUM_REG_WAIT * 6 + ++ 10 + 10 + 10, /* sdma_v4_0_ring_emit_fence x3 for user fence, vm fence */ ++ .emit_ib_size = 7 + 6, /* sdma_v4_0_ring_emit_ib */ ++ .emit_ib = sdma_v4_0_ring_emit_ib, ++ .emit_fence = sdma_v4_0_ring_emit_fence, ++ .emit_pipeline_sync = sdma_v4_0_ring_emit_pipeline_sync, ++ .emit_vm_flush = sdma_v4_0_ring_emit_vm_flush, ++ .emit_hdp_flush = sdma_v4_0_ring_emit_hdp_flush, ++ .test_ring = sdma_v4_0_ring_test_ring, ++ .test_ib = sdma_v4_0_ring_test_ib, ++ .insert_nop = sdma_v4_0_ring_insert_nop, ++ .pad_ib = sdma_v4_0_ring_pad_ib, ++ .emit_wreg = sdma_v4_0_ring_emit_wreg, ++ .emit_reg_wait = sdma_v4_0_ring_emit_reg_wait, ++ .emit_reg_write_reg_wait = amdgpu_ring_emit_reg_write_reg_wait_helper, ++}; ++ ++static const struct amdgpu_ring_funcs sdma_v4_0_page_ring_funcs_2nd_mmhub = { ++ .type = AMDGPU_RING_TYPE_SDMA, ++ .align_mask = 0xf, ++ .nop = SDMA_PKT_NOP_HEADER_OP(SDMA_OP_NOP), ++ .support_64bit_ptrs = true, ++ .vmhub = AMDGPU_MMHUB_1, + .get_rptr = sdma_v4_0_ring_get_rptr, + .get_wptr = sdma_v4_0_page_ring_get_wptr, + .set_wptr = sdma_v4_0_page_ring_set_wptr, +@@ -2298,7 +2333,12 @@ static void sdma_v4_0_set_ring_funcs(struct amdgpu_device *adev) + &sdma_v4_0_ring_funcs; + adev->sdma.instance[i].ring.me = i; + if (adev->sdma.has_page_queue) { +- adev->sdma.instance[i].page.funcs = &sdma_v4_0_page_ring_funcs; ++ if (adev->asic_type == CHIP_ARCTURUS && i >=5) ++ adev->sdma.instance[i].page.funcs = ++ &sdma_v4_0_page_ring_funcs_2nd_mmhub; ++ else ++ adev->sdma.instance[i].page.funcs = ++ &sdma_v4_0_page_ring_funcs; + adev->sdma.instance[i].page.me = i; + } + } +-- +2.17.1 + |