aboutsummaryrefslogtreecommitdiffstats
path: root/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/0017-drm-amdgpu-gfx9-adjust-mqd-allocation-size.patch
diff options
context:
space:
mode:
Diffstat (limited to 'meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/0017-drm-amdgpu-gfx9-adjust-mqd-allocation-size.patch')
-rw-r--r--meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/0017-drm-amdgpu-gfx9-adjust-mqd-allocation-size.patch120
1 files changed, 120 insertions, 0 deletions
diff --git a/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/0017-drm-amdgpu-gfx9-adjust-mqd-allocation-size.patch b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/0017-drm-amdgpu-gfx9-adjust-mqd-allocation-size.patch
new file mode 100644
index 00000000..bbaab96a
--- /dev/null
+++ b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/0017-drm-amdgpu-gfx9-adjust-mqd-allocation-size.patch
@@ -0,0 +1,120 @@
+From 41b947b00a8b8ea722faeeb79556d75402aa93e6 Mon Sep 17 00:00:00 2001
+From: Alex Deucher <alexander.deucher@amd.com>
+Date: Fri, 18 Aug 2017 23:52:37 -0400
+Subject: [PATCH 0017/4131] drm/amdgpu/gfx9: adjust mqd allocation size
+
+To allocate additional space for the dynamic cu masks.
+Confirmed with the hw team that we only need 1 dword
+for the mask. The mask is the same for each SE so
+you only need 1 dword.
+
+Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com>
+Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
+---
+ drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c | 25 ++++++++++++++++++-------
+ drivers/gpu/drm/amd/include/v9_structs.h | 8 ++++++++
+ 2 files changed, 26 insertions(+), 7 deletions(-)
+
+diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c b/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
+index 3981915..e483825 100644
+--- a/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
++++ b/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
+@@ -1346,7 +1346,7 @@ static int gfx_v9_0_sw_init(void *handle)
+ return r;
+
+ /* create MQD for all compute queues as wel as KIQ for SRIOV case */
+- r = amdgpu_gfx_compute_mqd_sw_init(adev, sizeof(struct v9_mqd));
++ r = amdgpu_gfx_compute_mqd_sw_init(adev, sizeof(struct v9_mqd_allocation));
+ if (r)
+ return r;
+
+@@ -2463,6 +2463,13 @@ static int gfx_v9_0_mqd_init(struct amdgpu_ring *ring)
+ mqd->compute_static_thread_mgmt_se3 = 0xffffffff;
+ mqd->compute_misc_reserved = 0x00000003;
+
++ mqd->dynamic_cu_mask_addr_lo =
++ lower_32_bits(ring->mqd_gpu_addr
++ + offsetof(struct v9_mqd_allocation, dynamic_cu_mask));
++ mqd->dynamic_cu_mask_addr_hi =
++ upper_32_bits(ring->mqd_gpu_addr
++ + offsetof(struct v9_mqd_allocation, dynamic_cu_mask));
++
+ eop_base_addr = ring->eop_gpu_addr >> 8;
+ mqd->cp_hqd_eop_base_addr_lo = eop_base_addr;
+ mqd->cp_hqd_eop_base_addr_hi = upper_32_bits(eop_base_addr);
+@@ -2695,7 +2702,7 @@ static int gfx_v9_0_kiq_init_queue(struct amdgpu_ring *ring)
+ if (adev->gfx.in_reset) { /* for GPU_RESET case */
+ /* reset MQD to a clean status */
+ if (adev->gfx.mec.mqd_backup[mqd_idx])
+- memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(*mqd));
++ memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct v9_mqd_allocation));
+
+ /* reset ring buffer */
+ ring->wptr = 0;
+@@ -2707,7 +2714,9 @@ static int gfx_v9_0_kiq_init_queue(struct amdgpu_ring *ring)
+ soc15_grbm_select(adev, 0, 0, 0, 0);
+ mutex_unlock(&adev->srbm_mutex);
+ } else {
+- memset((void *)mqd, 0, sizeof(*mqd));
++ memset((void *)mqd, 0, sizeof(struct v9_mqd_allocation));
++ ((struct v9_mqd_allocation *)mqd)->dynamic_cu_mask = 0xFFFFFFFF;
++ ((struct v9_mqd_allocation *)mqd)->dynamic_rb_mask = 0xFFFFFFFF;
+ mutex_lock(&adev->srbm_mutex);
+ soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
+ gfx_v9_0_mqd_init(ring);
+@@ -2716,7 +2725,7 @@ static int gfx_v9_0_kiq_init_queue(struct amdgpu_ring *ring)
+ mutex_unlock(&adev->srbm_mutex);
+
+ if (adev->gfx.mec.mqd_backup[mqd_idx])
+- memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(*mqd));
++ memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct v9_mqd_allocation));
+ }
+
+ return 0;
+@@ -2729,7 +2738,9 @@ static int gfx_v9_0_kcq_init_queue(struct amdgpu_ring *ring)
+ int mqd_idx = ring - &adev->gfx.compute_ring[0];
+
+ if (!adev->gfx.in_reset && !adev->gfx.in_suspend) {
+- memset((void *)mqd, 0, sizeof(*mqd));
++ memset((void *)mqd, 0, sizeof(struct v9_mqd_allocation));
++ ((struct v9_mqd_allocation *)mqd)->dynamic_cu_mask = 0xFFFFFFFF;
++ ((struct v9_mqd_allocation *)mqd)->dynamic_rb_mask = 0xFFFFFFFF;
+ mutex_lock(&adev->srbm_mutex);
+ soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
+ gfx_v9_0_mqd_init(ring);
+@@ -2737,11 +2748,11 @@ static int gfx_v9_0_kcq_init_queue(struct amdgpu_ring *ring)
+ mutex_unlock(&adev->srbm_mutex);
+
+ if (adev->gfx.mec.mqd_backup[mqd_idx])
+- memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(*mqd));
++ memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct v9_mqd_allocation));
+ } else if (adev->gfx.in_reset) { /* for GPU_RESET case */
+ /* reset MQD to a clean status */
+ if (adev->gfx.mec.mqd_backup[mqd_idx])
+- memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(*mqd));
++ memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct v9_mqd_allocation));
+
+ /* reset ring buffer */
+ ring->wptr = 0;
+diff --git a/drivers/gpu/drm/amd/include/v9_structs.h b/drivers/gpu/drm/amd/include/v9_structs.h
+index 56d79db..2fb25ab 100644
+--- a/drivers/gpu/drm/amd/include/v9_structs.h
++++ b/drivers/gpu/drm/amd/include/v9_structs.h
+@@ -672,6 +672,14 @@ struct v9_mqd {
+ uint32_t reserved_511;
+ };
+
++struct v9_mqd_allocation {
++ struct v9_mqd mqd;
++ uint32_t wptr_poll_mem;
++ uint32_t rptr_report_mem;
++ uint32_t dynamic_cu_mask;
++ uint32_t dynamic_rb_mask;
++};
++
+ /* from vega10 all CSA format is shifted to chain ib compatible mode */
+ struct v9_ce_ib_state {
+ /* section of non chained ib part */
+--
+2.7.4
+