diff options
Diffstat (limited to 'common/recipes-kernel/linux/linux-yocto-4.14.71/4431-drm-amdgpu-vg20-Restruct-uvd.idle_work-to-support-mu.patch')
-rw-r--r-- | common/recipes-kernel/linux/linux-yocto-4.14.71/4431-drm-amdgpu-vg20-Restruct-uvd.idle_work-to-support-mu.patch | 124 |
1 files changed, 124 insertions, 0 deletions
diff --git a/common/recipes-kernel/linux/linux-yocto-4.14.71/4431-drm-amdgpu-vg20-Restruct-uvd.idle_work-to-support-mu.patch b/common/recipes-kernel/linux/linux-yocto-4.14.71/4431-drm-amdgpu-vg20-Restruct-uvd.idle_work-to-support-mu.patch new file mode 100644 index 00000000..c42be53d --- /dev/null +++ b/common/recipes-kernel/linux/linux-yocto-4.14.71/4431-drm-amdgpu-vg20-Restruct-uvd.idle_work-to-support-mu.patch @@ -0,0 +1,124 @@ +From 324b2b32fef0f7032c728cfe6709e978220e2d2e Mon Sep 17 00:00:00 2001 +From: Chaudhary Amit Kumar <chaudharyamit.kumar@amd.com> +Date: Wed, 9 Jan 2019 19:31:08 +0530 +Subject: [PATCH 4431/5725] drm/amdgpu/vg20:Restruct uvd.idle_work to support + multiple instance (v2) + +Vega20 dual-UVD Hardware need two idle_works, restruct to support +multiple instance. + +v2: squash in indentation fix + +Signed-off-by: James Zhu <James.Zhu@amd.com> +Reviewed-by: Alex Deucher <alexander.deucher@amd.com> +Signed-off-by: Alex Deucher <alexander.deucher@amd.com> +Signed-off-by: Chaudhary Amit Kumar <chaudharyamit.kumar@amd.com> +--- + drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c | 17 +++++++++-------- + drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h | 7 ++++++- + 2 files changed, 15 insertions(+), 9 deletions(-) + +diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c +index c9ed917..3c040f3 100644 +--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c ++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c +@@ -129,8 +129,6 @@ int amdgpu_uvd_sw_init(struct amdgpu_device *adev) + unsigned version_major, version_minor, family_id; + int i, j, r; + +- INIT_DELAYED_WORK(&adev->uvd.inst->idle_work, amdgpu_uvd_idle_work_handler); +- + switch (adev->asic_type) { + #ifdef CONFIG_DRM_AMDGPU_CIK + case CHIP_BONAIRE: +@@ -237,6 +235,8 @@ int amdgpu_uvd_sw_init(struct amdgpu_device *adev) + bo_size += AMDGPU_GPU_PAGE_ALIGN(le32_to_cpu(hdr->ucode_size_bytes) + 8); + + for (j = 0; j < adev->uvd.num_uvd_inst; j++) { ++ adev->uvd.inst[j].delayed_work.ip_instance = j; ++ INIT_DELAYED_WORK(&adev->uvd.inst[j].delayed_work.idle_work, amdgpu_uvd_idle_work_handler); + + r = amdgpu_bo_create_kernel(adev, bo_size, PAGE_SIZE, + AMDGPU_GEM_DOMAIN_VRAM, &adev->uvd.inst[j].vcpu_bo, +@@ -317,7 +317,7 @@ int amdgpu_uvd_suspend(struct amdgpu_device *adev) + if (adev->uvd.inst[j].vcpu_bo == NULL) + continue; + +- cancel_delayed_work_sync(&adev->uvd.inst[j].idle_work); ++ cancel_delayed_work_sync(&adev->uvd.inst[j].delayed_work.idle_work); + + /* only valid for physical mode */ + if (adev->asic_type < CHIP_POLARIS10) { +@@ -1144,9 +1144,10 @@ int amdgpu_uvd_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, + + static void amdgpu_uvd_idle_work_handler(struct work_struct *work) + { ++ struct amdgpu_delayed_work *my_work = (struct amdgpu_delayed_work *)work; + struct amdgpu_device *adev = +- container_of(work, struct amdgpu_device, uvd.inst->idle_work.work); +- unsigned fences = amdgpu_fence_count_emitted(&adev->uvd.inst->ring); ++ container_of(work, struct amdgpu_device, uvd.inst[my_work->ip_instance].delayed_work.idle_work.work); ++ unsigned fences = amdgpu_fence_count_emitted(&adev->uvd.inst[my_work->ip_instance].ring); + + if (fences == 0) { + if (adev->pm.dpm_enabled) { +@@ -1160,7 +1161,7 @@ static void amdgpu_uvd_idle_work_handler(struct work_struct *work) + AMD_CG_STATE_GATE); + } + } else { +- schedule_delayed_work(&adev->uvd.inst->idle_work, UVD_IDLE_TIMEOUT); ++ schedule_delayed_work(&adev->uvd.inst[my_work->ip_instance].delayed_work.idle_work, UVD_IDLE_TIMEOUT); + } + } + +@@ -1172,7 +1173,7 @@ void amdgpu_uvd_ring_begin_use(struct amdgpu_ring *ring) + if (amdgpu_sriov_vf(adev)) + return; + +- set_clocks = !cancel_delayed_work_sync(&adev->uvd.inst->idle_work); ++ set_clocks = !cancel_delayed_work_sync(&adev->uvd.inst[ring->me].delayed_work.idle_work); + if (set_clocks) { + if (adev->pm.dpm_enabled) { + amdgpu_dpm_enable_uvd(adev, true); +@@ -1189,7 +1190,7 @@ void amdgpu_uvd_ring_begin_use(struct amdgpu_ring *ring) + void amdgpu_uvd_ring_end_use(struct amdgpu_ring *ring) + { + if (!amdgpu_sriov_vf(ring->adev)) +- schedule_delayed_work(&ring->adev->uvd.inst->idle_work, UVD_IDLE_TIMEOUT); ++ schedule_delayed_work(&ring->adev->uvd.inst[ring->me].delayed_work.idle_work, UVD_IDLE_TIMEOUT); + } + + /** +diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h +index b1579fb..7801eb8 100644 +--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h ++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.h +@@ -37,6 +37,11 @@ + (AMDGPU_GPU_PAGE_ALIGN(le32_to_cpu(((const struct common_firmware_header *)(adev)->uvd.fw->data)->ucode_size_bytes) + \ + 8) - AMDGPU_UVD_FIRMWARE_OFFSET) + ++struct amdgpu_delayed_work{ ++ struct delayed_work idle_work; ++ unsigned ip_instance; ++}; ++ + struct amdgpu_uvd_inst { + struct amdgpu_bo *vcpu_bo; + void *cpu_addr; +@@ -44,12 +49,12 @@ struct amdgpu_uvd_inst { + void *saved_bo; + atomic_t handles[AMDGPU_MAX_UVD_HANDLES]; + struct drm_file *filp[AMDGPU_MAX_UVD_HANDLES]; +- struct delayed_work idle_work; + struct amdgpu_ring ring; + struct amdgpu_ring ring_enc[AMDGPU_MAX_UVD_ENC_RINGS]; + struct amdgpu_irq_src irq; + struct drm_sched_entity entity; + struct drm_sched_entity entity_enc; ++ struct amdgpu_delayed_work delayed_work; + uint32_t srbm_soft_reset; + }; + +-- +2.7.4 + |