aboutsummaryrefslogtreecommitdiffstats
path: root/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/2055-drm-amdgpu-cleanup-job-reset-routine-v2.patch
diff options
context:
space:
mode:
Diffstat (limited to 'meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/2055-drm-amdgpu-cleanup-job-reset-routine-v2.patch')
-rw-r--r--meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/2055-drm-amdgpu-cleanup-job-reset-routine-v2.patch114
1 files changed, 114 insertions, 0 deletions
diff --git a/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/2055-drm-amdgpu-cleanup-job-reset-routine-v2.patch b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/2055-drm-amdgpu-cleanup-job-reset-routine-v2.patch
new file mode 100644
index 00000000..dbe748ce
--- /dev/null
+++ b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/2055-drm-amdgpu-cleanup-job-reset-routine-v2.patch
@@ -0,0 +1,114 @@
+From 22c6aecb8e3ec8a1b6c1fb43076e86a7de4c3780 Mon Sep 17 00:00:00 2001
+From: Monk Liu <Monk.Liu@amd.com>
+Date: Mon, 16 Oct 2017 19:46:43 +0800
+Subject: [PATCH 2055/4131] drm/amdgpu:cleanup job reset routine(v2)
+
+merge the setting guilty on context into this function
+to avoid implement extra routine.
+
+v2:
+go through entity list and compare the fence_ctx
+before operate on the entity, otherwise the entity
+may be just a wild pointer
+
+Change-Id: I7a0063464fdc85d5ac9080046380e745565ff540
+Signed-off-by: Monk Liu <Monk.Liu@amd.com>
+Reviewed-by: Chunming Zhou <David1.Zhou@amd.com>
+
+ Conflicts:
+ drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
+---
+ drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 4 ++--
+ drivers/gpu/drm/amd/scheduler/gpu_scheduler.c | 31 ++++++++++++++++++++++++++-
+ drivers/gpu/drm/amd/scheduler/gpu_scheduler.h | 2 +-
+ 3 files changed, 33 insertions(+), 4 deletions(-)
+
+diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
+index d0aaed4..29349bf 100644
+--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
+@@ -2896,7 +2896,7 @@ int amdgpu_sriov_gpu_reset(struct amdgpu_device *adev, struct amdgpu_job *job)
+ amd_sched_job_kickout(&job->base);
+
+ /* only do job_reset on the hang ring if @job not NULL */
+- amd_sched_hw_job_reset(&ring->sched);
++ amd_sched_hw_job_reset(&ring->sched, NULL);
+
+ /* after all hw jobs are reset, hw fence is meaningless, so force_completion */
+ amdgpu_fence_driver_force_completion(ring);
+@@ -3017,7 +3017,7 @@ int amdgpu_gpu_reset(struct amdgpu_device *adev)
+ if (!ring || !ring->sched.thread)
+ continue;
+ kthread_park(ring->sched.thread);
+- amd_sched_hw_job_reset(&ring->sched);
++ amd_sched_hw_job_reset(&ring->sched, NULL);
+ /* after all hw jobs are reset, hw fence is meaningless, so force_completion */
+ amdgpu_fence_driver_force_completion(ring);
+ }
+diff --git a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
+index 041510e..2634846 100644
+--- a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
++++ b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.c
+@@ -443,9 +443,18 @@ static void amd_sched_job_timedout(struct work_struct *work)
+ job->sched->ops->timedout_job(job);
+ }
+
+-void amd_sched_hw_job_reset(struct amd_gpu_scheduler *sched)
++static void amd_sched_set_guilty(struct amd_sched_job *s_job)
++{
++ if (atomic_inc_return(&s_job->karma) > s_job->sched->hang_limit)
++ if (s_job->s_entity->guilty)
++ atomic_set(s_job->s_entity->guilty, 1);
++}
++
++void amd_sched_hw_job_reset(struct amd_gpu_scheduler *sched, struct amd_sched_job *bad)
+ {
+ struct amd_sched_job *s_job;
++ struct amd_sched_entity *entity, *tmp;
++ int i;;
+
+ spin_lock(&sched->job_list_lock);
+ list_for_each_entry_reverse(s_job, &sched->ring_mirror_list, node) {
+@@ -458,6 +467,26 @@ void amd_sched_hw_job_reset(struct amd_gpu_scheduler *sched)
+ }
+ }
+ spin_unlock(&sched->job_list_lock);
++
++ if (bad) {
++ bool found = false;
++
++ for (i = AMD_SCHED_PRIORITY_MIN; i < AMD_SCHED_PRIORITY_MAX; i++ ) {
++ struct amd_sched_rq *rq = &sched->sched_rq[i];
++
++ spin_lock(&rq->lock);
++ list_for_each_entry_safe(entity, tmp, &rq->entities, list) {
++ if (bad->s_fence->scheduled.context == entity->fence_context) {
++ found = true;
++ amd_sched_set_guilty(bad);
++ break;
++ }
++ }
++ spin_unlock(&rq->lock);
++ if (found)
++ break;
++ }
++ }
+ }
+
+ void amd_sched_job_kickout(struct amd_sched_job *s_job)
+diff --git a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
+index 7342763..64c887f 100644
+--- a/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
++++ b/drivers/gpu/drm/amd/scheduler/gpu_scheduler.h
+@@ -173,7 +173,7 @@ int amd_sched_job_init(struct amd_sched_job *job,
+ struct amd_gpu_scheduler *sched,
+ struct amd_sched_entity *entity,
+ void *owner);
+-void amd_sched_hw_job_reset(struct amd_gpu_scheduler *sched);
++void amd_sched_hw_job_reset(struct amd_gpu_scheduler *sched, struct amd_sched_job *job);
+ void amd_sched_job_recovery(struct amd_gpu_scheduler *sched);
+ bool amd_sched_dependency_optimized(struct dma_fence* fence,
+ struct amd_sched_entity *entity);
+--
+2.7.4
+