diff options
Diffstat (limited to 'meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3356-drm-amdkfd-Avoid-submitting-an-unnecessary-packet-to.patch')
-rw-r--r-- | meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3356-drm-amdkfd-Avoid-submitting-an-unnecessary-packet-to.patch | 107 |
1 files changed, 107 insertions, 0 deletions
diff --git a/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3356-drm-amdkfd-Avoid-submitting-an-unnecessary-packet-to.patch b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3356-drm-amdkfd-Avoid-submitting-an-unnecessary-packet-to.patch new file mode 100644 index 00000000..56484f5b --- /dev/null +++ b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3356-drm-amdkfd-Avoid-submitting-an-unnecessary-packet-to.patch @@ -0,0 +1,107 @@ +From 02b05b431eaed710dc89896c50e9700d2360e2cf Mon Sep 17 00:00:00 2001 +From: Yong Zhao <yong.zhao@amd.com> +Date: Wed, 27 Sep 2017 00:09:51 -0400 +Subject: [PATCH 3356/4131] drm/amdkfd: Avoid submitting an unnecessary packet + to HWS + +v2: +Make queue mapping interfaces more consistent by passing unmap filter +parameters directly to execute_queues_cpsch, same as unmap_queues_cpsch. + +Signed-off-by: Yong Zhao <yong.zhao@amd.com> +Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com> +Reviewed-by: Oded Gabbay <oded.gabbay@gmail.com> +Signed-off-by: Oded Gabbay <oded.gabbay@gmail.com> +--- + .../gpu/drm/amd/amdkfd/kfd_device_queue_manager.c | 26 ++++++++++++---------- + 1 file changed, 14 insertions(+), 12 deletions(-) + +diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c +index dccb493..f5a5988 100644 +--- a/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c ++++ b/drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c +@@ -44,7 +44,9 @@ static int create_compute_queue_nocpsch(struct device_queue_manager *dqm, + struct queue *q, + struct qcm_process_device *qpd); + +-static int execute_queues_cpsch(struct device_queue_manager *dqm); ++static int execute_queues_cpsch(struct device_queue_manager *dqm, ++ enum kfd_unmap_queues_filter filter, ++ uint32_t filter_param); + static int unmap_queues_cpsch(struct device_queue_manager *dqm, + enum kfd_unmap_queues_filter filter, + uint32_t filter_param); +@@ -729,7 +731,7 @@ static int start_cpsch(struct device_queue_manager *dqm) + init_interrupts(dqm); + + mutex_lock(&dqm->lock); +- execute_queues_cpsch(dqm); ++ execute_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0); + mutex_unlock(&dqm->lock); + + return 0; +@@ -775,7 +777,7 @@ static int create_kernel_queue_cpsch(struct device_queue_manager *dqm, + list_add(&kq->list, &qpd->priv_queue_list); + dqm->queue_count++; + qpd->is_debug = true; +- execute_queues_cpsch(dqm); ++ execute_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0); + mutex_unlock(&dqm->lock); + + return 0; +@@ -786,12 +788,10 @@ static void destroy_kernel_queue_cpsch(struct device_queue_manager *dqm, + struct qcm_process_device *qpd) + { + mutex_lock(&dqm->lock); +- /* here we actually preempt the DIQ */ +- unmap_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES, 0); + list_del(&kq->list); + dqm->queue_count--; + qpd->is_debug = false; +- execute_queues_cpsch(dqm); ++ execute_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES, 0); + /* + * Unconditionally decrement this counter, regardless of the queue's + * type. +@@ -850,7 +850,8 @@ static int create_queue_cpsch(struct device_queue_manager *dqm, struct queue *q, + list_add(&q->list, &qpd->queues_list); + if (q->properties.is_active) { + dqm->queue_count++; +- retval = execute_queues_cpsch(dqm); ++ retval = execute_queues_cpsch(dqm, ++ KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0); + } + + if (q->properties.type == KFD_QUEUE_TYPE_SDMA) +@@ -960,14 +961,15 @@ static int unmap_queues_cpsch(struct device_queue_manager *dqm, + } + + /* dqm->lock mutex has to be locked before calling this function */ +-static int execute_queues_cpsch(struct device_queue_manager *dqm) ++static int execute_queues_cpsch(struct device_queue_manager *dqm, ++ enum kfd_unmap_queues_filter filter, ++ uint32_t filter_param) + { + int retval; + +- retval = unmap_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, +- 0); ++ retval = unmap_queues_cpsch(dqm, filter, filter_param); + if (retval) { +- pr_err("The cp might be in an unrecoverable state due to an unsuccessful queues preemption"); ++ pr_err("The cp might be in an unrecoverable state due to an unsuccessful queues preemption\n"); + return retval; + } + +@@ -1013,7 +1015,7 @@ static int destroy_queue_cpsch(struct device_queue_manager *dqm, + if (q->properties.is_active) + dqm->queue_count--; + +- execute_queues_cpsch(dqm); ++ execute_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0); + + mqd->uninit_mqd(mqd, q->mqd, q->mqd_mem_obj); + +-- +2.7.4 + |