aboutsummaryrefslogtreecommitdiffstats
path: root/common/recipes-kernel/linux/linux-yocto-4.19.8/1466-drm-amd-powerplay-Unify-smu-handle-task-function-v2.patch
diff options
context:
space:
mode:
Diffstat (limited to 'common/recipes-kernel/linux/linux-yocto-4.19.8/1466-drm-amd-powerplay-Unify-smu-handle-task-function-v2.patch')
-rw-r--r--common/recipes-kernel/linux/linux-yocto-4.19.8/1466-drm-amd-powerplay-Unify-smu-handle-task-function-v2.patch672
1 files changed, 672 insertions, 0 deletions
diff --git a/common/recipes-kernel/linux/linux-yocto-4.19.8/1466-drm-amd-powerplay-Unify-smu-handle-task-function-v2.patch b/common/recipes-kernel/linux/linux-yocto-4.19.8/1466-drm-amd-powerplay-Unify-smu-handle-task-function-v2.patch
new file mode 100644
index 00000000..84acc1c9
--- /dev/null
+++ b/common/recipes-kernel/linux/linux-yocto-4.19.8/1466-drm-amd-powerplay-Unify-smu-handle-task-function-v2.patch
@@ -0,0 +1,672 @@
+From 00a3e0a369c55fcc8fa511076362056ca0807128 Mon Sep 17 00:00:00 2001
+From: Likun Gao <Likun.Gao@amd.com>
+Date: Thu, 24 Jan 2019 19:53:40 +0800
+Subject: [PATCH 1466/2940] drm/amd/powerplay: Unify smu handle task function
+ (v2)
+
+Unify power stade adjust function into smu_handle_task by the judgment
+of task_id.
+Move functions which have no relationship with smu version into the file
+of amdgpu_smu.
+Modified the function of smu_display_config_changed into two part.
+Unify some similiar function.
+
+v2: Correct the operation of upload dpm level when force dpm limit value.
+
+Signed-off-by: Likun Gao <Likun.Gao@amd.com>
+Reviewed-by: Huang Rui <ray.huang@amd.com>
+Acked-by: Alex Deucher <alexander.deucher@amd.com>
+---
+ drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c | 46 ++--
+ drivers/gpu/drm/amd/powerplay/amdgpu_smu.c | 127 +++++++++
+ .../gpu/drm/amd/powerplay/inc/amdgpu_smu.h | 31 +++
+ drivers/gpu/drm/amd/powerplay/vega20_ppt.c | 258 +++++-------------
+ 4 files changed, 250 insertions(+), 212 deletions(-)
+
+diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c
+index b83981284a7c..47d2ba528a0f 100644
+--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c
++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c
+@@ -2608,28 +2608,38 @@ void amdgpu_pm_compute_clocks(struct amdgpu_device *adev)
+ amdgpu_fence_wait_empty(ring);
+ }
+
+- if (adev->powerplay.pp_funcs->dispatch_tasks) {
+- if (!amdgpu_device_has_dc_support(adev)) {
++ if (is_support_sw_smu(adev)) {
++ struct smu_context *smu = &adev->smu;
++ struct smu_dpm_context *smu_dpm = &adev->smu.smu_dpm;
++ mutex_lock(&(smu->mutex));
++ smu_handle_task(&adev->smu,
++ smu_dpm->dpm_level,
++ AMD_PP_TASK_DISPLAY_CONFIG_CHANGE);
++ mutex_unlock(&(smu->mutex));
++ } else {
++ if (adev->powerplay.pp_funcs->dispatch_tasks) {
++ if (!amdgpu_device_has_dc_support(adev)) {
++ mutex_lock(&adev->pm.mutex);
++ amdgpu_dpm_get_active_displays(adev);
++ adev->pm.pm_display_cfg.num_display = adev->pm.dpm.new_active_crtc_count;
++ adev->pm.pm_display_cfg.vrefresh = amdgpu_dpm_get_vrefresh(adev);
++ adev->pm.pm_display_cfg.min_vblank_time = amdgpu_dpm_get_vblank_time(adev);
++ /* we have issues with mclk switching with refresh rates over 120 hz on the non-DC code. */
++ if (adev->pm.pm_display_cfg.vrefresh > 120)
++ adev->pm.pm_display_cfg.min_vblank_time = 0;
++ if (adev->powerplay.pp_funcs->display_configuration_change)
++ adev->powerplay.pp_funcs->display_configuration_change(
++ adev->powerplay.pp_handle,
++ &adev->pm.pm_display_cfg);
++ mutex_unlock(&adev->pm.mutex);
++ }
++ amdgpu_dpm_dispatch_task(adev, AMD_PP_TASK_DISPLAY_CONFIG_CHANGE, NULL);
++ } else {
+ mutex_lock(&adev->pm.mutex);
+ amdgpu_dpm_get_active_displays(adev);
+- adev->pm.pm_display_cfg.num_display = adev->pm.dpm.new_active_crtc_count;
+- adev->pm.pm_display_cfg.vrefresh = amdgpu_dpm_get_vrefresh(adev);
+- adev->pm.pm_display_cfg.min_vblank_time = amdgpu_dpm_get_vblank_time(adev);
+- /* we have issues with mclk switching with refresh rates over 120 hz on the non-DC code. */
+- if (adev->pm.pm_display_cfg.vrefresh > 120)
+- adev->pm.pm_display_cfg.min_vblank_time = 0;
+- if (adev->powerplay.pp_funcs->display_configuration_change)
+- adev->powerplay.pp_funcs->display_configuration_change(
+- adev->powerplay.pp_handle,
+- &adev->pm.pm_display_cfg);
++ amdgpu_dpm_change_power_state_locked(adev);
+ mutex_unlock(&adev->pm.mutex);
+ }
+- amdgpu_dpm_dispatch_task(adev, AMD_PP_TASK_DISPLAY_CONFIG_CHANGE, NULL);
+- } else {
+- mutex_lock(&adev->pm.mutex);
+- amdgpu_dpm_get_active_displays(adev);
+- amdgpu_dpm_change_power_state_locked(adev);
+- mutex_unlock(&adev->pm.mutex);
+ }
+ }
+
+diff --git a/drivers/gpu/drm/amd/powerplay/amdgpu_smu.c b/drivers/gpu/drm/amd/powerplay/amdgpu_smu.c
+index d6578be92196..48b346f95cd6 100644
+--- a/drivers/gpu/drm/amd/powerplay/amdgpu_smu.c
++++ b/drivers/gpu/drm/amd/powerplay/amdgpu_smu.c
+@@ -1041,6 +1041,133 @@ static int smu_enable_umd_pstate(void *handle,
+ return 0;
+ }
+
++int smu_unforce_dpm_levels(struct smu_context *smu)
++{
++ int ret = 0;
++
++ ret = smu_upload_dpm_level(smu, false);
++ if (ret) {
++ pr_err("Failed to upload DPM Bootup Levels!");
++ return ret;
++ }
++
++ ret = smu_upload_dpm_level(smu, true);
++ if (ret) {
++ pr_err("Failed to upload DPM Max Levels!");
++ return ret;
++ }
++
++ return ret;
++}
++
++int smu_adjust_power_state_dynamic(struct smu_context *smu,
++ enum amd_dpm_forced_level level,
++ bool skip_display_settings)
++{
++ int ret = 0;
++ int index = 0;
++ uint32_t sclk_mask, mclk_mask, soc_mask;
++ long workload;
++ struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm);
++
++ if (!skip_display_settings) {
++ ret = smu_display_config_changed(smu);
++ if (ret) {
++ pr_err("Failed to change display config!");
++ return ret;
++ }
++ }
++
++ ret = smu_apply_clocks_adjust_rules(smu);
++ if (ret) {
++ pr_err("Failed to apply clocks adjust rules!");
++ return ret;
++ }
++
++ if (!skip_display_settings) {
++ ret = smu_notify_smc_dispaly_config(smu);
++ if (ret) {
++ pr_err("Failed to notify smc display config!");
++ return ret;
++ }
++ }
++
++ if (smu_dpm_ctx->dpm_level != level) {
++ switch (level) {
++ case AMD_DPM_FORCED_LEVEL_HIGH:
++ ret = smu_force_dpm_limit_value(smu, true);
++ break;
++ case AMD_DPM_FORCED_LEVEL_LOW:
++ ret = smu_force_dpm_limit_value(smu, false);
++ break;
++
++ case AMD_DPM_FORCED_LEVEL_AUTO:
++ ret = smu_unforce_dpm_levels(smu);
++ break;
++
++ case AMD_DPM_FORCED_LEVEL_PROFILE_STANDARD:
++ case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_SCLK:
++ case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_MCLK:
++ case AMD_DPM_FORCED_LEVEL_PROFILE_PEAK:
++ ret = smu_get_profiling_clk_mask(smu, level,
++ &sclk_mask,
++ &mclk_mask,
++ &soc_mask);
++ if (ret)
++ return ret;
++ smu_force_clk_levels(smu, PP_SCLK, 1 << sclk_mask);
++ smu_force_clk_levels(smu, PP_MCLK, 1 << mclk_mask);
++ break;
++
++ case AMD_DPM_FORCED_LEVEL_MANUAL:
++ case AMD_DPM_FORCED_LEVEL_PROFILE_EXIT:
++ default:
++ break;
++ }
++
++ if (!ret)
++ smu_dpm_ctx->dpm_level = level;
++ }
++
++ if (smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) {
++ index = fls(smu->workload_mask);
++ index = index > 0 && index <= WORKLOAD_POLICY_MAX ? index - 1 : 0;
++ workload = smu->workload_setting[index];
++
++ if (smu->power_profile_mode != workload)
++ smu_set_power_profile_mode(smu, &workload, 0);
++ }
++
++ return ret;
++}
++
++int smu_handle_task(struct smu_context *smu,
++ enum amd_dpm_forced_level level,
++ enum amd_pp_task task_id)
++{
++ int ret = 0;
++
++ switch (task_id) {
++ case AMD_PP_TASK_DISPLAY_CONFIG_CHANGE:
++ ret = smu_pre_display_config_changed(smu);
++ if (ret)
++ return ret;
++ ret = smu_set_cpu_power_state(smu);
++ if (ret)
++ return ret;
++ ret = smu_adjust_power_state_dynamic(smu, level, false);
++ break;
++ case AMD_PP_TASK_COMPLETE_INIT:
++ case AMD_PP_TASK_READJUST_POWER_STATE:
++ ret = smu_adjust_power_state_dynamic(smu, level, true);
++ break;
++ default:
++ break;
++ }
++
++ return ret;
++}
++
+ const struct amd_ip_funcs smu_ip_funcs = {
+ .name = "smu",
+ .early_init = smu_early_init,
+diff --git a/drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h b/drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h
+index 53ca9530ed1f..db050978020f 100644
+--- a/drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h
++++ b/drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h
+@@ -437,6 +437,18 @@ struct pptable_funcs {
+ int (*set_power_profile_mode)(struct smu_context *smu, long *input, uint32_t size);
+ enum amd_dpm_forced_level (*get_performance_level)(struct smu_context *smu);
+ int (*force_performance_level)(struct smu_context *smu, enum amd_dpm_forced_level level);
++ int (*pre_display_config_changed)(struct smu_context *smu);
++ int (*display_config_changed)(struct smu_context *smu);
++ int (*apply_clocks_adjust_rules)(struct smu_context *smu);
++ int (*notify_smc_dispaly_config)(struct smu_context *smu);
++ int (*force_dpm_limit_value)(struct smu_context *smu, bool highest);
++ int (*upload_dpm_level)(struct smu_context *smu, bool max);
++ int (*get_profiling_clk_mask)(struct smu_context *smu,
++ enum amd_dpm_forced_level level,
++ uint32_t *sclk_mask,
++ uint32_t *mclk_mask,
++ uint32_t *soc_mask);
++ int (*set_cpu_power_state)(struct smu_context *smu);
+ };
+
+ struct smu_funcs
+@@ -628,6 +640,22 @@ struct smu_funcs
+ ((smu)->ppt_funcs->get_performance_level ? (smu)->ppt_funcs->get_performance_level((smu)) : 0)
+ #define smu_force_performance_level(smu, level) \
+ ((smu)->ppt_funcs->force_performance_level ? (smu)->ppt_funcs->force_performance_level((smu), (level)) : 0)
++#define smu_pre_display_config_changed(smu) \
++ ((smu)->ppt_funcs->pre_display_config_changed ? (smu)->ppt_funcs->pre_display_config_changed((smu)) : 0)
++#define smu_display_config_changed(smu) \
++ ((smu)->ppt_funcs->display_config_changed ? (smu)->ppt_funcs->display_config_changed((smu)) : 0)
++#define smu_apply_clocks_adjust_rules(smu) \
++ ((smu)->ppt_funcs->apply_clocks_adjust_rules ? (smu)->ppt_funcs->apply_clocks_adjust_rules((smu)) : 0)
++#define smu_notify_smc_dispaly_config(smu) \
++ ((smu)->ppt_funcs->notify_smc_dispaly_config ? (smu)->ppt_funcs->notify_smc_dispaly_config((smu)) : 0)
++#define smu_force_dpm_limit_value(smu, highest) \
++ ((smu)->ppt_funcs->force_dpm_limit_value ? (smu)->ppt_funcs->force_dpm_limit_value((smu), (highest)) : 0)
++#define smu_upload_dpm_level(smu, max) \
++ ((smu)->ppt_funcs->upload_dpm_level ? (smu)->ppt_funcs->upload_dpm_level((smu), (max)) : 0)
++#define smu_get_profiling_clk_mask(smu, level, sclk_mask, mclk_mask, soc_mask) \
++ ((smu)->ppt_funcs->get_profiling_clk_mask ? (smu)->ppt_funcs->get_profiling_clk_mask((smu), (level), (sclk_mask), (mclk_mask), (soc_mask)) : 0)
++#define smu_set_cpu_power_state(smu) \
++ ((smu)->ppt_funcs->set_cpu_power_state ? (smu)->ppt_funcs->set_cpu_power_state((smu)) : 0)
+
+ #define smu_msg_get_index(smu, msg) \
+ ((smu)->ppt_funcs? ((smu)->ppt_funcs->get_smu_msg_index? (smu)->ppt_funcs->get_smu_msg_index((smu), (msg)) : -EINVAL) : -EINVAL)
+@@ -699,4 +727,7 @@ extern int smu_display_configuration_change(struct smu_context *smu, const
+ extern int smu_get_current_clocks(struct smu_context *smu,
+ struct amd_pp_clock_info *clocks);
+ extern int smu_dpm_set_power_gate(struct smu_context *smu,uint32_t block_type, bool gate);
++extern int smu_handle_task(struct smu_context *smu,
++ enum amd_dpm_forced_level level,
++ enum amd_pp_task task_id);
+ #endif
+diff --git a/drivers/gpu/drm/amd/powerplay/vega20_ppt.c b/drivers/gpu/drm/amd/powerplay/vega20_ppt.c
+index a90bf77dd9eb..300462aff83a 100644
+--- a/drivers/gpu/drm/amd/powerplay/vega20_ppt.c
++++ b/drivers/gpu/drm/amd/powerplay/vega20_ppt.c
+@@ -877,71 +877,39 @@ static int vega20_print_clk_levels(struct smu_context *smu,
+ return size;
+ }
+
+-static int vega20_upload_dpm_min_level(struct smu_context *smu)
++static int vega20_upload_dpm_level(struct smu_context *smu, bool max)
+ {
+ struct vega20_dpm_table *dpm_table;
+ struct vega20_single_dpm_table *single_dpm_table;
+- uint32_t min_freq;
++ uint32_t freq;
+ int ret = 0;
+
+ dpm_table = smu->smu_dpm.dpm_context;
+
+ if (smu_feature_is_enabled(smu, FEATURE_DPM_GFXCLK_BIT)) {
+ single_dpm_table = &(dpm_table->gfx_table);
+- min_freq = single_dpm_table->dpm_state.soft_min_level;
++ freq = max ? single_dpm_table->dpm_state.soft_max_level :
++ single_dpm_table->dpm_state.soft_min_level;
+ ret = smu_send_smc_msg_with_param(smu,
+- SMU_MSG_SetSoftMinByFreq,
+- (PPCLK_GFXCLK << 16) | (min_freq & 0xffff));
++ (max ? SMU_MSG_SetSoftMaxByFreq : SMU_MSG_SetSoftMinByFreq),
++ (PPCLK_GFXCLK << 16) | (freq & 0xffff));
+ if (ret) {
+- pr_err("Failed to set soft min gfxclk !\n");
++ pr_err("Failed to set soft %s gfxclk !\n",
++ max ? "max" : "min");
+ return ret;
+ }
+ }
+
+ if (smu_feature_is_enabled(smu, FEATURE_DPM_UCLK_BIT)) {
+ single_dpm_table = &(dpm_table->mem_table);
+- min_freq = single_dpm_table->dpm_state.soft_min_level;
++ freq = max ? single_dpm_table->dpm_state.soft_max_level :
++ single_dpm_table->dpm_state.soft_min_level;
+ ret = smu_send_smc_msg_with_param(smu,
+- SMU_MSG_SetSoftMinByFreq,
+- (PPCLK_UCLK << 16) | (min_freq & 0xffff));
++ (max ? SMU_MSG_SetSoftMaxByFreq : SMU_MSG_SetSoftMinByFreq),
++ (PPCLK_UCLK << 16) | (freq & 0xffff));
+ if (ret) {
+- pr_err("Failed to set soft min memclk !\n");
+- return ret;
+- }
+- }
+-
+- return ret;
+-}
+-
+-static int vega20_upload_dpm_max_level(struct smu_context *smu)
+-{
+- struct vega20_dpm_table *dpm_table;
+- struct vega20_single_dpm_table *single_dpm_table;
+- uint32_t max_freq;
+- int ret = 0;
+-
+- dpm_table = smu->smu_dpm.dpm_context;
+-
+- if (smu_feature_is_enabled(smu, FEATURE_DPM_GFXCLK_BIT)) {
+- single_dpm_table = &(dpm_table->gfx_table);
+- max_freq = single_dpm_table->dpm_state.soft_max_level;
+- ret = smu_send_smc_msg_with_param(smu,
+- SMU_MSG_SetSoftMaxByFreq,
+- (PPCLK_GFXCLK << 16) | (max_freq & 0xffff));
+- if (ret) {
+- pr_err("Failed to set soft max gfxclk !\n");
+- return ret;
+- }
+- }
+-
+- if (smu_feature_is_enabled(smu, FEATURE_DPM_UCLK_BIT)) {
+- single_dpm_table = &(dpm_table->mem_table);
+- max_freq = single_dpm_table->dpm_state.soft_max_level;
+- ret = smu_send_smc_msg_with_param(smu,
+- SMU_MSG_SetSoftMaxByFreq,
+- (PPCLK_UCLK << 16) | (max_freq & 0xffff));
+- if (ret) {
+- pr_err("Failed to set soft max memclk !\n");
++ pr_err("Failed to set soft %s memclk !\n",
++ max ? "max" : "min");
+ return ret;
+ }
+ }
+@@ -986,13 +954,13 @@ static int vega20_force_clk_levels(struct smu_context *smu,
+ single_dpm_table->dpm_state.soft_max_level =
+ single_dpm_table->dpm_levels[soft_max_level].value;
+
+- ret = vega20_upload_dpm_min_level(smu);
++ ret = vega20_upload_dpm_level(smu, false);
+ if (ret) {
+ pr_err("Failed to upload boot level to lowest!\n");
+ break;
+ }
+
+- ret = vega20_upload_dpm_max_level(smu);
++ ret = vega20_upload_dpm_level(smu, true);
+ if (ret)
+ pr_err("Failed to upload dpm max level to highest!\n");
+
+@@ -1013,13 +981,13 @@ static int vega20_force_clk_levels(struct smu_context *smu,
+ single_dpm_table->dpm_state.soft_max_level =
+ single_dpm_table->dpm_levels[soft_max_level].value;
+
+- ret = vega20_upload_dpm_min_level(smu);
++ ret = vega20_upload_dpm_level(smu, false);
+ if (ret) {
+ pr_err("Failed to upload boot level to lowest!\n");
+ break;
+ }
+
+- ret = vega20_upload_dpm_max_level(smu);
++ ret = vega20_upload_dpm_level(smu, true);
+ if (ret)
+ pr_err("Failed to upload dpm max level to highest!\n");
+
+@@ -1389,11 +1357,26 @@ vega20_set_uclk_to_highest_dpm_level(struct smu_context *smu,
+ return ret;
+ }
+
+-static int vega20_display_config_changed(struct smu_context *smu)
++static int vega20_pre_display_config_changed(struct smu_context *smu)
+ {
+ int ret = 0;
+ struct vega20_dpm_table *dpm_table = smu->smu_dpm.dpm_context;
+
++ if (!smu->smu_dpm.dpm_context)
++ return -EINVAL;
++
++ smu_send_smc_msg_with_param(smu, SMU_MSG_NumOfDisplays, 0);
++ ret = vega20_set_uclk_to_highest_dpm_level(smu,
++ &dpm_table->mem_table);
++ if (ret)
++ pr_err("Failed to set uclk to highest dpm level");
++ return ret;
++}
++
++static int vega20_display_config_changed(struct smu_context *smu)
++{
++ int ret = 0;
++
+ if (!smu->funcs)
+ return -EINVAL;
+
+@@ -1402,14 +1385,6 @@ static int vega20_display_config_changed(struct smu_context *smu)
+ !smu->smu_table.tables[TABLE_WATERMARKS].cpu_addr)
+ return -EINVAL;
+
+- smu_send_smc_msg_with_param(smu, SMU_MSG_NumOfDisplays, 0);
+- ret = vega20_set_uclk_to_highest_dpm_level(smu,
+- &dpm_table->mem_table);
+- if (ret) {
+- pr_err("Failed to set uclk to highest dpm level");
+- return ret;
+- }
+-
+ if ((smu->watermarks_bitmap & WATERMARKS_EXIST) &&
+ !(smu->watermarks_bitmap & WATERMARKS_LOADED)) {
+ ret = smu->funcs->write_watermarks_table(smu);
+@@ -1672,85 +1647,42 @@ static uint32_t vega20_find_highest_dpm_level(struct vega20_single_dpm_table *ta
+ return i;
+ }
+
+-static int vega20_force_dpm_highest(struct smu_context *smu)
++static int vega20_force_dpm_limit_value(struct smu_context *smu, bool highest)
+ {
+ uint32_t soft_level;
+ int ret = 0;
+- struct vega20_dpm_table *dpm_table = (struct vega20_dpm_table *)smu->smu_dpm.dpm_context;
++ struct vega20_dpm_table *dpm_table =
++ (struct vega20_dpm_table *)smu->smu_dpm.dpm_context;
+
+- soft_level = vega20_find_highest_dpm_level(&(dpm_table->gfx_table));
++ if (highest)
++ soft_level = vega20_find_highest_dpm_level(&(dpm_table->gfx_table));
++ else
++ soft_level = vega20_find_lowest_dpm_level(&(dpm_table->gfx_table));
+
+ dpm_table->gfx_table.dpm_state.soft_min_level =
+ dpm_table->gfx_table.dpm_state.soft_max_level =
+ dpm_table->gfx_table.dpm_levels[soft_level].value;
+
+- soft_level = vega20_find_highest_dpm_level(&(dpm_table->mem_table));
++ if (highest)
++ soft_level = vega20_find_highest_dpm_level(&(dpm_table->mem_table));
++ else
++ soft_level = vega20_find_lowest_dpm_level(&(dpm_table->mem_table));
+
+ dpm_table->mem_table.dpm_state.soft_min_level =
+ dpm_table->mem_table.dpm_state.soft_max_level =
+ dpm_table->mem_table.dpm_levels[soft_level].value;
+
+- ret = vega20_upload_dpm_min_level(smu);
++ ret = vega20_upload_dpm_level(smu, false);
+ if (ret) {
+- pr_err("Failed to upload boot level to highest!");
++ pr_err("Failed to upload boot level to %s!\n",
++ highest ? "highest" : "lowest");
+ return ret;
+ }
+
+- ret = vega20_upload_dpm_max_level(smu);
++ ret = vega20_upload_dpm_level(smu, true);
+ if (ret) {
+- pr_err("Failed to upload dpm max level to highest!");
+- return ret;
+- }
+-
+- return ret;
+-}
+-
+-static int vega20_force_dpm_lowest(struct smu_context *smu)
+-{
+- uint32_t soft_level;
+- int ret = 0;
+- struct vega20_dpm_table *dpm_table = (struct vega20_dpm_table *)smu->smu_dpm.dpm_context;
+-
+- soft_level = vega20_find_lowest_dpm_level(&(dpm_table->gfx_table));
+-
+- dpm_table->gfx_table.dpm_state.soft_min_level =
+- dpm_table->gfx_table.dpm_state.soft_max_level =
+- dpm_table->gfx_table.dpm_levels[soft_level].value;
+-
+- soft_level = vega20_find_lowest_dpm_level(&(dpm_table->mem_table));
+-
+- dpm_table->mem_table.dpm_state.soft_min_level =
+- dpm_table->mem_table.dpm_state.soft_max_level =
+- dpm_table->mem_table.dpm_levels[soft_level].value;
+-
+- ret = vega20_upload_dpm_min_level(smu);
+- if (ret) {
+- pr_err("Failed to upload boot level to lowest!");
+- return ret;
+- }
+-
+- ret = vega20_upload_dpm_max_level(smu);
+- if (ret) {
+- pr_err("Failed to upload dpm max level to lowest!");
+- return ret;
+- }
+-
+- return ret;
+-}
+-
+-static int vega20_unforce_dpm_levels(struct smu_context *smu)
+-{
+- int ret = 0;
+-
+- ret = vega20_upload_dpm_min_level(smu);
+- if (ret) {
+- pr_err("Failed to upload DPM Bootup Levels!");
+- return ret;
+- }
+-
+- ret = vega20_upload_dpm_max_level(smu);
+- if (ret) {
+- pr_err("Failed to upload DPM Max Levels!");
++ pr_err("Failed to upload dpm max level to %s!\n!",
++ highest ? "highest" : "lowest");
+ return ret;
+ }
+
+@@ -1771,78 +1703,6 @@ static enum amd_dpm_forced_level vega20_get_performance_level(struct smu_context
+ return smu_dpm_ctx->dpm_level;
+ }
+
+-static int vega20_adjust_power_state_dynamic(struct smu_context *smu,
+- enum amd_dpm_forced_level level)
+-{
+- int ret = 0;
+- int index = 0;
+- uint32_t sclk_mask, mclk_mask, soc_mask;
+- long workload;
+- struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm);
+-
+- ret = vega20_display_config_changed(smu);
+- if (ret) {
+- pr_err("Failed to change display config!");
+- return ret;
+- }
+- ret = vega20_apply_clocks_adjust_rules(smu);
+- if (ret) {
+- pr_err("Failed to apply clocks adjust rules!");
+- return ret;
+- }
+- ret = vega20_notify_smc_dispaly_config(smu);
+- if (ret) {
+- pr_err("Failed to notify smc display config!");
+- return ret;
+- }
+-
+- switch (level) {
+- case AMD_DPM_FORCED_LEVEL_HIGH:
+- ret = vega20_force_dpm_highest(smu);
+- break;
+- case AMD_DPM_FORCED_LEVEL_LOW:
+- ret = vega20_force_dpm_lowest(smu);
+- break;
+-
+- case AMD_DPM_FORCED_LEVEL_AUTO:
+- ret = vega20_unforce_dpm_levels(smu);
+- break;
+-
+- case AMD_DPM_FORCED_LEVEL_PROFILE_STANDARD:
+- case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_SCLK:
+- case AMD_DPM_FORCED_LEVEL_PROFILE_MIN_MCLK:
+- case AMD_DPM_FORCED_LEVEL_PROFILE_PEAK:
+- ret = vega20_get_profiling_clk_mask(smu, level,
+- &sclk_mask,
+- &mclk_mask,
+- &soc_mask);
+- if (ret)
+- return ret;
+- vega20_force_clk_levels(smu, PP_SCLK, 1 << sclk_mask);
+- vega20_force_clk_levels(smu, PP_MCLK, 1 << mclk_mask);
+- break;
+-
+- case AMD_DPM_FORCED_LEVEL_MANUAL:
+- case AMD_DPM_FORCED_LEVEL_PROFILE_EXIT:
+- default:
+- break;
+- }
+-
+- if (!ret)
+- smu_dpm_ctx->dpm_level = level;
+-
+- if (smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL) {
+- index = fls(smu->workload_mask);
+- index = index > 0 && index <= WORKLOAD_POLICY_MAX ? index - 1 : 0;
+- workload = smu->workload_setting[index];
+-
+- if (smu->power_profile_mode != workload)
+- smu->funcs->set_power_profile_mode(smu, &workload, 0);
+- }
+-
+- return ret;
+-}
+-
+ static int
+ vega20_force_performance_level(struct smu_context *smu, enum amd_dpm_forced_level level)
+ {
+@@ -1861,7 +1721,8 @@ vega20_force_performance_level(struct smu_context *smu, enum amd_dpm_forced_leve
+ mutex_lock(&smu->mutex);
+
+ smu->adev->ip_blocks[i].version->funcs->enable_umd_pstate(smu, &level);
+- ret = vega20_adjust_power_state_dynamic(smu, level);
++ ret = smu_handle_task(smu, level,
++ AMD_PP_TASK_READJUST_POWER_STATE);
+
+ mutex_unlock(&smu->mutex);
+
+@@ -2009,7 +1870,8 @@ static int vega20_set_od_percentage(struct smu_context *smu,
+ single_dpm_table->dpm_levels[0].value = smu->smu_table.boot_values.gfxclk / 100;
+ }
+
+- ret = vega20_adjust_power_state_dynamic(smu, smu_dpm->dpm_level);
++ ret = smu_handle_task(smu, smu_dpm->dpm_level,
++ AMD_PP_TASK_READJUST_POWER_STATE);
+
+ set_od_failed:
+ mutex_unlock(&(smu->mutex));
+@@ -2239,7 +2101,8 @@ static int vega20_odn_edit_dpm_table(struct smu_context *smu,
+
+ if (type == PP_OD_COMMIT_DPM_TABLE) {
+ mutex_lock(&(smu->mutex));
+- ret = vega20_adjust_power_state_dynamic(smu, smu_dpm->dpm_level);
++ ret = smu_handle_task(smu, smu_dpm->dpm_level,
++ AMD_PP_TASK_READJUST_POWER_STATE);
+ mutex_unlock(&(smu->mutex));
+ }
+
+@@ -2268,6 +2131,13 @@ static const struct pptable_funcs vega20_ppt_funcs = {
+ .update_specified_od8_value = vega20_update_specified_od8_value,
+ .set_od_percentage = vega20_set_od_percentage,
+ .od_edit_dpm_table = vega20_odn_edit_dpm_table,
++ .pre_display_config_changed = vega20_pre_display_config_changed,
++ .display_config_changed = vega20_display_config_changed,
++ .apply_clocks_adjust_rules = vega20_apply_clocks_adjust_rules,
++ .notify_smc_dispaly_config = vega20_notify_smc_dispaly_config,
++ .force_dpm_limit_value = vega20_force_dpm_limit_value,
++ .upload_dpm_level = vega20_upload_dpm_level,
++ .get_profiling_clk_mask = vega20_get_profiling_clk_mask,
+ };
+
+ void vega20_set_ppt_funcs(struct smu_context *smu)
+--
+2.17.1
+