aboutsummaryrefslogtreecommitdiffstats
path: root/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3599-drm-amd-pp-Remove-dead-error-checking-code-on-Vega10.patch
diff options
context:
space:
mode:
Diffstat (limited to 'meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3599-drm-amd-pp-Remove-dead-error-checking-code-on-Vega10.patch')
-rw-r--r--meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3599-drm-amd-pp-Remove-dead-error-checking-code-on-Vega10.patch464
1 files changed, 464 insertions, 0 deletions
diff --git a/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3599-drm-amd-pp-Remove-dead-error-checking-code-on-Vega10.patch b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3599-drm-amd-pp-Remove-dead-error-checking-code-on-Vega10.patch
new file mode 100644
index 00000000..49c58398
--- /dev/null
+++ b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71/3599-drm-amd-pp-Remove-dead-error-checking-code-on-Vega10.patch
@@ -0,0 +1,464 @@
+From 063c41dc5141db6b68b371b00de5b71d51cf65ef Mon Sep 17 00:00:00 2001
+From: Rex Zhu <Rex.Zhu@amd.com>
+Date: Thu, 22 Feb 2018 17:20:53 +0800
+Subject: [PATCH 3599/4131] drm/amd/pp: Remove dead error checking code on
+ Vega10
+
+when smu failed, print out the error info immediately
+for debug. smum_send_msg_to_smu always return true,
+so no need to check return value.
+
+Reviewed-by: Alex Deucher <alexander.deucher@amd.com>
+Signed-off-by: Rex Zhu <Rex.Zhu@amd.com>
+Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
+---
+ drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c | 122 ++++++---------------
+ .../gpu/drm/amd/powerplay/hwmgr/vega10_powertune.c | 22 ++--
+ .../gpu/drm/amd/powerplay/hwmgr/vega10_thermal.c | 10 +-
+ .../gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c | 55 ++++------
+ 4 files changed, 67 insertions(+), 142 deletions(-)
+
+diff --git a/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c b/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c
+index 1f12114..8a39ebd 100644
+--- a/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c
++++ b/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c
+@@ -2868,11 +2868,8 @@ static int vega10_enable_dpm_tasks(struct pp_hwmgr *hwmgr)
+ (struct vega10_hwmgr *)(hwmgr->backend);
+ int tmp_result, result = 0;
+
+- tmp_result = smum_send_msg_to_smc_with_parameter(hwmgr,
+- PPSMC_MSG_ConfigureTelemetry, data->config_telemetry);
+- PP_ASSERT_WITH_CODE(!tmp_result,
+- "Failed to configure telemetry!",
+- return tmp_result);
++ smum_send_msg_to_smc_with_parameter(hwmgr,
++ PPSMC_MSG_ConfigureTelemetry, data->config_telemetry);
+
+ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_NumOfDisplays, 0);
+@@ -2883,13 +2880,9 @@ static int vega10_enable_dpm_tasks(struct pp_hwmgr *hwmgr)
+ return 0);
+
+ if ((hwmgr->smu_version == 0x001c2c00) ||
+- (hwmgr->smu_version == 0x001c2d00)) {
+- tmp_result = smum_send_msg_to_smc_with_parameter(hwmgr,
++ (hwmgr->smu_version == 0x001c2d00))
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_UpdatePkgPwrPidAlpha, 1);
+- PP_ASSERT_WITH_CODE(!tmp_result,
+- "Failed to set package power PID!",
+- return tmp_result);
+- }
+
+ tmp_result = vega10_construct_voltage_tables(hwmgr);
+ PP_ASSERT_WITH_CODE(!tmp_result,
+@@ -3641,12 +3634,9 @@ static int vega10_upload_dpm_bootup_level(struct pp_hwmgr *hwmgr)
+ if (!data->registry_data.sclk_dpm_key_disabled) {
+ if (data->smc_state_table.gfx_boot_level !=
+ data->dpm_table.gfx_table.dpm_state.soft_min_level) {
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc_with_parameter(
+- hwmgr,
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetSoftMinGfxclkByIndex,
+- data->smc_state_table.gfx_boot_level),
+- "Failed to set soft min sclk index!",
+- return -EINVAL);
++ data->smc_state_table.gfx_boot_level);
+ data->dpm_table.gfx_table.dpm_state.soft_min_level =
+ data->smc_state_table.gfx_boot_level;
+ }
+@@ -3657,19 +3647,13 @@ static int vega10_upload_dpm_bootup_level(struct pp_hwmgr *hwmgr)
+ data->dpm_table.mem_table.dpm_state.soft_min_level) {
+ if (data->smc_state_table.mem_boot_level == NUM_UCLK_DPM_LEVELS - 1) {
+ socclk_idx = vega10_get_soc_index_for_max_uclk(hwmgr);
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc_with_parameter(
+- hwmgr,
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetSoftMinSocclkByIndex,
+- socclk_idx),
+- "Failed to set soft min uclk index!",
+- return -EINVAL);
++ socclk_idx);
+ } else {
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc_with_parameter(
+- hwmgr,
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetSoftMinUclkByIndex,
+- data->smc_state_table.mem_boot_level),
+- "Failed to set soft min uclk index!",
+- return -EINVAL);
++ data->smc_state_table.mem_boot_level);
+ }
+ data->dpm_table.mem_table.dpm_state.soft_min_level =
+ data->smc_state_table.mem_boot_level;
+@@ -3688,13 +3672,10 @@ static int vega10_upload_dpm_max_level(struct pp_hwmgr *hwmgr)
+
+ if (!data->registry_data.sclk_dpm_key_disabled) {
+ if (data->smc_state_table.gfx_max_level !=
+- data->dpm_table.gfx_table.dpm_state.soft_max_level) {
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc_with_parameter(
+- hwmgr,
++ data->dpm_table.gfx_table.dpm_state.soft_max_level) {
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetSoftMaxGfxclkByIndex,
+- data->smc_state_table.gfx_max_level),
+- "Failed to set soft max sclk index!",
+- return -EINVAL);
++ data->smc_state_table.gfx_max_level);
+ data->dpm_table.gfx_table.dpm_state.soft_max_level =
+ data->smc_state_table.gfx_max_level;
+ }
+@@ -3702,13 +3683,10 @@ static int vega10_upload_dpm_max_level(struct pp_hwmgr *hwmgr)
+
+ if (!data->registry_data.mclk_dpm_key_disabled) {
+ if (data->smc_state_table.mem_max_level !=
+- data->dpm_table.mem_table.dpm_state.soft_max_level) {
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc_with_parameter(
+- hwmgr,
+- PPSMC_MSG_SetSoftMaxUclkByIndex,
+- data->smc_state_table.mem_max_level),
+- "Failed to set soft max mclk index!",
+- return -EINVAL);
++ data->dpm_table.mem_table.dpm_state.soft_max_level) {
++ smum_send_msg_to_smc_with_parameter(hwmgr,
++ PPSMC_MSG_SetSoftMaxUclkByIndex,
++ data->smc_state_table.mem_max_level);
+ data->dpm_table.mem_table.dpm_state.soft_max_level =
+ data->smc_state_table.mem_max_level;
+ }
+@@ -3778,7 +3756,6 @@ static int vega10_update_sclk_threshold(struct pp_hwmgr *hwmgr)
+ {
+ struct vega10_hwmgr *data =
+ (struct vega10_hwmgr *)(hwmgr->backend);
+- int result = 0;
+ uint32_t low_sclk_interrupt_threshold = 0;
+
+ if (PP_CAP(PHM_PlatformCaps_SclkThrottleLowNotification) &&
+@@ -3790,12 +3767,12 @@ static int vega10_update_sclk_threshold(struct pp_hwmgr *hwmgr)
+ cpu_to_le32(low_sclk_interrupt_threshold);
+
+ /* This message will also enable SmcToHost Interrupt */
+- result = smum_send_msg_to_smc_with_parameter(hwmgr,
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetLowGfxclkInterruptThreshold,
+ (uint32_t)low_sclk_interrupt_threshold);
+ }
+
+- return result;
++ return 0;
+ }
+
+ static int vega10_set_power_state_tasks(struct pp_hwmgr *hwmgr,
+@@ -3886,11 +3863,7 @@ static int vega10_get_gpu_power(struct pp_hwmgr *hwmgr,
+ {
+ uint32_t value;
+
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc(hwmgr,
+- PPSMC_MSG_GetCurrPkgPwr),
+- "Failed to get current package power!",
+- return -EINVAL);
+-
++ smum_send_msg_to_smc(hwmgr, PPSMC_MSG_GetCurrPkgPwr);
+ vega10_read_arg_from_smc(hwmgr, &value);
+
+ /* power value is an integer */
+@@ -3973,10 +3946,10 @@ static int vega10_read_sensor(struct pp_hwmgr *hwmgr, int idx,
+ return ret;
+ }
+
+-static int vega10_notify_smc_display_change(struct pp_hwmgr *hwmgr,
++static void vega10_notify_smc_display_change(struct pp_hwmgr *hwmgr,
+ bool has_disp)
+ {
+- return smum_send_msg_to_smc_with_parameter(hwmgr,
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetUclkFastSwitch,
+ has_disp ? 0 : 1);
+ }
+@@ -4011,7 +3984,7 @@ int vega10_display_clock_voltage_request(struct pp_hwmgr *hwmgr,
+
+ if (!result) {
+ clk_request = (clk_freq << 16) | clk_select;
+- result = smum_send_msg_to_smc_with_parameter(hwmgr,
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_RequestDisplayClockByFreq,
+ clk_request);
+ }
+@@ -4080,10 +4053,9 @@ static int vega10_notify_smc_display_config_after_ps_adjustment(
+ clock_req.clock_type = amd_pp_dcef_clock;
+ clock_req.clock_freq_in_khz = dpm_table->dpm_levels[i].value;
+ if (!vega10_display_clock_voltage_request(hwmgr, &clock_req)) {
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc_with_parameter(
++ smum_send_msg_to_smc_with_parameter(
+ hwmgr, PPSMC_MSG_SetMinDeepSleepDcefclk,
+- min_clocks.dcefClockInSR /100),
+- "Attempt to set divider for DCEFCLK Failed!",);
++ min_clocks.dcefClockInSR / 100);
+ } else {
+ pr_info("Attempt to set Hard Min for DCEFCLK Failed!");
+ }
+@@ -4563,14 +4535,8 @@ static int vega10_print_clock_levels(struct pp_hwmgr *hwmgr,
+ if (data->registry_data.sclk_dpm_key_disabled)
+ break;
+
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc(hwmgr,
+- PPSMC_MSG_GetCurrentGfxclkIndex),
+- "Attempt to get current sclk index Failed!",
+- return -1);
+- PP_ASSERT_WITH_CODE(!vega10_read_arg_from_smc(hwmgr,
+- &now),
+- "Attempt to read sclk index Failed!",
+- return -1);
++ smum_send_msg_to_smc(hwmgr, PPSMC_MSG_GetCurrentGfxclkIndex);
++ vega10_read_arg_from_smc(hwmgr, &now);
+
+ for (i = 0; i < sclk_table->count; i++)
+ size += sprintf(buf + size, "%d: %uMhz %s\n",
+@@ -4581,14 +4547,8 @@ static int vega10_print_clock_levels(struct pp_hwmgr *hwmgr,
+ if (data->registry_data.mclk_dpm_key_disabled)
+ break;
+
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc(hwmgr,
+- PPSMC_MSG_GetCurrentUclkIndex),
+- "Attempt to get current mclk index Failed!",
+- return -1);
+- PP_ASSERT_WITH_CODE(!vega10_read_arg_from_smc(hwmgr,
+- &now),
+- "Attempt to read mclk index Failed!",
+- return -1);
++ smum_send_msg_to_smc(hwmgr, PPSMC_MSG_GetCurrentUclkIndex);
++ vega10_read_arg_from_smc(hwmgr, &now);
+
+ for (i = 0; i < mclk_table->count; i++)
+ size += sprintf(buf + size, "%d: %uMhz %s\n",
+@@ -4596,14 +4556,8 @@ static int vega10_print_clock_levels(struct pp_hwmgr *hwmgr,
+ (i == now) ? "*" : "");
+ break;
+ case PP_PCIE:
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc(hwmgr,
+- PPSMC_MSG_GetCurrentLinkIndex),
+- "Attempt to get current mclk index Failed!",
+- return -1);
+- PP_ASSERT_WITH_CODE(!vega10_read_arg_from_smc(hwmgr,
+- &now),
+- "Attempt to read mclk index Failed!",
+- return -1);
++ smum_send_msg_to_smc(hwmgr, PPSMC_MSG_GetCurrentLinkIndex);
++ vega10_read_arg_from_smc(hwmgr, &now);
+
+ for (i = 0; i < pcie_table->count; i++)
+ size += sprintf(buf + size, "%d: %s %s\n", i,
+@@ -4835,24 +4789,18 @@ static int vega10_set_power_profile_state(struct pp_hwmgr *hwmgr,
+
+ if (sclk_idx != ~0) {
+ if (!data->registry_data.sclk_dpm_key_disabled)
+- PP_ASSERT_WITH_CODE(
+- !smum_send_msg_to_smc_with_parameter(
++ smum_send_msg_to_smc_with_parameter(
+ hwmgr,
+ PPSMC_MSG_SetSoftMinGfxclkByIndex,
+- sclk_idx),
+- "Failed to set soft min sclk index!",
+- return -EINVAL);
++ sclk_idx);
+ }
+
+ if (mclk_idx != ~0) {
+ if (!data->registry_data.mclk_dpm_key_disabled)
+- PP_ASSERT_WITH_CODE(
+- !smum_send_msg_to_smc_with_parameter(
++ smum_send_msg_to_smc_with_parameter(
+ hwmgr,
+ PPSMC_MSG_SetSoftMinUclkByIndex,
+- mclk_idx),
+- "Failed to set soft min mclk index!",
+- return -EINVAL);
++ mclk_idx);
+ }
+
+ return 0;
+diff --git a/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_powertune.c b/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_powertune.c
+index 981c9e5..f5ed171 100644
+--- a/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_powertune.c
++++ b/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_powertune.c
+@@ -850,7 +850,6 @@ static int vega10_program_gc_didt_config_registers(struct pp_hwmgr *hwmgr, const
+ static void vega10_didt_set_mask(struct pp_hwmgr *hwmgr, const bool enable)
+ {
+ uint32_t data;
+- int result;
+ uint32_t en = (enable ? 1 : 0);
+ uint32_t didt_block_info = SQ_IR_MASK | TCP_IR_MASK | TD_PCC_MASK;
+
+@@ -924,11 +923,10 @@ static void vega10_didt_set_mask(struct pp_hwmgr *hwmgr, const bool enable)
+ }
+ }
+
+- if (enable) {
+- /* For Vega10, SMC does not support any mask yet. */
+- result = smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_ConfigureGfxDidt, didt_block_info);
+- PP_ASSERT((0 == result), "[EnableDiDtConfig] SMC Configure Gfx Didt Failed!");
+- }
++ /* For Vega10, SMC does not support any mask yet. */
++ if (enable)
++ smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_ConfigureGfxDidt, didt_block_info);
++
+ }
+
+ static int vega10_enable_cac_driving_se_didt_config(struct pp_hwmgr *hwmgr)
+@@ -1344,7 +1342,7 @@ int vega10_set_power_limit(struct pp_hwmgr *hwmgr, uint32_t n)
+ (struct vega10_hwmgr *)(hwmgr->backend);
+
+ if (data->registry_data.enable_pkg_pwr_tracking_feature)
+- return smum_send_msg_to_smc_with_parameter(hwmgr,
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetPptLimit, n);
+
+ return 0;
+@@ -1406,24 +1404,24 @@ int vega10_disable_power_containment(struct pp_hwmgr *hwmgr)
+ return 0;
+ }
+
+-static int vega10_set_overdrive_target_percentage(struct pp_hwmgr *hwmgr,
++static void vega10_set_overdrive_target_percentage(struct pp_hwmgr *hwmgr,
+ uint32_t adjust_percent)
+ {
+- return smum_send_msg_to_smc_with_parameter(hwmgr,
++ smum_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_OverDriveSetPercentage, adjust_percent);
+ }
+
+ int vega10_power_control_set_level(struct pp_hwmgr *hwmgr)
+ {
+- int adjust_percent, result = 0;
++ int adjust_percent;
+
+ if (PP_CAP(PHM_PlatformCaps_PowerContainment)) {
+ adjust_percent =
+ hwmgr->platform_descriptor.TDPAdjustmentPolarity ?
+ hwmgr->platform_descriptor.TDPAdjustment :
+ (-1 * hwmgr->platform_descriptor.TDPAdjustment);
+- result = vega10_set_overdrive_target_percentage(hwmgr,
++ vega10_set_overdrive_target_percentage(hwmgr,
+ (uint32_t)adjust_percent);
+ }
+- return result;
++ return 0;
+ }
+diff --git a/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_thermal.c b/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_thermal.c
+index eb6e965..fc2325e 100644
+--- a/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_thermal.c
++++ b/drivers/gpu/drm/amd/powerplay/hwmgr/vega10_thermal.c
+@@ -31,14 +31,8 @@
+
+ static int vega10_get_current_rpm(struct pp_hwmgr *hwmgr, uint32_t *current_rpm)
+ {
+- PP_ASSERT_WITH_CODE(!smum_send_msg_to_smc(hwmgr,
+- PPSMC_MSG_GetCurrentRpm),
+- "Attempt to get current RPM from SMC Failed!",
+- return -1);
+- PP_ASSERT_WITH_CODE(!vega10_read_arg_from_smc(hwmgr,
+- current_rpm),
+- "Attempt to read current RPM from SMC Failed!",
+- return -1);
++ smum_send_msg_to_smc(hwmgr, PPSMC_MSG_GetCurrentRpm);
++ vega10_read_arg_from_smc(hwmgr, current_rpm);
+ return 0;
+ }
+
+diff --git a/drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c b/drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c
+index b4c487f..dd842ae 100644
+--- a/drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c
++++ b/drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c
+@@ -228,20 +228,15 @@ int vega10_copy_table_from_smc(struct pp_hwmgr *hwmgr,
+ "Invalid SMU Table version!", return -EINVAL);
+ PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0,
+ "Invalid SMU Table Length!", return -EINVAL);
+- PP_ASSERT_WITH_CODE(vega10_send_msg_to_smc_with_parameter(hwmgr,
++ vega10_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetDriverDramAddrHigh,
+- priv->smu_tables.entry[table_id].table_addr_high) == 0,
+- "[CopyTableFromSMC] Attempt to Set Dram Addr High Failed!", return -EINVAL);
+- PP_ASSERT_WITH_CODE(vega10_send_msg_to_smc_with_parameter(hwmgr,
++ priv->smu_tables.entry[table_id].table_addr_high);
++ vega10_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetDriverDramAddrLow,
+- priv->smu_tables.entry[table_id].table_addr_low) == 0,
+- "[CopyTableFromSMC] Attempt to Set Dram Addr Low Failed!",
+- return -EINVAL);
+- PP_ASSERT_WITH_CODE(vega10_send_msg_to_smc_with_parameter(hwmgr,
++ priv->smu_tables.entry[table_id].table_addr_low);
++ vega10_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_TransferTableSmu2Dram,
+- priv->smu_tables.entry[table_id].table_id) == 0,
+- "[CopyTableFromSMC] Attempt to Transfer Table From SMU Failed!",
+- return -EINVAL);
++ priv->smu_tables.entry[table_id].table_id);
+
+ memcpy(table, priv->smu_tables.entry[table_id].table,
+ priv->smu_tables.entry[table_id].size);
+@@ -270,21 +265,15 @@ int vega10_copy_table_to_smc(struct pp_hwmgr *hwmgr,
+ memcpy(priv->smu_tables.entry[table_id].table, table,
+ priv->smu_tables.entry[table_id].size);
+
+- PP_ASSERT_WITH_CODE(vega10_send_msg_to_smc_with_parameter(hwmgr,
++ vega10_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetDriverDramAddrHigh,
+- priv->smu_tables.entry[table_id].table_addr_high) == 0,
+- "[CopyTableToSMC] Attempt to Set Dram Addr High Failed!",
+- return -EINVAL;);
+- PP_ASSERT_WITH_CODE(vega10_send_msg_to_smc_with_parameter(hwmgr,
++ priv->smu_tables.entry[table_id].table_addr_high);
++ vega10_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetDriverDramAddrLow,
+- priv->smu_tables.entry[table_id].table_addr_low) == 0,
+- "[CopyTableToSMC] Attempt to Set Dram Addr Low Failed!",
+- return -EINVAL);
+- PP_ASSERT_WITH_CODE(vega10_send_msg_to_smc_with_parameter(hwmgr,
++ priv->smu_tables.entry[table_id].table_addr_low);
++ vega10_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_TransferTableDram2Smu,
+- priv->smu_tables.entry[table_id].table_id) == 0,
+- "[CopyTableToSMC] Attempt to Transfer Table To SMU Failed!",
+- return -EINVAL);
++ priv->smu_tables.entry[table_id].table_id);
+
+ return 0;
+ }
+@@ -323,13 +312,9 @@ int vega10_get_smc_features(struct pp_hwmgr *hwmgr,
+ if (features_enabled == NULL)
+ return -EINVAL;
+
+- if (!vega10_send_msg_to_smc(hwmgr,
+- PPSMC_MSG_GetEnabledSmuFeatures)) {
+- vega10_read_arg_from_smc(hwmgr, features_enabled);
+- return 0;
+- }
+-
+- return -EINVAL;
++ vega10_send_msg_to_smc(hwmgr, PPSMC_MSG_GetEnabledSmuFeatures);
++ vega10_read_arg_from_smc(hwmgr, features_enabled);
++ return 0;
+ }
+
+ int vega10_set_tools_address(struct pp_hwmgr *hwmgr)
+@@ -339,12 +324,12 @@ int vega10_set_tools_address(struct pp_hwmgr *hwmgr)
+
+ if (priv->smu_tables.entry[TOOLSTABLE].table_addr_high ||
+ priv->smu_tables.entry[TOOLSTABLE].table_addr_low) {
+- if (!vega10_send_msg_to_smc_with_parameter(hwmgr,
++ vega10_send_msg_to_smc_with_parameter(hwmgr,
+ PPSMC_MSG_SetToolsDramAddrHigh,
+- priv->smu_tables.entry[TOOLSTABLE].table_addr_high))
+- vega10_send_msg_to_smc_with_parameter(hwmgr,
+- PPSMC_MSG_SetToolsDramAddrLow,
+- priv->smu_tables.entry[TOOLSTABLE].table_addr_low);
++ priv->smu_tables.entry[TOOLSTABLE].table_addr_high);
++ vega10_send_msg_to_smc_with_parameter(hwmgr,
++ PPSMC_MSG_SetToolsDramAddrLow,
++ priv->smu_tables.entry[TOOLSTABLE].table_addr_low);
+ }
+ return 0;
+ }
+--
+2.7.4
+