diff options
Diffstat (limited to 'common/recipes-kernel/linux/linux-yocto-4.14.71/4790-drm-amd-display-Moving-powerplay-functions-to-a-sepa.patch')
-rw-r--r-- | common/recipes-kernel/linux/linux-yocto-4.14.71/4790-drm-amd-display-Moving-powerplay-functions-to-a-sepa.patch | 1035 |
1 files changed, 1035 insertions, 0 deletions
diff --git a/common/recipes-kernel/linux/linux-yocto-4.14.71/4790-drm-amd-display-Moving-powerplay-functions-to-a-sepa.patch b/common/recipes-kernel/linux/linux-yocto-4.14.71/4790-drm-amd-display-Moving-powerplay-functions-to-a-sepa.patch new file mode 100644 index 00000000..c61fa4ea --- /dev/null +++ b/common/recipes-kernel/linux/linux-yocto-4.14.71/4790-drm-amd-display-Moving-powerplay-functions-to-a-sepa.patch @@ -0,0 +1,1035 @@ +From dd3df5a2d35287242a63c93f88219a29892ac44a Mon Sep 17 00:00:00 2001 +From: Mikita Lipski <mikita.lipski@amd.com> +Date: Fri, 1 Jun 2018 15:02:55 -0400 +Subject: [PATCH 4790/5725] drm/amd/display: Moving powerplay functions to a + separate class + +Moving powerplay functions to a new amdgpu_dm_pp_smu class +and cleaning dm_services class from unused headers. + +Signed-off-by: Mikita Lipski <mikita.lipski@amd.com> +Reviewed-by: Tony Cheng <Tony.Cheng@amd.com> +Acked-by: Harry Wentland <harry.wentland@amd.com> +Signed-off-by: Alex Deucher <alexander.deucher@amd.com> +Signed-off-by: Raveendra Talabattula <raveendra.talabattula@amd.com> +--- + drivers/gpu/drm/amd/display/amdgpu_dm/Makefile | 2 +- + .../drm/amd/display/amdgpu_dm/amdgpu_dm_pp_smu.c | 525 +++++++++++++++++++++ + .../drm/amd/display/amdgpu_dm/amdgpu_dm_services.c | 452 ------------------ + 3 files changed, 526 insertions(+), 453 deletions(-) + create mode 100644 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_pp_smu.c + +diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/Makefile b/drivers/gpu/drm/amd/display/amdgpu_dm/Makefile +index c54685e..98854bc 100644 +--- a/drivers/gpu/drm/amd/display/amdgpu_dm/Makefile ++++ b/drivers/gpu/drm/amd/display/amdgpu_dm/Makefile +@@ -7,7 +7,7 @@ + AMDGPUDM = amdgpu_dm.o amdgpu_dm_irq.o amdgpu_dm_mst_types.o amdgpu_dm_color.o + + ifneq ($(CONFIG_DRM_AMD_DC),) +-AMDGPUDM += amdgpu_dm_services.o amdgpu_dm_helpers.o ++AMDGPUDM += amdgpu_dm_services.o amdgpu_dm_helpers.o amdgpu_dm_pp_smu.o + endif + + ifneq ($(CONFIG_DEBUG_FS),) +diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_pp_smu.c b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_pp_smu.c +new file mode 100644 +index 0000000..ad96e2a +--- /dev/null ++++ b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_pp_smu.c +@@ -0,0 +1,525 @@ ++/* ++ * Copyright 2018 Advanced Micro Devices, Inc. ++ * ++ * Permission is hereby granted, free of charge, to any person obtaining a ++ * copy of this software and associated documentation files (the "Software"), ++ * to deal in the Software without restriction, including without limitation ++ * the rights to use, copy, modify, merge, publish, distribute, sublicense, ++ * and/or sell copies of the Software, and to permit persons to whom the ++ * Software is furnished to do so, subject to the following conditions: ++ * ++ * The above copyright notice and this permission notice shall be included in ++ * all copies or substantial portions of the Software. ++ * ++ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR ++ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, ++ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL ++ * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR ++ * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ++ * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR ++ * OTHER DEALINGS IN THE SOFTWARE. ++ * ++ * Authors: AMD ++ */ ++#include <linux/string.h> ++#include <linux/acpi.h> ++ ++#include <drm/drmP.h> ++#include <drm/drm_crtc_helper.h> ++#include <drm/amdgpu_drm.h> ++#include "dm_services.h" ++#include "amdgpu.h" ++#include "amdgpu_dm.h" ++#include "amdgpu_dm_irq.h" ++#include "amdgpu_pm.h" ++#include "dm_pp_smu.h" ++#include "../../powerplay/inc/hwmgr.h" ++#include "../../powerplay/hwmgr/smu10_hwmgr.h" ++ ++ ++bool dm_pp_apply_display_requirements( ++ const struct dc_context *ctx, ++ const struct dm_pp_display_configuration *pp_display_cfg) ++{ ++ struct amdgpu_device *adev = ctx->driver_context; ++ ++ if (adev->pm.dpm_enabled) { ++ ++ memset(&adev->pm.pm_display_cfg, 0, ++ sizeof(adev->pm.pm_display_cfg)); ++ ++ adev->pm.pm_display_cfg.cpu_cc6_disable = ++ pp_display_cfg->cpu_cc6_disable; ++ ++ adev->pm.pm_display_cfg.cpu_pstate_disable = ++ pp_display_cfg->cpu_pstate_disable; ++ ++ adev->pm.pm_display_cfg.cpu_pstate_separation_time = ++ pp_display_cfg->cpu_pstate_separation_time; ++ ++ adev->pm.pm_display_cfg.nb_pstate_switch_disable = ++ pp_display_cfg->nb_pstate_switch_disable; ++ ++ adev->pm.pm_display_cfg.num_display = ++ pp_display_cfg->display_count; ++ adev->pm.pm_display_cfg.num_path_including_non_display = ++ pp_display_cfg->display_count; ++ ++ adev->pm.pm_display_cfg.min_core_set_clock = ++ pp_display_cfg->min_engine_clock_khz/10; ++ adev->pm.pm_display_cfg.min_core_set_clock_in_sr = ++ pp_display_cfg->min_engine_clock_deep_sleep_khz/10; ++ adev->pm.pm_display_cfg.min_mem_set_clock = ++ pp_display_cfg->min_memory_clock_khz/10; ++ ++ adev->pm.pm_display_cfg.multi_monitor_in_sync = ++ pp_display_cfg->all_displays_in_sync; ++ adev->pm.pm_display_cfg.min_vblank_time = ++ pp_display_cfg->avail_mclk_switch_time_us; ++ ++ adev->pm.pm_display_cfg.display_clk = ++ pp_display_cfg->disp_clk_khz/10; ++ ++ adev->pm.pm_display_cfg.dce_tolerable_mclk_in_active_latency = ++ pp_display_cfg->avail_mclk_switch_time_in_disp_active_us; ++ ++ adev->pm.pm_display_cfg.crtc_index = pp_display_cfg->crtc_index; ++ adev->pm.pm_display_cfg.line_time_in_us = ++ pp_display_cfg->line_time_in_us; ++ ++ adev->pm.pm_display_cfg.vrefresh = pp_display_cfg->disp_configs[0].v_refresh; ++ adev->pm.pm_display_cfg.crossfire_display_index = -1; ++ adev->pm.pm_display_cfg.min_bus_bandwidth = 0; ++ ++ /* TODO: complete implementation of ++ * pp_display_configuration_change(). ++ * Follow example of: ++ * PHM_StoreDALConfigurationData - powerplay\hwmgr\hardwaremanager.c ++ * PP_IRI_DisplayConfigurationChange - powerplay\eventmgr\iri.c */ ++ if (adev->powerplay.pp_funcs->display_configuration_change) ++ adev->powerplay.pp_funcs->display_configuration_change( ++ adev->powerplay.pp_handle, ++ &adev->pm.pm_display_cfg); ++ ++ /* TODO: replace by a separate call to 'apply display cfg'? */ ++ amdgpu_pm_compute_clocks(adev); ++ } ++ ++ return true; ++} ++ ++static void get_default_clock_levels( ++ enum dm_pp_clock_type clk_type, ++ struct dm_pp_clock_levels *clks) ++{ ++ uint32_t disp_clks_in_khz[6] = { ++ 300000, 400000, 496560, 626090, 685720, 757900 }; ++ uint32_t sclks_in_khz[6] = { ++ 300000, 360000, 423530, 514290, 626090, 720000 }; ++ uint32_t mclks_in_khz[2] = { 333000, 800000 }; ++ ++ switch (clk_type) { ++ case DM_PP_CLOCK_TYPE_DISPLAY_CLK: ++ clks->num_levels = 6; ++ memmove(clks->clocks_in_khz, disp_clks_in_khz, ++ sizeof(disp_clks_in_khz)); ++ break; ++ case DM_PP_CLOCK_TYPE_ENGINE_CLK: ++ clks->num_levels = 6; ++ memmove(clks->clocks_in_khz, sclks_in_khz, ++ sizeof(sclks_in_khz)); ++ break; ++ case DM_PP_CLOCK_TYPE_MEMORY_CLK: ++ clks->num_levels = 2; ++ memmove(clks->clocks_in_khz, mclks_in_khz, ++ sizeof(mclks_in_khz)); ++ break; ++ default: ++ clks->num_levels = 0; ++ break; ++ } ++} ++ ++static enum amd_pp_clock_type dc_to_pp_clock_type( ++ enum dm_pp_clock_type dm_pp_clk_type) ++{ ++ enum amd_pp_clock_type amd_pp_clk_type = 0; ++ ++ switch (dm_pp_clk_type) { ++ case DM_PP_CLOCK_TYPE_DISPLAY_CLK: ++ amd_pp_clk_type = amd_pp_disp_clock; ++ break; ++ case DM_PP_CLOCK_TYPE_ENGINE_CLK: ++ amd_pp_clk_type = amd_pp_sys_clock; ++ break; ++ case DM_PP_CLOCK_TYPE_MEMORY_CLK: ++ amd_pp_clk_type = amd_pp_mem_clock; ++ break; ++ case DM_PP_CLOCK_TYPE_DCEFCLK: ++ amd_pp_clk_type = amd_pp_dcef_clock; ++ break; ++ case DM_PP_CLOCK_TYPE_DCFCLK: ++ amd_pp_clk_type = amd_pp_dcf_clock; ++ break; ++ case DM_PP_CLOCK_TYPE_PIXELCLK: ++ amd_pp_clk_type = amd_pp_pixel_clock; ++ break; ++ case DM_PP_CLOCK_TYPE_FCLK: ++ amd_pp_clk_type = amd_pp_f_clock; ++ break; ++ case DM_PP_CLOCK_TYPE_DISPLAYPHYCLK: ++ amd_pp_clk_type = amd_pp_dpp_clock; ++ break; ++ default: ++ DRM_ERROR("DM_PPLIB: invalid clock type: %d!\n", ++ dm_pp_clk_type); ++ break; ++ } ++ ++ return amd_pp_clk_type; ++} ++ ++static void pp_to_dc_clock_levels( ++ const struct amd_pp_clocks *pp_clks, ++ struct dm_pp_clock_levels *dc_clks, ++ enum dm_pp_clock_type dc_clk_type) ++{ ++ uint32_t i; ++ ++ if (pp_clks->count > DM_PP_MAX_CLOCK_LEVELS) { ++ DRM_INFO("DM_PPLIB: Warning: %s clock: number of levels %d exceeds maximum of %d!\n", ++ DC_DECODE_PP_CLOCK_TYPE(dc_clk_type), ++ pp_clks->count, ++ DM_PP_MAX_CLOCK_LEVELS); ++ ++ dc_clks->num_levels = DM_PP_MAX_CLOCK_LEVELS; ++ } else ++ dc_clks->num_levels = pp_clks->count; ++ ++ DRM_INFO("DM_PPLIB: values for %s clock\n", ++ DC_DECODE_PP_CLOCK_TYPE(dc_clk_type)); ++ ++ for (i = 0; i < dc_clks->num_levels; i++) { ++ DRM_INFO("DM_PPLIB:\t %d\n", pp_clks->clock[i]); ++ /* translate 10kHz to kHz */ ++ dc_clks->clocks_in_khz[i] = pp_clks->clock[i] * 10; ++ } ++} ++ ++static void pp_to_dc_clock_levels_with_latency( ++ const struct pp_clock_levels_with_latency *pp_clks, ++ struct dm_pp_clock_levels_with_latency *clk_level_info, ++ enum dm_pp_clock_type dc_clk_type) ++{ ++ uint32_t i; ++ ++ if (pp_clks->num_levels > DM_PP_MAX_CLOCK_LEVELS) { ++ DRM_INFO("DM_PPLIB: Warning: %s clock: number of levels %d exceeds maximum of %d!\n", ++ DC_DECODE_PP_CLOCK_TYPE(dc_clk_type), ++ pp_clks->num_levels, ++ DM_PP_MAX_CLOCK_LEVELS); ++ ++ clk_level_info->num_levels = DM_PP_MAX_CLOCK_LEVELS; ++ } else ++ clk_level_info->num_levels = pp_clks->num_levels; ++ ++ DRM_DEBUG("DM_PPLIB: values for %s clock\n", ++ DC_DECODE_PP_CLOCK_TYPE(dc_clk_type)); ++ ++ for (i = 0; i < clk_level_info->num_levels; i++) { ++ DRM_DEBUG("DM_PPLIB:\t %d in 10kHz\n", pp_clks->data[i].clocks_in_khz); ++ /* translate 10kHz to kHz */ ++ clk_level_info->data[i].clocks_in_khz = pp_clks->data[i].clocks_in_khz * 10; ++ clk_level_info->data[i].latency_in_us = pp_clks->data[i].latency_in_us; ++ } ++} ++ ++static void pp_to_dc_clock_levels_with_voltage( ++ const struct pp_clock_levels_with_voltage *pp_clks, ++ struct dm_pp_clock_levels_with_voltage *clk_level_info, ++ enum dm_pp_clock_type dc_clk_type) ++{ ++ uint32_t i; ++ ++ if (pp_clks->num_levels > DM_PP_MAX_CLOCK_LEVELS) { ++ DRM_INFO("DM_PPLIB: Warning: %s clock: number of levels %d exceeds maximum of %d!\n", ++ DC_DECODE_PP_CLOCK_TYPE(dc_clk_type), ++ pp_clks->num_levels, ++ DM_PP_MAX_CLOCK_LEVELS); ++ ++ clk_level_info->num_levels = DM_PP_MAX_CLOCK_LEVELS; ++ } else ++ clk_level_info->num_levels = pp_clks->num_levels; ++ ++ DRM_INFO("DM_PPLIB: values for %s clock\n", ++ DC_DECODE_PP_CLOCK_TYPE(dc_clk_type)); ++ ++ for (i = 0; i < clk_level_info->num_levels; i++) { ++ DRM_INFO("DM_PPLIB:\t %d in 10kHz\n", pp_clks->data[i].clocks_in_khz); ++ /* translate 10kHz to kHz */ ++ clk_level_info->data[i].clocks_in_khz = pp_clks->data[i].clocks_in_khz * 10; ++ clk_level_info->data[i].voltage_in_mv = pp_clks->data[i].voltage_in_mv; ++ } ++} ++ ++bool dm_pp_get_clock_levels_by_type( ++ const struct dc_context *ctx, ++ enum dm_pp_clock_type clk_type, ++ struct dm_pp_clock_levels *dc_clks) ++{ ++ struct amdgpu_device *adev = ctx->driver_context; ++ void *pp_handle = adev->powerplay.pp_handle; ++ struct amd_pp_clocks pp_clks = { 0 }; ++ struct amd_pp_simple_clock_info validation_clks = { 0 }; ++ uint32_t i; ++ ++ if (adev->powerplay.pp_funcs->get_clock_by_type) { ++ if (adev->powerplay.pp_funcs->get_clock_by_type(pp_handle, ++ dc_to_pp_clock_type(clk_type), &pp_clks)) { ++ /* Error in pplib. Provide default values. */ ++ get_default_clock_levels(clk_type, dc_clks); ++ return true; ++ } ++ } ++ ++ pp_to_dc_clock_levels(&pp_clks, dc_clks, clk_type); ++ ++ if (adev->powerplay.pp_funcs->get_display_mode_validation_clocks) { ++ if (adev->powerplay.pp_funcs->get_display_mode_validation_clocks( ++ pp_handle, &validation_clks)) { ++ /* Error in pplib. Provide default values. */ ++ DRM_INFO("DM_PPLIB: Warning: using default validation clocks!\n"); ++ validation_clks.engine_max_clock = 72000; ++ validation_clks.memory_max_clock = 80000; ++ validation_clks.level = 0; ++ } ++ } ++ ++ DRM_INFO("DM_PPLIB: Validation clocks:\n"); ++ DRM_INFO("DM_PPLIB: engine_max_clock: %d\n", ++ validation_clks.engine_max_clock); ++ DRM_INFO("DM_PPLIB: memory_max_clock: %d\n", ++ validation_clks.memory_max_clock); ++ DRM_INFO("DM_PPLIB: level : %d\n", ++ validation_clks.level); ++ ++ /* Translate 10 kHz to kHz. */ ++ validation_clks.engine_max_clock *= 10; ++ validation_clks.memory_max_clock *= 10; ++ ++ /* Determine the highest non-boosted level from the Validation Clocks */ ++ if (clk_type == DM_PP_CLOCK_TYPE_ENGINE_CLK) { ++ for (i = 0; i < dc_clks->num_levels; i++) { ++ if (dc_clks->clocks_in_khz[i] > validation_clks.engine_max_clock) { ++ /* This clock is higher the validation clock. ++ * Than means the previous one is the highest ++ * non-boosted one. */ ++ DRM_INFO("DM_PPLIB: reducing engine clock level from %d to %d\n", ++ dc_clks->num_levels, i); ++ dc_clks->num_levels = i > 0 ? i : 1; ++ break; ++ } ++ } ++ } else if (clk_type == DM_PP_CLOCK_TYPE_MEMORY_CLK) { ++ for (i = 0; i < dc_clks->num_levels; i++) { ++ if (dc_clks->clocks_in_khz[i] > validation_clks.memory_max_clock) { ++ DRM_INFO("DM_PPLIB: reducing memory clock level from %d to %d\n", ++ dc_clks->num_levels, i); ++ dc_clks->num_levels = i > 0 ? i : 1; ++ break; ++ } ++ } ++ } ++ ++ return true; ++} ++ ++bool dm_pp_get_clock_levels_by_type_with_latency( ++ const struct dc_context *ctx, ++ enum dm_pp_clock_type clk_type, ++ struct dm_pp_clock_levels_with_latency *clk_level_info) ++{ ++ struct amdgpu_device *adev = ctx->driver_context; ++ void *pp_handle = adev->powerplay.pp_handle; ++ struct pp_clock_levels_with_latency pp_clks = { 0 }; ++ const struct amd_pm_funcs *pp_funcs = adev->powerplay.pp_funcs; ++ ++ if (!pp_funcs || !pp_funcs->get_clock_by_type_with_latency) ++ return false; ++ ++ if (pp_funcs->get_clock_by_type_with_latency(pp_handle, ++ dc_to_pp_clock_type(clk_type), ++ &pp_clks)) ++ return false; ++ ++ pp_to_dc_clock_levels_with_latency(&pp_clks, clk_level_info, clk_type); ++ ++ return true; ++} ++ ++bool dm_pp_get_clock_levels_by_type_with_voltage( ++ const struct dc_context *ctx, ++ enum dm_pp_clock_type clk_type, ++ struct dm_pp_clock_levels_with_voltage *clk_level_info) ++{ ++ struct amdgpu_device *adev = ctx->driver_context; ++ void *pp_handle = adev->powerplay.pp_handle; ++ struct pp_clock_levels_with_voltage pp_clk_info = {0}; ++ const struct amd_pm_funcs *pp_funcs = adev->powerplay.pp_funcs; ++ ++ if (pp_funcs->get_clock_by_type_with_voltage(pp_handle, ++ dc_to_pp_clock_type(clk_type), ++ &pp_clk_info)) ++ return false; ++ ++ pp_to_dc_clock_levels_with_voltage(&pp_clk_info, clk_level_info, clk_type); ++ ++ return true; ++} ++ ++bool dm_pp_notify_wm_clock_changes( ++ const struct dc_context *ctx, ++ struct dm_pp_wm_sets_with_clock_ranges *wm_with_clock_ranges) ++{ ++ /* TODO: to be implemented */ ++ return false; ++} ++ ++bool dm_pp_apply_power_level_change_request( ++ const struct dc_context *ctx, ++ struct dm_pp_power_level_change_request *level_change_req) ++{ ++ /* TODO: to be implemented */ ++ return false; ++} ++ ++bool dm_pp_apply_clock_for_voltage_request( ++ const struct dc_context *ctx, ++ struct dm_pp_clock_for_voltage_req *clock_for_voltage_req) ++{ ++ struct amdgpu_device *adev = ctx->driver_context; ++ struct pp_display_clock_request pp_clock_request = {0}; ++ int ret = 0; ++ ++ pp_clock_request.clock_type = dc_to_pp_clock_type(clock_for_voltage_req->clk_type); ++ pp_clock_request.clock_freq_in_khz = clock_for_voltage_req->clocks_in_khz; ++ ++ if (!pp_clock_request.clock_type) ++ return false; ++ ++ if (adev->powerplay.pp_funcs->display_clock_voltage_request) ++ ret = adev->powerplay.pp_funcs->display_clock_voltage_request( ++ adev->powerplay.pp_handle, ++ &pp_clock_request); ++ if (ret) ++ return false; ++ return true; ++} ++ ++bool dm_pp_get_static_clocks( ++ const struct dc_context *ctx, ++ struct dm_pp_static_clock_info *static_clk_info) ++{ ++ struct amdgpu_device *adev = ctx->driver_context; ++ struct amd_pp_clock_info pp_clk_info = {0}; ++ int ret = 0; ++ ++ if (adev->powerplay.pp_funcs->get_current_clocks) ++ ret = adev->powerplay.pp_funcs->get_current_clocks( ++ adev->powerplay.pp_handle, ++ &pp_clk_info); ++ if (ret) ++ return false; ++ ++ static_clk_info->max_clocks_state = pp_clk_info.max_clocks_state; ++ /* translate 10kHz to kHz */ ++ static_clk_info->max_mclk_khz = pp_clk_info.max_memory_clock * 10; ++ static_clk_info->max_sclk_khz = pp_clk_info.max_engine_clock * 10; ++ ++ return true; ++} ++ ++void pp_rv_set_display_requirement(struct pp_smu *pp, ++ struct pp_smu_display_requirement_rv *req) ++{ ++ struct amdgpu_device *adev = pp->ctx->driver_context; ++ struct pp_hwmgr *hwmgr = adev->powerplay.pp_handle; ++ int ret = 0; ++ if (hwmgr->hwmgr_func->set_deep_sleep_dcefclk) ++ ret = hwmgr->hwmgr_func->set_deep_sleep_dcefclk(hwmgr, req->hard_min_dcefclk_khz/10); ++ if (hwmgr->hwmgr_func->set_active_display_count) ++ ret = hwmgr->hwmgr_func->set_active_display_count(hwmgr, req->display_count); ++ ++ //store_cc6 is not yet implemented in SMU level ++} ++ ++void pp_rv_set_wm_ranges(struct pp_smu *pp, ++ struct pp_smu_wm_range_sets *ranges) ++{ ++ struct amdgpu_device *adev = pp->ctx->driver_context; ++ struct pp_hwmgr *hwmgr = adev->powerplay.pp_handle; ++ struct pp_wm_sets_with_clock_ranges_soc15 ranges_soc15 = {0}; ++ int i = 0; ++ ++ if (!hwmgr->hwmgr_func->set_watermarks_for_clocks_ranges || ++ !pp || !ranges) ++ return; ++ ++ //not entirely sure if thats a correct assignment ++ ranges_soc15.num_wm_sets_dmif = ranges->num_reader_wm_sets; ++ ranges_soc15.num_wm_sets_mcif = ranges->num_writer_wm_sets; ++ ++ for (i = 0; i < ranges_soc15.num_wm_sets_dmif; i++) { ++ if (ranges->reader_wm_sets[i].wm_inst > 3) ++ ranges_soc15.wm_sets_dmif[i].wm_set_id = DC_WM_SET_A; ++ else ++ ranges_soc15.wm_sets_dmif[i].wm_set_id = ++ ranges->reader_wm_sets[i].wm_inst; ++ ranges_soc15.wm_sets_dmif[i].wm_max_dcefclk_in_khz = ++ ranges->reader_wm_sets[i].max_drain_clk_khz; ++ ranges_soc15.wm_sets_dmif[i].wm_min_dcefclk_in_khz = ++ ranges->reader_wm_sets[i].min_drain_clk_khz; ++ ranges_soc15.wm_sets_dmif[i].wm_max_memclk_in_khz = ++ ranges->reader_wm_sets[i].max_fill_clk_khz; ++ ranges_soc15.wm_sets_dmif[i].wm_min_memclk_in_khz = ++ ranges->reader_wm_sets[i].min_fill_clk_khz; ++ } ++ ++ for (i = 0; i < ranges_soc15.num_wm_sets_mcif; i++) { ++ if (ranges->writer_wm_sets[i].wm_inst > 3) ++ ranges_soc15.wm_sets_dmif[i].wm_set_id = DC_WM_SET_A; ++ else ++ ranges_soc15.wm_sets_mcif[i].wm_set_id = ++ ranges->writer_wm_sets[i].wm_inst; ++ ranges_soc15.wm_sets_mcif[i].wm_max_socclk_in_khz = ++ ranges->writer_wm_sets[i].max_fill_clk_khz; ++ ranges_soc15.wm_sets_mcif[i].wm_min_socclk_in_khz = ++ ranges->writer_wm_sets[i].min_fill_clk_khz; ++ ranges_soc15.wm_sets_mcif[i].wm_max_memclk_in_khz = ++ ranges->writer_wm_sets[i].max_fill_clk_khz; ++ ranges_soc15.wm_sets_mcif[i].wm_min_memclk_in_khz = ++ ranges->writer_wm_sets[i].min_fill_clk_khz; ++ } ++ ++ hwmgr->hwmgr_func->set_watermarks_for_clocks_ranges(hwmgr, &ranges_soc15); ++ ++} ++ ++void pp_rv_set_pme_wa_enable(struct pp_smu *pp) ++{ ++ struct amdgpu_device *adev = pp->ctx->driver_context; ++ struct pp_hwmgr *hwmgr = adev->powerplay.pp_handle; ++ ++ if (hwmgr->hwmgr_func->smus_notify_pwe) ++ hwmgr->hwmgr_func->smus_notify_pwe(hwmgr); ++} ++ ++void dm_pp_get_funcs_rv( ++ struct dc_context *ctx, ++ struct pp_smu_funcs_rv *funcs) ++{ ++ funcs->pp_smu.ctx = ctx; ++ funcs->set_display_requirement = pp_rv_set_display_requirement; ++ funcs->set_wm_ranges = pp_rv_set_wm_ranges; ++ funcs->set_pme_wa_enable = pp_rv_set_pme_wa_enable; ++} +diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_services.c b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_services.c +index 4f86f6f..19e7faa 100644 +--- a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_services.c ++++ b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_services.c +@@ -34,9 +34,6 @@ + #include "amdgpu_dm.h" + #include "amdgpu_dm_irq.h" + #include "amdgpu_pm.h" +-#include "dm_pp_smu.h" +-#include "../../powerplay/inc/hwmgr.h" +-#include "../../powerplay/hwmgr/smu10_hwmgr.h" + + + +@@ -76,452 +73,3 @@ bool dm_read_persistent_data(struct dc_context *ctx, + } + + /**** power component interfaces ****/ +- +-bool dm_pp_apply_display_requirements( +- const struct dc_context *ctx, +- const struct dm_pp_display_configuration *pp_display_cfg) +-{ +- struct amdgpu_device *adev = ctx->driver_context; +- +- if (adev->pm.dpm_enabled) { +- +- memset(&adev->pm.pm_display_cfg, 0, +- sizeof(adev->pm.pm_display_cfg)); +- +- adev->pm.pm_display_cfg.cpu_cc6_disable = +- pp_display_cfg->cpu_cc6_disable; +- +- adev->pm.pm_display_cfg.cpu_pstate_disable = +- pp_display_cfg->cpu_pstate_disable; +- +- adev->pm.pm_display_cfg.cpu_pstate_separation_time = +- pp_display_cfg->cpu_pstate_separation_time; +- +- adev->pm.pm_display_cfg.nb_pstate_switch_disable = +- pp_display_cfg->nb_pstate_switch_disable; +- +- adev->pm.pm_display_cfg.num_display = +- pp_display_cfg->display_count; +- adev->pm.pm_display_cfg.num_path_including_non_display = +- pp_display_cfg->display_count; +- +- adev->pm.pm_display_cfg.min_core_set_clock = +- pp_display_cfg->min_engine_clock_khz/10; +- adev->pm.pm_display_cfg.min_core_set_clock_in_sr = +- pp_display_cfg->min_engine_clock_deep_sleep_khz/10; +- adev->pm.pm_display_cfg.min_mem_set_clock = +- pp_display_cfg->min_memory_clock_khz/10; +- +- adev->pm.pm_display_cfg.multi_monitor_in_sync = +- pp_display_cfg->all_displays_in_sync; +- adev->pm.pm_display_cfg.min_vblank_time = +- pp_display_cfg->avail_mclk_switch_time_us; +- +- adev->pm.pm_display_cfg.display_clk = +- pp_display_cfg->disp_clk_khz/10; +- +- adev->pm.pm_display_cfg.dce_tolerable_mclk_in_active_latency = +- pp_display_cfg->avail_mclk_switch_time_in_disp_active_us; +- +- adev->pm.pm_display_cfg.crtc_index = pp_display_cfg->crtc_index; +- adev->pm.pm_display_cfg.line_time_in_us = +- pp_display_cfg->line_time_in_us; +- +- adev->pm.pm_display_cfg.vrefresh = pp_display_cfg->disp_configs[0].v_refresh; +- adev->pm.pm_display_cfg.crossfire_display_index = -1; +- adev->pm.pm_display_cfg.min_bus_bandwidth = 0; +- +- /* TODO: complete implementation of +- * pp_display_configuration_change(). +- * Follow example of: +- * PHM_StoreDALConfigurationData - powerplay\hwmgr\hardwaremanager.c +- * PP_IRI_DisplayConfigurationChange - powerplay\eventmgr\iri.c */ +- if (adev->powerplay.pp_funcs->display_configuration_change) +- adev->powerplay.pp_funcs->display_configuration_change( +- adev->powerplay.pp_handle, +- &adev->pm.pm_display_cfg); +- +- /* TODO: replace by a separate call to 'apply display cfg'? */ +- amdgpu_pm_compute_clocks(adev); +- } +- +- return true; +-} +- +-static void get_default_clock_levels( +- enum dm_pp_clock_type clk_type, +- struct dm_pp_clock_levels *clks) +-{ +- uint32_t disp_clks_in_khz[6] = { +- 300000, 400000, 496560, 626090, 685720, 757900 }; +- uint32_t sclks_in_khz[6] = { +- 300000, 360000, 423530, 514290, 626090, 720000 }; +- uint32_t mclks_in_khz[2] = { 333000, 800000 }; +- +- switch (clk_type) { +- case DM_PP_CLOCK_TYPE_DISPLAY_CLK: +- clks->num_levels = 6; +- memmove(clks->clocks_in_khz, disp_clks_in_khz, +- sizeof(disp_clks_in_khz)); +- break; +- case DM_PP_CLOCK_TYPE_ENGINE_CLK: +- clks->num_levels = 6; +- memmove(clks->clocks_in_khz, sclks_in_khz, +- sizeof(sclks_in_khz)); +- break; +- case DM_PP_CLOCK_TYPE_MEMORY_CLK: +- clks->num_levels = 2; +- memmove(clks->clocks_in_khz, mclks_in_khz, +- sizeof(mclks_in_khz)); +- break; +- default: +- clks->num_levels = 0; +- break; +- } +-} +- +-static enum amd_pp_clock_type dc_to_pp_clock_type( +- enum dm_pp_clock_type dm_pp_clk_type) +-{ +- enum amd_pp_clock_type amd_pp_clk_type = 0; +- +- switch (dm_pp_clk_type) { +- case DM_PP_CLOCK_TYPE_DISPLAY_CLK: +- amd_pp_clk_type = amd_pp_disp_clock; +- break; +- case DM_PP_CLOCK_TYPE_ENGINE_CLK: +- amd_pp_clk_type = amd_pp_sys_clock; +- break; +- case DM_PP_CLOCK_TYPE_MEMORY_CLK: +- amd_pp_clk_type = amd_pp_mem_clock; +- break; +- case DM_PP_CLOCK_TYPE_DCEFCLK: +- amd_pp_clk_type = amd_pp_dcef_clock; +- break; +- case DM_PP_CLOCK_TYPE_DCFCLK: +- amd_pp_clk_type = amd_pp_dcf_clock; +- break; +- case DM_PP_CLOCK_TYPE_PIXELCLK: +- amd_pp_clk_type = amd_pp_pixel_clock; +- break; +- case DM_PP_CLOCK_TYPE_FCLK: +- amd_pp_clk_type = amd_pp_f_clock; +- break; +- case DM_PP_CLOCK_TYPE_DISPLAYPHYCLK: +- amd_pp_clk_type = amd_pp_dpp_clock; +- break; +- default: +- DRM_ERROR("DM_PPLIB: invalid clock type: %d!\n", +- dm_pp_clk_type); +- break; +- } +- +- return amd_pp_clk_type; +-} +- +-static void pp_to_dc_clock_levels( +- const struct amd_pp_clocks *pp_clks, +- struct dm_pp_clock_levels *dc_clks, +- enum dm_pp_clock_type dc_clk_type) +-{ +- uint32_t i; +- +- if (pp_clks->count > DM_PP_MAX_CLOCK_LEVELS) { +- DRM_INFO("DM_PPLIB: Warning: %s clock: number of levels %d exceeds maximum of %d!\n", +- DC_DECODE_PP_CLOCK_TYPE(dc_clk_type), +- pp_clks->count, +- DM_PP_MAX_CLOCK_LEVELS); +- +- dc_clks->num_levels = DM_PP_MAX_CLOCK_LEVELS; +- } else +- dc_clks->num_levels = pp_clks->count; +- +- DRM_INFO("DM_PPLIB: values for %s clock\n", +- DC_DECODE_PP_CLOCK_TYPE(dc_clk_type)); +- +- for (i = 0; i < dc_clks->num_levels; i++) { +- DRM_INFO("DM_PPLIB:\t %d\n", pp_clks->clock[i]); +- /* translate 10kHz to kHz */ +- dc_clks->clocks_in_khz[i] = pp_clks->clock[i] * 10; +- } +-} +- +-static void pp_to_dc_clock_levels_with_voltage( +- const struct pp_clock_levels_with_voltage *pp_clks, +- struct dm_pp_clock_levels_with_voltage *clk_level_info, +- enum dm_pp_clock_type dc_clk_type) +-{ +- uint32_t i; +- +- if (pp_clks->num_levels > DM_PP_MAX_CLOCK_LEVELS) { +- DRM_INFO("DM_PPLIB: Warning: %s clock: number of levels %d exceeds maximum of %d!\n", +- DC_DECODE_PP_CLOCK_TYPE(dc_clk_type), +- pp_clks->num_levels, +- DM_PP_MAX_CLOCK_LEVELS); +- +- clk_level_info->num_levels = DM_PP_MAX_CLOCK_LEVELS; +- } else +- clk_level_info->num_levels = pp_clks->num_levels; +- +- DRM_INFO("DM_PPLIB: values for %s clock\n", +- DC_DECODE_PP_CLOCK_TYPE(dc_clk_type)); +- +- for (i = 0; i < clk_level_info->num_levels; i++) { +- DRM_INFO("DM_PPLIB:\t %d in 10kHz\n", pp_clks->data[i].clocks_in_khz); +- /* translate 10kHz to kHz */ +- clk_level_info->data[i].clocks_in_khz = pp_clks->data[i].clocks_in_khz * 10; +- clk_level_info->data[i].voltage_in_mv = pp_clks->data[i].voltage_in_mv; +- } +-} +- +- +-bool dm_pp_get_clock_levels_by_type( +- const struct dc_context *ctx, +- enum dm_pp_clock_type clk_type, +- struct dm_pp_clock_levels *dc_clks) +-{ +- struct amdgpu_device *adev = ctx->driver_context; +- void *pp_handle = adev->powerplay.pp_handle; +- struct amd_pp_clocks pp_clks = { 0 }; +- struct amd_pp_simple_clock_info validation_clks = { 0 }; +- uint32_t i; +- +- if (adev->powerplay.pp_funcs->get_clock_by_type) { +- if (adev->powerplay.pp_funcs->get_clock_by_type(pp_handle, +- dc_to_pp_clock_type(clk_type), &pp_clks)) { +- /* Error in pplib. Provide default values. */ +- get_default_clock_levels(clk_type, dc_clks); +- return true; +- } +- } +- +- pp_to_dc_clock_levels(&pp_clks, dc_clks, clk_type); +- +- if (adev->powerplay.pp_funcs->get_display_mode_validation_clocks) { +- if (adev->powerplay.pp_funcs->get_display_mode_validation_clocks( +- pp_handle, &validation_clks)) { +- /* Error in pplib. Provide default values. */ +- DRM_INFO("DM_PPLIB: Warning: using default validation clocks!\n"); +- validation_clks.engine_max_clock = 72000; +- validation_clks.memory_max_clock = 80000; +- validation_clks.level = 0; +- } +- } +- +- DRM_INFO("DM_PPLIB: Validation clocks:\n"); +- DRM_INFO("DM_PPLIB: engine_max_clock: %d\n", +- validation_clks.engine_max_clock); +- DRM_INFO("DM_PPLIB: memory_max_clock: %d\n", +- validation_clks.memory_max_clock); +- DRM_INFO("DM_PPLIB: level : %d\n", +- validation_clks.level); +- +- /* Translate 10 kHz to kHz. */ +- validation_clks.engine_max_clock *= 10; +- validation_clks.memory_max_clock *= 10; +- +- /* Determine the highest non-boosted level from the Validation Clocks */ +- if (clk_type == DM_PP_CLOCK_TYPE_ENGINE_CLK) { +- for (i = 0; i < dc_clks->num_levels; i++) { +- if (dc_clks->clocks_in_khz[i] > validation_clks.engine_max_clock) { +- /* This clock is higher the validation clock. +- * Than means the previous one is the highest +- * non-boosted one. */ +- DRM_INFO("DM_PPLIB: reducing engine clock level from %d to %d\n", +- dc_clks->num_levels, i); +- dc_clks->num_levels = i > 0 ? i : 1; +- break; +- } +- } +- } else if (clk_type == DM_PP_CLOCK_TYPE_MEMORY_CLK) { +- for (i = 0; i < dc_clks->num_levels; i++) { +- if (dc_clks->clocks_in_khz[i] > validation_clks.memory_max_clock) { +- DRM_INFO("DM_PPLIB: reducing memory clock level from %d to %d\n", +- dc_clks->num_levels, i); +- dc_clks->num_levels = i > 0 ? i : 1; +- break; +- } +- } +- } +- +- return true; +-} +- +-bool dm_pp_get_clock_levels_by_type_with_latency( +- const struct dc_context *ctx, +- enum dm_pp_clock_type clk_type, +- struct dm_pp_clock_levels_with_latency *clk_level_info) +-{ +- /* TODO: to be implemented */ +- return false; +-} +- +-bool dm_pp_get_clock_levels_by_type_with_voltage( +- const struct dc_context *ctx, +- enum dm_pp_clock_type clk_type, +- struct dm_pp_clock_levels_with_voltage *clk_level_info) +-{ +- struct amdgpu_device *adev = ctx->driver_context; +- void *pp_handle = adev->powerplay.pp_handle; +- struct pp_clock_levels_with_voltage pp_clk_info = {0}; +- const struct amd_pm_funcs *pp_funcs = adev->powerplay.pp_funcs; +- +- if (pp_funcs->get_clock_by_type_with_voltage(pp_handle, +- dc_to_pp_clock_type(clk_type), +- &pp_clk_info)) +- return false; +- +- pp_to_dc_clock_levels_with_voltage(&pp_clk_info, clk_level_info, clk_type); +- +- return true; +-} +- +-bool dm_pp_notify_wm_clock_changes( +- const struct dc_context *ctx, +- struct dm_pp_wm_sets_with_clock_ranges *wm_with_clock_ranges) +-{ +- /* TODO: to be implemented */ +- return false; +-} +- +-bool dm_pp_apply_power_level_change_request( +- const struct dc_context *ctx, +- struct dm_pp_power_level_change_request *level_change_req) +-{ +- /* TODO: to be implemented */ +- return false; +-} +- +-bool dm_pp_apply_clock_for_voltage_request( +- const struct dc_context *ctx, +- struct dm_pp_clock_for_voltage_req *clock_for_voltage_req) +-{ +- struct amdgpu_device *adev = ctx->driver_context; +- struct pp_display_clock_request pp_clock_request = {0}; +- int ret = 0; +- +- pp_clock_request.clock_type = dc_to_pp_clock_type(clock_for_voltage_req->clk_type); +- pp_clock_request.clock_freq_in_khz = clock_for_voltage_req->clocks_in_khz; +- +- if (!pp_clock_request.clock_type) +- return false; +- +- if (adev->powerplay.pp_funcs->display_clock_voltage_request) +- ret = adev->powerplay.pp_funcs->display_clock_voltage_request( +- adev->powerplay.pp_handle, +- &pp_clock_request); +- if (ret) +- return false; +- return true; +-} +- +-bool dm_pp_get_static_clocks( +- const struct dc_context *ctx, +- struct dm_pp_static_clock_info *static_clk_info) +-{ +- struct amdgpu_device *adev = ctx->driver_context; +- struct amd_pp_clock_info pp_clk_info = {0}; +- int ret = 0; +- +- if (adev->powerplay.pp_funcs->get_current_clocks) +- ret = adev->powerplay.pp_funcs->get_current_clocks( +- adev->powerplay.pp_handle, +- &pp_clk_info); +- if (ret) +- return false; +- +- static_clk_info->max_clocks_state = pp_clk_info.max_clocks_state; +- /* translate 10kHz to kHz */ +- static_clk_info->max_mclk_khz = pp_clk_info.max_memory_clock * 10; +- static_clk_info->max_sclk_khz = pp_clk_info.max_engine_clock * 10; +- +- return true; +-} +- +-void pp_rv_set_display_requirement(struct pp_smu *pp, +- struct pp_smu_display_requirement_rv *req) +-{ +- struct amdgpu_device *adev = pp->ctx->driver_context; +- struct pp_hwmgr *hwmgr = adev->powerplay.pp_handle; +- int ret = 0; +- if (hwmgr->hwmgr_func->set_deep_sleep_dcefclk) +- ret = hwmgr->hwmgr_func->set_deep_sleep_dcefclk(hwmgr, req->hard_min_dcefclk_khz/10); +- if (hwmgr->hwmgr_func->set_active_display_count) +- ret = hwmgr->hwmgr_func->set_active_display_count(hwmgr, req->display_count); +- +- //store_cc6 is not yet implemented in SMU level +-} +- +-void pp_rv_set_wm_ranges(struct pp_smu *pp, +- struct pp_smu_wm_range_sets *ranges) +-{ +- struct amdgpu_device *adev = pp->ctx->driver_context; +- struct pp_hwmgr *hwmgr = adev->powerplay.pp_handle; +- struct pp_wm_sets_with_clock_ranges_soc15 ranges_soc15 = {0}; +- int i = 0; +- +- if (!hwmgr->hwmgr_func->set_watermarks_for_clocks_ranges || +- !pp || !ranges) +- return; +- +- //not entirely sure if thats a correct assignment +- ranges_soc15.num_wm_sets_dmif = ranges->num_reader_wm_sets; +- ranges_soc15.num_wm_sets_mcif = ranges->num_writer_wm_sets; +- +- for (i = 0; i < ranges_soc15.num_wm_sets_dmif; i++) { +- if (ranges->reader_wm_sets[i].wm_inst > 3) +- ranges_soc15.wm_sets_dmif[i].wm_set_id = DC_WM_SET_A; +- else +- ranges_soc15.wm_sets_dmif[i].wm_set_id = +- ranges->reader_wm_sets[i].wm_inst; +- ranges_soc15.wm_sets_dmif[i].wm_max_dcefclk_in_khz = +- ranges->reader_wm_sets[i].max_drain_clk_khz; +- ranges_soc15.wm_sets_dmif[i].wm_min_dcefclk_in_khz = +- ranges->reader_wm_sets[i].min_drain_clk_khz; +- ranges_soc15.wm_sets_dmif[i].wm_max_memclk_in_khz = +- ranges->reader_wm_sets[i].max_fill_clk_khz; +- ranges_soc15.wm_sets_dmif[i].wm_min_memclk_in_khz = +- ranges->reader_wm_sets[i].min_fill_clk_khz; +- } +- +- for (i = 0; i < ranges_soc15.num_wm_sets_mcif; i++) { +- if (ranges->writer_wm_sets[i].wm_inst > 3) +- ranges_soc15.wm_sets_dmif[i].wm_set_id = DC_WM_SET_A; +- else +- ranges_soc15.wm_sets_mcif[i].wm_set_id = +- ranges->writer_wm_sets[i].wm_inst; +- ranges_soc15.wm_sets_mcif[i].wm_max_socclk_in_khz = +- ranges->writer_wm_sets[i].max_fill_clk_khz; +- ranges_soc15.wm_sets_mcif[i].wm_min_socclk_in_khz = +- ranges->writer_wm_sets[i].min_fill_clk_khz; +- ranges_soc15.wm_sets_mcif[i].wm_max_memclk_in_khz = +- ranges->writer_wm_sets[i].max_fill_clk_khz; +- ranges_soc15.wm_sets_mcif[i].wm_min_memclk_in_khz = +- ranges->writer_wm_sets[i].min_fill_clk_khz; +- } +- +- hwmgr->hwmgr_func->set_watermarks_for_clocks_ranges(hwmgr, &ranges_soc15); +- +-} +- +-void pp_rv_set_pme_wa_enable(struct pp_smu *pp) +-{ +- struct amdgpu_device *adev = pp->ctx->driver_context; +- struct pp_hwmgr *hwmgr = adev->powerplay.pp_handle; +- +- if (hwmgr->hwmgr_func->smus_notify_pwe) +- hwmgr->hwmgr_func->smus_notify_pwe(hwmgr); +-} +- +-void dm_pp_get_funcs_rv( +- struct dc_context *ctx, +- struct pp_smu_funcs_rv *funcs) +-{ +- funcs->pp_smu.ctx = ctx; +- funcs->set_display_requirement = pp_rv_set_display_requirement; +- funcs->set_wm_ranges = pp_rv_set_wm_ranges; +- funcs->set_pme_wa_enable = pp_rv_set_pme_wa_enable; +-} +- +- +-/**** end of power component interfaces ****/ +-- +2.7.4 + |