aboutsummaryrefslogtreecommitdiffstats
path: root/common/recipes-kernel/linux/linux-yocto-4.14.71/0668-drm-amd-display-break-up-plane-disable-and-disconnec.patch
diff options
context:
space:
mode:
Diffstat (limited to 'common/recipes-kernel/linux/linux-yocto-4.14.71/0668-drm-amd-display-break-up-plane-disable-and-disconnec.patch')
-rw-r--r--common/recipes-kernel/linux/linux-yocto-4.14.71/0668-drm-amd-display-break-up-plane-disable-and-disconnec.patch282
1 files changed, 282 insertions, 0 deletions
diff --git a/common/recipes-kernel/linux/linux-yocto-4.14.71/0668-drm-amd-display-break-up-plane-disable-and-disconnec.patch b/common/recipes-kernel/linux/linux-yocto-4.14.71/0668-drm-amd-display-break-up-plane-disable-and-disconnec.patch
new file mode 100644
index 00000000..59fbfc47
--- /dev/null
+++ b/common/recipes-kernel/linux/linux-yocto-4.14.71/0668-drm-amd-display-break-up-plane-disable-and-disconnec.patch
@@ -0,0 +1,282 @@
+From 81466cab7c571e6ed0cc5022fd16deb43a722180 Mon Sep 17 00:00:00 2001
+From: Eric Yang <Eric.Yang2@amd.com>
+Date: Mon, 24 Jul 2017 10:47:02 -0400
+Subject: [PATCH 0668/4131] drm/amd/display: break up plane disable and
+ disconnect in set mode
+
+Signed-off-by: Eric Yang <Eric.Yang2@amd.com>
+Reviewed-by: Tony Cheng <Tony.Cheng@amd.com>
+Acked-by: Harry Wentland <Harry.Wentland@amd.com>
+Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
+---
+ drivers/gpu/drm/amd/display/dc/core/dc.c | 23 +++--
+ .../amd/display/dc/dce110/dce110_hw_sequencer.c | 5 +-
+ .../drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c | 102 ++++++++++++++++++---
+ drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h | 4 +-
+ 4 files changed, 111 insertions(+), 23 deletions(-)
+
+diff --git a/drivers/gpu/drm/amd/display/dc/core/dc.c b/drivers/gpu/drm/amd/display/dc/core/dc.c
+index d3873ac..7921880 100644
+--- a/drivers/gpu/drm/amd/display/dc/core/dc.c
++++ b/drivers/gpu/drm/amd/display/dc/core/dc.c
+@@ -932,7 +932,7 @@ static bool dc_commit_context_no_check(struct dc *dc, struct validate_context *c
+
+ for (i = 0; i < core_dc->res_pool->pipe_count; i++) {
+ pipe = &context->res_ctx.pipe_ctx[i];
+- core_dc->hwss.wait_for_mpcc_disconnect(core_dc->res_pool, pipe);
++ core_dc->hwss.wait_for_mpcc_disconnect(core_dc, core_dc->res_pool, pipe);
+ }
+ result = core_dc->hwss.apply_ctx_to_hw(core_dc, context);
+
+@@ -1364,6 +1364,16 @@ void dc_update_surfaces_and_stream(struct dc *dc,
+ struct core_stream *stream = DC_STREAM_TO_CORE(dc_stream);
+ struct dc_context *dc_ctx = core_dc->ctx;
+
++ /* Currently this function do not result in any HW programming
++ * when called with 0 surface. But proceeding will cause
++ * SW state to be updated in validate_context. So we might as
++ * well make it not do anything at all until the hw programming
++ * is implemented properly to handle 0 surface case.
++ * TODO: fix hw programming then remove this early return
++ */
++ if (surface_count == 0)
++ return;
++
+ stream_status = dc_stream_get_status(dc_stream);
+ ASSERT(stream_status);
+ if (!stream_status)
+@@ -1535,15 +1545,10 @@ void dc_update_surfaces_and_stream(struct dc *dc,
+ }
+
+ if (update_type > UPDATE_TYPE_FAST) {
+- for (i = 0; i < surface_count; i++) {
+- for (j = 0; j < core_dc->res_pool->pipe_count; j++) {
+- struct pipe_ctx *pipe_ctx = &context->res_ctx.pipe_ctx[j];
+-
+- if (pipe_ctx->surface != srf_updates[i].surface)
+- continue;
++ for (j = 0; j < core_dc->res_pool->pipe_count; j++) {
++ struct pipe_ctx *pipe_ctx = &context->res_ctx.pipe_ctx[j];
+
+- core_dc->hwss.wait_for_mpcc_disconnect(core_dc->res_pool, pipe_ctx);
+- }
++ core_dc->hwss.wait_for_mpcc_disconnect(core_dc, core_dc->res_pool, pipe_ctx);
+ }
+ }
+
+diff --git a/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c b/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c
+index 58701fd..aa4777a 100644
+--- a/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c
++++ b/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c
+@@ -2583,7 +2583,10 @@ static void dce110_power_down_fe(struct core_dc *dc, int fe_idx)
+ dc->res_pool->transforms[fe_idx]);
+ }
+
+-static void dce110_wait_for_mpcc_disconnect(struct resource_pool *res_pool, struct pipe_ctx *pipe_ctx)
++static void dce110_wait_for_mpcc_disconnect(
++ struct core_dc *dc,
++ struct resource_pool *res_pool,
++ struct pipe_ctx *pipe_ctx)
+ {
+ /* do nothing*/
+ }
+diff --git a/drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c b/drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c
+index 3a0fcaa..bbe3e87 100644
+--- a/drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c
++++ b/drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c
+@@ -831,21 +831,32 @@ static void plane_atomic_disconnect(struct core_dc *dc,
+ if (opp_id == 0xf)
+ return;
+
++ if (dc->public.debug.sanity_checks)
++ verify_allow_pstate_change_high(dc->hwseq);
++
+ mi->funcs->dcc_control(mi, false, false);
+
++ if (dc->public.debug.sanity_checks)
++ verify_allow_pstate_change_high(dc->hwseq);
++
+ mpcc_cfg.opp_id = 0xf;
+ mpcc_cfg.top_dpp_id = 0xf;
+ mpcc_cfg.bot_mpcc_id = 0xf;
+ mpcc_cfg.top_of_tree = tg->inst == mpcc->inst;
+ mpcc->funcs->set(mpcc, &mpcc_cfg);
+
+- /* Hack to preserve old opp_id for plane_atomic_disable
+- * to find the correct otg */
++ /*
++ * Hack to preserve old opp_id for plane_atomic_disable
++ * to find the correct otg
++ */
+ mpcc->opp_id = opp_id_cached;
+
+ /* todo:call remove pipe from tree */
+ /* flag mpcc idle pending */
+
++ /*dm_logger_write(dc->ctx->logger, LOG_ERROR,
++ "[debug_mpo: plane_atomic_disconnect pending on mpcc %d]\n",
++ fe_idx);*/
+ xfm->funcs->transform_reset(xfm);
+ }
+
+@@ -864,6 +875,10 @@ static void plane_atomic_disable(struct core_dc *dc,
+ return;
+
+ mpcc->funcs->wait_for_idle(mpcc);
++ dc->res_pool->opps[opp_id]->mpcc_disconnect_pending[mpcc->inst] = false;
++ /*dm_logger_write(dc->ctx->logger, LOG_ERROR,
++ "[debug_mpo: atomic disable finished on mpcc %d]\n",
++ fe_idx);*/
+
+ mi->funcs->set_blank(mi, true);
+
+@@ -885,13 +900,27 @@ static void plane_atomic_disable(struct core_dc *dc,
+ verify_allow_pstate_change_high(dc->hwseq);
+ }
+
+-/* kill power to plane hw
++/*
++ * kill power to plane hw
+ * note: cannot power down until plane is disable
+-static void plane_atomic_power_down()
++ */
++static void plane_atomic_power_down(struct core_dc *dc, int fe_idx)
+ {
++ struct dce_hwseq *hws = dc->hwseq;
++
++ REG_SET(DC_IP_REQUEST_CNTL, 0,
++ IP_REQUEST_EN, 1);
++ dpp_pg_control(hws, fe_idx, false);
++ hubp_pg_control(hws, fe_idx, false);
++ REG_SET(DC_IP_REQUEST_CNTL, 0,
++ IP_REQUEST_EN, 0);
++ dm_logger_write(dc->ctx->logger, LOG_DC,
++ "Power gated front end %d\n", fe_idx);
+
++ if (dc->public.debug.sanity_checks)
++ verify_allow_pstate_change_high(dc->hwseq);
+ }
+-*/
++
+
+ static void reset_front_end(
+ struct core_dc *dc,
+@@ -953,6 +982,37 @@ static void reset_hw_ctx_wrap(
+ int i;
+
+ /* Reset Front End*/
++ /* Lock*/
++ for (i = 0; i < dc->res_pool->pipe_count; i++) {
++ struct pipe_ctx *cur_pipe_ctx = &dc->current_context->res_ctx.pipe_ctx[i];
++ struct timing_generator *tg = cur_pipe_ctx->tg;
++
++ if (cur_pipe_ctx->stream)
++ tg->funcs->lock(tg);
++ }
++ /* Disconnect*/
++ for (i = dc->res_pool->pipe_count - 1; i >= 0 ; i--) {
++ struct pipe_ctx *pipe_ctx_old =
++ &dc->current_context->res_ctx.pipe_ctx[i];
++ struct pipe_ctx *pipe_ctx = &context->res_ctx.pipe_ctx[i];
++
++ if (!pipe_ctx->stream ||
++ !pipe_ctx->surface ||
++ pipe_need_reprogram(pipe_ctx_old, pipe_ctx)) {
++
++ plane_atomic_disconnect(dc, i);
++ }
++ }
++ /* Unlock*/
++ for (i = dc->res_pool->pipe_count - 1; i >= 0; i--) {
++ struct pipe_ctx *cur_pipe_ctx = &dc->current_context->res_ctx.pipe_ctx[i];
++ struct timing_generator *tg = cur_pipe_ctx->tg;
++
++ if (cur_pipe_ctx->stream)
++ tg->funcs->unlock(tg);
++ }
++
++ /* Disable and Powerdown*/
+ for (i = dc->res_pool->pipe_count - 1; i >= 0 ; i--) {
+ struct pipe_ctx *pipe_ctx_old =
+ &dc->current_context->res_ctx.pipe_ctx[i];
+@@ -961,11 +1021,16 @@ static void reset_hw_ctx_wrap(
+ /*if (!pipe_ctx_old->stream)
+ continue;*/
+
++ if (pipe_ctx->stream && pipe_ctx->surface
++ && !pipe_need_reprogram(pipe_ctx_old, pipe_ctx))
++ continue;
++
++ plane_atomic_disable(dc, i);
++
+ if (!pipe_ctx->stream || !pipe_ctx->surface)
+- dcn10_power_down_fe(dc, i);
+- else if (pipe_need_reprogram(pipe_ctx_old, pipe_ctx))
+- reset_front_end(dc, i);
++ plane_atomic_power_down(dc, i);
+ }
++
+ /* Reset Back End*/
+ for (i = dc->res_pool->pipe_count - 1; i >= 0 ; i--) {
+ struct pipe_ctx *pipe_ctx_old =
+@@ -2079,6 +2144,10 @@ static void dcn10_apply_ctx_for_surface(
+ old_pipe_ctx->mpcc->funcs->set(old_pipe_ctx->mpcc, &mpcc_cfg);
+ old_pipe_ctx->top_pipe->opp->mpcc_disconnect_pending[old_pipe_ctx->mpcc->inst] = true;
+
++ /*dm_logger_write(dc->ctx->logger, LOG_ERROR,
++ "[debug_mpo: apply_ctx disconnect pending on mpcc %d]\n",
++ old_pipe_ctx->mpcc->inst);*/
++
+ if (dc->public.debug.sanity_checks)
+ verify_allow_pstate_change_high(dc->hwseq);
+
+@@ -2351,18 +2420,27 @@ static void dcn10_log_hw_state(struct core_dc *dc)
+ */
+ }
+
+-static void dcn10_wait_for_mpcc_disconnect(struct resource_pool *res_pool, struct pipe_ctx *pipe_ctx)
++static void dcn10_wait_for_mpcc_disconnect(
++ struct core_dc *dc,
++ struct resource_pool *res_pool,
++ struct pipe_ctx *pipe_ctx)
+ {
+ int i;
+- for (i = 0; i < MAX_PIPES; i++) {
+- if (!pipe_ctx->opp || !pipe_ctx->mpcc)
+- continue;
+
++ if (!pipe_ctx->opp || !pipe_ctx->mpcc)
++ return;
++
++ for (i = 0; i < MAX_PIPES; i++) {
+ if (pipe_ctx->opp->mpcc_disconnect_pending[i]) {
+ pipe_ctx->mpcc->funcs->wait_for_idle(res_pool->mpcc[i]);
+ pipe_ctx->opp->mpcc_disconnect_pending[i] = false;
++ res_pool->mis[i]->funcs->set_blank(res_pool->mis[i], true);
++ /*dm_logger_write(dc->ctx->logger, LOG_ERROR,
++ "[debug_mpo: wait_for_mpcc finished waiting on mpcc %d]\n",
++ i);*/
+ }
+ }
++
+ }
+
+ static bool dcn10_dummy_display_power_gating(
+diff --git a/drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h b/drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h
+index 2343beb..a3eec07 100644
+--- a/drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h
++++ b/drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h
+@@ -161,7 +161,9 @@ struct hw_sequencer_funcs {
+
+ void (*log_hw_state)(struct core_dc *dc);
+
+- void (*wait_for_mpcc_disconnect)(struct resource_pool *res_pool, struct pipe_ctx *pipe_ctx);
++ void (*wait_for_mpcc_disconnect)(struct core_dc *dc,
++ struct resource_pool *res_pool,
++ struct pipe_ctx *pipe_ctx);
+ };
+
+ void color_space_to_black_color(
+--
+2.7.4
+