From 81466cab7c571e6ed0cc5022fd16deb43a722180 Mon Sep 17 00:00:00 2001 From: Eric Yang Date: Mon, 24 Jul 2017 10:47:02 -0400 Subject: [PATCH 0668/4131] drm/amd/display: break up plane disable and disconnect in set mode Signed-off-by: Eric Yang Reviewed-by: Tony Cheng Acked-by: Harry Wentland Signed-off-by: Alex Deucher --- drivers/gpu/drm/amd/display/dc/core/dc.c | 23 +++-- .../amd/display/dc/dce110/dce110_hw_sequencer.c | 5 +- .../drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c | 102 ++++++++++++++++++--- drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h | 4 +- 4 files changed, 111 insertions(+), 23 deletions(-) diff --git a/drivers/gpu/drm/amd/display/dc/core/dc.c b/drivers/gpu/drm/amd/display/dc/core/dc.c index d3873ac..7921880 100644 --- a/drivers/gpu/drm/amd/display/dc/core/dc.c +++ b/drivers/gpu/drm/amd/display/dc/core/dc.c @@ -932,7 +932,7 @@ static bool dc_commit_context_no_check(struct dc *dc, struct validate_context *c for (i = 0; i < core_dc->res_pool->pipe_count; i++) { pipe = &context->res_ctx.pipe_ctx[i]; - core_dc->hwss.wait_for_mpcc_disconnect(core_dc->res_pool, pipe); + core_dc->hwss.wait_for_mpcc_disconnect(core_dc, core_dc->res_pool, pipe); } result = core_dc->hwss.apply_ctx_to_hw(core_dc, context); @@ -1364,6 +1364,16 @@ void dc_update_surfaces_and_stream(struct dc *dc, struct core_stream *stream = DC_STREAM_TO_CORE(dc_stream); struct dc_context *dc_ctx = core_dc->ctx; + /* Currently this function do not result in any HW programming + * when called with 0 surface. But proceeding will cause + * SW state to be updated in validate_context. So we might as + * well make it not do anything at all until the hw programming + * is implemented properly to handle 0 surface case. + * TODO: fix hw programming then remove this early return + */ + if (surface_count == 0) + return; + stream_status = dc_stream_get_status(dc_stream); ASSERT(stream_status); if (!stream_status) @@ -1535,15 +1545,10 @@ void dc_update_surfaces_and_stream(struct dc *dc, } if (update_type > UPDATE_TYPE_FAST) { - for (i = 0; i < surface_count; i++) { - for (j = 0; j < core_dc->res_pool->pipe_count; j++) { - struct pipe_ctx *pipe_ctx = &context->res_ctx.pipe_ctx[j]; - - if (pipe_ctx->surface != srf_updates[i].surface) - continue; + for (j = 0; j < core_dc->res_pool->pipe_count; j++) { + struct pipe_ctx *pipe_ctx = &context->res_ctx.pipe_ctx[j]; - core_dc->hwss.wait_for_mpcc_disconnect(core_dc->res_pool, pipe_ctx); - } + core_dc->hwss.wait_for_mpcc_disconnect(core_dc, core_dc->res_pool, pipe_ctx); } } diff --git a/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c b/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c index 58701fd..aa4777a 100644 --- a/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c +++ b/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c @@ -2583,7 +2583,10 @@ static void dce110_power_down_fe(struct core_dc *dc, int fe_idx) dc->res_pool->transforms[fe_idx]); } -static void dce110_wait_for_mpcc_disconnect(struct resource_pool *res_pool, struct pipe_ctx *pipe_ctx) +static void dce110_wait_for_mpcc_disconnect( + struct core_dc *dc, + struct resource_pool *res_pool, + struct pipe_ctx *pipe_ctx) { /* do nothing*/ } diff --git a/drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c b/drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c index 3a0fcaa..bbe3e87 100644 --- a/drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c +++ b/drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c @@ -831,21 +831,32 @@ static void plane_atomic_disconnect(struct core_dc *dc, if (opp_id == 0xf) return; + if (dc->public.debug.sanity_checks) + verify_allow_pstate_change_high(dc->hwseq); + mi->funcs->dcc_control(mi, false, false); + if (dc->public.debug.sanity_checks) + verify_allow_pstate_change_high(dc->hwseq); + mpcc_cfg.opp_id = 0xf; mpcc_cfg.top_dpp_id = 0xf; mpcc_cfg.bot_mpcc_id = 0xf; mpcc_cfg.top_of_tree = tg->inst == mpcc->inst; mpcc->funcs->set(mpcc, &mpcc_cfg); - /* Hack to preserve old opp_id for plane_atomic_disable - * to find the correct otg */ + /* + * Hack to preserve old opp_id for plane_atomic_disable + * to find the correct otg + */ mpcc->opp_id = opp_id_cached; /* todo:call remove pipe from tree */ /* flag mpcc idle pending */ + /*dm_logger_write(dc->ctx->logger, LOG_ERROR, + "[debug_mpo: plane_atomic_disconnect pending on mpcc %d]\n", + fe_idx);*/ xfm->funcs->transform_reset(xfm); } @@ -864,6 +875,10 @@ static void plane_atomic_disable(struct core_dc *dc, return; mpcc->funcs->wait_for_idle(mpcc); + dc->res_pool->opps[opp_id]->mpcc_disconnect_pending[mpcc->inst] = false; + /*dm_logger_write(dc->ctx->logger, LOG_ERROR, + "[debug_mpo: atomic disable finished on mpcc %d]\n", + fe_idx);*/ mi->funcs->set_blank(mi, true); @@ -885,13 +900,27 @@ static void plane_atomic_disable(struct core_dc *dc, verify_allow_pstate_change_high(dc->hwseq); } -/* kill power to plane hw +/* + * kill power to plane hw * note: cannot power down until plane is disable -static void plane_atomic_power_down() + */ +static void plane_atomic_power_down(struct core_dc *dc, int fe_idx) { + struct dce_hwseq *hws = dc->hwseq; + + REG_SET(DC_IP_REQUEST_CNTL, 0, + IP_REQUEST_EN, 1); + dpp_pg_control(hws, fe_idx, false); + hubp_pg_control(hws, fe_idx, false); + REG_SET(DC_IP_REQUEST_CNTL, 0, + IP_REQUEST_EN, 0); + dm_logger_write(dc->ctx->logger, LOG_DC, + "Power gated front end %d\n", fe_idx); + if (dc->public.debug.sanity_checks) + verify_allow_pstate_change_high(dc->hwseq); } -*/ + static void reset_front_end( struct core_dc *dc, @@ -953,6 +982,37 @@ static void reset_hw_ctx_wrap( int i; /* Reset Front End*/ + /* Lock*/ + for (i = 0; i < dc->res_pool->pipe_count; i++) { + struct pipe_ctx *cur_pipe_ctx = &dc->current_context->res_ctx.pipe_ctx[i]; + struct timing_generator *tg = cur_pipe_ctx->tg; + + if (cur_pipe_ctx->stream) + tg->funcs->lock(tg); + } + /* Disconnect*/ + for (i = dc->res_pool->pipe_count - 1; i >= 0 ; i--) { + struct pipe_ctx *pipe_ctx_old = + &dc->current_context->res_ctx.pipe_ctx[i]; + struct pipe_ctx *pipe_ctx = &context->res_ctx.pipe_ctx[i]; + + if (!pipe_ctx->stream || + !pipe_ctx->surface || + pipe_need_reprogram(pipe_ctx_old, pipe_ctx)) { + + plane_atomic_disconnect(dc, i); + } + } + /* Unlock*/ + for (i = dc->res_pool->pipe_count - 1; i >= 0; i--) { + struct pipe_ctx *cur_pipe_ctx = &dc->current_context->res_ctx.pipe_ctx[i]; + struct timing_generator *tg = cur_pipe_ctx->tg; + + if (cur_pipe_ctx->stream) + tg->funcs->unlock(tg); + } + + /* Disable and Powerdown*/ for (i = dc->res_pool->pipe_count - 1; i >= 0 ; i--) { struct pipe_ctx *pipe_ctx_old = &dc->current_context->res_ctx.pipe_ctx[i]; @@ -961,11 +1021,16 @@ static void reset_hw_ctx_wrap( /*if (!pipe_ctx_old->stream) continue;*/ + if (pipe_ctx->stream && pipe_ctx->surface + && !pipe_need_reprogram(pipe_ctx_old, pipe_ctx)) + continue; + + plane_atomic_disable(dc, i); + if (!pipe_ctx->stream || !pipe_ctx->surface) - dcn10_power_down_fe(dc, i); - else if (pipe_need_reprogram(pipe_ctx_old, pipe_ctx)) - reset_front_end(dc, i); + plane_atomic_power_down(dc, i); } + /* Reset Back End*/ for (i = dc->res_pool->pipe_count - 1; i >= 0 ; i--) { struct pipe_ctx *pipe_ctx_old = @@ -2079,6 +2144,10 @@ static void dcn10_apply_ctx_for_surface( old_pipe_ctx->mpcc->funcs->set(old_pipe_ctx->mpcc, &mpcc_cfg); old_pipe_ctx->top_pipe->opp->mpcc_disconnect_pending[old_pipe_ctx->mpcc->inst] = true; + /*dm_logger_write(dc->ctx->logger, LOG_ERROR, + "[debug_mpo: apply_ctx disconnect pending on mpcc %d]\n", + old_pipe_ctx->mpcc->inst);*/ + if (dc->public.debug.sanity_checks) verify_allow_pstate_change_high(dc->hwseq); @@ -2351,18 +2420,27 @@ static void dcn10_log_hw_state(struct core_dc *dc) */ } -static void dcn10_wait_for_mpcc_disconnect(struct resource_pool *res_pool, struct pipe_ctx *pipe_ctx) +static void dcn10_wait_for_mpcc_disconnect( + struct core_dc *dc, + struct resource_pool *res_pool, + struct pipe_ctx *pipe_ctx) { int i; - for (i = 0; i < MAX_PIPES; i++) { - if (!pipe_ctx->opp || !pipe_ctx->mpcc) - continue; + if (!pipe_ctx->opp || !pipe_ctx->mpcc) + return; + + for (i = 0; i < MAX_PIPES; i++) { if (pipe_ctx->opp->mpcc_disconnect_pending[i]) { pipe_ctx->mpcc->funcs->wait_for_idle(res_pool->mpcc[i]); pipe_ctx->opp->mpcc_disconnect_pending[i] = false; + res_pool->mis[i]->funcs->set_blank(res_pool->mis[i], true); + /*dm_logger_write(dc->ctx->logger, LOG_ERROR, + "[debug_mpo: wait_for_mpcc finished waiting on mpcc %d]\n", + i);*/ } } + } static bool dcn10_dummy_display_power_gating( diff --git a/drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h b/drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h index 2343beb..a3eec07 100644 --- a/drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h +++ b/drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h @@ -161,7 +161,9 @@ struct hw_sequencer_funcs { void (*log_hw_state)(struct core_dc *dc); - void (*wait_for_mpcc_disconnect)(struct resource_pool *res_pool, struct pipe_ctx *pipe_ctx); + void (*wait_for_mpcc_disconnect)(struct core_dc *dc, + struct resource_pool *res_pool, + struct pipe_ctx *pipe_ctx); }; void color_space_to_black_color( -- 2.7.4