struct dc_state *old_state;
bool subvp_prev_use = false;
- dc_allow_idle_optimizations(dc, false);
dc_z10_restore(dc);
+ dc_allow_idle_optimizations(dc, false);
for (i = 0; i < dc->res_pool->pipe_count; i++) {
struct pipe_ctx *old_pipe = &dc->current_state->res_ctx.pipe_ctx[i];
int i, j;
struct pipe_ctx *top_pipe_to_program = NULL;
struct dc_stream_status *stream_status = NULL;
- if (dc->caps.ips_support)
- dc_allow_idle_optimizations(dc, false);
-
dc_z10_restore(dc);
top_pipe_to_program = resource_get_otg_master_for_stream(
// dc->current_state anymore, so we have to cache it before we apply
// the new SubVP context
subvp_prev_use = false;
- if (dc->caps.ips_support)
- dc_allow_idle_optimizations(dc, false);
-
dc_z10_restore(dc);
if (update_type == UPDATE_TYPE_FULL)
wait_for_outstanding_hw_updates(dc, context);
case DC_ACPI_CM_POWER_STATE_D0:
dc_state_construct(dc, dc->current_state);
- if (dc->caps.ips_support)
- dc_allow_idle_optimizations(dc, false);
-
dc_z10_restore(dc);
dc->hwss.init_hw(dc);
stream->cursor_attributes = *attributes;
+ dc_z10_restore(dc);
/* disable idle optimizations while updating cursor */
if (dc->idle_optimizations_allowed) {
dc_allow_idle_optimizations(dc, false);
}
dc = stream->ctx->dc;
+ dc_z10_restore(dc);
/* disable idle optimizations if enabling cursor */
- if (dc->idle_optimizations_allowed &&
- (!stream->cursor_position.enable || dc->debug.exit_idle_opt_for_cursor_updates ||
- dc->caps.ips_support) &&
- position->enable) {
+ if (dc->idle_optimizations_allowed && (!stream->cursor_position.enable || dc->debug.exit_idle_opt_for_cursor_updates)
+ && position->enable) {
dc_allow_idle_optimizations(dc, false);
- dc_z10_restore(dc);
reset_idle_optimizations = true;
}