Merge tag 'nfs-for-4.20-2' of git://git.linux-nfs.org/projects/trondmy/linux-nfs
[linux-2.6-block.git] / drivers / gpu / drm / msm / disp / dpu1 / dpu_crtc.c
1 /*
2  * Copyright (c) 2014-2018 The Linux Foundation. All rights reserved.
3  * Copyright (C) 2013 Red Hat
4  * Author: Rob Clark <robdclark@gmail.com>
5  *
6  * This program is free software; you can redistribute it and/or modify it
7  * under the terms of the GNU General Public License version 2 as published by
8  * the Free Software Foundation.
9  *
10  * This program is distributed in the hope that it will be useful, but WITHOUT
11  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
12  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
13  * more details.
14  *
15  * You should have received a copy of the GNU General Public License along with
16  * this program.  If not, see <http://www.gnu.org/licenses/>.
17  */
18
19 #define pr_fmt(fmt)     "[drm:%s:%d] " fmt, __func__, __LINE__
20 #include <linux/sort.h>
21 #include <linux/debugfs.h>
22 #include <linux/ktime.h>
23 #include <drm/drm_mode.h>
24 #include <drm/drm_crtc.h>
25 #include <drm/drm_crtc_helper.h>
26 #include <drm/drm_flip_work.h>
27 #include <drm/drm_rect.h>
28
29 #include "dpu_kms.h"
30 #include "dpu_hw_lm.h"
31 #include "dpu_hw_ctl.h"
32 #include "dpu_crtc.h"
33 #include "dpu_plane.h"
34 #include "dpu_encoder.h"
35 #include "dpu_vbif.h"
36 #include "dpu_power_handle.h"
37 #include "dpu_core_perf.h"
38 #include "dpu_trace.h"
39
40 #define DPU_DRM_BLEND_OP_NOT_DEFINED    0
41 #define DPU_DRM_BLEND_OP_OPAQUE         1
42 #define DPU_DRM_BLEND_OP_PREMULTIPLIED  2
43 #define DPU_DRM_BLEND_OP_COVERAGE       3
44 #define DPU_DRM_BLEND_OP_MAX            4
45
46 /* layer mixer index on dpu_crtc */
47 #define LEFT_MIXER 0
48 #define RIGHT_MIXER 1
49
50 static inline int _dpu_crtc_get_mixer_width(struct dpu_crtc_state *cstate,
51                                             struct drm_display_mode *mode)
52 {
53         return mode->hdisplay / cstate->num_mixers;
54 }
55
56 static inline struct dpu_kms *_dpu_crtc_get_kms(struct drm_crtc *crtc)
57 {
58         struct msm_drm_private *priv = crtc->dev->dev_private;
59
60         return to_dpu_kms(priv->kms);
61 }
62
63 static void dpu_crtc_destroy(struct drm_crtc *crtc)
64 {
65         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
66
67         DPU_DEBUG("\n");
68
69         if (!crtc)
70                 return;
71
72         dpu_crtc->phandle = NULL;
73
74         drm_crtc_cleanup(crtc);
75         mutex_destroy(&dpu_crtc->crtc_lock);
76         kfree(dpu_crtc);
77 }
78
79 static void _dpu_crtc_setup_blend_cfg(struct dpu_crtc_mixer *mixer,
80                 struct dpu_plane_state *pstate, struct dpu_format *format)
81 {
82         struct dpu_hw_mixer *lm = mixer->hw_lm;
83         uint32_t blend_op;
84         struct drm_format_name_buf format_name;
85
86         /* default to opaque blending */
87         blend_op = DPU_BLEND_FG_ALPHA_FG_CONST |
88                 DPU_BLEND_BG_ALPHA_BG_CONST;
89
90         if (format->alpha_enable) {
91                 /* coverage blending */
92                 blend_op = DPU_BLEND_FG_ALPHA_FG_PIXEL |
93                         DPU_BLEND_BG_ALPHA_FG_PIXEL |
94                         DPU_BLEND_BG_INV_ALPHA;
95         }
96
97         lm->ops.setup_blend_config(lm, pstate->stage,
98                                 0xFF, 0, blend_op);
99
100         DPU_DEBUG("format:%s, alpha_en:%u blend_op:0x%x\n",
101                 drm_get_format_name(format->base.pixel_format, &format_name),
102                 format->alpha_enable, blend_op);
103 }
104
105 static void _dpu_crtc_program_lm_output_roi(struct drm_crtc *crtc)
106 {
107         struct dpu_crtc *dpu_crtc;
108         struct dpu_crtc_state *crtc_state;
109         int lm_idx, lm_horiz_position;
110
111         dpu_crtc = to_dpu_crtc(crtc);
112         crtc_state = to_dpu_crtc_state(crtc->state);
113
114         lm_horiz_position = 0;
115         for (lm_idx = 0; lm_idx < crtc_state->num_mixers; lm_idx++) {
116                 const struct drm_rect *lm_roi = &crtc_state->lm_bounds[lm_idx];
117                 struct dpu_hw_mixer *hw_lm = crtc_state->mixers[lm_idx].hw_lm;
118                 struct dpu_hw_mixer_cfg cfg;
119
120                 if (!lm_roi || !drm_rect_visible(lm_roi))
121                         continue;
122
123                 cfg.out_width = drm_rect_width(lm_roi);
124                 cfg.out_height = drm_rect_height(lm_roi);
125                 cfg.right_mixer = lm_horiz_position++;
126                 cfg.flags = 0;
127                 hw_lm->ops.setup_mixer_out(hw_lm, &cfg);
128         }
129 }
130
131 static void _dpu_crtc_blend_setup_mixer(struct drm_crtc *crtc,
132         struct dpu_crtc *dpu_crtc, struct dpu_crtc_mixer *mixer)
133 {
134         struct drm_plane *plane;
135         struct drm_framebuffer *fb;
136         struct drm_plane_state *state;
137         struct dpu_crtc_state *cstate = to_dpu_crtc_state(crtc->state);
138         struct dpu_plane_state *pstate = NULL;
139         struct dpu_format *format;
140         struct dpu_hw_ctl *ctl = mixer->lm_ctl;
141         struct dpu_hw_stage_cfg *stage_cfg = &dpu_crtc->stage_cfg;
142
143         u32 flush_mask;
144         uint32_t stage_idx, lm_idx;
145         int zpos_cnt[DPU_STAGE_MAX + 1] = { 0 };
146         bool bg_alpha_enable = false;
147
148         drm_atomic_crtc_for_each_plane(plane, crtc) {
149                 state = plane->state;
150                 if (!state)
151                         continue;
152
153                 pstate = to_dpu_plane_state(state);
154                 fb = state->fb;
155
156                 dpu_plane_get_ctl_flush(plane, ctl, &flush_mask);
157
158                 DPU_DEBUG("crtc %d stage:%d - plane %d sspp %d fb %d\n",
159                                 crtc->base.id,
160                                 pstate->stage,
161                                 plane->base.id,
162                                 dpu_plane_pipe(plane) - SSPP_VIG0,
163                                 state->fb ? state->fb->base.id : -1);
164
165                 format = to_dpu_format(msm_framebuffer_format(pstate->base.fb));
166
167                 if (pstate->stage == DPU_STAGE_BASE && format->alpha_enable)
168                         bg_alpha_enable = true;
169
170                 stage_idx = zpos_cnt[pstate->stage]++;
171                 stage_cfg->stage[pstate->stage][stage_idx] =
172                                         dpu_plane_pipe(plane);
173                 stage_cfg->multirect_index[pstate->stage][stage_idx] =
174                                         pstate->multirect_index;
175
176                 trace_dpu_crtc_setup_mixer(DRMID(crtc), DRMID(plane),
177                                            state, pstate, stage_idx,
178                                            dpu_plane_pipe(plane) - SSPP_VIG0,
179                                            format->base.pixel_format,
180                                            fb ? fb->modifier : 0);
181
182                 /* blend config update */
183                 for (lm_idx = 0; lm_idx < cstate->num_mixers; lm_idx++) {
184                         _dpu_crtc_setup_blend_cfg(mixer + lm_idx,
185                                                 pstate, format);
186
187                         mixer[lm_idx].flush_mask |= flush_mask;
188
189                         if (bg_alpha_enable && !format->alpha_enable)
190                                 mixer[lm_idx].mixer_op_mode = 0;
191                         else
192                                 mixer[lm_idx].mixer_op_mode |=
193                                                 1 << pstate->stage;
194                 }
195         }
196
197          _dpu_crtc_program_lm_output_roi(crtc);
198 }
199
200 /**
201  * _dpu_crtc_blend_setup - configure crtc mixers
202  * @crtc: Pointer to drm crtc structure
203  */
204 static void _dpu_crtc_blend_setup(struct drm_crtc *crtc)
205 {
206         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
207         struct dpu_crtc_state *cstate = to_dpu_crtc_state(crtc->state);
208         struct dpu_crtc_mixer *mixer = cstate->mixers;
209         struct dpu_hw_ctl *ctl;
210         struct dpu_hw_mixer *lm;
211         int i;
212
213         DPU_DEBUG("%s\n", dpu_crtc->name);
214
215         for (i = 0; i < cstate->num_mixers; i++) {
216                 if (!mixer[i].hw_lm || !mixer[i].lm_ctl) {
217                         DPU_ERROR("invalid lm or ctl assigned to mixer\n");
218                         return;
219                 }
220                 mixer[i].mixer_op_mode = 0;
221                 mixer[i].flush_mask = 0;
222                 if (mixer[i].lm_ctl->ops.clear_all_blendstages)
223                         mixer[i].lm_ctl->ops.clear_all_blendstages(
224                                         mixer[i].lm_ctl);
225         }
226
227         /* initialize stage cfg */
228         memset(&dpu_crtc->stage_cfg, 0, sizeof(struct dpu_hw_stage_cfg));
229
230         _dpu_crtc_blend_setup_mixer(crtc, dpu_crtc, mixer);
231
232         for (i = 0; i < cstate->num_mixers; i++) {
233                 ctl = mixer[i].lm_ctl;
234                 lm = mixer[i].hw_lm;
235
236                 lm->ops.setup_alpha_out(lm, mixer[i].mixer_op_mode);
237
238                 mixer[i].flush_mask |= ctl->ops.get_bitmask_mixer(ctl,
239                         mixer[i].hw_lm->idx);
240
241                 /* stage config flush mask */
242                 ctl->ops.update_pending_flush(ctl, mixer[i].flush_mask);
243
244                 DPU_DEBUG("lm %d, op_mode 0x%X, ctl %d, flush mask 0x%x\n",
245                         mixer[i].hw_lm->idx - LM_0,
246                         mixer[i].mixer_op_mode,
247                         ctl->idx - CTL_0,
248                         mixer[i].flush_mask);
249
250                 ctl->ops.setup_blendstage(ctl, mixer[i].hw_lm->idx,
251                         &dpu_crtc->stage_cfg);
252         }
253 }
254
255 /**
256  *  _dpu_crtc_complete_flip - signal pending page_flip events
257  * Any pending vblank events are added to the vblank_event_list
258  * so that the next vblank interrupt shall signal them.
259  * However PAGE_FLIP events are not handled through the vblank_event_list.
260  * This API signals any pending PAGE_FLIP events requested through
261  * DRM_IOCTL_MODE_PAGE_FLIP and are cached in the dpu_crtc->event.
262  * @crtc: Pointer to drm crtc structure
263  */
264 static void _dpu_crtc_complete_flip(struct drm_crtc *crtc)
265 {
266         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
267         struct drm_device *dev = crtc->dev;
268         unsigned long flags;
269
270         spin_lock_irqsave(&dev->event_lock, flags);
271         if (dpu_crtc->event) {
272                 DRM_DEBUG_VBL("%s: send event: %pK\n", dpu_crtc->name,
273                               dpu_crtc->event);
274                 trace_dpu_crtc_complete_flip(DRMID(crtc));
275                 drm_crtc_send_vblank_event(crtc, dpu_crtc->event);
276                 dpu_crtc->event = NULL;
277         }
278         spin_unlock_irqrestore(&dev->event_lock, flags);
279 }
280
281 enum dpu_intf_mode dpu_crtc_get_intf_mode(struct drm_crtc *crtc)
282 {
283         struct drm_encoder *encoder;
284
285         if (!crtc || !crtc->dev) {
286                 DPU_ERROR("invalid crtc\n");
287                 return INTF_MODE_NONE;
288         }
289
290         drm_for_each_encoder(encoder, crtc->dev)
291                 if (encoder->crtc == crtc)
292                         return dpu_encoder_get_intf_mode(encoder);
293
294         return INTF_MODE_NONE;
295 }
296
297 static void dpu_crtc_vblank_cb(void *data)
298 {
299         struct drm_crtc *crtc = (struct drm_crtc *)data;
300         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
301
302         /* keep statistics on vblank callback - with auto reset via debugfs */
303         if (ktime_compare(dpu_crtc->vblank_cb_time, ktime_set(0, 0)) == 0)
304                 dpu_crtc->vblank_cb_time = ktime_get();
305         else
306                 dpu_crtc->vblank_cb_count++;
307         _dpu_crtc_complete_flip(crtc);
308         drm_crtc_handle_vblank(crtc);
309         trace_dpu_crtc_vblank_cb(DRMID(crtc));
310 }
311
312 static void dpu_crtc_frame_event_work(struct kthread_work *work)
313 {
314         struct dpu_crtc_frame_event *fevent = container_of(work,
315                         struct dpu_crtc_frame_event, work);
316         struct drm_crtc *crtc = fevent->crtc;
317         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
318         unsigned long flags;
319         bool frame_done = false;
320
321         DPU_ATRACE_BEGIN("crtc_frame_event");
322
323         DRM_DEBUG_KMS("crtc%d event:%u ts:%lld\n", crtc->base.id, fevent->event,
324                         ktime_to_ns(fevent->ts));
325
326         if (fevent->event & (DPU_ENCODER_FRAME_EVENT_DONE
327                                 | DPU_ENCODER_FRAME_EVENT_ERROR
328                                 | DPU_ENCODER_FRAME_EVENT_PANEL_DEAD)) {
329
330                 if (atomic_read(&dpu_crtc->frame_pending) < 1) {
331                         /* this should not happen */
332                         DRM_ERROR("crtc%d ev:%u ts:%lld frame_pending:%d\n",
333                                         crtc->base.id,
334                                         fevent->event,
335                                         ktime_to_ns(fevent->ts),
336                                         atomic_read(&dpu_crtc->frame_pending));
337                 } else if (atomic_dec_return(&dpu_crtc->frame_pending) == 0) {
338                         /* release bandwidth and other resources */
339                         trace_dpu_crtc_frame_event_done(DRMID(crtc),
340                                                         fevent->event);
341                         dpu_core_perf_crtc_release_bw(crtc);
342                 } else {
343                         trace_dpu_crtc_frame_event_more_pending(DRMID(crtc),
344                                                                 fevent->event);
345                 }
346
347                 if (fevent->event & DPU_ENCODER_FRAME_EVENT_DONE)
348                         dpu_core_perf_crtc_update(crtc, 0, false);
349
350                 if (fevent->event & (DPU_ENCODER_FRAME_EVENT_DONE
351                                         | DPU_ENCODER_FRAME_EVENT_ERROR))
352                         frame_done = true;
353         }
354
355         if (fevent->event & DPU_ENCODER_FRAME_EVENT_PANEL_DEAD)
356                 DPU_ERROR("crtc%d ts:%lld received panel dead event\n",
357                                 crtc->base.id, ktime_to_ns(fevent->ts));
358
359         if (frame_done)
360                 complete_all(&dpu_crtc->frame_done_comp);
361
362         spin_lock_irqsave(&dpu_crtc->spin_lock, flags);
363         list_add_tail(&fevent->list, &dpu_crtc->frame_event_list);
364         spin_unlock_irqrestore(&dpu_crtc->spin_lock, flags);
365         DPU_ATRACE_END("crtc_frame_event");
366 }
367
368 /*
369  * dpu_crtc_frame_event_cb - crtc frame event callback API. CRTC module
370  * registers this API to encoder for all frame event callbacks like
371  * frame_error, frame_done, idle_timeout, etc. Encoder may call different events
372  * from different context - IRQ, user thread, commit_thread, etc. Each event
373  * should be carefully reviewed and should be processed in proper task context
374  * to avoid schedulin delay or properly manage the irq context's bottom half
375  * processing.
376  */
377 static void dpu_crtc_frame_event_cb(void *data, u32 event)
378 {
379         struct drm_crtc *crtc = (struct drm_crtc *)data;
380         struct dpu_crtc *dpu_crtc;
381         struct msm_drm_private *priv;
382         struct dpu_crtc_frame_event *fevent;
383         unsigned long flags;
384         u32 crtc_id;
385
386         /* Nothing to do on idle event */
387         if (event & DPU_ENCODER_FRAME_EVENT_IDLE)
388                 return;
389
390         dpu_crtc = to_dpu_crtc(crtc);
391         priv = crtc->dev->dev_private;
392         crtc_id = drm_crtc_index(crtc);
393
394         trace_dpu_crtc_frame_event_cb(DRMID(crtc), event);
395
396         spin_lock_irqsave(&dpu_crtc->spin_lock, flags);
397         fevent = list_first_entry_or_null(&dpu_crtc->frame_event_list,
398                         struct dpu_crtc_frame_event, list);
399         if (fevent)
400                 list_del_init(&fevent->list);
401         spin_unlock_irqrestore(&dpu_crtc->spin_lock, flags);
402
403         if (!fevent) {
404                 DRM_ERROR("crtc%d event %d overflow\n", crtc->base.id, event);
405                 return;
406         }
407
408         fevent->event = event;
409         fevent->crtc = crtc;
410         fevent->ts = ktime_get();
411         kthread_queue_work(&priv->event_thread[crtc_id].worker, &fevent->work);
412 }
413
414 void dpu_crtc_complete_commit(struct drm_crtc *crtc,
415                 struct drm_crtc_state *old_state)
416 {
417         if (!crtc || !crtc->state) {
418                 DPU_ERROR("invalid crtc\n");
419                 return;
420         }
421         trace_dpu_crtc_complete_commit(DRMID(crtc));
422 }
423
424 static void _dpu_crtc_setup_mixer_for_encoder(
425                 struct drm_crtc *crtc,
426                 struct drm_encoder *enc)
427 {
428         struct dpu_crtc_state *cstate = to_dpu_crtc_state(crtc->state);
429         struct dpu_kms *dpu_kms = _dpu_crtc_get_kms(crtc);
430         struct dpu_rm *rm = &dpu_kms->rm;
431         struct dpu_crtc_mixer *mixer;
432         struct dpu_hw_ctl *last_valid_ctl = NULL;
433         int i;
434         struct dpu_rm_hw_iter lm_iter, ctl_iter;
435
436         dpu_rm_init_hw_iter(&lm_iter, enc->base.id, DPU_HW_BLK_LM);
437         dpu_rm_init_hw_iter(&ctl_iter, enc->base.id, DPU_HW_BLK_CTL);
438
439         /* Set up all the mixers and ctls reserved by this encoder */
440         for (i = cstate->num_mixers; i < ARRAY_SIZE(cstate->mixers); i++) {
441                 mixer = &cstate->mixers[i];
442
443                 if (!dpu_rm_get_hw(rm, &lm_iter))
444                         break;
445                 mixer->hw_lm = (struct dpu_hw_mixer *)lm_iter.hw;
446
447                 /* CTL may be <= LMs, if <, multiple LMs controlled by 1 CTL */
448                 if (!dpu_rm_get_hw(rm, &ctl_iter)) {
449                         DPU_DEBUG("no ctl assigned to lm %d, using previous\n",
450                                         mixer->hw_lm->idx - LM_0);
451                         mixer->lm_ctl = last_valid_ctl;
452                 } else {
453                         mixer->lm_ctl = (struct dpu_hw_ctl *)ctl_iter.hw;
454                         last_valid_ctl = mixer->lm_ctl;
455                 }
456
457                 /* Shouldn't happen, mixers are always >= ctls */
458                 if (!mixer->lm_ctl) {
459                         DPU_ERROR("no valid ctls found for lm %d\n",
460                                         mixer->hw_lm->idx - LM_0);
461                         return;
462                 }
463
464                 mixer->encoder = enc;
465
466                 cstate->num_mixers++;
467                 DPU_DEBUG("setup mixer %d: lm %d\n",
468                                 i, mixer->hw_lm->idx - LM_0);
469                 DPU_DEBUG("setup mixer %d: ctl %d\n",
470                                 i, mixer->lm_ctl->idx - CTL_0);
471         }
472 }
473
474 static void _dpu_crtc_setup_mixers(struct drm_crtc *crtc)
475 {
476         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
477         struct drm_encoder *enc;
478
479         mutex_lock(&dpu_crtc->crtc_lock);
480         /* Check for mixers on all encoders attached to this crtc */
481         list_for_each_entry(enc, &crtc->dev->mode_config.encoder_list, head) {
482                 if (enc->crtc != crtc)
483                         continue;
484
485                 _dpu_crtc_setup_mixer_for_encoder(crtc, enc);
486         }
487
488         mutex_unlock(&dpu_crtc->crtc_lock);
489 }
490
491 static void _dpu_crtc_setup_lm_bounds(struct drm_crtc *crtc,
492                 struct drm_crtc_state *state)
493 {
494         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
495         struct dpu_crtc_state *cstate = to_dpu_crtc_state(state);
496         struct drm_display_mode *adj_mode = &state->adjusted_mode;
497         u32 crtc_split_width = _dpu_crtc_get_mixer_width(cstate, adj_mode);
498         int i;
499
500         for (i = 0; i < cstate->num_mixers; i++) {
501                 struct drm_rect *r = &cstate->lm_bounds[i];
502                 r->x1 = crtc_split_width * i;
503                 r->y1 = 0;
504                 r->x2 = r->x1 + crtc_split_width;
505                 r->y2 = dpu_crtc_get_mixer_height(dpu_crtc, cstate, adj_mode);
506
507                 trace_dpu_crtc_setup_lm_bounds(DRMID(crtc), i, r);
508         }
509
510         drm_mode_debug_printmodeline(adj_mode);
511 }
512
513 static void dpu_crtc_atomic_begin(struct drm_crtc *crtc,
514                 struct drm_crtc_state *old_state)
515 {
516         struct dpu_crtc *dpu_crtc;
517         struct dpu_crtc_state *cstate;
518         struct drm_encoder *encoder;
519         struct drm_device *dev;
520         unsigned long flags;
521         struct dpu_crtc_smmu_state_data *smmu_state;
522
523         if (!crtc) {
524                 DPU_ERROR("invalid crtc\n");
525                 return;
526         }
527
528         if (!crtc->state->enable) {
529                 DPU_DEBUG("crtc%d -> enable %d, skip atomic_begin\n",
530                                 crtc->base.id, crtc->state->enable);
531                 return;
532         }
533
534         DPU_DEBUG("crtc%d\n", crtc->base.id);
535
536         dpu_crtc = to_dpu_crtc(crtc);
537         cstate = to_dpu_crtc_state(crtc->state);
538         dev = crtc->dev;
539         smmu_state = &dpu_crtc->smmu_state;
540
541         if (!cstate->num_mixers) {
542                 _dpu_crtc_setup_mixers(crtc);
543                 _dpu_crtc_setup_lm_bounds(crtc, crtc->state);
544         }
545
546         if (dpu_crtc->event) {
547                 WARN_ON(dpu_crtc->event);
548         } else {
549                 spin_lock_irqsave(&dev->event_lock, flags);
550                 dpu_crtc->event = crtc->state->event;
551                 crtc->state->event = NULL;
552                 spin_unlock_irqrestore(&dev->event_lock, flags);
553         }
554
555         list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
556                 if (encoder->crtc != crtc)
557                         continue;
558
559                 /* encoder will trigger pending mask now */
560                 dpu_encoder_trigger_kickoff_pending(encoder);
561         }
562
563         /*
564          * If no mixers have been allocated in dpu_crtc_atomic_check(),
565          * it means we are trying to flush a CRTC whose state is disabled:
566          * nothing else needs to be done.
567          */
568         if (unlikely(!cstate->num_mixers))
569                 return;
570
571         _dpu_crtc_blend_setup(crtc);
572
573         /*
574          * PP_DONE irq is only used by command mode for now.
575          * It is better to request pending before FLUSH and START trigger
576          * to make sure no pp_done irq missed.
577          * This is safe because no pp_done will happen before SW trigger
578          * in command mode.
579          */
580 }
581
582 static void dpu_crtc_atomic_flush(struct drm_crtc *crtc,
583                 struct drm_crtc_state *old_crtc_state)
584 {
585         struct dpu_crtc *dpu_crtc;
586         struct drm_device *dev;
587         struct drm_plane *plane;
588         struct msm_drm_private *priv;
589         struct msm_drm_thread *event_thread;
590         unsigned long flags;
591         struct dpu_crtc_state *cstate;
592
593         if (!crtc->state->enable) {
594                 DPU_DEBUG("crtc%d -> enable %d, skip atomic_flush\n",
595                                 crtc->base.id, crtc->state->enable);
596                 return;
597         }
598
599         DPU_DEBUG("crtc%d\n", crtc->base.id);
600
601         dpu_crtc = to_dpu_crtc(crtc);
602         cstate = to_dpu_crtc_state(crtc->state);
603         dev = crtc->dev;
604         priv = dev->dev_private;
605
606         if (crtc->index >= ARRAY_SIZE(priv->event_thread)) {
607                 DPU_ERROR("invalid crtc index[%d]\n", crtc->index);
608                 return;
609         }
610
611         event_thread = &priv->event_thread[crtc->index];
612
613         if (dpu_crtc->event) {
614                 DPU_DEBUG("already received dpu_crtc->event\n");
615         } else {
616                 spin_lock_irqsave(&dev->event_lock, flags);
617                 dpu_crtc->event = crtc->state->event;
618                 crtc->state->event = NULL;
619                 spin_unlock_irqrestore(&dev->event_lock, flags);
620         }
621
622         /*
623          * If no mixers has been allocated in dpu_crtc_atomic_check(),
624          * it means we are trying to flush a CRTC whose state is disabled:
625          * nothing else needs to be done.
626          */
627         if (unlikely(!cstate->num_mixers))
628                 return;
629
630         /*
631          * For planes without commit update, drm framework will not add
632          * those planes to current state since hardware update is not
633          * required. However, if those planes were power collapsed since
634          * last commit cycle, driver has to restore the hardware state
635          * of those planes explicitly here prior to plane flush.
636          */
637         drm_atomic_crtc_for_each_plane(plane, crtc)
638                 dpu_plane_restore(plane);
639
640         /* update performance setting before crtc kickoff */
641         dpu_core_perf_crtc_update(crtc, 1, false);
642
643         /*
644          * Final plane updates: Give each plane a chance to complete all
645          *                      required writes/flushing before crtc's "flush
646          *                      everything" call below.
647          */
648         drm_atomic_crtc_for_each_plane(plane, crtc) {
649                 if (dpu_crtc->smmu_state.transition_error)
650                         dpu_plane_set_error(plane, true);
651                 dpu_plane_flush(plane);
652         }
653
654         /* Kickoff will be scheduled by outer layer */
655 }
656
657 /**
658  * dpu_crtc_destroy_state - state destroy hook
659  * @crtc: drm CRTC
660  * @state: CRTC state object to release
661  */
662 static void dpu_crtc_destroy_state(struct drm_crtc *crtc,
663                 struct drm_crtc_state *state)
664 {
665         struct dpu_crtc *dpu_crtc;
666         struct dpu_crtc_state *cstate;
667
668         if (!crtc || !state) {
669                 DPU_ERROR("invalid argument(s)\n");
670                 return;
671         }
672
673         dpu_crtc = to_dpu_crtc(crtc);
674         cstate = to_dpu_crtc_state(state);
675
676         DPU_DEBUG("crtc%d\n", crtc->base.id);
677
678         __drm_atomic_helper_crtc_destroy_state(state);
679
680         kfree(cstate);
681 }
682
683 static int _dpu_crtc_wait_for_frame_done(struct drm_crtc *crtc)
684 {
685         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
686         int ret, rc = 0;
687
688         if (!atomic_read(&dpu_crtc->frame_pending)) {
689                 DPU_DEBUG("no frames pending\n");
690                 return 0;
691         }
692
693         DPU_ATRACE_BEGIN("frame done completion wait");
694         ret = wait_for_completion_timeout(&dpu_crtc->frame_done_comp,
695                         msecs_to_jiffies(DPU_FRAME_DONE_TIMEOUT));
696         if (!ret) {
697                 DRM_ERROR("frame done wait timed out, ret:%d\n", ret);
698                 rc = -ETIMEDOUT;
699         }
700         DPU_ATRACE_END("frame done completion wait");
701
702         return rc;
703 }
704
705 void dpu_crtc_commit_kickoff(struct drm_crtc *crtc)
706 {
707         struct drm_encoder *encoder;
708         struct drm_device *dev = crtc->dev;
709         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
710         struct dpu_kms *dpu_kms = _dpu_crtc_get_kms(crtc);
711         struct dpu_crtc_state *cstate = to_dpu_crtc_state(crtc->state);
712         int ret;
713
714         /*
715          * If no mixers has been allocated in dpu_crtc_atomic_check(),
716          * it means we are trying to start a CRTC whose state is disabled:
717          * nothing else needs to be done.
718          */
719         if (unlikely(!cstate->num_mixers))
720                 return;
721
722         DPU_ATRACE_BEGIN("crtc_commit");
723
724         list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
725                 struct dpu_encoder_kickoff_params params = { 0 };
726
727                 if (encoder->crtc != crtc)
728                         continue;
729
730                 /*
731                  * Encoder will flush/start now, unless it has a tx pending.
732                  * If so, it may delay and flush at an irq event (e.g. ppdone)
733                  */
734                 dpu_encoder_prepare_for_kickoff(encoder, &params);
735         }
736
737         /* wait for frame_event_done completion */
738         DPU_ATRACE_BEGIN("wait_for_frame_done_event");
739         ret = _dpu_crtc_wait_for_frame_done(crtc);
740         DPU_ATRACE_END("wait_for_frame_done_event");
741         if (ret) {
742                 DPU_ERROR("crtc%d wait for frame done failed;frame_pending%d\n",
743                                 crtc->base.id,
744                                 atomic_read(&dpu_crtc->frame_pending));
745                 goto end;
746         }
747
748         if (atomic_inc_return(&dpu_crtc->frame_pending) == 1) {
749                 /* acquire bandwidth and other resources */
750                 DPU_DEBUG("crtc%d first commit\n", crtc->base.id);
751         } else
752                 DPU_DEBUG("crtc%d commit\n", crtc->base.id);
753
754         dpu_crtc->play_count++;
755
756         dpu_vbif_clear_errors(dpu_kms);
757
758         list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) {
759                 if (encoder->crtc != crtc)
760                         continue;
761
762                 dpu_encoder_kickoff(encoder);
763         }
764
765 end:
766         reinit_completion(&dpu_crtc->frame_done_comp);
767         DPU_ATRACE_END("crtc_commit");
768 }
769
770 /**
771  * _dpu_crtc_vblank_enable_no_lock - update power resource and vblank request
772  * @dpu_crtc: Pointer to dpu crtc structure
773  * @enable: Whether to enable/disable vblanks
774  */
775 static void _dpu_crtc_vblank_enable_no_lock(
776                 struct dpu_crtc *dpu_crtc, bool enable)
777 {
778         struct drm_crtc *crtc = &dpu_crtc->base;
779         struct drm_device *dev = crtc->dev;
780         struct drm_encoder *enc;
781
782         if (enable) {
783                 /* drop lock since power crtc cb may try to re-acquire lock */
784                 mutex_unlock(&dpu_crtc->crtc_lock);
785                 pm_runtime_get_sync(dev->dev);
786                 mutex_lock(&dpu_crtc->crtc_lock);
787
788                 list_for_each_entry(enc, &dev->mode_config.encoder_list, head) {
789                         if (enc->crtc != crtc)
790                                 continue;
791
792                         trace_dpu_crtc_vblank_enable(DRMID(&dpu_crtc->base),
793                                                      DRMID(enc), enable,
794                                                      dpu_crtc);
795
796                         dpu_encoder_register_vblank_callback(enc,
797                                         dpu_crtc_vblank_cb, (void *)crtc);
798                 }
799         } else {
800                 list_for_each_entry(enc, &dev->mode_config.encoder_list, head) {
801                         if (enc->crtc != crtc)
802                                 continue;
803
804                         trace_dpu_crtc_vblank_enable(DRMID(&dpu_crtc->base),
805                                                      DRMID(enc), enable,
806                                                      dpu_crtc);
807
808                         dpu_encoder_register_vblank_callback(enc, NULL, NULL);
809                 }
810
811                 /* drop lock since power crtc cb may try to re-acquire lock */
812                 mutex_unlock(&dpu_crtc->crtc_lock);
813                 pm_runtime_put_sync(dev->dev);
814                 mutex_lock(&dpu_crtc->crtc_lock);
815         }
816 }
817
818 /**
819  * _dpu_crtc_set_suspend - notify crtc of suspend enable/disable
820  * @crtc: Pointer to drm crtc object
821  * @enable: true to enable suspend, false to indicate resume
822  */
823 static void _dpu_crtc_set_suspend(struct drm_crtc *crtc, bool enable)
824 {
825         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
826
827         DRM_DEBUG_KMS("crtc%d suspend = %d\n", crtc->base.id, enable);
828
829         mutex_lock(&dpu_crtc->crtc_lock);
830
831         /*
832          * If the vblank is enabled, release a power reference on suspend
833          * and take it back during resume (if it is still enabled).
834          */
835         trace_dpu_crtc_set_suspend(DRMID(&dpu_crtc->base), enable, dpu_crtc);
836         if (dpu_crtc->suspend == enable)
837                 DPU_DEBUG("crtc%d suspend already set to %d, ignoring update\n",
838                                 crtc->base.id, enable);
839         else if (dpu_crtc->enabled && dpu_crtc->vblank_requested) {
840                 _dpu_crtc_vblank_enable_no_lock(dpu_crtc, !enable);
841         }
842
843         dpu_crtc->suspend = enable;
844         mutex_unlock(&dpu_crtc->crtc_lock);
845 }
846
847 /**
848  * dpu_crtc_duplicate_state - state duplicate hook
849  * @crtc: Pointer to drm crtc structure
850  * @Returns: Pointer to new drm_crtc_state structure
851  */
852 static struct drm_crtc_state *dpu_crtc_duplicate_state(struct drm_crtc *crtc)
853 {
854         struct dpu_crtc *dpu_crtc;
855         struct dpu_crtc_state *cstate, *old_cstate;
856
857         if (!crtc || !crtc->state) {
858                 DPU_ERROR("invalid argument(s)\n");
859                 return NULL;
860         }
861
862         dpu_crtc = to_dpu_crtc(crtc);
863         old_cstate = to_dpu_crtc_state(crtc->state);
864         cstate = kmemdup(old_cstate, sizeof(*old_cstate), GFP_KERNEL);
865         if (!cstate) {
866                 DPU_ERROR("failed to allocate state\n");
867                 return NULL;
868         }
869
870         /* duplicate base helper */
871         __drm_atomic_helper_crtc_duplicate_state(crtc, &cstate->base);
872
873         return &cstate->base;
874 }
875
876 /**
877  * dpu_crtc_reset - reset hook for CRTCs
878  * Resets the atomic state for @crtc by freeing the state pointer (which might
879  * be NULL, e.g. at driver load time) and allocating a new empty state object.
880  * @crtc: Pointer to drm crtc structure
881  */
882 static void dpu_crtc_reset(struct drm_crtc *crtc)
883 {
884         struct dpu_crtc *dpu_crtc;
885         struct dpu_crtc_state *cstate;
886
887         if (!crtc) {
888                 DPU_ERROR("invalid crtc\n");
889                 return;
890         }
891
892         /* revert suspend actions, if necessary */
893         if (dpu_kms_is_suspend_state(crtc->dev))
894                 _dpu_crtc_set_suspend(crtc, false);
895
896         /* remove previous state, if present */
897         if (crtc->state) {
898                 dpu_crtc_destroy_state(crtc, crtc->state);
899                 crtc->state = 0;
900         }
901
902         dpu_crtc = to_dpu_crtc(crtc);
903         cstate = kzalloc(sizeof(*cstate), GFP_KERNEL);
904         if (!cstate) {
905                 DPU_ERROR("failed to allocate state\n");
906                 return;
907         }
908
909         cstate->base.crtc = crtc;
910         crtc->state = &cstate->base;
911 }
912
913 static void dpu_crtc_handle_power_event(u32 event_type, void *arg)
914 {
915         struct drm_crtc *crtc = arg;
916         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
917         struct drm_encoder *encoder;
918
919         mutex_lock(&dpu_crtc->crtc_lock);
920
921         trace_dpu_crtc_handle_power_event(DRMID(crtc), event_type);
922
923         /* restore encoder; crtc will be programmed during commit */
924         drm_for_each_encoder(encoder, crtc->dev) {
925                 if (encoder->crtc != crtc)
926                         continue;
927
928                 dpu_encoder_virt_restore(encoder);
929         }
930
931         mutex_unlock(&dpu_crtc->crtc_lock);
932 }
933
934 static void dpu_crtc_disable(struct drm_crtc *crtc)
935 {
936         struct dpu_crtc *dpu_crtc;
937         struct dpu_crtc_state *cstate;
938         struct drm_display_mode *mode;
939         struct drm_encoder *encoder;
940         struct msm_drm_private *priv;
941         unsigned long flags;
942
943         if (!crtc || !crtc->dev || !crtc->dev->dev_private || !crtc->state) {
944                 DPU_ERROR("invalid crtc\n");
945                 return;
946         }
947         dpu_crtc = to_dpu_crtc(crtc);
948         cstate = to_dpu_crtc_state(crtc->state);
949         mode = &cstate->base.adjusted_mode;
950         priv = crtc->dev->dev_private;
951
952         DRM_DEBUG_KMS("crtc%d\n", crtc->base.id);
953
954         if (dpu_kms_is_suspend_state(crtc->dev))
955                 _dpu_crtc_set_suspend(crtc, true);
956
957         /* Disable/save vblank irq handling */
958         drm_crtc_vblank_off(crtc);
959
960         mutex_lock(&dpu_crtc->crtc_lock);
961
962         /* wait for frame_event_done completion */
963         if (_dpu_crtc_wait_for_frame_done(crtc))
964                 DPU_ERROR("crtc%d wait for frame done failed;frame_pending%d\n",
965                                 crtc->base.id,
966                                 atomic_read(&dpu_crtc->frame_pending));
967
968         trace_dpu_crtc_disable(DRMID(crtc), false, dpu_crtc);
969         if (dpu_crtc->enabled && !dpu_crtc->suspend &&
970                         dpu_crtc->vblank_requested) {
971                 _dpu_crtc_vblank_enable_no_lock(dpu_crtc, false);
972         }
973         dpu_crtc->enabled = false;
974
975         if (atomic_read(&dpu_crtc->frame_pending)) {
976                 trace_dpu_crtc_disable_frame_pending(DRMID(crtc),
977                                      atomic_read(&dpu_crtc->frame_pending));
978                 dpu_core_perf_crtc_release_bw(crtc);
979                 atomic_set(&dpu_crtc->frame_pending, 0);
980         }
981
982         dpu_core_perf_crtc_update(crtc, 0, true);
983
984         drm_for_each_encoder(encoder, crtc->dev) {
985                 if (encoder->crtc != crtc)
986                         continue;
987                 dpu_encoder_register_frame_event_callback(encoder, NULL, NULL);
988         }
989
990         if (dpu_crtc->power_event)
991                 dpu_power_handle_unregister_event(dpu_crtc->phandle,
992                                 dpu_crtc->power_event);
993
994         memset(cstate->mixers, 0, sizeof(cstate->mixers));
995         cstate->num_mixers = 0;
996
997         /* disable clk & bw control until clk & bw properties are set */
998         cstate->bw_control = false;
999         cstate->bw_split_vote = false;
1000
1001         mutex_unlock(&dpu_crtc->crtc_lock);
1002
1003         if (crtc->state->event && !crtc->state->active) {
1004                 spin_lock_irqsave(&crtc->dev->event_lock, flags);
1005                 drm_crtc_send_vblank_event(crtc, crtc->state->event);
1006                 crtc->state->event = NULL;
1007                 spin_unlock_irqrestore(&crtc->dev->event_lock, flags);
1008         }
1009 }
1010
1011 static void dpu_crtc_enable(struct drm_crtc *crtc,
1012                 struct drm_crtc_state *old_crtc_state)
1013 {
1014         struct dpu_crtc *dpu_crtc;
1015         struct drm_encoder *encoder;
1016         struct msm_drm_private *priv;
1017
1018         if (!crtc || !crtc->dev || !crtc->dev->dev_private) {
1019                 DPU_ERROR("invalid crtc\n");
1020                 return;
1021         }
1022         priv = crtc->dev->dev_private;
1023
1024         DRM_DEBUG_KMS("crtc%d\n", crtc->base.id);
1025         dpu_crtc = to_dpu_crtc(crtc);
1026
1027         drm_for_each_encoder(encoder, crtc->dev) {
1028                 if (encoder->crtc != crtc)
1029                         continue;
1030                 dpu_encoder_register_frame_event_callback(encoder,
1031                                 dpu_crtc_frame_event_cb, (void *)crtc);
1032         }
1033
1034         mutex_lock(&dpu_crtc->crtc_lock);
1035         trace_dpu_crtc_enable(DRMID(crtc), true, dpu_crtc);
1036         if (!dpu_crtc->enabled && !dpu_crtc->suspend &&
1037                         dpu_crtc->vblank_requested) {
1038                 _dpu_crtc_vblank_enable_no_lock(dpu_crtc, true);
1039         }
1040         dpu_crtc->enabled = true;
1041
1042         mutex_unlock(&dpu_crtc->crtc_lock);
1043
1044         /* Enable/restore vblank irq handling */
1045         drm_crtc_vblank_on(crtc);
1046
1047         dpu_crtc->power_event = dpu_power_handle_register_event(
1048                 dpu_crtc->phandle, DPU_POWER_EVENT_ENABLE,
1049                 dpu_crtc_handle_power_event, crtc, dpu_crtc->name);
1050
1051 }
1052
1053 struct plane_state {
1054         struct dpu_plane_state *dpu_pstate;
1055         const struct drm_plane_state *drm_pstate;
1056         int stage;
1057         u32 pipe_id;
1058 };
1059
1060 static int dpu_crtc_atomic_check(struct drm_crtc *crtc,
1061                 struct drm_crtc_state *state)
1062 {
1063         struct dpu_crtc *dpu_crtc;
1064         struct plane_state *pstates;
1065         struct dpu_crtc_state *cstate;
1066
1067         const struct drm_plane_state *pstate;
1068         struct drm_plane *plane;
1069         struct drm_display_mode *mode;
1070
1071         int cnt = 0, rc = 0, mixer_width, i, z_pos;
1072
1073         struct dpu_multirect_plane_states multirect_plane[DPU_STAGE_MAX * 2];
1074         int multirect_count = 0;
1075         const struct drm_plane_state *pipe_staged[SSPP_MAX];
1076         int left_zpos_cnt = 0, right_zpos_cnt = 0;
1077         struct drm_rect crtc_rect = { 0 };
1078
1079         if (!crtc) {
1080                 DPU_ERROR("invalid crtc\n");
1081                 return -EINVAL;
1082         }
1083
1084         pstates = kzalloc(sizeof(*pstates) * DPU_STAGE_MAX * 4, GFP_KERNEL);
1085
1086         dpu_crtc = to_dpu_crtc(crtc);
1087         cstate = to_dpu_crtc_state(state);
1088
1089         if (!state->enable || !state->active) {
1090                 DPU_DEBUG("crtc%d -> enable %d, active %d, skip atomic_check\n",
1091                                 crtc->base.id, state->enable, state->active);
1092                 goto end;
1093         }
1094
1095         mode = &state->adjusted_mode;
1096         DPU_DEBUG("%s: check", dpu_crtc->name);
1097
1098         /* force a full mode set if active state changed */
1099         if (state->active_changed)
1100                 state->mode_changed = true;
1101
1102         memset(pipe_staged, 0, sizeof(pipe_staged));
1103
1104         mixer_width = _dpu_crtc_get_mixer_width(cstate, mode);
1105
1106         _dpu_crtc_setup_lm_bounds(crtc, state);
1107
1108         crtc_rect.x2 = mode->hdisplay;
1109         crtc_rect.y2 = mode->vdisplay;
1110
1111          /* get plane state for all drm planes associated with crtc state */
1112         drm_atomic_crtc_state_for_each_plane_state(plane, pstate, state) {
1113                 struct drm_rect dst, clip = crtc_rect;
1114
1115                 if (IS_ERR_OR_NULL(pstate)) {
1116                         rc = PTR_ERR(pstate);
1117                         DPU_ERROR("%s: failed to get plane%d state, %d\n",
1118                                         dpu_crtc->name, plane->base.id, rc);
1119                         goto end;
1120                 }
1121                 if (cnt >= DPU_STAGE_MAX * 4)
1122                         continue;
1123
1124                 pstates[cnt].dpu_pstate = to_dpu_plane_state(pstate);
1125                 pstates[cnt].drm_pstate = pstate;
1126                 pstates[cnt].stage = pstate->normalized_zpos;
1127                 pstates[cnt].pipe_id = dpu_plane_pipe(plane);
1128
1129                 if (pipe_staged[pstates[cnt].pipe_id]) {
1130                         multirect_plane[multirect_count].r0 =
1131                                 pipe_staged[pstates[cnt].pipe_id];
1132                         multirect_plane[multirect_count].r1 = pstate;
1133                         multirect_count++;
1134
1135                         pipe_staged[pstates[cnt].pipe_id] = NULL;
1136                 } else {
1137                         pipe_staged[pstates[cnt].pipe_id] = pstate;
1138                 }
1139
1140                 cnt++;
1141
1142                 dst = drm_plane_state_dest(pstate);
1143                 if (!drm_rect_intersect(&clip, &dst)) {
1144                         DPU_ERROR("invalid vertical/horizontal destination\n");
1145                         DPU_ERROR("display: " DRM_RECT_FMT " plane: "
1146                                   DRM_RECT_FMT "\n", DRM_RECT_ARG(&crtc_rect),
1147                                   DRM_RECT_ARG(&dst));
1148                         rc = -E2BIG;
1149                         goto end;
1150                 }
1151         }
1152
1153         for (i = 1; i < SSPP_MAX; i++) {
1154                 if (pipe_staged[i]) {
1155                         dpu_plane_clear_multirect(pipe_staged[i]);
1156
1157                         if (is_dpu_plane_virtual(pipe_staged[i]->plane)) {
1158                                 DPU_ERROR(
1159                                         "r1 only virt plane:%d not supported\n",
1160                                         pipe_staged[i]->plane->base.id);
1161                                 rc  = -EINVAL;
1162                                 goto end;
1163                         }
1164                 }
1165         }
1166
1167         z_pos = -1;
1168         for (i = 0; i < cnt; i++) {
1169                 /* reset counts at every new blend stage */
1170                 if (pstates[i].stage != z_pos) {
1171                         left_zpos_cnt = 0;
1172                         right_zpos_cnt = 0;
1173                         z_pos = pstates[i].stage;
1174                 }
1175
1176                 /* verify z_pos setting before using it */
1177                 if (z_pos >= DPU_STAGE_MAX - DPU_STAGE_0) {
1178                         DPU_ERROR("> %d plane stages assigned\n",
1179                                         DPU_STAGE_MAX - DPU_STAGE_0);
1180                         rc = -EINVAL;
1181                         goto end;
1182                 } else if (pstates[i].drm_pstate->crtc_x < mixer_width) {
1183                         if (left_zpos_cnt == 2) {
1184                                 DPU_ERROR("> 2 planes @ stage %d on left\n",
1185                                         z_pos);
1186                                 rc = -EINVAL;
1187                                 goto end;
1188                         }
1189                         left_zpos_cnt++;
1190
1191                 } else {
1192                         if (right_zpos_cnt == 2) {
1193                                 DPU_ERROR("> 2 planes @ stage %d on right\n",
1194                                         z_pos);
1195                                 rc = -EINVAL;
1196                                 goto end;
1197                         }
1198                         right_zpos_cnt++;
1199                 }
1200
1201                 pstates[i].dpu_pstate->stage = z_pos + DPU_STAGE_0;
1202                 DPU_DEBUG("%s: zpos %d", dpu_crtc->name, z_pos);
1203         }
1204
1205         for (i = 0; i < multirect_count; i++) {
1206                 if (dpu_plane_validate_multirect_v2(&multirect_plane[i])) {
1207                         DPU_ERROR(
1208                         "multirect validation failed for planes (%d - %d)\n",
1209                                         multirect_plane[i].r0->plane->base.id,
1210                                         multirect_plane[i].r1->plane->base.id);
1211                         rc = -EINVAL;
1212                         goto end;
1213                 }
1214         }
1215
1216         rc = dpu_core_perf_crtc_check(crtc, state);
1217         if (rc) {
1218                 DPU_ERROR("crtc%d failed performance check %d\n",
1219                                 crtc->base.id, rc);
1220                 goto end;
1221         }
1222
1223         /* validate source split:
1224          * use pstates sorted by stage to check planes on same stage
1225          * we assume that all pipes are in source split so its valid to compare
1226          * without taking into account left/right mixer placement
1227          */
1228         for (i = 1; i < cnt; i++) {
1229                 struct plane_state *prv_pstate, *cur_pstate;
1230                 struct drm_rect left_rect, right_rect;
1231                 int32_t left_pid, right_pid;
1232                 int32_t stage;
1233
1234                 prv_pstate = &pstates[i - 1];
1235                 cur_pstate = &pstates[i];
1236                 if (prv_pstate->stage != cur_pstate->stage)
1237                         continue;
1238
1239                 stage = cur_pstate->stage;
1240
1241                 left_pid = prv_pstate->dpu_pstate->base.plane->base.id;
1242                 left_rect = drm_plane_state_dest(prv_pstate->drm_pstate);
1243
1244                 right_pid = cur_pstate->dpu_pstate->base.plane->base.id;
1245                 right_rect = drm_plane_state_dest(cur_pstate->drm_pstate);
1246
1247                 if (right_rect.x1 < left_rect.x1) {
1248                         swap(left_pid, right_pid);
1249                         swap(left_rect, right_rect);
1250                 }
1251
1252                 /**
1253                  * - planes are enumerated in pipe-priority order such that
1254                  *   planes with lower drm_id must be left-most in a shared
1255                  *   blend-stage when using source split.
1256                  * - planes in source split must be contiguous in width
1257                  * - planes in source split must have same dest yoff and height
1258                  */
1259                 if (right_pid < left_pid) {
1260                         DPU_ERROR(
1261                                 "invalid src split cfg. priority mismatch. stage: %d left: %d right: %d\n",
1262                                 stage, left_pid, right_pid);
1263                         rc = -EINVAL;
1264                         goto end;
1265                 } else if (right_rect.x1 != drm_rect_width(&left_rect)) {
1266                         DPU_ERROR("non-contiguous coordinates for src split. "
1267                                   "stage: %d left: " DRM_RECT_FMT " right: "
1268                                   DRM_RECT_FMT "\n", stage,
1269                                   DRM_RECT_ARG(&left_rect),
1270                                   DRM_RECT_ARG(&right_rect));
1271                         rc = -EINVAL;
1272                         goto end;
1273                 } else if (left_rect.y1 != right_rect.y1 ||
1274                            drm_rect_height(&left_rect) != drm_rect_height(&right_rect)) {
1275                         DPU_ERROR("source split at stage: %d. invalid "
1276                                   "yoff/height: left: " DRM_RECT_FMT " right: "
1277                                   DRM_RECT_FMT "\n", stage,
1278                                   DRM_RECT_ARG(&left_rect),
1279                                   DRM_RECT_ARG(&right_rect));
1280                         rc = -EINVAL;
1281                         goto end;
1282                 }
1283         }
1284
1285 end:
1286         kfree(pstates);
1287         return rc;
1288 }
1289
1290 int dpu_crtc_vblank(struct drm_crtc *crtc, bool en)
1291 {
1292         struct dpu_crtc *dpu_crtc;
1293
1294         if (!crtc) {
1295                 DPU_ERROR("invalid crtc\n");
1296                 return -EINVAL;
1297         }
1298         dpu_crtc = to_dpu_crtc(crtc);
1299
1300         mutex_lock(&dpu_crtc->crtc_lock);
1301         trace_dpu_crtc_vblank(DRMID(&dpu_crtc->base), en, dpu_crtc);
1302         if (dpu_crtc->enabled && !dpu_crtc->suspend) {
1303                 _dpu_crtc_vblank_enable_no_lock(dpu_crtc, en);
1304         }
1305         dpu_crtc->vblank_requested = en;
1306         mutex_unlock(&dpu_crtc->crtc_lock);
1307
1308         return 0;
1309 }
1310
1311 #ifdef CONFIG_DEBUG_FS
1312 static int _dpu_debugfs_status_show(struct seq_file *s, void *data)
1313 {
1314         struct dpu_crtc *dpu_crtc;
1315         struct dpu_plane_state *pstate = NULL;
1316         struct dpu_crtc_mixer *m;
1317
1318         struct drm_crtc *crtc;
1319         struct drm_plane *plane;
1320         struct drm_display_mode *mode;
1321         struct drm_framebuffer *fb;
1322         struct drm_plane_state *state;
1323         struct dpu_crtc_state *cstate;
1324
1325         int i, out_width;
1326
1327         if (!s || !s->private)
1328                 return -EINVAL;
1329
1330         dpu_crtc = s->private;
1331         crtc = &dpu_crtc->base;
1332
1333         drm_modeset_lock_all(crtc->dev);
1334         cstate = to_dpu_crtc_state(crtc->state);
1335
1336         mutex_lock(&dpu_crtc->crtc_lock);
1337         mode = &crtc->state->adjusted_mode;
1338         out_width = _dpu_crtc_get_mixer_width(cstate, mode);
1339
1340         seq_printf(s, "crtc:%d width:%d height:%d\n", crtc->base.id,
1341                                 mode->hdisplay, mode->vdisplay);
1342
1343         seq_puts(s, "\n");
1344
1345         for (i = 0; i < cstate->num_mixers; ++i) {
1346                 m = &cstate->mixers[i];
1347                 if (!m->hw_lm)
1348                         seq_printf(s, "\tmixer[%d] has no lm\n", i);
1349                 else if (!m->lm_ctl)
1350                         seq_printf(s, "\tmixer[%d] has no ctl\n", i);
1351                 else
1352                         seq_printf(s, "\tmixer:%d ctl:%d width:%d height:%d\n",
1353                                 m->hw_lm->idx - LM_0, m->lm_ctl->idx - CTL_0,
1354                                 out_width, mode->vdisplay);
1355         }
1356
1357         seq_puts(s, "\n");
1358
1359         drm_atomic_crtc_for_each_plane(plane, crtc) {
1360                 pstate = to_dpu_plane_state(plane->state);
1361                 state = plane->state;
1362
1363                 if (!pstate || !state)
1364                         continue;
1365
1366                 seq_printf(s, "\tplane:%u stage:%d\n", plane->base.id,
1367                         pstate->stage);
1368
1369                 if (plane->state->fb) {
1370                         fb = plane->state->fb;
1371
1372                         seq_printf(s, "\tfb:%d image format:%4.4s wxh:%ux%u ",
1373                                 fb->base.id, (char *) &fb->format->format,
1374                                 fb->width, fb->height);
1375                         for (i = 0; i < ARRAY_SIZE(fb->format->cpp); ++i)
1376                                 seq_printf(s, "cpp[%d]:%u ",
1377                                                 i, fb->format->cpp[i]);
1378                         seq_puts(s, "\n\t");
1379
1380                         seq_printf(s, "modifier:%8llu ", fb->modifier);
1381                         seq_puts(s, "\n");
1382
1383                         seq_puts(s, "\t");
1384                         for (i = 0; i < ARRAY_SIZE(fb->pitches); i++)
1385                                 seq_printf(s, "pitches[%d]:%8u ", i,
1386                                                         fb->pitches[i]);
1387                         seq_puts(s, "\n");
1388
1389                         seq_puts(s, "\t");
1390                         for (i = 0; i < ARRAY_SIZE(fb->offsets); i++)
1391                                 seq_printf(s, "offsets[%d]:%8u ", i,
1392                                                         fb->offsets[i]);
1393                         seq_puts(s, "\n");
1394                 }
1395
1396                 seq_printf(s, "\tsrc_x:%4d src_y:%4d src_w:%4d src_h:%4d\n",
1397                         state->src_x, state->src_y, state->src_w, state->src_h);
1398
1399                 seq_printf(s, "\tdst x:%4d dst_y:%4d dst_w:%4d dst_h:%4d\n",
1400                         state->crtc_x, state->crtc_y, state->crtc_w,
1401                         state->crtc_h);
1402                 seq_printf(s, "\tmultirect: mode: %d index: %d\n",
1403                         pstate->multirect_mode, pstate->multirect_index);
1404
1405                 seq_puts(s, "\n");
1406         }
1407         if (dpu_crtc->vblank_cb_count) {
1408                 ktime_t diff = ktime_sub(ktime_get(), dpu_crtc->vblank_cb_time);
1409                 s64 diff_ms = ktime_to_ms(diff);
1410                 s64 fps = diff_ms ? div_s64(
1411                                 dpu_crtc->vblank_cb_count * 1000, diff_ms) : 0;
1412
1413                 seq_printf(s,
1414                         "vblank fps:%lld count:%u total:%llums total_framecount:%llu\n",
1415                                 fps, dpu_crtc->vblank_cb_count,
1416                                 ktime_to_ms(diff), dpu_crtc->play_count);
1417
1418                 /* reset time & count for next measurement */
1419                 dpu_crtc->vblank_cb_count = 0;
1420                 dpu_crtc->vblank_cb_time = ktime_set(0, 0);
1421         }
1422
1423         seq_printf(s, "vblank_enable:%d\n", dpu_crtc->vblank_requested);
1424
1425         mutex_unlock(&dpu_crtc->crtc_lock);
1426         drm_modeset_unlock_all(crtc->dev);
1427
1428         return 0;
1429 }
1430
1431 static int _dpu_debugfs_status_open(struct inode *inode, struct file *file)
1432 {
1433         return single_open(file, _dpu_debugfs_status_show, inode->i_private);
1434 }
1435
1436 #define DEFINE_DPU_DEBUGFS_SEQ_FOPS(__prefix)                          \
1437 static int __prefix ## _open(struct inode *inode, struct file *file)    \
1438 {                                                                       \
1439         return single_open(file, __prefix ## _show, inode->i_private);  \
1440 }                                                                       \
1441 static const struct file_operations __prefix ## _fops = {               \
1442         .owner = THIS_MODULE,                                           \
1443         .open = __prefix ## _open,                                      \
1444         .release = single_release,                                      \
1445         .read = seq_read,                                               \
1446         .llseek = seq_lseek,                                            \
1447 }
1448
1449 static int dpu_crtc_debugfs_state_show(struct seq_file *s, void *v)
1450 {
1451         struct drm_crtc *crtc = (struct drm_crtc *) s->private;
1452         struct dpu_crtc *dpu_crtc = to_dpu_crtc(crtc);
1453         int i;
1454
1455         seq_printf(s, "client type: %d\n", dpu_crtc_get_client_type(crtc));
1456         seq_printf(s, "intf_mode: %d\n", dpu_crtc_get_intf_mode(crtc));
1457         seq_printf(s, "core_clk_rate: %llu\n",
1458                         dpu_crtc->cur_perf.core_clk_rate);
1459         for (i = DPU_POWER_HANDLE_DBUS_ID_MNOC;
1460                         i < DPU_POWER_HANDLE_DBUS_ID_MAX; i++) {
1461                 seq_printf(s, "bw_ctl[%s]: %llu\n",
1462                                 dpu_power_handle_get_dbus_name(i),
1463                                 dpu_crtc->cur_perf.bw_ctl[i]);
1464                 seq_printf(s, "max_per_pipe_ib[%s]: %llu\n",
1465                                 dpu_power_handle_get_dbus_name(i),
1466                                 dpu_crtc->cur_perf.max_per_pipe_ib[i]);
1467         }
1468
1469         return 0;
1470 }
1471 DEFINE_DPU_DEBUGFS_SEQ_FOPS(dpu_crtc_debugfs_state);
1472
1473 static int _dpu_crtc_init_debugfs(struct drm_crtc *crtc)
1474 {
1475         struct dpu_crtc *dpu_crtc;
1476         struct dpu_kms *dpu_kms;
1477
1478         static const struct file_operations debugfs_status_fops = {
1479                 .open =         _dpu_debugfs_status_open,
1480                 .read =         seq_read,
1481                 .llseek =       seq_lseek,
1482                 .release =      single_release,
1483         };
1484
1485         if (!crtc)
1486                 return -EINVAL;
1487         dpu_crtc = to_dpu_crtc(crtc);
1488
1489         dpu_kms = _dpu_crtc_get_kms(crtc);
1490
1491         dpu_crtc->debugfs_root = debugfs_create_dir(dpu_crtc->name,
1492                         crtc->dev->primary->debugfs_root);
1493         if (!dpu_crtc->debugfs_root)
1494                 return -ENOMEM;
1495
1496         /* don't error check these */
1497         debugfs_create_file("status", 0400,
1498                         dpu_crtc->debugfs_root,
1499                         dpu_crtc, &debugfs_status_fops);
1500         debugfs_create_file("state", 0600,
1501                         dpu_crtc->debugfs_root,
1502                         &dpu_crtc->base,
1503                         &dpu_crtc_debugfs_state_fops);
1504
1505         return 0;
1506 }
1507
1508 static void _dpu_crtc_destroy_debugfs(struct drm_crtc *crtc)
1509 {
1510         struct dpu_crtc *dpu_crtc;
1511
1512         if (!crtc)
1513                 return;
1514         dpu_crtc = to_dpu_crtc(crtc);
1515         debugfs_remove_recursive(dpu_crtc->debugfs_root);
1516 }
1517 #else
1518 static int _dpu_crtc_init_debugfs(struct drm_crtc *crtc)
1519 {
1520         return 0;
1521 }
1522
1523 static void _dpu_crtc_destroy_debugfs(struct drm_crtc *crtc)
1524 {
1525 }
1526 #endif /* CONFIG_DEBUG_FS */
1527
1528 static int dpu_crtc_late_register(struct drm_crtc *crtc)
1529 {
1530         return _dpu_crtc_init_debugfs(crtc);
1531 }
1532
1533 static void dpu_crtc_early_unregister(struct drm_crtc *crtc)
1534 {
1535         _dpu_crtc_destroy_debugfs(crtc);
1536 }
1537
1538 static const struct drm_crtc_funcs dpu_crtc_funcs = {
1539         .set_config = drm_atomic_helper_set_config,
1540         .destroy = dpu_crtc_destroy,
1541         .page_flip = drm_atomic_helper_page_flip,
1542         .reset = dpu_crtc_reset,
1543         .atomic_duplicate_state = dpu_crtc_duplicate_state,
1544         .atomic_destroy_state = dpu_crtc_destroy_state,
1545         .late_register = dpu_crtc_late_register,
1546         .early_unregister = dpu_crtc_early_unregister,
1547 };
1548
1549 static const struct drm_crtc_helper_funcs dpu_crtc_helper_funcs = {
1550         .disable = dpu_crtc_disable,
1551         .atomic_enable = dpu_crtc_enable,
1552         .atomic_check = dpu_crtc_atomic_check,
1553         .atomic_begin = dpu_crtc_atomic_begin,
1554         .atomic_flush = dpu_crtc_atomic_flush,
1555 };
1556
1557 /* initialize crtc */
1558 struct drm_crtc *dpu_crtc_init(struct drm_device *dev, struct drm_plane *plane,
1559                                 struct drm_plane *cursor)
1560 {
1561         struct drm_crtc *crtc = NULL;
1562         struct dpu_crtc *dpu_crtc = NULL;
1563         struct msm_drm_private *priv = NULL;
1564         struct dpu_kms *kms = NULL;
1565         int i;
1566
1567         priv = dev->dev_private;
1568         kms = to_dpu_kms(priv->kms);
1569
1570         dpu_crtc = kzalloc(sizeof(*dpu_crtc), GFP_KERNEL);
1571         if (!dpu_crtc)
1572                 return ERR_PTR(-ENOMEM);
1573
1574         crtc = &dpu_crtc->base;
1575         crtc->dev = dev;
1576
1577         mutex_init(&dpu_crtc->crtc_lock);
1578         spin_lock_init(&dpu_crtc->spin_lock);
1579         atomic_set(&dpu_crtc->frame_pending, 0);
1580
1581         init_completion(&dpu_crtc->frame_done_comp);
1582
1583         INIT_LIST_HEAD(&dpu_crtc->frame_event_list);
1584
1585         for (i = 0; i < ARRAY_SIZE(dpu_crtc->frame_events); i++) {
1586                 INIT_LIST_HEAD(&dpu_crtc->frame_events[i].list);
1587                 list_add(&dpu_crtc->frame_events[i].list,
1588                                 &dpu_crtc->frame_event_list);
1589                 kthread_init_work(&dpu_crtc->frame_events[i].work,
1590                                 dpu_crtc_frame_event_work);
1591         }
1592
1593         drm_crtc_init_with_planes(dev, crtc, plane, cursor, &dpu_crtc_funcs,
1594                                 NULL);
1595
1596         drm_crtc_helper_add(crtc, &dpu_crtc_helper_funcs);
1597         plane->crtc = crtc;
1598
1599         /* save user friendly CRTC name for later */
1600         snprintf(dpu_crtc->name, DPU_CRTC_NAME_SIZE, "crtc%u", crtc->base.id);
1601
1602         /* initialize event handling */
1603         spin_lock_init(&dpu_crtc->event_lock);
1604
1605         dpu_crtc->phandle = &kms->phandle;
1606
1607         DPU_DEBUG("%s: successfully initialized crtc\n", dpu_crtc->name);
1608         return crtc;
1609 }