blob: 0dc96e12fd06dc17a9f6c992705a12a06c4df9c3 [file] [log] [blame]
Rob Clark16ea9752013-01-08 15:04:28 -06001/*
2 * Copyright (C) 2012 Texas Instruments
3 * Author: Rob Clark <robdclark@gmail.com>
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
12 * more details.
13 *
14 * You should have received a copy of the GNU General Public License along with
15 * this program. If not, see <http://www.gnu.org/licenses/>.
16 */
17
Sean Paulce2f2c32016-09-21 06:14:53 -070018#include <drm/drm_atomic.h>
Jyri Sarha305198d2016-04-07 15:05:16 +030019#include <drm/drm_atomic_helper.h>
Sean Paulce2f2c32016-09-21 06:14:53 -070020#include <drm/drm_crtc.h>
21#include <drm/drm_flip_work.h>
22#include <drm/drm_plane_helper.h>
Jyri Sarha4e910c72016-09-06 22:55:33 +030023#include <linux/workqueue.h>
Bartosz Golaszewski93452352016-10-31 15:19:26 +010024#include <linux/completion.h>
25#include <linux/dma-mapping.h>
Rob Clark16ea9752013-01-08 15:04:28 -060026
27#include "tilcdc_drv.h"
28#include "tilcdc_regs.h"
29
Bartosz Golaszewski93452352016-10-31 15:19:26 +010030#define TILCDC_VBLANK_SAFETY_THRESHOLD_US 1000
Jyri Sarha55e165c2016-11-15 23:37:24 +020031#define TILCDC_PALETTE_SIZE 32
32#define TILCDC_PALETTE_FIRST_ENTRY 0x4000
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020033
Rob Clark16ea9752013-01-08 15:04:28 -060034struct tilcdc_crtc {
35 struct drm_crtc base;
36
Jyri Sarha47f571c2016-04-07 15:04:18 +030037 struct drm_plane primary;
Rob Clark16ea9752013-01-08 15:04:28 -060038 const struct tilcdc_panel_info *info;
Rob Clark16ea9752013-01-08 15:04:28 -060039 struct drm_pending_vblank_event *event;
Jyri Sarha2d53a182016-10-25 12:27:31 +030040 struct mutex enable_lock;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +030041 bool enabled;
Jyri Sarha2d53a182016-10-25 12:27:31 +030042 bool shutdown;
Rob Clark16ea9752013-01-08 15:04:28 -060043 wait_queue_head_t frame_done_wq;
44 bool frame_done;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020045 spinlock_t irq_lock;
46
Jyri Sarha642e5162016-09-06 16:19:54 +030047 unsigned int lcd_fck_rate;
48
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020049 ktime_t last_vblank;
Rob Clark16ea9752013-01-08 15:04:28 -060050
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030051 struct drm_framebuffer *curr_fb;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020052 struct drm_framebuffer *next_fb;
Rob Clark16ea9752013-01-08 15:04:28 -060053
54 /* for deferred fb unref's: */
Rob Clarka464d612013-08-07 13:41:20 -040055 struct drm_flip_work unref_work;
Jyri Sarha103cd8b2015-02-10 14:13:23 +020056
57 /* Only set if an external encoder is connected */
58 bool simulate_vesa_sync;
Jyri Sarha5895d082016-01-08 14:33:09 +020059
60 int sync_lost_count;
61 bool frame_intact;
Jyri Sarha13b3d722016-04-06 14:02:38 +030062 struct work_struct recover_work;
Bartosz Golaszewski93452352016-10-31 15:19:26 +010063
64 dma_addr_t palette_dma_handle;
Jyri Sarha55e165c2016-11-15 23:37:24 +020065 u16 *palette_base;
Bartosz Golaszewski93452352016-10-31 15:19:26 +010066 struct completion palette_loaded;
Rob Clark16ea9752013-01-08 15:04:28 -060067};
68#define to_tilcdc_crtc(x) container_of(x, struct tilcdc_crtc, base)
69
Rob Clarka464d612013-08-07 13:41:20 -040070static void unref_worker(struct drm_flip_work *work, void *val)
Rob Clark16ea9752013-01-08 15:04:28 -060071{
Darren Etheridgef7b45752013-06-21 13:52:26 -050072 struct tilcdc_crtc *tilcdc_crtc =
Rob Clarka464d612013-08-07 13:41:20 -040073 container_of(work, struct tilcdc_crtc, unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -060074 struct drm_device *dev = tilcdc_crtc->base.dev;
Rob Clark16ea9752013-01-08 15:04:28 -060075
76 mutex_lock(&dev->mode_config.mutex);
Rob Clarka464d612013-08-07 13:41:20 -040077 drm_framebuffer_unreference(val);
Rob Clark16ea9752013-01-08 15:04:28 -060078 mutex_unlock(&dev->mode_config.mutex);
79}
80
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030081static void set_scanout(struct drm_crtc *crtc, struct drm_framebuffer *fb)
Rob Clark16ea9752013-01-08 15:04:28 -060082{
83 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
84 struct drm_device *dev = crtc->dev;
Daniel Schultz4c268d62016-10-28 13:52:41 +020085 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -060086 struct drm_gem_cma_object *gem;
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030087 dma_addr_t start, end;
Jyri Sarha7eb9f062016-08-26 15:10:14 +030088 u64 dma_base_and_ceiling;
Rob Clark16ea9752013-01-08 15:04:28 -060089
Rob Clark16ea9752013-01-08 15:04:28 -060090 gem = drm_fb_cma_get_gem_obj(fb, 0);
91
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030092 start = gem->paddr + fb->offsets[0] +
93 crtc->y * fb->pitches[0] +
Ville Syrjälä353c8592016-12-14 23:30:57 +020094 crtc->x * fb->format->cpp[0];
Rob Clark16ea9752013-01-08 15:04:28 -060095
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030096 end = start + (crtc->mode.vdisplay * fb->pitches[0]);
Rob Clark16ea9752013-01-08 15:04:28 -060097
Jyri Sarha7eb9f062016-08-26 15:10:14 +030098 /* Write LCDC_DMA_FB_BASE_ADDR_0_REG and LCDC_DMA_FB_CEILING_ADDR_0_REG
99 * with a single insruction, if available. This should make it more
100 * unlikely that LCDC would fetch the DMA addresses in the middle of
101 * an update.
102 */
Daniel Schultz4c268d62016-10-28 13:52:41 +0200103 if (priv->rev == 1)
104 end -= 1;
105
106 dma_base_and_ceiling = (u64)end << 32 | start;
Jyri Sarha7eb9f062016-08-26 15:10:14 +0300107 tilcdc_write64(dev, LCDC_DMA_FB_BASE_ADDR_0_REG, dma_base_and_ceiling);
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300108
109 if (tilcdc_crtc->curr_fb)
110 drm_flip_work_queue(&tilcdc_crtc->unref_work,
111 tilcdc_crtc->curr_fb);
112
113 tilcdc_crtc->curr_fb = fb;
Rob Clark16ea9752013-01-08 15:04:28 -0600114}
115
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100116/*
Jyri Sarha55e165c2016-11-15 23:37:24 +0200117 * The driver currently only supports only true color formats. For
118 * true color the palette block is bypassed, but a 32 byte palette
119 * should still be loaded. The first 16-bit entry must be 0x4000 while
120 * all other entries must be zeroed.
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100121 */
122static void tilcdc_crtc_load_palette(struct drm_crtc *crtc)
123{
Jyri Sarha55e165c2016-11-15 23:37:24 +0200124 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
125 struct drm_device *dev = crtc->dev;
126 struct tilcdc_drm_private *priv = dev->dev_private;
Jyri Sarhae59f5af2016-11-17 18:46:16 +0200127 int ret;
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100128
Jyri Sarha274c34d2016-11-15 23:57:42 +0200129 reinit_completion(&tilcdc_crtc->palette_loaded);
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100130
131 /* Tell the LCDC where the palette is located. */
132 tilcdc_write(dev, LCDC_DMA_FB_BASE_ADDR_0_REG,
133 tilcdc_crtc->palette_dma_handle);
134 tilcdc_write(dev, LCDC_DMA_FB_CEILING_ADDR_0_REG,
Jyri Sarha55e165c2016-11-15 23:37:24 +0200135 (u32) tilcdc_crtc->palette_dma_handle +
136 TILCDC_PALETTE_SIZE - 1);
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100137
Jyri Sarha55e165c2016-11-15 23:37:24 +0200138 /* Set dma load mode for palette loading only. */
139 tilcdc_write_mask(dev, LCDC_RASTER_CTRL_REG,
140 LCDC_PALETTE_LOAD_MODE(PALETTE_ONLY),
141 LCDC_PALETTE_LOAD_MODE_MASK);
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100142
Jyri Sarha55e165c2016-11-15 23:37:24 +0200143 /* Enable DMA Palette Loaded Interrupt */
144 if (priv->rev == 1)
145 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_V1_PL_INT_ENA);
146 else
147 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG, LCDC_V2_PL_INT_ENA);
148
149 /* Enable LCDC DMA and wait for palette to be loaded. */
150 tilcdc_clear_irqstatus(dev, 0xffffffff);
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100151 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
152
Jyri Sarhae59f5af2016-11-17 18:46:16 +0200153 ret = wait_for_completion_timeout(&tilcdc_crtc->palette_loaded,
154 msecs_to_jiffies(50));
155 if (ret == 0)
156 dev_err(dev->dev, "%s: Palette loading timeout", __func__);
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100157
Jyri Sarha55e165c2016-11-15 23:37:24 +0200158 /* Disable LCDC DMA and DMA Palette Loaded Interrupt. */
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100159 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarha55e165c2016-11-15 23:37:24 +0200160 if (priv->rev == 1)
161 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_V1_PL_INT_ENA);
162 else
163 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG, LCDC_V2_PL_INT_ENA);
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100164}
165
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300166static void tilcdc_crtc_enable_irqs(struct drm_device *dev)
167{
168 struct tilcdc_drm_private *priv = dev->dev_private;
169
170 tilcdc_clear_irqstatus(dev, 0xffffffff);
171
172 if (priv->rev == 1) {
173 tilcdc_set(dev, LCDC_RASTER_CTRL_REG,
Jyri Sarha36725832016-11-21 18:30:19 +0200174 LCDC_V1_SYNC_LOST_INT_ENA | LCDC_V1_FRAME_DONE_INT_ENA |
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300175 LCDC_V1_UNDERFLOW_INT_ENA);
Karl Beldan8d6c3f72016-08-23 12:57:00 +0000176 tilcdc_set(dev, LCDC_DMA_CTRL_REG,
177 LCDC_V1_END_OF_FRAME_INT_ENA);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300178 } else {
179 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG,
180 LCDC_V2_UNDERFLOW_INT_ENA |
181 LCDC_V2_END_OF_FRAME0_INT_ENA |
182 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
183 }
184}
185
186static void tilcdc_crtc_disable_irqs(struct drm_device *dev)
187{
188 struct tilcdc_drm_private *priv = dev->dev_private;
189
190 /* disable irqs that we might have enabled: */
191 if (priv->rev == 1) {
192 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
Jyri Sarha36725832016-11-21 18:30:19 +0200193 LCDC_V1_SYNC_LOST_INT_ENA | LCDC_V1_FRAME_DONE_INT_ENA |
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300194 LCDC_V1_UNDERFLOW_INT_ENA | LCDC_V1_PL_INT_ENA);
195 tilcdc_clear(dev, LCDC_DMA_CTRL_REG,
196 LCDC_V1_END_OF_FRAME_INT_ENA);
197 } else {
198 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
199 LCDC_V2_UNDERFLOW_INT_ENA | LCDC_V2_PL_INT_ENA |
200 LCDC_V2_END_OF_FRAME0_INT_ENA |
201 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
202 }
203}
204
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300205static void reset(struct drm_crtc *crtc)
Rob Clark16ea9752013-01-08 15:04:28 -0600206{
207 struct drm_device *dev = crtc->dev;
208 struct tilcdc_drm_private *priv = dev->dev_private;
209
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300210 if (priv->rev != 2)
211 return;
212
213 tilcdc_set(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
214 usleep_range(250, 1000);
215 tilcdc_clear(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
216}
217
Jyri Sarha75d7f272016-11-24 23:25:08 +0200218/*
219 * Calculate the percentage difference between the requested pixel clock rate
220 * and the effective rate resulting from calculating the clock divider value.
221 */
222static unsigned int tilcdc_pclk_diff(unsigned long rate,
223 unsigned long real_rate)
224{
225 int r = rate / 100, rr = real_rate / 100;
226
227 return (unsigned int)(abs(((rr - r) * 100) / r));
228}
229
230static void tilcdc_crtc_set_clk(struct drm_crtc *crtc)
231{
232 struct drm_device *dev = crtc->dev;
233 struct tilcdc_drm_private *priv = dev->dev_private;
234 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
235 unsigned long clk_rate, real_rate, req_rate;
236 unsigned int clkdiv;
237 int ret;
238
239 clkdiv = 2; /* first try using a standard divider of 2 */
240
241 /* mode.clock is in KHz, set_rate wants parameter in Hz */
242 req_rate = crtc->mode.clock * 1000;
243
244 ret = clk_set_rate(priv->clk, req_rate * clkdiv);
245 clk_rate = clk_get_rate(priv->clk);
246 if (ret < 0) {
247 /*
248 * If we fail to set the clock rate (some architectures don't
249 * use the common clock framework yet and may not implement
250 * all the clk API calls for every clock), try the next best
251 * thing: adjusting the clock divider, unless clk_get_rate()
252 * failed as well.
253 */
254 if (!clk_rate) {
255 /* Nothing more we can do. Just bail out. */
256 dev_err(dev->dev,
257 "failed to set the pixel clock - unable to read current lcdc clock rate\n");
258 return;
259 }
260
261 clkdiv = DIV_ROUND_CLOSEST(clk_rate, req_rate);
262
263 /*
264 * Emit a warning if the real clock rate resulting from the
265 * calculated divider differs much from the requested rate.
266 *
267 * 5% is an arbitrary value - LCDs are usually quite tolerant
268 * about pixel clock rates.
269 */
270 real_rate = clkdiv * req_rate;
271
272 if (tilcdc_pclk_diff(clk_rate, real_rate) > 5) {
273 dev_warn(dev->dev,
274 "effective pixel clock rate (%luHz) differs from the calculated rate (%luHz)\n",
275 clk_rate, real_rate);
276 }
277 }
278
279 tilcdc_crtc->lcd_fck_rate = clk_rate;
280
281 DBG("lcd_clk=%u, mode clock=%d, div=%u",
282 tilcdc_crtc->lcd_fck_rate, crtc->mode.clock, clkdiv);
283
284 /* Configure the LCD clock divisor. */
285 tilcdc_write(dev, LCDC_CTRL_REG, LCDC_CLK_DIVISOR(clkdiv) |
286 LCDC_RASTER_MODE);
287
288 if (priv->rev == 2)
289 tilcdc_set(dev, LCDC_CLK_ENABLE_REG,
290 LCDC_V2_DMA_CLK_EN | LCDC_V2_LIDD_CLK_EN |
291 LCDC_V2_CORE_CLK_EN);
292}
293
294static void tilcdc_crtc_set_mode(struct drm_crtc *crtc)
295{
296 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
297 struct drm_device *dev = crtc->dev;
298 struct tilcdc_drm_private *priv = dev->dev_private;
299 const struct tilcdc_panel_info *info = tilcdc_crtc->info;
300 uint32_t reg, hbp, hfp, hsw, vbp, vfp, vsw;
301 struct drm_display_mode *mode = &crtc->state->adjusted_mode;
302 struct drm_framebuffer *fb = crtc->primary->state->fb;
303
304 if (WARN_ON(!info))
305 return;
306
307 if (WARN_ON(!fb))
308 return;
309
310 /* Configure the Burst Size and fifo threshold of DMA: */
311 reg = tilcdc_read(dev, LCDC_DMA_CTRL_REG) & ~0x00000770;
312 switch (info->dma_burst_sz) {
313 case 1:
314 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_1);
315 break;
316 case 2:
317 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_2);
318 break;
319 case 4:
320 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_4);
321 break;
322 case 8:
323 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_8);
324 break;
325 case 16:
326 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_16);
327 break;
328 default:
329 dev_err(dev->dev, "invalid burst size\n");
330 return;
331 }
332 reg |= (info->fifo_th << 8);
333 tilcdc_write(dev, LCDC_DMA_CTRL_REG, reg);
334
335 /* Configure timings: */
336 hbp = mode->htotal - mode->hsync_end;
337 hfp = mode->hsync_start - mode->hdisplay;
338 hsw = mode->hsync_end - mode->hsync_start;
339 vbp = mode->vtotal - mode->vsync_end;
340 vfp = mode->vsync_start - mode->vdisplay;
341 vsw = mode->vsync_end - mode->vsync_start;
342
343 DBG("%dx%d, hbp=%u, hfp=%u, hsw=%u, vbp=%u, vfp=%u, vsw=%u",
344 mode->hdisplay, mode->vdisplay, hbp, hfp, hsw, vbp, vfp, vsw);
345
346 /* Set AC Bias Period and Number of Transitions per Interrupt: */
347 reg = tilcdc_read(dev, LCDC_RASTER_TIMING_2_REG) & ~0x000fff00;
348 reg |= LCDC_AC_BIAS_FREQUENCY(info->ac_bias) |
349 LCDC_AC_BIAS_TRANSITIONS_PER_INT(info->ac_bias_intrpt);
350
351 /*
352 * subtract one from hfp, hbp, hsw because the hardware uses
353 * a value of 0 as 1
354 */
355 if (priv->rev == 2) {
356 /* clear bits we're going to set */
357 reg &= ~0x78000033;
358 reg |= ((hfp-1) & 0x300) >> 8;
359 reg |= ((hbp-1) & 0x300) >> 4;
360 reg |= ((hsw-1) & 0x3c0) << 21;
361 }
362 tilcdc_write(dev, LCDC_RASTER_TIMING_2_REG, reg);
363
364 reg = (((mode->hdisplay >> 4) - 1) << 4) |
365 (((hbp-1) & 0xff) << 24) |
366 (((hfp-1) & 0xff) << 16) |
367 (((hsw-1) & 0x3f) << 10);
368 if (priv->rev == 2)
369 reg |= (((mode->hdisplay >> 4) - 1) & 0x40) >> 3;
370 tilcdc_write(dev, LCDC_RASTER_TIMING_0_REG, reg);
371
372 reg = ((mode->vdisplay - 1) & 0x3ff) |
373 ((vbp & 0xff) << 24) |
374 ((vfp & 0xff) << 16) |
375 (((vsw-1) & 0x3f) << 10);
376 tilcdc_write(dev, LCDC_RASTER_TIMING_1_REG, reg);
377
378 /*
379 * be sure to set Bit 10 for the V2 LCDC controller,
380 * otherwise limited to 1024 pixels width, stopping
381 * 1920x1080 being supported.
382 */
383 if (priv->rev == 2) {
384 if ((mode->vdisplay - 1) & 0x400) {
385 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG,
386 LCDC_LPP_B10);
387 } else {
388 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG,
389 LCDC_LPP_B10);
390 }
391 }
392
393 /* Configure display type: */
394 reg = tilcdc_read(dev, LCDC_RASTER_CTRL_REG) &
395 ~(LCDC_TFT_MODE | LCDC_MONO_8BIT_MODE | LCDC_MONOCHROME_MODE |
396 LCDC_V2_TFT_24BPP_MODE | LCDC_V2_TFT_24BPP_UNPACK |
397 0x000ff000 /* Palette Loading Delay bits */);
398 reg |= LCDC_TFT_MODE; /* no monochrome/passive support */
399 if (info->tft_alt_mode)
400 reg |= LCDC_TFT_ALT_ENABLE;
401 if (priv->rev == 2) {
402 switch (fb->pixel_format) {
403 case DRM_FORMAT_BGR565:
404 case DRM_FORMAT_RGB565:
405 break;
406 case DRM_FORMAT_XBGR8888:
407 case DRM_FORMAT_XRGB8888:
408 reg |= LCDC_V2_TFT_24BPP_UNPACK;
409 /* fallthrough */
410 case DRM_FORMAT_BGR888:
411 case DRM_FORMAT_RGB888:
412 reg |= LCDC_V2_TFT_24BPP_MODE;
413 break;
414 default:
415 dev_err(dev->dev, "invalid pixel format\n");
416 return;
417 }
418 }
419 reg |= info->fdd < 12;
420 tilcdc_write(dev, LCDC_RASTER_CTRL_REG, reg);
421
422 if (info->invert_pxl_clk)
423 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
424 else
425 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
426
427 if (info->sync_ctrl)
428 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
429 else
430 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
431
432 if (info->sync_edge)
433 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
434 else
435 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
436
437 if (mode->flags & DRM_MODE_FLAG_NHSYNC)
438 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
439 else
440 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
441
442 if (mode->flags & DRM_MODE_FLAG_NVSYNC)
443 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
444 else
445 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
446
447 if (info->raster_order)
448 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
449 else
450 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
451
452 tilcdc_crtc_set_clk(crtc);
453
454 tilcdc_crtc_load_palette(crtc);
455
456 set_scanout(crtc, fb);
457
458 drm_framebuffer_reference(fb);
459
460 crtc->hwmode = crtc->state->adjusted_mode;
461}
462
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300463static void tilcdc_crtc_enable(struct drm_crtc *crtc)
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300464{
465 struct drm_device *dev = crtc->dev;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300466 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
467
Jyri Sarha2e0965b2016-09-06 17:25:08 +0300468 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
Jyri Sarha2d53a182016-10-25 12:27:31 +0300469 mutex_lock(&tilcdc_crtc->enable_lock);
470 if (tilcdc_crtc->enabled || tilcdc_crtc->shutdown) {
471 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300472 return;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300473 }
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300474
475 pm_runtime_get_sync(dev->dev);
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300476
477 reset(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600478
Jyri Sarha75d7f272016-11-24 23:25:08 +0200479 tilcdc_crtc_set_mode(crtc);
480
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300481 tilcdc_crtc_enable_irqs(dev);
482
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300483 tilcdc_clear(dev, LCDC_DMA_CTRL_REG, LCDC_DUAL_FRAME_BUFFER_ENABLE);
Jyri Sarhaf13e0882016-11-19 18:00:32 +0200484 tilcdc_write_mask(dev, LCDC_RASTER_CTRL_REG,
485 LCDC_PALETTE_LOAD_MODE(DATA_ONLY),
486 LCDC_PALETTE_LOAD_MODE_MASK);
Rob Clark16ea9752013-01-08 15:04:28 -0600487 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300488
489 drm_crtc_vblank_on(crtc);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300490
491 tilcdc_crtc->enabled = true;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300492 mutex_unlock(&tilcdc_crtc->enable_lock);
Rob Clark16ea9752013-01-08 15:04:28 -0600493}
494
Jyri Sarha2d53a182016-10-25 12:27:31 +0300495static void tilcdc_crtc_off(struct drm_crtc *crtc, bool shutdown)
Rob Clark16ea9752013-01-08 15:04:28 -0600496{
Jyri Sarha2d5be882016-04-07 20:20:23 +0300497 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600498 struct drm_device *dev = crtc->dev;
Jyri Sarha2d5be882016-04-07 20:20:23 +0300499 struct tilcdc_drm_private *priv = dev->dev_private;
Jyri Sarha75d7f272016-11-24 23:25:08 +0200500 int ret;
Rob Clark16ea9752013-01-08 15:04:28 -0600501
Jyri Sarha2d53a182016-10-25 12:27:31 +0300502 mutex_lock(&tilcdc_crtc->enable_lock);
503 if (shutdown)
504 tilcdc_crtc->shutdown = true;
505 if (!tilcdc_crtc->enabled) {
506 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300507 return;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300508 }
Jyri Sarha2d5be882016-04-07 20:20:23 +0300509 tilcdc_crtc->frame_done = false;
Rob Clark16ea9752013-01-08 15:04:28 -0600510 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarha2d5be882016-04-07 20:20:23 +0300511
512 /*
Jyri Sarha75d7f272016-11-24 23:25:08 +0200513 * Wait for framedone irq which will still come before putting
514 * things to sleep..
Jyri Sarha2d5be882016-04-07 20:20:23 +0300515 */
Jyri Sarha75d7f272016-11-24 23:25:08 +0200516 ret = wait_event_timeout(tilcdc_crtc->frame_done_wq,
517 tilcdc_crtc->frame_done,
518 msecs_to_jiffies(500));
519 if (ret == 0)
520 dev_err(dev->dev, "%s: timeout waiting for framedone\n",
521 __func__);
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300522
523 drm_crtc_vblank_off(crtc);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300524
525 tilcdc_crtc_disable_irqs(dev);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300526
527 pm_runtime_put_sync(dev->dev);
528
529 if (tilcdc_crtc->next_fb) {
530 drm_flip_work_queue(&tilcdc_crtc->unref_work,
531 tilcdc_crtc->next_fb);
532 tilcdc_crtc->next_fb = NULL;
533 }
534
535 if (tilcdc_crtc->curr_fb) {
536 drm_flip_work_queue(&tilcdc_crtc->unref_work,
537 tilcdc_crtc->curr_fb);
538 tilcdc_crtc->curr_fb = NULL;
539 }
540
541 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
542 tilcdc_crtc->last_vblank = ktime_set(0, 0);
543
544 tilcdc_crtc->enabled = false;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300545 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300546}
547
Jyri Sarha9e79e062016-10-18 23:23:27 +0300548static void tilcdc_crtc_disable(struct drm_crtc *crtc)
549{
550 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
Jyri Sarha2d53a182016-10-25 12:27:31 +0300551 tilcdc_crtc_off(crtc, false);
552}
553
554void tilcdc_crtc_shutdown(struct drm_crtc *crtc)
555{
556 tilcdc_crtc_off(crtc, true);
Jyri Sarha9e79e062016-10-18 23:23:27 +0300557}
558
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300559static bool tilcdc_crtc_is_on(struct drm_crtc *crtc)
560{
561 return crtc->state && crtc->state->enable && crtc->state->active;
Rob Clark16ea9752013-01-08 15:04:28 -0600562}
563
Jyri Sarha13b3d722016-04-06 14:02:38 +0300564static void tilcdc_crtc_recover_work(struct work_struct *work)
565{
566 struct tilcdc_crtc *tilcdc_crtc =
567 container_of(work, struct tilcdc_crtc, recover_work);
568 struct drm_crtc *crtc = &tilcdc_crtc->base;
569
570 dev_info(crtc->dev->dev, "%s: Reset CRTC", __func__);
571
572 drm_modeset_lock_crtc(crtc, NULL);
573
574 if (!tilcdc_crtc_is_on(crtc))
575 goto out;
576
577 tilcdc_crtc_disable(crtc);
578 tilcdc_crtc_enable(crtc);
579out:
580 drm_modeset_unlock_crtc(crtc);
581}
582
Rob Clark16ea9752013-01-08 15:04:28 -0600583static void tilcdc_crtc_destroy(struct drm_crtc *crtc)
584{
585 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Jyri Sarha4e910c72016-09-06 22:55:33 +0300586 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600587
Jyri Sarha6c94c712016-09-07 11:46:40 +0300588 drm_modeset_lock_crtc(crtc, NULL);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300589 tilcdc_crtc_disable(crtc);
Jyri Sarha6c94c712016-09-07 11:46:40 +0300590 drm_modeset_unlock_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600591
Jyri Sarha4e910c72016-09-06 22:55:33 +0300592 flush_workqueue(priv->wq);
Rob Clark16ea9752013-01-08 15:04:28 -0600593
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300594 of_node_put(crtc->port);
Rob Clark16ea9752013-01-08 15:04:28 -0600595 drm_crtc_cleanup(crtc);
Rob Clarka464d612013-08-07 13:41:20 -0400596 drm_flip_work_cleanup(&tilcdc_crtc->unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -0600597}
598
Jyri Sarhae0e344e2016-06-22 17:21:06 +0300599int tilcdc_crtc_update_fb(struct drm_crtc *crtc,
Rob Clark16ea9752013-01-08 15:04:28 -0600600 struct drm_framebuffer *fb,
Jyri Sarhae0e344e2016-06-22 17:21:06 +0300601 struct drm_pending_vblank_event *event)
Rob Clark16ea9752013-01-08 15:04:28 -0600602{
603 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
604 struct drm_device *dev = crtc->dev;
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300605 unsigned long flags;
Tomi Valkeinen6f206e92014-02-07 17:37:07 +0000606
Jyri Sarha2e0965b2016-09-06 17:25:08 +0300607 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
608
Rob Clark16ea9752013-01-08 15:04:28 -0600609 if (tilcdc_crtc->event) {
610 dev_err(dev->dev, "already pending page flip!\n");
611 return -EBUSY;
612 }
613
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300614 drm_framebuffer_reference(fb);
615
Matt Roperf4510a22014-04-01 15:22:40 -0700616 crtc->primary->fb = fb;
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300617
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200618 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300619
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300620 if (crtc->hwmode.vrefresh && ktime_to_ns(tilcdc_crtc->last_vblank)) {
621 ktime_t next_vblank;
622 s64 tdiff;
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300623
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300624 next_vblank = ktime_add_us(tilcdc_crtc->last_vblank,
625 1000000 / crtc->hwmode.vrefresh);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200626
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300627 tdiff = ktime_to_us(ktime_sub(next_vblank, ktime_get()));
628
629 if (tdiff < TILCDC_VBLANK_SAFETY_THRESHOLD_US)
630 tilcdc_crtc->next_fb = fb;
631 }
632
633 if (tilcdc_crtc->next_fb != fb)
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200634 set_scanout(crtc, fb);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200635
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300636 tilcdc_crtc->event = event;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200637
638 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600639
640 return 0;
641}
642
Rob Clark16ea9752013-01-08 15:04:28 -0600643static bool tilcdc_crtc_mode_fixup(struct drm_crtc *crtc,
644 const struct drm_display_mode *mode,
645 struct drm_display_mode *adjusted_mode)
646{
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200647 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
648
649 if (!tilcdc_crtc->simulate_vesa_sync)
650 return true;
651
652 /*
653 * tilcdc does not generate VESA-compliant sync but aligns
654 * VS on the second edge of HS instead of first edge.
655 * We use adjusted_mode, to fixup sync by aligning both rising
656 * edges and add HSKEW offset to fix the sync.
657 */
658 adjusted_mode->hskew = mode->hsync_end - mode->hsync_start;
659 adjusted_mode->flags |= DRM_MODE_FLAG_HSKEW;
660
661 if (mode->flags & DRM_MODE_FLAG_NHSYNC) {
662 adjusted_mode->flags |= DRM_MODE_FLAG_PHSYNC;
663 adjusted_mode->flags &= ~DRM_MODE_FLAG_NHSYNC;
664 } else {
665 adjusted_mode->flags |= DRM_MODE_FLAG_NHSYNC;
666 adjusted_mode->flags &= ~DRM_MODE_FLAG_PHSYNC;
667 }
668
Rob Clark16ea9752013-01-08 15:04:28 -0600669 return true;
670}
671
Jyri Sarhadb380c52016-04-07 15:10:23 +0300672static int tilcdc_crtc_atomic_check(struct drm_crtc *crtc,
673 struct drm_crtc_state *state)
674{
675 struct drm_display_mode *mode = &state->mode;
676 int ret;
677
678 /* If we are not active we don't care */
679 if (!state->active)
680 return 0;
681
682 if (state->state->planes[0].ptr != crtc->primary ||
683 state->state->planes[0].state == NULL ||
684 state->state->planes[0].state->crtc != crtc) {
685 dev_dbg(crtc->dev->dev, "CRTC primary plane must be present");
686 return -EINVAL;
687 }
688
689 ret = tilcdc_crtc_mode_valid(crtc, mode);
690 if (ret) {
691 dev_dbg(crtc->dev->dev, "Mode \"%s\" not valid", mode->name);
692 return -EINVAL;
693 }
694
695 return 0;
696}
697
Rob Clark16ea9752013-01-08 15:04:28 -0600698static const struct drm_crtc_funcs tilcdc_crtc_funcs = {
Jyri Sarha305198d2016-04-07 15:05:16 +0300699 .destroy = tilcdc_crtc_destroy,
700 .set_config = drm_atomic_helper_set_config,
701 .page_flip = drm_atomic_helper_page_flip,
702 .reset = drm_atomic_helper_crtc_reset,
703 .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state,
704 .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state,
Rob Clark16ea9752013-01-08 15:04:28 -0600705};
706
707static const struct drm_crtc_helper_funcs tilcdc_crtc_helper_funcs = {
Rob Clark16ea9752013-01-08 15:04:28 -0600708 .mode_fixup = tilcdc_crtc_mode_fixup,
Jyri Sarha305198d2016-04-07 15:05:16 +0300709 .enable = tilcdc_crtc_enable,
710 .disable = tilcdc_crtc_disable,
Jyri Sarhadb380c52016-04-07 15:10:23 +0300711 .atomic_check = tilcdc_crtc_atomic_check,
Rob Clark16ea9752013-01-08 15:04:28 -0600712};
713
714int tilcdc_crtc_max_width(struct drm_crtc *crtc)
715{
716 struct drm_device *dev = crtc->dev;
717 struct tilcdc_drm_private *priv = dev->dev_private;
718 int max_width = 0;
719
720 if (priv->rev == 1)
721 max_width = 1024;
722 else if (priv->rev == 2)
723 max_width = 2048;
724
725 return max_width;
726}
727
728int tilcdc_crtc_mode_valid(struct drm_crtc *crtc, struct drm_display_mode *mode)
729{
730 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
731 unsigned int bandwidth;
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500732 uint32_t hbp, hfp, hsw, vbp, vfp, vsw;
Rob Clark16ea9752013-01-08 15:04:28 -0600733
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500734 /*
735 * check to see if the width is within the range that
736 * the LCD Controller physically supports
737 */
Rob Clark16ea9752013-01-08 15:04:28 -0600738 if (mode->hdisplay > tilcdc_crtc_max_width(crtc))
739 return MODE_VIRTUAL_X;
740
741 /* width must be multiple of 16 */
742 if (mode->hdisplay & 0xf)
743 return MODE_VIRTUAL_X;
744
745 if (mode->vdisplay > 2048)
746 return MODE_VIRTUAL_Y;
747
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500748 DBG("Processing mode %dx%d@%d with pixel clock %d",
749 mode->hdisplay, mode->vdisplay,
750 drm_mode_vrefresh(mode), mode->clock);
751
752 hbp = mode->htotal - mode->hsync_end;
753 hfp = mode->hsync_start - mode->hdisplay;
754 hsw = mode->hsync_end - mode->hsync_start;
755 vbp = mode->vtotal - mode->vsync_end;
756 vfp = mode->vsync_start - mode->vdisplay;
757 vsw = mode->vsync_end - mode->vsync_start;
758
759 if ((hbp-1) & ~0x3ff) {
760 DBG("Pruning mode: Horizontal Back Porch out of range");
761 return MODE_HBLANK_WIDE;
762 }
763
764 if ((hfp-1) & ~0x3ff) {
765 DBG("Pruning mode: Horizontal Front Porch out of range");
766 return MODE_HBLANK_WIDE;
767 }
768
769 if ((hsw-1) & ~0x3ff) {
770 DBG("Pruning mode: Horizontal Sync Width out of range");
771 return MODE_HSYNC_WIDE;
772 }
773
774 if (vbp & ~0xff) {
775 DBG("Pruning mode: Vertical Back Porch out of range");
776 return MODE_VBLANK_WIDE;
777 }
778
779 if (vfp & ~0xff) {
780 DBG("Pruning mode: Vertical Front Porch out of range");
781 return MODE_VBLANK_WIDE;
782 }
783
784 if ((vsw-1) & ~0x3f) {
785 DBG("Pruning mode: Vertical Sync Width out of range");
786 return MODE_VSYNC_WIDE;
787 }
788
Darren Etheridge4e564342013-06-21 13:52:23 -0500789 /*
790 * some devices have a maximum allowed pixel clock
791 * configured from the DT
792 */
793 if (mode->clock > priv->max_pixelclock) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500794 DBG("Pruning mode: pixel clock too high");
Darren Etheridge4e564342013-06-21 13:52:23 -0500795 return MODE_CLOCK_HIGH;
796 }
797
798 /*
799 * some devices further limit the max horizontal resolution
800 * configured from the DT
801 */
802 if (mode->hdisplay > priv->max_width)
803 return MODE_BAD_WIDTH;
804
Rob Clark16ea9752013-01-08 15:04:28 -0600805 /* filter out modes that would require too much memory bandwidth: */
Darren Etheridge4e564342013-06-21 13:52:23 -0500806 bandwidth = mode->hdisplay * mode->vdisplay *
807 drm_mode_vrefresh(mode);
808 if (bandwidth > priv->max_bandwidth) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500809 DBG("Pruning mode: exceeds defined bandwidth limit");
Rob Clark16ea9752013-01-08 15:04:28 -0600810 return MODE_BAD;
Darren Etheridge4e564342013-06-21 13:52:23 -0500811 }
Rob Clark16ea9752013-01-08 15:04:28 -0600812
813 return MODE_OK;
814}
815
816void tilcdc_crtc_set_panel_info(struct drm_crtc *crtc,
817 const struct tilcdc_panel_info *info)
818{
819 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
820 tilcdc_crtc->info = info;
821}
822
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200823void tilcdc_crtc_set_simulate_vesa_sync(struct drm_crtc *crtc,
824 bool simulate_vesa_sync)
825{
826 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
827
828 tilcdc_crtc->simulate_vesa_sync = simulate_vesa_sync;
829}
830
Rob Clark16ea9752013-01-08 15:04:28 -0600831void tilcdc_crtc_update_clk(struct drm_crtc *crtc)
832{
Rob Clark16ea9752013-01-08 15:04:28 -0600833 struct drm_device *dev = crtc->dev;
834 struct tilcdc_drm_private *priv = dev->dev_private;
Jyri Sarha642e5162016-09-06 16:19:54 +0300835 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600836
Jyri Sarha642e5162016-09-06 16:19:54 +0300837 drm_modeset_lock_crtc(crtc, NULL);
838 if (tilcdc_crtc->lcd_fck_rate != clk_get_rate(priv->clk)) {
839 if (tilcdc_crtc_is_on(crtc)) {
840 pm_runtime_get_sync(dev->dev);
841 tilcdc_crtc_disable(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600842
Jyri Sarha642e5162016-09-06 16:19:54 +0300843 tilcdc_crtc_set_clk(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600844
Jyri Sarha642e5162016-09-06 16:19:54 +0300845 tilcdc_crtc_enable(crtc);
846 pm_runtime_put_sync(dev->dev);
847 }
Rob Clark16ea9752013-01-08 15:04:28 -0600848 }
Jyri Sarha642e5162016-09-06 16:19:54 +0300849 drm_modeset_unlock_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600850}
851
Jyri Sarha5895d082016-01-08 14:33:09 +0200852#define SYNC_LOST_COUNT_LIMIT 50
853
Rob Clark16ea9752013-01-08 15:04:28 -0600854irqreturn_t tilcdc_crtc_irq(struct drm_crtc *crtc)
855{
856 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
857 struct drm_device *dev = crtc->dev;
858 struct tilcdc_drm_private *priv = dev->dev_private;
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300859 uint32_t stat;
Rob Clark16ea9752013-01-08 15:04:28 -0600860
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300861 stat = tilcdc_read_irqstatus(dev);
862 tilcdc_clear_irqstatus(dev, stat);
863
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300864 if (stat & LCDC_END_OF_FRAME0) {
Rob Clark16ea9752013-01-08 15:04:28 -0600865 unsigned long flags;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200866 bool skip_event = false;
867 ktime_t now;
868
869 now = ktime_get();
Rob Clark16ea9752013-01-08 15:04:28 -0600870
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300871 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
Rob Clark16ea9752013-01-08 15:04:28 -0600872
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200873 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600874
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200875 tilcdc_crtc->last_vblank = now;
Rob Clark16ea9752013-01-08 15:04:28 -0600876
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200877 if (tilcdc_crtc->next_fb) {
878 set_scanout(crtc, tilcdc_crtc->next_fb);
879 tilcdc_crtc->next_fb = NULL;
880 skip_event = true;
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300881 }
882
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200883 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
884
Gustavo Padovan099ede82016-07-04 21:04:52 -0300885 drm_crtc_handle_vblank(crtc);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200886
887 if (!skip_event) {
888 struct drm_pending_vblank_event *event;
889
890 spin_lock_irqsave(&dev->event_lock, flags);
891
892 event = tilcdc_crtc->event;
893 tilcdc_crtc->event = NULL;
894 if (event)
Gustavo Padovandfebc152016-04-14 10:48:22 -0700895 drm_crtc_send_vblank_event(crtc, event);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200896
897 spin_unlock_irqrestore(&dev->event_lock, flags);
898 }
Jyri Sarha5895d082016-01-08 14:33:09 +0200899
900 if (tilcdc_crtc->frame_intact)
901 tilcdc_crtc->sync_lost_count = 0;
902 else
903 tilcdc_crtc->frame_intact = true;
Rob Clark16ea9752013-01-08 15:04:28 -0600904 }
905
Jyri Sarha14944112016-04-07 20:36:48 +0300906 if (stat & LCDC_FIFO_UNDERFLOW)
Daniel Schultzd7014532016-10-28 13:52:42 +0200907 dev_err_ratelimited(dev->dev, "%s(0x%08x): FIFO underflow",
Jyri Sarha14944112016-04-07 20:36:48 +0300908 __func__, stat);
909
Jyri Sarha55e165c2016-11-15 23:37:24 +0200910 if (stat & LCDC_PL_LOAD_DONE) {
911 complete(&tilcdc_crtc->palette_loaded);
912 if (priv->rev == 1)
913 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
914 LCDC_V1_PL_INT_ENA);
915 else
916 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
917 LCDC_V2_PL_INT_ENA);
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100918 }
919
Jyri Sarhacba88442016-11-16 00:12:27 +0200920 if (stat & LCDC_SYNC_LOST) {
921 dev_err_ratelimited(dev->dev, "%s(0x%08x): Sync lost",
922 __func__, stat);
923 tilcdc_crtc->frame_intact = false;
924 if (tilcdc_crtc->sync_lost_count++ >
925 SYNC_LOST_COUNT_LIMIT) {
926 dev_err(dev->dev, "%s(0x%08x): Sync lost flood detected, recovering", __func__, stat);
927 queue_work(system_wq, &tilcdc_crtc->recover_work);
928 if (priv->rev == 1)
929 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
930 LCDC_V1_SYNC_LOST_INT_ENA);
931 else
932 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
933 LCDC_SYNC_LOST);
934 tilcdc_crtc->sync_lost_count = 0;
935 }
936 }
937
Jyri Sarha36725832016-11-21 18:30:19 +0200938 if (stat & LCDC_FRAME_DONE) {
939 tilcdc_crtc->frame_done = true;
940 wake_up(&tilcdc_crtc->frame_done_wq);
941 /* rev 1 lcdc appears to hang if irq is not disbaled here */
942 if (priv->rev == 1)
943 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
944 LCDC_V1_FRAME_DONE_INT_ENA);
945 }
946
Jyri Sarha14944112016-04-07 20:36:48 +0300947 /* For revision 2 only */
Rob Clark16ea9752013-01-08 15:04:28 -0600948 if (priv->rev == 2) {
Jyri Sarha14944112016-04-07 20:36:48 +0300949 /* Indicate to LCDC that the interrupt service routine has
950 * completed, see 13.3.6.1.6 in AM335x TRM.
951 */
952 tilcdc_write(dev, LCDC_END_OF_INT_IND_REG, 0);
953 }
Jyri Sarhac0c2baa2015-12-18 13:07:52 +0200954
Rob Clark16ea9752013-01-08 15:04:28 -0600955 return IRQ_HANDLED;
956}
957
Jyri Sarha9963d362016-11-15 22:56:46 +0200958int tilcdc_crtc_create(struct drm_device *dev)
Rob Clark16ea9752013-01-08 15:04:28 -0600959{
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300960 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600961 struct tilcdc_crtc *tilcdc_crtc;
962 struct drm_crtc *crtc;
963 int ret;
964
Jyri Sarhad0ec32c2016-02-23 12:44:27 +0200965 tilcdc_crtc = devm_kzalloc(dev->dev, sizeof(*tilcdc_crtc), GFP_KERNEL);
Rob Clark16ea9752013-01-08 15:04:28 -0600966 if (!tilcdc_crtc) {
967 dev_err(dev->dev, "allocation failed\n");
Jyri Sarha9963d362016-11-15 22:56:46 +0200968 return -ENOMEM;
Rob Clark16ea9752013-01-08 15:04:28 -0600969 }
970
Jyri Sarha55e165c2016-11-15 23:37:24 +0200971 init_completion(&tilcdc_crtc->palette_loaded);
972 tilcdc_crtc->palette_base = dmam_alloc_coherent(dev->dev,
973 TILCDC_PALETTE_SIZE,
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100974 &tilcdc_crtc->palette_dma_handle,
975 GFP_KERNEL | __GFP_ZERO);
Jyri Sarha55e165c2016-11-15 23:37:24 +0200976 if (!tilcdc_crtc->palette_base)
977 return -ENOMEM;
978 *tilcdc_crtc->palette_base = TILCDC_PALETTE_FIRST_ENTRY;
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100979
Rob Clark16ea9752013-01-08 15:04:28 -0600980 crtc = &tilcdc_crtc->base;
981
Jyri Sarha47f571c2016-04-07 15:04:18 +0300982 ret = tilcdc_plane_init(dev, &tilcdc_crtc->primary);
983 if (ret < 0)
984 goto fail;
985
Jyri Sarha2d53a182016-10-25 12:27:31 +0300986 mutex_init(&tilcdc_crtc->enable_lock);
987
Rob Clark16ea9752013-01-08 15:04:28 -0600988 init_waitqueue_head(&tilcdc_crtc->frame_done_wq);
989
Boris BREZILLONd7f8db52014-11-14 19:30:30 +0100990 drm_flip_work_init(&tilcdc_crtc->unref_work,
Rob Clarka464d612013-08-07 13:41:20 -0400991 "unref", unref_worker);
Rob Clark16ea9752013-01-08 15:04:28 -0600992
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200993 spin_lock_init(&tilcdc_crtc->irq_lock);
Jyri Sarha13b3d722016-04-06 14:02:38 +0300994 INIT_WORK(&tilcdc_crtc->recover_work, tilcdc_crtc_recover_work);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200995
Jyri Sarha47f571c2016-04-07 15:04:18 +0300996 ret = drm_crtc_init_with_planes(dev, crtc,
997 &tilcdc_crtc->primary,
998 NULL,
999 &tilcdc_crtc_funcs,
1000 "tilcdc crtc");
Rob Clark16ea9752013-01-08 15:04:28 -06001001 if (ret < 0)
1002 goto fail;
1003
1004 drm_crtc_helper_add(crtc, &tilcdc_crtc_helper_funcs);
1005
Jyri Sarhad66284fb2015-05-27 11:58:37 +03001006 if (priv->is_componentized) {
1007 struct device_node *ports =
1008 of_get_child_by_name(dev->dev->of_node, "ports");
1009
1010 if (ports) {
1011 crtc->port = of_get_child_by_name(ports, "port");
1012 of_node_put(ports);
1013 } else {
1014 crtc->port =
1015 of_get_child_by_name(dev->dev->of_node, "port");
1016 }
1017 if (!crtc->port) { /* This should never happen */
1018 dev_err(dev->dev, "Port node not found in %s\n",
1019 dev->dev->of_node->full_name);
Jyri Sarha9963d362016-11-15 22:56:46 +02001020 ret = -EINVAL;
Jyri Sarhad66284fb2015-05-27 11:58:37 +03001021 goto fail;
1022 }
1023 }
1024
Jyri Sarha9963d362016-11-15 22:56:46 +02001025 priv->crtc = crtc;
1026 return 0;
Rob Clark16ea9752013-01-08 15:04:28 -06001027
1028fail:
1029 tilcdc_crtc_destroy(crtc);
Jyri Sarha9963d362016-11-15 22:56:46 +02001030 return -ENOMEM;
Rob Clark16ea9752013-01-08 15:04:28 -06001031}