blob: 4735251a394d8be03784c7b8dc8b5636744f750d [file] [log] [blame]
Thomas Gleixnercaab2772019-06-03 07:44:50 +02001/* SPDX-License-Identifier: GPL-2.0-only */
Rob Clarkdd2da6e2013-11-30 16:12:10 -05002/*
Jeykumar Sankaran25fdd592018-06-27 15:26:09 -04003 * Copyright (c) 2016-2018, The Linux Foundation. All rights reserved.
Rob Clarkdd2da6e2013-11-30 16:12:10 -05004 * Copyright (C) 2013 Red Hat
5 * Author: Rob Clark <robdclark@gmail.com>
Rob Clarkdd2da6e2013-11-30 16:12:10 -05006 */
7
8#ifndef __MSM_KMS_H__
9#define __MSM_KMS_H__
10
11#include <linux/clk.h>
12#include <linux/regulator/consumer.h>
13
14#include "msm_drv.h"
15
Stephane Viau7ca12712014-12-08 10:48:57 -050016#define MAX_PLANE 4
17
Rob Clarkdd2da6e2013-11-30 16:12:10 -050018/* As there are different display controller blocks depending on the
19 * snapdragon version, the kms support is split out and the appropriate
20 * implementation is loaded at runtime. The kms module is responsible
21 * for constructing the appropriate planes/crtcs/encoders/connectors.
22 */
23struct msm_kms_funcs {
24 /* hw initialization: */
25 int (*hw_init)(struct msm_kms *kms);
26 /* irq handling: */
27 void (*irq_preinstall)(struct msm_kms *kms);
28 int (*irq_postinstall)(struct msm_kms *kms);
29 void (*irq_uninstall)(struct msm_kms *kms);
30 irqreturn_t (*irq)(struct msm_kms *kms);
31 int (*enable_vblank)(struct msm_kms *kms, struct drm_crtc *crtc);
32 void (*disable_vblank)(struct msm_kms *kms, struct drm_crtc *crtc);
Rob Clark9f6b6562019-08-29 09:45:14 -070033
34 /*
35 * Atomic commit handling:
Rob Clark2d99ced2019-08-29 09:45:16 -070036 *
37 * Note that in the case of async commits, the funcs which take
38 * a crtc_mask (ie. ->flush_commit(), and ->complete_commit())
39 * might not be evenly balanced with ->prepare_commit(), however
40 * each crtc that effected by a ->prepare_commit() (potentially
41 * multiple times) will eventually (at end of vsync period) be
42 * flushed and completed.
43 *
44 * This has some implications about tracking of cleanup state,
45 * for example SMP blocks to release after commit completes. Ie.
46 * cleanup state should be also duplicated in the various
47 * duplicate_state() methods, as the current cleanup state at
48 * ->complete_commit() time may have accumulated cleanup work
49 * from multiple commits.
Rob Clark9f6b6562019-08-29 09:45:14 -070050 */
51
52 /**
Rob Clarke35a29d2019-08-29 09:45:15 -070053 * Enable/disable power/clks needed for hw access done in other
54 * commit related methods.
55 *
56 * If mdp4 is migrated to runpm, we could probably drop these
57 * and use runpm directly.
58 */
59 void (*enable_commit)(struct msm_kms *kms);
60 void (*disable_commit)(struct msm_kms *kms);
61
62 /**
Rob Clark2d99ced2019-08-29 09:45:16 -070063 * If the kms backend supports async commit, it should implement
64 * this method to return the time of the next vsync. This is
65 * used to determine a time slightly before vsync, for the async
66 * commit timer to run and complete an async commit.
67 */
68 ktime_t (*vsync_time)(struct msm_kms *kms, struct drm_crtc *crtc);
69
70 /**
Rob Clark9f6b6562019-08-29 09:45:14 -070071 * Prepare for atomic commit. This is called after any previous
72 * (async or otherwise) commit has completed.
73 */
Rob Clark0b776d42015-01-30 17:04:45 -050074 void (*prepare_commit)(struct msm_kms *kms, struct drm_atomic_state *state);
Rob Clark9f6b6562019-08-29 09:45:14 -070075
76 /**
77 * Flush an atomic commit. This is called after the hardware
78 * updates have already been pushed down to effected planes/
79 * crtcs/encoders/connectors.
80 */
81 void (*flush_commit)(struct msm_kms *kms, unsigned crtc_mask);
82
Rob Clark9f6b6562019-08-29 09:45:14 -070083 /**
84 * Wait for any in-progress flush to complete on the specified
85 * crtcs. This should not block if there is no in-progress
86 * commit (ie. don't just wait for a vblank), as it will also
87 * be called before ->prepare_commit() to ensure any potential
88 * "async" commit has completed.
89 */
Rob Clarkd4d2c602019-08-29 09:45:12 -070090 void (*wait_flush)(struct msm_kms *kms, unsigned crtc_mask);
91
Rob Clark9f6b6562019-08-29 09:45:14 -070092 /**
93 * Clean up after commit is completed. This is called after
94 * ->wait_flush(), to give the backend a chance to do any
95 * post-commit cleanup.
96 */
97 void (*complete_commit)(struct msm_kms *kms, unsigned crtc_mask);
98
99 /*
100 * Format handling:
101 */
102
Jeykumar Sankaran4e490232018-02-13 12:42:44 -0500103 /* get msm_format w/ optional format modifiers from drm_mode_fb_cmd2 */
104 const struct msm_format *(*get_format)(struct msm_kms *kms,
105 const uint32_t format,
106 const uint64_t modifiers);
Jeykumar Sankaran25fdd592018-06-27 15:26:09 -0400107 /* do format checking on format modified through fb_cmd2 modifiers */
108 int (*check_modified_format)(const struct msm_kms *kms,
109 const struct msm_format *msm_fmt,
110 const struct drm_mode_fb_cmd2 *cmd,
111 struct drm_gem_object **bos);
Rob Clark9f6b6562019-08-29 09:45:14 -0700112
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500113 /* misc: */
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500114 long (*round_pixclk)(struct msm_kms *kms, unsigned long rate,
115 struct drm_encoder *encoder);
Hai Li7a6dc952015-03-26 19:25:15 -0400116 int (*set_split_display)(struct msm_kms *kms,
117 struct drm_encoder *encoder,
118 struct drm_encoder *slave_encoder,
119 bool is_cmd_mode);
Archit Taneja9c9f6f82016-12-05 15:24:53 +0530120 void (*set_encoder_mode)(struct msm_kms *kms,
121 struct drm_encoder *encoder,
122 bool cmd_mode);
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500123 /* cleanup: */
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500124 void (*destroy)(struct msm_kms *kms);
Rob Clarkbc5289e2016-10-26 14:06:55 -0400125#ifdef CONFIG_DEBUG_FS
126 /* debugfs: */
127 int (*debugfs_init)(struct msm_kms *kms, struct drm_minor *minor);
Rob Clarkbc5289e2016-10-26 14:06:55 -0400128#endif
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500129};
130
Rob Clark2d99ced2019-08-29 09:45:16 -0700131struct msm_kms;
132
133/*
134 * A per-crtc timer for pending async atomic flushes. Scheduled to expire
135 * shortly before vblank to flush pending async updates.
136 */
137struct msm_pending_timer {
138 struct hrtimer timer;
Rob Clark363bcec2020-10-19 14:10:53 -0700139 struct kthread_work work;
140 struct kthread_worker *worker;
Rob Clark2d99ced2019-08-29 09:45:16 -0700141 struct msm_kms *kms;
142 unsigned crtc_idx;
143};
144
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500145struct msm_kms {
146 const struct msm_kms_funcs *funcs;
Rob Clark2d99ced2019-08-29 09:45:16 -0700147 struct drm_device *dev;
Archit Tanejaa2b3a552016-05-18 15:06:03 +0530148
149 /* irq number to be passed on to drm_irq_install */
150 int irq;
Rob Clarkf59f62d2017-06-13 10:22:37 -0400151
152 /* mapper-id used to request GEM buffer mapped for scanout: */
Rob Clarkf59f62d2017-06-13 10:22:37 -0400153 struct msm_gem_address_space *aspace;
Rob Clark2d99ced2019-08-29 09:45:16 -0700154
155 /*
156 * For async commit, where ->flush_commit() and later happens
157 * from the crtc's pending_timer close to end of the frame:
158 */
Krishna Manikandanb3d91802020-10-16 19:40:43 +0530159 struct mutex commit_lock[MAX_CRTCS];
Stephen Boyd6ec93512021-01-28 09:09:29 -0800160 struct lock_class_key commit_lock_keys[MAX_CRTCS];
Rob Clark2d99ced2019-08-29 09:45:16 -0700161 unsigned pending_crtc_mask;
162 struct msm_pending_timer pending_timers[MAX_CRTCS];
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500163};
164
Rob Clarkffe71112020-10-19 14:10:52 -0700165static inline int msm_kms_init(struct msm_kms *kms,
Rob Clark9e0efa62013-11-30 17:24:22 -0500166 const struct msm_kms_funcs *funcs)
167{
Rob Clark363bcec2020-10-19 14:10:53 -0700168 unsigned i, ret;
Rob Clark2d99ced2019-08-29 09:45:16 -0700169
Stephen Boyd6ec93512021-01-28 09:09:29 -0800170 for (i = 0; i < ARRAY_SIZE(kms->commit_lock); i++) {
171 lockdep_register_key(&kms->commit_lock_keys[i]);
172 __mutex_init(&kms->commit_lock[i], "&kms->commit_lock[i]",
173 &kms->commit_lock_keys[i]);
174 }
Krishna Manikandanb3d91802020-10-16 19:40:43 +0530175
Rob Clark9e0efa62013-11-30 17:24:22 -0500176 kms->funcs = funcs;
Rob Clark2d99ced2019-08-29 09:45:16 -0700177
Rob Clark363bcec2020-10-19 14:10:53 -0700178 for (i = 0; i < ARRAY_SIZE(kms->pending_timers); i++) {
179 ret = msm_atomic_init_pending_timer(&kms->pending_timers[i], kms, i);
180 if (ret) {
181 return ret;
182 }
183 }
Rob Clarkffe71112020-10-19 14:10:52 -0700184
185 return 0;
186}
187
188static inline void msm_kms_destroy(struct msm_kms *kms)
189{
Rob Clark363bcec2020-10-19 14:10:53 -0700190 unsigned i;
191
192 for (i = 0; i < ARRAY_SIZE(kms->pending_timers); i++)
193 msm_atomic_destroy_pending_timer(&kms->pending_timers[i]);
Rob Clark9e0efa62013-11-30 17:24:22 -0500194}
195
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500196struct msm_kms *mdp4_kms_init(struct drm_device *dev);
Rob Clark06c0dd92013-11-30 17:51:47 -0500197struct msm_kms *mdp5_kms_init(struct drm_device *dev);
Jeykumar Sankaran25fdd592018-06-27 15:26:09 -0400198struct msm_kms *dpu_kms_init(struct drm_device *dev);
Rajesh Yadavbc3220b2018-06-21 16:06:10 -0400199
200struct msm_mdss_funcs {
201 int (*enable)(struct msm_mdss *mdss);
202 int (*disable)(struct msm_mdss *mdss);
203 void (*destroy)(struct drm_device *dev);
204};
205
206struct msm_mdss {
207 struct drm_device *dev;
208 const struct msm_mdss_funcs *funcs;
209};
210
211int mdp5_mdss_init(struct drm_device *dev);
Jeykumar Sankaran25fdd592018-06-27 15:26:09 -0400212int dpu_mdss_init(struct drm_device *dev);
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500213
Rob Clarkd4d2c602019-08-29 09:45:12 -0700214#define for_each_crtc_mask(dev, crtc, crtc_mask) \
215 drm_for_each_crtc(crtc, dev) \
216 for_each_if (drm_crtc_mask(crtc) & (crtc_mask))
217
Rob Clarkcb21f3f2020-10-20 15:26:00 -0700218#define for_each_crtc_mask_reverse(dev, crtc, crtc_mask) \
219 drm_for_each_crtc_reverse(crtc, dev) \
220 for_each_if (drm_crtc_mask(crtc) & (crtc_mask))
221
Rob Clarkdd2da6e2013-11-30 16:12:10 -0500222#endif /* __MSM_KMS_H__ */