blob: 386e8c84be0af24c951526554b37e24cfbc4ae79 [file] [log] [blame]
Thomas Gleixner1802d0b2019-05-27 08:55:21 +02001// SPDX-License-Identifier: GPL-2.0-only
Leilk Liua5682312015-08-07 15:19:50 +08002/*
3 * Copyright (c) 2015 MediaTek Inc.
4 * Author: Leilk Liu <leilk.liu@mediatek.com>
Leilk Liua5682312015-08-07 15:19:50 +08005 */
6
7#include <linux/clk.h>
8#include <linux/device.h>
9#include <linux/err.h>
10#include <linux/interrupt.h>
Leilk Liudd69a0a2015-08-24 11:45:15 +080011#include <linux/io.h>
Leilk Liua5682312015-08-07 15:19:50 +080012#include <linux/ioport.h>
13#include <linux/module.h>
14#include <linux/of.h>
Leilk Liu37457602015-10-26 16:09:44 +080015#include <linux/of_gpio.h>
Leilk Liua5682312015-08-07 15:19:50 +080016#include <linux/platform_device.h>
17#include <linux/platform_data/spi-mt65xx.h>
18#include <linux/pm_runtime.h>
19#include <linux/spi/spi.h>
luhua.xufdeae8f2019-09-11 05:55:31 -040020#include <linux/dma-mapping.h>
Leilk Liua5682312015-08-07 15:19:50 +080021
22#define SPI_CFG0_REG 0x0000
23#define SPI_CFG1_REG 0x0004
24#define SPI_TX_SRC_REG 0x0008
25#define SPI_RX_DST_REG 0x000c
26#define SPI_TX_DATA_REG 0x0010
27#define SPI_RX_DATA_REG 0x0014
28#define SPI_CMD_REG 0x0018
29#define SPI_STATUS0_REG 0x001c
30#define SPI_PAD_SEL_REG 0x0024
Leilk Liu058fe492017-06-12 09:24:39 +080031#define SPI_CFG2_REG 0x0028
luhua.xufdeae8f2019-09-11 05:55:31 -040032#define SPI_TX_SRC_REG_64 0x002c
33#define SPI_RX_DST_REG_64 0x0030
Leilk Liua5682312015-08-07 15:19:50 +080034
35#define SPI_CFG0_SCK_HIGH_OFFSET 0
36#define SPI_CFG0_SCK_LOW_OFFSET 8
37#define SPI_CFG0_CS_HOLD_OFFSET 16
38#define SPI_CFG0_CS_SETUP_OFFSET 24
Leilk Liu058fe492017-06-12 09:24:39 +080039#define SPI_ADJUST_CFG0_CS_HOLD_OFFSET 0
40#define SPI_ADJUST_CFG0_CS_SETUP_OFFSET 16
Leilk Liua5682312015-08-07 15:19:50 +080041
42#define SPI_CFG1_CS_IDLE_OFFSET 0
43#define SPI_CFG1_PACKET_LOOP_OFFSET 8
44#define SPI_CFG1_PACKET_LENGTH_OFFSET 16
Mason Zhangf84d8662021-07-13 19:40:49 +080045#define SPI_CFG1_GET_TICK_DLY_OFFSET 29
Leilk Liua5682312015-08-07 15:19:50 +080046
Mason Zhangf84d8662021-07-13 19:40:49 +080047#define SPI_CFG1_GET_TICK_DLY_MASK 0xe0000000
Leilk Liua5682312015-08-07 15:19:50 +080048#define SPI_CFG1_CS_IDLE_MASK 0xff
49#define SPI_CFG1_PACKET_LOOP_MASK 0xff00
50#define SPI_CFG1_PACKET_LENGTH_MASK 0x3ff0000
leilk.liu44b37eb2020-07-01 17:00:20 +080051#define SPI_CFG2_SCK_HIGH_OFFSET 0
52#define SPI_CFG2_SCK_LOW_OFFSET 16
Leilk Liua5682312015-08-07 15:19:50 +080053
Leilk Liua71d6ea2015-08-20 17:19:08 +080054#define SPI_CMD_ACT BIT(0)
55#define SPI_CMD_RESUME BIT(1)
Leilk Liua5682312015-08-07 15:19:50 +080056#define SPI_CMD_RST BIT(2)
57#define SPI_CMD_PAUSE_EN BIT(4)
58#define SPI_CMD_DEASSERT BIT(5)
Leilk Liu058fe492017-06-12 09:24:39 +080059#define SPI_CMD_SAMPLE_SEL BIT(6)
60#define SPI_CMD_CS_POL BIT(7)
Leilk Liua5682312015-08-07 15:19:50 +080061#define SPI_CMD_CPHA BIT(8)
62#define SPI_CMD_CPOL BIT(9)
63#define SPI_CMD_RX_DMA BIT(10)
64#define SPI_CMD_TX_DMA BIT(11)
65#define SPI_CMD_TXMSBF BIT(12)
66#define SPI_CMD_RXMSBF BIT(13)
67#define SPI_CMD_RX_ENDIAN BIT(14)
68#define SPI_CMD_TX_ENDIAN BIT(15)
69#define SPI_CMD_FINISH_IE BIT(16)
70#define SPI_CMD_PAUSE_IE BIT(17)
71
Leilk Liua5682312015-08-07 15:19:50 +080072#define MT8173_SPI_MAX_PAD_SEL 3
73
Leilk Liu50f8fec2015-08-24 11:45:16 +080074#define MTK_SPI_PAUSE_INT_STATUS 0x2
75
Leilk Liua5682312015-08-07 15:19:50 +080076#define MTK_SPI_IDLE 0
77#define MTK_SPI_PAUSED 1
78
Daniel Kurtz1ce24862017-01-27 00:21:54 +080079#define MTK_SPI_MAX_FIFO_SIZE 32U
Leilk Liua5682312015-08-07 15:19:50 +080080#define MTK_SPI_PACKET_SIZE 1024
luhua.xufdeae8f2019-09-11 05:55:31 -040081#define MTK_SPI_32BITS_MASK (0xffffffff)
82
83#define DMA_ADDR_EXT_BITS (36)
84#define DMA_ADDR_DEF_BITS (32)
Leilk Liua5682312015-08-07 15:19:50 +080085
86struct mtk_spi_compatible {
Leilk Liuaf579372015-08-20 17:19:07 +080087 bool need_pad_sel;
88 /* Must explicitly send dummy Tx bytes to do Rx only transfer */
89 bool must_tx;
Leilk Liu058fe492017-06-12 09:24:39 +080090 /* some IC design adjust cfg register to enhance time accuracy */
91 bool enhance_timing;
luhua.xufdeae8f2019-09-11 05:55:31 -040092 /* some IC support DMA addr extension */
93 bool dma_ext;
Mason Zhang162a31e2021-06-29 18:08:15 +080094 /* some IC no need unprepare SPI clk */
95 bool no_need_unprepare;
Leilk Liua5682312015-08-07 15:19:50 +080096};
97
98struct mtk_spi {
99 void __iomem *base;
100 u32 state;
Leilk Liu37457602015-10-26 16:09:44 +0800101 int pad_num;
102 u32 *pad_sel;
Leilk Liuadcbcfe2015-08-31 21:18:57 +0800103 struct clk *parent_clk, *sel_clk, *spi_clk;
Leilk Liua5682312015-08-07 15:19:50 +0800104 struct spi_transfer *cur_transfer;
105 u32 xfer_len;
Peter Shih00bca732018-09-10 11:54:21 +0800106 u32 num_xfered;
Leilk Liua5682312015-08-07 15:19:50 +0800107 struct scatterlist *tx_sgl, *rx_sgl;
108 u32 tx_sgl_len, rx_sgl_len;
109 const struct mtk_spi_compatible *dev_comp;
Mason Zhang162a31e2021-06-29 18:08:15 +0800110 u32 spi_clk_hz;
Leilk Liua5682312015-08-07 15:19:50 +0800111};
112
Leilk Liu4eaf6f72015-12-31 10:59:00 +0800113static const struct mtk_spi_compatible mtk_common_compat;
Leilk Liufc4f2262017-06-12 09:24:40 +0800114
leilk.liu@mediatek.comb6b1f2d2017-06-20 16:21:07 +0800115static const struct mtk_spi_compatible mt2712_compat = {
116 .must_tx = true,
117};
118
luhua.xu2c231e02019-09-11 05:55:30 -0400119static const struct mtk_spi_compatible mt6765_compat = {
120 .need_pad_sel = true,
121 .must_tx = true,
122 .enhance_timing = true,
luhua.xufdeae8f2019-09-11 05:55:31 -0400123 .dma_ext = true,
luhua.xu2c231e02019-09-11 05:55:30 -0400124};
125
Leilk Liufc4f2262017-06-12 09:24:40 +0800126static const struct mtk_spi_compatible mt7622_compat = {
127 .must_tx = true,
128 .enhance_timing = true,
129};
130
Leilk Liua5682312015-08-07 15:19:50 +0800131static const struct mtk_spi_compatible mt8173_compat = {
Leilk Liuaf579372015-08-20 17:19:07 +0800132 .need_pad_sel = true,
133 .must_tx = true,
Leilk Liua5682312015-08-07 15:19:50 +0800134};
135
Leilk Liub654aa62018-11-01 14:02:19 +0800136static const struct mtk_spi_compatible mt8183_compat = {
137 .need_pad_sel = true,
138 .must_tx = true,
139 .enhance_timing = true,
140};
141
Mason Zhang162a31e2021-06-29 18:08:15 +0800142static const struct mtk_spi_compatible mt6893_compat = {
143 .need_pad_sel = true,
144 .must_tx = true,
145 .enhance_timing = true,
146 .dma_ext = true,
147 .no_need_unprepare = true,
148};
149
Leilk Liua5682312015-08-07 15:19:50 +0800150/*
151 * A piece of default chip info unless the platform
152 * supplies it.
153 */
154static const struct mtk_chip_config mtk_default_chip_info = {
Leilk Liu058fe492017-06-12 09:24:39 +0800155 .sample_sel = 0,
Mason Zhangf84d8662021-07-13 19:40:49 +0800156 .tick_delay = 0,
Leilk Liua5682312015-08-07 15:19:50 +0800157};
158
159static const struct of_device_id mtk_spi_of_match[] = {
Leilk Liu15bcdefd2015-12-31 10:59:01 +0800160 { .compatible = "mediatek,mt2701-spi",
161 .data = (void *)&mtk_common_compat,
162 },
leilk.liu@mediatek.comb6b1f2d2017-06-20 16:21:07 +0800163 { .compatible = "mediatek,mt2712-spi",
164 .data = (void *)&mt2712_compat,
165 },
Leilk Liu4eaf6f72015-12-31 10:59:00 +0800166 { .compatible = "mediatek,mt6589-spi",
167 .data = (void *)&mtk_common_compat,
168 },
luhua.xu2c231e02019-09-11 05:55:30 -0400169 { .compatible = "mediatek,mt6765-spi",
170 .data = (void *)&mt6765_compat,
171 },
Leilk Liufc4f2262017-06-12 09:24:40 +0800172 { .compatible = "mediatek,mt7622-spi",
173 .data = (void *)&mt7622_compat,
174 },
Leilk Liu942779c2018-11-20 16:41:08 +0800175 { .compatible = "mediatek,mt7629-spi",
176 .data = (void *)&mt7622_compat,
177 },
Leilk Liu4eaf6f72015-12-31 10:59:00 +0800178 { .compatible = "mediatek,mt8135-spi",
179 .data = (void *)&mtk_common_compat,
180 },
181 { .compatible = "mediatek,mt8173-spi",
182 .data = (void *)&mt8173_compat,
183 },
Leilk Liub654aa62018-11-01 14:02:19 +0800184 { .compatible = "mediatek,mt8183-spi",
185 .data = (void *)&mt8183_compat,
186 },
leilk.liu8cf125c2020-07-21 20:24:36 +0800187 { .compatible = "mediatek,mt8192-spi",
188 .data = (void *)&mt6765_compat,
189 },
Mason Zhang162a31e2021-06-29 18:08:15 +0800190 { .compatible = "mediatek,mt6893-spi",
191 .data = (void *)&mt6893_compat,
192 },
Leilk Liua5682312015-08-07 15:19:50 +0800193 {}
194};
195MODULE_DEVICE_TABLE(of, mtk_spi_of_match);
196
197static void mtk_spi_reset(struct mtk_spi *mdata)
198{
199 u32 reg_val;
200
201 /* set the software reset bit in SPI_CMD_REG. */
202 reg_val = readl(mdata->base + SPI_CMD_REG);
203 reg_val |= SPI_CMD_RST;
204 writel(reg_val, mdata->base + SPI_CMD_REG);
205
206 reg_val = readl(mdata->base + SPI_CMD_REG);
207 reg_val &= ~SPI_CMD_RST;
208 writel(reg_val, mdata->base + SPI_CMD_REG);
209}
210
Mason Zhang04e6bb02021-08-04 21:37:47 +0800211static int mtk_spi_set_hw_cs_timing(struct spi_device *spi)
212{
213 struct mtk_spi *mdata = spi_master_get_devdata(spi->master);
214 struct spi_delay *cs_setup = &spi->cs_setup;
215 struct spi_delay *cs_hold = &spi->cs_hold;
216 struct spi_delay *cs_inactive = &spi->cs_inactive;
Mason Zhang5c842e52021-08-09 13:59:12 +0800217 u32 setup, hold, inactive;
Mason Zhang04e6bb02021-08-04 21:37:47 +0800218 u32 reg_val;
219 int delay;
220
221 delay = spi_delay_to_ns(cs_setup, NULL);
222 if (delay < 0)
223 return delay;
224 setup = (delay * DIV_ROUND_UP(mdata->spi_clk_hz, 1000000)) / 1000;
225
226 delay = spi_delay_to_ns(cs_hold, NULL);
227 if (delay < 0)
228 return delay;
229 hold = (delay * DIV_ROUND_UP(mdata->spi_clk_hz, 1000000)) / 1000;
230
231 delay = spi_delay_to_ns(cs_inactive, NULL);
232 if (delay < 0)
233 return delay;
234 inactive = (delay * DIV_ROUND_UP(mdata->spi_clk_hz, 1000000)) / 1000;
235
236 setup = setup ? setup : 1;
237 hold = hold ? hold : 1;
238 inactive = inactive ? inactive : 1;
239
240 reg_val = readl(mdata->base + SPI_CFG0_REG);
241 if (mdata->dev_comp->enhance_timing) {
Mason Zhang5c842e52021-08-09 13:59:12 +0800242 hold = min_t(u32, hold, 0x10000);
243 setup = min_t(u32, setup, 0x10000);
Mason Zhang04e6bb02021-08-04 21:37:47 +0800244 reg_val &= ~(0xffff << SPI_ADJUST_CFG0_CS_HOLD_OFFSET);
245 reg_val |= (((hold - 1) & 0xffff)
246 << SPI_ADJUST_CFG0_CS_HOLD_OFFSET);
247 reg_val &= ~(0xffff << SPI_ADJUST_CFG0_CS_SETUP_OFFSET);
248 reg_val |= (((setup - 1) & 0xffff)
249 << SPI_ADJUST_CFG0_CS_SETUP_OFFSET);
250 } else {
Mason Zhang5c842e52021-08-09 13:59:12 +0800251 hold = min_t(u32, hold, 0x100);
252 setup = min_t(u32, setup, 0x100);
Mason Zhang04e6bb02021-08-04 21:37:47 +0800253 reg_val &= ~(0xff << SPI_CFG0_CS_HOLD_OFFSET);
254 reg_val |= (((hold - 1) & 0xff) << SPI_CFG0_CS_HOLD_OFFSET);
255 reg_val &= ~(0xff << SPI_CFG0_CS_SETUP_OFFSET);
256 reg_val |= (((setup - 1) & 0xff)
257 << SPI_CFG0_CS_SETUP_OFFSET);
258 }
259 writel(reg_val, mdata->base + SPI_CFG0_REG);
260
Mason Zhang5c842e52021-08-09 13:59:12 +0800261 inactive = min_t(u32, inactive, 0x100);
Mason Zhang04e6bb02021-08-04 21:37:47 +0800262 reg_val = readl(mdata->base + SPI_CFG1_REG);
263 reg_val &= ~SPI_CFG1_CS_IDLE_MASK;
264 reg_val |= (((inactive - 1) & 0xff) << SPI_CFG1_CS_IDLE_OFFSET);
265 writel(reg_val, mdata->base + SPI_CFG1_REG);
266
267 return 0;
268}
269
Leilk Liu79b5d3f2015-10-26 16:09:41 +0800270static int mtk_spi_prepare_message(struct spi_master *master,
271 struct spi_message *msg)
Leilk Liua5682312015-08-07 15:19:50 +0800272{
Leilk Liu79b5d3f2015-10-26 16:09:41 +0800273 u16 cpha, cpol;
Leilk Liua5682312015-08-07 15:19:50 +0800274 u32 reg_val;
Leilk Liu79b5d3f2015-10-26 16:09:41 +0800275 struct spi_device *spi = msg->spi;
Leilk Liu58a984c72015-10-26 16:09:43 +0800276 struct mtk_chip_config *chip_config = spi->controller_data;
Leilk Liu79b5d3f2015-10-26 16:09:41 +0800277 struct mtk_spi *mdata = spi_master_get_devdata(master);
278
279 cpha = spi->mode & SPI_CPHA ? 1 : 0;
280 cpol = spi->mode & SPI_CPOL ? 1 : 0;
281
Leilk Liu79b5d3f2015-10-26 16:09:41 +0800282 reg_val = readl(mdata->base + SPI_CMD_REG);
283 if (cpha)
284 reg_val |= SPI_CMD_CPHA;
285 else
286 reg_val &= ~SPI_CMD_CPHA;
287 if (cpol)
288 reg_val |= SPI_CMD_CPOL;
289 else
290 reg_val &= ~SPI_CMD_CPOL;
Leilk Liua5682312015-08-07 15:19:50 +0800291
292 /* set the mlsbx and mlsbtx */
Leilk Liu3e582c62019-06-05 11:07:04 +0800293 if (spi->mode & SPI_LSB_FIRST) {
Leilk Liua71d6ea2015-08-20 17:19:08 +0800294 reg_val &= ~SPI_CMD_TXMSBF;
Leilk Liua71d6ea2015-08-20 17:19:08 +0800295 reg_val &= ~SPI_CMD_RXMSBF;
Leilk Liu3e582c62019-06-05 11:07:04 +0800296 } else {
297 reg_val |= SPI_CMD_TXMSBF;
298 reg_val |= SPI_CMD_RXMSBF;
299 }
Leilk Liua5682312015-08-07 15:19:50 +0800300
301 /* set the tx/rx endian */
Leilk Liu44f636d2015-08-20 17:19:06 +0800302#ifdef __LITTLE_ENDIAN
303 reg_val &= ~SPI_CMD_TX_ENDIAN;
304 reg_val &= ~SPI_CMD_RX_ENDIAN;
305#else
306 reg_val |= SPI_CMD_TX_ENDIAN;
307 reg_val |= SPI_CMD_RX_ENDIAN;
308#endif
Leilk Liua5682312015-08-07 15:19:50 +0800309
Leilk Liu058fe492017-06-12 09:24:39 +0800310 if (mdata->dev_comp->enhance_timing) {
Luhua Xuae7c2d32019-11-18 12:57:16 +0800311 /* set CS polarity */
312 if (spi->mode & SPI_CS_HIGH)
Leilk Liu058fe492017-06-12 09:24:39 +0800313 reg_val |= SPI_CMD_CS_POL;
314 else
315 reg_val &= ~SPI_CMD_CS_POL;
Luhua Xuae7c2d32019-11-18 12:57:16 +0800316
Leilk Liu058fe492017-06-12 09:24:39 +0800317 if (chip_config->sample_sel)
318 reg_val |= SPI_CMD_SAMPLE_SEL;
319 else
320 reg_val &= ~SPI_CMD_SAMPLE_SEL;
321 }
322
Leilk Liua5682312015-08-07 15:19:50 +0800323 /* set finish and pause interrupt always enable */
Leilk Liu15293322015-08-27 21:09:04 +0800324 reg_val |= SPI_CMD_FINISH_IE | SPI_CMD_PAUSE_IE;
Leilk Liua5682312015-08-07 15:19:50 +0800325
326 /* disable dma mode */
327 reg_val &= ~(SPI_CMD_TX_DMA | SPI_CMD_RX_DMA);
328
329 /* disable deassert mode */
330 reg_val &= ~SPI_CMD_DEASSERT;
331
332 writel(reg_val, mdata->base + SPI_CMD_REG);
333
334 /* pad select */
335 if (mdata->dev_comp->need_pad_sel)
Leilk Liu37457602015-10-26 16:09:44 +0800336 writel(mdata->pad_sel[spi->chip_select],
337 mdata->base + SPI_PAD_SEL_REG);
Leilk Liua5682312015-08-07 15:19:50 +0800338
Mason Zhangf84d8662021-07-13 19:40:49 +0800339 /* tick delay */
340 reg_val = readl(mdata->base + SPI_CFG1_REG);
341 reg_val &= ~SPI_CFG1_GET_TICK_DLY_MASK;
342 reg_val |= ((chip_config->tick_delay & 0x7)
343 << SPI_CFG1_GET_TICK_DLY_OFFSET);
344 writel(reg_val, mdata->base + SPI_CFG1_REG);
345
Mason Zhang04e6bb02021-08-04 21:37:47 +0800346 /* set hw cs timing */
347 mtk_spi_set_hw_cs_timing(spi);
Leilk Liua5682312015-08-07 15:19:50 +0800348 return 0;
349}
350
351static void mtk_spi_set_cs(struct spi_device *spi, bool enable)
352{
353 u32 reg_val;
354 struct mtk_spi *mdata = spi_master_get_devdata(spi->master);
355
Luhua Xuae7c2d32019-11-18 12:57:16 +0800356 if (spi->mode & SPI_CS_HIGH)
357 enable = !enable;
358
Leilk Liua5682312015-08-07 15:19:50 +0800359 reg_val = readl(mdata->base + SPI_CMD_REG);
Leilk Liu6583d202015-09-07 19:37:57 +0800360 if (!enable) {
Leilk Liua5682312015-08-07 15:19:50 +0800361 reg_val |= SPI_CMD_PAUSE_EN;
Leilk Liu6583d202015-09-07 19:37:57 +0800362 writel(reg_val, mdata->base + SPI_CMD_REG);
363 } else {
Leilk Liua5682312015-08-07 15:19:50 +0800364 reg_val &= ~SPI_CMD_PAUSE_EN;
Leilk Liu6583d202015-09-07 19:37:57 +0800365 writel(reg_val, mdata->base + SPI_CMD_REG);
366 mdata->state = MTK_SPI_IDLE;
367 mtk_spi_reset(mdata);
368 }
Leilk Liua5682312015-08-07 15:19:50 +0800369}
370
371static void mtk_spi_prepare_transfer(struct spi_master *master,
372 struct spi_transfer *xfer)
373{
Mason Zhang162a31e2021-06-29 18:08:15 +0800374 u32 div, sck_time, reg_val;
Leilk Liua5682312015-08-07 15:19:50 +0800375 struct mtk_spi *mdata = spi_master_get_devdata(master);
376
Mason Zhang162a31e2021-06-29 18:08:15 +0800377 if (xfer->speed_hz < mdata->spi_clk_hz / 2)
378 div = DIV_ROUND_UP(mdata->spi_clk_hz, xfer->speed_hz);
Leilk Liua5682312015-08-07 15:19:50 +0800379 else
380 div = 1;
381
Leilk Liu2ce0acf2015-08-24 11:45:18 +0800382 sck_time = (div + 1) / 2;
Leilk Liua5682312015-08-07 15:19:50 +0800383
Leilk Liu058fe492017-06-12 09:24:39 +0800384 if (mdata->dev_comp->enhance_timing) {
leilk.liu9f6e7e82021-02-07 11:09:53 +0800385 reg_val = readl(mdata->base + SPI_CFG2_REG);
386 reg_val &= ~(0xffff << SPI_CFG2_SCK_HIGH_OFFSET);
387 reg_val |= (((sck_time - 1) & 0xffff)
leilk.liu44b37eb2020-07-01 17:00:20 +0800388 << SPI_CFG2_SCK_HIGH_OFFSET);
leilk.liu9f6e7e82021-02-07 11:09:53 +0800389 reg_val &= ~(0xffff << SPI_CFG2_SCK_LOW_OFFSET);
Leilk Liu058fe492017-06-12 09:24:39 +0800390 reg_val |= (((sck_time - 1) & 0xffff)
leilk.liu44b37eb2020-07-01 17:00:20 +0800391 << SPI_CFG2_SCK_LOW_OFFSET);
Leilk Liu058fe492017-06-12 09:24:39 +0800392 writel(reg_val, mdata->base + SPI_CFG2_REG);
Leilk Liu058fe492017-06-12 09:24:39 +0800393 } else {
leilk.liu9f6e7e82021-02-07 11:09:53 +0800394 reg_val = readl(mdata->base + SPI_CFG0_REG);
395 reg_val &= ~(0xff << SPI_CFG0_SCK_HIGH_OFFSET);
396 reg_val |= (((sck_time - 1) & 0xff)
Leilk Liu058fe492017-06-12 09:24:39 +0800397 << SPI_CFG0_SCK_HIGH_OFFSET);
leilk.liu9f6e7e82021-02-07 11:09:53 +0800398 reg_val &= ~(0xff << SPI_CFG0_SCK_LOW_OFFSET);
Leilk Liu058fe492017-06-12 09:24:39 +0800399 reg_val |= (((sck_time - 1) & 0xff) << SPI_CFG0_SCK_LOW_OFFSET);
Leilk Liu058fe492017-06-12 09:24:39 +0800400 writel(reg_val, mdata->base + SPI_CFG0_REG);
401 }
Leilk Liua5682312015-08-07 15:19:50 +0800402}
403
404static void mtk_spi_setup_packet(struct spi_master *master)
405{
406 u32 packet_size, packet_loop, reg_val;
407 struct mtk_spi *mdata = spi_master_get_devdata(master);
408
Leilk Liu50f8fec2015-08-24 11:45:16 +0800409 packet_size = min_t(u32, mdata->xfer_len, MTK_SPI_PACKET_SIZE);
Leilk Liua5682312015-08-07 15:19:50 +0800410 packet_loop = mdata->xfer_len / packet_size;
411
412 reg_val = readl(mdata->base + SPI_CFG1_REG);
Leilk Liu50f8fec2015-08-24 11:45:16 +0800413 reg_val &= ~(SPI_CFG1_PACKET_LENGTH_MASK | SPI_CFG1_PACKET_LOOP_MASK);
Leilk Liua5682312015-08-07 15:19:50 +0800414 reg_val |= (packet_size - 1) << SPI_CFG1_PACKET_LENGTH_OFFSET;
415 reg_val |= (packet_loop - 1) << SPI_CFG1_PACKET_LOOP_OFFSET;
416 writel(reg_val, mdata->base + SPI_CFG1_REG);
417}
418
419static void mtk_spi_enable_transfer(struct spi_master *master)
420{
Leilk Liu50f8fec2015-08-24 11:45:16 +0800421 u32 cmd;
Leilk Liua5682312015-08-07 15:19:50 +0800422 struct mtk_spi *mdata = spi_master_get_devdata(master);
423
424 cmd = readl(mdata->base + SPI_CMD_REG);
425 if (mdata->state == MTK_SPI_IDLE)
Leilk Liua71d6ea2015-08-20 17:19:08 +0800426 cmd |= SPI_CMD_ACT;
Leilk Liua5682312015-08-07 15:19:50 +0800427 else
Leilk Liua71d6ea2015-08-20 17:19:08 +0800428 cmd |= SPI_CMD_RESUME;
Leilk Liua5682312015-08-07 15:19:50 +0800429 writel(cmd, mdata->base + SPI_CMD_REG);
430}
431
Leilk Liu50f8fec2015-08-24 11:45:16 +0800432static int mtk_spi_get_mult_delta(u32 xfer_len)
Leilk Liua5682312015-08-07 15:19:50 +0800433{
Leilk Liu50f8fec2015-08-24 11:45:16 +0800434 u32 mult_delta;
Leilk Liua5682312015-08-07 15:19:50 +0800435
436 if (xfer_len > MTK_SPI_PACKET_SIZE)
437 mult_delta = xfer_len % MTK_SPI_PACKET_SIZE;
438 else
439 mult_delta = 0;
440
441 return mult_delta;
442}
443
444static void mtk_spi_update_mdata_len(struct spi_master *master)
445{
446 int mult_delta;
447 struct mtk_spi *mdata = spi_master_get_devdata(master);
448
449 if (mdata->tx_sgl_len && mdata->rx_sgl_len) {
450 if (mdata->tx_sgl_len > mdata->rx_sgl_len) {
451 mult_delta = mtk_spi_get_mult_delta(mdata->rx_sgl_len);
452 mdata->xfer_len = mdata->rx_sgl_len - mult_delta;
453 mdata->rx_sgl_len = mult_delta;
454 mdata->tx_sgl_len -= mdata->xfer_len;
455 } else {
456 mult_delta = mtk_spi_get_mult_delta(mdata->tx_sgl_len);
457 mdata->xfer_len = mdata->tx_sgl_len - mult_delta;
458 mdata->tx_sgl_len = mult_delta;
459 mdata->rx_sgl_len -= mdata->xfer_len;
460 }
461 } else if (mdata->tx_sgl_len) {
462 mult_delta = mtk_spi_get_mult_delta(mdata->tx_sgl_len);
463 mdata->xfer_len = mdata->tx_sgl_len - mult_delta;
464 mdata->tx_sgl_len = mult_delta;
465 } else if (mdata->rx_sgl_len) {
466 mult_delta = mtk_spi_get_mult_delta(mdata->rx_sgl_len);
467 mdata->xfer_len = mdata->rx_sgl_len - mult_delta;
468 mdata->rx_sgl_len = mult_delta;
469 }
470}
471
472static void mtk_spi_setup_dma_addr(struct spi_master *master,
473 struct spi_transfer *xfer)
474{
475 struct mtk_spi *mdata = spi_master_get_devdata(master);
476
luhua.xufdeae8f2019-09-11 05:55:31 -0400477 if (mdata->tx_sgl) {
478 writel((u32)(xfer->tx_dma & MTK_SPI_32BITS_MASK),
479 mdata->base + SPI_TX_SRC_REG);
480#ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
481 if (mdata->dev_comp->dma_ext)
482 writel((u32)(xfer->tx_dma >> 32),
483 mdata->base + SPI_TX_SRC_REG_64);
484#endif
485 }
486
487 if (mdata->rx_sgl) {
488 writel((u32)(xfer->rx_dma & MTK_SPI_32BITS_MASK),
489 mdata->base + SPI_RX_DST_REG);
490#ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
491 if (mdata->dev_comp->dma_ext)
492 writel((u32)(xfer->rx_dma >> 32),
493 mdata->base + SPI_RX_DST_REG_64);
494#endif
495 }
Leilk Liua5682312015-08-07 15:19:50 +0800496}
497
498static int mtk_spi_fifo_transfer(struct spi_master *master,
499 struct spi_device *spi,
500 struct spi_transfer *xfer)
501{
Nicolas Boichatde327e42015-12-27 18:17:06 +0800502 int cnt, remainder;
503 u32 reg_val;
Leilk Liua5682312015-08-07 15:19:50 +0800504 struct mtk_spi *mdata = spi_master_get_devdata(master);
505
506 mdata->cur_transfer = xfer;
Daniel Kurtz1ce24862017-01-27 00:21:54 +0800507 mdata->xfer_len = min(MTK_SPI_MAX_FIFO_SIZE, xfer->len);
Peter Shih00bca732018-09-10 11:54:21 +0800508 mdata->num_xfered = 0;
Leilk Liua5682312015-08-07 15:19:50 +0800509 mtk_spi_prepare_transfer(master, xfer);
510 mtk_spi_setup_packet(master);
511
Guenter Roeck0d5c3952021-08-01 20:00:23 -0700512 if (xfer->tx_buf) {
513 cnt = xfer->len / 4;
Peter Hess3a70dd22021-07-06 14:16:09 +0200514 iowrite32_rep(mdata->base + SPI_TX_DATA_REG, xfer->tx_buf, cnt);
Guenter Roeck0d5c3952021-08-01 20:00:23 -0700515 remainder = xfer->len % 4;
516 if (remainder > 0) {
517 reg_val = 0;
Peter Hess3a70dd22021-07-06 14:16:09 +0200518 memcpy(&reg_val, xfer->tx_buf + (cnt * 4), remainder);
519 writel(reg_val, mdata->base + SPI_TX_DATA_REG);
520 }
Nicolas Boichatde327e42015-12-27 18:17:06 +0800521 }
522
Leilk Liua5682312015-08-07 15:19:50 +0800523 mtk_spi_enable_transfer(master);
524
525 return 1;
526}
527
528static int mtk_spi_dma_transfer(struct spi_master *master,
529 struct spi_device *spi,
530 struct spi_transfer *xfer)
531{
532 int cmd;
533 struct mtk_spi *mdata = spi_master_get_devdata(master);
534
535 mdata->tx_sgl = NULL;
536 mdata->rx_sgl = NULL;
537 mdata->tx_sgl_len = 0;
538 mdata->rx_sgl_len = 0;
539 mdata->cur_transfer = xfer;
Peter Shih00bca732018-09-10 11:54:21 +0800540 mdata->num_xfered = 0;
Leilk Liua5682312015-08-07 15:19:50 +0800541
542 mtk_spi_prepare_transfer(master, xfer);
543
544 cmd = readl(mdata->base + SPI_CMD_REG);
545 if (xfer->tx_buf)
546 cmd |= SPI_CMD_TX_DMA;
547 if (xfer->rx_buf)
548 cmd |= SPI_CMD_RX_DMA;
549 writel(cmd, mdata->base + SPI_CMD_REG);
550
551 if (xfer->tx_buf)
552 mdata->tx_sgl = xfer->tx_sg.sgl;
553 if (xfer->rx_buf)
554 mdata->rx_sgl = xfer->rx_sg.sgl;
555
556 if (mdata->tx_sgl) {
557 xfer->tx_dma = sg_dma_address(mdata->tx_sgl);
558 mdata->tx_sgl_len = sg_dma_len(mdata->tx_sgl);
559 }
560 if (mdata->rx_sgl) {
561 xfer->rx_dma = sg_dma_address(mdata->rx_sgl);
562 mdata->rx_sgl_len = sg_dma_len(mdata->rx_sgl);
563 }
564
565 mtk_spi_update_mdata_len(master);
566 mtk_spi_setup_packet(master);
567 mtk_spi_setup_dma_addr(master, xfer);
568 mtk_spi_enable_transfer(master);
569
570 return 1;
571}
572
573static int mtk_spi_transfer_one(struct spi_master *master,
574 struct spi_device *spi,
575 struct spi_transfer *xfer)
576{
577 if (master->can_dma(master, spi, xfer))
578 return mtk_spi_dma_transfer(master, spi, xfer);
579 else
580 return mtk_spi_fifo_transfer(master, spi, xfer);
581}
582
583static bool mtk_spi_can_dma(struct spi_master *master,
584 struct spi_device *spi,
585 struct spi_transfer *xfer)
586{
Daniel Kurtz1ce24862017-01-27 00:21:54 +0800587 /* Buffers for DMA transactions must be 4-byte aligned */
588 return (xfer->len > MTK_SPI_MAX_FIFO_SIZE &&
589 (unsigned long)xfer->tx_buf % 4 == 0 &&
590 (unsigned long)xfer->rx_buf % 4 == 0);
Leilk Liua5682312015-08-07 15:19:50 +0800591}
592
Leilk Liu58a984c72015-10-26 16:09:43 +0800593static int mtk_spi_setup(struct spi_device *spi)
594{
595 struct mtk_spi *mdata = spi_master_get_devdata(spi->master);
596
597 if (!spi->controller_data)
598 spi->controller_data = (void *)&mtk_default_chip_info;
599
Nicolas Boichat98c8dcc2015-11-09 12:14:51 +0800600 if (mdata->dev_comp->need_pad_sel && gpio_is_valid(spi->cs_gpio))
Leilk Liu37457602015-10-26 16:09:44 +0800601 gpio_direction_output(spi->cs_gpio, !(spi->mode & SPI_CS_HIGH));
602
Leilk Liu58a984c72015-10-26 16:09:43 +0800603 return 0;
604}
605
Leilk Liua5682312015-08-07 15:19:50 +0800606static irqreturn_t mtk_spi_interrupt(int irq, void *dev_id)
607{
Peter Shih00bca732018-09-10 11:54:21 +0800608 u32 cmd, reg_val, cnt, remainder, len;
Leilk Liua5682312015-08-07 15:19:50 +0800609 struct spi_master *master = dev_id;
610 struct mtk_spi *mdata = spi_master_get_devdata(master);
611 struct spi_transfer *trans = mdata->cur_transfer;
612
613 reg_val = readl(mdata->base + SPI_STATUS0_REG);
Leilk Liu50f8fec2015-08-24 11:45:16 +0800614 if (reg_val & MTK_SPI_PAUSE_INT_STATUS)
Leilk Liua5682312015-08-07 15:19:50 +0800615 mdata->state = MTK_SPI_PAUSED;
616 else
617 mdata->state = MTK_SPI_IDLE;
618
619 if (!master->can_dma(master, master->cur_msg->spi, trans)) {
Leilk Liua5682312015-08-07 15:19:50 +0800620 if (trans->rx_buf) {
Nicolas Boichatde327e42015-12-27 18:17:06 +0800621 cnt = mdata->xfer_len / 4;
Leilk Liu44f636d2015-08-20 17:19:06 +0800622 ioread32_rep(mdata->base + SPI_RX_DATA_REG,
Peter Shih00bca732018-09-10 11:54:21 +0800623 trans->rx_buf + mdata->num_xfered, cnt);
Nicolas Boichatde327e42015-12-27 18:17:06 +0800624 remainder = mdata->xfer_len % 4;
625 if (remainder > 0) {
626 reg_val = readl(mdata->base + SPI_RX_DATA_REG);
Peter Shih00bca732018-09-10 11:54:21 +0800627 memcpy(trans->rx_buf +
628 mdata->num_xfered +
629 (cnt * 4),
630 &reg_val,
631 remainder);
Nicolas Boichatde327e42015-12-27 18:17:06 +0800632 }
Leilk Liua5682312015-08-07 15:19:50 +0800633 }
Daniel Kurtz1ce24862017-01-27 00:21:54 +0800634
Peter Shih00bca732018-09-10 11:54:21 +0800635 mdata->num_xfered += mdata->xfer_len;
636 if (mdata->num_xfered == trans->len) {
Daniel Kurtz1ce24862017-01-27 00:21:54 +0800637 spi_finalize_current_transfer(master);
638 return IRQ_HANDLED;
639 }
640
Peter Shih00bca732018-09-10 11:54:21 +0800641 len = trans->len - mdata->num_xfered;
642 mdata->xfer_len = min(MTK_SPI_MAX_FIFO_SIZE, len);
Daniel Kurtz1ce24862017-01-27 00:21:54 +0800643 mtk_spi_setup_packet(master);
644
Leilk Liua4d8f642018-10-31 16:49:16 +0800645 cnt = mdata->xfer_len / 4;
Peter Shih00bca732018-09-10 11:54:21 +0800646 iowrite32_rep(mdata->base + SPI_TX_DATA_REG,
647 trans->tx_buf + mdata->num_xfered, cnt);
Daniel Kurtz1ce24862017-01-27 00:21:54 +0800648
Leilk Liua4d8f642018-10-31 16:49:16 +0800649 remainder = mdata->xfer_len % 4;
Daniel Kurtz1ce24862017-01-27 00:21:54 +0800650 if (remainder > 0) {
651 reg_val = 0;
Peter Shih00bca732018-09-10 11:54:21 +0800652 memcpy(&reg_val,
653 trans->tx_buf + (cnt * 4) + mdata->num_xfered,
654 remainder);
Daniel Kurtz1ce24862017-01-27 00:21:54 +0800655 writel(reg_val, mdata->base + SPI_TX_DATA_REG);
656 }
657
658 mtk_spi_enable_transfer(master);
659
Leilk Liua5682312015-08-07 15:19:50 +0800660 return IRQ_HANDLED;
661 }
662
663 if (mdata->tx_sgl)
664 trans->tx_dma += mdata->xfer_len;
665 if (mdata->rx_sgl)
666 trans->rx_dma += mdata->xfer_len;
667
668 if (mdata->tx_sgl && (mdata->tx_sgl_len == 0)) {
669 mdata->tx_sgl = sg_next(mdata->tx_sgl);
670 if (mdata->tx_sgl) {
671 trans->tx_dma = sg_dma_address(mdata->tx_sgl);
672 mdata->tx_sgl_len = sg_dma_len(mdata->tx_sgl);
673 }
674 }
675 if (mdata->rx_sgl && (mdata->rx_sgl_len == 0)) {
676 mdata->rx_sgl = sg_next(mdata->rx_sgl);
677 if (mdata->rx_sgl) {
678 trans->rx_dma = sg_dma_address(mdata->rx_sgl);
679 mdata->rx_sgl_len = sg_dma_len(mdata->rx_sgl);
680 }
681 }
682
683 if (!mdata->tx_sgl && !mdata->rx_sgl) {
684 /* spi disable dma */
685 cmd = readl(mdata->base + SPI_CMD_REG);
686 cmd &= ~SPI_CMD_TX_DMA;
687 cmd &= ~SPI_CMD_RX_DMA;
688 writel(cmd, mdata->base + SPI_CMD_REG);
689
690 spi_finalize_current_transfer(master);
691 return IRQ_HANDLED;
692 }
693
694 mtk_spi_update_mdata_len(master);
695 mtk_spi_setup_packet(master);
696 mtk_spi_setup_dma_addr(master, trans);
697 mtk_spi_enable_transfer(master);
698
699 return IRQ_HANDLED;
700}
701
702static int mtk_spi_probe(struct platform_device *pdev)
703{
704 struct spi_master *master;
705 struct mtk_spi *mdata;
706 const struct of_device_id *of_id;
luhua.xufdeae8f2019-09-11 05:55:31 -0400707 int i, irq, ret, addr_bits;
Leilk Liua5682312015-08-07 15:19:50 +0800708
709 master = spi_alloc_master(&pdev->dev, sizeof(*mdata));
710 if (!master) {
711 dev_err(&pdev->dev, "failed to alloc spi master\n");
712 return -ENOMEM;
713 }
714
715 master->auto_runtime_pm = true;
716 master->dev.of_node = pdev->dev.of_node;
Leilk Liu3e582c62019-06-05 11:07:04 +0800717 master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_LSB_FIRST;
Leilk Liua5682312015-08-07 15:19:50 +0800718
719 master->set_cs = mtk_spi_set_cs;
Leilk Liua5682312015-08-07 15:19:50 +0800720 master->prepare_message = mtk_spi_prepare_message;
721 master->transfer_one = mtk_spi_transfer_one;
722 master->can_dma = mtk_spi_can_dma;
Leilk Liu58a984c72015-10-26 16:09:43 +0800723 master->setup = mtk_spi_setup;
leilk.liu9f6e7e82021-02-07 11:09:53 +0800724 master->set_cs_timing = mtk_spi_set_hw_cs_timing;
Leilk Liua5682312015-08-07 15:19:50 +0800725
726 of_id = of_match_node(mtk_spi_of_match, pdev->dev.of_node);
727 if (!of_id) {
728 dev_err(&pdev->dev, "failed to probe of_node\n");
729 ret = -EINVAL;
730 goto err_put_master;
731 }
732
733 mdata = spi_master_get_devdata(master);
734 mdata->dev_comp = of_id->data;
Luhua Xuae7c2d32019-11-18 12:57:16 +0800735
736 if (mdata->dev_comp->enhance_timing)
737 master->mode_bits |= SPI_CS_HIGH;
738
Leilk Liua5682312015-08-07 15:19:50 +0800739 if (mdata->dev_comp->must_tx)
740 master->flags = SPI_MASTER_MUST_TX;
741
742 if (mdata->dev_comp->need_pad_sel) {
Leilk Liu37457602015-10-26 16:09:44 +0800743 mdata->pad_num = of_property_count_u32_elems(
744 pdev->dev.of_node,
745 "mediatek,pad-select");
746 if (mdata->pad_num < 0) {
747 dev_err(&pdev->dev,
748 "No 'mediatek,pad-select' property\n");
749 ret = -EINVAL;
Leilk Liua5682312015-08-07 15:19:50 +0800750 goto err_put_master;
751 }
752
Leilk Liu37457602015-10-26 16:09:44 +0800753 mdata->pad_sel = devm_kmalloc_array(&pdev->dev, mdata->pad_num,
754 sizeof(u32), GFP_KERNEL);
755 if (!mdata->pad_sel) {
756 ret = -ENOMEM;
Leilk Liua5682312015-08-07 15:19:50 +0800757 goto err_put_master;
758 }
Leilk Liu37457602015-10-26 16:09:44 +0800759
760 for (i = 0; i < mdata->pad_num; i++) {
761 of_property_read_u32_index(pdev->dev.of_node,
762 "mediatek,pad-select",
763 i, &mdata->pad_sel[i]);
764 if (mdata->pad_sel[i] > MT8173_SPI_MAX_PAD_SEL) {
765 dev_err(&pdev->dev, "wrong pad-sel[%d]: %u\n",
766 i, mdata->pad_sel[i]);
767 ret = -EINVAL;
768 goto err_put_master;
769 }
770 }
Leilk Liua5682312015-08-07 15:19:50 +0800771 }
772
773 platform_set_drvdata(pdev, master);
Markus Elfring5dd381e2019-09-21 14:45:40 +0200774 mdata->base = devm_platform_ioremap_resource(pdev, 0);
Leilk Liua5682312015-08-07 15:19:50 +0800775 if (IS_ERR(mdata->base)) {
776 ret = PTR_ERR(mdata->base);
777 goto err_put_master;
778 }
779
780 irq = platform_get_irq(pdev, 0);
781 if (irq < 0) {
Leilk Liua5682312015-08-07 15:19:50 +0800782 ret = irq;
783 goto err_put_master;
784 }
785
786 if (!pdev->dev.dma_mask)
787 pdev->dev.dma_mask = &pdev->dev.coherent_dma_mask;
788
789 ret = devm_request_irq(&pdev->dev, irq, mtk_spi_interrupt,
790 IRQF_TRIGGER_NONE, dev_name(&pdev->dev), master);
791 if (ret) {
792 dev_err(&pdev->dev, "failed to register irq (%d)\n", ret);
793 goto err_put_master;
794 }
795
Leilk Liua5682312015-08-07 15:19:50 +0800796 mdata->parent_clk = devm_clk_get(&pdev->dev, "parent-clk");
797 if (IS_ERR(mdata->parent_clk)) {
798 ret = PTR_ERR(mdata->parent_clk);
799 dev_err(&pdev->dev, "failed to get parent-clk: %d\n", ret);
800 goto err_put_master;
801 }
802
Leilk Liuadcbcfe2015-08-31 21:18:57 +0800803 mdata->sel_clk = devm_clk_get(&pdev->dev, "sel-clk");
804 if (IS_ERR(mdata->sel_clk)) {
Javier Martinez Canillase26d15f2015-09-15 14:46:45 +0200805 ret = PTR_ERR(mdata->sel_clk);
Leilk Liuadcbcfe2015-08-31 21:18:57 +0800806 dev_err(&pdev->dev, "failed to get sel-clk: %d\n", ret);
807 goto err_put_master;
808 }
809
810 mdata->spi_clk = devm_clk_get(&pdev->dev, "spi-clk");
811 if (IS_ERR(mdata->spi_clk)) {
Javier Martinez Canillase26d15f2015-09-15 14:46:45 +0200812 ret = PTR_ERR(mdata->spi_clk);
Leilk Liuadcbcfe2015-08-31 21:18:57 +0800813 dev_err(&pdev->dev, "failed to get spi-clk: %d\n", ret);
814 goto err_put_master;
815 }
816
Leilk Liua5682312015-08-07 15:19:50 +0800817 ret = clk_prepare_enable(mdata->spi_clk);
818 if (ret < 0) {
819 dev_err(&pdev->dev, "failed to enable spi_clk (%d)\n", ret);
820 goto err_put_master;
821 }
822
Leilk Liuadcbcfe2015-08-31 21:18:57 +0800823 ret = clk_set_parent(mdata->sel_clk, mdata->parent_clk);
Leilk Liua5682312015-08-07 15:19:50 +0800824 if (ret < 0) {
825 dev_err(&pdev->dev, "failed to clk_set_parent (%d)\n", ret);
Leilk Liue38da372015-11-25 17:50:38 +0800826 clk_disable_unprepare(mdata->spi_clk);
827 goto err_put_master;
Leilk Liua5682312015-08-07 15:19:50 +0800828 }
829
Mason Zhang162a31e2021-06-29 18:08:15 +0800830 mdata->spi_clk_hz = clk_get_rate(mdata->spi_clk);
831
832 if (mdata->dev_comp->no_need_unprepare)
833 clk_disable(mdata->spi_clk);
834 else
835 clk_disable_unprepare(mdata->spi_clk);
Leilk Liua5682312015-08-07 15:19:50 +0800836
837 pm_runtime_enable(&pdev->dev);
838
Leilk Liu37457602015-10-26 16:09:44 +0800839 if (mdata->dev_comp->need_pad_sel) {
840 if (mdata->pad_num != master->num_chipselect) {
841 dev_err(&pdev->dev,
842 "pad_num does not match num_chipselect(%d != %d)\n",
843 mdata->pad_num, master->num_chipselect);
844 ret = -EINVAL;
Leilk Liue38da372015-11-25 17:50:38 +0800845 goto err_disable_runtime_pm;
Leilk Liu37457602015-10-26 16:09:44 +0800846 }
847
Nicolas Boichat98c8dcc2015-11-09 12:14:51 +0800848 if (!master->cs_gpios && master->num_chipselect > 1) {
849 dev_err(&pdev->dev,
850 "cs_gpios not specified and num_chipselect > 1\n");
851 ret = -EINVAL;
Leilk Liue38da372015-11-25 17:50:38 +0800852 goto err_disable_runtime_pm;
Nicolas Boichat98c8dcc2015-11-09 12:14:51 +0800853 }
854
855 if (master->cs_gpios) {
856 for (i = 0; i < master->num_chipselect; i++) {
857 ret = devm_gpio_request(&pdev->dev,
858 master->cs_gpios[i],
859 dev_name(&pdev->dev));
860 if (ret) {
861 dev_err(&pdev->dev,
862 "can't get CS GPIO %i\n", i);
Leilk Liue38da372015-11-25 17:50:38 +0800863 goto err_disable_runtime_pm;
Nicolas Boichat98c8dcc2015-11-09 12:14:51 +0800864 }
Leilk Liu37457602015-10-26 16:09:44 +0800865 }
866 }
867 }
868
luhua.xufdeae8f2019-09-11 05:55:31 -0400869 if (mdata->dev_comp->dma_ext)
870 addr_bits = DMA_ADDR_EXT_BITS;
871 else
872 addr_bits = DMA_ADDR_DEF_BITS;
873 ret = dma_set_mask(&pdev->dev, DMA_BIT_MASK(addr_bits));
874 if (ret)
875 dev_notice(&pdev->dev, "SPI dma_set_mask(%d) failed, ret:%d\n",
876 addr_bits, ret);
877
Mason Zhangc934fec2021-07-13 19:42:48 +0800878 ret = devm_spi_register_master(&pdev->dev, master);
879 if (ret) {
880 dev_err(&pdev->dev, "failed to register master (%d)\n", ret);
881 goto err_disable_runtime_pm;
882 }
883
Leilk Liua5682312015-08-07 15:19:50 +0800884 return 0;
885
Leilk Liue38da372015-11-25 17:50:38 +0800886err_disable_runtime_pm:
887 pm_runtime_disable(&pdev->dev);
Leilk Liua5682312015-08-07 15:19:50 +0800888err_put_master:
889 spi_master_put(master);
890
891 return ret;
892}
893
894static int mtk_spi_remove(struct platform_device *pdev)
895{
896 struct spi_master *master = platform_get_drvdata(pdev);
897 struct mtk_spi *mdata = spi_master_get_devdata(master);
898
899 pm_runtime_disable(&pdev->dev);
900
901 mtk_spi_reset(mdata);
Leilk Liua5682312015-08-07 15:19:50 +0800902
Mason Zhang162a31e2021-06-29 18:08:15 +0800903 if (mdata->dev_comp->no_need_unprepare)
904 clk_unprepare(mdata->spi_clk);
905
Leilk Liua5682312015-08-07 15:19:50 +0800906 return 0;
907}
908
909#ifdef CONFIG_PM_SLEEP
910static int mtk_spi_suspend(struct device *dev)
911{
912 int ret;
913 struct spi_master *master = dev_get_drvdata(dev);
914 struct mtk_spi *mdata = spi_master_get_devdata(master);
915
916 ret = spi_master_suspend(master);
917 if (ret)
918 return ret;
919
920 if (!pm_runtime_suspended(dev))
921 clk_disable_unprepare(mdata->spi_clk);
922
923 return ret;
924}
925
926static int mtk_spi_resume(struct device *dev)
927{
928 int ret;
929 struct spi_master *master = dev_get_drvdata(dev);
930 struct mtk_spi *mdata = spi_master_get_devdata(master);
931
932 if (!pm_runtime_suspended(dev)) {
933 ret = clk_prepare_enable(mdata->spi_clk);
Leilk Liu13da5a02015-08-24 11:45:17 +0800934 if (ret < 0) {
935 dev_err(dev, "failed to enable spi_clk (%d)\n", ret);
Leilk Liua5682312015-08-07 15:19:50 +0800936 return ret;
Leilk Liu13da5a02015-08-24 11:45:17 +0800937 }
Leilk Liua5682312015-08-07 15:19:50 +0800938 }
939
940 ret = spi_master_resume(master);
941 if (ret < 0)
942 clk_disable_unprepare(mdata->spi_clk);
943
944 return ret;
945}
946#endif /* CONFIG_PM_SLEEP */
947
948#ifdef CONFIG_PM
949static int mtk_spi_runtime_suspend(struct device *dev)
950{
951 struct spi_master *master = dev_get_drvdata(dev);
952 struct mtk_spi *mdata = spi_master_get_devdata(master);
953
Mason Zhang162a31e2021-06-29 18:08:15 +0800954 if (mdata->dev_comp->no_need_unprepare)
955 clk_disable(mdata->spi_clk);
956 else
957 clk_disable_unprepare(mdata->spi_clk);
Leilk Liua5682312015-08-07 15:19:50 +0800958
959 return 0;
960}
961
962static int mtk_spi_runtime_resume(struct device *dev)
963{
964 struct spi_master *master = dev_get_drvdata(dev);
965 struct mtk_spi *mdata = spi_master_get_devdata(master);
Leilk Liu13da5a02015-08-24 11:45:17 +0800966 int ret;
Leilk Liua5682312015-08-07 15:19:50 +0800967
Mason Zhang162a31e2021-06-29 18:08:15 +0800968 if (mdata->dev_comp->no_need_unprepare)
969 ret = clk_enable(mdata->spi_clk);
970 else
971 ret = clk_prepare_enable(mdata->spi_clk);
Leilk Liu13da5a02015-08-24 11:45:17 +0800972 if (ret < 0) {
973 dev_err(dev, "failed to enable spi_clk (%d)\n", ret);
974 return ret;
975 }
976
977 return 0;
Leilk Liua5682312015-08-07 15:19:50 +0800978}
979#endif /* CONFIG_PM */
980
981static const struct dev_pm_ops mtk_spi_pm = {
982 SET_SYSTEM_SLEEP_PM_OPS(mtk_spi_suspend, mtk_spi_resume)
983 SET_RUNTIME_PM_OPS(mtk_spi_runtime_suspend,
984 mtk_spi_runtime_resume, NULL)
985};
986
kbuild test robot4299aaa2015-08-07 22:33:11 +0800987static struct platform_driver mtk_spi_driver = {
Leilk Liua5682312015-08-07 15:19:50 +0800988 .driver = {
989 .name = "mtk-spi",
990 .pm = &mtk_spi_pm,
991 .of_match_table = mtk_spi_of_match,
992 },
993 .probe = mtk_spi_probe,
994 .remove = mtk_spi_remove,
995};
996
997module_platform_driver(mtk_spi_driver);
998
999MODULE_DESCRIPTION("MTK SPI Controller driver");
1000MODULE_AUTHOR("Leilk Liu <leilk.liu@mediatek.com>");
1001MODULE_LICENSE("GPL v2");
Axel Line4001882015-08-11 09:15:30 +08001002MODULE_ALIAS("platform:mtk-spi");