blob: 9cd738682aab14e5d1d0e858e186640717647a0c [file] [log] [blame]
Radu Pireae1892542018-07-13 19:47:35 +03001// SPDX-License-Identifier: GPL-2.0
2//
3// Driver for AT91 USART Controllers as SPI
4//
5// Copyright (C) 2018 Microchip Technology Inc.
6//
7// Author: Radu Pirea <radu.pirea@microchip.com>
8
9#include <linux/clk.h>
10#include <linux/delay.h>
Radu Pireac3fdefc2019-05-05 21:06:46 +030011#include <linux/dmaengine.h>
12#include <linux/dma-direction.h>
Radu Pireae1892542018-07-13 19:47:35 +030013#include <linux/interrupt.h>
14#include <linux/kernel.h>
15#include <linux/module.h>
Radu Pireac3fdefc2019-05-05 21:06:46 +030016#include <linux/of_platform.h>
MaĆ­ra Canal3340ec42021-10-19 15:05:24 -030017#include <linux/gpio/consumer.h>
Randy Dunlapc2b142cc2018-11-29 14:34:15 -080018#include <linux/pinctrl/consumer.h>
Radu Pireae1892542018-07-13 19:47:35 +030019#include <linux/platform_device.h>
Radu Pirea96ed3ec2018-11-21 13:27:30 +020020#include <linux/pm_runtime.h>
Radu Pireae1892542018-07-13 19:47:35 +030021
22#include <linux/spi/spi.h>
23
24#define US_CR 0x00
25#define US_MR 0x04
26#define US_IER 0x08
27#define US_IDR 0x0C
28#define US_CSR 0x14
29#define US_RHR 0x18
30#define US_THR 0x1C
31#define US_BRGR 0x20
32#define US_VERSION 0xFC
33
34#define US_CR_RSTRX BIT(2)
35#define US_CR_RSTTX BIT(3)
36#define US_CR_RXEN BIT(4)
37#define US_CR_RXDIS BIT(5)
38#define US_CR_TXEN BIT(6)
39#define US_CR_TXDIS BIT(7)
40
41#define US_MR_SPI_MASTER 0x0E
42#define US_MR_CHRL GENMASK(7, 6)
43#define US_MR_CPHA BIT(8)
44#define US_MR_CPOL BIT(16)
45#define US_MR_CLKO BIT(18)
46#define US_MR_WRDBT BIT(20)
47#define US_MR_LOOP BIT(15)
48
49#define US_IR_RXRDY BIT(0)
50#define US_IR_TXRDY BIT(1)
51#define US_IR_OVRE BIT(5)
52
53#define US_BRGR_SIZE BIT(16)
54
55#define US_MIN_CLK_DIV 0x06
56#define US_MAX_CLK_DIV BIT(16)
57
58#define US_RESET (US_CR_RSTRX | US_CR_RSTTX)
59#define US_DISABLE (US_CR_RXDIS | US_CR_TXDIS)
60#define US_ENABLE (US_CR_RXEN | US_CR_TXEN)
61#define US_OVRE_RXRDY_IRQS (US_IR_OVRE | US_IR_RXRDY)
62
63#define US_INIT \
64 (US_MR_SPI_MASTER | US_MR_CHRL | US_MR_CLKO | US_MR_WRDBT)
Radu Pireac3fdefc2019-05-05 21:06:46 +030065#define US_DMA_MIN_BYTES 16
66#define US_DMA_TIMEOUT (msecs_to_jiffies(1000))
Radu Pireae1892542018-07-13 19:47:35 +030067
68/* Register access macros */
69#define at91_usart_spi_readl(port, reg) \
70 readl_relaxed((port)->regs + US_##reg)
71#define at91_usart_spi_writel(port, reg, value) \
72 writel_relaxed((value), (port)->regs + US_##reg)
73
74#define at91_usart_spi_readb(port, reg) \
75 readb_relaxed((port)->regs + US_##reg)
76#define at91_usart_spi_writeb(port, reg, value) \
77 writeb_relaxed((value), (port)->regs + US_##reg)
78
79struct at91_usart_spi {
Radu Pireac3fdefc2019-05-05 21:06:46 +030080 struct platform_device *mpdev;
Radu Pireae1892542018-07-13 19:47:35 +030081 struct spi_transfer *current_transfer;
82 void __iomem *regs;
83 struct device *dev;
84 struct clk *clk;
85
Radu Pireac3fdefc2019-05-05 21:06:46 +030086 struct completion xfer_completion;
87
Radu Pireae1892542018-07-13 19:47:35 +030088 /*used in interrupt to protect data reading*/
89 spinlock_t lock;
90
Radu Pireac3fdefc2019-05-05 21:06:46 +030091 phys_addr_t phybase;
92
Radu Pireae1892542018-07-13 19:47:35 +030093 int irq;
94 unsigned int current_tx_remaining_bytes;
95 unsigned int current_rx_remaining_bytes;
96
97 u32 spi_clk;
98 u32 status;
99
100 bool xfer_failed;
Radu Pireac3fdefc2019-05-05 21:06:46 +0300101 bool use_dma;
Radu Pireae1892542018-07-13 19:47:35 +0300102};
103
Radu Pireac3fdefc2019-05-05 21:06:46 +0300104static void dma_callback(void *data)
105{
106 struct spi_controller *ctlr = data;
107 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr);
108
109 at91_usart_spi_writel(aus, IER, US_IR_RXRDY);
110 aus->current_rx_remaining_bytes = 0;
111 complete(&aus->xfer_completion);
112}
113
114static bool at91_usart_spi_can_dma(struct spi_controller *ctrl,
115 struct spi_device *spi,
116 struct spi_transfer *xfer)
117{
118 struct at91_usart_spi *aus = spi_master_get_devdata(ctrl);
119
120 return aus->use_dma && xfer->len >= US_DMA_MIN_BYTES;
121}
122
123static int at91_usart_spi_configure_dma(struct spi_controller *ctlr,
124 struct at91_usart_spi *aus)
125{
126 struct dma_slave_config slave_config;
127 struct device *dev = &aus->mpdev->dev;
128 phys_addr_t phybase = aus->phybase;
129 dma_cap_mask_t mask;
130 int err = 0;
131
132 dma_cap_zero(mask);
133 dma_cap_set(DMA_SLAVE, mask);
134
Peter Ujfalusicef76e52019-11-13 11:42:48 +0200135 ctlr->dma_tx = dma_request_chan(dev, "tx");
Radu Pireac3fdefc2019-05-05 21:06:46 +0300136 if (IS_ERR_OR_NULL(ctlr->dma_tx)) {
137 if (IS_ERR(ctlr->dma_tx)) {
138 err = PTR_ERR(ctlr->dma_tx);
139 goto at91_usart_spi_error_clear;
140 }
141
142 dev_dbg(dev,
143 "DMA TX channel not available, SPI unable to use DMA\n");
144 err = -EBUSY;
145 goto at91_usart_spi_error_clear;
146 }
147
Peter Ujfalusicef76e52019-11-13 11:42:48 +0200148 ctlr->dma_rx = dma_request_chan(dev, "rx");
Radu Pireac3fdefc2019-05-05 21:06:46 +0300149 if (IS_ERR_OR_NULL(ctlr->dma_rx)) {
150 if (IS_ERR(ctlr->dma_rx)) {
151 err = PTR_ERR(ctlr->dma_rx);
152 goto at91_usart_spi_error;
153 }
154
155 dev_dbg(dev,
156 "DMA RX channel not available, SPI unable to use DMA\n");
157 err = -EBUSY;
158 goto at91_usart_spi_error;
159 }
160
161 slave_config.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
162 slave_config.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
163 slave_config.dst_addr = (dma_addr_t)phybase + US_THR;
164 slave_config.src_addr = (dma_addr_t)phybase + US_RHR;
165 slave_config.src_maxburst = 1;
166 slave_config.dst_maxburst = 1;
167 slave_config.device_fc = false;
168
169 slave_config.direction = DMA_DEV_TO_MEM;
170 if (dmaengine_slave_config(ctlr->dma_rx, &slave_config)) {
171 dev_err(&ctlr->dev,
172 "failed to configure rx dma channel\n");
173 err = -EINVAL;
174 goto at91_usart_spi_error;
175 }
176
177 slave_config.direction = DMA_MEM_TO_DEV;
178 if (dmaengine_slave_config(ctlr->dma_tx, &slave_config)) {
179 dev_err(&ctlr->dev,
180 "failed to configure tx dma channel\n");
181 err = -EINVAL;
182 goto at91_usart_spi_error;
183 }
184
185 aus->use_dma = true;
186 return 0;
187
188at91_usart_spi_error:
189 if (!IS_ERR_OR_NULL(ctlr->dma_tx))
190 dma_release_channel(ctlr->dma_tx);
191 if (!IS_ERR_OR_NULL(ctlr->dma_rx))
192 dma_release_channel(ctlr->dma_rx);
193 ctlr->dma_tx = NULL;
194 ctlr->dma_rx = NULL;
195
196at91_usart_spi_error_clear:
197 return err;
198}
199
200static void at91_usart_spi_release_dma(struct spi_controller *ctlr)
201{
202 if (ctlr->dma_rx)
203 dma_release_channel(ctlr->dma_rx);
204 if (ctlr->dma_tx)
205 dma_release_channel(ctlr->dma_tx);
206}
207
208static void at91_usart_spi_stop_dma(struct spi_controller *ctlr)
209{
210 if (ctlr->dma_rx)
211 dmaengine_terminate_all(ctlr->dma_rx);
212 if (ctlr->dma_tx)
213 dmaengine_terminate_all(ctlr->dma_tx);
214}
215
216static int at91_usart_spi_dma_transfer(struct spi_controller *ctlr,
217 struct spi_transfer *xfer)
218{
219 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr);
220 struct dma_chan *rxchan = ctlr->dma_rx;
221 struct dma_chan *txchan = ctlr->dma_tx;
222 struct dma_async_tx_descriptor *rxdesc;
223 struct dma_async_tx_descriptor *txdesc;
224 dma_cookie_t cookie;
225
226 /* Disable RX interrupt */
227 at91_usart_spi_writel(aus, IDR, US_IR_RXRDY);
228
229 rxdesc = dmaengine_prep_slave_sg(rxchan,
230 xfer->rx_sg.sgl,
231 xfer->rx_sg.nents,
232 DMA_DEV_TO_MEM,
233 DMA_PREP_INTERRUPT |
234 DMA_CTRL_ACK);
235 if (!rxdesc)
236 goto at91_usart_spi_err_dma;
237
238 txdesc = dmaengine_prep_slave_sg(txchan,
239 xfer->tx_sg.sgl,
240 xfer->tx_sg.nents,
241 DMA_MEM_TO_DEV,
242 DMA_PREP_INTERRUPT |
243 DMA_CTRL_ACK);
244 if (!txdesc)
245 goto at91_usart_spi_err_dma;
246
247 rxdesc->callback = dma_callback;
248 rxdesc->callback_param = ctlr;
249
250 cookie = rxdesc->tx_submit(rxdesc);
251 if (dma_submit_error(cookie))
252 goto at91_usart_spi_err_dma;
253
254 cookie = txdesc->tx_submit(txdesc);
255 if (dma_submit_error(cookie))
256 goto at91_usart_spi_err_dma;
257
258 rxchan->device->device_issue_pending(rxchan);
259 txchan->device->device_issue_pending(txchan);
260
261 return 0;
262
263at91_usart_spi_err_dma:
264 /* Enable RX interrupt if something fails and fallback to PIO */
265 at91_usart_spi_writel(aus, IER, US_IR_RXRDY);
266 at91_usart_spi_stop_dma(ctlr);
267
268 return -ENOMEM;
269}
270
271static unsigned long at91_usart_spi_dma_timeout(struct at91_usart_spi *aus)
272{
273 return wait_for_completion_timeout(&aus->xfer_completion,
274 US_DMA_TIMEOUT);
275}
276
Radu Pireae1892542018-07-13 19:47:35 +0300277static inline u32 at91_usart_spi_tx_ready(struct at91_usart_spi *aus)
278{
279 return aus->status & US_IR_TXRDY;
280}
281
282static inline u32 at91_usart_spi_rx_ready(struct at91_usart_spi *aus)
283{
284 return aus->status & US_IR_RXRDY;
285}
286
287static inline u32 at91_usart_spi_check_overrun(struct at91_usart_spi *aus)
288{
289 return aus->status & US_IR_OVRE;
290}
291
292static inline u32 at91_usart_spi_read_status(struct at91_usart_spi *aus)
293{
294 aus->status = at91_usart_spi_readl(aus, CSR);
295 return aus->status;
296}
297
298static inline void at91_usart_spi_tx(struct at91_usart_spi *aus)
299{
300 unsigned int len = aus->current_transfer->len;
301 unsigned int remaining = aus->current_tx_remaining_bytes;
302 const u8 *tx_buf = aus->current_transfer->tx_buf;
303
304 if (!remaining)
305 return;
306
307 if (at91_usart_spi_tx_ready(aus)) {
308 at91_usart_spi_writeb(aus, THR, tx_buf[len - remaining]);
309 aus->current_tx_remaining_bytes--;
310 }
311}
312
313static inline void at91_usart_spi_rx(struct at91_usart_spi *aus)
314{
315 int len = aus->current_transfer->len;
316 int remaining = aus->current_rx_remaining_bytes;
317 u8 *rx_buf = aus->current_transfer->rx_buf;
318
319 if (!remaining)
320 return;
321
322 rx_buf[len - remaining] = at91_usart_spi_readb(aus, RHR);
323 aus->current_rx_remaining_bytes--;
324}
325
326static inline void
327at91_usart_spi_set_xfer_speed(struct at91_usart_spi *aus,
328 struct spi_transfer *xfer)
329{
330 at91_usart_spi_writel(aus, BRGR,
331 DIV_ROUND_UP(aus->spi_clk, xfer->speed_hz));
332}
333
334static irqreturn_t at91_usart_spi_interrupt(int irq, void *dev_id)
335{
336 struct spi_controller *controller = dev_id;
337 struct at91_usart_spi *aus = spi_master_get_devdata(controller);
338
339 spin_lock(&aus->lock);
340 at91_usart_spi_read_status(aus);
341
342 if (at91_usart_spi_check_overrun(aus)) {
343 aus->xfer_failed = true;
344 at91_usart_spi_writel(aus, IDR, US_IR_OVRE | US_IR_RXRDY);
345 spin_unlock(&aus->lock);
346 return IRQ_HANDLED;
347 }
348
349 if (at91_usart_spi_rx_ready(aus)) {
350 at91_usart_spi_rx(aus);
351 spin_unlock(&aus->lock);
352 return IRQ_HANDLED;
353 }
354
355 spin_unlock(&aus->lock);
356
357 return IRQ_NONE;
358}
359
360static int at91_usart_spi_setup(struct spi_device *spi)
361{
362 struct at91_usart_spi *aus = spi_master_get_devdata(spi->controller);
363 u32 *ausd = spi->controller_state;
364 unsigned int mr = at91_usart_spi_readl(aus, MR);
Radu Pireae1892542018-07-13 19:47:35 +0300365
366 if (spi->mode & SPI_CPOL)
367 mr |= US_MR_CPOL;
368 else
369 mr &= ~US_MR_CPOL;
370
371 if (spi->mode & SPI_CPHA)
372 mr |= US_MR_CPHA;
373 else
374 mr &= ~US_MR_CPHA;
375
376 if (spi->mode & SPI_LOOP)
377 mr |= US_MR_LOOP;
378 else
379 mr &= ~US_MR_LOOP;
380
381 if (!ausd) {
382 ausd = kzalloc(sizeof(*ausd), GFP_KERNEL);
383 if (!ausd)
384 return -ENOMEM;
385
386 spi->controller_state = ausd;
387 }
388
389 *ausd = mr;
390
391 dev_dbg(&spi->dev,
392 "setup: bpw %u mode 0x%x -> mr %d %08x\n",
Axel Lin72dca1f2019-03-21 23:53:48 +0800393 spi->bits_per_word, spi->mode, spi->chip_select, mr);
Radu Pireae1892542018-07-13 19:47:35 +0300394
395 return 0;
396}
397
Lee Jones259e0a02018-09-11 11:38:50 +0100398static int at91_usart_spi_transfer_one(struct spi_controller *ctlr,
399 struct spi_device *spi,
400 struct spi_transfer *xfer)
Radu Pireae1892542018-07-13 19:47:35 +0300401{
402 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr);
Radu Pireac3fdefc2019-05-05 21:06:46 +0300403 unsigned long dma_timeout = 0;
404 int ret = 0;
Radu Pireae1892542018-07-13 19:47:35 +0300405
406 at91_usart_spi_set_xfer_speed(aus, xfer);
407 aus->xfer_failed = false;
408 aus->current_transfer = xfer;
409 aus->current_tx_remaining_bytes = xfer->len;
410 aus->current_rx_remaining_bytes = xfer->len;
411
412 while ((aus->current_tx_remaining_bytes ||
413 aus->current_rx_remaining_bytes) && !aus->xfer_failed) {
Radu Pireac3fdefc2019-05-05 21:06:46 +0300414 reinit_completion(&aus->xfer_completion);
415 if (at91_usart_spi_can_dma(ctlr, spi, xfer) &&
416 !ret) {
417 ret = at91_usart_spi_dma_transfer(ctlr, xfer);
418 if (ret)
419 continue;
420
421 dma_timeout = at91_usart_spi_dma_timeout(aus);
422
423 if (WARN_ON(dma_timeout == 0)) {
424 dev_err(&spi->dev, "DMA transfer timeout\n");
425 return -EIO;
426 }
427 aus->current_tx_remaining_bytes = 0;
428 } else {
429 at91_usart_spi_read_status(aus);
430 at91_usart_spi_tx(aus);
431 }
432
Radu Pireae1892542018-07-13 19:47:35 +0300433 cpu_relax();
434 }
435
436 if (aus->xfer_failed) {
437 dev_err(aus->dev, "Overrun!\n");
438 return -EIO;
439 }
440
441 return 0;
442}
443
Lee Jones259e0a02018-09-11 11:38:50 +0100444static int at91_usart_spi_prepare_message(struct spi_controller *ctlr,
445 struct spi_message *message)
Radu Pireae1892542018-07-13 19:47:35 +0300446{
447 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr);
448 struct spi_device *spi = message->spi;
449 u32 *ausd = spi->controller_state;
450
451 at91_usart_spi_writel(aus, CR, US_ENABLE);
452 at91_usart_spi_writel(aus, IER, US_OVRE_RXRDY_IRQS);
453 at91_usart_spi_writel(aus, MR, *ausd);
454
455 return 0;
456}
457
Lee Jones259e0a02018-09-11 11:38:50 +0100458static int at91_usart_spi_unprepare_message(struct spi_controller *ctlr,
459 struct spi_message *message)
Radu Pireae1892542018-07-13 19:47:35 +0300460{
461 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr);
462
463 at91_usart_spi_writel(aus, CR, US_RESET | US_DISABLE);
464 at91_usart_spi_writel(aus, IDR, US_OVRE_RXRDY_IRQS);
465
466 return 0;
467}
468
469static void at91_usart_spi_cleanup(struct spi_device *spi)
470{
471 struct at91_usart_spi_device *ausd = spi->controller_state;
472
473 spi->controller_state = NULL;
474 kfree(ausd);
475}
476
477static void at91_usart_spi_init(struct at91_usart_spi *aus)
478{
479 at91_usart_spi_writel(aus, MR, US_INIT);
480 at91_usart_spi_writel(aus, CR, US_RESET | US_DISABLE);
481}
482
483static int at91_usart_gpio_setup(struct platform_device *pdev)
484{
MaĆ­ra Canal3340ec42021-10-19 15:05:24 -0300485 struct gpio_descs *cs_gpios;
Radu Pireae1892542018-07-13 19:47:35 +0300486
MaĆ­ra Canal3340ec42021-10-19 15:05:24 -0300487 cs_gpios = devm_gpiod_get_array_optional(&pdev->dev, "cs", GPIOD_OUT_LOW);
Radu Pireae1892542018-07-13 19:47:35 +0300488
MaĆ­ra Canal3340ec42021-10-19 15:05:24 -0300489 if (IS_ERR(cs_gpios))
490 return PTR_ERR(cs_gpios);
Radu Pireae1892542018-07-13 19:47:35 +0300491
492 return 0;
493}
494
495static int at91_usart_spi_probe(struct platform_device *pdev)
496{
497 struct resource *regs;
498 struct spi_controller *controller;
499 struct at91_usart_spi *aus;
500 struct clk *clk;
501 int irq;
502 int ret;
503
504 regs = platform_get_resource(to_platform_device(pdev->dev.parent),
505 IORESOURCE_MEM, 0);
506 if (!regs)
507 return -EINVAL;
508
509 irq = platform_get_irq(to_platform_device(pdev->dev.parent), 0);
510 if (irq < 0)
511 return irq;
512
513 clk = devm_clk_get(pdev->dev.parent, "usart");
514 if (IS_ERR(clk))
515 return PTR_ERR(clk);
516
517 ret = -ENOMEM;
518 controller = spi_alloc_master(&pdev->dev, sizeof(*aus));
519 if (!controller)
520 goto at91_usart_spi_probe_fail;
521
522 ret = at91_usart_gpio_setup(pdev);
523 if (ret)
524 goto at91_usart_spi_probe_fail;
525
526 controller->mode_bits = SPI_CPOL | SPI_CPHA | SPI_LOOP | SPI_CS_HIGH;
527 controller->dev.of_node = pdev->dev.parent->of_node;
528 controller->bits_per_word_mask = SPI_BPW_MASK(8);
529 controller->setup = at91_usart_spi_setup;
530 controller->flags = SPI_MASTER_MUST_RX | SPI_MASTER_MUST_TX;
531 controller->transfer_one = at91_usart_spi_transfer_one;
532 controller->prepare_message = at91_usart_spi_prepare_message;
533 controller->unprepare_message = at91_usart_spi_unprepare_message;
Radu Pireac3fdefc2019-05-05 21:06:46 +0300534 controller->can_dma = at91_usart_spi_can_dma;
Radu Pireae1892542018-07-13 19:47:35 +0300535 controller->cleanup = at91_usart_spi_cleanup;
536 controller->max_speed_hz = DIV_ROUND_UP(clk_get_rate(clk),
537 US_MIN_CLK_DIV);
538 controller->min_speed_hz = DIV_ROUND_UP(clk_get_rate(clk),
539 US_MAX_CLK_DIV);
540 platform_set_drvdata(pdev, controller);
541
542 aus = spi_master_get_devdata(controller);
543
544 aus->dev = &pdev->dev;
545 aus->regs = devm_ioremap_resource(&pdev->dev, regs);
546 if (IS_ERR(aus->regs)) {
547 ret = PTR_ERR(aus->regs);
548 goto at91_usart_spi_probe_fail;
549 }
550
551 aus->irq = irq;
552 aus->clk = clk;
553
554 ret = devm_request_irq(&pdev->dev, irq, at91_usart_spi_interrupt, 0,
555 dev_name(&pdev->dev), controller);
556 if (ret)
557 goto at91_usart_spi_probe_fail;
558
559 ret = clk_prepare_enable(clk);
560 if (ret)
561 goto at91_usart_spi_probe_fail;
562
563 aus->spi_clk = clk_get_rate(clk);
564 at91_usart_spi_init(aus);
565
Radu Pireac3fdefc2019-05-05 21:06:46 +0300566 aus->phybase = regs->start;
567
568 aus->mpdev = to_platform_device(pdev->dev.parent);
569
570 ret = at91_usart_spi_configure_dma(controller, aus);
571 if (ret)
572 goto at91_usart_fail_dma;
573
Radu Pireae1892542018-07-13 19:47:35 +0300574 spin_lock_init(&aus->lock);
Radu Pireac3fdefc2019-05-05 21:06:46 +0300575 init_completion(&aus->xfer_completion);
576
Radu Pireae1892542018-07-13 19:47:35 +0300577 ret = devm_spi_register_master(&pdev->dev, controller);
578 if (ret)
579 goto at91_usart_fail_register_master;
580
581 dev_info(&pdev->dev,
582 "AT91 USART SPI Controller version 0x%x at %pa (irq %d)\n",
583 at91_usart_spi_readl(aus, VERSION),
584 &regs->start, irq);
585
586 return 0;
587
588at91_usart_fail_register_master:
Radu Pireac3fdefc2019-05-05 21:06:46 +0300589 at91_usart_spi_release_dma(controller);
590at91_usart_fail_dma:
Radu Pireae1892542018-07-13 19:47:35 +0300591 clk_disable_unprepare(clk);
592at91_usart_spi_probe_fail:
593 spi_master_put(controller);
594 return ret;
595}
596
Radu Pirea96ed3ec2018-11-21 13:27:30 +0200597__maybe_unused static int at91_usart_spi_runtime_suspend(struct device *dev)
598{
599 struct spi_controller *ctlr = dev_get_drvdata(dev);
600 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr);
601
602 clk_disable_unprepare(aus->clk);
603 pinctrl_pm_select_sleep_state(dev);
604
605 return 0;
606}
607
608__maybe_unused static int at91_usart_spi_runtime_resume(struct device *dev)
609{
610 struct spi_controller *ctrl = dev_get_drvdata(dev);
611 struct at91_usart_spi *aus = spi_master_get_devdata(ctrl);
612
613 pinctrl_pm_select_default_state(dev);
614
615 return clk_prepare_enable(aus->clk);
616}
617
618__maybe_unused static int at91_usart_spi_suspend(struct device *dev)
619{
620 struct spi_controller *ctrl = dev_get_drvdata(dev);
621 int ret;
622
623 ret = spi_controller_suspend(ctrl);
624 if (ret)
625 return ret;
626
627 if (!pm_runtime_suspended(dev))
628 at91_usart_spi_runtime_suspend(dev);
629
630 return 0;
631}
632
633__maybe_unused static int at91_usart_spi_resume(struct device *dev)
634{
635 struct spi_controller *ctrl = dev_get_drvdata(dev);
636 struct at91_usart_spi *aus = spi_master_get_devdata(ctrl);
637 int ret;
638
639 if (!pm_runtime_suspended(dev)) {
640 ret = at91_usart_spi_runtime_resume(dev);
641 if (ret)
642 return ret;
643 }
644
645 at91_usart_spi_init(aus);
646
647 return spi_controller_resume(ctrl);
648}
649
Radu Pireae1892542018-07-13 19:47:35 +0300650static int at91_usart_spi_remove(struct platform_device *pdev)
651{
652 struct spi_controller *ctlr = platform_get_drvdata(pdev);
653 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr);
654
Radu Pireac3fdefc2019-05-05 21:06:46 +0300655 at91_usart_spi_release_dma(ctlr);
Radu Pireae1892542018-07-13 19:47:35 +0300656 clk_disable_unprepare(aus->clk);
657
658 return 0;
659}
660
Radu Pirea96ed3ec2018-11-21 13:27:30 +0200661static const struct dev_pm_ops at91_usart_spi_pm_ops = {
662 SET_SYSTEM_SLEEP_PM_OPS(at91_usart_spi_suspend, at91_usart_spi_resume)
663 SET_RUNTIME_PM_OPS(at91_usart_spi_runtime_suspend,
664 at91_usart_spi_runtime_resume, NULL)
665};
666
Radu Pireae1892542018-07-13 19:47:35 +0300667static struct platform_driver at91_usart_spi_driver = {
668 .driver = {
669 .name = "at91_usart_spi",
Radu Pirea96ed3ec2018-11-21 13:27:30 +0200670 .pm = &at91_usart_spi_pm_ops,
Radu Pireae1892542018-07-13 19:47:35 +0300671 },
672 .probe = at91_usart_spi_probe,
673 .remove = at91_usart_spi_remove,
674};
675
676module_platform_driver(at91_usart_spi_driver);
677
678MODULE_DESCRIPTION("Microchip AT91 USART SPI Controller driver");
679MODULE_AUTHOR("Radu Pirea <radu.pirea@microchip.com>");
680MODULE_LICENSE("GPL v2");
681MODULE_ALIAS("platform:at91_usart_spi");