blob: cfbee2cfba6b0103e1da409e931ea155763fdb59 [file] [log] [blame]
Thomas Gleixnerd2912cb2019-06-04 10:11:33 +02001// SPDX-License-Identifier: GPL-2.0-only
Andy Shevchenko4b45efe2015-07-27 18:04:03 +03002/*
3 * Intel Sunrisepoint LPSS core support.
4 *
5 * Copyright (C) 2015, Intel Corporation
6 *
7 * Authors: Andy Shevchenko <andriy.shevchenko@linux.intel.com>
8 * Mika Westerberg <mika.westerberg@linux.intel.com>
9 * Heikki Krogerus <heikki.krogerus@linux.intel.com>
10 * Jarkko Nikula <jarkko.nikula@linux.intel.com>
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030011 */
12
13#include <linux/clk.h>
14#include <linux/clkdev.h>
15#include <linux/clk-provider.h>
16#include <linux/debugfs.h>
17#include <linux/idr.h>
Stephen Boyd62e59c42019-04-18 15:20:22 -070018#include <linux/io.h>
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030019#include <linux/ioport.h>
20#include <linux/kernel.h>
21#include <linux/module.h>
22#include <linux/mfd/core.h>
23#include <linux/pm_qos.h>
24#include <linux/pm_runtime.h>
Mika Westerberge15ad212015-11-30 17:11:41 +020025#include <linux/property.h>
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030026#include <linux/seq_file.h>
Linus Torvalds9cf5c092015-11-06 14:22:15 -080027#include <linux/io-64-nonatomic-lo-hi.h>
Andy Shevchenko689d4452015-09-14 11:32:48 +030028
Andy Shevchenkoffcfc202019-04-09 18:02:19 +030029#include <linux/dma/idma64.h>
30
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030031#include "intel-lpss.h"
32
33#define LPSS_DEV_OFFSET 0x000
34#define LPSS_DEV_SIZE 0x200
35#define LPSS_PRIV_OFFSET 0x200
36#define LPSS_PRIV_SIZE 0x100
Heikki Krogerus41a3da22016-04-18 15:14:56 +030037#define LPSS_PRIV_REG_COUNT (LPSS_PRIV_SIZE / 4)
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030038#define LPSS_IDMA64_OFFSET 0x800
39#define LPSS_IDMA64_SIZE 0x800
40
41/* Offsets from lpss->priv */
42#define LPSS_PRIV_RESETS 0x04
Andy Shevchenkof3b23e5a2018-04-24 18:05:12 +030043#define LPSS_PRIV_RESETS_IDMA BIT(2)
44#define LPSS_PRIV_RESETS_FUNC 0x3
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030045
46#define LPSS_PRIV_ACTIVELTR 0x10
47#define LPSS_PRIV_IDLELTR 0x14
48
49#define LPSS_PRIV_LTR_REQ BIT(15)
Andy Shevchenkocbd1c5c2019-08-16 20:33:42 +030050#define LPSS_PRIV_LTR_SCALE_MASK GENMASK(11, 10)
51#define LPSS_PRIV_LTR_SCALE_1US (2 << 10)
52#define LPSS_PRIV_LTR_SCALE_32US (3 << 10)
53#define LPSS_PRIV_LTR_VALUE_MASK GENMASK(9, 0)
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030054
55#define LPSS_PRIV_SSP_REG 0x20
56#define LPSS_PRIV_SSP_REG_DIS_DMA_FIN BIT(0)
57
Andy Shevchenko689d4452015-09-14 11:32:48 +030058#define LPSS_PRIV_REMAP_ADDR 0x40
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030059
60#define LPSS_PRIV_CAPS 0xfc
61#define LPSS_PRIV_CAPS_NO_IDMA BIT(8)
Andy Shevchenkocbd1c5c2019-08-16 20:33:42 +030062#define LPSS_PRIV_CAPS_TYPE_MASK GENMASK(7, 4)
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030063#define LPSS_PRIV_CAPS_TYPE_SHIFT 4
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030064
65/* This matches the type field in CAPS register */
66enum intel_lpss_dev_type {
67 LPSS_DEV_I2C = 0,
68 LPSS_DEV_UART,
69 LPSS_DEV_SPI,
70};
71
72struct intel_lpss {
73 const struct intel_lpss_platform_info *info;
74 enum intel_lpss_dev_type type;
75 struct clk *clk;
76 struct clk_lookup *clock;
Mika Westerberge15ad212015-11-30 17:11:41 +020077 struct mfd_cell *cell;
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030078 struct device *dev;
79 void __iomem *priv;
Heikki Krogerus41a3da22016-04-18 15:14:56 +030080 u32 priv_ctx[LPSS_PRIV_REG_COUNT];
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030081 int devid;
82 u32 caps;
83 u32 active_ltr;
84 u32 idle_ltr;
85 struct dentry *debugfs;
86};
87
88static const struct resource intel_lpss_dev_resources[] = {
89 DEFINE_RES_MEM_NAMED(LPSS_DEV_OFFSET, LPSS_DEV_SIZE, "lpss_dev"),
90 DEFINE_RES_MEM_NAMED(LPSS_PRIV_OFFSET, LPSS_PRIV_SIZE, "lpss_priv"),
91 DEFINE_RES_IRQ(0),
92};
93
94static const struct resource intel_lpss_idma64_resources[] = {
95 DEFINE_RES_MEM(LPSS_IDMA64_OFFSET, LPSS_IDMA64_SIZE),
96 DEFINE_RES_IRQ(0),
97};
98
Andy Shevchenko4b45efe2015-07-27 18:04:03 +030099/*
100 * Cells needs to be ordered so that the iDMA is created first. This is
101 * because we need to be sure the DMA is available when the host controller
102 * driver is probed.
103 */
104static const struct mfd_cell intel_lpss_idma64_cell = {
105 .name = LPSS_IDMA64_DRIVER_NAME,
106 .num_resources = ARRAY_SIZE(intel_lpss_idma64_resources),
107 .resources = intel_lpss_idma64_resources,
108};
109
110static const struct mfd_cell intel_lpss_i2c_cell = {
111 .name = "i2c_designware",
112 .num_resources = ARRAY_SIZE(intel_lpss_dev_resources),
113 .resources = intel_lpss_dev_resources,
114};
115
116static const struct mfd_cell intel_lpss_uart_cell = {
117 .name = "dw-apb-uart",
118 .num_resources = ARRAY_SIZE(intel_lpss_dev_resources),
119 .resources = intel_lpss_dev_resources,
120};
121
122static const struct mfd_cell intel_lpss_spi_cell = {
123 .name = "pxa2xx-spi",
124 .num_resources = ARRAY_SIZE(intel_lpss_dev_resources),
125 .resources = intel_lpss_dev_resources,
126};
127
128static DEFINE_IDA(intel_lpss_devid_ida);
129static struct dentry *intel_lpss_debugfs;
130
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300131static void intel_lpss_cache_ltr(struct intel_lpss *lpss)
132{
133 lpss->active_ltr = readl(lpss->priv + LPSS_PRIV_ACTIVELTR);
134 lpss->idle_ltr = readl(lpss->priv + LPSS_PRIV_IDLELTR);
135}
136
137static int intel_lpss_debugfs_add(struct intel_lpss *lpss)
138{
139 struct dentry *dir;
140
141 dir = debugfs_create_dir(dev_name(lpss->dev), intel_lpss_debugfs);
142 if (IS_ERR(dir))
143 return PTR_ERR(dir);
144
145 /* Cache the values into lpss structure */
146 intel_lpss_cache_ltr(lpss);
147
148 debugfs_create_x32("capabilities", S_IRUGO, dir, &lpss->caps);
149 debugfs_create_x32("active_ltr", S_IRUGO, dir, &lpss->active_ltr);
150 debugfs_create_x32("idle_ltr", S_IRUGO, dir, &lpss->idle_ltr);
151
152 lpss->debugfs = dir;
153 return 0;
154}
155
156static void intel_lpss_debugfs_remove(struct intel_lpss *lpss)
157{
158 debugfs_remove_recursive(lpss->debugfs);
159}
160
161static void intel_lpss_ltr_set(struct device *dev, s32 val)
162{
163 struct intel_lpss *lpss = dev_get_drvdata(dev);
164 u32 ltr;
165
166 /*
167 * Program latency tolerance (LTR) accordingly what has been asked
168 * by the PM QoS layer or disable it in case we were passed
169 * negative value or PM_QOS_LATENCY_ANY.
170 */
171 ltr = readl(lpss->priv + LPSS_PRIV_ACTIVELTR);
172
173 if (val == PM_QOS_LATENCY_ANY || val < 0) {
174 ltr &= ~LPSS_PRIV_LTR_REQ;
175 } else {
176 ltr |= LPSS_PRIV_LTR_REQ;
177 ltr &= ~LPSS_PRIV_LTR_SCALE_MASK;
178 ltr &= ~LPSS_PRIV_LTR_VALUE_MASK;
179
180 if (val > LPSS_PRIV_LTR_VALUE_MASK)
181 ltr |= LPSS_PRIV_LTR_SCALE_32US | val >> 5;
182 else
183 ltr |= LPSS_PRIV_LTR_SCALE_1US | val;
184 }
185
186 if (ltr == lpss->active_ltr)
187 return;
188
189 writel(ltr, lpss->priv + LPSS_PRIV_ACTIVELTR);
190 writel(ltr, lpss->priv + LPSS_PRIV_IDLELTR);
191
192 /* Cache the values into lpss structure */
193 intel_lpss_cache_ltr(lpss);
194}
195
196static void intel_lpss_ltr_expose(struct intel_lpss *lpss)
197{
198 lpss->dev->power.set_latency_tolerance = intel_lpss_ltr_set;
199 dev_pm_qos_expose_latency_tolerance(lpss->dev);
200}
201
202static void intel_lpss_ltr_hide(struct intel_lpss *lpss)
203{
204 dev_pm_qos_hide_latency_tolerance(lpss->dev);
205 lpss->dev->power.set_latency_tolerance = NULL;
206}
207
208static int intel_lpss_assign_devs(struct intel_lpss *lpss)
209{
Mika Westerberge15ad212015-11-30 17:11:41 +0200210 const struct mfd_cell *cell;
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300211 unsigned int type;
212
213 type = lpss->caps & LPSS_PRIV_CAPS_TYPE_MASK;
214 type >>= LPSS_PRIV_CAPS_TYPE_SHIFT;
215
216 switch (type) {
217 case LPSS_DEV_I2C:
Mika Westerberge15ad212015-11-30 17:11:41 +0200218 cell = &intel_lpss_i2c_cell;
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300219 break;
220 case LPSS_DEV_UART:
Mika Westerberge15ad212015-11-30 17:11:41 +0200221 cell = &intel_lpss_uart_cell;
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300222 break;
223 case LPSS_DEV_SPI:
Mika Westerberge15ad212015-11-30 17:11:41 +0200224 cell = &intel_lpss_spi_cell;
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300225 break;
226 default:
227 return -ENODEV;
228 }
229
Mika Westerberge15ad212015-11-30 17:11:41 +0200230 lpss->cell = devm_kmemdup(lpss->dev, cell, sizeof(*cell), GFP_KERNEL);
231 if (!lpss->cell)
232 return -ENOMEM;
233
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300234 lpss->type = type;
235
236 return 0;
237}
238
239static bool intel_lpss_has_idma(const struct intel_lpss *lpss)
240{
241 return (lpss->caps & LPSS_PRIV_CAPS_NO_IDMA) == 0;
242}
243
244static void intel_lpss_set_remap_addr(const struct intel_lpss *lpss)
245{
246 resource_size_t addr = lpss->info->mem->start;
247
Andy Shevchenko689d4452015-09-14 11:32:48 +0300248 lo_hi_writeq(addr, lpss->priv + LPSS_PRIV_REMAP_ADDR);
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300249}
250
251static void intel_lpss_deassert_reset(const struct intel_lpss *lpss)
252{
253 u32 value = LPSS_PRIV_RESETS_FUNC | LPSS_PRIV_RESETS_IDMA;
254
255 /* Bring out the device from reset */
256 writel(value, lpss->priv + LPSS_PRIV_RESETS);
257}
258
259static void intel_lpss_init_dev(const struct intel_lpss *lpss)
260{
261 u32 value = LPSS_PRIV_SSP_REG_DIS_DMA_FIN;
262
Binbin Wudad06532019-04-08 16:09:10 +0800263 /* Set the device in reset state */
264 writel(0, lpss->priv + LPSS_PRIV_RESETS);
265
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300266 intel_lpss_deassert_reset(lpss);
267
Andy Shevchenkod28b6252018-04-24 18:00:10 +0300268 intel_lpss_set_remap_addr(lpss);
269
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300270 if (!intel_lpss_has_idma(lpss))
271 return;
272
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300273 /* Make sure that SPI multiblock DMA transfers are re-enabled */
274 if (lpss->type == LPSS_DEV_SPI)
275 writel(value, lpss->priv + LPSS_PRIV_SSP_REG);
276}
277
278static void intel_lpss_unregister_clock_tree(struct clk *clk)
279{
280 struct clk *parent;
281
282 while (clk) {
283 parent = clk_get_parent(clk);
284 clk_unregister(clk);
285 clk = parent;
286 }
287}
288
289static int intel_lpss_register_clock_divider(struct intel_lpss *lpss,
290 const char *devname,
291 struct clk **clk)
292{
293 char name[32];
294 struct clk *tmp = *clk;
295
296 snprintf(name, sizeof(name), "%s-enable", devname);
297 tmp = clk_register_gate(NULL, name, __clk_get_name(tmp), 0,
298 lpss->priv, 0, 0, NULL);
299 if (IS_ERR(tmp))
300 return PTR_ERR(tmp);
301
302 snprintf(name, sizeof(name), "%s-div", devname);
303 tmp = clk_register_fractional_divider(NULL, name, __clk_get_name(tmp),
Andy Shevchenko82f53f92021-08-12 20:00:24 +0300304 CLK_FRAC_DIVIDER_POWER_OF_TWO_PS,
305 lpss->priv, 1, 15, 16, 15, 0,
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300306 NULL);
307 if (IS_ERR(tmp))
308 return PTR_ERR(tmp);
309 *clk = tmp;
310
311 snprintf(name, sizeof(name), "%s-update", devname);
312 tmp = clk_register_gate(NULL, name, __clk_get_name(tmp),
313 CLK_SET_RATE_PARENT, lpss->priv, 31, 0, NULL);
314 if (IS_ERR(tmp))
315 return PTR_ERR(tmp);
316 *clk = tmp;
317
318 return 0;
319}
320
321static int intel_lpss_register_clock(struct intel_lpss *lpss)
322{
323 const struct mfd_cell *cell = lpss->cell;
324 struct clk *clk;
325 char devname[24];
326 int ret;
327
328 if (!lpss->info->clk_rate)
329 return 0;
330
331 /* Root clock */
Stephen Boyd0f7e70e2016-04-19 18:20:49 -0700332 clk = clk_register_fixed_rate(NULL, dev_name(lpss->dev), NULL, 0,
333 lpss->info->clk_rate);
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300334 if (IS_ERR(clk))
335 return PTR_ERR(clk);
336
337 snprintf(devname, sizeof(devname), "%s.%d", cell->name, lpss->devid);
338
339 /*
340 * Support for clock divider only if it has some preset value.
341 * Otherwise we assume that the divider is not used.
342 */
343 if (lpss->type != LPSS_DEV_I2C) {
344 ret = intel_lpss_register_clock_divider(lpss, devname, &clk);
345 if (ret)
346 goto err_clk_register;
347 }
348
349 ret = -ENOMEM;
350
351 /* Clock for the host controller */
352 lpss->clock = clkdev_create(clk, lpss->info->clk_con_id, "%s", devname);
353 if (!lpss->clock)
354 goto err_clk_register;
355
356 lpss->clk = clk;
357
358 return 0;
359
360err_clk_register:
361 intel_lpss_unregister_clock_tree(clk);
362
363 return ret;
364}
365
366static void intel_lpss_unregister_clock(struct intel_lpss *lpss)
367{
368 if (IS_ERR_OR_NULL(lpss->clk))
369 return;
370
371 clkdev_drop(lpss->clock);
372 intel_lpss_unregister_clock_tree(lpss->clk);
373}
374
375int intel_lpss_probe(struct device *dev,
376 const struct intel_lpss_platform_info *info)
377{
378 struct intel_lpss *lpss;
379 int ret;
380
381 if (!info || !info->mem || info->irq <= 0)
382 return -EINVAL;
383
384 lpss = devm_kzalloc(dev, sizeof(*lpss), GFP_KERNEL);
385 if (!lpss)
386 return -ENOMEM;
387
Tuowen Zhaoa8ff78f2019-10-16 15:06:29 -0600388 lpss->priv = devm_ioremap_uc(dev, info->mem->start + LPSS_PRIV_OFFSET,
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300389 LPSS_PRIV_SIZE);
390 if (!lpss->priv)
391 return -ENOMEM;
392
393 lpss->info = info;
394 lpss->dev = dev;
395 lpss->caps = readl(lpss->priv + LPSS_PRIV_CAPS);
396
397 dev_set_drvdata(dev, lpss);
398
399 ret = intel_lpss_assign_devs(lpss);
400 if (ret)
401 return ret;
402
Heikki Krogerus03152e32021-03-01 16:42:21 +0200403 lpss->cell->swnode = info->swnode;
Hans de Goedee6b14202021-12-03 12:51:08 +0100404 lpss->cell->ignore_resource_conflicts = info->ignore_resource_conflicts;
Mika Westerberge15ad212015-11-30 17:11:41 +0200405
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300406 intel_lpss_init_dev(lpss);
407
408 lpss->devid = ida_simple_get(&intel_lpss_devid_ida, 0, 0, GFP_KERNEL);
409 if (lpss->devid < 0)
410 return lpss->devid;
411
412 ret = intel_lpss_register_clock(lpss);
413 if (ret)
414 goto err_clk_register;
415
416 intel_lpss_ltr_expose(lpss);
417
418 ret = intel_lpss_debugfs_add(lpss);
419 if (ret)
420 dev_warn(dev, "Failed to create debugfs entries\n");
421
422 if (intel_lpss_has_idma(lpss)) {
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300423 ret = mfd_add_devices(dev, lpss->devid, &intel_lpss_idma64_cell,
424 1, info->mem, info->irq, NULL);
425 if (ret)
426 dev_warn(dev, "Failed to add %s, fallback to PIO\n",
427 LPSS_IDMA64_DRIVER_NAME);
428 }
429
430 ret = mfd_add_devices(dev, lpss->devid, lpss->cell,
431 1, info->mem, info->irq, NULL);
432 if (ret)
433 goto err_remove_ltr;
434
Rafael J. Wysocki8425ec72018-01-03 01:34:53 +0100435 dev_pm_set_driver_flags(dev, DPM_FLAG_SMART_SUSPEND);
436
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300437 return 0;
438
439err_remove_ltr:
440 intel_lpss_debugfs_remove(lpss);
441 intel_lpss_ltr_hide(lpss);
Andy Shevchenko84cb36c2016-01-22 16:48:46 +0200442 intel_lpss_unregister_clock(lpss);
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300443
444err_clk_register:
445 ida_simple_remove(&intel_lpss_devid_ida, lpss->devid);
446
447 return ret;
448}
449EXPORT_SYMBOL_GPL(intel_lpss_probe);
450
451void intel_lpss_remove(struct device *dev)
452{
453 struct intel_lpss *lpss = dev_get_drvdata(dev);
454
455 mfd_remove_devices(dev);
456 intel_lpss_debugfs_remove(lpss);
457 intel_lpss_ltr_hide(lpss);
458 intel_lpss_unregister_clock(lpss);
459 ida_simple_remove(&intel_lpss_devid_ida, lpss->devid);
460}
461EXPORT_SYMBOL_GPL(intel_lpss_remove);
462
463static int resume_lpss_device(struct device *dev, void *data)
464{
Rafael J. Wysocki8425ec72018-01-03 01:34:53 +0100465 if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND))
466 pm_runtime_resume(dev);
467
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300468 return 0;
469}
470
471int intel_lpss_prepare(struct device *dev)
472{
473 /*
474 * Resume both child devices before entering system sleep. This
475 * ensures that they are in proper state before they get suspended.
476 */
477 device_for_each_child_reverse(dev, NULL, resume_lpss_device);
478 return 0;
479}
480EXPORT_SYMBOL_GPL(intel_lpss_prepare);
481
482int intel_lpss_suspend(struct device *dev)
483{
Heikki Krogerus41a3da22016-04-18 15:14:56 +0300484 struct intel_lpss *lpss = dev_get_drvdata(dev);
485 unsigned int i;
486
487 /* Save device context */
488 for (i = 0; i < LPSS_PRIV_REG_COUNT; i++)
489 lpss->priv_ctx[i] = readl(lpss->priv + i * 4);
490
Furquan Shaikh0b471aa2017-07-23 23:02:19 -0700491 /*
492 * If the device type is not UART, then put the controller into
493 * reset. UART cannot be put into reset since S3/S0ix fail when
494 * no_console_suspend flag is enabled.
495 */
496 if (lpss->type != LPSS_DEV_UART)
497 writel(0, lpss->priv + LPSS_PRIV_RESETS);
498
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300499 return 0;
500}
501EXPORT_SYMBOL_GPL(intel_lpss_suspend);
502
503int intel_lpss_resume(struct device *dev)
504{
505 struct intel_lpss *lpss = dev_get_drvdata(dev);
Heikki Krogerus41a3da22016-04-18 15:14:56 +0300506 unsigned int i;
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300507
Heikki Krogerus41a3da22016-04-18 15:14:56 +0300508 intel_lpss_deassert_reset(lpss);
509
510 /* Restore device context */
511 for (i = 0; i < LPSS_PRIV_REG_COUNT; i++)
512 writel(lpss->priv_ctx[i], lpss->priv + i * 4);
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300513
514 return 0;
515}
516EXPORT_SYMBOL_GPL(intel_lpss_resume);
517
518static int __init intel_lpss_init(void)
519{
520 intel_lpss_debugfs = debugfs_create_dir("intel_lpss", NULL);
521 return 0;
522}
523module_init(intel_lpss_init);
524
525static void __exit intel_lpss_exit(void)
526{
Andy Shevchenko02f36912019-06-21 15:56:31 +0300527 ida_destroy(&intel_lpss_devid_ida);
Andy Shevchenko4b45efe2015-07-27 18:04:03 +0300528 debugfs_remove(intel_lpss_debugfs);
529}
530module_exit(intel_lpss_exit);
531
532MODULE_AUTHOR("Andy Shevchenko <andriy.shevchenko@linux.intel.com>");
533MODULE_AUTHOR("Mika Westerberg <mika.westerberg@linux.intel.com>");
534MODULE_AUTHOR("Heikki Krogerus <heikki.krogerus@linux.intel.com>");
535MODULE_AUTHOR("Jarkko Nikula <jarkko.nikula@linux.intel.com>");
536MODULE_DESCRIPTION("Intel LPSS core driver");
537MODULE_LICENSE("GPL v2");
Andy Shevchenko569fac72019-08-21 11:37:12 +0300538/*
539 * Ensure the DMA driver is loaded before the host controller device appears,
540 * so that the host controller driver can request its DMA channels as early
541 * as possible.
542 *
543 * If the DMA module is not there that's OK as well.
544 */
545MODULE_SOFTDEP("pre: platform:" LPSS_IDMA64_DRIVER_NAME);