Thomas Gleixner | d2912cb | 2019-06-04 10:11:33 +0200 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0-only |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 2 | /* |
| 3 | * Intel Sunrisepoint LPSS core support. |
| 4 | * |
| 5 | * Copyright (C) 2015, Intel Corporation |
| 6 | * |
| 7 | * Authors: Andy Shevchenko <andriy.shevchenko@linux.intel.com> |
| 8 | * Mika Westerberg <mika.westerberg@linux.intel.com> |
| 9 | * Heikki Krogerus <heikki.krogerus@linux.intel.com> |
| 10 | * Jarkko Nikula <jarkko.nikula@linux.intel.com> |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 11 | */ |
| 12 | |
| 13 | #include <linux/clk.h> |
| 14 | #include <linux/clkdev.h> |
| 15 | #include <linux/clk-provider.h> |
| 16 | #include <linux/debugfs.h> |
| 17 | #include <linux/idr.h> |
Stephen Boyd | 62e59c4 | 2019-04-18 15:20:22 -0700 | [diff] [blame] | 18 | #include <linux/io.h> |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 19 | #include <linux/ioport.h> |
| 20 | #include <linux/kernel.h> |
| 21 | #include <linux/module.h> |
| 22 | #include <linux/mfd/core.h> |
| 23 | #include <linux/pm_qos.h> |
| 24 | #include <linux/pm_runtime.h> |
Mika Westerberg | e15ad21 | 2015-11-30 17:11:41 +0200 | [diff] [blame] | 25 | #include <linux/property.h> |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 26 | #include <linux/seq_file.h> |
Linus Torvalds | 9cf5c09 | 2015-11-06 14:22:15 -0800 | [diff] [blame] | 27 | #include <linux/io-64-nonatomic-lo-hi.h> |
Andy Shevchenko | 689d445 | 2015-09-14 11:32:48 +0300 | [diff] [blame] | 28 | |
Andy Shevchenko | ffcfc20 | 2019-04-09 18:02:19 +0300 | [diff] [blame] | 29 | #include <linux/dma/idma64.h> |
| 30 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 31 | #include "intel-lpss.h" |
| 32 | |
| 33 | #define LPSS_DEV_OFFSET 0x000 |
| 34 | #define LPSS_DEV_SIZE 0x200 |
| 35 | #define LPSS_PRIV_OFFSET 0x200 |
| 36 | #define LPSS_PRIV_SIZE 0x100 |
Heikki Krogerus | 41a3da2 | 2016-04-18 15:14:56 +0300 | [diff] [blame] | 37 | #define LPSS_PRIV_REG_COUNT (LPSS_PRIV_SIZE / 4) |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 38 | #define LPSS_IDMA64_OFFSET 0x800 |
| 39 | #define LPSS_IDMA64_SIZE 0x800 |
| 40 | |
| 41 | /* Offsets from lpss->priv */ |
| 42 | #define LPSS_PRIV_RESETS 0x04 |
Andy Shevchenko | f3b23e5a | 2018-04-24 18:05:12 +0300 | [diff] [blame] | 43 | #define LPSS_PRIV_RESETS_IDMA BIT(2) |
| 44 | #define LPSS_PRIV_RESETS_FUNC 0x3 |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 45 | |
| 46 | #define LPSS_PRIV_ACTIVELTR 0x10 |
| 47 | #define LPSS_PRIV_IDLELTR 0x14 |
| 48 | |
| 49 | #define LPSS_PRIV_LTR_REQ BIT(15) |
Andy Shevchenko | cbd1c5c | 2019-08-16 20:33:42 +0300 | [diff] [blame] | 50 | #define LPSS_PRIV_LTR_SCALE_MASK GENMASK(11, 10) |
| 51 | #define LPSS_PRIV_LTR_SCALE_1US (2 << 10) |
| 52 | #define LPSS_PRIV_LTR_SCALE_32US (3 << 10) |
| 53 | #define LPSS_PRIV_LTR_VALUE_MASK GENMASK(9, 0) |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 54 | |
| 55 | #define LPSS_PRIV_SSP_REG 0x20 |
| 56 | #define LPSS_PRIV_SSP_REG_DIS_DMA_FIN BIT(0) |
| 57 | |
Andy Shevchenko | 689d445 | 2015-09-14 11:32:48 +0300 | [diff] [blame] | 58 | #define LPSS_PRIV_REMAP_ADDR 0x40 |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 59 | |
| 60 | #define LPSS_PRIV_CAPS 0xfc |
| 61 | #define LPSS_PRIV_CAPS_NO_IDMA BIT(8) |
Andy Shevchenko | cbd1c5c | 2019-08-16 20:33:42 +0300 | [diff] [blame] | 62 | #define LPSS_PRIV_CAPS_TYPE_MASK GENMASK(7, 4) |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 63 | #define LPSS_PRIV_CAPS_TYPE_SHIFT 4 |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 64 | |
| 65 | /* This matches the type field in CAPS register */ |
| 66 | enum intel_lpss_dev_type { |
| 67 | LPSS_DEV_I2C = 0, |
| 68 | LPSS_DEV_UART, |
| 69 | LPSS_DEV_SPI, |
| 70 | }; |
| 71 | |
| 72 | struct intel_lpss { |
| 73 | const struct intel_lpss_platform_info *info; |
| 74 | enum intel_lpss_dev_type type; |
| 75 | struct clk *clk; |
| 76 | struct clk_lookup *clock; |
Mika Westerberg | e15ad21 | 2015-11-30 17:11:41 +0200 | [diff] [blame] | 77 | struct mfd_cell *cell; |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 78 | struct device *dev; |
| 79 | void __iomem *priv; |
Heikki Krogerus | 41a3da2 | 2016-04-18 15:14:56 +0300 | [diff] [blame] | 80 | u32 priv_ctx[LPSS_PRIV_REG_COUNT]; |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 81 | int devid; |
| 82 | u32 caps; |
| 83 | u32 active_ltr; |
| 84 | u32 idle_ltr; |
| 85 | struct dentry *debugfs; |
| 86 | }; |
| 87 | |
| 88 | static const struct resource intel_lpss_dev_resources[] = { |
| 89 | DEFINE_RES_MEM_NAMED(LPSS_DEV_OFFSET, LPSS_DEV_SIZE, "lpss_dev"), |
| 90 | DEFINE_RES_MEM_NAMED(LPSS_PRIV_OFFSET, LPSS_PRIV_SIZE, "lpss_priv"), |
| 91 | DEFINE_RES_IRQ(0), |
| 92 | }; |
| 93 | |
| 94 | static const struct resource intel_lpss_idma64_resources[] = { |
| 95 | DEFINE_RES_MEM(LPSS_IDMA64_OFFSET, LPSS_IDMA64_SIZE), |
| 96 | DEFINE_RES_IRQ(0), |
| 97 | }; |
| 98 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 99 | /* |
| 100 | * Cells needs to be ordered so that the iDMA is created first. This is |
| 101 | * because we need to be sure the DMA is available when the host controller |
| 102 | * driver is probed. |
| 103 | */ |
| 104 | static const struct mfd_cell intel_lpss_idma64_cell = { |
| 105 | .name = LPSS_IDMA64_DRIVER_NAME, |
| 106 | .num_resources = ARRAY_SIZE(intel_lpss_idma64_resources), |
| 107 | .resources = intel_lpss_idma64_resources, |
| 108 | }; |
| 109 | |
| 110 | static const struct mfd_cell intel_lpss_i2c_cell = { |
| 111 | .name = "i2c_designware", |
| 112 | .num_resources = ARRAY_SIZE(intel_lpss_dev_resources), |
| 113 | .resources = intel_lpss_dev_resources, |
| 114 | }; |
| 115 | |
| 116 | static const struct mfd_cell intel_lpss_uart_cell = { |
| 117 | .name = "dw-apb-uart", |
| 118 | .num_resources = ARRAY_SIZE(intel_lpss_dev_resources), |
| 119 | .resources = intel_lpss_dev_resources, |
| 120 | }; |
| 121 | |
| 122 | static const struct mfd_cell intel_lpss_spi_cell = { |
| 123 | .name = "pxa2xx-spi", |
| 124 | .num_resources = ARRAY_SIZE(intel_lpss_dev_resources), |
| 125 | .resources = intel_lpss_dev_resources, |
| 126 | }; |
| 127 | |
| 128 | static DEFINE_IDA(intel_lpss_devid_ida); |
| 129 | static struct dentry *intel_lpss_debugfs; |
| 130 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 131 | static void intel_lpss_cache_ltr(struct intel_lpss *lpss) |
| 132 | { |
| 133 | lpss->active_ltr = readl(lpss->priv + LPSS_PRIV_ACTIVELTR); |
| 134 | lpss->idle_ltr = readl(lpss->priv + LPSS_PRIV_IDLELTR); |
| 135 | } |
| 136 | |
| 137 | static int intel_lpss_debugfs_add(struct intel_lpss *lpss) |
| 138 | { |
| 139 | struct dentry *dir; |
| 140 | |
| 141 | dir = debugfs_create_dir(dev_name(lpss->dev), intel_lpss_debugfs); |
| 142 | if (IS_ERR(dir)) |
| 143 | return PTR_ERR(dir); |
| 144 | |
| 145 | /* Cache the values into lpss structure */ |
| 146 | intel_lpss_cache_ltr(lpss); |
| 147 | |
| 148 | debugfs_create_x32("capabilities", S_IRUGO, dir, &lpss->caps); |
| 149 | debugfs_create_x32("active_ltr", S_IRUGO, dir, &lpss->active_ltr); |
| 150 | debugfs_create_x32("idle_ltr", S_IRUGO, dir, &lpss->idle_ltr); |
| 151 | |
| 152 | lpss->debugfs = dir; |
| 153 | return 0; |
| 154 | } |
| 155 | |
| 156 | static void intel_lpss_debugfs_remove(struct intel_lpss *lpss) |
| 157 | { |
| 158 | debugfs_remove_recursive(lpss->debugfs); |
| 159 | } |
| 160 | |
| 161 | static void intel_lpss_ltr_set(struct device *dev, s32 val) |
| 162 | { |
| 163 | struct intel_lpss *lpss = dev_get_drvdata(dev); |
| 164 | u32 ltr; |
| 165 | |
| 166 | /* |
| 167 | * Program latency tolerance (LTR) accordingly what has been asked |
| 168 | * by the PM QoS layer or disable it in case we were passed |
| 169 | * negative value or PM_QOS_LATENCY_ANY. |
| 170 | */ |
| 171 | ltr = readl(lpss->priv + LPSS_PRIV_ACTIVELTR); |
| 172 | |
| 173 | if (val == PM_QOS_LATENCY_ANY || val < 0) { |
| 174 | ltr &= ~LPSS_PRIV_LTR_REQ; |
| 175 | } else { |
| 176 | ltr |= LPSS_PRIV_LTR_REQ; |
| 177 | ltr &= ~LPSS_PRIV_LTR_SCALE_MASK; |
| 178 | ltr &= ~LPSS_PRIV_LTR_VALUE_MASK; |
| 179 | |
| 180 | if (val > LPSS_PRIV_LTR_VALUE_MASK) |
| 181 | ltr |= LPSS_PRIV_LTR_SCALE_32US | val >> 5; |
| 182 | else |
| 183 | ltr |= LPSS_PRIV_LTR_SCALE_1US | val; |
| 184 | } |
| 185 | |
| 186 | if (ltr == lpss->active_ltr) |
| 187 | return; |
| 188 | |
| 189 | writel(ltr, lpss->priv + LPSS_PRIV_ACTIVELTR); |
| 190 | writel(ltr, lpss->priv + LPSS_PRIV_IDLELTR); |
| 191 | |
| 192 | /* Cache the values into lpss structure */ |
| 193 | intel_lpss_cache_ltr(lpss); |
| 194 | } |
| 195 | |
| 196 | static void intel_lpss_ltr_expose(struct intel_lpss *lpss) |
| 197 | { |
| 198 | lpss->dev->power.set_latency_tolerance = intel_lpss_ltr_set; |
| 199 | dev_pm_qos_expose_latency_tolerance(lpss->dev); |
| 200 | } |
| 201 | |
| 202 | static void intel_lpss_ltr_hide(struct intel_lpss *lpss) |
| 203 | { |
| 204 | dev_pm_qos_hide_latency_tolerance(lpss->dev); |
| 205 | lpss->dev->power.set_latency_tolerance = NULL; |
| 206 | } |
| 207 | |
| 208 | static int intel_lpss_assign_devs(struct intel_lpss *lpss) |
| 209 | { |
Mika Westerberg | e15ad21 | 2015-11-30 17:11:41 +0200 | [diff] [blame] | 210 | const struct mfd_cell *cell; |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 211 | unsigned int type; |
| 212 | |
| 213 | type = lpss->caps & LPSS_PRIV_CAPS_TYPE_MASK; |
| 214 | type >>= LPSS_PRIV_CAPS_TYPE_SHIFT; |
| 215 | |
| 216 | switch (type) { |
| 217 | case LPSS_DEV_I2C: |
Mika Westerberg | e15ad21 | 2015-11-30 17:11:41 +0200 | [diff] [blame] | 218 | cell = &intel_lpss_i2c_cell; |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 219 | break; |
| 220 | case LPSS_DEV_UART: |
Mika Westerberg | e15ad21 | 2015-11-30 17:11:41 +0200 | [diff] [blame] | 221 | cell = &intel_lpss_uart_cell; |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 222 | break; |
| 223 | case LPSS_DEV_SPI: |
Mika Westerberg | e15ad21 | 2015-11-30 17:11:41 +0200 | [diff] [blame] | 224 | cell = &intel_lpss_spi_cell; |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 225 | break; |
| 226 | default: |
| 227 | return -ENODEV; |
| 228 | } |
| 229 | |
Mika Westerberg | e15ad21 | 2015-11-30 17:11:41 +0200 | [diff] [blame] | 230 | lpss->cell = devm_kmemdup(lpss->dev, cell, sizeof(*cell), GFP_KERNEL); |
| 231 | if (!lpss->cell) |
| 232 | return -ENOMEM; |
| 233 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 234 | lpss->type = type; |
| 235 | |
| 236 | return 0; |
| 237 | } |
| 238 | |
| 239 | static bool intel_lpss_has_idma(const struct intel_lpss *lpss) |
| 240 | { |
| 241 | return (lpss->caps & LPSS_PRIV_CAPS_NO_IDMA) == 0; |
| 242 | } |
| 243 | |
| 244 | static void intel_lpss_set_remap_addr(const struct intel_lpss *lpss) |
| 245 | { |
| 246 | resource_size_t addr = lpss->info->mem->start; |
| 247 | |
Andy Shevchenko | 689d445 | 2015-09-14 11:32:48 +0300 | [diff] [blame] | 248 | lo_hi_writeq(addr, lpss->priv + LPSS_PRIV_REMAP_ADDR); |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 249 | } |
| 250 | |
| 251 | static void intel_lpss_deassert_reset(const struct intel_lpss *lpss) |
| 252 | { |
| 253 | u32 value = LPSS_PRIV_RESETS_FUNC | LPSS_PRIV_RESETS_IDMA; |
| 254 | |
| 255 | /* Bring out the device from reset */ |
| 256 | writel(value, lpss->priv + LPSS_PRIV_RESETS); |
| 257 | } |
| 258 | |
| 259 | static void intel_lpss_init_dev(const struct intel_lpss *lpss) |
| 260 | { |
| 261 | u32 value = LPSS_PRIV_SSP_REG_DIS_DMA_FIN; |
| 262 | |
Binbin Wu | dad0653 | 2019-04-08 16:09:10 +0800 | [diff] [blame] | 263 | /* Set the device in reset state */ |
| 264 | writel(0, lpss->priv + LPSS_PRIV_RESETS); |
| 265 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 266 | intel_lpss_deassert_reset(lpss); |
| 267 | |
Andy Shevchenko | d28b625 | 2018-04-24 18:00:10 +0300 | [diff] [blame] | 268 | intel_lpss_set_remap_addr(lpss); |
| 269 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 270 | if (!intel_lpss_has_idma(lpss)) |
| 271 | return; |
| 272 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 273 | /* Make sure that SPI multiblock DMA transfers are re-enabled */ |
| 274 | if (lpss->type == LPSS_DEV_SPI) |
| 275 | writel(value, lpss->priv + LPSS_PRIV_SSP_REG); |
| 276 | } |
| 277 | |
| 278 | static void intel_lpss_unregister_clock_tree(struct clk *clk) |
| 279 | { |
| 280 | struct clk *parent; |
| 281 | |
| 282 | while (clk) { |
| 283 | parent = clk_get_parent(clk); |
| 284 | clk_unregister(clk); |
| 285 | clk = parent; |
| 286 | } |
| 287 | } |
| 288 | |
| 289 | static int intel_lpss_register_clock_divider(struct intel_lpss *lpss, |
| 290 | const char *devname, |
| 291 | struct clk **clk) |
| 292 | { |
| 293 | char name[32]; |
| 294 | struct clk *tmp = *clk; |
| 295 | |
| 296 | snprintf(name, sizeof(name), "%s-enable", devname); |
| 297 | tmp = clk_register_gate(NULL, name, __clk_get_name(tmp), 0, |
| 298 | lpss->priv, 0, 0, NULL); |
| 299 | if (IS_ERR(tmp)) |
| 300 | return PTR_ERR(tmp); |
| 301 | |
| 302 | snprintf(name, sizeof(name), "%s-div", devname); |
| 303 | tmp = clk_register_fractional_divider(NULL, name, __clk_get_name(tmp), |
Andy Shevchenko | 82f53f9 | 2021-08-12 20:00:24 +0300 | [diff] [blame] | 304 | CLK_FRAC_DIVIDER_POWER_OF_TWO_PS, |
| 305 | lpss->priv, 1, 15, 16, 15, 0, |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 306 | NULL); |
| 307 | if (IS_ERR(tmp)) |
| 308 | return PTR_ERR(tmp); |
| 309 | *clk = tmp; |
| 310 | |
| 311 | snprintf(name, sizeof(name), "%s-update", devname); |
| 312 | tmp = clk_register_gate(NULL, name, __clk_get_name(tmp), |
| 313 | CLK_SET_RATE_PARENT, lpss->priv, 31, 0, NULL); |
| 314 | if (IS_ERR(tmp)) |
| 315 | return PTR_ERR(tmp); |
| 316 | *clk = tmp; |
| 317 | |
| 318 | return 0; |
| 319 | } |
| 320 | |
| 321 | static int intel_lpss_register_clock(struct intel_lpss *lpss) |
| 322 | { |
| 323 | const struct mfd_cell *cell = lpss->cell; |
| 324 | struct clk *clk; |
| 325 | char devname[24]; |
| 326 | int ret; |
| 327 | |
| 328 | if (!lpss->info->clk_rate) |
| 329 | return 0; |
| 330 | |
| 331 | /* Root clock */ |
Stephen Boyd | 0f7e70e | 2016-04-19 18:20:49 -0700 | [diff] [blame] | 332 | clk = clk_register_fixed_rate(NULL, dev_name(lpss->dev), NULL, 0, |
| 333 | lpss->info->clk_rate); |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 334 | if (IS_ERR(clk)) |
| 335 | return PTR_ERR(clk); |
| 336 | |
| 337 | snprintf(devname, sizeof(devname), "%s.%d", cell->name, lpss->devid); |
| 338 | |
| 339 | /* |
| 340 | * Support for clock divider only if it has some preset value. |
| 341 | * Otherwise we assume that the divider is not used. |
| 342 | */ |
| 343 | if (lpss->type != LPSS_DEV_I2C) { |
| 344 | ret = intel_lpss_register_clock_divider(lpss, devname, &clk); |
| 345 | if (ret) |
| 346 | goto err_clk_register; |
| 347 | } |
| 348 | |
| 349 | ret = -ENOMEM; |
| 350 | |
| 351 | /* Clock for the host controller */ |
| 352 | lpss->clock = clkdev_create(clk, lpss->info->clk_con_id, "%s", devname); |
| 353 | if (!lpss->clock) |
| 354 | goto err_clk_register; |
| 355 | |
| 356 | lpss->clk = clk; |
| 357 | |
| 358 | return 0; |
| 359 | |
| 360 | err_clk_register: |
| 361 | intel_lpss_unregister_clock_tree(clk); |
| 362 | |
| 363 | return ret; |
| 364 | } |
| 365 | |
| 366 | static void intel_lpss_unregister_clock(struct intel_lpss *lpss) |
| 367 | { |
| 368 | if (IS_ERR_OR_NULL(lpss->clk)) |
| 369 | return; |
| 370 | |
| 371 | clkdev_drop(lpss->clock); |
| 372 | intel_lpss_unregister_clock_tree(lpss->clk); |
| 373 | } |
| 374 | |
| 375 | int intel_lpss_probe(struct device *dev, |
| 376 | const struct intel_lpss_platform_info *info) |
| 377 | { |
| 378 | struct intel_lpss *lpss; |
| 379 | int ret; |
| 380 | |
| 381 | if (!info || !info->mem || info->irq <= 0) |
| 382 | return -EINVAL; |
| 383 | |
| 384 | lpss = devm_kzalloc(dev, sizeof(*lpss), GFP_KERNEL); |
| 385 | if (!lpss) |
| 386 | return -ENOMEM; |
| 387 | |
Tuowen Zhao | a8ff78f | 2019-10-16 15:06:29 -0600 | [diff] [blame] | 388 | lpss->priv = devm_ioremap_uc(dev, info->mem->start + LPSS_PRIV_OFFSET, |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 389 | LPSS_PRIV_SIZE); |
| 390 | if (!lpss->priv) |
| 391 | return -ENOMEM; |
| 392 | |
| 393 | lpss->info = info; |
| 394 | lpss->dev = dev; |
| 395 | lpss->caps = readl(lpss->priv + LPSS_PRIV_CAPS); |
| 396 | |
| 397 | dev_set_drvdata(dev, lpss); |
| 398 | |
| 399 | ret = intel_lpss_assign_devs(lpss); |
| 400 | if (ret) |
| 401 | return ret; |
| 402 | |
Heikki Krogerus | 03152e3 | 2021-03-01 16:42:21 +0200 | [diff] [blame] | 403 | lpss->cell->swnode = info->swnode; |
Hans de Goede | e6b1420 | 2021-12-03 12:51:08 +0100 | [diff] [blame] | 404 | lpss->cell->ignore_resource_conflicts = info->ignore_resource_conflicts; |
Mika Westerberg | e15ad21 | 2015-11-30 17:11:41 +0200 | [diff] [blame] | 405 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 406 | intel_lpss_init_dev(lpss); |
| 407 | |
| 408 | lpss->devid = ida_simple_get(&intel_lpss_devid_ida, 0, 0, GFP_KERNEL); |
| 409 | if (lpss->devid < 0) |
| 410 | return lpss->devid; |
| 411 | |
| 412 | ret = intel_lpss_register_clock(lpss); |
| 413 | if (ret) |
| 414 | goto err_clk_register; |
| 415 | |
| 416 | intel_lpss_ltr_expose(lpss); |
| 417 | |
| 418 | ret = intel_lpss_debugfs_add(lpss); |
| 419 | if (ret) |
| 420 | dev_warn(dev, "Failed to create debugfs entries\n"); |
| 421 | |
| 422 | if (intel_lpss_has_idma(lpss)) { |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 423 | ret = mfd_add_devices(dev, lpss->devid, &intel_lpss_idma64_cell, |
| 424 | 1, info->mem, info->irq, NULL); |
| 425 | if (ret) |
| 426 | dev_warn(dev, "Failed to add %s, fallback to PIO\n", |
| 427 | LPSS_IDMA64_DRIVER_NAME); |
| 428 | } |
| 429 | |
| 430 | ret = mfd_add_devices(dev, lpss->devid, lpss->cell, |
| 431 | 1, info->mem, info->irq, NULL); |
| 432 | if (ret) |
| 433 | goto err_remove_ltr; |
| 434 | |
Rafael J. Wysocki | 8425ec7 | 2018-01-03 01:34:53 +0100 | [diff] [blame] | 435 | dev_pm_set_driver_flags(dev, DPM_FLAG_SMART_SUSPEND); |
| 436 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 437 | return 0; |
| 438 | |
| 439 | err_remove_ltr: |
| 440 | intel_lpss_debugfs_remove(lpss); |
| 441 | intel_lpss_ltr_hide(lpss); |
Andy Shevchenko | 84cb36c | 2016-01-22 16:48:46 +0200 | [diff] [blame] | 442 | intel_lpss_unregister_clock(lpss); |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 443 | |
| 444 | err_clk_register: |
| 445 | ida_simple_remove(&intel_lpss_devid_ida, lpss->devid); |
| 446 | |
| 447 | return ret; |
| 448 | } |
| 449 | EXPORT_SYMBOL_GPL(intel_lpss_probe); |
| 450 | |
| 451 | void intel_lpss_remove(struct device *dev) |
| 452 | { |
| 453 | struct intel_lpss *lpss = dev_get_drvdata(dev); |
| 454 | |
| 455 | mfd_remove_devices(dev); |
| 456 | intel_lpss_debugfs_remove(lpss); |
| 457 | intel_lpss_ltr_hide(lpss); |
| 458 | intel_lpss_unregister_clock(lpss); |
| 459 | ida_simple_remove(&intel_lpss_devid_ida, lpss->devid); |
| 460 | } |
| 461 | EXPORT_SYMBOL_GPL(intel_lpss_remove); |
| 462 | |
| 463 | static int resume_lpss_device(struct device *dev, void *data) |
| 464 | { |
Rafael J. Wysocki | 8425ec7 | 2018-01-03 01:34:53 +0100 | [diff] [blame] | 465 | if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND)) |
| 466 | pm_runtime_resume(dev); |
| 467 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 468 | return 0; |
| 469 | } |
| 470 | |
| 471 | int intel_lpss_prepare(struct device *dev) |
| 472 | { |
| 473 | /* |
| 474 | * Resume both child devices before entering system sleep. This |
| 475 | * ensures that they are in proper state before they get suspended. |
| 476 | */ |
| 477 | device_for_each_child_reverse(dev, NULL, resume_lpss_device); |
| 478 | return 0; |
| 479 | } |
| 480 | EXPORT_SYMBOL_GPL(intel_lpss_prepare); |
| 481 | |
| 482 | int intel_lpss_suspend(struct device *dev) |
| 483 | { |
Heikki Krogerus | 41a3da2 | 2016-04-18 15:14:56 +0300 | [diff] [blame] | 484 | struct intel_lpss *lpss = dev_get_drvdata(dev); |
| 485 | unsigned int i; |
| 486 | |
| 487 | /* Save device context */ |
| 488 | for (i = 0; i < LPSS_PRIV_REG_COUNT; i++) |
| 489 | lpss->priv_ctx[i] = readl(lpss->priv + i * 4); |
| 490 | |
Furquan Shaikh | 0b471aa | 2017-07-23 23:02:19 -0700 | [diff] [blame] | 491 | /* |
| 492 | * If the device type is not UART, then put the controller into |
| 493 | * reset. UART cannot be put into reset since S3/S0ix fail when |
| 494 | * no_console_suspend flag is enabled. |
| 495 | */ |
| 496 | if (lpss->type != LPSS_DEV_UART) |
| 497 | writel(0, lpss->priv + LPSS_PRIV_RESETS); |
| 498 | |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 499 | return 0; |
| 500 | } |
| 501 | EXPORT_SYMBOL_GPL(intel_lpss_suspend); |
| 502 | |
| 503 | int intel_lpss_resume(struct device *dev) |
| 504 | { |
| 505 | struct intel_lpss *lpss = dev_get_drvdata(dev); |
Heikki Krogerus | 41a3da2 | 2016-04-18 15:14:56 +0300 | [diff] [blame] | 506 | unsigned int i; |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 507 | |
Heikki Krogerus | 41a3da2 | 2016-04-18 15:14:56 +0300 | [diff] [blame] | 508 | intel_lpss_deassert_reset(lpss); |
| 509 | |
| 510 | /* Restore device context */ |
| 511 | for (i = 0; i < LPSS_PRIV_REG_COUNT; i++) |
| 512 | writel(lpss->priv_ctx[i], lpss->priv + i * 4); |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 513 | |
| 514 | return 0; |
| 515 | } |
| 516 | EXPORT_SYMBOL_GPL(intel_lpss_resume); |
| 517 | |
| 518 | static int __init intel_lpss_init(void) |
| 519 | { |
| 520 | intel_lpss_debugfs = debugfs_create_dir("intel_lpss", NULL); |
| 521 | return 0; |
| 522 | } |
| 523 | module_init(intel_lpss_init); |
| 524 | |
| 525 | static void __exit intel_lpss_exit(void) |
| 526 | { |
Andy Shevchenko | 02f3691 | 2019-06-21 15:56:31 +0300 | [diff] [blame] | 527 | ida_destroy(&intel_lpss_devid_ida); |
Andy Shevchenko | 4b45efe | 2015-07-27 18:04:03 +0300 | [diff] [blame] | 528 | debugfs_remove(intel_lpss_debugfs); |
| 529 | } |
| 530 | module_exit(intel_lpss_exit); |
| 531 | |
| 532 | MODULE_AUTHOR("Andy Shevchenko <andriy.shevchenko@linux.intel.com>"); |
| 533 | MODULE_AUTHOR("Mika Westerberg <mika.westerberg@linux.intel.com>"); |
| 534 | MODULE_AUTHOR("Heikki Krogerus <heikki.krogerus@linux.intel.com>"); |
| 535 | MODULE_AUTHOR("Jarkko Nikula <jarkko.nikula@linux.intel.com>"); |
| 536 | MODULE_DESCRIPTION("Intel LPSS core driver"); |
| 537 | MODULE_LICENSE("GPL v2"); |
Andy Shevchenko | 569fac7 | 2019-08-21 11:37:12 +0300 | [diff] [blame] | 538 | /* |
| 539 | * Ensure the DMA driver is loaded before the host controller device appears, |
| 540 | * so that the host controller driver can request its DMA channels as early |
| 541 | * as possible. |
| 542 | * |
| 543 | * If the DMA module is not there that's OK as well. |
| 544 | */ |
| 545 | MODULE_SOFTDEP("pre: platform:" LPSS_IDMA64_DRIVER_NAME); |