Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 1 | /* |
| 2 | * Copyright © 2014 Intel Corporation |
| 3 | * |
| 4 | * Permission is hereby granted, free of charge, to any person obtaining a |
| 5 | * copy of this software and associated documentation files (the "Software"), |
| 6 | * to deal in the Software without restriction, including without limitation |
| 7 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| 8 | * and/or sell copies of the Software, and to permit persons to whom the |
| 9 | * Software is furnished to do so, subject to the following conditions: |
| 10 | * |
| 11 | * The above copyright notice and this permission notice (including the next |
| 12 | * paragraph) shall be included in all copies or substantial portions of the |
| 13 | * Software. |
| 14 | * |
| 15 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| 16 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| 17 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| 18 | * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| 19 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
| 20 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS |
| 21 | * IN THE SOFTWARE. |
| 22 | * |
| 23 | */ |
| 24 | #include <linux/firmware.h> |
| 25 | #include "i915_drv.h" |
| 26 | #include "i915_reg.h" |
| 27 | |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 28 | /** |
| 29 | * DOC: csr support for dmc |
| 30 | * |
| 31 | * Display Context Save and Restore (CSR) firmware support added from gen9 |
| 32 | * onwards to drive newly added DMC (Display microcontroller) in display |
| 33 | * engine to save and restore the state of display engine when it enter into |
| 34 | * low-power state and comes back to normal. |
| 35 | * |
| 36 | * Firmware loading status will be one of the below states: FW_UNINITIALIZED, |
| 37 | * FW_LOADED, FW_FAILED. |
| 38 | * |
| 39 | * Once the firmware is written into the registers status will be moved from |
| 40 | * FW_UNINITIALIZED to FW_LOADED and for any erroneous condition status will |
| 41 | * be moved to FW_FAILED. |
| 42 | */ |
| 43 | |
Rodrigo Vivi | a4a027a | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 44 | #define I915_CSR_KBL "i915/kbl_dmc_ver1.bin" |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 45 | MODULE_FIRMWARE(I915_CSR_KBL); |
| 46 | #define KBL_CSR_VERSION_REQUIRED CSR_VERSION(1, 1) |
| 47 | |
Rodrigo Vivi | bf546f8 | 2015-06-03 16:50:19 -0700 | [diff] [blame] | 48 | #define I915_CSR_SKL "i915/skl_dmc_ver1.bin" |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 49 | MODULE_FIRMWARE(I915_CSR_SKL); |
Rodrigo Vivi | a4a027a | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 50 | #define SKL_CSR_VERSION_REQUIRED CSR_VERSION(1, 23) |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 51 | |
Animesh Manna | 18c237c | 2015-08-04 22:02:41 +0530 | [diff] [blame] | 52 | #define I915_CSR_BXT "i915/bxt_dmc_ver1.bin" |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 53 | MODULE_FIRMWARE(I915_CSR_BXT); |
| 54 | #define BXT_CSR_VERSION_REQUIRED CSR_VERSION(1, 7) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 55 | |
Chris Wilson | cbfc2d2 | 2016-01-13 17:38:15 +0000 | [diff] [blame] | 56 | #define FIRMWARE_URL "https://01.org/linuxgraphics/intel-linux-graphics-firmwares" |
| 57 | |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 58 | |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 59 | |
Mika Kuoppala | 9c5308e | 2015-10-30 17:52:16 +0200 | [diff] [blame] | 60 | |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 61 | #define CSR_MAX_FW_SIZE 0x2FFF |
| 62 | #define CSR_DEFAULT_FW_OFFSET 0xFFFFFFFF |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 63 | |
| 64 | struct intel_css_header { |
| 65 | /* 0x09 for DMC */ |
| 66 | uint32_t module_type; |
| 67 | |
| 68 | /* Includes the DMC specific header in dwords */ |
| 69 | uint32_t header_len; |
| 70 | |
| 71 | /* always value would be 0x10000 */ |
| 72 | uint32_t header_ver; |
| 73 | |
| 74 | /* Not used */ |
| 75 | uint32_t module_id; |
| 76 | |
| 77 | /* Not used */ |
| 78 | uint32_t module_vendor; |
| 79 | |
| 80 | /* in YYYYMMDD format */ |
| 81 | uint32_t date; |
| 82 | |
| 83 | /* Size in dwords (CSS_Headerlen + PackageHeaderLen + dmc FWsLen)/4 */ |
| 84 | uint32_t size; |
| 85 | |
| 86 | /* Not used */ |
| 87 | uint32_t key_size; |
| 88 | |
| 89 | /* Not used */ |
| 90 | uint32_t modulus_size; |
| 91 | |
| 92 | /* Not used */ |
| 93 | uint32_t exponent_size; |
| 94 | |
| 95 | /* Not used */ |
| 96 | uint32_t reserved1[12]; |
| 97 | |
| 98 | /* Major Minor */ |
| 99 | uint32_t version; |
| 100 | |
| 101 | /* Not used */ |
| 102 | uint32_t reserved2[8]; |
| 103 | |
| 104 | /* Not used */ |
| 105 | uint32_t kernel_header_info; |
| 106 | } __packed; |
| 107 | |
| 108 | struct intel_fw_info { |
| 109 | uint16_t reserved1; |
| 110 | |
| 111 | /* Stepping (A, B, C, ..., *). * is a wildcard */ |
| 112 | char stepping; |
| 113 | |
| 114 | /* Sub-stepping (0, 1, ..., *). * is a wildcard */ |
| 115 | char substepping; |
| 116 | |
| 117 | uint32_t offset; |
| 118 | uint32_t reserved2; |
| 119 | } __packed; |
| 120 | |
| 121 | struct intel_package_header { |
| 122 | /* DMC container header length in dwords */ |
| 123 | unsigned char header_len; |
| 124 | |
| 125 | /* always value would be 0x01 */ |
| 126 | unsigned char header_ver; |
| 127 | |
| 128 | unsigned char reserved[10]; |
| 129 | |
| 130 | /* Number of valid entries in the FWInfo array below */ |
| 131 | uint32_t num_entries; |
| 132 | |
| 133 | struct intel_fw_info fw_info[20]; |
| 134 | } __packed; |
| 135 | |
| 136 | struct intel_dmc_header { |
| 137 | /* always value would be 0x40403E3E */ |
| 138 | uint32_t signature; |
| 139 | |
| 140 | /* DMC binary header length */ |
| 141 | unsigned char header_len; |
| 142 | |
| 143 | /* 0x01 */ |
| 144 | unsigned char header_ver; |
| 145 | |
| 146 | /* Reserved */ |
| 147 | uint16_t dmcc_ver; |
| 148 | |
| 149 | /* Major, Minor */ |
| 150 | uint32_t project; |
| 151 | |
| 152 | /* Firmware program size (excluding header) in dwords */ |
| 153 | uint32_t fw_size; |
| 154 | |
| 155 | /* Major Minor version */ |
| 156 | uint32_t fw_version; |
| 157 | |
| 158 | /* Number of valid MMIO cycles present. */ |
| 159 | uint32_t mmio_count; |
| 160 | |
| 161 | /* MMIO address */ |
| 162 | uint32_t mmioaddr[8]; |
| 163 | |
| 164 | /* MMIO data */ |
| 165 | uint32_t mmiodata[8]; |
| 166 | |
| 167 | /* FW filename */ |
| 168 | unsigned char dfile[32]; |
| 169 | |
| 170 | uint32_t reserved1[2]; |
| 171 | } __packed; |
| 172 | |
| 173 | struct stepping_info { |
| 174 | char stepping; |
| 175 | char substepping; |
| 176 | }; |
| 177 | |
Rodrigo Vivi | a25c9f0 | 2015-12-09 07:51:59 -0800 | [diff] [blame] | 178 | static const struct stepping_info kbl_stepping_info[] = { |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 179 | {'A', '0'}, {'B', '0'}, {'C', '0'}, |
| 180 | {'D', '0'}, {'E', '0'}, {'F', '0'}, |
| 181 | {'G', '0'}, {'H', '0'}, {'I', '0'}, |
Rodrigo Vivi | a25c9f0 | 2015-12-09 07:51:59 -0800 | [diff] [blame] | 182 | }; |
| 183 | |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 184 | static const struct stepping_info skl_stepping_info[] = { |
Jani Nikula | 84cb00e | 2015-10-20 15:38:31 +0300 | [diff] [blame] | 185 | {'A', '0'}, {'B', '0'}, {'C', '0'}, |
| 186 | {'D', '0'}, {'E', '0'}, {'F', '0'}, |
Mat Martineau | a41c888 | 2016-01-28 15:19:23 -0800 | [diff] [blame] | 187 | {'G', '0'}, {'H', '0'}, {'I', '0'}, |
| 188 | {'J', '0'}, {'K', '0'} |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 189 | }; |
| 190 | |
Jani Nikula | b9cd5bfd | 2015-10-20 15:38:32 +0300 | [diff] [blame] | 191 | static const struct stepping_info bxt_stepping_info[] = { |
Animesh Manna | cff765f | 2015-08-04 22:02:43 +0530 | [diff] [blame] | 192 | {'A', '0'}, {'A', '1'}, {'A', '2'}, |
| 193 | {'B', '0'}, {'B', '1'}, {'B', '2'} |
| 194 | }; |
| 195 | |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 196 | static const struct stepping_info no_stepping_info = { '*', '*' }; |
| 197 | |
| 198 | static const struct stepping_info * |
| 199 | intel_get_stepping_info(struct drm_i915_private *dev_priv) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 200 | { |
Jani Nikula | b1a14c6 | 2015-10-20 15:38:33 +0300 | [diff] [blame] | 201 | const struct stepping_info *si; |
| 202 | unsigned int size; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 203 | |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 204 | if (IS_KABYLAKE(dev_priv)) { |
Rodrigo Vivi | a25c9f0 | 2015-12-09 07:51:59 -0800 | [diff] [blame] | 205 | size = ARRAY_SIZE(kbl_stepping_info); |
| 206 | si = kbl_stepping_info; |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 207 | } else if (IS_SKYLAKE(dev_priv)) { |
Jani Nikula | b1a14c6 | 2015-10-20 15:38:33 +0300 | [diff] [blame] | 208 | size = ARRAY_SIZE(skl_stepping_info); |
| 209 | si = skl_stepping_info; |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 210 | } else if (IS_BROXTON(dev_priv)) { |
Jani Nikula | b1a14c6 | 2015-10-20 15:38:33 +0300 | [diff] [blame] | 211 | size = ARRAY_SIZE(bxt_stepping_info); |
| 212 | si = bxt_stepping_info; |
| 213 | } else { |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 214 | size = 0; |
Jani Nikula | b1a14c6 | 2015-10-20 15:38:33 +0300 | [diff] [blame] | 215 | } |
| 216 | |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 217 | if (INTEL_REVID(dev_priv) < size) |
| 218 | return si + INTEL_REVID(dev_priv); |
Jani Nikula | b1a14c6 | 2015-10-20 15:38:33 +0300 | [diff] [blame] | 219 | |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 220 | return &no_stepping_info; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 221 | } |
| 222 | |
Imre Deak | 2abc525 | 2016-03-04 21:57:41 +0200 | [diff] [blame] | 223 | static void gen9_set_dc_state_debugmask(struct drm_i915_private *dev_priv) |
| 224 | { |
| 225 | uint32_t val, mask; |
| 226 | |
| 227 | mask = DC_STATE_DEBUG_MASK_MEMORY_UP; |
| 228 | |
| 229 | if (IS_BROXTON(dev_priv)) |
| 230 | mask |= DC_STATE_DEBUG_MASK_CORES; |
| 231 | |
| 232 | /* The below bit doesn't need to be cleared ever afterwards */ |
| 233 | val = I915_READ(DC_STATE_DEBUG); |
| 234 | if ((val & mask) != mask) { |
| 235 | val |= mask; |
| 236 | I915_WRITE(DC_STATE_DEBUG, val); |
| 237 | POSTING_READ(DC_STATE_DEBUG); |
| 238 | } |
| 239 | } |
| 240 | |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 241 | /** |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 242 | * intel_csr_load_program() - write the firmware from memory to register. |
Daniel Vetter | f444837 | 2015-10-28 23:59:02 +0200 | [diff] [blame] | 243 | * @dev_priv: i915 drm device. |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 244 | * |
| 245 | * CSR firmware is read from a .bin file and kept in internal memory one time. |
| 246 | * Everytime display comes back from low power state this function is called to |
| 247 | * copy the firmware from internal memory to registers. |
| 248 | */ |
Imre Deak | 2abc525 | 2016-03-04 21:57:41 +0200 | [diff] [blame] | 249 | void intel_csr_load_program(struct drm_i915_private *dev_priv) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 250 | { |
Animesh Manna | a7f749f | 2015-08-03 21:55:32 +0530 | [diff] [blame] | 251 | u32 *payload = dev_priv->csr.dmc_payload; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 252 | uint32_t i, fw_size; |
| 253 | |
Daniel Vetter | f444837 | 2015-10-28 23:59:02 +0200 | [diff] [blame] | 254 | if (!IS_GEN9(dev_priv)) { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 255 | DRM_ERROR("No CSR support available for this platform\n"); |
Imre Deak | 2abc525 | 2016-03-04 21:57:41 +0200 | [diff] [blame] | 256 | return; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 257 | } |
| 258 | |
Patrik Jakobsson | fc131bf | 2015-11-09 16:48:16 +0100 | [diff] [blame] | 259 | if (!dev_priv->csr.dmc_payload) { |
| 260 | DRM_ERROR("Tried to program CSR with empty payload\n"); |
Imre Deak | 2abc525 | 2016-03-04 21:57:41 +0200 | [diff] [blame] | 261 | return; |
Patrik Jakobsson | fc131bf | 2015-11-09 16:48:16 +0100 | [diff] [blame] | 262 | } |
Animesh Manna | 4b7ab5f | 2015-08-26 01:36:05 +0530 | [diff] [blame] | 263 | |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 264 | fw_size = dev_priv->csr.dmc_fw_size; |
| 265 | for (i = 0; i < fw_size; i++) |
Ville Syrjälä | d2aa5ae | 2015-09-18 20:03:23 +0300 | [diff] [blame] | 266 | I915_WRITE(CSR_PROGRAM(i), payload[i]); |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 267 | |
| 268 | for (i = 0; i < dev_priv->csr.mmio_count; i++) { |
| 269 | I915_WRITE(dev_priv->csr.mmioaddr[i], |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 270 | dev_priv->csr.mmiodata[i]); |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 271 | } |
Patrik Jakobsson | 832dba8 | 2016-02-18 17:21:11 +0200 | [diff] [blame] | 272 | |
| 273 | dev_priv->csr.dc_state = 0; |
Mika Kuoppala | 1e657ad | 2016-02-18 17:21:14 +0200 | [diff] [blame] | 274 | |
Imre Deak | 2abc525 | 2016-03-04 21:57:41 +0200 | [diff] [blame] | 275 | gen9_set_dc_state_debugmask(dev_priv); |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 276 | } |
| 277 | |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 278 | static uint32_t *parse_csr_fw(struct drm_i915_private *dev_priv, |
| 279 | const struct firmware *fw) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 280 | { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 281 | struct intel_css_header *css_header; |
| 282 | struct intel_package_header *package_header; |
| 283 | struct intel_dmc_header *dmc_header; |
| 284 | struct intel_csr *csr = &dev_priv->csr; |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 285 | const struct stepping_info *si = intel_get_stepping_info(dev_priv); |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 286 | uint32_t dmc_offset = CSR_DEFAULT_FW_OFFSET, readcount = 0, nbytes; |
| 287 | uint32_t i; |
Animesh Manna | a7f749f | 2015-08-03 21:55:32 +0530 | [diff] [blame] | 288 | uint32_t *dmc_payload; |
Patrik Jakobsson | 4aa7fb9 | 2016-05-16 11:30:57 +0200 | [diff] [blame] | 289 | uint32_t required_version; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 290 | |
Mika Kuoppala | 9c5308e | 2015-10-30 17:52:16 +0200 | [diff] [blame] | 291 | if (!fw) |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 292 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 293 | |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 294 | /* Extract CSS Header information*/ |
| 295 | css_header = (struct intel_css_header *)fw->data; |
| 296 | if (sizeof(struct intel_css_header) != |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 297 | (css_header->header_len * 4)) { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 298 | DRM_ERROR("Firmware has wrong CSS header length %u bytes\n", |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 299 | (css_header->header_len * 4)); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 300 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 301 | } |
Damien Lespiau | b6e7d89 | 2015-10-27 14:46:59 +0200 | [diff] [blame] | 302 | |
| 303 | csr->version = css_header->version; |
| 304 | |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 305 | if (IS_KABYLAKE(dev_priv)) { |
Patrik Jakobsson | 4aa7fb9 | 2016-05-16 11:30:57 +0200 | [diff] [blame] | 306 | required_version = KBL_CSR_VERSION_REQUIRED; |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 307 | } else if (IS_SKYLAKE(dev_priv)) { |
Patrik Jakobsson | 4aa7fb9 | 2016-05-16 11:30:57 +0200 | [diff] [blame] | 308 | required_version = SKL_CSR_VERSION_REQUIRED; |
Imre Deak | e796853 | 2016-04-01 16:02:32 +0300 | [diff] [blame] | 309 | } else if (IS_BROXTON(dev_priv)) { |
Patrik Jakobsson | 4aa7fb9 | 2016-05-16 11:30:57 +0200 | [diff] [blame] | 310 | required_version = BXT_CSR_VERSION_REQUIRED; |
Imre Deak | e796853 | 2016-04-01 16:02:32 +0300 | [diff] [blame] | 311 | } else { |
| 312 | MISSING_CASE(INTEL_REVID(dev_priv)); |
Patrik Jakobsson | 4aa7fb9 | 2016-05-16 11:30:57 +0200 | [diff] [blame] | 313 | required_version = 0; |
Imre Deak | e796853 | 2016-04-01 16:02:32 +0300 | [diff] [blame] | 314 | } |
| 315 | |
Patrik Jakobsson | 4aa7fb9 | 2016-05-16 11:30:57 +0200 | [diff] [blame] | 316 | if (csr->version != required_version) { |
| 317 | DRM_INFO("Refusing to load DMC firmware v%u.%u," |
| 318 | " please use v%u.%u [" FIRMWARE_URL "].\n", |
Mika Kuoppala | 9c5308e | 2015-10-30 17:52:16 +0200 | [diff] [blame] | 319 | CSR_VERSION_MAJOR(csr->version), |
| 320 | CSR_VERSION_MINOR(csr->version), |
Patrik Jakobsson | 4aa7fb9 | 2016-05-16 11:30:57 +0200 | [diff] [blame] | 321 | CSR_VERSION_MAJOR(required_version), |
| 322 | CSR_VERSION_MINOR(required_version)); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 323 | return NULL; |
Mika Kuoppala | 9c5308e | 2015-10-30 17:52:16 +0200 | [diff] [blame] | 324 | } |
| 325 | |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 326 | readcount += sizeof(struct intel_css_header); |
| 327 | |
| 328 | /* Extract Package Header information*/ |
| 329 | package_header = (struct intel_package_header *) |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 330 | &fw->data[readcount]; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 331 | if (sizeof(struct intel_package_header) != |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 332 | (package_header->header_len * 4)) { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 333 | DRM_ERROR("Firmware has wrong package header length %u bytes\n", |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 334 | (package_header->header_len * 4)); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 335 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 336 | } |
| 337 | readcount += sizeof(struct intel_package_header); |
| 338 | |
| 339 | /* Search for dmc_offset to find firware binary. */ |
| 340 | for (i = 0; i < package_header->num_entries; i++) { |
| 341 | if (package_header->fw_info[i].substepping == '*' && |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 342 | si->stepping == package_header->fw_info[i].stepping) { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 343 | dmc_offset = package_header->fw_info[i].offset; |
| 344 | break; |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 345 | } else if (si->stepping == package_header->fw_info[i].stepping && |
| 346 | si->substepping == package_header->fw_info[i].substepping) { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 347 | dmc_offset = package_header->fw_info[i].offset; |
| 348 | break; |
| 349 | } else if (package_header->fw_info[i].stepping == '*' && |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 350 | package_header->fw_info[i].substepping == '*') |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 351 | dmc_offset = package_header->fw_info[i].offset; |
| 352 | } |
| 353 | if (dmc_offset == CSR_DEFAULT_FW_OFFSET) { |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 354 | DRM_ERROR("Firmware not supported for %c stepping\n", |
| 355 | si->stepping); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 356 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 357 | } |
| 358 | readcount += dmc_offset; |
| 359 | |
| 360 | /* Extract dmc_header information. */ |
| 361 | dmc_header = (struct intel_dmc_header *)&fw->data[readcount]; |
| 362 | if (sizeof(struct intel_dmc_header) != (dmc_header->header_len)) { |
| 363 | DRM_ERROR("Firmware has wrong dmc header length %u bytes\n", |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 364 | (dmc_header->header_len)); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 365 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 366 | } |
| 367 | readcount += sizeof(struct intel_dmc_header); |
| 368 | |
| 369 | /* Cache the dmc header info. */ |
| 370 | if (dmc_header->mmio_count > ARRAY_SIZE(csr->mmioaddr)) { |
| 371 | DRM_ERROR("Firmware has wrong mmio count %u\n", |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 372 | dmc_header->mmio_count); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 373 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 374 | } |
| 375 | csr->mmio_count = dmc_header->mmio_count; |
| 376 | for (i = 0; i < dmc_header->mmio_count; i++) { |
Takashi Iwai | 982b0b2 | 2015-09-09 16:52:09 +0200 | [diff] [blame] | 377 | if (dmc_header->mmioaddr[i] < CSR_MMIO_START_RANGE || |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 378 | dmc_header->mmioaddr[i] > CSR_MMIO_END_RANGE) { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 379 | DRM_ERROR(" Firmware has wrong mmio address 0x%x\n", |
Daniel Vetter | f98f70d | 2015-10-28 23:58:59 +0200 | [diff] [blame] | 380 | dmc_header->mmioaddr[i]); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 381 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 382 | } |
Ville Syrjälä | f0f59a0 | 2015-11-18 15:33:26 +0200 | [diff] [blame] | 383 | csr->mmioaddr[i] = _MMIO(dmc_header->mmioaddr[i]); |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 384 | csr->mmiodata[i] = dmc_header->mmiodata[i]; |
| 385 | } |
| 386 | |
| 387 | /* fw_size is in dwords, so multiplied by 4 to convert into bytes. */ |
| 388 | nbytes = dmc_header->fw_size * 4; |
| 389 | if (nbytes > CSR_MAX_FW_SIZE) { |
| 390 | DRM_ERROR("CSR firmware too big (%u) bytes\n", nbytes); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 391 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 392 | } |
| 393 | csr->dmc_fw_size = dmc_header->fw_size; |
| 394 | |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 395 | dmc_payload = kmalloc(nbytes, GFP_KERNEL); |
| 396 | if (!dmc_payload) { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 397 | DRM_ERROR("Memory allocation failed for dmc payload\n"); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 398 | return NULL; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 399 | } |
| 400 | |
Chris Wilson | 1bb4308 | 2016-03-07 12:05:57 +0000 | [diff] [blame] | 401 | return memcpy(dmc_payload, &fw->data[readcount], nbytes); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 402 | } |
| 403 | |
Daniel Vetter | 8144ac5 | 2015-10-28 23:59:04 +0200 | [diff] [blame] | 404 | static void csr_load_work_fn(struct work_struct *work) |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 405 | { |
Daniel Vetter | 8144ac5 | 2015-10-28 23:59:04 +0200 | [diff] [blame] | 406 | struct drm_i915_private *dev_priv; |
| 407 | struct intel_csr *csr; |
| 408 | const struct firmware *fw; |
| 409 | int ret; |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 410 | |
Daniel Vetter | 8144ac5 | 2015-10-28 23:59:04 +0200 | [diff] [blame] | 411 | dev_priv = container_of(work, typeof(*dev_priv), csr.work); |
| 412 | csr = &dev_priv->csr; |
| 413 | |
| 414 | ret = request_firmware(&fw, dev_priv->csr.fw_path, |
Chris Wilson | 91c8a32 | 2016-07-05 10:40:23 +0100 | [diff] [blame] | 415 | &dev_priv->drm.pdev->dev); |
Imre Deak | 2abc525 | 2016-03-04 21:57:41 +0200 | [diff] [blame] | 416 | if (fw) |
| 417 | dev_priv->csr.dmc_payload = parse_csr_fw(dev_priv, fw); |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 418 | |
Daniel Vetter | 6a6582b | 2015-11-12 17:11:29 +0200 | [diff] [blame] | 419 | if (dev_priv->csr.dmc_payload) { |
Imre Deak | 2abc525 | 2016-03-04 21:57:41 +0200 | [diff] [blame] | 420 | intel_csr_load_program(dev_priv); |
| 421 | |
Daniel Vetter | 01a6908 | 2015-10-28 23:58:56 +0200 | [diff] [blame] | 422 | intel_display_power_put(dev_priv, POWER_DOMAIN_INIT); |
Mika Kuoppala | 9c5308e | 2015-10-30 17:52:16 +0200 | [diff] [blame] | 423 | |
| 424 | DRM_INFO("Finished loading %s (v%u.%u)\n", |
| 425 | dev_priv->csr.fw_path, |
| 426 | CSR_VERSION_MAJOR(csr->version), |
| 427 | CSR_VERSION_MINOR(csr->version)); |
| 428 | } else { |
Chris Wilson | 91c8a32 | 2016-07-05 10:40:23 +0100 | [diff] [blame] | 429 | dev_notice(dev_priv->drm.dev, |
Chris Wilson | cbfc2d2 | 2016-01-13 17:38:15 +0000 | [diff] [blame] | 430 | "Failed to load DMC firmware" |
| 431 | " [" FIRMWARE_URL "]," |
| 432 | " disabling runtime power management.\n"); |
Mika Kuoppala | 9c5308e | 2015-10-30 17:52:16 +0200 | [diff] [blame] | 433 | } |
| 434 | |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 435 | release_firmware(fw); |
| 436 | } |
| 437 | |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 438 | /** |
| 439 | * intel_csr_ucode_init() - initialize the firmware loading. |
Daniel Vetter | f444837 | 2015-10-28 23:59:02 +0200 | [diff] [blame] | 440 | * @dev_priv: i915 drm device. |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 441 | * |
| 442 | * This function is called at the time of loading the display driver to read |
| 443 | * firmware from a .bin file and copied into a internal memory. |
| 444 | */ |
Daniel Vetter | f444837 | 2015-10-28 23:59:02 +0200 | [diff] [blame] | 445 | void intel_csr_ucode_init(struct drm_i915_private *dev_priv) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 446 | { |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 447 | struct intel_csr *csr = &dev_priv->csr; |
Daniel Vetter | 8144ac5 | 2015-10-28 23:59:04 +0200 | [diff] [blame] | 448 | |
| 449 | INIT_WORK(&dev_priv->csr.work, csr_load_work_fn); |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 450 | |
Daniel Vetter | f444837 | 2015-10-28 23:59:02 +0200 | [diff] [blame] | 451 | if (!HAS_CSR(dev_priv)) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 452 | return; |
| 453 | |
Rodrigo Vivi | 4922d49 | 2016-04-26 14:59:51 -0700 | [diff] [blame] | 454 | if (IS_KABYLAKE(dev_priv)) |
| 455 | csr->fw_path = I915_CSR_KBL; |
| 456 | else if (IS_SKYLAKE(dev_priv)) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 457 | csr->fw_path = I915_CSR_SKL; |
Animesh Manna | 18c237c | 2015-08-04 22:02:41 +0530 | [diff] [blame] | 458 | else if (IS_BROXTON(dev_priv)) |
| 459 | csr->fw_path = I915_CSR_BXT; |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 460 | else { |
| 461 | DRM_ERROR("Unexpected: no known CSR firmware for platform\n"); |
| 462 | return; |
| 463 | } |
| 464 | |
Damien Lespiau | abd41dc | 2015-06-04 16:42:16 +0100 | [diff] [blame] | 465 | DRM_DEBUG_KMS("Loading %s\n", csr->fw_path); |
| 466 | |
Suketu Shah | dc17430 | 2015-04-17 19:46:16 +0530 | [diff] [blame] | 467 | /* |
| 468 | * Obtain a runtime pm reference, until CSR is loaded, |
| 469 | * to avoid entering runtime-suspend. |
| 470 | */ |
Daniel Vetter | 01a6908 | 2015-10-28 23:58:56 +0200 | [diff] [blame] | 471 | intel_display_power_get(dev_priv, POWER_DOMAIN_INIT); |
Suketu Shah | dc17430 | 2015-04-17 19:46:16 +0530 | [diff] [blame] | 472 | |
Daniel Vetter | 8144ac5 | 2015-10-28 23:59:04 +0200 | [diff] [blame] | 473 | schedule_work(&dev_priv->csr.work); |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 474 | } |
| 475 | |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 476 | /** |
Imre Deak | f74ed08 | 2016-04-18 14:48:21 +0300 | [diff] [blame] | 477 | * intel_csr_ucode_suspend() - prepare CSR firmware before system suspend |
| 478 | * @dev_priv: i915 drm device |
| 479 | * |
| 480 | * Prepare the DMC firmware before entering system suspend. This includes |
| 481 | * flushing pending work items and releasing any resources acquired during |
| 482 | * init. |
| 483 | */ |
| 484 | void intel_csr_ucode_suspend(struct drm_i915_private *dev_priv) |
| 485 | { |
| 486 | if (!HAS_CSR(dev_priv)) |
| 487 | return; |
| 488 | |
| 489 | flush_work(&dev_priv->csr.work); |
| 490 | |
| 491 | /* Drop the reference held in case DMC isn't loaded. */ |
| 492 | if (!dev_priv->csr.dmc_payload) |
| 493 | intel_display_power_put(dev_priv, POWER_DOMAIN_INIT); |
| 494 | } |
| 495 | |
| 496 | /** |
| 497 | * intel_csr_ucode_resume() - init CSR firmware during system resume |
| 498 | * @dev_priv: i915 drm device |
| 499 | * |
| 500 | * Reinitialize the DMC firmware during system resume, reacquiring any |
| 501 | * resources released in intel_csr_ucode_suspend(). |
| 502 | */ |
| 503 | void intel_csr_ucode_resume(struct drm_i915_private *dev_priv) |
| 504 | { |
| 505 | if (!HAS_CSR(dev_priv)) |
| 506 | return; |
| 507 | |
| 508 | /* |
| 509 | * Reacquire the reference to keep RPM disabled in case DMC isn't |
| 510 | * loaded. |
| 511 | */ |
| 512 | if (!dev_priv->csr.dmc_payload) |
| 513 | intel_display_power_get(dev_priv, POWER_DOMAIN_INIT); |
| 514 | } |
| 515 | |
| 516 | /** |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 517 | * intel_csr_ucode_fini() - unload the CSR firmware. |
Daniel Vetter | f444837 | 2015-10-28 23:59:02 +0200 | [diff] [blame] | 518 | * @dev_priv: i915 drm device. |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 519 | * |
Imre Deak | f74ed08 | 2016-04-18 14:48:21 +0300 | [diff] [blame] | 520 | * Firmmware unloading includes freeing the internal memory and reset the |
Animesh Manna | aa9145c | 2015-05-13 22:13:29 +0530 | [diff] [blame] | 521 | * firmware loading status. |
| 522 | */ |
Daniel Vetter | f444837 | 2015-10-28 23:59:02 +0200 | [diff] [blame] | 523 | void intel_csr_ucode_fini(struct drm_i915_private *dev_priv) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 524 | { |
Daniel Vetter | f444837 | 2015-10-28 23:59:02 +0200 | [diff] [blame] | 525 | if (!HAS_CSR(dev_priv)) |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 526 | return; |
| 527 | |
Imre Deak | f74ed08 | 2016-04-18 14:48:21 +0300 | [diff] [blame] | 528 | intel_csr_ucode_suspend(dev_priv); |
Animesh Manna | 15e72c1 | 2015-10-28 23:59:05 +0200 | [diff] [blame] | 529 | |
Daniel Vetter | eb80562 | 2015-05-04 14:58:44 +0200 | [diff] [blame] | 530 | kfree(dev_priv->csr.dmc_payload); |
| 531 | } |