Ard Biesheuvel | 4febfb8 | 2019-02-02 10:41:15 +0100 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0 |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 2 | /* |
| 3 | * EFI capsule support. |
| 4 | * |
| 5 | * Copyright 2013 Intel Corporation; author Matt Fleming |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 6 | */ |
| 7 | |
| 8 | #define pr_fmt(fmt) "efi: " fmt |
| 9 | |
| 10 | #include <linux/slab.h> |
| 11 | #include <linux/mutex.h> |
| 12 | #include <linux/highmem.h> |
| 13 | #include <linux/efi.h> |
| 14 | #include <linux/vmalloc.h> |
Ard Biesheuvel | 4dbe44f | 2020-12-07 18:40:53 +0100 | [diff] [blame^] | 15 | #include <asm/efi.h> |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 16 | #include <asm/io.h> |
| 17 | |
| 18 | typedef struct { |
| 19 | u64 length; |
| 20 | u64 data; |
| 21 | } efi_capsule_block_desc_t; |
| 22 | |
| 23 | static bool capsule_pending; |
Matt Fleming | 62075e5 | 2016-05-06 22:39:27 +0100 | [diff] [blame] | 24 | static bool stop_capsules; |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 25 | static int efi_reset_type = -1; |
| 26 | |
| 27 | /* |
| 28 | * capsule_mutex serialises access to both capsule_pending and |
Matt Fleming | 62075e5 | 2016-05-06 22:39:27 +0100 | [diff] [blame] | 29 | * efi_reset_type and stop_capsules. |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 30 | */ |
| 31 | static DEFINE_MUTEX(capsule_mutex); |
| 32 | |
| 33 | /** |
| 34 | * efi_capsule_pending - has a capsule been passed to the firmware? |
| 35 | * @reset_type: store the type of EFI reset if capsule is pending |
| 36 | * |
| 37 | * To ensure that the registered capsule is processed correctly by the |
| 38 | * firmware we need to perform a specific type of reset. If a capsule is |
| 39 | * pending return the reset type in @reset_type. |
| 40 | * |
| 41 | * This function will race with callers of efi_capsule_update(), for |
| 42 | * example, calling this function while somebody else is in |
| 43 | * efi_capsule_update() but hasn't reached efi_capsue_update_locked() |
| 44 | * will miss the updates to capsule_pending and efi_reset_type after |
| 45 | * efi_capsule_update_locked() completes. |
| 46 | * |
| 47 | * A non-racy use is from platform reboot code because we use |
| 48 | * system_state to ensure no capsules can be sent to the firmware once |
| 49 | * we're at SYSTEM_RESTART. See efi_capsule_update_locked(). |
| 50 | */ |
| 51 | bool efi_capsule_pending(int *reset_type) |
| 52 | { |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 53 | if (!capsule_pending) |
Matt Fleming | 62075e5 | 2016-05-06 22:39:27 +0100 | [diff] [blame] | 54 | return false; |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 55 | |
| 56 | if (reset_type) |
| 57 | *reset_type = efi_reset_type; |
Matt Fleming | 62075e5 | 2016-05-06 22:39:27 +0100 | [diff] [blame] | 58 | |
| 59 | return true; |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 60 | } |
| 61 | |
| 62 | /* |
| 63 | * Whitelist of EFI capsule flags that we support. |
| 64 | * |
| 65 | * We do not handle EFI_CAPSULE_INITIATE_RESET because that would |
| 66 | * require us to prepare the kernel for reboot. Refuse to load any |
| 67 | * capsules with that flag and any other flags that we do not know how |
| 68 | * to handle. |
| 69 | */ |
| 70 | #define EFI_CAPSULE_SUPPORTED_FLAG_MASK \ |
| 71 | (EFI_CAPSULE_PERSIST_ACROSS_RESET | EFI_CAPSULE_POPULATE_SYSTEM_TABLE) |
| 72 | |
| 73 | /** |
| 74 | * efi_capsule_supported - does the firmware support the capsule? |
| 75 | * @guid: vendor guid of capsule |
| 76 | * @flags: capsule flags |
| 77 | * @size: size of capsule data |
| 78 | * @reset: the reset type required for this capsule |
| 79 | * |
| 80 | * Check whether a capsule with @flags is supported by the firmware |
| 81 | * and that @size doesn't exceed the maximum size for a capsule. |
| 82 | * |
| 83 | * No attempt is made to check @reset against the reset type required |
| 84 | * by any pending capsules because of the races involved. |
| 85 | */ |
| 86 | int efi_capsule_supported(efi_guid_t guid, u32 flags, size_t size, int *reset) |
| 87 | { |
Matt Fleming | fb7a84c | 2016-05-06 22:39:29 +0100 | [diff] [blame] | 88 | efi_capsule_header_t capsule; |
| 89 | efi_capsule_header_t *cap_list[] = { &capsule }; |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 90 | efi_status_t status; |
| 91 | u64 max_size; |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 92 | |
| 93 | if (flags & ~EFI_CAPSULE_SUPPORTED_FLAG_MASK) |
| 94 | return -EINVAL; |
| 95 | |
Matt Fleming | fb7a84c | 2016-05-06 22:39:29 +0100 | [diff] [blame] | 96 | capsule.headersize = capsule.imagesize = sizeof(capsule); |
| 97 | memcpy(&capsule.guid, &guid, sizeof(efi_guid_t)); |
| 98 | capsule.flags = flags; |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 99 | |
Matt Fleming | fb7a84c | 2016-05-06 22:39:29 +0100 | [diff] [blame] | 100 | status = efi.query_capsule_caps(cap_list, 1, &max_size, reset); |
| 101 | if (status != EFI_SUCCESS) |
| 102 | return efi_status_to_err(status); |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 103 | |
| 104 | if (size > max_size) |
Matt Fleming | fb7a84c | 2016-05-06 22:39:29 +0100 | [diff] [blame] | 105 | return -ENOSPC; |
| 106 | |
| 107 | return 0; |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 108 | } |
| 109 | EXPORT_SYMBOL_GPL(efi_capsule_supported); |
| 110 | |
| 111 | /* |
| 112 | * Every scatter gather list (block descriptor) page must end with a |
| 113 | * continuation pointer. The last continuation pointer of the last |
| 114 | * page must be zero to mark the end of the chain. |
| 115 | */ |
| 116 | #define SGLIST_PER_PAGE ((PAGE_SIZE / sizeof(efi_capsule_block_desc_t)) - 1) |
| 117 | |
| 118 | /* |
| 119 | * How many scatter gather list (block descriptor) pages do we need |
| 120 | * to map @count pages? |
| 121 | */ |
| 122 | static inline unsigned int sg_pages_num(unsigned int count) |
| 123 | { |
| 124 | return DIV_ROUND_UP(count, SGLIST_PER_PAGE); |
| 125 | } |
| 126 | |
| 127 | /** |
| 128 | * efi_capsule_update_locked - pass a single capsule to the firmware |
| 129 | * @capsule: capsule to send to the firmware |
| 130 | * @sg_pages: array of scatter gather (block descriptor) pages |
| 131 | * @reset: the reset type required for @capsule |
| 132 | * |
| 133 | * Since this function must be called under capsule_mutex check |
| 134 | * whether efi_reset_type will conflict with @reset, and atomically |
| 135 | * set it and capsule_pending if a capsule was successfully sent to |
| 136 | * the firmware. |
| 137 | * |
| 138 | * We also check to see if the system is about to restart, and if so, |
| 139 | * abort. This avoids races between efi_capsule_update() and |
| 140 | * efi_capsule_pending(). |
| 141 | */ |
| 142 | static int |
| 143 | efi_capsule_update_locked(efi_capsule_header_t *capsule, |
| 144 | struct page **sg_pages, int reset) |
| 145 | { |
| 146 | efi_physical_addr_t sglist_phys; |
| 147 | efi_status_t status; |
| 148 | |
| 149 | lockdep_assert_held(&capsule_mutex); |
| 150 | |
| 151 | /* |
| 152 | * If someone has already registered a capsule that requires a |
| 153 | * different reset type, we're out of luck and must abort. |
| 154 | */ |
| 155 | if (efi_reset_type >= 0 && efi_reset_type != reset) { |
| 156 | pr_err("Conflicting capsule reset type %d (%d).\n", |
| 157 | reset, efi_reset_type); |
| 158 | return -EINVAL; |
| 159 | } |
| 160 | |
| 161 | /* |
| 162 | * If the system is getting ready to restart it may have |
| 163 | * called efi_capsule_pending() to make decisions (such as |
| 164 | * whether to force an EFI reboot), and we're racing against |
| 165 | * that call. Abort in that case. |
| 166 | */ |
Matt Fleming | 62075e5 | 2016-05-06 22:39:27 +0100 | [diff] [blame] | 167 | if (unlikely(stop_capsules)) { |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 168 | pr_warn("Capsule update raced with reboot, aborting.\n"); |
| 169 | return -EINVAL; |
| 170 | } |
| 171 | |
| 172 | sglist_phys = page_to_phys(sg_pages[0]); |
| 173 | |
| 174 | status = efi.update_capsule(&capsule, 1, sglist_phys); |
| 175 | if (status == EFI_SUCCESS) { |
| 176 | capsule_pending = true; |
| 177 | efi_reset_type = reset; |
| 178 | } |
| 179 | |
| 180 | return efi_status_to_err(status); |
| 181 | } |
| 182 | |
| 183 | /** |
| 184 | * efi_capsule_update - send a capsule to the firmware |
| 185 | * @capsule: capsule to send to firmware |
| 186 | * @pages: an array of capsule data pages |
| 187 | * |
| 188 | * Build a scatter gather list with EFI capsule block descriptors to |
| 189 | * map the capsule described by @capsule with its data in @pages and |
| 190 | * send it to the firmware via the UpdateCapsule() runtime service. |
| 191 | * |
Austin Christ | 6862e6a | 2016-08-11 11:42:00 +0100 | [diff] [blame] | 192 | * @capsule must be a virtual mapping of the complete capsule update in the |
| 193 | * kernel address space, as the capsule can be consumed immediately. |
| 194 | * A capsule_header_t that describes the entire contents of the capsule |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 195 | * must be at the start of the first data page. |
| 196 | * |
| 197 | * Even though this function will validate that the firmware supports |
| 198 | * the capsule guid, users will likely want to check that |
| 199 | * efi_capsule_supported() returns true before calling this function |
| 200 | * because it makes it easier to print helpful error messages. |
| 201 | * |
| 202 | * If the capsule is successfully submitted to the firmware, any |
| 203 | * subsequent calls to efi_capsule_pending() will return true. @pages |
| 204 | * must not be released or modified if this function returns |
| 205 | * successfully. |
| 206 | * |
| 207 | * Callers must be prepared for this function to fail, which can |
| 208 | * happen if we raced with system reboot or if there is already a |
| 209 | * pending capsule that has a reset type that conflicts with the one |
| 210 | * required by @capsule. Do NOT use efi_capsule_pending() to detect |
| 211 | * this conflict since that would be racy. Instead, submit the capsule |
| 212 | * to efi_capsule_update() and check the return value. |
| 213 | * |
| 214 | * Return 0 on success, a converted EFI status code on failure. |
| 215 | */ |
Ard Biesheuvel | 2a457fb | 2017-06-02 13:52:03 +0000 | [diff] [blame] | 216 | int efi_capsule_update(efi_capsule_header_t *capsule, phys_addr_t *pages) |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 217 | { |
| 218 | u32 imagesize = capsule->imagesize; |
| 219 | efi_guid_t guid = capsule->guid; |
| 220 | unsigned int count, sg_count; |
| 221 | u32 flags = capsule->flags; |
| 222 | struct page **sg_pages; |
| 223 | int rv, reset_type; |
| 224 | int i, j; |
| 225 | |
| 226 | rv = efi_capsule_supported(guid, flags, imagesize, &reset_type); |
| 227 | if (rv) |
| 228 | return rv; |
| 229 | |
| 230 | count = DIV_ROUND_UP(imagesize, PAGE_SIZE); |
| 231 | sg_count = sg_pages_num(count); |
| 232 | |
Kees Cook | 6396bb2 | 2018-06-12 14:03:40 -0700 | [diff] [blame] | 233 | sg_pages = kcalloc(sg_count, sizeof(*sg_pages), GFP_KERNEL); |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 234 | if (!sg_pages) |
| 235 | return -ENOMEM; |
| 236 | |
| 237 | for (i = 0; i < sg_count; i++) { |
| 238 | sg_pages[i] = alloc_page(GFP_KERNEL); |
| 239 | if (!sg_pages[i]) { |
| 240 | rv = -ENOMEM; |
| 241 | goto out; |
| 242 | } |
| 243 | } |
| 244 | |
| 245 | for (i = 0; i < sg_count; i++) { |
| 246 | efi_capsule_block_desc_t *sglist; |
| 247 | |
Ard Biesheuvel | 91c1c09 | 2020-12-07 17:33:33 +0100 | [diff] [blame] | 248 | sglist = kmap_atomic(sg_pages[i]); |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 249 | |
| 250 | for (j = 0; j < SGLIST_PER_PAGE && count > 0; j++) { |
Ard Biesheuvel | 2a457fb | 2017-06-02 13:52:03 +0000 | [diff] [blame] | 251 | u64 sz = min_t(u64, imagesize, |
| 252 | PAGE_SIZE - (u64)*pages % PAGE_SIZE); |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 253 | |
| 254 | sglist[j].length = sz; |
Ard Biesheuvel | 2a457fb | 2017-06-02 13:52:03 +0000 | [diff] [blame] | 255 | sglist[j].data = *pages++; |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 256 | |
| 257 | imagesize -= sz; |
| 258 | count--; |
| 259 | } |
| 260 | |
| 261 | /* Continuation pointer */ |
| 262 | sglist[j].length = 0; |
| 263 | |
| 264 | if (i + 1 == sg_count) |
| 265 | sglist[j].data = 0; |
| 266 | else |
| 267 | sglist[j].data = page_to_phys(sg_pages[i + 1]); |
| 268 | |
Ard Biesheuvel | 4dbe44f | 2020-12-07 18:40:53 +0100 | [diff] [blame^] | 269 | #if defined(CONFIG_ARM) || defined(CONFIG_ARM64) |
| 270 | /* |
| 271 | * At runtime, the firmware has no way to find out where the |
| 272 | * sglist elements are mapped, if they are mapped in the first |
| 273 | * place. Therefore, on architectures that can only perform |
| 274 | * cache maintenance by virtual address, the firmware is unable |
| 275 | * to perform this maintenance, and so it is up to the OS to do |
| 276 | * it instead. |
| 277 | */ |
| 278 | efi_capsule_flush_cache_range(sglist, PAGE_SIZE); |
| 279 | #endif |
Ard Biesheuvel | 91c1c09 | 2020-12-07 17:33:33 +0100 | [diff] [blame] | 280 | kunmap_atomic(sglist); |
Matt Fleming | f0133f3 | 2016-04-25 21:06:59 +0100 | [diff] [blame] | 281 | } |
| 282 | |
| 283 | mutex_lock(&capsule_mutex); |
| 284 | rv = efi_capsule_update_locked(capsule, sg_pages, reset_type); |
| 285 | mutex_unlock(&capsule_mutex); |
| 286 | |
| 287 | out: |
| 288 | for (i = 0; rv && i < sg_count; i++) { |
| 289 | if (sg_pages[i]) |
| 290 | __free_page(sg_pages[i]); |
| 291 | } |
| 292 | |
| 293 | kfree(sg_pages); |
| 294 | return rv; |
| 295 | } |
| 296 | EXPORT_SYMBOL_GPL(efi_capsule_update); |
Matt Fleming | 62075e5 | 2016-05-06 22:39:27 +0100 | [diff] [blame] | 297 | |
| 298 | static int capsule_reboot_notify(struct notifier_block *nb, unsigned long event, void *cmd) |
| 299 | { |
| 300 | mutex_lock(&capsule_mutex); |
| 301 | stop_capsules = true; |
| 302 | mutex_unlock(&capsule_mutex); |
| 303 | |
| 304 | return NOTIFY_DONE; |
| 305 | } |
| 306 | |
| 307 | static struct notifier_block capsule_reboot_nb = { |
| 308 | .notifier_call = capsule_reboot_notify, |
| 309 | }; |
| 310 | |
| 311 | static int __init capsule_reboot_register(void) |
| 312 | { |
| 313 | return register_reboot_notifier(&capsule_reboot_nb); |
| 314 | } |
| 315 | core_initcall(capsule_reboot_register); |