blob: 7c3bd2c3cdca9c08fd77f1e8ad48e99080e4f178 [file] [log] [blame]
Thomas Gleixnercaab2772019-06-03 07:44:50 +02001// SPDX-License-Identifier: GPL-2.0-only
Robin Murphye5fc9752016-01-26 17:13:13 +00002/*
3 * CPU-agnostic ARM page table allocator.
4 *
5 * ARMv7 Short-descriptor format, supporting
6 * - Basic memory attributes
7 * - Simplified access permissions (AP[2:1] model)
8 * - Backwards-compatible TEX remap
9 * - Large pages/supersections (if indicated by the caller)
10 *
11 * Not supporting:
12 * - Legacy access permissions (AP[2:0] model)
13 *
14 * Almost certainly never supporting:
15 * - PXN
16 * - Domains
17 *
Robin Murphye5fc9752016-01-26 17:13:13 +000018 * Copyright (C) 2014-2015 ARM Limited
19 * Copyright (c) 2014-2015 MediaTek Inc.
20 */
21
22#define pr_fmt(fmt) "arm-v7s io-pgtable: " fmt
23
Robin Murphy119ff302017-06-22 16:53:55 +010024#include <linux/atomic.h>
Robin Murphye5fc9752016-01-26 17:13:13 +000025#include <linux/dma-mapping.h>
26#include <linux/gfp.h>
Rob Herringb77cf112019-02-05 10:37:31 -060027#include <linux/io-pgtable.h>
Robin Murphye5fc9752016-01-26 17:13:13 +000028#include <linux/iommu.h>
29#include <linux/kernel.h>
30#include <linux/kmemleak.h>
31#include <linux/sizes.h>
32#include <linux/slab.h>
Robin Murphy119ff302017-06-22 16:53:55 +010033#include <linux/spinlock.h>
Robin Murphye5fc9752016-01-26 17:13:13 +000034#include <linux/types.h>
35
36#include <asm/barrier.h>
37
Robin Murphye5fc9752016-01-26 17:13:13 +000038/* Struct accessors */
39#define io_pgtable_to_data(x) \
40 container_of((x), struct arm_v7s_io_pgtable, iop)
41
42#define io_pgtable_ops_to_data(x) \
43 io_pgtable_to_data(io_pgtable_ops_to_pgtable(x))
44
45/*
46 * We have 32 bits total; 12 bits resolved at level 1, 8 bits at level 2,
47 * and 12 bits in a page. With some carefully-chosen coefficients we can
48 * hide the ugly inconsistencies behind these macros and at least let the
49 * rest of the code pretend to be somewhat sane.
50 */
51#define ARM_V7S_ADDR_BITS 32
52#define _ARM_V7S_LVL_BITS(lvl) (16 - (lvl) * 4)
53#define ARM_V7S_LVL_SHIFT(lvl) (ARM_V7S_ADDR_BITS - (4 + 8 * (lvl)))
54#define ARM_V7S_TABLE_SHIFT 10
55
56#define ARM_V7S_PTES_PER_LVL(lvl) (1 << _ARM_V7S_LVL_BITS(lvl))
57#define ARM_V7S_TABLE_SIZE(lvl) \
58 (ARM_V7S_PTES_PER_LVL(lvl) * sizeof(arm_v7s_iopte))
59
60#define ARM_V7S_BLOCK_SIZE(lvl) (1UL << ARM_V7S_LVL_SHIFT(lvl))
61#define ARM_V7S_LVL_MASK(lvl) ((u32)(~0U << ARM_V7S_LVL_SHIFT(lvl)))
62#define ARM_V7S_TABLE_MASK ((u32)(~0U << ARM_V7S_TABLE_SHIFT))
63#define _ARM_V7S_IDX_MASK(lvl) (ARM_V7S_PTES_PER_LVL(lvl) - 1)
64#define ARM_V7S_LVL_IDX(addr, lvl) ({ \
65 int _l = lvl; \
66 ((u32)(addr) >> ARM_V7S_LVL_SHIFT(_l)) & _ARM_V7S_IDX_MASK(_l); \
67})
68
69/*
70 * Large page/supersection entries are effectively a block of 16 page/section
71 * entries, along the lines of the LPAE contiguous hint, but all with the
72 * same output address. For want of a better common name we'll call them
73 * "contiguous" versions of their respective page/section entries here, but
74 * noting the distinction (WRT to TLB maintenance) that they represent *one*
75 * entry repeated 16 times, not 16 separate entries (as in the LPAE case).
76 */
77#define ARM_V7S_CONT_PAGES 16
78
79/* PTE type bits: these are all mixed up with XN/PXN bits in most cases */
80#define ARM_V7S_PTE_TYPE_TABLE 0x1
81#define ARM_V7S_PTE_TYPE_PAGE 0x2
82#define ARM_V7S_PTE_TYPE_CONT_PAGE 0x1
83
84#define ARM_V7S_PTE_IS_VALID(pte) (((pte) & 0x3) != 0)
Robin Murphy9db829d2017-06-22 16:53:50 +010085#define ARM_V7S_PTE_IS_TABLE(pte, lvl) \
86 ((lvl) == 1 && (((pte) & 0x3) == ARM_V7S_PTE_TYPE_TABLE))
Robin Murphye5fc9752016-01-26 17:13:13 +000087
88/* Page table bits */
89#define ARM_V7S_ATTR_XN(lvl) BIT(4 * (2 - (lvl)))
90#define ARM_V7S_ATTR_B BIT(2)
91#define ARM_V7S_ATTR_C BIT(3)
92#define ARM_V7S_ATTR_NS_TABLE BIT(3)
93#define ARM_V7S_ATTR_NS_SECTION BIT(19)
94
95#define ARM_V7S_CONT_SECTION BIT(18)
96#define ARM_V7S_CONT_PAGE_XN_SHIFT 15
97
98/*
99 * The attribute bits are consistently ordered*, but occupy bits [17:10] of
100 * a level 1 PTE vs. bits [11:4] at level 2. Thus we define the individual
101 * fields relative to that 8-bit block, plus a total shift relative to the PTE.
102 */
103#define ARM_V7S_ATTR_SHIFT(lvl) (16 - (lvl) * 6)
104
105#define ARM_V7S_ATTR_MASK 0xff
106#define ARM_V7S_ATTR_AP0 BIT(0)
107#define ARM_V7S_ATTR_AP1 BIT(1)
108#define ARM_V7S_ATTR_AP2 BIT(5)
109#define ARM_V7S_ATTR_S BIT(6)
110#define ARM_V7S_ATTR_NG BIT(7)
111#define ARM_V7S_TEX_SHIFT 2
112#define ARM_V7S_TEX_MASK 0x7
113#define ARM_V7S_ATTR_TEX(val) (((val) & ARM_V7S_TEX_MASK) << ARM_V7S_TEX_SHIFT)
114
Yong Wu4c019de2019-08-24 11:01:54 +0800115/* MediaTek extend the two bits for PA 32bit/33bit */
116#define ARM_V7S_ATTR_MTK_PA_BIT32 BIT(9)
117#define ARM_V7S_ATTR_MTK_PA_BIT33 BIT(4)
Yong Wu1afe2312016-03-14 06:01:10 +0800118
Robin Murphye5fc9752016-01-26 17:13:13 +0000119/* *well, except for TEX on level 2 large pages, of course :( */
120#define ARM_V7S_CONT_PAGE_TEX_SHIFT 6
121#define ARM_V7S_CONT_PAGE_TEX_MASK (ARM_V7S_TEX_MASK << ARM_V7S_CONT_PAGE_TEX_SHIFT)
122
123/* Simplified access permissions */
124#define ARM_V7S_PTE_AF ARM_V7S_ATTR_AP0
125#define ARM_V7S_PTE_AP_UNPRIV ARM_V7S_ATTR_AP1
126#define ARM_V7S_PTE_AP_RDONLY ARM_V7S_ATTR_AP2
127
128/* Register bits */
129#define ARM_V7S_RGN_NC 0
130#define ARM_V7S_RGN_WBWA 1
131#define ARM_V7S_RGN_WT 2
132#define ARM_V7S_RGN_WB 3
133
134#define ARM_V7S_PRRR_TYPE_DEVICE 1
135#define ARM_V7S_PRRR_TYPE_NORMAL 2
136#define ARM_V7S_PRRR_TR(n, type) (((type) & 0x3) << ((n) * 2))
137#define ARM_V7S_PRRR_DS0 BIT(16)
138#define ARM_V7S_PRRR_DS1 BIT(17)
139#define ARM_V7S_PRRR_NS0 BIT(18)
140#define ARM_V7S_PRRR_NS1 BIT(19)
141#define ARM_V7S_PRRR_NOS(n) BIT((n) + 24)
142
143#define ARM_V7S_NMRR_IR(n, attr) (((attr) & 0x3) << ((n) * 2))
144#define ARM_V7S_NMRR_OR(n, attr) (((attr) & 0x3) << ((n) * 2 + 16))
145
146#define ARM_V7S_TTBR_S BIT(1)
147#define ARM_V7S_TTBR_NOS BIT(5)
148#define ARM_V7S_TTBR_ORGN_ATTR(attr) (((attr) & 0x3) << 3)
149#define ARM_V7S_TTBR_IRGN_ATTR(attr) \
150 ((((attr) & 0x1) << 6) | (((attr) & 0x2) >> 1))
151
152#define ARM_V7S_TCR_PD1 BIT(5)
153
Nicolas Boichat0a352552019-03-28 20:43:46 -0700154#ifdef CONFIG_ZONE_DMA32
155#define ARM_V7S_TABLE_GFP_DMA GFP_DMA32
156#define ARM_V7S_TABLE_SLAB_FLAGS SLAB_CACHE_DMA32
157#else
158#define ARM_V7S_TABLE_GFP_DMA GFP_DMA
159#define ARM_V7S_TABLE_SLAB_FLAGS SLAB_CACHE_DMA
160#endif
161
Robin Murphye5fc9752016-01-26 17:13:13 +0000162typedef u32 arm_v7s_iopte;
163
164static bool selftest_running;
165
166struct arm_v7s_io_pgtable {
167 struct io_pgtable iop;
168
169 arm_v7s_iopte *pgd;
170 struct kmem_cache *l2_tables;
Robin Murphy119ff302017-06-22 16:53:55 +0100171 spinlock_t split_lock;
Robin Murphye5fc9752016-01-26 17:13:13 +0000172};
173
Yong Wu5950b952019-08-24 11:01:51 +0800174static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl);
175
Robin Murphye5fc9752016-01-26 17:13:13 +0000176static dma_addr_t __arm_v7s_dma_addr(void *pages)
177{
178 return (dma_addr_t)virt_to_phys(pages);
179}
180
Yong Wu4c019de2019-08-24 11:01:54 +0800181static bool arm_v7s_is_mtk_enabled(struct io_pgtable_cfg *cfg)
182{
183 return IS_ENABLED(CONFIG_PHYS_ADDR_T_64BIT) &&
184 (cfg->quirks & IO_PGTABLE_QUIRK_ARM_MTK_EXT);
185}
186
Yong Wu5950b952019-08-24 11:01:51 +0800187static arm_v7s_iopte paddr_to_iopte(phys_addr_t paddr, int lvl,
188 struct io_pgtable_cfg *cfg)
Robin Murphye5fc9752016-01-26 17:13:13 +0000189{
Yong Wu4c019de2019-08-24 11:01:54 +0800190 arm_v7s_iopte pte = paddr & ARM_V7S_LVL_MASK(lvl);
191
192 if (!arm_v7s_is_mtk_enabled(cfg))
193 return pte;
194
195 if (paddr & BIT_ULL(32))
196 pte |= ARM_V7S_ATTR_MTK_PA_BIT32;
197 if (paddr & BIT_ULL(33))
198 pte |= ARM_V7S_ATTR_MTK_PA_BIT33;
199 return pte;
Yong Wu5950b952019-08-24 11:01:51 +0800200}
201
202static phys_addr_t iopte_to_paddr(arm_v7s_iopte pte, int lvl,
203 struct io_pgtable_cfg *cfg)
204{
205 arm_v7s_iopte mask;
Yong Wu4c019de2019-08-24 11:01:54 +0800206 phys_addr_t paddr;
Yong Wu5950b952019-08-24 11:01:51 +0800207
Robin Murphye5fc9752016-01-26 17:13:13 +0000208 if (ARM_V7S_PTE_IS_TABLE(pte, lvl))
Yong Wu5950b952019-08-24 11:01:51 +0800209 mask = ARM_V7S_TABLE_MASK;
210 else if (arm_v7s_pte_is_cont(pte, lvl))
211 mask = ARM_V7S_LVL_MASK(lvl) * ARM_V7S_CONT_PAGES;
Robin Murphye5fc9752016-01-26 17:13:13 +0000212 else
Yong Wu5950b952019-08-24 11:01:51 +0800213 mask = ARM_V7S_LVL_MASK(lvl);
214
Yong Wu4c019de2019-08-24 11:01:54 +0800215 paddr = pte & mask;
216 if (!arm_v7s_is_mtk_enabled(cfg))
217 return paddr;
218
219 if (pte & ARM_V7S_ATTR_MTK_PA_BIT32)
220 paddr |= BIT_ULL(32);
221 if (pte & ARM_V7S_ATTR_MTK_PA_BIT33)
222 paddr |= BIT_ULL(33);
223 return paddr;
Yong Wu5950b952019-08-24 11:01:51 +0800224}
225
226static arm_v7s_iopte *iopte_deref(arm_v7s_iopte pte, int lvl,
227 struct arm_v7s_io_pgtable *data)
228{
229 return phys_to_virt(iopte_to_paddr(pte, lvl, &data->iop.cfg));
Robin Murphye5fc9752016-01-26 17:13:13 +0000230}
231
232static void *__arm_v7s_alloc_table(int lvl, gfp_t gfp,
233 struct arm_v7s_io_pgtable *data)
234{
Robin Murphy81b3c252017-06-22 16:53:53 +0100235 struct io_pgtable_cfg *cfg = &data->iop.cfg;
236 struct device *dev = cfg->iommu_dev;
Jean-Philippe Brucker29859ae2018-06-19 13:52:24 +0100237 phys_addr_t phys;
Robin Murphye5fc9752016-01-26 17:13:13 +0000238 dma_addr_t dma;
239 size_t size = ARM_V7S_TABLE_SIZE(lvl);
240 void *table = NULL;
241
242 if (lvl == 1)
Nicolas Boichat0a352552019-03-28 20:43:46 -0700243 table = (void *)__get_free_pages(
244 __GFP_ZERO | ARM_V7S_TABLE_GFP_DMA, get_order(size));
Robin Murphye5fc9752016-01-26 17:13:13 +0000245 else if (lvl == 2)
Nicolas Boichat0a352552019-03-28 20:43:46 -0700246 table = kmem_cache_zalloc(data->l2_tables, gfp);
Jean-Philippe Brucker29859ae2018-06-19 13:52:24 +0100247 phys = virt_to_phys(table);
Nicolas Boichat0a352552019-03-28 20:43:46 -0700248 if (phys != (arm_v7s_iopte)phys) {
Jean-Philippe Brucker29859ae2018-06-19 13:52:24 +0100249 /* Doesn't fit in PTE */
Nicolas Boichat0a352552019-03-28 20:43:46 -0700250 dev_err(dev, "Page table does not fit in PTE: %pa", &phys);
Jean-Philippe Brucker29859ae2018-06-19 13:52:24 +0100251 goto out_free;
Nicolas Boichat0a352552019-03-28 20:43:46 -0700252 }
Will Deacon4f418452019-06-25 12:51:25 +0100253 if (table && !cfg->coherent_walk) {
Robin Murphye5fc9752016-01-26 17:13:13 +0000254 dma = dma_map_single(dev, table, size, DMA_TO_DEVICE);
255 if (dma_mapping_error(dev, dma))
256 goto out_free;
257 /*
258 * We depend on the IOMMU being able to work with any physical
259 * address directly, so if the DMA layer suggests otherwise by
260 * translating or truncating them, that bodes very badly...
261 */
Jean-Philippe Brucker29859ae2018-06-19 13:52:24 +0100262 if (dma != phys)
Robin Murphye5fc9752016-01-26 17:13:13 +0000263 goto out_unmap;
264 }
Nicolas Boichat032ebd82019-01-28 17:43:01 +0800265 if (lvl == 2)
266 kmemleak_ignore(table);
Robin Murphye5fc9752016-01-26 17:13:13 +0000267 return table;
268
269out_unmap:
270 dev_err(dev, "Cannot accommodate DMA translation for IOMMU page tables\n");
271 dma_unmap_single(dev, dma, size, DMA_TO_DEVICE);
272out_free:
273 if (lvl == 1)
274 free_pages((unsigned long)table, get_order(size));
275 else
276 kmem_cache_free(data->l2_tables, table);
277 return NULL;
278}
279
280static void __arm_v7s_free_table(void *table, int lvl,
281 struct arm_v7s_io_pgtable *data)
282{
Robin Murphy81b3c252017-06-22 16:53:53 +0100283 struct io_pgtable_cfg *cfg = &data->iop.cfg;
284 struct device *dev = cfg->iommu_dev;
Robin Murphye5fc9752016-01-26 17:13:13 +0000285 size_t size = ARM_V7S_TABLE_SIZE(lvl);
286
Will Deacon4f418452019-06-25 12:51:25 +0100287 if (!cfg->coherent_walk)
Robin Murphye5fc9752016-01-26 17:13:13 +0000288 dma_unmap_single(dev, __arm_v7s_dma_addr(table), size,
289 DMA_TO_DEVICE);
290 if (lvl == 1)
291 free_pages((unsigned long)table, get_order(size));
292 else
293 kmem_cache_free(data->l2_tables, table);
294}
295
296static void __arm_v7s_pte_sync(arm_v7s_iopte *ptep, int num_entries,
297 struct io_pgtable_cfg *cfg)
298{
Will Deacon4f418452019-06-25 12:51:25 +0100299 if (cfg->coherent_walk)
Robin Murphye5fc9752016-01-26 17:13:13 +0000300 return;
301
302 dma_sync_single_for_device(cfg->iommu_dev, __arm_v7s_dma_addr(ptep),
303 num_entries * sizeof(*ptep), DMA_TO_DEVICE);
304}
305static void __arm_v7s_set_pte(arm_v7s_iopte *ptep, arm_v7s_iopte pte,
306 int num_entries, struct io_pgtable_cfg *cfg)
307{
308 int i;
309
310 for (i = 0; i < num_entries; i++)
311 ptep[i] = pte;
312
313 __arm_v7s_pte_sync(ptep, num_entries, cfg);
314}
315
316static arm_v7s_iopte arm_v7s_prot_to_pte(int prot, int lvl,
317 struct io_pgtable_cfg *cfg)
318{
319 bool ap = !(cfg->quirks & IO_PGTABLE_QUIRK_NO_PERMS);
Robin Murphye88ccab2016-04-05 12:39:32 +0100320 arm_v7s_iopte pte = ARM_V7S_ATTR_NG | ARM_V7S_ATTR_S;
Robin Murphye5fc9752016-01-26 17:13:13 +0000321
Robin Murphye88ccab2016-04-05 12:39:32 +0100322 if (!(prot & IOMMU_MMIO))
323 pte |= ARM_V7S_ATTR_TEX(1);
Robin Murphye5fc9752016-01-26 17:13:13 +0000324 if (ap) {
Robin Murphy5baf1e92017-01-06 18:58:10 +0530325 pte |= ARM_V7S_PTE_AF;
326 if (!(prot & IOMMU_PRIV))
327 pte |= ARM_V7S_PTE_AP_UNPRIV;
Robin Murphye5fc9752016-01-26 17:13:13 +0000328 if (!(prot & IOMMU_WRITE))
329 pte |= ARM_V7S_PTE_AP_RDONLY;
330 }
331 pte <<= ARM_V7S_ATTR_SHIFT(lvl);
332
333 if ((prot & IOMMU_NOEXEC) && ap)
334 pte |= ARM_V7S_ATTR_XN(lvl);
Robin Murphye88ccab2016-04-05 12:39:32 +0100335 if (prot & IOMMU_MMIO)
336 pte |= ARM_V7S_ATTR_B;
337 else if (prot & IOMMU_CACHE)
Robin Murphye5fc9752016-01-26 17:13:13 +0000338 pte |= ARM_V7S_ATTR_B | ARM_V7S_ATTR_C;
339
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100340 pte |= ARM_V7S_PTE_TYPE_PAGE;
341 if (lvl == 1 && (cfg->quirks & IO_PGTABLE_QUIRK_ARM_NS))
342 pte |= ARM_V7S_ATTR_NS_SECTION;
343
Robin Murphye5fc9752016-01-26 17:13:13 +0000344 return pte;
345}
346
347static int arm_v7s_pte_to_prot(arm_v7s_iopte pte, int lvl)
348{
349 int prot = IOMMU_READ;
Robin Murphye88ccab2016-04-05 12:39:32 +0100350 arm_v7s_iopte attr = pte >> ARM_V7S_ATTR_SHIFT(lvl);
Robin Murphye5fc9752016-01-26 17:13:13 +0000351
Robin Murphye633fc72016-08-11 17:44:05 +0100352 if (!(attr & ARM_V7S_PTE_AP_RDONLY))
Robin Murphye5fc9752016-01-26 17:13:13 +0000353 prot |= IOMMU_WRITE;
Robin Murphy5baf1e92017-01-06 18:58:10 +0530354 if (!(attr & ARM_V7S_PTE_AP_UNPRIV))
355 prot |= IOMMU_PRIV;
Robin Murphye88ccab2016-04-05 12:39:32 +0100356 if ((attr & (ARM_V7S_TEX_MASK << ARM_V7S_TEX_SHIFT)) == 0)
357 prot |= IOMMU_MMIO;
358 else if (pte & ARM_V7S_ATTR_C)
Robin Murphye5fc9752016-01-26 17:13:13 +0000359 prot |= IOMMU_CACHE;
Robin Murphye633fc72016-08-11 17:44:05 +0100360 if (pte & ARM_V7S_ATTR_XN(lvl))
361 prot |= IOMMU_NOEXEC;
Robin Murphye5fc9752016-01-26 17:13:13 +0000362
363 return prot;
364}
365
366static arm_v7s_iopte arm_v7s_pte_to_cont(arm_v7s_iopte pte, int lvl)
367{
368 if (lvl == 1) {
369 pte |= ARM_V7S_CONT_SECTION;
370 } else if (lvl == 2) {
371 arm_v7s_iopte xn = pte & ARM_V7S_ATTR_XN(lvl);
372 arm_v7s_iopte tex = pte & ARM_V7S_CONT_PAGE_TEX_MASK;
373
374 pte ^= xn | tex | ARM_V7S_PTE_TYPE_PAGE;
375 pte |= (xn << ARM_V7S_CONT_PAGE_XN_SHIFT) |
376 (tex << ARM_V7S_CONT_PAGE_TEX_SHIFT) |
377 ARM_V7S_PTE_TYPE_CONT_PAGE;
378 }
379 return pte;
380}
381
382static arm_v7s_iopte arm_v7s_cont_to_pte(arm_v7s_iopte pte, int lvl)
383{
384 if (lvl == 1) {
385 pte &= ~ARM_V7S_CONT_SECTION;
386 } else if (lvl == 2) {
387 arm_v7s_iopte xn = pte & BIT(ARM_V7S_CONT_PAGE_XN_SHIFT);
388 arm_v7s_iopte tex = pte & (ARM_V7S_CONT_PAGE_TEX_MASK <<
389 ARM_V7S_CONT_PAGE_TEX_SHIFT);
390
391 pte ^= xn | tex | ARM_V7S_PTE_TYPE_CONT_PAGE;
392 pte |= (xn >> ARM_V7S_CONT_PAGE_XN_SHIFT) |
393 (tex >> ARM_V7S_CONT_PAGE_TEX_SHIFT) |
394 ARM_V7S_PTE_TYPE_PAGE;
395 }
396 return pte;
397}
398
399static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl)
400{
401 if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte, lvl))
402 return pte & ARM_V7S_CONT_SECTION;
403 else if (lvl == 2)
404 return !(pte & ARM_V7S_PTE_TYPE_PAGE);
405 return false;
406}
407
Will Deacon3951c412019-07-02 16:45:15 +0100408static size_t __arm_v7s_unmap(struct arm_v7s_io_pgtable *,
409 struct iommu_iotlb_gather *, unsigned long,
Vivek Gautam193e67c2018-02-05 23:29:19 +0530410 size_t, int, arm_v7s_iopte *);
Robin Murphye5fc9752016-01-26 17:13:13 +0000411
412static int arm_v7s_init_pte(struct arm_v7s_io_pgtable *data,
413 unsigned long iova, phys_addr_t paddr, int prot,
414 int lvl, int num_entries, arm_v7s_iopte *ptep)
415{
416 struct io_pgtable_cfg *cfg = &data->iop.cfg;
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100417 arm_v7s_iopte pte;
Robin Murphye5fc9752016-01-26 17:13:13 +0000418 int i;
419
420 for (i = 0; i < num_entries; i++)
421 if (ARM_V7S_PTE_IS_TABLE(ptep[i], lvl)) {
422 /*
423 * We need to unmap and free the old table before
424 * overwriting it with a block entry.
425 */
426 arm_v7s_iopte *tblp;
427 size_t sz = ARM_V7S_BLOCK_SIZE(lvl);
428
429 tblp = ptep - ARM_V7S_LVL_IDX(iova, lvl);
Will Deacon3951c412019-07-02 16:45:15 +0100430 if (WARN_ON(__arm_v7s_unmap(data, NULL, iova + i * sz,
Robin Murphye5fc9752016-01-26 17:13:13 +0000431 sz, lvl, tblp) != sz))
432 return -EINVAL;
433 } else if (ptep[i]) {
434 /* We require an unmap first */
435 WARN_ON(!selftest_running);
436 return -EEXIST;
437 }
438
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100439 pte = arm_v7s_prot_to_pte(prot, lvl, cfg);
Robin Murphye5fc9752016-01-26 17:13:13 +0000440 if (num_entries > 1)
441 pte = arm_v7s_pte_to_cont(pte, lvl);
442
Yong Wu5950b952019-08-24 11:01:51 +0800443 pte |= paddr_to_iopte(paddr, lvl, cfg);
Robin Murphye5fc9752016-01-26 17:13:13 +0000444
445 __arm_v7s_set_pte(ptep, pte, num_entries, cfg);
446 return 0;
447}
448
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100449static arm_v7s_iopte arm_v7s_install_table(arm_v7s_iopte *table,
450 arm_v7s_iopte *ptep,
Robin Murphy119ff302017-06-22 16:53:55 +0100451 arm_v7s_iopte curr,
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100452 struct io_pgtable_cfg *cfg)
453{
Robin Murphy119ff302017-06-22 16:53:55 +0100454 arm_v7s_iopte old, new;
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100455
456 new = virt_to_phys(table) | ARM_V7S_PTE_TYPE_TABLE;
457 if (cfg->quirks & IO_PGTABLE_QUIRK_ARM_NS)
458 new |= ARM_V7S_ATTR_NS_TABLE;
459
Will Deacon77f34452017-06-23 12:02:38 +0100460 /*
461 * Ensure the table itself is visible before its PTE can be.
462 * Whilst we could get away with cmpxchg64_release below, this
463 * doesn't have any ordering semantics when !CONFIG_SMP.
464 */
465 dma_wmb();
Robin Murphy119ff302017-06-22 16:53:55 +0100466
467 old = cmpxchg_relaxed(ptep, curr, new);
468 __arm_v7s_pte_sync(ptep, 1, cfg);
469
470 return old;
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100471}
472
Robin Murphye5fc9752016-01-26 17:13:13 +0000473static int __arm_v7s_map(struct arm_v7s_io_pgtable *data, unsigned long iova,
474 phys_addr_t paddr, size_t size, int prot,
475 int lvl, arm_v7s_iopte *ptep)
476{
477 struct io_pgtable_cfg *cfg = &data->iop.cfg;
478 arm_v7s_iopte pte, *cptep;
479 int num_entries = size >> ARM_V7S_LVL_SHIFT(lvl);
480
481 /* Find our entry at the current level */
482 ptep += ARM_V7S_LVL_IDX(iova, lvl);
483
484 /* If we can install a leaf entry at this level, then do so */
485 if (num_entries)
486 return arm_v7s_init_pte(data, iova, paddr, prot,
487 lvl, num_entries, ptep);
488
489 /* We can't allocate tables at the final level */
490 if (WARN_ON(lvl == 2))
491 return -EINVAL;
492
493 /* Grab a pointer to the next level */
Robin Murphy119ff302017-06-22 16:53:55 +0100494 pte = READ_ONCE(*ptep);
Robin Murphye5fc9752016-01-26 17:13:13 +0000495 if (!pte) {
496 cptep = __arm_v7s_alloc_table(lvl + 1, GFP_ATOMIC, data);
497 if (!cptep)
498 return -ENOMEM;
499
Robin Murphy119ff302017-06-22 16:53:55 +0100500 pte = arm_v7s_install_table(cptep, ptep, 0, cfg);
501 if (pte)
502 __arm_v7s_free_table(cptep, lvl + 1, data);
Oleksandr Tyshchenkoa03849e2017-02-27 14:30:26 +0200503 } else {
Robin Murphy119ff302017-06-22 16:53:55 +0100504 /* We've no easy way of knowing if it's synced yet, so... */
505 __arm_v7s_pte_sync(ptep, 1, cfg);
506 }
507
508 if (ARM_V7S_PTE_IS_TABLE(pte, lvl)) {
Yong Wu5950b952019-08-24 11:01:51 +0800509 cptep = iopte_deref(pte, lvl, data);
Robin Murphy119ff302017-06-22 16:53:55 +0100510 } else if (pte) {
Oleksandr Tyshchenkoa03849e2017-02-27 14:30:26 +0200511 /* We require an unmap first */
512 WARN_ON(!selftest_running);
513 return -EEXIST;
Robin Murphye5fc9752016-01-26 17:13:13 +0000514 }
515
516 /* Rinse, repeat */
517 return __arm_v7s_map(data, iova, paddr, size, prot, lvl + 1, cptep);
518}
519
520static int arm_v7s_map(struct io_pgtable_ops *ops, unsigned long iova,
521 phys_addr_t paddr, size_t size, int prot)
522{
523 struct arm_v7s_io_pgtable *data = io_pgtable_ops_to_data(ops);
Robin Murphy507e4c92016-01-26 17:13:14 +0000524 struct io_pgtable *iop = &data->iop;
Robin Murphye5fc9752016-01-26 17:13:13 +0000525 int ret;
526
527 /* If no access, then nothing to do */
528 if (!(prot & (IOMMU_READ | IOMMU_WRITE)))
529 return 0;
530
Yong Wu7f315c92019-08-24 11:01:52 +0800531 if (WARN_ON(iova >= (1ULL << data->iop.cfg.ias) ||
532 paddr >= (1ULL << data->iop.cfg.oas)))
Robin Murphy76557392017-07-03 14:52:24 +0100533 return -ERANGE;
534
Robin Murphye5fc9752016-01-26 17:13:13 +0000535 ret = __arm_v7s_map(data, iova, paddr, size, prot, 1, data->pgd);
536 /*
537 * Synchronise all PTE updates for the new mapping before there's
538 * a chance for anything to kick off a table walk for the new iova.
539 */
Robin Murphy507e4c92016-01-26 17:13:14 +0000540 if (iop->cfg.quirks & IO_PGTABLE_QUIRK_TLBI_ON_MAP) {
Will Deacon10b7a7d2019-07-02 16:44:32 +0100541 io_pgtable_tlb_flush_walk(iop, iova, size,
542 ARM_V7S_BLOCK_SIZE(2));
Robin Murphye5fc9752016-01-26 17:13:13 +0000543 } else {
544 wmb();
545 }
546
547 return ret;
548}
549
550static void arm_v7s_free_pgtable(struct io_pgtable *iop)
551{
552 struct arm_v7s_io_pgtable *data = io_pgtable_to_data(iop);
553 int i;
554
555 for (i = 0; i < ARM_V7S_PTES_PER_LVL(1); i++) {
556 arm_v7s_iopte pte = data->pgd[i];
557
558 if (ARM_V7S_PTE_IS_TABLE(pte, 1))
Yong Wu5950b952019-08-24 11:01:51 +0800559 __arm_v7s_free_table(iopte_deref(pte, 1, data),
560 2, data);
Robin Murphye5fc9752016-01-26 17:13:13 +0000561 }
562 __arm_v7s_free_table(data->pgd, 1, data);
563 kmem_cache_destroy(data->l2_tables);
564 kfree(data);
565}
566
Robin Murphy119ff302017-06-22 16:53:55 +0100567static arm_v7s_iopte arm_v7s_split_cont(struct arm_v7s_io_pgtable *data,
568 unsigned long iova, int idx, int lvl,
569 arm_v7s_iopte *ptep)
Robin Murphye5fc9752016-01-26 17:13:13 +0000570{
Robin Murphy507e4c92016-01-26 17:13:14 +0000571 struct io_pgtable *iop = &data->iop;
Robin Murphye5fc9752016-01-26 17:13:13 +0000572 arm_v7s_iopte pte;
573 size_t size = ARM_V7S_BLOCK_SIZE(lvl);
574 int i;
575
Robin Murphy119ff302017-06-22 16:53:55 +0100576 /* Check that we didn't lose a race to get the lock */
577 pte = *ptep;
578 if (!arm_v7s_pte_is_cont(pte, lvl))
579 return pte;
580
Robin Murphye5fc9752016-01-26 17:13:13 +0000581 ptep -= idx & (ARM_V7S_CONT_PAGES - 1);
Robin Murphy119ff302017-06-22 16:53:55 +0100582 pte = arm_v7s_cont_to_pte(pte, lvl);
583 for (i = 0; i < ARM_V7S_CONT_PAGES; i++)
584 ptep[i] = pte + i * size;
Robin Murphye5fc9752016-01-26 17:13:13 +0000585
Robin Murphy507e4c92016-01-26 17:13:14 +0000586 __arm_v7s_pte_sync(ptep, ARM_V7S_CONT_PAGES, &iop->cfg);
Robin Murphye5fc9752016-01-26 17:13:13 +0000587
588 size *= ARM_V7S_CONT_PAGES;
Will Deacon10b7a7d2019-07-02 16:44:32 +0100589 io_pgtable_tlb_flush_leaf(iop, iova, size, size);
Robin Murphy119ff302017-06-22 16:53:55 +0100590 return pte;
Robin Murphye5fc9752016-01-26 17:13:13 +0000591}
592
Vivek Gautam193e67c2018-02-05 23:29:19 +0530593static size_t arm_v7s_split_blk_unmap(struct arm_v7s_io_pgtable *data,
Will Deacon3951c412019-07-02 16:45:15 +0100594 struct iommu_iotlb_gather *gather,
Vivek Gautam193e67c2018-02-05 23:29:19 +0530595 unsigned long iova, size_t size,
596 arm_v7s_iopte blk_pte,
597 arm_v7s_iopte *ptep)
Robin Murphye5fc9752016-01-26 17:13:13 +0000598{
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100599 struct io_pgtable_cfg *cfg = &data->iop.cfg;
600 arm_v7s_iopte pte, *tablep;
601 int i, unmap_idx, num_entries, num_ptes;
Robin Murphye5fc9752016-01-26 17:13:13 +0000602
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100603 tablep = __arm_v7s_alloc_table(2, GFP_ATOMIC, data);
604 if (!tablep)
605 return 0; /* Bytes unmapped */
Robin Murphye5fc9752016-01-26 17:13:13 +0000606
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100607 num_ptes = ARM_V7S_PTES_PER_LVL(2);
608 num_entries = size >> ARM_V7S_LVL_SHIFT(2);
609 unmap_idx = ARM_V7S_LVL_IDX(iova, 2);
Robin Murphye5fc9752016-01-26 17:13:13 +0000610
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100611 pte = arm_v7s_prot_to_pte(arm_v7s_pte_to_prot(blk_pte, 1), 2, cfg);
612 if (num_entries > 1)
613 pte = arm_v7s_pte_to_cont(pte, 2);
614
615 for (i = 0; i < num_ptes; i += num_entries, pte += size) {
Robin Murphye5fc9752016-01-26 17:13:13 +0000616 /* Unmap! */
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100617 if (i == unmap_idx)
Robin Murphye5fc9752016-01-26 17:13:13 +0000618 continue;
619
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100620 __arm_v7s_set_pte(&tablep[i], pte, num_entries, cfg);
Robin Murphye5fc9752016-01-26 17:13:13 +0000621 }
622
Robin Murphy119ff302017-06-22 16:53:55 +0100623 pte = arm_v7s_install_table(tablep, ptep, blk_pte, cfg);
624 if (pte != blk_pte) {
625 __arm_v7s_free_table(tablep, 2, data);
626
627 if (!ARM_V7S_PTE_IS_TABLE(pte, 1))
628 return 0;
629
Yong Wu5950b952019-08-24 11:01:51 +0800630 tablep = iopte_deref(pte, 1, data);
Will Deacon3951c412019-07-02 16:45:15 +0100631 return __arm_v7s_unmap(data, gather, iova, size, 2, tablep);
Robin Murphy119ff302017-06-22 16:53:55 +0100632 }
Robin Murphyb9f1ef32017-06-22 16:53:52 +0100633
Will Deacon3951c412019-07-02 16:45:15 +0100634 io_pgtable_tlb_add_page(&data->iop, gather, iova, size);
Robin Murphye5fc9752016-01-26 17:13:13 +0000635 return size;
636}
637
Vivek Gautam193e67c2018-02-05 23:29:19 +0530638static size_t __arm_v7s_unmap(struct arm_v7s_io_pgtable *data,
Will Deacon3951c412019-07-02 16:45:15 +0100639 struct iommu_iotlb_gather *gather,
Vivek Gautam193e67c2018-02-05 23:29:19 +0530640 unsigned long iova, size_t size, int lvl,
641 arm_v7s_iopte *ptep)
Robin Murphye5fc9752016-01-26 17:13:13 +0000642{
643 arm_v7s_iopte pte[ARM_V7S_CONT_PAGES];
Robin Murphy507e4c92016-01-26 17:13:14 +0000644 struct io_pgtable *iop = &data->iop;
Robin Murphye5fc9752016-01-26 17:13:13 +0000645 int idx, i = 0, num_entries = size >> ARM_V7S_LVL_SHIFT(lvl);
646
647 /* Something went horribly wrong and we ran out of page table */
648 if (WARN_ON(lvl > 2))
649 return 0;
650
651 idx = ARM_V7S_LVL_IDX(iova, lvl);
652 ptep += idx;
653 do {
Robin Murphy119ff302017-06-22 16:53:55 +0100654 pte[i] = READ_ONCE(ptep[i]);
655 if (WARN_ON(!ARM_V7S_PTE_IS_VALID(pte[i])))
Robin Murphye5fc9752016-01-26 17:13:13 +0000656 return 0;
Robin Murphye5fc9752016-01-26 17:13:13 +0000657 } while (++i < num_entries);
658
659 /*
660 * If we've hit a contiguous 'large page' entry at this level, it
661 * needs splitting first, unless we're unmapping the whole lot.
Robin Murphy119ff302017-06-22 16:53:55 +0100662 *
663 * For splitting, we can't rewrite 16 PTEs atomically, and since we
664 * can't necessarily assume TEX remap we don't have a software bit to
665 * mark live entries being split. In practice (i.e. DMA API code), we
666 * will never be splitting large pages anyway, so just wrap this edge
667 * case in a lock for the sake of correctness and be done with it.
Robin Murphye5fc9752016-01-26 17:13:13 +0000668 */
Robin Murphy119ff302017-06-22 16:53:55 +0100669 if (num_entries <= 1 && arm_v7s_pte_is_cont(pte[0], lvl)) {
670 unsigned long flags;
671
672 spin_lock_irqsave(&data->split_lock, flags);
673 pte[0] = arm_v7s_split_cont(data, iova, idx, lvl, ptep);
674 spin_unlock_irqrestore(&data->split_lock, flags);
675 }
Robin Murphye5fc9752016-01-26 17:13:13 +0000676
677 /* If the size matches this level, we're in the right place */
678 if (num_entries) {
679 size_t blk_size = ARM_V7S_BLOCK_SIZE(lvl);
680
Robin Murphy507e4c92016-01-26 17:13:14 +0000681 __arm_v7s_set_pte(ptep, 0, num_entries, &iop->cfg);
Robin Murphye5fc9752016-01-26 17:13:13 +0000682
683 for (i = 0; i < num_entries; i++) {
684 if (ARM_V7S_PTE_IS_TABLE(pte[i], lvl)) {
685 /* Also flush any partial walks */
Will Deacon10b7a7d2019-07-02 16:44:32 +0100686 io_pgtable_tlb_flush_walk(iop, iova, blk_size,
687 ARM_V7S_BLOCK_SIZE(lvl + 1));
Yong Wu5950b952019-08-24 11:01:51 +0800688 ptep = iopte_deref(pte[i], lvl, data);
Robin Murphye5fc9752016-01-26 17:13:13 +0000689 __arm_v7s_free_table(ptep, lvl + 1, data);
Robin Murphyb2dfeba2018-09-20 17:10:26 +0100690 } else if (iop->cfg.quirks & IO_PGTABLE_QUIRK_NON_STRICT) {
691 /*
692 * Order the PTE update against queueing the IOVA, to
693 * guarantee that a flush callback from a different CPU
694 * has observed it before the TLBIALL can be issued.
695 */
696 smp_wmb();
Robin Murphye5fc9752016-01-26 17:13:13 +0000697 } else {
Will Deacon3951c412019-07-02 16:45:15 +0100698 io_pgtable_tlb_add_page(iop, gather, iova, blk_size);
Robin Murphye5fc9752016-01-26 17:13:13 +0000699 }
700 iova += blk_size;
701 }
702 return size;
703 } else if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte[0], lvl)) {
704 /*
705 * Insert a table at the next level to map the old region,
706 * minus the part we want to unmap
707 */
Will Deacon3951c412019-07-02 16:45:15 +0100708 return arm_v7s_split_blk_unmap(data, gather, iova, size, pte[0],
709 ptep);
Robin Murphye5fc9752016-01-26 17:13:13 +0000710 }
711
712 /* Keep on walkin' */
Yong Wu5950b952019-08-24 11:01:51 +0800713 ptep = iopte_deref(pte[0], lvl, data);
Will Deacon3951c412019-07-02 16:45:15 +0100714 return __arm_v7s_unmap(data, gather, iova, size, lvl + 1, ptep);
Robin Murphye5fc9752016-01-26 17:13:13 +0000715}
716
Vivek Gautam193e67c2018-02-05 23:29:19 +0530717static size_t arm_v7s_unmap(struct io_pgtable_ops *ops, unsigned long iova,
Will Deacona2d3a382019-07-02 16:44:58 +0100718 size_t size, struct iommu_iotlb_gather *gather)
Robin Murphye5fc9752016-01-26 17:13:13 +0000719{
Robin Murphye5fc9752016-01-26 17:13:13 +0000720 struct arm_v7s_io_pgtable *data = io_pgtable_ops_to_data(ops);
Robin Murphye5fc9752016-01-26 17:13:13 +0000721
Robin Murphy76557392017-07-03 14:52:24 +0100722 if (WARN_ON(upper_32_bits(iova)))
723 return 0;
724
Will Deacon3951c412019-07-02 16:45:15 +0100725 return __arm_v7s_unmap(data, gather, iova, size, 1, data->pgd);
Robin Murphye5fc9752016-01-26 17:13:13 +0000726}
727
728static phys_addr_t arm_v7s_iova_to_phys(struct io_pgtable_ops *ops,
729 unsigned long iova)
730{
731 struct arm_v7s_io_pgtable *data = io_pgtable_ops_to_data(ops);
732 arm_v7s_iopte *ptep = data->pgd, pte;
733 int lvl = 0;
734 u32 mask;
735
736 do {
Robin Murphy119ff302017-06-22 16:53:55 +0100737 ptep += ARM_V7S_LVL_IDX(iova, ++lvl);
738 pte = READ_ONCE(*ptep);
Yong Wu5950b952019-08-24 11:01:51 +0800739 ptep = iopte_deref(pte, lvl, data);
Robin Murphye5fc9752016-01-26 17:13:13 +0000740 } while (ARM_V7S_PTE_IS_TABLE(pte, lvl));
741
742 if (!ARM_V7S_PTE_IS_VALID(pte))
743 return 0;
744
745 mask = ARM_V7S_LVL_MASK(lvl);
746 if (arm_v7s_pte_is_cont(pte, lvl))
747 mask *= ARM_V7S_CONT_PAGES;
Yong Wu5950b952019-08-24 11:01:51 +0800748 return iopte_to_paddr(pte, lvl, &data->iop.cfg) | (iova & ~mask);
Robin Murphye5fc9752016-01-26 17:13:13 +0000749}
750
751static struct io_pgtable *arm_v7s_alloc_pgtable(struct io_pgtable_cfg *cfg,
752 void *cookie)
753{
754 struct arm_v7s_io_pgtable *data;
755
Yong Wu4c019de2019-08-24 11:01:54 +0800756 if (cfg->ias > ARM_V7S_ADDR_BITS)
757 return NULL;
758
759 if (cfg->oas > (arm_v7s_is_mtk_enabled(cfg) ? 34 : ARM_V7S_ADDR_BITS))
Robin Murphye5fc9752016-01-26 17:13:13 +0000760 return NULL;
761
Robin Murphy3850db42016-02-12 17:09:46 +0000762 if (cfg->quirks & ~(IO_PGTABLE_QUIRK_ARM_NS |
763 IO_PGTABLE_QUIRK_NO_PERMS |
Yong Wu1afe2312016-03-14 06:01:10 +0800764 IO_PGTABLE_QUIRK_TLBI_ON_MAP |
Yong Wu73d50812019-08-24 11:01:53 +0800765 IO_PGTABLE_QUIRK_ARM_MTK_EXT |
Robin Murphyb2dfeba2018-09-20 17:10:26 +0100766 IO_PGTABLE_QUIRK_NON_STRICT))
Robin Murphy3850db42016-02-12 17:09:46 +0000767 return NULL;
768
Yong Wu1afe2312016-03-14 06:01:10 +0800769 /* If ARM_MTK_4GB is enabled, the NO_PERMS is also expected. */
Yong Wu73d50812019-08-24 11:01:53 +0800770 if (cfg->quirks & IO_PGTABLE_QUIRK_ARM_MTK_EXT &&
Yong Wu1afe2312016-03-14 06:01:10 +0800771 !(cfg->quirks & IO_PGTABLE_QUIRK_NO_PERMS))
772 return NULL;
773
Robin Murphye5fc9752016-01-26 17:13:13 +0000774 data = kmalloc(sizeof(*data), GFP_KERNEL);
775 if (!data)
776 return NULL;
777
Robin Murphy119ff302017-06-22 16:53:55 +0100778 spin_lock_init(&data->split_lock);
Robin Murphye5fc9752016-01-26 17:13:13 +0000779 data->l2_tables = kmem_cache_create("io-pgtable_armv7s_l2",
780 ARM_V7S_TABLE_SIZE(2),
781 ARM_V7S_TABLE_SIZE(2),
Nicolas Boichat0a352552019-03-28 20:43:46 -0700782 ARM_V7S_TABLE_SLAB_FLAGS, NULL);
Robin Murphye5fc9752016-01-26 17:13:13 +0000783 if (!data->l2_tables)
784 goto out_free_data;
785
786 data->iop.ops = (struct io_pgtable_ops) {
787 .map = arm_v7s_map,
788 .unmap = arm_v7s_unmap,
789 .iova_to_phys = arm_v7s_iova_to_phys,
790 };
791
792 /* We have to do this early for __arm_v7s_alloc_table to work... */
793 data->iop.cfg = *cfg;
794
795 /*
796 * Unless the IOMMU driver indicates supersection support by
797 * having SZ_16M set in the initial bitmap, they won't be used.
798 */
799 cfg->pgsize_bitmap &= SZ_4K | SZ_64K | SZ_1M | SZ_16M;
800
801 /* TCR: T0SZ=0, disable TTBR1 */
802 cfg->arm_v7s_cfg.tcr = ARM_V7S_TCR_PD1;
803
804 /*
805 * TEX remap: the indices used map to the closest equivalent types
806 * under the non-TEX-remap interpretation of those attribute bits,
807 * excepting various implementation-defined aspects of shareability.
808 */
809 cfg->arm_v7s_cfg.prrr = ARM_V7S_PRRR_TR(1, ARM_V7S_PRRR_TYPE_DEVICE) |
810 ARM_V7S_PRRR_TR(4, ARM_V7S_PRRR_TYPE_NORMAL) |
811 ARM_V7S_PRRR_TR(7, ARM_V7S_PRRR_TYPE_NORMAL) |
812 ARM_V7S_PRRR_DS0 | ARM_V7S_PRRR_DS1 |
813 ARM_V7S_PRRR_NS1 | ARM_V7S_PRRR_NOS(7);
814 cfg->arm_v7s_cfg.nmrr = ARM_V7S_NMRR_IR(7, ARM_V7S_RGN_WBWA) |
815 ARM_V7S_NMRR_OR(7, ARM_V7S_RGN_WBWA);
816
817 /* Looking good; allocate a pgd */
818 data->pgd = __arm_v7s_alloc_table(1, GFP_KERNEL, data);
819 if (!data->pgd)
820 goto out_free_data;
821
822 /* Ensure the empty pgd is visible before any actual TTBR write */
823 wmb();
824
825 /* TTBRs */
826 cfg->arm_v7s_cfg.ttbr[0] = virt_to_phys(data->pgd) |
827 ARM_V7S_TTBR_S | ARM_V7S_TTBR_NOS |
Bjorn Andersson9e6ea592019-05-15 16:32:34 -0700828 (cfg->coherent_walk ?
829 (ARM_V7S_TTBR_IRGN_ATTR(ARM_V7S_RGN_WBWA) |
830 ARM_V7S_TTBR_ORGN_ATTR(ARM_V7S_RGN_WBWA)) :
831 (ARM_V7S_TTBR_IRGN_ATTR(ARM_V7S_RGN_NC) |
832 ARM_V7S_TTBR_ORGN_ATTR(ARM_V7S_RGN_NC)));
Robin Murphye5fc9752016-01-26 17:13:13 +0000833 cfg->arm_v7s_cfg.ttbr[1] = 0;
834 return &data->iop;
835
836out_free_data:
837 kmem_cache_destroy(data->l2_tables);
838 kfree(data);
839 return NULL;
840}
841
842struct io_pgtable_init_fns io_pgtable_arm_v7s_init_fns = {
843 .alloc = arm_v7s_alloc_pgtable,
844 .free = arm_v7s_free_pgtable,
845};
846
847#ifdef CONFIG_IOMMU_IO_PGTABLE_ARMV7S_SELFTEST
848
Robin Murphyb5813c12019-10-25 19:08:30 +0100849static struct io_pgtable_cfg *cfg_cookie __initdata;
Robin Murphye5fc9752016-01-26 17:13:13 +0000850
Robin Murphyb5813c12019-10-25 19:08:30 +0100851static void __init dummy_tlb_flush_all(void *cookie)
Robin Murphye5fc9752016-01-26 17:13:13 +0000852{
853 WARN_ON(cookie != cfg_cookie);
854}
855
Robin Murphyb5813c12019-10-25 19:08:30 +0100856static void __init dummy_tlb_flush(unsigned long iova, size_t size,
857 size_t granule, void *cookie)
Robin Murphye5fc9752016-01-26 17:13:13 +0000858{
859 WARN_ON(cookie != cfg_cookie);
860 WARN_ON(!(size & cfg_cookie->pgsize_bitmap));
861}
862
Robin Murphyb5813c12019-10-25 19:08:30 +0100863static void __init dummy_tlb_add_page(struct iommu_iotlb_gather *gather,
864 unsigned long iova, size_t granule,
865 void *cookie)
Robin Murphye5fc9752016-01-26 17:13:13 +0000866{
Will Deaconabfd6fe2019-07-02 16:44:41 +0100867 dummy_tlb_flush(iova, granule, granule, cookie);
Robin Murphye5fc9752016-01-26 17:13:13 +0000868}
869
Robin Murphyb5813c12019-10-25 19:08:30 +0100870static const struct iommu_flush_ops dummy_tlb_ops __initconst = {
Robin Murphye5fc9752016-01-26 17:13:13 +0000871 .tlb_flush_all = dummy_tlb_flush_all,
Will Deacon10b7a7d2019-07-02 16:44:32 +0100872 .tlb_flush_walk = dummy_tlb_flush,
873 .tlb_flush_leaf = dummy_tlb_flush,
Will Deaconabfd6fe2019-07-02 16:44:41 +0100874 .tlb_add_page = dummy_tlb_add_page,
Robin Murphye5fc9752016-01-26 17:13:13 +0000875};
876
877#define __FAIL(ops) ({ \
878 WARN(1, "selftest: test failed\n"); \
879 selftest_running = false; \
880 -EFAULT; \
881})
882
883static int __init arm_v7s_do_selftests(void)
884{
885 struct io_pgtable_ops *ops;
886 struct io_pgtable_cfg cfg = {
887 .tlb = &dummy_tlb_ops,
888 .oas = 32,
889 .ias = 32,
Will Deacon4f418452019-06-25 12:51:25 +0100890 .coherent_walk = true,
891 .quirks = IO_PGTABLE_QUIRK_ARM_NS,
Robin Murphye5fc9752016-01-26 17:13:13 +0000892 .pgsize_bitmap = SZ_4K | SZ_64K | SZ_1M | SZ_16M,
893 };
894 unsigned int iova, size, iova_start;
895 unsigned int i, loopnr = 0;
896
897 selftest_running = true;
898
899 cfg_cookie = &cfg;
900
901 ops = alloc_io_pgtable_ops(ARM_V7S, &cfg, &cfg);
902 if (!ops) {
903 pr_err("selftest: failed to allocate io pgtable ops\n");
904 return -EINVAL;
905 }
906
907 /*
908 * Initial sanity checks.
909 * Empty page tables shouldn't provide any translations.
910 */
911 if (ops->iova_to_phys(ops, 42))
912 return __FAIL(ops);
913
914 if (ops->iova_to_phys(ops, SZ_1G + 42))
915 return __FAIL(ops);
916
917 if (ops->iova_to_phys(ops, SZ_2G + 42))
918 return __FAIL(ops);
919
920 /*
921 * Distinct mappings of different granule sizes.
922 */
923 iova = 0;
Kefeng Wang4ae8a5c2016-09-21 13:41:31 +0800924 for_each_set_bit(i, &cfg.pgsize_bitmap, BITS_PER_LONG) {
Robin Murphye5fc9752016-01-26 17:13:13 +0000925 size = 1UL << i;
926 if (ops->map(ops, iova, iova, size, IOMMU_READ |
927 IOMMU_WRITE |
928 IOMMU_NOEXEC |
929 IOMMU_CACHE))
930 return __FAIL(ops);
931
932 /* Overlapping mappings */
933 if (!ops->map(ops, iova, iova + size, size,
934 IOMMU_READ | IOMMU_NOEXEC))
935 return __FAIL(ops);
936
937 if (ops->iova_to_phys(ops, iova + 42) != (iova + 42))
938 return __FAIL(ops);
939
940 iova += SZ_16M;
Robin Murphye5fc9752016-01-26 17:13:13 +0000941 loopnr++;
942 }
943
944 /* Partial unmap */
945 i = 1;
946 size = 1UL << __ffs(cfg.pgsize_bitmap);
947 while (i < loopnr) {
948 iova_start = i * SZ_16M;
Will Deacona2d3a382019-07-02 16:44:58 +0100949 if (ops->unmap(ops, iova_start + size, size, NULL) != size)
Robin Murphye5fc9752016-01-26 17:13:13 +0000950 return __FAIL(ops);
951
952 /* Remap of partial unmap */
953 if (ops->map(ops, iova_start + size, size, size, IOMMU_READ))
954 return __FAIL(ops);
955
956 if (ops->iova_to_phys(ops, iova_start + size + 42)
957 != (size + 42))
958 return __FAIL(ops);
959 i++;
960 }
961
962 /* Full unmap */
963 iova = 0;
YueHaibingf793b132018-04-26 12:49:29 +0800964 for_each_set_bit(i, &cfg.pgsize_bitmap, BITS_PER_LONG) {
Robin Murphye5fc9752016-01-26 17:13:13 +0000965 size = 1UL << i;
966
Will Deacona2d3a382019-07-02 16:44:58 +0100967 if (ops->unmap(ops, iova, size, NULL) != size)
Robin Murphye5fc9752016-01-26 17:13:13 +0000968 return __FAIL(ops);
969
970 if (ops->iova_to_phys(ops, iova + 42))
971 return __FAIL(ops);
972
973 /* Remap full block */
974 if (ops->map(ops, iova, iova, size, IOMMU_WRITE))
975 return __FAIL(ops);
976
977 if (ops->iova_to_phys(ops, iova + 42) != (iova + 42))
978 return __FAIL(ops);
979
980 iova += SZ_16M;
Robin Murphye5fc9752016-01-26 17:13:13 +0000981 }
982
983 free_io_pgtable_ops(ops);
984
985 selftest_running = false;
986
987 pr_info("self test ok\n");
988 return 0;
989}
990subsys_initcall(arm_v7s_do_selftests);
991#endif