blob: 8b3e5b5cd8fa5d10f3a639d2e107bd28b53dc01b [file] [log] [blame]
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -08001#include <linux/kernel.h>
2#include <linux/mm.h>
3#include <linux/slab.h>
4#include <linux/uaccess.h>
5#include <linux/ktime.h>
6#include <linux/debugfs.h>
7
8#define GUP_FAST_BENCHMARK _IOWR('g', 1, struct gup_benchmark)
Barry Song657d4f72020-10-13 16:51:54 -07009#define GUP_BENCHMARK _IOWR('g', 2, struct gup_benchmark)
10#define PIN_FAST_BENCHMARK _IOWR('g', 3, struct gup_benchmark)
11#define PIN_BENCHMARK _IOWR('g', 4, struct gup_benchmark)
12#define PIN_LONGTERM_BENCHMARK _IOWR('g', 5, struct gup_benchmark)
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080013
14struct gup_benchmark {
Keith Busch26db3d02018-10-26 15:09:52 -070015 __u64 get_delta_usec;
16 __u64 put_delta_usec;
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080017 __u64 addr;
18 __u64 size;
19 __u32 nr_pages_per_call;
20 __u32 flags;
Keith Busch26db3d02018-10-26 15:09:52 -070021 __u64 expansion[10]; /* For future use */
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080022};
23
John Hubbard41c45d32020-04-01 21:05:41 -070024static void put_back_pages(unsigned int cmd, struct page **pages,
25 unsigned long nr_pages)
26{
27 unsigned long i;
28
29 switch (cmd) {
30 case GUP_FAST_BENCHMARK:
John Hubbard41c45d32020-04-01 21:05:41 -070031 case GUP_BENCHMARK:
32 for (i = 0; i < nr_pages; i++)
33 put_page(pages[i]);
34 break;
35
36 case PIN_FAST_BENCHMARK:
37 case PIN_BENCHMARK:
Barry Song657d4f72020-10-13 16:51:54 -070038 case PIN_LONGTERM_BENCHMARK:
John Hubbard41c45d32020-04-01 21:05:41 -070039 unpin_user_pages(pages, nr_pages);
40 break;
41 }
42}
43
44static void verify_dma_pinned(unsigned int cmd, struct page **pages,
45 unsigned long nr_pages)
46{
47 unsigned long i;
48 struct page *page;
49
50 switch (cmd) {
51 case PIN_FAST_BENCHMARK:
52 case PIN_BENCHMARK:
Barry Song657d4f72020-10-13 16:51:54 -070053 case PIN_LONGTERM_BENCHMARK:
John Hubbard41c45d32020-04-01 21:05:41 -070054 for (i = 0; i < nr_pages; i++) {
55 page = pages[i];
56 if (WARN(!page_maybe_dma_pinned(page),
57 "pages[%lu] is NOT dma-pinned\n", i)) {
58
59 dump_page(page, "gup_benchmark failure");
60 break;
61 }
62 }
63 break;
64 }
65}
66
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080067static int __gup_benchmark_ioctl(unsigned int cmd,
68 struct gup_benchmark *gup)
69{
70 ktime_t start_time, end_time;
YueHaibing51896862018-10-05 15:51:44 -070071 unsigned long i, nr_pages, addr, next;
72 int nr;
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080073 struct page **pages;
Navid Emamdoosta7c46c02020-01-04 13:00:12 -080074 int ret = 0;
Jann Hornf3964592020-10-17 16:14:12 -070075 bool needs_mmap_lock =
76 cmd != GUP_FAST_BENCHMARK && cmd != PIN_FAST_BENCHMARK;
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080077
Dan Carpenter4b408c72018-10-30 15:04:32 -070078 if (gup->size > ULONG_MAX)
79 return -EINVAL;
80
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080081 nr_pages = gup->size / PAGE_SIZE;
Kees Cook778e1cd2018-06-12 14:04:48 -070082 pages = kvcalloc(nr_pages, sizeof(void *), GFP_KERNEL);
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080083 if (!pages)
84 return -ENOMEM;
85
Jann Hornf3964592020-10-17 16:14:12 -070086 if (needs_mmap_lock && mmap_read_lock_killable(current->mm)) {
87 ret = -EINTR;
88 goto free_pages;
89 }
90
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -080091 i = 0;
92 nr = gup->nr_pages_per_call;
93 start_time = ktime_get();
94 for (addr = gup->addr; addr < gup->addr + gup->size; addr = next) {
95 if (nr != gup->nr_pages_per_call)
96 break;
97
98 next = addr + nr * PAGE_SIZE;
99 if (next > gup->addr + gup->size) {
100 next = gup->addr + gup->size;
101 nr = (next - addr) / PAGE_SIZE;
102 }
103
John Hubbardbdffe232020-01-30 22:13:32 -0800104 /* Filter out most gup flags: only allow a tiny subset here: */
105 gup->flags &= FOLL_WRITE;
106
Keith Busch714a3a12018-10-26 15:09:56 -0700107 switch (cmd) {
108 case GUP_FAST_BENCHMARK:
John Hubbardbdffe232020-01-30 22:13:32 -0800109 nr = get_user_pages_fast(addr, nr, gup->flags,
Keith Busch714a3a12018-10-26 15:09:56 -0700110 pages + i);
111 break;
Keith Busch714a3a12018-10-26 15:09:56 -0700112 case GUP_BENCHMARK:
John Hubbardbdffe232020-01-30 22:13:32 -0800113 nr = get_user_pages(addr, nr, gup->flags, pages + i,
Keith Busch714a3a12018-10-26 15:09:56 -0700114 NULL);
115 break;
John Hubbard41c45d32020-04-01 21:05:41 -0700116 case PIN_FAST_BENCHMARK:
117 nr = pin_user_pages_fast(addr, nr, gup->flags,
118 pages + i);
119 break;
120 case PIN_BENCHMARK:
121 nr = pin_user_pages(addr, nr, gup->flags, pages + i,
122 NULL);
123 break;
Barry Song657d4f72020-10-13 16:51:54 -0700124 case PIN_LONGTERM_BENCHMARK:
125 nr = pin_user_pages(addr, nr,
126 gup->flags | FOLL_LONGTERM,
127 pages + i, NULL);
128 break;
Keith Busch714a3a12018-10-26 15:09:56 -0700129 default:
Navid Emamdoosta7c46c02020-01-04 13:00:12 -0800130 ret = -EINVAL;
Jann Hornf3964592020-10-17 16:14:12 -0700131 goto unlock;
Keith Busch714a3a12018-10-26 15:09:56 -0700132 }
133
Michael S. Tsirkin09e35a42018-04-13 15:35:16 -0700134 if (nr <= 0)
135 break;
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -0800136 i += nr;
137 }
138 end_time = ktime_get();
139
John Hubbard41c45d32020-04-01 21:05:41 -0700140 /* Shifting the meaning of nr_pages: now it is actual number pinned: */
141 nr_pages = i;
142
Keith Busch26db3d02018-10-26 15:09:52 -0700143 gup->get_delta_usec = ktime_us_delta(end_time, start_time);
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -0800144 gup->size = addr - gup->addr;
145
John Hubbard41c45d32020-04-01 21:05:41 -0700146 /*
147 * Take an un-benchmark-timed moment to verify DMA pinned
148 * state: print a warning if any non-dma-pinned pages are found:
149 */
150 verify_dma_pinned(cmd, pages, nr_pages);
151
Keith Busch26db3d02018-10-26 15:09:52 -0700152 start_time = ktime_get();
John Hubbard41c45d32020-04-01 21:05:41 -0700153
154 put_back_pages(cmd, pages, nr_pages);
155
Keith Busch26db3d02018-10-26 15:09:52 -0700156 end_time = ktime_get();
157 gup->put_delta_usec = ktime_us_delta(end_time, start_time);
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -0800158
Jann Hornf3964592020-10-17 16:14:12 -0700159unlock:
160 if (needs_mmap_lock)
161 mmap_read_unlock(current->mm);
162free_pages:
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -0800163 kvfree(pages);
Navid Emamdoosta7c46c02020-01-04 13:00:12 -0800164 return ret;
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -0800165}
166
167static long gup_benchmark_ioctl(struct file *filep, unsigned int cmd,
168 unsigned long arg)
169{
170 struct gup_benchmark gup;
171 int ret;
172
Keith Busch714a3a12018-10-26 15:09:56 -0700173 switch (cmd) {
174 case GUP_FAST_BENCHMARK:
Keith Busch714a3a12018-10-26 15:09:56 -0700175 case GUP_BENCHMARK:
John Hubbard41c45d32020-04-01 21:05:41 -0700176 case PIN_FAST_BENCHMARK:
177 case PIN_BENCHMARK:
Barry Song657d4f72020-10-13 16:51:54 -0700178 case PIN_LONGTERM_BENCHMARK:
Keith Busch714a3a12018-10-26 15:09:56 -0700179 break;
180 default:
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -0800181 return -EINVAL;
Keith Busch714a3a12018-10-26 15:09:56 -0700182 }
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -0800183
184 if (copy_from_user(&gup, (void __user *)arg, sizeof(gup)))
185 return -EFAULT;
186
187 ret = __gup_benchmark_ioctl(cmd, &gup);
188 if (ret)
189 return ret;
190
191 if (copy_to_user((void __user *)arg, &gup, sizeof(gup)))
192 return -EFAULT;
193
194 return 0;
195}
196
197static const struct file_operations gup_benchmark_fops = {
198 .open = nonseekable_open,
199 .unlocked_ioctl = gup_benchmark_ioctl,
200};
201
202static int gup_benchmark_init(void)
203{
Greg Kroah-Hartmand9f79792019-03-05 15:46:09 -0800204 debugfs_create_file_unsafe("gup_benchmark", 0600, NULL, NULL,
205 &gup_benchmark_fops);
Kirill A. Shutemov64c349f2017-11-17 15:31:22 -0800206
207 return 0;
208}
209
210late_initcall(gup_benchmark_init);