blob: 0c5cb2d6436a4772994e3cf75ee60e88e2d2a5ef [file] [log] [blame]
Thomas Gleixner5b497af2019-05-29 07:18:09 -07001// SPDX-License-Identifier: GPL-2.0-only
Alexei Starovoitov64a89462014-05-08 14:10:52 -07002/*
3 * Testsuite for BPF interpreter and BPF JIT compiler
4 *
5 * Copyright (c) 2011-2014 PLUMgrid, http://plumgrid.com
Alexei Starovoitov64a89462014-05-08 14:10:52 -07006 */
7
8#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
9
10#include <linux/init.h>
11#include <linux/module.h>
12#include <linux/filter.h>
Alexei Starovoitov4d9c5c52015-07-20 20:34:19 -070013#include <linux/bpf.h>
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014#include <linux/skbuff.h>
15#include <linux/netdevice.h>
16#include <linux/if_vlan.h>
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020017#include <linux/random.h>
Nicolas Schichanbac142a2015-08-04 15:19:08 +020018#include <linux/highmem.h>
Eric Dumazetd40bc962018-02-26 10:52:46 -080019#include <linux/sched.h>
Alexei Starovoitov64a89462014-05-08 14:10:52 -070020
Daniel Borkmann10f18e02014-05-23 18:44:00 +020021/* General test specific settings */
Alexei Starovoitov64a89462014-05-08 14:10:52 -070022#define MAX_SUBTESTS 3
Eric Dumazet9960d762018-02-28 08:39:20 -080023#define MAX_TESTRUNS 1000
Alexei Starovoitov64a89462014-05-08 14:10:52 -070024#define MAX_DATA 128
25#define MAX_INSNS 512
26#define MAX_K 0xffffFFFF
27
Daniel Borkmann10f18e02014-05-23 18:44:00 +020028/* Few constants used to init test 'skb' */
Alexei Starovoitov64a89462014-05-08 14:10:52 -070029#define SKB_TYPE 3
30#define SKB_MARK 0x1234aaaa
31#define SKB_HASH 0x1234aaab
32#define SKB_QUEUE_MAP 123
33#define SKB_VLAN_TCI 0xffff
Michał Mirosław0c4b2d32018-11-10 19:58:36 +010034#define SKB_VLAN_PRESENT 1
Alexei Starovoitov64a89462014-05-08 14:10:52 -070035#define SKB_DEV_IFINDEX 577
36#define SKB_DEV_TYPE 588
37
Daniel Borkmann10f18e02014-05-23 18:44:00 +020038/* Redefine REGs to make tests less verbose */
39#define R0 BPF_REG_0
40#define R1 BPF_REG_1
41#define R2 BPF_REG_2
42#define R3 BPF_REG_3
43#define R4 BPF_REG_4
44#define R5 BPF_REG_5
45#define R6 BPF_REG_6
46#define R7 BPF_REG_7
47#define R8 BPF_REG_8
48#define R9 BPF_REG_9
49#define R10 BPF_REG_10
50
51/* Flags that can be passed to test cases */
52#define FLAG_NO_DATA BIT(0)
53#define FLAG_EXPECTED_FAIL BIT(1)
Nicolas Schichanbac142a2015-08-04 15:19:08 +020054#define FLAG_SKB_FRAG BIT(2)
Johan Almbladh27cc6da2021-09-14 11:18:36 +020055#define FLAG_VERIFIER_ZEXT BIT(3)
Daniel Borkmann10f18e02014-05-23 18:44:00 +020056
57enum {
58 CLASSIC = BIT(6), /* Old BPF instructions only. */
59 INTERNAL = BIT(7), /* Extended instruction set. */
60};
61
62#define TEST_TYPE_MASK (CLASSIC | INTERNAL)
Alexei Starovoitov64a89462014-05-08 14:10:52 -070063
64struct bpf_test {
65 const char *descr;
66 union {
67 struct sock_filter insns[MAX_INSNS];
Alexei Starovoitov2695fb52014-07-24 16:38:21 -070068 struct bpf_insn insns_int[MAX_INSNS];
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020069 struct {
70 void *insns;
71 unsigned int len;
72 } ptr;
Andrew Mortonece80492014-05-22 10:16:46 -070073 } u;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020074 __u8 aux;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070075 __u8 data[MAX_DATA];
76 struct {
77 int data_size;
78 __u32 result;
79 } test[MAX_SUBTESTS];
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020080 int (*fill_helper)(struct bpf_test *self);
Yonghong Song09584b42018-02-02 22:37:15 -080081 int expected_errcode; /* used when FLAG_EXPECTED_FAIL is set in the aux */
Nicolas Schichanbac142a2015-08-04 15:19:08 +020082 __u8 frag_data[MAX_DATA];
Alexei Starovoitov105c0362017-05-30 13:31:32 -070083 int stack_depth; /* for eBPF only, since tests don't call verifier */
Johan Almbladhc2a228d2021-09-14 11:18:29 +020084 int nr_testruns; /* Custom run count, defaults to MAX_TESTRUNS if 0 */
Alexei Starovoitov64a89462014-05-08 14:10:52 -070085};
86
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020087/* Large test cases need separate allocation and fill handler. */
88
89static int bpf_fill_maxinsns1(struct bpf_test *self)
90{
91 unsigned int len = BPF_MAXINSNS;
92 struct sock_filter *insn;
93 __u32 k = ~0;
94 int i;
95
96 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
97 if (!insn)
98 return -ENOMEM;
99
100 for (i = 0; i < len; i++, k--)
101 insn[i] = __BPF_STMT(BPF_RET | BPF_K, k);
102
103 self->u.ptr.insns = insn;
104 self->u.ptr.len = len;
105
106 return 0;
107}
108
109static int bpf_fill_maxinsns2(struct bpf_test *self)
110{
111 unsigned int len = BPF_MAXINSNS;
112 struct sock_filter *insn;
113 int i;
114
115 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
116 if (!insn)
117 return -ENOMEM;
118
119 for (i = 0; i < len; i++)
120 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe);
121
122 self->u.ptr.insns = insn;
123 self->u.ptr.len = len;
124
125 return 0;
126}
127
128static int bpf_fill_maxinsns3(struct bpf_test *self)
129{
130 unsigned int len = BPF_MAXINSNS;
131 struct sock_filter *insn;
132 struct rnd_state rnd;
133 int i;
134
135 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
136 if (!insn)
137 return -ENOMEM;
138
139 prandom_seed_state(&rnd, 3141592653589793238ULL);
140
141 for (i = 0; i < len - 1; i++) {
142 __u32 k = prandom_u32_state(&rnd);
143
144 insn[i] = __BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, k);
145 }
146
147 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0);
148
149 self->u.ptr.insns = insn;
150 self->u.ptr.len = len;
151
152 return 0;
153}
154
155static int bpf_fill_maxinsns4(struct bpf_test *self)
156{
157 unsigned int len = BPF_MAXINSNS + 1;
158 struct sock_filter *insn;
159 int i;
160
161 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
162 if (!insn)
163 return -ENOMEM;
164
165 for (i = 0; i < len; i++)
166 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe);
167
168 self->u.ptr.insns = insn;
169 self->u.ptr.len = len;
170
171 return 0;
172}
173
174static int bpf_fill_maxinsns5(struct bpf_test *self)
175{
176 unsigned int len = BPF_MAXINSNS;
177 struct sock_filter *insn;
178 int i;
179
180 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
181 if (!insn)
182 return -ENOMEM;
183
184 insn[0] = __BPF_JUMP(BPF_JMP | BPF_JA, len - 2, 0, 0);
185
186 for (i = 1; i < len - 1; i++)
187 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe);
188
189 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xabababab);
190
191 self->u.ptr.insns = insn;
192 self->u.ptr.len = len;
193
194 return 0;
195}
196
197static int bpf_fill_maxinsns6(struct bpf_test *self)
198{
199 unsigned int len = BPF_MAXINSNS;
200 struct sock_filter *insn;
201 int i;
202
203 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
204 if (!insn)
205 return -ENOMEM;
206
207 for (i = 0; i < len - 1; i++)
208 insn[i] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF +
209 SKF_AD_VLAN_TAG_PRESENT);
210
211 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0);
212
213 self->u.ptr.insns = insn;
214 self->u.ptr.len = len;
215
216 return 0;
217}
218
219static int bpf_fill_maxinsns7(struct bpf_test *self)
220{
221 unsigned int len = BPF_MAXINSNS;
222 struct sock_filter *insn;
223 int i;
224
225 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
226 if (!insn)
227 return -ENOMEM;
228
229 for (i = 0; i < len - 4; i++)
230 insn[i] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF +
231 SKF_AD_CPU);
232
233 insn[len - 4] = __BPF_STMT(BPF_MISC | BPF_TAX, 0);
234 insn[len - 3] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF +
235 SKF_AD_CPU);
236 insn[len - 2] = __BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0);
237 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0);
238
239 self->u.ptr.insns = insn;
240 self->u.ptr.len = len;
241
242 return 0;
243}
244
245static int bpf_fill_maxinsns8(struct bpf_test *self)
246{
247 unsigned int len = BPF_MAXINSNS;
248 struct sock_filter *insn;
249 int i, jmp_off = len - 3;
250
251 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
252 if (!insn)
253 return -ENOMEM;
254
255 insn[0] = __BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff);
256
257 for (i = 1; i < len - 1; i++)
258 insn[i] = __BPF_JUMP(BPF_JMP | BPF_JGT, 0xffffffff, jmp_off--, 0);
259
260 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0);
261
262 self->u.ptr.insns = insn;
263 self->u.ptr.len = len;
264
265 return 0;
266}
267
Daniel Borkmann3b529602015-05-23 01:10:07 +0200268static int bpf_fill_maxinsns9(struct bpf_test *self)
269{
270 unsigned int len = BPF_MAXINSNS;
271 struct bpf_insn *insn;
272 int i;
273
274 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
275 if (!insn)
276 return -ENOMEM;
277
278 insn[0] = BPF_JMP_IMM(BPF_JA, 0, 0, len - 2);
279 insn[1] = BPF_ALU32_IMM(BPF_MOV, R0, 0xcbababab);
280 insn[2] = BPF_EXIT_INSN();
281
282 for (i = 3; i < len - 2; i++)
283 insn[i] = BPF_ALU32_IMM(BPF_MOV, R0, 0xfefefefe);
284
285 insn[len - 2] = BPF_EXIT_INSN();
286 insn[len - 1] = BPF_JMP_IMM(BPF_JA, 0, 0, -(len - 1));
287
288 self->u.ptr.insns = insn;
289 self->u.ptr.len = len;
290
291 return 0;
292}
293
294static int bpf_fill_maxinsns10(struct bpf_test *self)
295{
296 unsigned int len = BPF_MAXINSNS, hlen = len - 2;
297 struct bpf_insn *insn;
298 int i;
299
300 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
301 if (!insn)
302 return -ENOMEM;
303
304 for (i = 0; i < hlen / 2; i++)
305 insn[i] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen - 2 - 2 * i);
306 for (i = hlen - 1; i > hlen / 2; i--)
307 insn[i] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen - 1 - 2 * i);
308
309 insn[hlen / 2] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen / 2 - 1);
310 insn[hlen] = BPF_ALU32_IMM(BPF_MOV, R0, 0xabababac);
311 insn[hlen + 1] = BPF_EXIT_INSN();
312
313 self->u.ptr.insns = insn;
314 self->u.ptr.len = len;
315
316 return 0;
317}
318
Daniel Borkmannbde28bc2015-05-26 22:35:43 +0200319static int __bpf_fill_ja(struct bpf_test *self, unsigned int len,
320 unsigned int plen)
321{
322 struct sock_filter *insn;
323 unsigned int rlen;
324 int i, j;
325
326 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
327 if (!insn)
328 return -ENOMEM;
329
330 rlen = (len % plen) - 1;
331
332 for (i = 0; i + plen < len; i += plen)
333 for (j = 0; j < plen; j++)
334 insn[i + j] = __BPF_JUMP(BPF_JMP | BPF_JA,
335 plen - 1 - j, 0, 0);
336 for (j = 0; j < rlen; j++)
337 insn[i + j] = __BPF_JUMP(BPF_JMP | BPF_JA, rlen - 1 - j,
338 0, 0);
339
340 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xababcbac);
341
342 self->u.ptr.insns = insn;
343 self->u.ptr.len = len;
344
345 return 0;
346}
347
348static int bpf_fill_maxinsns11(struct bpf_test *self)
349{
Gary Lin16a660e2021-01-19 18:25:00 +0800350 /* Hits 70 passes on x86_64 and triggers NOPs padding. */
Daniel Borkmannbde28bc2015-05-26 22:35:43 +0200351 return __bpf_fill_ja(self, BPF_MAXINSNS, 68);
352}
353
Daniel Borkmannbe088152018-06-02 23:06:32 +0200354static int bpf_fill_maxinsns12(struct bpf_test *self)
355{
356 unsigned int len = BPF_MAXINSNS;
357 struct sock_filter *insn;
358 int i = 0;
359
360 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
361 if (!insn)
362 return -ENOMEM;
363
364 insn[0] = __BPF_JUMP(BPF_JMP | BPF_JA, len - 2, 0, 0);
365
366 for (i = 1; i < len - 1; i++)
367 insn[i] = __BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0);
368
369 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xabababab);
370
371 self->u.ptr.insns = insn;
372 self->u.ptr.len = len;
373
374 return 0;
375}
376
377static int bpf_fill_maxinsns13(struct bpf_test *self)
378{
379 unsigned int len = BPF_MAXINSNS;
380 struct sock_filter *insn;
381 int i = 0;
382
383 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
384 if (!insn)
385 return -ENOMEM;
386
387 for (i = 0; i < len - 3; i++)
388 insn[i] = __BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0);
389
390 insn[len - 3] = __BPF_STMT(BPF_LD | BPF_IMM, 0xabababab);
391 insn[len - 2] = __BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0);
392 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0);
393
394 self->u.ptr.insns = insn;
395 self->u.ptr.len = len;
396
397 return 0;
398}
399
Daniel Borkmannbde28bc2015-05-26 22:35:43 +0200400static int bpf_fill_ja(struct bpf_test *self)
401{
402 /* Hits exactly 11 passes on x86_64 JIT. */
403 return __bpf_fill_ja(self, 12, 9);
404}
405
Alexei Starovoitov4d9c5c52015-07-20 20:34:19 -0700406static int bpf_fill_ld_abs_get_processor_id(struct bpf_test *self)
407{
408 unsigned int len = BPF_MAXINSNS;
409 struct sock_filter *insn;
410 int i;
411
412 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
413 if (!insn)
414 return -ENOMEM;
415
416 for (i = 0; i < len - 1; i += 2) {
417 insn[i] = __BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 0);
418 insn[i + 1] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
419 SKF_AD_OFF + SKF_AD_CPU);
420 }
421
422 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xbee);
423
424 self->u.ptr.insns = insn;
425 self->u.ptr.len = len;
426
427 return 0;
428}
429
Daniel Borkmann85f68fe2017-05-01 02:57:20 +0200430static int __bpf_fill_stxdw(struct bpf_test *self, int size)
431{
432 unsigned int len = BPF_MAXINSNS;
433 struct bpf_insn *insn;
434 int i;
435
436 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
437 if (!insn)
438 return -ENOMEM;
439
440 insn[0] = BPF_ALU32_IMM(BPF_MOV, R0, 1);
441 insn[1] = BPF_ST_MEM(size, R10, -40, 42);
442
443 for (i = 2; i < len - 2; i++)
444 insn[i] = BPF_STX_XADD(size, R10, R0, -40);
445
446 insn[len - 2] = BPF_LDX_MEM(size, R0, R10, -40);
447 insn[len - 1] = BPF_EXIT_INSN();
448
449 self->u.ptr.insns = insn;
450 self->u.ptr.len = len;
Alexei Starovoitov105c0362017-05-30 13:31:32 -0700451 self->stack_depth = 40;
Daniel Borkmann85f68fe2017-05-01 02:57:20 +0200452
453 return 0;
454}
455
456static int bpf_fill_stxw(struct bpf_test *self)
457{
458 return __bpf_fill_stxdw(self, BPF_W);
459}
460
461static int bpf_fill_stxdw(struct bpf_test *self)
462{
463 return __bpf_fill_stxdw(self, BPF_DW);
464}
465
Johan Almbladh68c956f2021-09-14 11:18:31 +0200466static int __bpf_ld_imm64(struct bpf_insn insns[2], u8 reg, s64 imm64)
467{
468 struct bpf_insn tmp[] = {BPF_LD_IMM64(reg, imm64)};
469
470 memcpy(insns, tmp, sizeof(tmp));
471 return 2;
472}
473
Johan Almbladhf1517eb2021-09-14 11:18:39 +0200474/*
475 * Branch conversion tests. Complex operations can expand to a lot
476 * of instructions when JITed. This in turn may cause jump offsets
477 * to overflow the field size of the native instruction, triggering
478 * a branch conversion mechanism in some JITs.
479 */
480static int __bpf_fill_max_jmp(struct bpf_test *self, int jmp, int imm)
481{
482 struct bpf_insn *insns;
483 int len = S16_MAX + 5;
484 int i;
485
486 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL);
487 if (!insns)
488 return -ENOMEM;
489
490 i = __bpf_ld_imm64(insns, R1, 0x0123456789abcdefULL);
491 insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1);
492 insns[i++] = BPF_JMP_IMM(jmp, R0, imm, S16_MAX);
493 insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 2);
494 insns[i++] = BPF_EXIT_INSN();
495
496 while (i < len - 1) {
497 static const int ops[] = {
498 BPF_LSH, BPF_RSH, BPF_ARSH, BPF_ADD,
499 BPF_SUB, BPF_MUL, BPF_DIV, BPF_MOD,
500 };
501 int op = ops[(i >> 1) % ARRAY_SIZE(ops)];
502
503 if (i & 1)
504 insns[i++] = BPF_ALU32_REG(op, R0, R1);
505 else
506 insns[i++] = BPF_ALU64_REG(op, R0, R1);
507 }
508
509 insns[i++] = BPF_EXIT_INSN();
510 self->u.ptr.insns = insns;
511 self->u.ptr.len = len;
512 BUG_ON(i != len);
513
514 return 0;
515}
516
517/* Branch taken by runtime decision */
518static int bpf_fill_max_jmp_taken(struct bpf_test *self)
519{
520 return __bpf_fill_max_jmp(self, BPF_JEQ, 1);
521}
522
523/* Branch not taken by runtime decision */
524static int bpf_fill_max_jmp_not_taken(struct bpf_test *self)
525{
526 return __bpf_fill_max_jmp(self, BPF_JEQ, 0);
527}
528
529/* Branch always taken, known at JIT time */
530static int bpf_fill_max_jmp_always_taken(struct bpf_test *self)
531{
532 return __bpf_fill_max_jmp(self, BPF_JGE, 0);
533}
534
535/* Branch never taken, known at JIT time */
536static int bpf_fill_max_jmp_never_taken(struct bpf_test *self)
537{
538 return __bpf_fill_max_jmp(self, BPF_JLT, 0);
539}
540
Johan Almbladhe42fc3c2021-10-01 15:03:44 +0200541/* ALU result computation used in tests */
542static bool __bpf_alu_result(u64 *res, u64 v1, u64 v2, u8 op)
543{
544 *res = 0;
545 switch (op) {
546 case BPF_MOV:
547 *res = v2;
548 break;
549 case BPF_AND:
550 *res = v1 & v2;
551 break;
552 case BPF_OR:
553 *res = v1 | v2;
554 break;
555 case BPF_XOR:
556 *res = v1 ^ v2;
557 break;
558 case BPF_LSH:
559 *res = v1 << v2;
560 break;
561 case BPF_RSH:
562 *res = v1 >> v2;
563 break;
564 case BPF_ARSH:
565 *res = v1 >> v2;
566 if (v2 > 0 && v1 > S64_MAX)
567 *res |= ~0ULL << (64 - v2);
568 break;
569 case BPF_ADD:
570 *res = v1 + v2;
571 break;
572 case BPF_SUB:
573 *res = v1 - v2;
574 break;
575 case BPF_MUL:
576 *res = v1 * v2;
577 break;
578 case BPF_DIV:
579 if (v2 == 0)
580 return false;
581 *res = div64_u64(v1, v2);
582 break;
583 case BPF_MOD:
584 if (v2 == 0)
585 return false;
586 div64_u64_rem(v1, v2, res);
587 break;
588 }
589 return true;
590}
591
Johan Almbladh68c956f2021-09-14 11:18:31 +0200592/* Test an ALU shift operation for all valid shift values */
593static int __bpf_fill_alu_shift(struct bpf_test *self, u8 op,
594 u8 mode, bool alu32)
595{
596 static const s64 regs[] = {
597 0x0123456789abcdefLL, /* dword > 0, word < 0 */
598 0xfedcba9876543210LL, /* dowrd < 0, word > 0 */
599 0xfedcba0198765432LL, /* dowrd < 0, word < 0 */
600 0x0123458967abcdefLL, /* dword > 0, word > 0 */
601 };
602 int bits = alu32 ? 32 : 64;
603 int len = (2 + 7 * bits) * ARRAY_SIZE(regs) + 3;
604 struct bpf_insn *insn;
605 int imm, k;
606 int i = 0;
607
608 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
609 if (!insn)
610 return -ENOMEM;
611
612 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0);
613
614 for (k = 0; k < ARRAY_SIZE(regs); k++) {
615 s64 reg = regs[k];
616
617 i += __bpf_ld_imm64(&insn[i], R3, reg);
618
619 for (imm = 0; imm < bits; imm++) {
620 u64 val;
621
622 /* Perform operation */
623 insn[i++] = BPF_ALU64_REG(BPF_MOV, R1, R3);
624 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R2, imm);
625 if (alu32) {
626 if (mode == BPF_K)
627 insn[i++] = BPF_ALU32_IMM(op, R1, imm);
628 else
629 insn[i++] = BPF_ALU32_REG(op, R1, R2);
Johan Almbladhe42fc3c2021-10-01 15:03:44 +0200630
631 if (op == BPF_ARSH)
632 reg = (s32)reg;
633 else
634 reg = (u32)reg;
635 __bpf_alu_result(&val, reg, imm, op);
636 val = (u32)val;
Johan Almbladh68c956f2021-09-14 11:18:31 +0200637 } else {
638 if (mode == BPF_K)
639 insn[i++] = BPF_ALU64_IMM(op, R1, imm);
640 else
641 insn[i++] = BPF_ALU64_REG(op, R1, R2);
Johan Almbladhe42fc3c2021-10-01 15:03:44 +0200642 __bpf_alu_result(&val, reg, imm, op);
Johan Almbladh68c956f2021-09-14 11:18:31 +0200643 }
644
645 /*
646 * When debugging a JIT that fails this test, one
647 * can write the immediate value to R0 here to find
648 * out which operand values that fail.
649 */
650
651 /* Load reference and check the result */
652 i += __bpf_ld_imm64(&insn[i], R4, val);
653 insn[i++] = BPF_JMP_REG(BPF_JEQ, R1, R4, 1);
654 insn[i++] = BPF_EXIT_INSN();
655 }
656 }
657
658 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1);
659 insn[i++] = BPF_EXIT_INSN();
660
661 self->u.ptr.insns = insn;
662 self->u.ptr.len = len;
Johan Almbladh68813602021-10-01 15:03:47 +0200663 BUG_ON(i != len);
Johan Almbladh68c956f2021-09-14 11:18:31 +0200664
665 return 0;
666}
667
Johan Almbladh68813602021-10-01 15:03:47 +0200668static int bpf_fill_alu64_lsh_imm(struct bpf_test *self)
Johan Almbladh68c956f2021-09-14 11:18:31 +0200669{
670 return __bpf_fill_alu_shift(self, BPF_LSH, BPF_K, false);
671}
672
Johan Almbladh68813602021-10-01 15:03:47 +0200673static int bpf_fill_alu64_rsh_imm(struct bpf_test *self)
Johan Almbladh68c956f2021-09-14 11:18:31 +0200674{
675 return __bpf_fill_alu_shift(self, BPF_RSH, BPF_K, false);
676}
677
Johan Almbladh68813602021-10-01 15:03:47 +0200678static int bpf_fill_alu64_arsh_imm(struct bpf_test *self)
Johan Almbladh68c956f2021-09-14 11:18:31 +0200679{
680 return __bpf_fill_alu_shift(self, BPF_ARSH, BPF_K, false);
681}
682
Johan Almbladh68813602021-10-01 15:03:47 +0200683static int bpf_fill_alu64_lsh_reg(struct bpf_test *self)
Johan Almbladh68c956f2021-09-14 11:18:31 +0200684{
685 return __bpf_fill_alu_shift(self, BPF_LSH, BPF_X, false);
686}
687
Johan Almbladh68813602021-10-01 15:03:47 +0200688static int bpf_fill_alu64_rsh_reg(struct bpf_test *self)
Johan Almbladh68c956f2021-09-14 11:18:31 +0200689{
690 return __bpf_fill_alu_shift(self, BPF_RSH, BPF_X, false);
691}
692
Johan Almbladh68813602021-10-01 15:03:47 +0200693static int bpf_fill_alu64_arsh_reg(struct bpf_test *self)
Johan Almbladh68c956f2021-09-14 11:18:31 +0200694{
695 return __bpf_fill_alu_shift(self, BPF_ARSH, BPF_X, false);
696}
697
698static int bpf_fill_alu32_lsh_imm(struct bpf_test *self)
699{
700 return __bpf_fill_alu_shift(self, BPF_LSH, BPF_K, true);
701}
702
703static int bpf_fill_alu32_rsh_imm(struct bpf_test *self)
704{
705 return __bpf_fill_alu_shift(self, BPF_RSH, BPF_K, true);
706}
707
708static int bpf_fill_alu32_arsh_imm(struct bpf_test *self)
709{
710 return __bpf_fill_alu_shift(self, BPF_ARSH, BPF_K, true);
711}
712
713static int bpf_fill_alu32_lsh_reg(struct bpf_test *self)
714{
715 return __bpf_fill_alu_shift(self, BPF_LSH, BPF_X, true);
716}
717
718static int bpf_fill_alu32_rsh_reg(struct bpf_test *self)
719{
720 return __bpf_fill_alu_shift(self, BPF_RSH, BPF_X, true);
721}
722
723static int bpf_fill_alu32_arsh_reg(struct bpf_test *self)
724{
725 return __bpf_fill_alu_shift(self, BPF_ARSH, BPF_X, true);
726}
727
Johan Almbladh9298e632021-09-14 11:18:32 +0200728/*
Johan Almbladh68813602021-10-01 15:03:47 +0200729 * Test an ALU register shift operation for all valid shift values
730 * for the case when the source and destination are the same.
731 */
732static int __bpf_fill_alu_shift_same_reg(struct bpf_test *self, u8 op,
733 bool alu32)
734{
735 int bits = alu32 ? 32 : 64;
736 int len = 3 + 6 * bits;
737 struct bpf_insn *insn;
738 int i = 0;
739 u64 val;
740
741 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
742 if (!insn)
743 return -ENOMEM;
744
745 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0);
746
747 for (val = 0; val < bits; val++) {
748 u64 res;
749
750 /* Perform operation */
751 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R1, val);
752 if (alu32)
753 insn[i++] = BPF_ALU32_REG(op, R1, R1);
754 else
755 insn[i++] = BPF_ALU64_REG(op, R1, R1);
756
757 /* Compute the reference result */
758 __bpf_alu_result(&res, val, val, op);
759 if (alu32)
760 res = (u32)res;
761 i += __bpf_ld_imm64(&insn[i], R2, res);
762
763 /* Check the actual result */
764 insn[i++] = BPF_JMP_REG(BPF_JEQ, R1, R2, 1);
765 insn[i++] = BPF_EXIT_INSN();
766 }
767
768 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1);
769 insn[i++] = BPF_EXIT_INSN();
770
771 self->u.ptr.insns = insn;
772 self->u.ptr.len = len;
773 BUG_ON(i != len);
774
775 return 0;
776}
777
778static int bpf_fill_alu64_lsh_same_reg(struct bpf_test *self)
779{
780 return __bpf_fill_alu_shift_same_reg(self, BPF_LSH, false);
781}
782
783static int bpf_fill_alu64_rsh_same_reg(struct bpf_test *self)
784{
785 return __bpf_fill_alu_shift_same_reg(self, BPF_RSH, false);
786}
787
788static int bpf_fill_alu64_arsh_same_reg(struct bpf_test *self)
789{
790 return __bpf_fill_alu_shift_same_reg(self, BPF_ARSH, false);
791}
792
793static int bpf_fill_alu32_lsh_same_reg(struct bpf_test *self)
794{
795 return __bpf_fill_alu_shift_same_reg(self, BPF_LSH, true);
796}
797
798static int bpf_fill_alu32_rsh_same_reg(struct bpf_test *self)
799{
800 return __bpf_fill_alu_shift_same_reg(self, BPF_RSH, true);
801}
802
803static int bpf_fill_alu32_arsh_same_reg(struct bpf_test *self)
804{
805 return __bpf_fill_alu_shift_same_reg(self, BPF_ARSH, true);
806}
807
808/*
Johan Almbladh9298e632021-09-14 11:18:32 +0200809 * Common operand pattern generator for exhaustive power-of-two magnitudes
810 * tests. The block size parameters can be adjusted to increase/reduce the
811 * number of combinatons tested and thereby execution speed and memory
812 * footprint.
813 */
814
815static inline s64 value(int msb, int delta, int sign)
816{
817 return sign * (1LL << msb) + delta;
818}
819
820static int __bpf_fill_pattern(struct bpf_test *self, void *arg,
821 int dbits, int sbits, int block1, int block2,
822 int (*emit)(struct bpf_test*, void*,
823 struct bpf_insn*, s64, s64))
824{
825 static const int sgn[][2] = {{1, 1}, {1, -1}, {-1, 1}, {-1, -1}};
826 struct bpf_insn *insns;
827 int di, si, bt, db, sb;
828 int count, len, k;
829 int extra = 1 + 2;
830 int i = 0;
831
832 /* Total number of iterations for the two pattern */
833 count = (dbits - 1) * (sbits - 1) * block1 * block1 * ARRAY_SIZE(sgn);
834 count += (max(dbits, sbits) - 1) * block2 * block2 * ARRAY_SIZE(sgn);
835
836 /* Compute the maximum number of insns and allocate the buffer */
837 len = extra + count * (*emit)(self, arg, NULL, 0, 0);
838 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL);
839 if (!insns)
840 return -ENOMEM;
841
842 /* Add head instruction(s) */
843 insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0);
844
845 /*
846 * Pattern 1: all combinations of power-of-two magnitudes and sign,
847 * and with a block of contiguous values around each magnitude.
848 */
849 for (di = 0; di < dbits - 1; di++) /* Dst magnitudes */
850 for (si = 0; si < sbits - 1; si++) /* Src magnitudes */
851 for (k = 0; k < ARRAY_SIZE(sgn); k++) /* Sign combos */
852 for (db = -(block1 / 2);
853 db < (block1 + 1) / 2; db++)
854 for (sb = -(block1 / 2);
855 sb < (block1 + 1) / 2; sb++) {
856 s64 dst, src;
857
858 dst = value(di, db, sgn[k][0]);
859 src = value(si, sb, sgn[k][1]);
860 i += (*emit)(self, arg,
861 &insns[i],
862 dst, src);
863 }
864 /*
865 * Pattern 2: all combinations for a larger block of values
866 * for each power-of-two magnitude and sign, where the magnitude is
867 * the same for both operands.
868 */
869 for (bt = 0; bt < max(dbits, sbits) - 1; bt++) /* Magnitude */
870 for (k = 0; k < ARRAY_SIZE(sgn); k++) /* Sign combos */
871 for (db = -(block2 / 2); db < (block2 + 1) / 2; db++)
872 for (sb = -(block2 / 2);
873 sb < (block2 + 1) / 2; sb++) {
874 s64 dst, src;
875
876 dst = value(bt % dbits, db, sgn[k][0]);
877 src = value(bt % sbits, sb, sgn[k][1]);
878 i += (*emit)(self, arg, &insns[i],
879 dst, src);
880 }
881
882 /* Append tail instructions */
883 insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1);
884 insns[i++] = BPF_EXIT_INSN();
885 BUG_ON(i > len);
886
887 self->u.ptr.insns = insns;
888 self->u.ptr.len = i;
889
890 return 0;
891}
892
893/*
894 * Block size parameters used in pattern tests below. une as needed to
895 * increase/reduce the number combinations tested, see following examples.
896 * block values per operand MSB
897 * ----------------------------------------
898 * 0 none
899 * 1 (1 << MSB)
900 * 2 (1 << MSB) + [-1, 0]
901 * 3 (1 << MSB) + [-1, 0, 1]
902 */
903#define PATTERN_BLOCK1 1
904#define PATTERN_BLOCK2 5
905
906/* Number of test runs for a pattern test */
907#define NR_PATTERN_RUNS 1
908
909/*
910 * Exhaustive tests of ALU operations for all combinations of power-of-two
911 * magnitudes of the operands, both for positive and negative values. The
Johan Almbladhf68e8ef2021-10-01 15:03:41 +0200912 * test is designed to verify e.g. the ALU and ALU64 operations for JITs that
Johan Almbladh9298e632021-09-14 11:18:32 +0200913 * emit different code depending on the magnitude of the immediate value.
914 */
Johan Almbladh9298e632021-09-14 11:18:32 +0200915static int __bpf_emit_alu64_imm(struct bpf_test *self, void *arg,
916 struct bpf_insn *insns, s64 dst, s64 imm)
917{
918 int op = *(int *)arg;
919 int i = 0;
920 u64 res;
921
922 if (!insns)
923 return 7;
924
925 if (__bpf_alu_result(&res, dst, (s32)imm, op)) {
926 i += __bpf_ld_imm64(&insns[i], R1, dst);
927 i += __bpf_ld_imm64(&insns[i], R3, res);
928 insns[i++] = BPF_ALU64_IMM(op, R1, imm);
929 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1);
930 insns[i++] = BPF_EXIT_INSN();
931 }
932
933 return i;
934}
935
936static int __bpf_emit_alu32_imm(struct bpf_test *self, void *arg,
937 struct bpf_insn *insns, s64 dst, s64 imm)
938{
939 int op = *(int *)arg;
940 int i = 0;
941 u64 res;
942
943 if (!insns)
944 return 7;
945
946 if (__bpf_alu_result(&res, (u32)dst, (u32)imm, op)) {
947 i += __bpf_ld_imm64(&insns[i], R1, dst);
948 i += __bpf_ld_imm64(&insns[i], R3, (u32)res);
949 insns[i++] = BPF_ALU32_IMM(op, R1, imm);
950 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1);
951 insns[i++] = BPF_EXIT_INSN();
952 }
953
954 return i;
955}
956
957static int __bpf_emit_alu64_reg(struct bpf_test *self, void *arg,
958 struct bpf_insn *insns, s64 dst, s64 src)
959{
960 int op = *(int *)arg;
961 int i = 0;
962 u64 res;
963
964 if (!insns)
965 return 9;
966
967 if (__bpf_alu_result(&res, dst, src, op)) {
968 i += __bpf_ld_imm64(&insns[i], R1, dst);
969 i += __bpf_ld_imm64(&insns[i], R2, src);
970 i += __bpf_ld_imm64(&insns[i], R3, res);
971 insns[i++] = BPF_ALU64_REG(op, R1, R2);
972 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1);
973 insns[i++] = BPF_EXIT_INSN();
974 }
975
976 return i;
977}
978
979static int __bpf_emit_alu32_reg(struct bpf_test *self, void *arg,
980 struct bpf_insn *insns, s64 dst, s64 src)
981{
982 int op = *(int *)arg;
983 int i = 0;
984 u64 res;
985
986 if (!insns)
987 return 9;
988
989 if (__bpf_alu_result(&res, (u32)dst, (u32)src, op)) {
990 i += __bpf_ld_imm64(&insns[i], R1, dst);
991 i += __bpf_ld_imm64(&insns[i], R2, src);
992 i += __bpf_ld_imm64(&insns[i], R3, (u32)res);
993 insns[i++] = BPF_ALU32_REG(op, R1, R2);
994 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1);
995 insns[i++] = BPF_EXIT_INSN();
996 }
997
998 return i;
999}
1000
1001static int __bpf_fill_alu64_imm(struct bpf_test *self, int op)
1002{
1003 return __bpf_fill_pattern(self, &op, 64, 32,
1004 PATTERN_BLOCK1, PATTERN_BLOCK2,
1005 &__bpf_emit_alu64_imm);
1006}
1007
1008static int __bpf_fill_alu32_imm(struct bpf_test *self, int op)
1009{
1010 return __bpf_fill_pattern(self, &op, 64, 32,
1011 PATTERN_BLOCK1, PATTERN_BLOCK2,
1012 &__bpf_emit_alu32_imm);
1013}
1014
1015static int __bpf_fill_alu64_reg(struct bpf_test *self, int op)
1016{
1017 return __bpf_fill_pattern(self, &op, 64, 64,
1018 PATTERN_BLOCK1, PATTERN_BLOCK2,
1019 &__bpf_emit_alu64_reg);
1020}
1021
1022static int __bpf_fill_alu32_reg(struct bpf_test *self, int op)
1023{
1024 return __bpf_fill_pattern(self, &op, 64, 64,
1025 PATTERN_BLOCK1, PATTERN_BLOCK2,
1026 &__bpf_emit_alu32_reg);
1027}
1028
1029/* ALU64 immediate operations */
1030static int bpf_fill_alu64_mov_imm(struct bpf_test *self)
1031{
1032 return __bpf_fill_alu64_imm(self, BPF_MOV);
1033}
1034
1035static int bpf_fill_alu64_and_imm(struct bpf_test *self)
1036{
1037 return __bpf_fill_alu64_imm(self, BPF_AND);
1038}
1039
1040static int bpf_fill_alu64_or_imm(struct bpf_test *self)
1041{
1042 return __bpf_fill_alu64_imm(self, BPF_OR);
1043}
1044
1045static int bpf_fill_alu64_xor_imm(struct bpf_test *self)
1046{
1047 return __bpf_fill_alu64_imm(self, BPF_XOR);
1048}
1049
1050static int bpf_fill_alu64_add_imm(struct bpf_test *self)
1051{
1052 return __bpf_fill_alu64_imm(self, BPF_ADD);
1053}
1054
1055static int bpf_fill_alu64_sub_imm(struct bpf_test *self)
1056{
1057 return __bpf_fill_alu64_imm(self, BPF_SUB);
1058}
1059
1060static int bpf_fill_alu64_mul_imm(struct bpf_test *self)
1061{
1062 return __bpf_fill_alu64_imm(self, BPF_MUL);
1063}
1064
1065static int bpf_fill_alu64_div_imm(struct bpf_test *self)
1066{
1067 return __bpf_fill_alu64_imm(self, BPF_DIV);
1068}
1069
1070static int bpf_fill_alu64_mod_imm(struct bpf_test *self)
1071{
1072 return __bpf_fill_alu64_imm(self, BPF_MOD);
1073}
1074
1075/* ALU32 immediate operations */
1076static int bpf_fill_alu32_mov_imm(struct bpf_test *self)
1077{
1078 return __bpf_fill_alu32_imm(self, BPF_MOV);
1079}
1080
1081static int bpf_fill_alu32_and_imm(struct bpf_test *self)
1082{
1083 return __bpf_fill_alu32_imm(self, BPF_AND);
1084}
1085
1086static int bpf_fill_alu32_or_imm(struct bpf_test *self)
1087{
1088 return __bpf_fill_alu32_imm(self, BPF_OR);
1089}
1090
1091static int bpf_fill_alu32_xor_imm(struct bpf_test *self)
1092{
1093 return __bpf_fill_alu32_imm(self, BPF_XOR);
1094}
1095
1096static int bpf_fill_alu32_add_imm(struct bpf_test *self)
1097{
1098 return __bpf_fill_alu32_imm(self, BPF_ADD);
1099}
1100
1101static int bpf_fill_alu32_sub_imm(struct bpf_test *self)
1102{
1103 return __bpf_fill_alu32_imm(self, BPF_SUB);
1104}
1105
1106static int bpf_fill_alu32_mul_imm(struct bpf_test *self)
1107{
1108 return __bpf_fill_alu32_imm(self, BPF_MUL);
1109}
1110
1111static int bpf_fill_alu32_div_imm(struct bpf_test *self)
1112{
1113 return __bpf_fill_alu32_imm(self, BPF_DIV);
1114}
1115
1116static int bpf_fill_alu32_mod_imm(struct bpf_test *self)
1117{
1118 return __bpf_fill_alu32_imm(self, BPF_MOD);
1119}
1120
1121/* ALU64 register operations */
1122static int bpf_fill_alu64_mov_reg(struct bpf_test *self)
1123{
1124 return __bpf_fill_alu64_reg(self, BPF_MOV);
1125}
1126
1127static int bpf_fill_alu64_and_reg(struct bpf_test *self)
1128{
1129 return __bpf_fill_alu64_reg(self, BPF_AND);
1130}
1131
1132static int bpf_fill_alu64_or_reg(struct bpf_test *self)
1133{
1134 return __bpf_fill_alu64_reg(self, BPF_OR);
1135}
1136
1137static int bpf_fill_alu64_xor_reg(struct bpf_test *self)
1138{
1139 return __bpf_fill_alu64_reg(self, BPF_XOR);
1140}
1141
1142static int bpf_fill_alu64_add_reg(struct bpf_test *self)
1143{
1144 return __bpf_fill_alu64_reg(self, BPF_ADD);
1145}
1146
1147static int bpf_fill_alu64_sub_reg(struct bpf_test *self)
1148{
1149 return __bpf_fill_alu64_reg(self, BPF_SUB);
1150}
1151
1152static int bpf_fill_alu64_mul_reg(struct bpf_test *self)
1153{
1154 return __bpf_fill_alu64_reg(self, BPF_MUL);
1155}
1156
1157static int bpf_fill_alu64_div_reg(struct bpf_test *self)
1158{
1159 return __bpf_fill_alu64_reg(self, BPF_DIV);
1160}
1161
1162static int bpf_fill_alu64_mod_reg(struct bpf_test *self)
1163{
1164 return __bpf_fill_alu64_reg(self, BPF_MOD);
1165}
1166
1167/* ALU32 register operations */
1168static int bpf_fill_alu32_mov_reg(struct bpf_test *self)
1169{
1170 return __bpf_fill_alu32_reg(self, BPF_MOV);
1171}
1172
1173static int bpf_fill_alu32_and_reg(struct bpf_test *self)
1174{
1175 return __bpf_fill_alu32_reg(self, BPF_AND);
1176}
1177
1178static int bpf_fill_alu32_or_reg(struct bpf_test *self)
1179{
1180 return __bpf_fill_alu32_reg(self, BPF_OR);
1181}
1182
1183static int bpf_fill_alu32_xor_reg(struct bpf_test *self)
1184{
1185 return __bpf_fill_alu32_reg(self, BPF_XOR);
1186}
1187
1188static int bpf_fill_alu32_add_reg(struct bpf_test *self)
1189{
1190 return __bpf_fill_alu32_reg(self, BPF_ADD);
1191}
1192
1193static int bpf_fill_alu32_sub_reg(struct bpf_test *self)
1194{
1195 return __bpf_fill_alu32_reg(self, BPF_SUB);
1196}
1197
1198static int bpf_fill_alu32_mul_reg(struct bpf_test *self)
1199{
1200 return __bpf_fill_alu32_reg(self, BPF_MUL);
1201}
1202
1203static int bpf_fill_alu32_div_reg(struct bpf_test *self)
1204{
1205 return __bpf_fill_alu32_reg(self, BPF_DIV);
1206}
1207
1208static int bpf_fill_alu32_mod_reg(struct bpf_test *self)
1209{
1210 return __bpf_fill_alu32_reg(self, BPF_MOD);
1211}
1212
Johan Almbladh2e807612021-09-14 11:18:35 +02001213/*
Johan Almbladhdaed6082021-10-01 15:03:45 +02001214 * Test JITs that implement complex ALU operations as function
1215 * calls, and must re-arrange operands for argument passing.
1216 */
1217static int __bpf_fill_alu_imm_regs(struct bpf_test *self, u8 op, bool alu32)
1218{
1219 int len = 2 + 10 * 10;
1220 struct bpf_insn *insns;
1221 u64 dst, res;
1222 int i = 0;
1223 u32 imm;
1224 int rd;
1225
1226 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL);
1227 if (!insns)
1228 return -ENOMEM;
1229
1230 /* Operand and result values according to operation */
1231 if (alu32)
1232 dst = 0x76543210U;
1233 else
1234 dst = 0x7edcba9876543210ULL;
1235 imm = 0x01234567U;
1236
1237 if (op == BPF_LSH || op == BPF_RSH || op == BPF_ARSH)
1238 imm &= 31;
1239
1240 __bpf_alu_result(&res, dst, imm, op);
1241
1242 if (alu32)
1243 res = (u32)res;
1244
1245 /* Check all operand registers */
1246 for (rd = R0; rd <= R9; rd++) {
1247 i += __bpf_ld_imm64(&insns[i], rd, dst);
1248
1249 if (alu32)
1250 insns[i++] = BPF_ALU32_IMM(op, rd, imm);
1251 else
1252 insns[i++] = BPF_ALU64_IMM(op, rd, imm);
1253
1254 insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, res, 2);
1255 insns[i++] = BPF_MOV64_IMM(R0, __LINE__);
1256 insns[i++] = BPF_EXIT_INSN();
1257
1258 insns[i++] = BPF_ALU64_IMM(BPF_RSH, rd, 32);
1259 insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, res >> 32, 2);
1260 insns[i++] = BPF_MOV64_IMM(R0, __LINE__);
1261 insns[i++] = BPF_EXIT_INSN();
1262 }
1263
1264 insns[i++] = BPF_MOV64_IMM(R0, 1);
1265 insns[i++] = BPF_EXIT_INSN();
1266
1267 self->u.ptr.insns = insns;
1268 self->u.ptr.len = len;
1269 BUG_ON(i != len);
1270
1271 return 0;
1272}
1273
1274/* ALU64 K registers */
1275static int bpf_fill_alu64_mov_imm_regs(struct bpf_test *self)
1276{
1277 return __bpf_fill_alu_imm_regs(self, BPF_MOV, false);
1278}
1279
1280static int bpf_fill_alu64_and_imm_regs(struct bpf_test *self)
1281{
1282 return __bpf_fill_alu_imm_regs(self, BPF_AND, false);
1283}
1284
1285static int bpf_fill_alu64_or_imm_regs(struct bpf_test *self)
1286{
1287 return __bpf_fill_alu_imm_regs(self, BPF_OR, false);
1288}
1289
1290static int bpf_fill_alu64_xor_imm_regs(struct bpf_test *self)
1291{
1292 return __bpf_fill_alu_imm_regs(self, BPF_XOR, false);
1293}
1294
1295static int bpf_fill_alu64_lsh_imm_regs(struct bpf_test *self)
1296{
1297 return __bpf_fill_alu_imm_regs(self, BPF_LSH, false);
1298}
1299
1300static int bpf_fill_alu64_rsh_imm_regs(struct bpf_test *self)
1301{
1302 return __bpf_fill_alu_imm_regs(self, BPF_RSH, false);
1303}
1304
1305static int bpf_fill_alu64_arsh_imm_regs(struct bpf_test *self)
1306{
1307 return __bpf_fill_alu_imm_regs(self, BPF_ARSH, false);
1308}
1309
1310static int bpf_fill_alu64_add_imm_regs(struct bpf_test *self)
1311{
1312 return __bpf_fill_alu_imm_regs(self, BPF_ADD, false);
1313}
1314
1315static int bpf_fill_alu64_sub_imm_regs(struct bpf_test *self)
1316{
1317 return __bpf_fill_alu_imm_regs(self, BPF_SUB, false);
1318}
1319
1320static int bpf_fill_alu64_mul_imm_regs(struct bpf_test *self)
1321{
1322 return __bpf_fill_alu_imm_regs(self, BPF_MUL, false);
1323}
1324
1325static int bpf_fill_alu64_div_imm_regs(struct bpf_test *self)
1326{
1327 return __bpf_fill_alu_imm_regs(self, BPF_DIV, false);
1328}
1329
1330static int bpf_fill_alu64_mod_imm_regs(struct bpf_test *self)
1331{
1332 return __bpf_fill_alu_imm_regs(self, BPF_MOD, false);
1333}
1334
1335/* ALU32 K registers */
1336static int bpf_fill_alu32_mov_imm_regs(struct bpf_test *self)
1337{
1338 return __bpf_fill_alu_imm_regs(self, BPF_MOV, true);
1339}
1340
1341static int bpf_fill_alu32_and_imm_regs(struct bpf_test *self)
1342{
1343 return __bpf_fill_alu_imm_regs(self, BPF_AND, true);
1344}
1345
1346static int bpf_fill_alu32_or_imm_regs(struct bpf_test *self)
1347{
1348 return __bpf_fill_alu_imm_regs(self, BPF_OR, true);
1349}
1350
1351static int bpf_fill_alu32_xor_imm_regs(struct bpf_test *self)
1352{
1353 return __bpf_fill_alu_imm_regs(self, BPF_XOR, true);
1354}
1355
1356static int bpf_fill_alu32_lsh_imm_regs(struct bpf_test *self)
1357{
1358 return __bpf_fill_alu_imm_regs(self, BPF_LSH, true);
1359}
1360
1361static int bpf_fill_alu32_rsh_imm_regs(struct bpf_test *self)
1362{
1363 return __bpf_fill_alu_imm_regs(self, BPF_RSH, true);
1364}
1365
1366static int bpf_fill_alu32_arsh_imm_regs(struct bpf_test *self)
1367{
1368 return __bpf_fill_alu_imm_regs(self, BPF_ARSH, true);
1369}
1370
1371static int bpf_fill_alu32_add_imm_regs(struct bpf_test *self)
1372{
1373 return __bpf_fill_alu_imm_regs(self, BPF_ADD, true);
1374}
1375
1376static int bpf_fill_alu32_sub_imm_regs(struct bpf_test *self)
1377{
1378 return __bpf_fill_alu_imm_regs(self, BPF_SUB, true);
1379}
1380
1381static int bpf_fill_alu32_mul_imm_regs(struct bpf_test *self)
1382{
1383 return __bpf_fill_alu_imm_regs(self, BPF_MUL, true);
1384}
1385
1386static int bpf_fill_alu32_div_imm_regs(struct bpf_test *self)
1387{
1388 return __bpf_fill_alu_imm_regs(self, BPF_DIV, true);
1389}
1390
1391static int bpf_fill_alu32_mod_imm_regs(struct bpf_test *self)
1392{
1393 return __bpf_fill_alu_imm_regs(self, BPF_MOD, true);
1394}
1395
1396/*
1397 * Test JITs that implement complex ALU operations as function
1398 * calls, and must re-arrange operands for argument passing.
1399 */
1400static int __bpf_fill_alu_reg_pairs(struct bpf_test *self, u8 op, bool alu32)
1401{
1402 int len = 2 + 10 * 10 * 12;
1403 u64 dst, src, res, same;
1404 struct bpf_insn *insns;
1405 int rd, rs;
1406 int i = 0;
1407
1408 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL);
1409 if (!insns)
1410 return -ENOMEM;
1411
1412 /* Operand and result values according to operation */
1413 if (alu32) {
1414 dst = 0x76543210U;
1415 src = 0x01234567U;
1416 } else {
1417 dst = 0x7edcba9876543210ULL;
1418 src = 0x0123456789abcdefULL;
1419 }
1420
1421 if (op == BPF_LSH || op == BPF_RSH || op == BPF_ARSH)
1422 src &= 31;
1423
1424 __bpf_alu_result(&res, dst, src, op);
1425 __bpf_alu_result(&same, src, src, op);
1426
1427 if (alu32) {
1428 res = (u32)res;
1429 same = (u32)same;
1430 }
1431
1432 /* Check all combinations of operand registers */
1433 for (rd = R0; rd <= R9; rd++) {
1434 for (rs = R0; rs <= R9; rs++) {
1435 u64 val = rd == rs ? same : res;
1436
1437 i += __bpf_ld_imm64(&insns[i], rd, dst);
1438 i += __bpf_ld_imm64(&insns[i], rs, src);
1439
1440 if (alu32)
1441 insns[i++] = BPF_ALU32_REG(op, rd, rs);
1442 else
1443 insns[i++] = BPF_ALU64_REG(op, rd, rs);
1444
1445 insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, val, 2);
1446 insns[i++] = BPF_MOV64_IMM(R0, __LINE__);
1447 insns[i++] = BPF_EXIT_INSN();
1448
1449 insns[i++] = BPF_ALU64_IMM(BPF_RSH, rd, 32);
1450 insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, val >> 32, 2);
1451 insns[i++] = BPF_MOV64_IMM(R0, __LINE__);
1452 insns[i++] = BPF_EXIT_INSN();
1453 }
1454 }
1455
1456 insns[i++] = BPF_MOV64_IMM(R0, 1);
1457 insns[i++] = BPF_EXIT_INSN();
1458
1459 self->u.ptr.insns = insns;
1460 self->u.ptr.len = len;
1461 BUG_ON(i != len);
1462
1463 return 0;
1464}
1465
1466/* ALU64 X register combinations */
1467static int bpf_fill_alu64_mov_reg_pairs(struct bpf_test *self)
1468{
1469 return __bpf_fill_alu_reg_pairs(self, BPF_MOV, false);
1470}
1471
1472static int bpf_fill_alu64_and_reg_pairs(struct bpf_test *self)
1473{
1474 return __bpf_fill_alu_reg_pairs(self, BPF_AND, false);
1475}
1476
1477static int bpf_fill_alu64_or_reg_pairs(struct bpf_test *self)
1478{
1479 return __bpf_fill_alu_reg_pairs(self, BPF_OR, false);
1480}
1481
1482static int bpf_fill_alu64_xor_reg_pairs(struct bpf_test *self)
1483{
1484 return __bpf_fill_alu_reg_pairs(self, BPF_XOR, false);
1485}
1486
1487static int bpf_fill_alu64_lsh_reg_pairs(struct bpf_test *self)
1488{
1489 return __bpf_fill_alu_reg_pairs(self, BPF_LSH, false);
1490}
1491
1492static int bpf_fill_alu64_rsh_reg_pairs(struct bpf_test *self)
1493{
1494 return __bpf_fill_alu_reg_pairs(self, BPF_RSH, false);
1495}
1496
1497static int bpf_fill_alu64_arsh_reg_pairs(struct bpf_test *self)
1498{
1499 return __bpf_fill_alu_reg_pairs(self, BPF_ARSH, false);
1500}
1501
1502static int bpf_fill_alu64_add_reg_pairs(struct bpf_test *self)
1503{
1504 return __bpf_fill_alu_reg_pairs(self, BPF_ADD, false);
1505}
1506
1507static int bpf_fill_alu64_sub_reg_pairs(struct bpf_test *self)
1508{
1509 return __bpf_fill_alu_reg_pairs(self, BPF_SUB, false);
1510}
1511
1512static int bpf_fill_alu64_mul_reg_pairs(struct bpf_test *self)
1513{
1514 return __bpf_fill_alu_reg_pairs(self, BPF_MUL, false);
1515}
1516
1517static int bpf_fill_alu64_div_reg_pairs(struct bpf_test *self)
1518{
1519 return __bpf_fill_alu_reg_pairs(self, BPF_DIV, false);
1520}
1521
1522static int bpf_fill_alu64_mod_reg_pairs(struct bpf_test *self)
1523{
1524 return __bpf_fill_alu_reg_pairs(self, BPF_MOD, false);
1525}
1526
1527/* ALU32 X register combinations */
1528static int bpf_fill_alu32_mov_reg_pairs(struct bpf_test *self)
1529{
1530 return __bpf_fill_alu_reg_pairs(self, BPF_MOV, true);
1531}
1532
1533static int bpf_fill_alu32_and_reg_pairs(struct bpf_test *self)
1534{
1535 return __bpf_fill_alu_reg_pairs(self, BPF_AND, true);
1536}
1537
1538static int bpf_fill_alu32_or_reg_pairs(struct bpf_test *self)
1539{
1540 return __bpf_fill_alu_reg_pairs(self, BPF_OR, true);
1541}
1542
1543static int bpf_fill_alu32_xor_reg_pairs(struct bpf_test *self)
1544{
1545 return __bpf_fill_alu_reg_pairs(self, BPF_XOR, true);
1546}
1547
1548static int bpf_fill_alu32_lsh_reg_pairs(struct bpf_test *self)
1549{
1550 return __bpf_fill_alu_reg_pairs(self, BPF_LSH, true);
1551}
1552
1553static int bpf_fill_alu32_rsh_reg_pairs(struct bpf_test *self)
1554{
1555 return __bpf_fill_alu_reg_pairs(self, BPF_RSH, true);
1556}
1557
1558static int bpf_fill_alu32_arsh_reg_pairs(struct bpf_test *self)
1559{
1560 return __bpf_fill_alu_reg_pairs(self, BPF_ARSH, true);
1561}
1562
1563static int bpf_fill_alu32_add_reg_pairs(struct bpf_test *self)
1564{
1565 return __bpf_fill_alu_reg_pairs(self, BPF_ADD, true);
1566}
1567
1568static int bpf_fill_alu32_sub_reg_pairs(struct bpf_test *self)
1569{
1570 return __bpf_fill_alu_reg_pairs(self, BPF_SUB, true);
1571}
1572
1573static int bpf_fill_alu32_mul_reg_pairs(struct bpf_test *self)
1574{
1575 return __bpf_fill_alu_reg_pairs(self, BPF_MUL, true);
1576}
1577
1578static int bpf_fill_alu32_div_reg_pairs(struct bpf_test *self)
1579{
1580 return __bpf_fill_alu_reg_pairs(self, BPF_DIV, true);
1581}
1582
1583static int bpf_fill_alu32_mod_reg_pairs(struct bpf_test *self)
1584{
1585 return __bpf_fill_alu_reg_pairs(self, BPF_MOD, true);
1586}
1587
1588/*
Johan Almbladhf68e8ef2021-10-01 15:03:41 +02001589 * Exhaustive tests of atomic operations for all power-of-two operand
1590 * magnitudes, both for positive and negative values.
1591 */
1592
1593static int __bpf_emit_atomic64(struct bpf_test *self, void *arg,
1594 struct bpf_insn *insns, s64 dst, s64 src)
1595{
1596 int op = *(int *)arg;
1597 u64 keep, fetch, res;
1598 int i = 0;
1599
1600 if (!insns)
1601 return 21;
1602
1603 switch (op) {
1604 case BPF_XCHG:
1605 res = src;
1606 break;
1607 default:
1608 __bpf_alu_result(&res, dst, src, BPF_OP(op));
1609 }
1610
1611 keep = 0x0123456789abcdefULL;
1612 if (op & BPF_FETCH)
1613 fetch = dst;
1614 else
1615 fetch = src;
1616
1617 i += __bpf_ld_imm64(&insns[i], R0, keep);
1618 i += __bpf_ld_imm64(&insns[i], R1, dst);
1619 i += __bpf_ld_imm64(&insns[i], R2, src);
1620 i += __bpf_ld_imm64(&insns[i], R3, res);
1621 i += __bpf_ld_imm64(&insns[i], R4, fetch);
1622 i += __bpf_ld_imm64(&insns[i], R5, keep);
1623
1624 insns[i++] = BPF_STX_MEM(BPF_DW, R10, R1, -8);
1625 insns[i++] = BPF_ATOMIC_OP(BPF_DW, op, R10, R2, -8);
1626 insns[i++] = BPF_LDX_MEM(BPF_DW, R1, R10, -8);
1627
1628 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1);
1629 insns[i++] = BPF_EXIT_INSN();
1630
1631 insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R4, 1);
1632 insns[i++] = BPF_EXIT_INSN();
1633
1634 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R5, 1);
1635 insns[i++] = BPF_EXIT_INSN();
1636
1637 return i;
1638}
1639
1640static int __bpf_emit_atomic32(struct bpf_test *self, void *arg,
1641 struct bpf_insn *insns, s64 dst, s64 src)
1642{
1643 int op = *(int *)arg;
1644 u64 keep, fetch, res;
1645 int i = 0;
1646
1647 if (!insns)
1648 return 21;
1649
1650 switch (op) {
1651 case BPF_XCHG:
1652 res = src;
1653 break;
1654 default:
1655 __bpf_alu_result(&res, (u32)dst, (u32)src, BPF_OP(op));
1656 }
1657
1658 keep = 0x0123456789abcdefULL;
1659 if (op & BPF_FETCH)
1660 fetch = (u32)dst;
1661 else
1662 fetch = src;
1663
1664 i += __bpf_ld_imm64(&insns[i], R0, keep);
1665 i += __bpf_ld_imm64(&insns[i], R1, (u32)dst);
1666 i += __bpf_ld_imm64(&insns[i], R2, src);
1667 i += __bpf_ld_imm64(&insns[i], R3, (u32)res);
1668 i += __bpf_ld_imm64(&insns[i], R4, fetch);
1669 i += __bpf_ld_imm64(&insns[i], R5, keep);
1670
1671 insns[i++] = BPF_STX_MEM(BPF_W, R10, R1, -4);
1672 insns[i++] = BPF_ATOMIC_OP(BPF_W, op, R10, R2, -4);
1673 insns[i++] = BPF_LDX_MEM(BPF_W, R1, R10, -4);
1674
1675 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1);
1676 insns[i++] = BPF_EXIT_INSN();
1677
1678 insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R4, 1);
1679 insns[i++] = BPF_EXIT_INSN();
1680
1681 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R5, 1);
1682 insns[i++] = BPF_EXIT_INSN();
1683
1684 return i;
1685}
1686
1687static int __bpf_emit_cmpxchg64(struct bpf_test *self, void *arg,
1688 struct bpf_insn *insns, s64 dst, s64 src)
1689{
1690 int i = 0;
1691
1692 if (!insns)
1693 return 23;
1694
1695 i += __bpf_ld_imm64(&insns[i], R0, ~dst);
1696 i += __bpf_ld_imm64(&insns[i], R1, dst);
1697 i += __bpf_ld_imm64(&insns[i], R2, src);
1698
1699 /* Result unsuccessful */
1700 insns[i++] = BPF_STX_MEM(BPF_DW, R10, R1, -8);
1701 insns[i++] = BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -8);
1702 insns[i++] = BPF_LDX_MEM(BPF_DW, R3, R10, -8);
1703
1704 insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 2);
1705 insns[i++] = BPF_MOV64_IMM(R0, __LINE__);
1706 insns[i++] = BPF_EXIT_INSN();
1707
1708 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R3, 2);
1709 insns[i++] = BPF_MOV64_IMM(R0, __LINE__);
1710 insns[i++] = BPF_EXIT_INSN();
1711
1712 /* Result successful */
1713 insns[i++] = BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -8);
1714 insns[i++] = BPF_LDX_MEM(BPF_DW, R3, R10, -8);
1715
1716 insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R3, 2);
1717 insns[i++] = BPF_MOV64_IMM(R0, __LINE__);
1718 insns[i++] = BPF_EXIT_INSN();
1719
1720 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R1, 2);
1721 insns[i++] = BPF_MOV64_IMM(R0, __LINE__);
1722 insns[i++] = BPF_EXIT_INSN();
1723
1724 return i;
1725}
1726
1727static int __bpf_emit_cmpxchg32(struct bpf_test *self, void *arg,
1728 struct bpf_insn *insns, s64 dst, s64 src)
1729{
1730 int i = 0;
1731
1732 if (!insns)
1733 return 27;
1734
1735 i += __bpf_ld_imm64(&insns[i], R0, ~dst);
1736 i += __bpf_ld_imm64(&insns[i], R1, (u32)dst);
1737 i += __bpf_ld_imm64(&insns[i], R2, src);
1738
1739 /* Result unsuccessful */
1740 insns[i++] = BPF_STX_MEM(BPF_W, R10, R1, -4);
1741 insns[i++] = BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R2, -4);
1742 insns[i++] = BPF_ZEXT_REG(R0), /* Zext always inserted by verifier */
1743 insns[i++] = BPF_LDX_MEM(BPF_W, R3, R10, -4);
1744
1745 insns[i++] = BPF_JMP32_REG(BPF_JEQ, R1, R3, 2);
1746 insns[i++] = BPF_MOV32_IMM(R0, __LINE__);
1747 insns[i++] = BPF_EXIT_INSN();
1748
1749 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R3, 2);
1750 insns[i++] = BPF_MOV32_IMM(R0, __LINE__);
1751 insns[i++] = BPF_EXIT_INSN();
1752
1753 /* Result successful */
1754 i += __bpf_ld_imm64(&insns[i], R0, dst);
1755 insns[i++] = BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R2, -4);
1756 insns[i++] = BPF_ZEXT_REG(R0), /* Zext always inserted by verifier */
1757 insns[i++] = BPF_LDX_MEM(BPF_W, R3, R10, -4);
1758
1759 insns[i++] = BPF_JMP32_REG(BPF_JEQ, R2, R3, 2);
1760 insns[i++] = BPF_MOV32_IMM(R0, __LINE__);
1761 insns[i++] = BPF_EXIT_INSN();
1762
1763 insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R1, 2);
1764 insns[i++] = BPF_MOV32_IMM(R0, __LINE__);
1765 insns[i++] = BPF_EXIT_INSN();
1766
1767 return i;
1768}
1769
1770static int __bpf_fill_atomic64(struct bpf_test *self, int op)
1771{
1772 return __bpf_fill_pattern(self, &op, 64, 64,
1773 0, PATTERN_BLOCK2,
1774 &__bpf_emit_atomic64);
1775}
1776
1777static int __bpf_fill_atomic32(struct bpf_test *self, int op)
1778{
1779 return __bpf_fill_pattern(self, &op, 64, 64,
1780 0, PATTERN_BLOCK2,
1781 &__bpf_emit_atomic32);
1782}
1783
1784/* 64-bit atomic operations */
1785static int bpf_fill_atomic64_add(struct bpf_test *self)
1786{
1787 return __bpf_fill_atomic64(self, BPF_ADD);
1788}
1789
1790static int bpf_fill_atomic64_and(struct bpf_test *self)
1791{
1792 return __bpf_fill_atomic64(self, BPF_AND);
1793}
1794
1795static int bpf_fill_atomic64_or(struct bpf_test *self)
1796{
1797 return __bpf_fill_atomic64(self, BPF_OR);
1798}
1799
1800static int bpf_fill_atomic64_xor(struct bpf_test *self)
1801{
1802 return __bpf_fill_atomic64(self, BPF_XOR);
1803}
1804
1805static int bpf_fill_atomic64_add_fetch(struct bpf_test *self)
1806{
1807 return __bpf_fill_atomic64(self, BPF_ADD | BPF_FETCH);
1808}
1809
1810static int bpf_fill_atomic64_and_fetch(struct bpf_test *self)
1811{
1812 return __bpf_fill_atomic64(self, BPF_AND | BPF_FETCH);
1813}
1814
1815static int bpf_fill_atomic64_or_fetch(struct bpf_test *self)
1816{
1817 return __bpf_fill_atomic64(self, BPF_OR | BPF_FETCH);
1818}
1819
1820static int bpf_fill_atomic64_xor_fetch(struct bpf_test *self)
1821{
1822 return __bpf_fill_atomic64(self, BPF_XOR | BPF_FETCH);
1823}
1824
1825static int bpf_fill_atomic64_xchg(struct bpf_test *self)
1826{
1827 return __bpf_fill_atomic64(self, BPF_XCHG);
1828}
1829
1830static int bpf_fill_cmpxchg64(struct bpf_test *self)
1831{
1832 return __bpf_fill_pattern(self, NULL, 64, 64, 0, PATTERN_BLOCK2,
1833 &__bpf_emit_cmpxchg64);
1834}
1835
1836/* 32-bit atomic operations */
1837static int bpf_fill_atomic32_add(struct bpf_test *self)
1838{
1839 return __bpf_fill_atomic32(self, BPF_ADD);
1840}
1841
1842static int bpf_fill_atomic32_and(struct bpf_test *self)
1843{
1844 return __bpf_fill_atomic32(self, BPF_AND);
1845}
1846
1847static int bpf_fill_atomic32_or(struct bpf_test *self)
1848{
1849 return __bpf_fill_atomic32(self, BPF_OR);
1850}
1851
1852static int bpf_fill_atomic32_xor(struct bpf_test *self)
1853{
1854 return __bpf_fill_atomic32(self, BPF_XOR);
1855}
1856
1857static int bpf_fill_atomic32_add_fetch(struct bpf_test *self)
1858{
1859 return __bpf_fill_atomic32(self, BPF_ADD | BPF_FETCH);
1860}
1861
1862static int bpf_fill_atomic32_and_fetch(struct bpf_test *self)
1863{
1864 return __bpf_fill_atomic32(self, BPF_AND | BPF_FETCH);
1865}
1866
1867static int bpf_fill_atomic32_or_fetch(struct bpf_test *self)
1868{
1869 return __bpf_fill_atomic32(self, BPF_OR | BPF_FETCH);
1870}
1871
1872static int bpf_fill_atomic32_xor_fetch(struct bpf_test *self)
1873{
1874 return __bpf_fill_atomic32(self, BPF_XOR | BPF_FETCH);
1875}
1876
1877static int bpf_fill_atomic32_xchg(struct bpf_test *self)
1878{
1879 return __bpf_fill_atomic32(self, BPF_XCHG);
1880}
1881
1882static int bpf_fill_cmpxchg32(struct bpf_test *self)
1883{
1884 return __bpf_fill_pattern(self, NULL, 64, 64, 0, PATTERN_BLOCK2,
1885 &__bpf_emit_cmpxchg32);
1886}
1887
1888/*
Johan Almbladh6fae2e82021-10-01 15:03:46 +02001889 * Test JITs that implement ATOMIC operations as function calls or
1890 * other primitives, and must re-arrange operands for argument passing.
1891 */
1892static int __bpf_fill_atomic_reg_pairs(struct bpf_test *self, u8 width, u8 op)
1893{
1894 struct bpf_insn *insn;
1895 int len = 2 + 34 * 10 * 10;
1896 u64 mem, upd, res;
1897 int rd, rs, i = 0;
1898
1899 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
1900 if (!insn)
1901 return -ENOMEM;
1902
1903 /* Operand and memory values */
1904 if (width == BPF_DW) {
1905 mem = 0x0123456789abcdefULL;
1906 upd = 0xfedcba9876543210ULL;
1907 } else { /* BPF_W */
1908 mem = 0x01234567U;
1909 upd = 0x76543210U;
1910 }
1911
1912 /* Memory updated according to operation */
1913 switch (op) {
1914 case BPF_XCHG:
1915 res = upd;
1916 break;
1917 case BPF_CMPXCHG:
1918 res = mem;
1919 break;
1920 default:
1921 __bpf_alu_result(&res, mem, upd, BPF_OP(op));
1922 }
1923
1924 /* Test all operand registers */
1925 for (rd = R0; rd <= R9; rd++) {
1926 for (rs = R0; rs <= R9; rs++) {
1927 u64 cmp, src;
1928
1929 /* Initialize value in memory */
1930 i += __bpf_ld_imm64(&insn[i], R0, mem);
1931 insn[i++] = BPF_STX_MEM(width, R10, R0, -8);
1932
1933 /* Initialize registers in order */
1934 i += __bpf_ld_imm64(&insn[i], R0, ~mem);
1935 i += __bpf_ld_imm64(&insn[i], rs, upd);
1936 insn[i++] = BPF_MOV64_REG(rd, R10);
1937
1938 /* Perform atomic operation */
1939 insn[i++] = BPF_ATOMIC_OP(width, op, rd, rs, -8);
1940 if (op == BPF_CMPXCHG && width == BPF_W)
1941 insn[i++] = BPF_ZEXT_REG(R0);
1942
1943 /* Check R0 register value */
1944 if (op == BPF_CMPXCHG)
1945 cmp = mem; /* Expect value from memory */
1946 else if (R0 == rd || R0 == rs)
1947 cmp = 0; /* Aliased, checked below */
1948 else
1949 cmp = ~mem; /* Expect value to be preserved */
1950 if (cmp) {
1951 insn[i++] = BPF_JMP32_IMM(BPF_JEQ, R0,
1952 (u32)cmp, 2);
1953 insn[i++] = BPF_MOV32_IMM(R0, __LINE__);
1954 insn[i++] = BPF_EXIT_INSN();
1955 insn[i++] = BPF_ALU64_IMM(BPF_RSH, R0, 32);
1956 insn[i++] = BPF_JMP32_IMM(BPF_JEQ, R0,
1957 cmp >> 32, 2);
1958 insn[i++] = BPF_MOV32_IMM(R0, __LINE__);
1959 insn[i++] = BPF_EXIT_INSN();
1960 }
1961
1962 /* Check source register value */
1963 if (rs == R0 && op == BPF_CMPXCHG)
1964 src = 0; /* Aliased with R0, checked above */
1965 else if (rs == rd && (op == BPF_CMPXCHG ||
1966 !(op & BPF_FETCH)))
1967 src = 0; /* Aliased with rd, checked below */
1968 else if (op == BPF_CMPXCHG)
1969 src = upd; /* Expect value to be preserved */
1970 else if (op & BPF_FETCH)
1971 src = mem; /* Expect fetched value from mem */
1972 else /* no fetch */
1973 src = upd; /* Expect value to be preserved */
1974 if (src) {
1975 insn[i++] = BPF_JMP32_IMM(BPF_JEQ, rs,
1976 (u32)src, 2);
1977 insn[i++] = BPF_MOV32_IMM(R0, __LINE__);
1978 insn[i++] = BPF_EXIT_INSN();
1979 insn[i++] = BPF_ALU64_IMM(BPF_RSH, rs, 32);
1980 insn[i++] = BPF_JMP32_IMM(BPF_JEQ, rs,
1981 src >> 32, 2);
1982 insn[i++] = BPF_MOV32_IMM(R0, __LINE__);
1983 insn[i++] = BPF_EXIT_INSN();
1984 }
1985
1986 /* Check destination register value */
1987 if (!(rd == R0 && op == BPF_CMPXCHG) &&
1988 !(rd == rs && (op & BPF_FETCH))) {
1989 insn[i++] = BPF_JMP_REG(BPF_JEQ, rd, R10, 2);
1990 insn[i++] = BPF_MOV32_IMM(R0, __LINE__);
1991 insn[i++] = BPF_EXIT_INSN();
1992 }
1993
1994 /* Check value in memory */
1995 if (rs != rd) { /* No aliasing */
1996 i += __bpf_ld_imm64(&insn[i], R1, res);
1997 } else if (op == BPF_XCHG) { /* Aliased, XCHG */
1998 insn[i++] = BPF_MOV64_REG(R1, R10);
1999 } else if (op == BPF_CMPXCHG) { /* Aliased, CMPXCHG */
2000 i += __bpf_ld_imm64(&insn[i], R1, mem);
2001 } else { /* Aliased, ALU oper */
2002 i += __bpf_ld_imm64(&insn[i], R1, mem);
2003 insn[i++] = BPF_ALU64_REG(BPF_OP(op), R1, R10);
2004 }
2005
2006 insn[i++] = BPF_LDX_MEM(width, R0, R10, -8);
2007 if (width == BPF_DW)
2008 insn[i++] = BPF_JMP_REG(BPF_JEQ, R0, R1, 2);
2009 else /* width == BPF_W */
2010 insn[i++] = BPF_JMP32_REG(BPF_JEQ, R0, R1, 2);
2011 insn[i++] = BPF_MOV32_IMM(R0, __LINE__);
2012 insn[i++] = BPF_EXIT_INSN();
2013 }
2014 }
2015
2016 insn[i++] = BPF_MOV64_IMM(R0, 1);
2017 insn[i++] = BPF_EXIT_INSN();
2018
2019 self->u.ptr.insns = insn;
2020 self->u.ptr.len = i;
2021 BUG_ON(i > len);
2022
2023 return 0;
2024}
2025
2026/* 64-bit atomic register tests */
2027static int bpf_fill_atomic64_add_reg_pairs(struct bpf_test *self)
2028{
2029 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_ADD);
2030}
2031
2032static int bpf_fill_atomic64_and_reg_pairs(struct bpf_test *self)
2033{
2034 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_AND);
2035}
2036
2037static int bpf_fill_atomic64_or_reg_pairs(struct bpf_test *self)
2038{
2039 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_OR);
2040}
2041
2042static int bpf_fill_atomic64_xor_reg_pairs(struct bpf_test *self)
2043{
2044 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_XOR);
2045}
2046
2047static int bpf_fill_atomic64_add_fetch_reg_pairs(struct bpf_test *self)
2048{
2049 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_ADD | BPF_FETCH);
2050}
2051
2052static int bpf_fill_atomic64_and_fetch_reg_pairs(struct bpf_test *self)
2053{
2054 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_AND | BPF_FETCH);
2055}
2056
2057static int bpf_fill_atomic64_or_fetch_reg_pairs(struct bpf_test *self)
2058{
2059 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_OR | BPF_FETCH);
2060}
2061
2062static int bpf_fill_atomic64_xor_fetch_reg_pairs(struct bpf_test *self)
2063{
2064 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_XOR | BPF_FETCH);
2065}
2066
2067static int bpf_fill_atomic64_xchg_reg_pairs(struct bpf_test *self)
2068{
2069 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_XCHG);
2070}
2071
2072static int bpf_fill_atomic64_cmpxchg_reg_pairs(struct bpf_test *self)
2073{
2074 return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_CMPXCHG);
2075}
2076
2077/* 32-bit atomic register tests */
2078static int bpf_fill_atomic32_add_reg_pairs(struct bpf_test *self)
2079{
2080 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_ADD);
2081}
2082
2083static int bpf_fill_atomic32_and_reg_pairs(struct bpf_test *self)
2084{
2085 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_AND);
2086}
2087
2088static int bpf_fill_atomic32_or_reg_pairs(struct bpf_test *self)
2089{
2090 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_OR);
2091}
2092
2093static int bpf_fill_atomic32_xor_reg_pairs(struct bpf_test *self)
2094{
2095 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_XOR);
2096}
2097
2098static int bpf_fill_atomic32_add_fetch_reg_pairs(struct bpf_test *self)
2099{
2100 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_ADD | BPF_FETCH);
2101}
2102
2103static int bpf_fill_atomic32_and_fetch_reg_pairs(struct bpf_test *self)
2104{
2105 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_AND | BPF_FETCH);
2106}
2107
2108static int bpf_fill_atomic32_or_fetch_reg_pairs(struct bpf_test *self)
2109{
2110 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_OR | BPF_FETCH);
2111}
2112
2113static int bpf_fill_atomic32_xor_fetch_reg_pairs(struct bpf_test *self)
2114{
2115 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_XOR | BPF_FETCH);
2116}
2117
2118static int bpf_fill_atomic32_xchg_reg_pairs(struct bpf_test *self)
2119{
2120 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_XCHG);
2121}
2122
2123static int bpf_fill_atomic32_cmpxchg_reg_pairs(struct bpf_test *self)
2124{
2125 return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_CMPXCHG);
2126}
2127
2128/*
Johan Almbladh2e807612021-09-14 11:18:35 +02002129 * Test the two-instruction 64-bit immediate load operation for all
2130 * power-of-two magnitudes of the immediate operand. For each MSB, a block
2131 * of immediate values centered around the power-of-two MSB are tested,
2132 * both for positive and negative values. The test is designed to verify
2133 * the operation for JITs that emit different code depending on the magnitude
2134 * of the immediate value. This is often the case if the native instruction
2135 * immediate field width is narrower than 32 bits.
2136 */
Johan Almbladh0eb4ef82021-10-07 16:30:06 +02002137static int bpf_fill_ld_imm64_magn(struct bpf_test *self)
Johan Almbladh2e807612021-09-14 11:18:35 +02002138{
2139 int block = 64; /* Increase for more tests per MSB position */
2140 int len = 3 + 8 * 63 * block * 2;
2141 struct bpf_insn *insn;
2142 int bit, adj, sign;
2143 int i = 0;
2144
2145 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
2146 if (!insn)
2147 return -ENOMEM;
2148
2149 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0);
2150
2151 for (bit = 0; bit <= 62; bit++) {
2152 for (adj = -block / 2; adj < block / 2; adj++) {
2153 for (sign = -1; sign <= 1; sign += 2) {
2154 s64 imm = sign * ((1LL << bit) + adj);
2155
2156 /* Perform operation */
2157 i += __bpf_ld_imm64(&insn[i], R1, imm);
2158
2159 /* Load reference */
2160 insn[i++] = BPF_ALU32_IMM(BPF_MOV, R2, imm);
2161 insn[i++] = BPF_ALU32_IMM(BPF_MOV, R3,
2162 (u32)(imm >> 32));
2163 insn[i++] = BPF_ALU64_IMM(BPF_LSH, R3, 32);
2164 insn[i++] = BPF_ALU64_REG(BPF_OR, R2, R3);
2165
2166 /* Check result */
2167 insn[i++] = BPF_JMP_REG(BPF_JEQ, R1, R2, 1);
2168 insn[i++] = BPF_EXIT_INSN();
2169 }
2170 }
2171 }
2172
2173 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1);
2174 insn[i++] = BPF_EXIT_INSN();
2175
2176 self->u.ptr.insns = insn;
2177 self->u.ptr.len = len;
2178 BUG_ON(i != len);
2179
2180 return 0;
2181}
Johan Almbladha5a36542021-09-14 11:18:33 +02002182
2183/*
Johan Almbladh0eb4ef82021-10-07 16:30:06 +02002184 * Test the two-instruction 64-bit immediate load operation for different
2185 * combinations of bytes. Each byte in the 64-bit word is constructed as
2186 * (base & mask) | (rand() & ~mask), where rand() is a deterministic LCG.
2187 * All patterns (base1, mask1) and (base2, mask2) bytes are tested.
2188 */
2189static int __bpf_fill_ld_imm64_bytes(struct bpf_test *self,
2190 u8 base1, u8 mask1,
2191 u8 base2, u8 mask2)
2192{
2193 struct bpf_insn *insn;
2194 int len = 3 + 8 * BIT(8);
2195 int pattern, index;
2196 u32 rand = 1;
2197 int i = 0;
2198
2199 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL);
2200 if (!insn)
2201 return -ENOMEM;
2202
2203 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0);
2204
2205 for (pattern = 0; pattern < BIT(8); pattern++) {
2206 u64 imm = 0;
2207
2208 for (index = 0; index < 8; index++) {
2209 int byte;
2210
2211 if (pattern & BIT(index))
2212 byte = (base1 & mask1) | (rand & ~mask1);
2213 else
2214 byte = (base2 & mask2) | (rand & ~mask2);
2215 imm = (imm << 8) | byte;
2216 }
2217
2218 /* Update our LCG */
2219 rand = rand * 1664525 + 1013904223;
2220
2221 /* Perform operation */
2222 i += __bpf_ld_imm64(&insn[i], R1, imm);
2223
2224 /* Load reference */
2225 insn[i++] = BPF_ALU32_IMM(BPF_MOV, R2, imm);
2226 insn[i++] = BPF_ALU32_IMM(BPF_MOV, R3, (u32)(imm >> 32));
2227 insn[i++] = BPF_ALU64_IMM(BPF_LSH, R3, 32);
2228 insn[i++] = BPF_ALU64_REG(BPF_OR, R2, R3);
2229
2230 /* Check result */
2231 insn[i++] = BPF_JMP_REG(BPF_JEQ, R1, R2, 1);
2232 insn[i++] = BPF_EXIT_INSN();
2233 }
2234
2235 insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1);
2236 insn[i++] = BPF_EXIT_INSN();
2237
2238 self->u.ptr.insns = insn;
2239 self->u.ptr.len = len;
2240 BUG_ON(i != len);
2241
2242 return 0;
2243}
2244
2245static int bpf_fill_ld_imm64_checker(struct bpf_test *self)
2246{
2247 return __bpf_fill_ld_imm64_bytes(self, 0, 0xff, 0xff, 0xff);
2248}
2249
2250static int bpf_fill_ld_imm64_pos_neg(struct bpf_test *self)
2251{
2252 return __bpf_fill_ld_imm64_bytes(self, 1, 0x81, 0x80, 0x80);
2253}
2254
2255static int bpf_fill_ld_imm64_pos_zero(struct bpf_test *self)
2256{
2257 return __bpf_fill_ld_imm64_bytes(self, 1, 0x81, 0, 0xff);
2258}
2259
2260static int bpf_fill_ld_imm64_neg_zero(struct bpf_test *self)
2261{
2262 return __bpf_fill_ld_imm64_bytes(self, 0x80, 0x80, 0, 0xff);
2263}
2264
2265/*
Johan Almbladha5a36542021-09-14 11:18:33 +02002266 * Exhaustive tests of JMP operations for all combinations of power-of-two
2267 * magnitudes of the operands, both for positive and negative values. The
2268 * test is designed to verify e.g. the JMP and JMP32 operations for JITs that
2269 * emit different code depending on the magnitude of the immediate value.
2270 */
2271
2272static bool __bpf_match_jmp_cond(s64 v1, s64 v2, u8 op)
2273{
2274 switch (op) {
2275 case BPF_JSET:
2276 return !!(v1 & v2);
2277 case BPF_JEQ:
2278 return v1 == v2;
2279 case BPF_JNE:
2280 return v1 != v2;
2281 case BPF_JGT:
2282 return (u64)v1 > (u64)v2;
2283 case BPF_JGE:
2284 return (u64)v1 >= (u64)v2;
2285 case BPF_JLT:
2286 return (u64)v1 < (u64)v2;
2287 case BPF_JLE:
2288 return (u64)v1 <= (u64)v2;
2289 case BPF_JSGT:
2290 return v1 > v2;
2291 case BPF_JSGE:
2292 return v1 >= v2;
2293 case BPF_JSLT:
2294 return v1 < v2;
2295 case BPF_JSLE:
2296 return v1 <= v2;
2297 }
2298 return false;
2299}
2300
2301static int __bpf_emit_jmp_imm(struct bpf_test *self, void *arg,
2302 struct bpf_insn *insns, s64 dst, s64 imm)
2303{
2304 int op = *(int *)arg;
2305
2306 if (insns) {
2307 bool match = __bpf_match_jmp_cond(dst, (s32)imm, op);
2308 int i = 0;
2309
2310 insns[i++] = BPF_ALU32_IMM(BPF_MOV, R0, match);
2311
2312 i += __bpf_ld_imm64(&insns[i], R1, dst);
2313 insns[i++] = BPF_JMP_IMM(op, R1, imm, 1);
2314 if (!match)
2315 insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1);
2316 insns[i++] = BPF_EXIT_INSN();
2317
2318 return i;
2319 }
2320
2321 return 5 + 1;
2322}
2323
2324static int __bpf_emit_jmp32_imm(struct bpf_test *self, void *arg,
2325 struct bpf_insn *insns, s64 dst, s64 imm)
2326{
2327 int op = *(int *)arg;
2328
2329 if (insns) {
2330 bool match = __bpf_match_jmp_cond((s32)dst, (s32)imm, op);
2331 int i = 0;
2332
2333 i += __bpf_ld_imm64(&insns[i], R1, dst);
2334 insns[i++] = BPF_JMP32_IMM(op, R1, imm, 1);
2335 if (!match)
2336 insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1);
2337 insns[i++] = BPF_EXIT_INSN();
2338
2339 return i;
2340 }
2341
2342 return 5;
2343}
2344
2345static int __bpf_emit_jmp_reg(struct bpf_test *self, void *arg,
2346 struct bpf_insn *insns, s64 dst, s64 src)
2347{
2348 int op = *(int *)arg;
2349
2350 if (insns) {
2351 bool match = __bpf_match_jmp_cond(dst, src, op);
2352 int i = 0;
2353
2354 i += __bpf_ld_imm64(&insns[i], R1, dst);
2355 i += __bpf_ld_imm64(&insns[i], R2, src);
2356 insns[i++] = BPF_JMP_REG(op, R1, R2, 1);
2357 if (!match)
2358 insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1);
2359 insns[i++] = BPF_EXIT_INSN();
2360
2361 return i;
2362 }
2363
2364 return 7;
2365}
2366
2367static int __bpf_emit_jmp32_reg(struct bpf_test *self, void *arg,
2368 struct bpf_insn *insns, s64 dst, s64 src)
2369{
2370 int op = *(int *)arg;
2371
2372 if (insns) {
2373 bool match = __bpf_match_jmp_cond((s32)dst, (s32)src, op);
2374 int i = 0;
2375
2376 i += __bpf_ld_imm64(&insns[i], R1, dst);
2377 i += __bpf_ld_imm64(&insns[i], R2, src);
2378 insns[i++] = BPF_JMP32_REG(op, R1, R2, 1);
2379 if (!match)
2380 insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1);
2381 insns[i++] = BPF_EXIT_INSN();
2382
2383 return i;
2384 }
2385
2386 return 7;
2387}
2388
2389static int __bpf_fill_jmp_imm(struct bpf_test *self, int op)
2390{
2391 return __bpf_fill_pattern(self, &op, 64, 32,
2392 PATTERN_BLOCK1, PATTERN_BLOCK2,
2393 &__bpf_emit_jmp_imm);
2394}
2395
2396static int __bpf_fill_jmp32_imm(struct bpf_test *self, int op)
2397{
2398 return __bpf_fill_pattern(self, &op, 64, 32,
2399 PATTERN_BLOCK1, PATTERN_BLOCK2,
2400 &__bpf_emit_jmp32_imm);
2401}
2402
2403static int __bpf_fill_jmp_reg(struct bpf_test *self, int op)
2404{
2405 return __bpf_fill_pattern(self, &op, 64, 64,
2406 PATTERN_BLOCK1, PATTERN_BLOCK2,
2407 &__bpf_emit_jmp_reg);
2408}
2409
2410static int __bpf_fill_jmp32_reg(struct bpf_test *self, int op)
2411{
2412 return __bpf_fill_pattern(self, &op, 64, 64,
2413 PATTERN_BLOCK1, PATTERN_BLOCK2,
2414 &__bpf_emit_jmp32_reg);
2415}
2416
2417/* JMP immediate tests */
2418static int bpf_fill_jmp_jset_imm(struct bpf_test *self)
2419{
2420 return __bpf_fill_jmp_imm(self, BPF_JSET);
2421}
2422
2423static int bpf_fill_jmp_jeq_imm(struct bpf_test *self)
2424{
2425 return __bpf_fill_jmp_imm(self, BPF_JEQ);
2426}
2427
2428static int bpf_fill_jmp_jne_imm(struct bpf_test *self)
2429{
2430 return __bpf_fill_jmp_imm(self, BPF_JNE);
2431}
2432
2433static int bpf_fill_jmp_jgt_imm(struct bpf_test *self)
2434{
2435 return __bpf_fill_jmp_imm(self, BPF_JGT);
2436}
2437
2438static int bpf_fill_jmp_jge_imm(struct bpf_test *self)
2439{
2440 return __bpf_fill_jmp_imm(self, BPF_JGE);
2441}
2442
2443static int bpf_fill_jmp_jlt_imm(struct bpf_test *self)
2444{
2445 return __bpf_fill_jmp_imm(self, BPF_JLT);
2446}
2447
2448static int bpf_fill_jmp_jle_imm(struct bpf_test *self)
2449{
2450 return __bpf_fill_jmp_imm(self, BPF_JLE);
2451}
2452
2453static int bpf_fill_jmp_jsgt_imm(struct bpf_test *self)
2454{
2455 return __bpf_fill_jmp_imm(self, BPF_JSGT);
2456}
2457
2458static int bpf_fill_jmp_jsge_imm(struct bpf_test *self)
2459{
2460 return __bpf_fill_jmp_imm(self, BPF_JSGE);
2461}
2462
2463static int bpf_fill_jmp_jslt_imm(struct bpf_test *self)
2464{
2465 return __bpf_fill_jmp_imm(self, BPF_JSLT);
2466}
2467
2468static int bpf_fill_jmp_jsle_imm(struct bpf_test *self)
2469{
2470 return __bpf_fill_jmp_imm(self, BPF_JSLE);
2471}
2472
2473/* JMP32 immediate tests */
2474static int bpf_fill_jmp32_jset_imm(struct bpf_test *self)
2475{
2476 return __bpf_fill_jmp32_imm(self, BPF_JSET);
2477}
2478
2479static int bpf_fill_jmp32_jeq_imm(struct bpf_test *self)
2480{
2481 return __bpf_fill_jmp32_imm(self, BPF_JEQ);
2482}
2483
2484static int bpf_fill_jmp32_jne_imm(struct bpf_test *self)
2485{
2486 return __bpf_fill_jmp32_imm(self, BPF_JNE);
2487}
2488
2489static int bpf_fill_jmp32_jgt_imm(struct bpf_test *self)
2490{
2491 return __bpf_fill_jmp32_imm(self, BPF_JGT);
2492}
2493
2494static int bpf_fill_jmp32_jge_imm(struct bpf_test *self)
2495{
2496 return __bpf_fill_jmp32_imm(self, BPF_JGE);
2497}
2498
2499static int bpf_fill_jmp32_jlt_imm(struct bpf_test *self)
2500{
2501 return __bpf_fill_jmp32_imm(self, BPF_JLT);
2502}
2503
2504static int bpf_fill_jmp32_jle_imm(struct bpf_test *self)
2505{
2506 return __bpf_fill_jmp32_imm(self, BPF_JLE);
2507}
2508
2509static int bpf_fill_jmp32_jsgt_imm(struct bpf_test *self)
2510{
2511 return __bpf_fill_jmp32_imm(self, BPF_JSGT);
2512}
2513
2514static int bpf_fill_jmp32_jsge_imm(struct bpf_test *self)
2515{
2516 return __bpf_fill_jmp32_imm(self, BPF_JSGE);
2517}
2518
2519static int bpf_fill_jmp32_jslt_imm(struct bpf_test *self)
2520{
2521 return __bpf_fill_jmp32_imm(self, BPF_JSLT);
2522}
2523
2524static int bpf_fill_jmp32_jsle_imm(struct bpf_test *self)
2525{
2526 return __bpf_fill_jmp32_imm(self, BPF_JSLE);
2527}
2528
2529/* JMP register tests */
2530static int bpf_fill_jmp_jset_reg(struct bpf_test *self)
2531{
2532 return __bpf_fill_jmp_reg(self, BPF_JSET);
2533}
2534
2535static int bpf_fill_jmp_jeq_reg(struct bpf_test *self)
2536{
2537 return __bpf_fill_jmp_reg(self, BPF_JEQ);
2538}
2539
2540static int bpf_fill_jmp_jne_reg(struct bpf_test *self)
2541{
2542 return __bpf_fill_jmp_reg(self, BPF_JNE);
2543}
2544
2545static int bpf_fill_jmp_jgt_reg(struct bpf_test *self)
2546{
2547 return __bpf_fill_jmp_reg(self, BPF_JGT);
2548}
2549
2550static int bpf_fill_jmp_jge_reg(struct bpf_test *self)
2551{
2552 return __bpf_fill_jmp_reg(self, BPF_JGE);
2553}
2554
2555static int bpf_fill_jmp_jlt_reg(struct bpf_test *self)
2556{
2557 return __bpf_fill_jmp_reg(self, BPF_JLT);
2558}
2559
2560static int bpf_fill_jmp_jle_reg(struct bpf_test *self)
2561{
2562 return __bpf_fill_jmp_reg(self, BPF_JLE);
2563}
2564
2565static int bpf_fill_jmp_jsgt_reg(struct bpf_test *self)
2566{
2567 return __bpf_fill_jmp_reg(self, BPF_JSGT);
2568}
2569
2570static int bpf_fill_jmp_jsge_reg(struct bpf_test *self)
2571{
2572 return __bpf_fill_jmp_reg(self, BPF_JSGE);
2573}
2574
2575static int bpf_fill_jmp_jslt_reg(struct bpf_test *self)
2576{
2577 return __bpf_fill_jmp_reg(self, BPF_JSLT);
2578}
2579
2580static int bpf_fill_jmp_jsle_reg(struct bpf_test *self)
2581{
2582 return __bpf_fill_jmp_reg(self, BPF_JSLE);
2583}
2584
2585/* JMP32 register tests */
2586static int bpf_fill_jmp32_jset_reg(struct bpf_test *self)
2587{
2588 return __bpf_fill_jmp32_reg(self, BPF_JSET);
2589}
2590
2591static int bpf_fill_jmp32_jeq_reg(struct bpf_test *self)
2592{
2593 return __bpf_fill_jmp32_reg(self, BPF_JEQ);
2594}
2595
2596static int bpf_fill_jmp32_jne_reg(struct bpf_test *self)
2597{
2598 return __bpf_fill_jmp32_reg(self, BPF_JNE);
2599}
2600
2601static int bpf_fill_jmp32_jgt_reg(struct bpf_test *self)
2602{
2603 return __bpf_fill_jmp32_reg(self, BPF_JGT);
2604}
2605
2606static int bpf_fill_jmp32_jge_reg(struct bpf_test *self)
2607{
2608 return __bpf_fill_jmp32_reg(self, BPF_JGE);
2609}
2610
2611static int bpf_fill_jmp32_jlt_reg(struct bpf_test *self)
2612{
2613 return __bpf_fill_jmp32_reg(self, BPF_JLT);
2614}
2615
2616static int bpf_fill_jmp32_jle_reg(struct bpf_test *self)
2617{
2618 return __bpf_fill_jmp32_reg(self, BPF_JLE);
2619}
2620
2621static int bpf_fill_jmp32_jsgt_reg(struct bpf_test *self)
2622{
2623 return __bpf_fill_jmp32_reg(self, BPF_JSGT);
2624}
2625
2626static int bpf_fill_jmp32_jsge_reg(struct bpf_test *self)
2627{
2628 return __bpf_fill_jmp32_reg(self, BPF_JSGE);
2629}
2630
2631static int bpf_fill_jmp32_jslt_reg(struct bpf_test *self)
2632{
2633 return __bpf_fill_jmp32_reg(self, BPF_JSLT);
2634}
2635
2636static int bpf_fill_jmp32_jsle_reg(struct bpf_test *self)
2637{
2638 return __bpf_fill_jmp32_reg(self, BPF_JSLE);
2639}
2640
Johan Almbladha7d2e752021-09-14 11:18:34 +02002641/*
2642 * Set up a sequence of staggered jumps, forwards and backwards with
2643 * increasing offset. This tests the conversion of relative jumps to
2644 * JITed native jumps. On some architectures, for example MIPS, a large
2645 * PC-relative jump offset may overflow the immediate field of the native
2646 * conditional branch instruction, triggering a conversion to use an
2647 * absolute jump instead. Since this changes the jump offsets, another
2648 * offset computation pass is necessary, and that may in turn trigger
2649 * another branch conversion. This jump sequence is particularly nasty
2650 * in that regard.
2651 *
2652 * The sequence generation is parameterized by size and jump type.
2653 * The size must be even, and the expected result is always size + 1.
2654 * Below is an example with size=8 and result=9.
2655 *
2656 * ________________________Start
2657 * R0 = 0
2658 * R1 = r1
2659 * R2 = r2
2660 * ,------- JMP +4 * 3______________Preamble: 4 insns
2661 * ,----------|-ind 0- if R0 != 7 JMP 8 * 3 + 1 <--------------------.
2662 * | | R0 = 8 |
2663 * | | JMP +7 * 3 ------------------------.
2664 * | ,--------|-----1- if R0 != 5 JMP 7 * 3 + 1 <--------------. | |
2665 * | | | R0 = 6 | | |
2666 * | | | JMP +5 * 3 ------------------. | |
2667 * | | ,------|-----2- if R0 != 3 JMP 6 * 3 + 1 <--------. | | | |
2668 * | | | | R0 = 4 | | | | |
2669 * | | | | JMP +3 * 3 ------------. | | | |
2670 * | | | ,----|-----3- if R0 != 1 JMP 5 * 3 + 1 <--. | | | | | |
2671 * | | | | | R0 = 2 | | | | | | |
2672 * | | | | | JMP +1 * 3 ------. | | | | | |
2673 * | | | | ,--t=====4> if R0 != 0 JMP 4 * 3 + 1 1 2 3 4 5 6 7 8 loc
2674 * | | | | | R0 = 1 -1 +2 -3 +4 -5 +6 -7 +8 off
2675 * | | | | | JMP -2 * 3 ---' | | | | | | |
2676 * | | | | | ,------5- if R0 != 2 JMP 3 * 3 + 1 <-----' | | | | | |
2677 * | | | | | | R0 = 3 | | | | | |
2678 * | | | | | | JMP -4 * 3 ---------' | | | | |
2679 * | | | | | | ,----6- if R0 != 4 JMP 2 * 3 + 1 <-----------' | | | |
2680 * | | | | | | | R0 = 5 | | | |
2681 * | | | | | | | JMP -6 * 3 ---------------' | | |
2682 * | | | | | | | ,--7- if R0 != 6 JMP 1 * 3 + 1 <-----------------' | |
2683 * | | | | | | | | R0 = 7 | |
2684 * | | Error | | | JMP -8 * 3 ---------------------' |
2685 * | | paths | | | ,8- if R0 != 8 JMP 0 * 3 + 1 <-----------------------'
2686 * | | | | | | | | | R0 = 9__________________Sequence: 3 * size - 1 insns
2687 * `-+-+-+-+-+-+-+-+-> EXIT____________________Return: 1 insn
2688 *
2689 */
2690
2691/* The maximum size parameter */
2692#define MAX_STAGGERED_JMP_SIZE ((0x7fff / 3) & ~1)
2693
2694/* We use a reduced number of iterations to get a reasonable execution time */
2695#define NR_STAGGERED_JMP_RUNS 10
2696
2697static int __bpf_fill_staggered_jumps(struct bpf_test *self,
2698 const struct bpf_insn *jmp,
2699 u64 r1, u64 r2)
2700{
2701 int size = self->test[0].result - 1;
2702 int len = 4 + 3 * (size + 1);
2703 struct bpf_insn *insns;
2704 int off, ind;
2705
2706 insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL);
2707 if (!insns)
2708 return -ENOMEM;
2709
2710 /* Preamble */
2711 insns[0] = BPF_ALU64_IMM(BPF_MOV, R0, 0);
2712 insns[1] = BPF_ALU64_IMM(BPF_MOV, R1, r1);
2713 insns[2] = BPF_ALU64_IMM(BPF_MOV, R2, r2);
2714 insns[3] = BPF_JMP_IMM(BPF_JA, 0, 0, 3 * size / 2);
2715
2716 /* Sequence */
2717 for (ind = 0, off = size; ind <= size; ind++, off -= 2) {
2718 struct bpf_insn *ins = &insns[4 + 3 * ind];
2719 int loc;
2720
2721 if (off == 0)
2722 off--;
2723
2724 loc = abs(off);
2725 ins[0] = BPF_JMP_IMM(BPF_JNE, R0, loc - 1,
2726 3 * (size - ind) + 1);
2727 ins[1] = BPF_ALU64_IMM(BPF_MOV, R0, loc);
2728 ins[2] = *jmp;
2729 ins[2].off = 3 * (off - 1);
2730 }
2731
2732 /* Return */
2733 insns[len - 1] = BPF_EXIT_INSN();
2734
2735 self->u.ptr.insns = insns;
2736 self->u.ptr.len = len;
2737
2738 return 0;
2739}
2740
2741/* 64-bit unconditional jump */
2742static int bpf_fill_staggered_ja(struct bpf_test *self)
2743{
2744 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JA, 0, 0, 0);
2745
2746 return __bpf_fill_staggered_jumps(self, &jmp, 0, 0);
2747}
2748
2749/* 64-bit immediate jumps */
2750static int bpf_fill_staggered_jeq_imm(struct bpf_test *self)
2751{
2752 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JEQ, R1, 1234, 0);
2753
2754 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0);
2755}
2756
2757static int bpf_fill_staggered_jne_imm(struct bpf_test *self)
2758{
2759 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JNE, R1, 1234, 0);
2760
2761 return __bpf_fill_staggered_jumps(self, &jmp, 4321, 0);
2762}
2763
2764static int bpf_fill_staggered_jset_imm(struct bpf_test *self)
2765{
2766 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSET, R1, 0x82, 0);
2767
2768 return __bpf_fill_staggered_jumps(self, &jmp, 0x86, 0);
2769}
2770
2771static int bpf_fill_staggered_jgt_imm(struct bpf_test *self)
2772{
2773 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JGT, R1, 1234, 0);
2774
2775 return __bpf_fill_staggered_jumps(self, &jmp, 0x80000000, 0);
2776}
2777
2778static int bpf_fill_staggered_jge_imm(struct bpf_test *self)
2779{
2780 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JGE, R1, 1234, 0);
2781
2782 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0);
2783}
2784
2785static int bpf_fill_staggered_jlt_imm(struct bpf_test *self)
2786{
2787 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JLT, R1, 0x80000000, 0);
2788
2789 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0);
2790}
2791
2792static int bpf_fill_staggered_jle_imm(struct bpf_test *self)
2793{
2794 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JLE, R1, 1234, 0);
2795
2796 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0);
2797}
2798
2799static int bpf_fill_staggered_jsgt_imm(struct bpf_test *self)
2800{
2801 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSGT, R1, -2, 0);
2802
2803 return __bpf_fill_staggered_jumps(self, &jmp, -1, 0);
2804}
2805
2806static int bpf_fill_staggered_jsge_imm(struct bpf_test *self)
2807{
2808 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSGE, R1, -2, 0);
2809
2810 return __bpf_fill_staggered_jumps(self, &jmp, -2, 0);
2811}
2812
2813static int bpf_fill_staggered_jslt_imm(struct bpf_test *self)
2814{
2815 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSLT, R1, -1, 0);
2816
2817 return __bpf_fill_staggered_jumps(self, &jmp, -2, 0);
2818}
2819
2820static int bpf_fill_staggered_jsle_imm(struct bpf_test *self)
2821{
2822 struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSLE, R1, -1, 0);
2823
2824 return __bpf_fill_staggered_jumps(self, &jmp, -1, 0);
2825}
2826
2827/* 64-bit register jumps */
2828static int bpf_fill_staggered_jeq_reg(struct bpf_test *self)
2829{
2830 struct bpf_insn jmp = BPF_JMP_REG(BPF_JEQ, R1, R2, 0);
2831
2832 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234);
2833}
2834
2835static int bpf_fill_staggered_jne_reg(struct bpf_test *self)
2836{
2837 struct bpf_insn jmp = BPF_JMP_REG(BPF_JNE, R1, R2, 0);
2838
2839 return __bpf_fill_staggered_jumps(self, &jmp, 4321, 1234);
2840}
2841
2842static int bpf_fill_staggered_jset_reg(struct bpf_test *self)
2843{
2844 struct bpf_insn jmp = BPF_JMP_REG(BPF_JSET, R1, R2, 0);
2845
2846 return __bpf_fill_staggered_jumps(self, &jmp, 0x86, 0x82);
2847}
2848
2849static int bpf_fill_staggered_jgt_reg(struct bpf_test *self)
2850{
2851 struct bpf_insn jmp = BPF_JMP_REG(BPF_JGT, R1, R2, 0);
2852
2853 return __bpf_fill_staggered_jumps(self, &jmp, 0x80000000, 1234);
2854}
2855
2856static int bpf_fill_staggered_jge_reg(struct bpf_test *self)
2857{
2858 struct bpf_insn jmp = BPF_JMP_REG(BPF_JGE, R1, R2, 0);
2859
2860 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234);
2861}
2862
2863static int bpf_fill_staggered_jlt_reg(struct bpf_test *self)
2864{
2865 struct bpf_insn jmp = BPF_JMP_REG(BPF_JLT, R1, R2, 0);
2866
2867 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0x80000000);
2868}
2869
2870static int bpf_fill_staggered_jle_reg(struct bpf_test *self)
2871{
2872 struct bpf_insn jmp = BPF_JMP_REG(BPF_JLE, R1, R2, 0);
2873
2874 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234);
2875}
2876
2877static int bpf_fill_staggered_jsgt_reg(struct bpf_test *self)
2878{
2879 struct bpf_insn jmp = BPF_JMP_REG(BPF_JSGT, R1, R2, 0);
2880
2881 return __bpf_fill_staggered_jumps(self, &jmp, -1, -2);
2882}
2883
2884static int bpf_fill_staggered_jsge_reg(struct bpf_test *self)
2885{
2886 struct bpf_insn jmp = BPF_JMP_REG(BPF_JSGE, R1, R2, 0);
2887
2888 return __bpf_fill_staggered_jumps(self, &jmp, -2, -2);
2889}
2890
2891static int bpf_fill_staggered_jslt_reg(struct bpf_test *self)
2892{
2893 struct bpf_insn jmp = BPF_JMP_REG(BPF_JSLT, R1, R2, 0);
2894
2895 return __bpf_fill_staggered_jumps(self, &jmp, -2, -1);
2896}
2897
2898static int bpf_fill_staggered_jsle_reg(struct bpf_test *self)
2899{
2900 struct bpf_insn jmp = BPF_JMP_REG(BPF_JSLE, R1, R2, 0);
2901
2902 return __bpf_fill_staggered_jumps(self, &jmp, -1, -1);
2903}
2904
2905/* 32-bit immediate jumps */
2906static int bpf_fill_staggered_jeq32_imm(struct bpf_test *self)
2907{
2908 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JEQ, R1, 1234, 0);
2909
2910 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0);
2911}
2912
2913static int bpf_fill_staggered_jne32_imm(struct bpf_test *self)
2914{
2915 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JNE, R1, 1234, 0);
2916
2917 return __bpf_fill_staggered_jumps(self, &jmp, 4321, 0);
2918}
2919
2920static int bpf_fill_staggered_jset32_imm(struct bpf_test *self)
2921{
2922 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSET, R1, 0x82, 0);
2923
2924 return __bpf_fill_staggered_jumps(self, &jmp, 0x86, 0);
2925}
2926
2927static int bpf_fill_staggered_jgt32_imm(struct bpf_test *self)
2928{
2929 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JGT, R1, 1234, 0);
2930
2931 return __bpf_fill_staggered_jumps(self, &jmp, 0x80000000, 0);
2932}
2933
2934static int bpf_fill_staggered_jge32_imm(struct bpf_test *self)
2935{
2936 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JGE, R1, 1234, 0);
2937
2938 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0);
2939}
2940
2941static int bpf_fill_staggered_jlt32_imm(struct bpf_test *self)
2942{
2943 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JLT, R1, 0x80000000, 0);
2944
2945 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0);
2946}
2947
2948static int bpf_fill_staggered_jle32_imm(struct bpf_test *self)
2949{
2950 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JLE, R1, 1234, 0);
2951
2952 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0);
2953}
2954
2955static int bpf_fill_staggered_jsgt32_imm(struct bpf_test *self)
2956{
2957 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSGT, R1, -2, 0);
2958
2959 return __bpf_fill_staggered_jumps(self, &jmp, -1, 0);
2960}
2961
2962static int bpf_fill_staggered_jsge32_imm(struct bpf_test *self)
2963{
2964 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSGE, R1, -2, 0);
2965
2966 return __bpf_fill_staggered_jumps(self, &jmp, -2, 0);
2967}
2968
2969static int bpf_fill_staggered_jslt32_imm(struct bpf_test *self)
2970{
2971 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSLT, R1, -1, 0);
2972
2973 return __bpf_fill_staggered_jumps(self, &jmp, -2, 0);
2974}
2975
2976static int bpf_fill_staggered_jsle32_imm(struct bpf_test *self)
2977{
2978 struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSLE, R1, -1, 0);
2979
2980 return __bpf_fill_staggered_jumps(self, &jmp, -1, 0);
2981}
2982
2983/* 32-bit register jumps */
2984static int bpf_fill_staggered_jeq32_reg(struct bpf_test *self)
2985{
2986 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JEQ, R1, R2, 0);
2987
2988 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234);
2989}
2990
2991static int bpf_fill_staggered_jne32_reg(struct bpf_test *self)
2992{
2993 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JNE, R1, R2, 0);
2994
2995 return __bpf_fill_staggered_jumps(self, &jmp, 4321, 1234);
2996}
2997
2998static int bpf_fill_staggered_jset32_reg(struct bpf_test *self)
2999{
3000 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSET, R1, R2, 0);
3001
3002 return __bpf_fill_staggered_jumps(self, &jmp, 0x86, 0x82);
3003}
3004
3005static int bpf_fill_staggered_jgt32_reg(struct bpf_test *self)
3006{
3007 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JGT, R1, R2, 0);
3008
3009 return __bpf_fill_staggered_jumps(self, &jmp, 0x80000000, 1234);
3010}
3011
3012static int bpf_fill_staggered_jge32_reg(struct bpf_test *self)
3013{
3014 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JGE, R1, R2, 0);
3015
3016 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234);
3017}
3018
3019static int bpf_fill_staggered_jlt32_reg(struct bpf_test *self)
3020{
3021 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JLT, R1, R2, 0);
3022
3023 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0x80000000);
3024}
3025
3026static int bpf_fill_staggered_jle32_reg(struct bpf_test *self)
3027{
3028 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JLE, R1, R2, 0);
3029
3030 return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234);
3031}
3032
3033static int bpf_fill_staggered_jsgt32_reg(struct bpf_test *self)
3034{
3035 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSGT, R1, R2, 0);
3036
3037 return __bpf_fill_staggered_jumps(self, &jmp, -1, -2);
3038}
3039
3040static int bpf_fill_staggered_jsge32_reg(struct bpf_test *self)
3041{
3042 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSGE, R1, R2, 0);
3043
3044 return __bpf_fill_staggered_jumps(self, &jmp, -2, -2);
3045}
3046
3047static int bpf_fill_staggered_jslt32_reg(struct bpf_test *self)
3048{
3049 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSLT, R1, R2, 0);
3050
3051 return __bpf_fill_staggered_jumps(self, &jmp, -2, -1);
3052}
3053
3054static int bpf_fill_staggered_jsle32_reg(struct bpf_test *self)
3055{
3056 struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSLE, R1, R2, 0);
3057
3058 return __bpf_fill_staggered_jumps(self, &jmp, -1, -1);
3059}
3060
Johan Almbladha5a36542021-09-14 11:18:33 +02003061
Alexei Starovoitov64a89462014-05-08 14:10:52 -07003062static struct bpf_test tests[] = {
3063 {
3064 "TAX",
Andrew Mortonece80492014-05-22 10:16:46 -07003065 .u.insns = {
Alexei Starovoitov64a89462014-05-08 14:10:52 -07003066 BPF_STMT(BPF_LD | BPF_IMM, 1),
3067 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3068 BPF_STMT(BPF_LD | BPF_IMM, 2),
3069 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3070 BPF_STMT(BPF_ALU | BPF_NEG, 0), /* A == -3 */
3071 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3072 BPF_STMT(BPF_LD | BPF_LEN, 0),
3073 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3074 BPF_STMT(BPF_MISC | BPF_TAX, 0), /* X == len - 3 */
3075 BPF_STMT(BPF_LD | BPF_B | BPF_IND, 1),
3076 BPF_STMT(BPF_RET | BPF_A, 0)
3077 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003078 CLASSIC,
Alexei Starovoitov64a89462014-05-08 14:10:52 -07003079 { 10, 20, 30, 40, 50 },
3080 { { 2, 10 }, { 3, 20 }, { 4, 30 } },
3081 },
3082 {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003083 "TXA",
Andrew Mortonece80492014-05-22 10:16:46 -07003084 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003085 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3086 BPF_STMT(BPF_MISC | BPF_TXA, 0),
3087 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3088 BPF_STMT(BPF_RET | BPF_A, 0) /* A == len * 2 */
3089 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003090 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003091 { 10, 20, 30, 40, 50 },
3092 { { 1, 2 }, { 3, 6 }, { 4, 8 } },
3093 },
3094 {
3095 "ADD_SUB_MUL_K",
Andrew Mortonece80492014-05-22 10:16:46 -07003096 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003097 BPF_STMT(BPF_LD | BPF_IMM, 1),
3098 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 2),
3099 BPF_STMT(BPF_LDX | BPF_IMM, 3),
3100 BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0),
3101 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 0xffffffff),
3102 BPF_STMT(BPF_ALU | BPF_MUL | BPF_K, 3),
3103 BPF_STMT(BPF_RET | BPF_A, 0)
3104 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003105 CLASSIC | FLAG_NO_DATA,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003106 { },
3107 { { 0, 0xfffffffd } }
3108 },
3109 {
Denis Kirjanov6867b172014-12-01 13:12:25 +03003110 "DIV_MOD_KX",
Andrew Mortonece80492014-05-22 10:16:46 -07003111 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003112 BPF_STMT(BPF_LD | BPF_IMM, 8),
3113 BPF_STMT(BPF_ALU | BPF_DIV | BPF_K, 2),
3114 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3115 BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff),
3116 BPF_STMT(BPF_ALU | BPF_DIV | BPF_X, 0),
3117 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3118 BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff),
3119 BPF_STMT(BPF_ALU | BPF_DIV | BPF_K, 0x70000000),
Denis Kirjanov6867b172014-12-01 13:12:25 +03003120 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3121 BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff),
3122 BPF_STMT(BPF_ALU | BPF_MOD | BPF_X, 0),
3123 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3124 BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff),
3125 BPF_STMT(BPF_ALU | BPF_MOD | BPF_K, 0x70000000),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003126 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3127 BPF_STMT(BPF_RET | BPF_A, 0)
3128 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003129 CLASSIC | FLAG_NO_DATA,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003130 { },
Denis Kirjanov6867b172014-12-01 13:12:25 +03003131 { { 0, 0x20000000 } }
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003132 },
3133 {
3134 "AND_OR_LSH_K",
Andrew Mortonece80492014-05-22 10:16:46 -07003135 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003136 BPF_STMT(BPF_LD | BPF_IMM, 0xff),
3137 BPF_STMT(BPF_ALU | BPF_AND | BPF_K, 0xf0),
3138 BPF_STMT(BPF_ALU | BPF_LSH | BPF_K, 27),
3139 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3140 BPF_STMT(BPF_LD | BPF_IMM, 0xf),
3141 BPF_STMT(BPF_ALU | BPF_OR | BPF_K, 0xf0),
3142 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3143 BPF_STMT(BPF_RET | BPF_A, 0)
3144 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003145 CLASSIC | FLAG_NO_DATA,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003146 { },
3147 { { 0, 0x800000ff }, { 1, 0x800000ff } },
3148 },
3149 {
Chema Gonzaleze9d94502014-05-30 10:15:12 -07003150 "LD_IMM_0",
3151 .u.insns = {
3152 BPF_STMT(BPF_LD | BPF_IMM, 0), /* ld #0 */
3153 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0, 1, 0),
3154 BPF_STMT(BPF_RET | BPF_K, 0),
3155 BPF_STMT(BPF_RET | BPF_K, 1),
3156 },
3157 CLASSIC,
3158 { },
3159 { { 1, 1 } },
3160 },
3161 {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003162 "LD_IND",
Andrew Mortonece80492014-05-22 10:16:46 -07003163 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003164 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3165 BPF_STMT(BPF_LD | BPF_H | BPF_IND, MAX_K),
3166 BPF_STMT(BPF_RET | BPF_K, 1)
3167 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003168 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003169 { },
3170 { { 1, 0 }, { 10, 0 }, { 60, 0 } },
3171 },
3172 {
3173 "LD_ABS",
Andrew Mortonece80492014-05-22 10:16:46 -07003174 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003175 BPF_STMT(BPF_LD | BPF_W | BPF_ABS, 1000),
3176 BPF_STMT(BPF_RET | BPF_K, 1)
3177 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003178 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003179 { },
3180 { { 1, 0 }, { 10, 0 }, { 60, 0 } },
3181 },
3182 {
3183 "LD_ABS_LL",
Andrew Mortonece80492014-05-22 10:16:46 -07003184 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003185 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, SKF_LL_OFF),
3186 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3187 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, SKF_LL_OFF + 1),
3188 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3189 BPF_STMT(BPF_RET | BPF_A, 0)
3190 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003191 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003192 { 1, 2, 3 },
3193 { { 1, 0 }, { 2, 3 } },
3194 },
3195 {
3196 "LD_IND_LL",
Andrew Mortonece80492014-05-22 10:16:46 -07003197 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003198 BPF_STMT(BPF_LD | BPF_IMM, SKF_LL_OFF - 1),
3199 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3200 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3201 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3202 BPF_STMT(BPF_LD | BPF_B | BPF_IND, 0),
3203 BPF_STMT(BPF_RET | BPF_A, 0)
3204 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003205 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003206 { 1, 2, 3, 0xff },
3207 { { 1, 1 }, { 3, 3 }, { 4, 0xff } },
3208 },
3209 {
3210 "LD_ABS_NET",
Andrew Mortonece80492014-05-22 10:16:46 -07003211 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003212 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, SKF_NET_OFF),
3213 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3214 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, SKF_NET_OFF + 1),
3215 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3216 BPF_STMT(BPF_RET | BPF_A, 0)
3217 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003218 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003219 { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 3 },
3220 { { 15, 0 }, { 16, 3 } },
3221 },
3222 {
3223 "LD_IND_NET",
Andrew Mortonece80492014-05-22 10:16:46 -07003224 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003225 BPF_STMT(BPF_LD | BPF_IMM, SKF_NET_OFF - 15),
3226 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3227 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
3228 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3229 BPF_STMT(BPF_LD | BPF_B | BPF_IND, 0),
3230 BPF_STMT(BPF_RET | BPF_A, 0)
3231 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003232 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003233 { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 3 },
3234 { { 14, 0 }, { 15, 1 }, { 17, 3 } },
3235 },
3236 {
3237 "LD_PKTTYPE",
Andrew Mortonece80492014-05-22 10:16:46 -07003238 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003239 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3240 SKF_AD_OFF + SKF_AD_PKTTYPE),
3241 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, SKB_TYPE, 1, 0),
3242 BPF_STMT(BPF_RET | BPF_K, 1),
3243 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3244 SKF_AD_OFF + SKF_AD_PKTTYPE),
3245 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, SKB_TYPE, 1, 0),
3246 BPF_STMT(BPF_RET | BPF_K, 1),
3247 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3248 SKF_AD_OFF + SKF_AD_PKTTYPE),
3249 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, SKB_TYPE, 1, 0),
3250 BPF_STMT(BPF_RET | BPF_K, 1),
3251 BPF_STMT(BPF_RET | BPF_A, 0)
3252 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003253 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003254 { },
3255 { { 1, 3 }, { 10, 3 } },
3256 },
3257 {
3258 "LD_MARK",
Andrew Mortonece80492014-05-22 10:16:46 -07003259 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003260 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3261 SKF_AD_OFF + SKF_AD_MARK),
3262 BPF_STMT(BPF_RET | BPF_A, 0)
3263 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003264 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003265 { },
3266 { { 1, SKB_MARK}, { 10, SKB_MARK} },
3267 },
3268 {
3269 "LD_RXHASH",
Andrew Mortonece80492014-05-22 10:16:46 -07003270 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003271 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3272 SKF_AD_OFF + SKF_AD_RXHASH),
3273 BPF_STMT(BPF_RET | BPF_A, 0)
3274 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003275 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003276 { },
3277 { { 1, SKB_HASH}, { 10, SKB_HASH} },
3278 },
3279 {
3280 "LD_QUEUE",
Andrew Mortonece80492014-05-22 10:16:46 -07003281 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003282 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3283 SKF_AD_OFF + SKF_AD_QUEUE),
3284 BPF_STMT(BPF_RET | BPF_A, 0)
3285 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003286 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003287 { },
3288 { { 1, SKB_QUEUE_MAP }, { 10, SKB_QUEUE_MAP } },
3289 },
3290 {
3291 "LD_PROTOCOL",
Andrew Mortonece80492014-05-22 10:16:46 -07003292 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003293 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 1),
3294 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 20, 1, 0),
3295 BPF_STMT(BPF_RET | BPF_K, 0),
3296 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3297 SKF_AD_OFF + SKF_AD_PROTOCOL),
3298 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3299 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2),
3300 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 30, 1, 0),
3301 BPF_STMT(BPF_RET | BPF_K, 0),
3302 BPF_STMT(BPF_MISC | BPF_TXA, 0),
3303 BPF_STMT(BPF_RET | BPF_A, 0)
3304 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003305 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003306 { 10, 20, 30 },
3307 { { 10, ETH_P_IP }, { 100, ETH_P_IP } },
3308 },
3309 {
3310 "LD_VLAN_TAG",
Andrew Mortonece80492014-05-22 10:16:46 -07003311 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003312 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3313 SKF_AD_OFF + SKF_AD_VLAN_TAG),
3314 BPF_STMT(BPF_RET | BPF_A, 0)
3315 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003316 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003317 { },
3318 {
Michał Mirosław0c4b2d32018-11-10 19:58:36 +01003319 { 1, SKB_VLAN_TCI },
3320 { 10, SKB_VLAN_TCI }
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003321 },
3322 },
3323 {
3324 "LD_VLAN_TAG_PRESENT",
Andrew Mortonece80492014-05-22 10:16:46 -07003325 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003326 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3327 SKF_AD_OFF + SKF_AD_VLAN_TAG_PRESENT),
3328 BPF_STMT(BPF_RET | BPF_A, 0)
3329 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003330 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003331 { },
3332 {
Michał Mirosław0c4b2d32018-11-10 19:58:36 +01003333 { 1, SKB_VLAN_PRESENT },
3334 { 10, SKB_VLAN_PRESENT }
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003335 },
3336 },
3337 {
3338 "LD_IFINDEX",
Andrew Mortonece80492014-05-22 10:16:46 -07003339 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003340 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3341 SKF_AD_OFF + SKF_AD_IFINDEX),
3342 BPF_STMT(BPF_RET | BPF_A, 0)
3343 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003344 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003345 { },
3346 { { 1, SKB_DEV_IFINDEX }, { 10, SKB_DEV_IFINDEX } },
3347 },
3348 {
3349 "LD_HATYPE",
Andrew Mortonece80492014-05-22 10:16:46 -07003350 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003351 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3352 SKF_AD_OFF + SKF_AD_HATYPE),
3353 BPF_STMT(BPF_RET | BPF_A, 0)
3354 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003355 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003356 { },
3357 { { 1, SKB_DEV_TYPE }, { 10, SKB_DEV_TYPE } },
3358 },
3359 {
3360 "LD_CPU",
Andrew Mortonece80492014-05-22 10:16:46 -07003361 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003362 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3363 SKF_AD_OFF + SKF_AD_CPU),
3364 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3365 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3366 SKF_AD_OFF + SKF_AD_CPU),
3367 BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0),
3368 BPF_STMT(BPF_RET | BPF_A, 0)
3369 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003370 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003371 { },
3372 { { 1, 0 }, { 10, 0 } },
3373 },
3374 {
3375 "LD_NLATTR",
Andrew Mortonece80492014-05-22 10:16:46 -07003376 .u.insns = {
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003377 BPF_STMT(BPF_LDX | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003378 BPF_STMT(BPF_MISC | BPF_TXA, 0),
3379 BPF_STMT(BPF_LDX | BPF_IMM, 3),
3380 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3381 SKF_AD_OFF + SKF_AD_NLATTR),
3382 BPF_STMT(BPF_RET | BPF_A, 0)
3383 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003384 CLASSIC,
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003385#ifdef __BIG_ENDIAN
3386 { 0xff, 0xff, 0, 4, 0, 2, 0, 4, 0, 3 },
3387#else
3388 { 0xff, 0xff, 4, 0, 2, 0, 4, 0, 3, 0 },
3389#endif
3390 { { 4, 0 }, { 20, 6 } },
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003391 },
3392 {
3393 "LD_NLATTR_NEST",
Andrew Mortonece80492014-05-22 10:16:46 -07003394 .u.insns = {
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003395 BPF_STMT(BPF_LD | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003396 BPF_STMT(BPF_LDX | BPF_IMM, 3),
3397 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3398 SKF_AD_OFF + SKF_AD_NLATTR_NEST),
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003399 BPF_STMT(BPF_LD | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003400 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3401 SKF_AD_OFF + SKF_AD_NLATTR_NEST),
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003402 BPF_STMT(BPF_LD | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003403 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3404 SKF_AD_OFF + SKF_AD_NLATTR_NEST),
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003405 BPF_STMT(BPF_LD | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003406 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3407 SKF_AD_OFF + SKF_AD_NLATTR_NEST),
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003408 BPF_STMT(BPF_LD | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003409 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3410 SKF_AD_OFF + SKF_AD_NLATTR_NEST),
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003411 BPF_STMT(BPF_LD | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003412 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3413 SKF_AD_OFF + SKF_AD_NLATTR_NEST),
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003414 BPF_STMT(BPF_LD | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003415 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3416 SKF_AD_OFF + SKF_AD_NLATTR_NEST),
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003417 BPF_STMT(BPF_LD | BPF_IMM, 2),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003418 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3419 SKF_AD_OFF + SKF_AD_NLATTR_NEST),
3420 BPF_STMT(BPF_RET | BPF_A, 0)
3421 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003422 CLASSIC,
Alexei Starovoitovdf6d0f9832014-06-06 17:48:20 -07003423#ifdef __BIG_ENDIAN
3424 { 0xff, 0xff, 0, 12, 0, 1, 0, 4, 0, 2, 0, 4, 0, 3 },
3425#else
3426 { 0xff, 0xff, 12, 0, 1, 0, 4, 0, 2, 0, 4, 0, 3, 0 },
3427#endif
3428 { { 4, 0 }, { 20, 10 } },
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003429 },
3430 {
3431 "LD_PAYLOAD_OFF",
Andrew Mortonece80492014-05-22 10:16:46 -07003432 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003433 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3434 SKF_AD_OFF + SKF_AD_PAY_OFFSET),
3435 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3436 SKF_AD_OFF + SKF_AD_PAY_OFFSET),
3437 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3438 SKF_AD_OFF + SKF_AD_PAY_OFFSET),
3439 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3440 SKF_AD_OFF + SKF_AD_PAY_OFFSET),
3441 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3442 SKF_AD_OFF + SKF_AD_PAY_OFFSET),
3443 BPF_STMT(BPF_RET | BPF_A, 0)
3444 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003445 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003446 /* 00:00:00:00:00:00 > 00:00:00:00:00:00, ethtype IPv4 (0x0800),
3447 * length 98: 127.0.0.1 > 127.0.0.1: ICMP echo request,
3448 * id 9737, seq 1, length 64
3449 */
3450 { 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
3451 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
3452 0x08, 0x00,
3453 0x45, 0x00, 0x00, 0x54, 0xac, 0x8b, 0x40, 0x00, 0x40,
3454 0x01, 0x90, 0x1b, 0x7f, 0x00, 0x00, 0x01 },
3455 { { 30, 0 }, { 100, 42 } },
3456 },
3457 {
3458 "LD_ANC_XOR",
Andrew Mortonece80492014-05-22 10:16:46 -07003459 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003460 BPF_STMT(BPF_LD | BPF_IMM, 10),
3461 BPF_STMT(BPF_LDX | BPF_IMM, 300),
3462 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
3463 SKF_AD_OFF + SKF_AD_ALU_XOR_X),
3464 BPF_STMT(BPF_RET | BPF_A, 0)
3465 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003466 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003467 { },
Nathan Chancellor09845432019-08-18 21:34:20 -07003468 { { 4, 0xA ^ 300 }, { 20, 0xA ^ 300 } },
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003469 },
3470 {
3471 "SPILL_FILL",
Andrew Mortonece80492014-05-22 10:16:46 -07003472 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003473 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3474 BPF_STMT(BPF_LD | BPF_IMM, 2),
3475 BPF_STMT(BPF_ALU | BPF_RSH, 1),
3476 BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0),
3477 BPF_STMT(BPF_ST, 1), /* M1 = 1 ^ len */
3478 BPF_STMT(BPF_ALU | BPF_XOR | BPF_K, 0x80000000),
3479 BPF_STMT(BPF_ST, 2), /* M2 = 1 ^ len ^ 0x80000000 */
3480 BPF_STMT(BPF_STX, 15), /* M3 = len */
3481 BPF_STMT(BPF_LDX | BPF_MEM, 1),
3482 BPF_STMT(BPF_LD | BPF_MEM, 2),
3483 BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0),
3484 BPF_STMT(BPF_LDX | BPF_MEM, 15),
3485 BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0),
3486 BPF_STMT(BPF_RET | BPF_A, 0)
3487 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003488 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003489 { },
3490 { { 1, 0x80000001 }, { 2, 0x80000002 }, { 60, 0x80000000 ^ 60 } }
3491 },
3492 {
3493 "JEQ",
Andrew Mortonece80492014-05-22 10:16:46 -07003494 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003495 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3496 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2),
3497 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_X, 0, 0, 1),
3498 BPF_STMT(BPF_RET | BPF_K, 1),
3499 BPF_STMT(BPF_RET | BPF_K, MAX_K)
3500 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003501 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003502 { 3, 3, 3, 3, 3 },
3503 { { 1, 0 }, { 3, 1 }, { 4, MAX_K } },
3504 },
3505 {
3506 "JGT",
Andrew Mortonece80492014-05-22 10:16:46 -07003507 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003508 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3509 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2),
3510 BPF_JUMP(BPF_JMP | BPF_JGT | BPF_X, 0, 0, 1),
3511 BPF_STMT(BPF_RET | BPF_K, 1),
3512 BPF_STMT(BPF_RET | BPF_K, MAX_K)
3513 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003514 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003515 { 4, 4, 4, 3, 3 },
3516 { { 2, 0 }, { 3, 1 }, { 4, MAX_K } },
3517 },
3518 {
Daniel Borkmann92b31a92017-08-10 01:39:55 +02003519 "JGE (jt 0), test 1",
3520 .u.insns = {
3521 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3522 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2),
3523 BPF_JUMP(BPF_JMP | BPF_JGE | BPF_X, 0, 0, 1),
3524 BPF_STMT(BPF_RET | BPF_K, 1),
3525 BPF_STMT(BPF_RET | BPF_K, MAX_K)
3526 },
3527 CLASSIC,
3528 { 4, 4, 4, 3, 3 },
3529 { { 2, 0 }, { 3, 1 }, { 4, 1 } },
3530 },
3531 {
3532 "JGE (jt 0), test 2",
3533 .u.insns = {
3534 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3535 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2),
3536 BPF_JUMP(BPF_JMP | BPF_JGE | BPF_X, 0, 0, 1),
3537 BPF_STMT(BPF_RET | BPF_K, 1),
3538 BPF_STMT(BPF_RET | BPF_K, MAX_K)
3539 },
3540 CLASSIC,
3541 { 4, 4, 5, 3, 3 },
3542 { { 4, 1 }, { 5, 1 }, { 6, MAX_K } },
3543 },
3544 {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003545 "JGE",
Andrew Mortonece80492014-05-22 10:16:46 -07003546 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003547 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3548 BPF_STMT(BPF_LD | BPF_B | BPF_IND, MAX_K),
3549 BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 1, 1, 0),
3550 BPF_STMT(BPF_RET | BPF_K, 10),
3551 BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 2, 1, 0),
3552 BPF_STMT(BPF_RET | BPF_K, 20),
3553 BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 3, 1, 0),
3554 BPF_STMT(BPF_RET | BPF_K, 30),
3555 BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 4, 1, 0),
3556 BPF_STMT(BPF_RET | BPF_K, 40),
3557 BPF_STMT(BPF_RET | BPF_K, MAX_K)
3558 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003559 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003560 { 1, 2, 3, 4, 5 },
3561 { { 1, 20 }, { 3, 40 }, { 5, MAX_K } },
3562 },
3563 {
3564 "JSET",
Andrew Mortonece80492014-05-22 10:16:46 -07003565 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003566 BPF_JUMP(BPF_JMP | BPF_JA, 0, 0, 0),
3567 BPF_JUMP(BPF_JMP | BPF_JA, 1, 1, 1),
3568 BPF_JUMP(BPF_JMP | BPF_JA, 0, 0, 0),
3569 BPF_JUMP(BPF_JMP | BPF_JA, 0, 0, 0),
3570 BPF_STMT(BPF_LDX | BPF_LEN, 0),
3571 BPF_STMT(BPF_MISC | BPF_TXA, 0),
3572 BPF_STMT(BPF_ALU | BPF_SUB | BPF_K, 4),
3573 BPF_STMT(BPF_MISC | BPF_TAX, 0),
3574 BPF_STMT(BPF_LD | BPF_W | BPF_IND, 0),
3575 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 1, 0, 1),
3576 BPF_STMT(BPF_RET | BPF_K, 10),
3577 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0x80000000, 0, 1),
3578 BPF_STMT(BPF_RET | BPF_K, 20),
3579 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0),
3580 BPF_STMT(BPF_RET | BPF_K, 30),
3581 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0),
3582 BPF_STMT(BPF_RET | BPF_K, 30),
3583 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0),
3584 BPF_STMT(BPF_RET | BPF_K, 30),
3585 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0),
3586 BPF_STMT(BPF_RET | BPF_K, 30),
3587 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0),
3588 BPF_STMT(BPF_RET | BPF_K, 30),
3589 BPF_STMT(BPF_RET | BPF_K, MAX_K)
3590 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003591 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003592 { 0, 0xAA, 0x55, 1 },
3593 { { 4, 10 }, { 5, 20 }, { 6, MAX_K } },
3594 },
3595 {
Alexei Starovoitov64a89462014-05-08 14:10:52 -07003596 "tcpdump port 22",
Andrew Mortonece80492014-05-22 10:16:46 -07003597 .u.insns = {
Daniel Borkmannce25b682014-05-26 20:17:35 +02003598 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 12),
3599 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x86dd, 0, 8), /* IPv6 */
3600 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 20),
3601 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x84, 2, 0),
3602 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x6, 1, 0),
3603 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x11, 0, 17),
3604 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 54),
3605 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 14, 0),
3606 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 56),
3607 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 12, 13),
3608 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x0800, 0, 12), /* IPv4 */
3609 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 23),
3610 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x84, 2, 0),
3611 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x6, 1, 0),
3612 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x11, 0, 8),
3613 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 20),
3614 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0x1fff, 6, 0),
3615 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 14),
3616 BPF_STMT(BPF_LD | BPF_H | BPF_IND, 14),
3617 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 2, 0),
3618 BPF_STMT(BPF_LD | BPF_H | BPF_IND, 16),
3619 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 0, 1),
3620 BPF_STMT(BPF_RET | BPF_K, 0xffff),
3621 BPF_STMT(BPF_RET | BPF_K, 0),
Alexei Starovoitov64a89462014-05-08 14:10:52 -07003622 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003623 CLASSIC,
Alexei Starovoitov64a89462014-05-08 14:10:52 -07003624 /* 3c:07:54:43:e5:76 > 10:bf:48:d6:43:d6, ethertype IPv4(0x0800)
3625 * length 114: 10.1.1.149.49700 > 10.1.2.10.22: Flags [P.],
3626 * seq 1305692979:1305693027, ack 3650467037, win 65535,
3627 * options [nop,nop,TS val 2502645400 ecr 3971138], length 48
3628 */
3629 { 0x10, 0xbf, 0x48, 0xd6, 0x43, 0xd6,
3630 0x3c, 0x07, 0x54, 0x43, 0xe5, 0x76,
3631 0x08, 0x00,
3632 0x45, 0x10, 0x00, 0x64, 0x75, 0xb5,
3633 0x40, 0x00, 0x40, 0x06, 0xad, 0x2e, /* IP header */
3634 0x0a, 0x01, 0x01, 0x95, /* ip src */
3635 0x0a, 0x01, 0x02, 0x0a, /* ip dst */
3636 0xc2, 0x24,
3637 0x00, 0x16 /* dst port */ },
3638 { { 10, 0 }, { 30, 0 }, { 100, 65535 } },
3639 },
3640 {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003641 "tcpdump complex",
Andrew Mortonece80492014-05-22 10:16:46 -07003642 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003643 /* tcpdump -nei eth0 'tcp port 22 and (((ip[2:2] -
3644 * ((ip[0]&0xf)<<2)) - ((tcp[12]&0xf0)>>2)) != 0) and
3645 * (len > 115 or len < 30000000000)' -d
3646 */
Daniel Borkmannce25b682014-05-26 20:17:35 +02003647 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 12),
3648 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x86dd, 30, 0),
3649 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x800, 0, 29),
3650 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 23),
3651 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x6, 0, 27),
3652 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 20),
3653 BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0x1fff, 25, 0),
3654 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 14),
3655 BPF_STMT(BPF_LD | BPF_H | BPF_IND, 14),
3656 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 2, 0),
3657 BPF_STMT(BPF_LD | BPF_H | BPF_IND, 16),
3658 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 0, 20),
3659 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 16),
3660 BPF_STMT(BPF_ST, 1),
3661 BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 14),
3662 BPF_STMT(BPF_ALU | BPF_AND | BPF_K, 0xf),
3663 BPF_STMT(BPF_ALU | BPF_LSH | BPF_K, 2),
3664 BPF_STMT(BPF_MISC | BPF_TAX, 0x5), /* libpcap emits K on TAX */
3665 BPF_STMT(BPF_LD | BPF_MEM, 1),
3666 BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0),
3667 BPF_STMT(BPF_ST, 5),
3668 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 14),
3669 BPF_STMT(BPF_LD | BPF_B | BPF_IND, 26),
3670 BPF_STMT(BPF_ALU | BPF_AND | BPF_K, 0xf0),
3671 BPF_STMT(BPF_ALU | BPF_RSH | BPF_K, 2),
3672 BPF_STMT(BPF_MISC | BPF_TAX, 0x9), /* libpcap emits K on TAX */
3673 BPF_STMT(BPF_LD | BPF_MEM, 5),
3674 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_X, 0, 4, 0),
3675 BPF_STMT(BPF_LD | BPF_LEN, 0),
3676 BPF_JUMP(BPF_JMP | BPF_JGT | BPF_K, 0x73, 1, 0),
3677 BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 0xfc23ac00, 1, 0),
3678 BPF_STMT(BPF_RET | BPF_K, 0xffff),
3679 BPF_STMT(BPF_RET | BPF_K, 0),
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003680 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003681 CLASSIC,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003682 { 0x10, 0xbf, 0x48, 0xd6, 0x43, 0xd6,
3683 0x3c, 0x07, 0x54, 0x43, 0xe5, 0x76,
3684 0x08, 0x00,
3685 0x45, 0x10, 0x00, 0x64, 0x75, 0xb5,
3686 0x40, 0x00, 0x40, 0x06, 0xad, 0x2e, /* IP header */
3687 0x0a, 0x01, 0x01, 0x95, /* ip src */
3688 0x0a, 0x01, 0x02, 0x0a, /* ip dst */
3689 0xc2, 0x24,
3690 0x00, 0x16 /* dst port */ },
3691 { { 10, 0 }, { 30, 0 }, { 100, 65535 } },
3692 },
3693 {
3694 "RET_A",
Andrew Mortonece80492014-05-22 10:16:46 -07003695 .u.insns = {
Zhen Lei53b0fe32021-07-07 18:07:28 -07003696 /* check that uninitialized X and A contain zeros */
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003697 BPF_STMT(BPF_MISC | BPF_TXA, 0),
3698 BPF_STMT(BPF_RET | BPF_A, 0)
3699 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003700 CLASSIC,
3701 { },
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003702 { {1, 0}, {2, 0} },
3703 },
3704 {
3705 "INT: ADD trivial",
Andrew Mortonece80492014-05-22 10:16:46 -07003706 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003707 BPF_ALU64_IMM(BPF_MOV, R1, 1),
3708 BPF_ALU64_IMM(BPF_ADD, R1, 2),
3709 BPF_ALU64_IMM(BPF_MOV, R2, 3),
3710 BPF_ALU64_REG(BPF_SUB, R1, R2),
3711 BPF_ALU64_IMM(BPF_ADD, R1, -1),
3712 BPF_ALU64_IMM(BPF_MUL, R1, 3),
3713 BPF_ALU64_REG(BPF_MOV, R0, R1),
3714 BPF_EXIT_INSN(),
3715 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003716 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003717 { },
3718 { { 0, 0xfffffffd } }
3719 },
3720 {
3721 "INT: MUL_X",
Andrew Mortonece80492014-05-22 10:16:46 -07003722 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003723 BPF_ALU64_IMM(BPF_MOV, R0, -1),
3724 BPF_ALU64_IMM(BPF_MOV, R1, -1),
3725 BPF_ALU64_IMM(BPF_MOV, R2, 3),
3726 BPF_ALU64_REG(BPF_MUL, R1, R2),
3727 BPF_JMP_IMM(BPF_JEQ, R1, 0xfffffffd, 1),
3728 BPF_EXIT_INSN(),
3729 BPF_ALU64_IMM(BPF_MOV, R0, 1),
3730 BPF_EXIT_INSN(),
3731 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003732 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003733 { },
3734 { { 0, 1 } }
3735 },
3736 {
3737 "INT: MUL_X2",
Andrew Mortonece80492014-05-22 10:16:46 -07003738 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003739 BPF_ALU32_IMM(BPF_MOV, R0, -1),
3740 BPF_ALU32_IMM(BPF_MOV, R1, -1),
3741 BPF_ALU32_IMM(BPF_MOV, R2, 3),
3742 BPF_ALU64_REG(BPF_MUL, R1, R2),
3743 BPF_ALU64_IMM(BPF_RSH, R1, 8),
3744 BPF_JMP_IMM(BPF_JEQ, R1, 0x2ffffff, 1),
3745 BPF_EXIT_INSN(),
3746 BPF_ALU32_IMM(BPF_MOV, R0, 1),
3747 BPF_EXIT_INSN(),
3748 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003749 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003750 { },
3751 { { 0, 1 } }
3752 },
3753 {
3754 "INT: MUL32_X",
Andrew Mortonece80492014-05-22 10:16:46 -07003755 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003756 BPF_ALU32_IMM(BPF_MOV, R0, -1),
3757 BPF_ALU64_IMM(BPF_MOV, R1, -1),
3758 BPF_ALU32_IMM(BPF_MOV, R2, 3),
3759 BPF_ALU32_REG(BPF_MUL, R1, R2),
3760 BPF_ALU64_IMM(BPF_RSH, R1, 8),
3761 BPF_JMP_IMM(BPF_JEQ, R1, 0xffffff, 1),
3762 BPF_EXIT_INSN(),
3763 BPF_ALU32_IMM(BPF_MOV, R0, 1),
3764 BPF_EXIT_INSN(),
3765 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003766 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003767 { },
3768 { { 0, 1 } }
3769 },
3770 {
3771 /* Have to test all register combinations, since
3772 * JITing of different registers will produce
3773 * different asm code.
3774 */
3775 "INT: ADD 64-bit",
Andrew Mortonece80492014-05-22 10:16:46 -07003776 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003777 BPF_ALU64_IMM(BPF_MOV, R0, 0),
3778 BPF_ALU64_IMM(BPF_MOV, R1, 1),
3779 BPF_ALU64_IMM(BPF_MOV, R2, 2),
3780 BPF_ALU64_IMM(BPF_MOV, R3, 3),
3781 BPF_ALU64_IMM(BPF_MOV, R4, 4),
3782 BPF_ALU64_IMM(BPF_MOV, R5, 5),
3783 BPF_ALU64_IMM(BPF_MOV, R6, 6),
3784 BPF_ALU64_IMM(BPF_MOV, R7, 7),
3785 BPF_ALU64_IMM(BPF_MOV, R8, 8),
3786 BPF_ALU64_IMM(BPF_MOV, R9, 9),
3787 BPF_ALU64_IMM(BPF_ADD, R0, 20),
3788 BPF_ALU64_IMM(BPF_ADD, R1, 20),
3789 BPF_ALU64_IMM(BPF_ADD, R2, 20),
3790 BPF_ALU64_IMM(BPF_ADD, R3, 20),
3791 BPF_ALU64_IMM(BPF_ADD, R4, 20),
3792 BPF_ALU64_IMM(BPF_ADD, R5, 20),
3793 BPF_ALU64_IMM(BPF_ADD, R6, 20),
3794 BPF_ALU64_IMM(BPF_ADD, R7, 20),
3795 BPF_ALU64_IMM(BPF_ADD, R8, 20),
3796 BPF_ALU64_IMM(BPF_ADD, R9, 20),
3797 BPF_ALU64_IMM(BPF_SUB, R0, 10),
3798 BPF_ALU64_IMM(BPF_SUB, R1, 10),
3799 BPF_ALU64_IMM(BPF_SUB, R2, 10),
3800 BPF_ALU64_IMM(BPF_SUB, R3, 10),
3801 BPF_ALU64_IMM(BPF_SUB, R4, 10),
3802 BPF_ALU64_IMM(BPF_SUB, R5, 10),
3803 BPF_ALU64_IMM(BPF_SUB, R6, 10),
3804 BPF_ALU64_IMM(BPF_SUB, R7, 10),
3805 BPF_ALU64_IMM(BPF_SUB, R8, 10),
3806 BPF_ALU64_IMM(BPF_SUB, R9, 10),
3807 BPF_ALU64_REG(BPF_ADD, R0, R0),
3808 BPF_ALU64_REG(BPF_ADD, R0, R1),
3809 BPF_ALU64_REG(BPF_ADD, R0, R2),
3810 BPF_ALU64_REG(BPF_ADD, R0, R3),
3811 BPF_ALU64_REG(BPF_ADD, R0, R4),
3812 BPF_ALU64_REG(BPF_ADD, R0, R5),
3813 BPF_ALU64_REG(BPF_ADD, R0, R6),
3814 BPF_ALU64_REG(BPF_ADD, R0, R7),
3815 BPF_ALU64_REG(BPF_ADD, R0, R8),
3816 BPF_ALU64_REG(BPF_ADD, R0, R9), /* R0 == 155 */
3817 BPF_JMP_IMM(BPF_JEQ, R0, 155, 1),
3818 BPF_EXIT_INSN(),
3819 BPF_ALU64_REG(BPF_ADD, R1, R0),
3820 BPF_ALU64_REG(BPF_ADD, R1, R1),
3821 BPF_ALU64_REG(BPF_ADD, R1, R2),
3822 BPF_ALU64_REG(BPF_ADD, R1, R3),
3823 BPF_ALU64_REG(BPF_ADD, R1, R4),
3824 BPF_ALU64_REG(BPF_ADD, R1, R5),
3825 BPF_ALU64_REG(BPF_ADD, R1, R6),
3826 BPF_ALU64_REG(BPF_ADD, R1, R7),
3827 BPF_ALU64_REG(BPF_ADD, R1, R8),
3828 BPF_ALU64_REG(BPF_ADD, R1, R9), /* R1 == 456 */
3829 BPF_JMP_IMM(BPF_JEQ, R1, 456, 1),
3830 BPF_EXIT_INSN(),
3831 BPF_ALU64_REG(BPF_ADD, R2, R0),
3832 BPF_ALU64_REG(BPF_ADD, R2, R1),
3833 BPF_ALU64_REG(BPF_ADD, R2, R2),
3834 BPF_ALU64_REG(BPF_ADD, R2, R3),
3835 BPF_ALU64_REG(BPF_ADD, R2, R4),
3836 BPF_ALU64_REG(BPF_ADD, R2, R5),
3837 BPF_ALU64_REG(BPF_ADD, R2, R6),
3838 BPF_ALU64_REG(BPF_ADD, R2, R7),
3839 BPF_ALU64_REG(BPF_ADD, R2, R8),
3840 BPF_ALU64_REG(BPF_ADD, R2, R9), /* R2 == 1358 */
3841 BPF_JMP_IMM(BPF_JEQ, R2, 1358, 1),
3842 BPF_EXIT_INSN(),
3843 BPF_ALU64_REG(BPF_ADD, R3, R0),
3844 BPF_ALU64_REG(BPF_ADD, R3, R1),
3845 BPF_ALU64_REG(BPF_ADD, R3, R2),
3846 BPF_ALU64_REG(BPF_ADD, R3, R3),
3847 BPF_ALU64_REG(BPF_ADD, R3, R4),
3848 BPF_ALU64_REG(BPF_ADD, R3, R5),
3849 BPF_ALU64_REG(BPF_ADD, R3, R6),
3850 BPF_ALU64_REG(BPF_ADD, R3, R7),
3851 BPF_ALU64_REG(BPF_ADD, R3, R8),
3852 BPF_ALU64_REG(BPF_ADD, R3, R9), /* R3 == 4063 */
3853 BPF_JMP_IMM(BPF_JEQ, R3, 4063, 1),
3854 BPF_EXIT_INSN(),
3855 BPF_ALU64_REG(BPF_ADD, R4, R0),
3856 BPF_ALU64_REG(BPF_ADD, R4, R1),
3857 BPF_ALU64_REG(BPF_ADD, R4, R2),
3858 BPF_ALU64_REG(BPF_ADD, R4, R3),
3859 BPF_ALU64_REG(BPF_ADD, R4, R4),
3860 BPF_ALU64_REG(BPF_ADD, R4, R5),
3861 BPF_ALU64_REG(BPF_ADD, R4, R6),
3862 BPF_ALU64_REG(BPF_ADD, R4, R7),
3863 BPF_ALU64_REG(BPF_ADD, R4, R8),
3864 BPF_ALU64_REG(BPF_ADD, R4, R9), /* R4 == 12177 */
3865 BPF_JMP_IMM(BPF_JEQ, R4, 12177, 1),
3866 BPF_EXIT_INSN(),
3867 BPF_ALU64_REG(BPF_ADD, R5, R0),
3868 BPF_ALU64_REG(BPF_ADD, R5, R1),
3869 BPF_ALU64_REG(BPF_ADD, R5, R2),
3870 BPF_ALU64_REG(BPF_ADD, R5, R3),
3871 BPF_ALU64_REG(BPF_ADD, R5, R4),
3872 BPF_ALU64_REG(BPF_ADD, R5, R5),
3873 BPF_ALU64_REG(BPF_ADD, R5, R6),
3874 BPF_ALU64_REG(BPF_ADD, R5, R7),
3875 BPF_ALU64_REG(BPF_ADD, R5, R8),
3876 BPF_ALU64_REG(BPF_ADD, R5, R9), /* R5 == 36518 */
3877 BPF_JMP_IMM(BPF_JEQ, R5, 36518, 1),
3878 BPF_EXIT_INSN(),
3879 BPF_ALU64_REG(BPF_ADD, R6, R0),
3880 BPF_ALU64_REG(BPF_ADD, R6, R1),
3881 BPF_ALU64_REG(BPF_ADD, R6, R2),
3882 BPF_ALU64_REG(BPF_ADD, R6, R3),
3883 BPF_ALU64_REG(BPF_ADD, R6, R4),
3884 BPF_ALU64_REG(BPF_ADD, R6, R5),
3885 BPF_ALU64_REG(BPF_ADD, R6, R6),
3886 BPF_ALU64_REG(BPF_ADD, R6, R7),
3887 BPF_ALU64_REG(BPF_ADD, R6, R8),
3888 BPF_ALU64_REG(BPF_ADD, R6, R9), /* R6 == 109540 */
3889 BPF_JMP_IMM(BPF_JEQ, R6, 109540, 1),
3890 BPF_EXIT_INSN(),
3891 BPF_ALU64_REG(BPF_ADD, R7, R0),
3892 BPF_ALU64_REG(BPF_ADD, R7, R1),
3893 BPF_ALU64_REG(BPF_ADD, R7, R2),
3894 BPF_ALU64_REG(BPF_ADD, R7, R3),
3895 BPF_ALU64_REG(BPF_ADD, R7, R4),
3896 BPF_ALU64_REG(BPF_ADD, R7, R5),
3897 BPF_ALU64_REG(BPF_ADD, R7, R6),
3898 BPF_ALU64_REG(BPF_ADD, R7, R7),
3899 BPF_ALU64_REG(BPF_ADD, R7, R8),
3900 BPF_ALU64_REG(BPF_ADD, R7, R9), /* R7 == 328605 */
3901 BPF_JMP_IMM(BPF_JEQ, R7, 328605, 1),
3902 BPF_EXIT_INSN(),
3903 BPF_ALU64_REG(BPF_ADD, R8, R0),
3904 BPF_ALU64_REG(BPF_ADD, R8, R1),
3905 BPF_ALU64_REG(BPF_ADD, R8, R2),
3906 BPF_ALU64_REG(BPF_ADD, R8, R3),
3907 BPF_ALU64_REG(BPF_ADD, R8, R4),
3908 BPF_ALU64_REG(BPF_ADD, R8, R5),
3909 BPF_ALU64_REG(BPF_ADD, R8, R6),
3910 BPF_ALU64_REG(BPF_ADD, R8, R7),
3911 BPF_ALU64_REG(BPF_ADD, R8, R8),
3912 BPF_ALU64_REG(BPF_ADD, R8, R9), /* R8 == 985799 */
3913 BPF_JMP_IMM(BPF_JEQ, R8, 985799, 1),
3914 BPF_EXIT_INSN(),
3915 BPF_ALU64_REG(BPF_ADD, R9, R0),
3916 BPF_ALU64_REG(BPF_ADD, R9, R1),
3917 BPF_ALU64_REG(BPF_ADD, R9, R2),
3918 BPF_ALU64_REG(BPF_ADD, R9, R3),
3919 BPF_ALU64_REG(BPF_ADD, R9, R4),
3920 BPF_ALU64_REG(BPF_ADD, R9, R5),
3921 BPF_ALU64_REG(BPF_ADD, R9, R6),
3922 BPF_ALU64_REG(BPF_ADD, R9, R7),
3923 BPF_ALU64_REG(BPF_ADD, R9, R8),
3924 BPF_ALU64_REG(BPF_ADD, R9, R9), /* R9 == 2957380 */
3925 BPF_ALU64_REG(BPF_MOV, R0, R9),
3926 BPF_EXIT_INSN(),
3927 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02003928 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003929 { },
3930 { { 0, 2957380 } }
3931 },
3932 {
3933 "INT: ADD 32-bit",
Andrew Mortonece80492014-05-22 10:16:46 -07003934 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07003935 BPF_ALU32_IMM(BPF_MOV, R0, 20),
3936 BPF_ALU32_IMM(BPF_MOV, R1, 1),
3937 BPF_ALU32_IMM(BPF_MOV, R2, 2),
3938 BPF_ALU32_IMM(BPF_MOV, R3, 3),
3939 BPF_ALU32_IMM(BPF_MOV, R4, 4),
3940 BPF_ALU32_IMM(BPF_MOV, R5, 5),
3941 BPF_ALU32_IMM(BPF_MOV, R6, 6),
3942 BPF_ALU32_IMM(BPF_MOV, R7, 7),
3943 BPF_ALU32_IMM(BPF_MOV, R8, 8),
3944 BPF_ALU32_IMM(BPF_MOV, R9, 9),
3945 BPF_ALU64_IMM(BPF_ADD, R1, 10),
3946 BPF_ALU64_IMM(BPF_ADD, R2, 10),
3947 BPF_ALU64_IMM(BPF_ADD, R3, 10),
3948 BPF_ALU64_IMM(BPF_ADD, R4, 10),
3949 BPF_ALU64_IMM(BPF_ADD, R5, 10),
3950 BPF_ALU64_IMM(BPF_ADD, R6, 10),
3951 BPF_ALU64_IMM(BPF_ADD, R7, 10),
3952 BPF_ALU64_IMM(BPF_ADD, R8, 10),
3953 BPF_ALU64_IMM(BPF_ADD, R9, 10),
3954 BPF_ALU32_REG(BPF_ADD, R0, R1),
3955 BPF_ALU32_REG(BPF_ADD, R0, R2),
3956 BPF_ALU32_REG(BPF_ADD, R0, R3),
3957 BPF_ALU32_REG(BPF_ADD, R0, R4),
3958 BPF_ALU32_REG(BPF_ADD, R0, R5),
3959 BPF_ALU32_REG(BPF_ADD, R0, R6),
3960 BPF_ALU32_REG(BPF_ADD, R0, R7),
3961 BPF_ALU32_REG(BPF_ADD, R0, R8),
3962 BPF_ALU32_REG(BPF_ADD, R0, R9), /* R0 == 155 */
3963 BPF_JMP_IMM(BPF_JEQ, R0, 155, 1),
3964 BPF_EXIT_INSN(),
3965 BPF_ALU32_REG(BPF_ADD, R1, R0),
3966 BPF_ALU32_REG(BPF_ADD, R1, R1),
3967 BPF_ALU32_REG(BPF_ADD, R1, R2),
3968 BPF_ALU32_REG(BPF_ADD, R1, R3),
3969 BPF_ALU32_REG(BPF_ADD, R1, R4),
3970 BPF_ALU32_REG(BPF_ADD, R1, R5),
3971 BPF_ALU32_REG(BPF_ADD, R1, R6),
3972 BPF_ALU32_REG(BPF_ADD, R1, R7),
3973 BPF_ALU32_REG(BPF_ADD, R1, R8),
3974 BPF_ALU32_REG(BPF_ADD, R1, R9), /* R1 == 456 */
3975 BPF_JMP_IMM(BPF_JEQ, R1, 456, 1),
3976 BPF_EXIT_INSN(),
3977 BPF_ALU32_REG(BPF_ADD, R2, R0),
3978 BPF_ALU32_REG(BPF_ADD, R2, R1),
3979 BPF_ALU32_REG(BPF_ADD, R2, R2),
3980 BPF_ALU32_REG(BPF_ADD, R2, R3),
3981 BPF_ALU32_REG(BPF_ADD, R2, R4),
3982 BPF_ALU32_REG(BPF_ADD, R2, R5),
3983 BPF_ALU32_REG(BPF_ADD, R2, R6),
3984 BPF_ALU32_REG(BPF_ADD, R2, R7),
3985 BPF_ALU32_REG(BPF_ADD, R2, R8),
3986 BPF_ALU32_REG(BPF_ADD, R2, R9), /* R2 == 1358 */
3987 BPF_JMP_IMM(BPF_JEQ, R2, 1358, 1),
3988 BPF_EXIT_INSN(),
3989 BPF_ALU32_REG(BPF_ADD, R3, R0),
3990 BPF_ALU32_REG(BPF_ADD, R3, R1),
3991 BPF_ALU32_REG(BPF_ADD, R3, R2),
3992 BPF_ALU32_REG(BPF_ADD, R3, R3),
3993 BPF_ALU32_REG(BPF_ADD, R3, R4),
3994 BPF_ALU32_REG(BPF_ADD, R3, R5),
3995 BPF_ALU32_REG(BPF_ADD, R3, R6),
3996 BPF_ALU32_REG(BPF_ADD, R3, R7),
3997 BPF_ALU32_REG(BPF_ADD, R3, R8),
3998 BPF_ALU32_REG(BPF_ADD, R3, R9), /* R3 == 4063 */
3999 BPF_JMP_IMM(BPF_JEQ, R3, 4063, 1),
4000 BPF_EXIT_INSN(),
4001 BPF_ALU32_REG(BPF_ADD, R4, R0),
4002 BPF_ALU32_REG(BPF_ADD, R4, R1),
4003 BPF_ALU32_REG(BPF_ADD, R4, R2),
4004 BPF_ALU32_REG(BPF_ADD, R4, R3),
4005 BPF_ALU32_REG(BPF_ADD, R4, R4),
4006 BPF_ALU32_REG(BPF_ADD, R4, R5),
4007 BPF_ALU32_REG(BPF_ADD, R4, R6),
4008 BPF_ALU32_REG(BPF_ADD, R4, R7),
4009 BPF_ALU32_REG(BPF_ADD, R4, R8),
4010 BPF_ALU32_REG(BPF_ADD, R4, R9), /* R4 == 12177 */
4011 BPF_JMP_IMM(BPF_JEQ, R4, 12177, 1),
4012 BPF_EXIT_INSN(),
4013 BPF_ALU32_REG(BPF_ADD, R5, R0),
4014 BPF_ALU32_REG(BPF_ADD, R5, R1),
4015 BPF_ALU32_REG(BPF_ADD, R5, R2),
4016 BPF_ALU32_REG(BPF_ADD, R5, R3),
4017 BPF_ALU32_REG(BPF_ADD, R5, R4),
4018 BPF_ALU32_REG(BPF_ADD, R5, R5),
4019 BPF_ALU32_REG(BPF_ADD, R5, R6),
4020 BPF_ALU32_REG(BPF_ADD, R5, R7),
4021 BPF_ALU32_REG(BPF_ADD, R5, R8),
4022 BPF_ALU32_REG(BPF_ADD, R5, R9), /* R5 == 36518 */
4023 BPF_JMP_IMM(BPF_JEQ, R5, 36518, 1),
4024 BPF_EXIT_INSN(),
4025 BPF_ALU32_REG(BPF_ADD, R6, R0),
4026 BPF_ALU32_REG(BPF_ADD, R6, R1),
4027 BPF_ALU32_REG(BPF_ADD, R6, R2),
4028 BPF_ALU32_REG(BPF_ADD, R6, R3),
4029 BPF_ALU32_REG(BPF_ADD, R6, R4),
4030 BPF_ALU32_REG(BPF_ADD, R6, R5),
4031 BPF_ALU32_REG(BPF_ADD, R6, R6),
4032 BPF_ALU32_REG(BPF_ADD, R6, R7),
4033 BPF_ALU32_REG(BPF_ADD, R6, R8),
4034 BPF_ALU32_REG(BPF_ADD, R6, R9), /* R6 == 109540 */
4035 BPF_JMP_IMM(BPF_JEQ, R6, 109540, 1),
4036 BPF_EXIT_INSN(),
4037 BPF_ALU32_REG(BPF_ADD, R7, R0),
4038 BPF_ALU32_REG(BPF_ADD, R7, R1),
4039 BPF_ALU32_REG(BPF_ADD, R7, R2),
4040 BPF_ALU32_REG(BPF_ADD, R7, R3),
4041 BPF_ALU32_REG(BPF_ADD, R7, R4),
4042 BPF_ALU32_REG(BPF_ADD, R7, R5),
4043 BPF_ALU32_REG(BPF_ADD, R7, R6),
4044 BPF_ALU32_REG(BPF_ADD, R7, R7),
4045 BPF_ALU32_REG(BPF_ADD, R7, R8),
4046 BPF_ALU32_REG(BPF_ADD, R7, R9), /* R7 == 328605 */
4047 BPF_JMP_IMM(BPF_JEQ, R7, 328605, 1),
4048 BPF_EXIT_INSN(),
4049 BPF_ALU32_REG(BPF_ADD, R8, R0),
4050 BPF_ALU32_REG(BPF_ADD, R8, R1),
4051 BPF_ALU32_REG(BPF_ADD, R8, R2),
4052 BPF_ALU32_REG(BPF_ADD, R8, R3),
4053 BPF_ALU32_REG(BPF_ADD, R8, R4),
4054 BPF_ALU32_REG(BPF_ADD, R8, R5),
4055 BPF_ALU32_REG(BPF_ADD, R8, R6),
4056 BPF_ALU32_REG(BPF_ADD, R8, R7),
4057 BPF_ALU32_REG(BPF_ADD, R8, R8),
4058 BPF_ALU32_REG(BPF_ADD, R8, R9), /* R8 == 985799 */
4059 BPF_JMP_IMM(BPF_JEQ, R8, 985799, 1),
4060 BPF_EXIT_INSN(),
4061 BPF_ALU32_REG(BPF_ADD, R9, R0),
4062 BPF_ALU32_REG(BPF_ADD, R9, R1),
4063 BPF_ALU32_REG(BPF_ADD, R9, R2),
4064 BPF_ALU32_REG(BPF_ADD, R9, R3),
4065 BPF_ALU32_REG(BPF_ADD, R9, R4),
4066 BPF_ALU32_REG(BPF_ADD, R9, R5),
4067 BPF_ALU32_REG(BPF_ADD, R9, R6),
4068 BPF_ALU32_REG(BPF_ADD, R9, R7),
4069 BPF_ALU32_REG(BPF_ADD, R9, R8),
4070 BPF_ALU32_REG(BPF_ADD, R9, R9), /* R9 == 2957380 */
4071 BPF_ALU32_REG(BPF_MOV, R0, R9),
4072 BPF_EXIT_INSN(),
4073 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004074 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004075 { },
4076 { { 0, 2957380 } }
4077 },
4078 { /* Mainly checking JIT here. */
4079 "INT: SUB",
Andrew Mortonece80492014-05-22 10:16:46 -07004080 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004081 BPF_ALU64_IMM(BPF_MOV, R0, 0),
4082 BPF_ALU64_IMM(BPF_MOV, R1, 1),
4083 BPF_ALU64_IMM(BPF_MOV, R2, 2),
4084 BPF_ALU64_IMM(BPF_MOV, R3, 3),
4085 BPF_ALU64_IMM(BPF_MOV, R4, 4),
4086 BPF_ALU64_IMM(BPF_MOV, R5, 5),
4087 BPF_ALU64_IMM(BPF_MOV, R6, 6),
4088 BPF_ALU64_IMM(BPF_MOV, R7, 7),
4089 BPF_ALU64_IMM(BPF_MOV, R8, 8),
4090 BPF_ALU64_IMM(BPF_MOV, R9, 9),
4091 BPF_ALU64_REG(BPF_SUB, R0, R0),
4092 BPF_ALU64_REG(BPF_SUB, R0, R1),
4093 BPF_ALU64_REG(BPF_SUB, R0, R2),
4094 BPF_ALU64_REG(BPF_SUB, R0, R3),
4095 BPF_ALU64_REG(BPF_SUB, R0, R4),
4096 BPF_ALU64_REG(BPF_SUB, R0, R5),
4097 BPF_ALU64_REG(BPF_SUB, R0, R6),
4098 BPF_ALU64_REG(BPF_SUB, R0, R7),
4099 BPF_ALU64_REG(BPF_SUB, R0, R8),
4100 BPF_ALU64_REG(BPF_SUB, R0, R9),
4101 BPF_ALU64_IMM(BPF_SUB, R0, 10),
4102 BPF_JMP_IMM(BPF_JEQ, R0, -55, 1),
4103 BPF_EXIT_INSN(),
4104 BPF_ALU64_REG(BPF_SUB, R1, R0),
4105 BPF_ALU64_REG(BPF_SUB, R1, R2),
4106 BPF_ALU64_REG(BPF_SUB, R1, R3),
4107 BPF_ALU64_REG(BPF_SUB, R1, R4),
4108 BPF_ALU64_REG(BPF_SUB, R1, R5),
4109 BPF_ALU64_REG(BPF_SUB, R1, R6),
4110 BPF_ALU64_REG(BPF_SUB, R1, R7),
4111 BPF_ALU64_REG(BPF_SUB, R1, R8),
4112 BPF_ALU64_REG(BPF_SUB, R1, R9),
4113 BPF_ALU64_IMM(BPF_SUB, R1, 10),
4114 BPF_ALU64_REG(BPF_SUB, R2, R0),
4115 BPF_ALU64_REG(BPF_SUB, R2, R1),
4116 BPF_ALU64_REG(BPF_SUB, R2, R3),
4117 BPF_ALU64_REG(BPF_SUB, R2, R4),
4118 BPF_ALU64_REG(BPF_SUB, R2, R5),
4119 BPF_ALU64_REG(BPF_SUB, R2, R6),
4120 BPF_ALU64_REG(BPF_SUB, R2, R7),
4121 BPF_ALU64_REG(BPF_SUB, R2, R8),
4122 BPF_ALU64_REG(BPF_SUB, R2, R9),
4123 BPF_ALU64_IMM(BPF_SUB, R2, 10),
4124 BPF_ALU64_REG(BPF_SUB, R3, R0),
4125 BPF_ALU64_REG(BPF_SUB, R3, R1),
4126 BPF_ALU64_REG(BPF_SUB, R3, R2),
4127 BPF_ALU64_REG(BPF_SUB, R3, R4),
4128 BPF_ALU64_REG(BPF_SUB, R3, R5),
4129 BPF_ALU64_REG(BPF_SUB, R3, R6),
4130 BPF_ALU64_REG(BPF_SUB, R3, R7),
4131 BPF_ALU64_REG(BPF_SUB, R3, R8),
4132 BPF_ALU64_REG(BPF_SUB, R3, R9),
4133 BPF_ALU64_IMM(BPF_SUB, R3, 10),
4134 BPF_ALU64_REG(BPF_SUB, R4, R0),
4135 BPF_ALU64_REG(BPF_SUB, R4, R1),
4136 BPF_ALU64_REG(BPF_SUB, R4, R2),
4137 BPF_ALU64_REG(BPF_SUB, R4, R3),
4138 BPF_ALU64_REG(BPF_SUB, R4, R5),
4139 BPF_ALU64_REG(BPF_SUB, R4, R6),
4140 BPF_ALU64_REG(BPF_SUB, R4, R7),
4141 BPF_ALU64_REG(BPF_SUB, R4, R8),
4142 BPF_ALU64_REG(BPF_SUB, R4, R9),
4143 BPF_ALU64_IMM(BPF_SUB, R4, 10),
4144 BPF_ALU64_REG(BPF_SUB, R5, R0),
4145 BPF_ALU64_REG(BPF_SUB, R5, R1),
4146 BPF_ALU64_REG(BPF_SUB, R5, R2),
4147 BPF_ALU64_REG(BPF_SUB, R5, R3),
4148 BPF_ALU64_REG(BPF_SUB, R5, R4),
4149 BPF_ALU64_REG(BPF_SUB, R5, R6),
4150 BPF_ALU64_REG(BPF_SUB, R5, R7),
4151 BPF_ALU64_REG(BPF_SUB, R5, R8),
4152 BPF_ALU64_REG(BPF_SUB, R5, R9),
4153 BPF_ALU64_IMM(BPF_SUB, R5, 10),
4154 BPF_ALU64_REG(BPF_SUB, R6, R0),
4155 BPF_ALU64_REG(BPF_SUB, R6, R1),
4156 BPF_ALU64_REG(BPF_SUB, R6, R2),
4157 BPF_ALU64_REG(BPF_SUB, R6, R3),
4158 BPF_ALU64_REG(BPF_SUB, R6, R4),
4159 BPF_ALU64_REG(BPF_SUB, R6, R5),
4160 BPF_ALU64_REG(BPF_SUB, R6, R7),
4161 BPF_ALU64_REG(BPF_SUB, R6, R8),
4162 BPF_ALU64_REG(BPF_SUB, R6, R9),
4163 BPF_ALU64_IMM(BPF_SUB, R6, 10),
4164 BPF_ALU64_REG(BPF_SUB, R7, R0),
4165 BPF_ALU64_REG(BPF_SUB, R7, R1),
4166 BPF_ALU64_REG(BPF_SUB, R7, R2),
4167 BPF_ALU64_REG(BPF_SUB, R7, R3),
4168 BPF_ALU64_REG(BPF_SUB, R7, R4),
4169 BPF_ALU64_REG(BPF_SUB, R7, R5),
4170 BPF_ALU64_REG(BPF_SUB, R7, R6),
4171 BPF_ALU64_REG(BPF_SUB, R7, R8),
4172 BPF_ALU64_REG(BPF_SUB, R7, R9),
4173 BPF_ALU64_IMM(BPF_SUB, R7, 10),
4174 BPF_ALU64_REG(BPF_SUB, R8, R0),
4175 BPF_ALU64_REG(BPF_SUB, R8, R1),
4176 BPF_ALU64_REG(BPF_SUB, R8, R2),
4177 BPF_ALU64_REG(BPF_SUB, R8, R3),
4178 BPF_ALU64_REG(BPF_SUB, R8, R4),
4179 BPF_ALU64_REG(BPF_SUB, R8, R5),
4180 BPF_ALU64_REG(BPF_SUB, R8, R6),
4181 BPF_ALU64_REG(BPF_SUB, R8, R7),
4182 BPF_ALU64_REG(BPF_SUB, R8, R9),
4183 BPF_ALU64_IMM(BPF_SUB, R8, 10),
4184 BPF_ALU64_REG(BPF_SUB, R9, R0),
4185 BPF_ALU64_REG(BPF_SUB, R9, R1),
4186 BPF_ALU64_REG(BPF_SUB, R9, R2),
4187 BPF_ALU64_REG(BPF_SUB, R9, R3),
4188 BPF_ALU64_REG(BPF_SUB, R9, R4),
4189 BPF_ALU64_REG(BPF_SUB, R9, R5),
4190 BPF_ALU64_REG(BPF_SUB, R9, R6),
4191 BPF_ALU64_REG(BPF_SUB, R9, R7),
4192 BPF_ALU64_REG(BPF_SUB, R9, R8),
4193 BPF_ALU64_IMM(BPF_SUB, R9, 10),
4194 BPF_ALU64_IMM(BPF_SUB, R0, 10),
4195 BPF_ALU64_IMM(BPF_NEG, R0, 0),
4196 BPF_ALU64_REG(BPF_SUB, R0, R1),
4197 BPF_ALU64_REG(BPF_SUB, R0, R2),
4198 BPF_ALU64_REG(BPF_SUB, R0, R3),
4199 BPF_ALU64_REG(BPF_SUB, R0, R4),
4200 BPF_ALU64_REG(BPF_SUB, R0, R5),
4201 BPF_ALU64_REG(BPF_SUB, R0, R6),
4202 BPF_ALU64_REG(BPF_SUB, R0, R7),
4203 BPF_ALU64_REG(BPF_SUB, R0, R8),
4204 BPF_ALU64_REG(BPF_SUB, R0, R9),
4205 BPF_EXIT_INSN(),
4206 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004207 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004208 { },
4209 { { 0, 11 } }
4210 },
4211 { /* Mainly checking JIT here. */
4212 "INT: XOR",
Andrew Mortonece80492014-05-22 10:16:46 -07004213 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004214 BPF_ALU64_REG(BPF_SUB, R0, R0),
4215 BPF_ALU64_REG(BPF_XOR, R1, R1),
4216 BPF_JMP_REG(BPF_JEQ, R0, R1, 1),
4217 BPF_EXIT_INSN(),
4218 BPF_ALU64_IMM(BPF_MOV, R0, 10),
4219 BPF_ALU64_IMM(BPF_MOV, R1, -1),
4220 BPF_ALU64_REG(BPF_SUB, R1, R1),
4221 BPF_ALU64_REG(BPF_XOR, R2, R2),
4222 BPF_JMP_REG(BPF_JEQ, R1, R2, 1),
4223 BPF_EXIT_INSN(),
4224 BPF_ALU64_REG(BPF_SUB, R2, R2),
4225 BPF_ALU64_REG(BPF_XOR, R3, R3),
4226 BPF_ALU64_IMM(BPF_MOV, R0, 10),
4227 BPF_ALU64_IMM(BPF_MOV, R1, -1),
4228 BPF_JMP_REG(BPF_JEQ, R2, R3, 1),
4229 BPF_EXIT_INSN(),
4230 BPF_ALU64_REG(BPF_SUB, R3, R3),
4231 BPF_ALU64_REG(BPF_XOR, R4, R4),
4232 BPF_ALU64_IMM(BPF_MOV, R2, 1),
4233 BPF_ALU64_IMM(BPF_MOV, R5, -1),
4234 BPF_JMP_REG(BPF_JEQ, R3, R4, 1),
4235 BPF_EXIT_INSN(),
4236 BPF_ALU64_REG(BPF_SUB, R4, R4),
4237 BPF_ALU64_REG(BPF_XOR, R5, R5),
4238 BPF_ALU64_IMM(BPF_MOV, R3, 1),
4239 BPF_ALU64_IMM(BPF_MOV, R7, -1),
4240 BPF_JMP_REG(BPF_JEQ, R5, R4, 1),
4241 BPF_EXIT_INSN(),
4242 BPF_ALU64_IMM(BPF_MOV, R5, 1),
4243 BPF_ALU64_REG(BPF_SUB, R5, R5),
4244 BPF_ALU64_REG(BPF_XOR, R6, R6),
4245 BPF_ALU64_IMM(BPF_MOV, R1, 1),
4246 BPF_ALU64_IMM(BPF_MOV, R8, -1),
4247 BPF_JMP_REG(BPF_JEQ, R5, R6, 1),
4248 BPF_EXIT_INSN(),
4249 BPF_ALU64_REG(BPF_SUB, R6, R6),
4250 BPF_ALU64_REG(BPF_XOR, R7, R7),
4251 BPF_JMP_REG(BPF_JEQ, R7, R6, 1),
4252 BPF_EXIT_INSN(),
4253 BPF_ALU64_REG(BPF_SUB, R7, R7),
4254 BPF_ALU64_REG(BPF_XOR, R8, R8),
4255 BPF_JMP_REG(BPF_JEQ, R7, R8, 1),
4256 BPF_EXIT_INSN(),
4257 BPF_ALU64_REG(BPF_SUB, R8, R8),
4258 BPF_ALU64_REG(BPF_XOR, R9, R9),
4259 BPF_JMP_REG(BPF_JEQ, R9, R8, 1),
4260 BPF_EXIT_INSN(),
4261 BPF_ALU64_REG(BPF_SUB, R9, R9),
4262 BPF_ALU64_REG(BPF_XOR, R0, R0),
4263 BPF_JMP_REG(BPF_JEQ, R9, R0, 1),
4264 BPF_EXIT_INSN(),
4265 BPF_ALU64_REG(BPF_SUB, R1, R1),
4266 BPF_ALU64_REG(BPF_XOR, R0, R0),
4267 BPF_JMP_REG(BPF_JEQ, R9, R0, 2),
4268 BPF_ALU64_IMM(BPF_MOV, R0, 0),
4269 BPF_EXIT_INSN(),
4270 BPF_ALU64_IMM(BPF_MOV, R0, 1),
4271 BPF_EXIT_INSN(),
4272 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004273 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004274 { },
4275 { { 0, 1 } }
4276 },
4277 { /* Mainly checking JIT here. */
4278 "INT: MUL",
Andrew Mortonece80492014-05-22 10:16:46 -07004279 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004280 BPF_ALU64_IMM(BPF_MOV, R0, 11),
4281 BPF_ALU64_IMM(BPF_MOV, R1, 1),
4282 BPF_ALU64_IMM(BPF_MOV, R2, 2),
4283 BPF_ALU64_IMM(BPF_MOV, R3, 3),
4284 BPF_ALU64_IMM(BPF_MOV, R4, 4),
4285 BPF_ALU64_IMM(BPF_MOV, R5, 5),
4286 BPF_ALU64_IMM(BPF_MOV, R6, 6),
4287 BPF_ALU64_IMM(BPF_MOV, R7, 7),
4288 BPF_ALU64_IMM(BPF_MOV, R8, 8),
4289 BPF_ALU64_IMM(BPF_MOV, R9, 9),
4290 BPF_ALU64_REG(BPF_MUL, R0, R0),
4291 BPF_ALU64_REG(BPF_MUL, R0, R1),
4292 BPF_ALU64_REG(BPF_MUL, R0, R2),
4293 BPF_ALU64_REG(BPF_MUL, R0, R3),
4294 BPF_ALU64_REG(BPF_MUL, R0, R4),
4295 BPF_ALU64_REG(BPF_MUL, R0, R5),
4296 BPF_ALU64_REG(BPF_MUL, R0, R6),
4297 BPF_ALU64_REG(BPF_MUL, R0, R7),
4298 BPF_ALU64_REG(BPF_MUL, R0, R8),
4299 BPF_ALU64_REG(BPF_MUL, R0, R9),
4300 BPF_ALU64_IMM(BPF_MUL, R0, 10),
4301 BPF_JMP_IMM(BPF_JEQ, R0, 439084800, 1),
4302 BPF_EXIT_INSN(),
4303 BPF_ALU64_REG(BPF_MUL, R1, R0),
4304 BPF_ALU64_REG(BPF_MUL, R1, R2),
4305 BPF_ALU64_REG(BPF_MUL, R1, R3),
4306 BPF_ALU64_REG(BPF_MUL, R1, R4),
4307 BPF_ALU64_REG(BPF_MUL, R1, R5),
4308 BPF_ALU64_REG(BPF_MUL, R1, R6),
4309 BPF_ALU64_REG(BPF_MUL, R1, R7),
4310 BPF_ALU64_REG(BPF_MUL, R1, R8),
4311 BPF_ALU64_REG(BPF_MUL, R1, R9),
4312 BPF_ALU64_IMM(BPF_MUL, R1, 10),
4313 BPF_ALU64_REG(BPF_MOV, R2, R1),
4314 BPF_ALU64_IMM(BPF_RSH, R2, 32),
4315 BPF_JMP_IMM(BPF_JEQ, R2, 0x5a924, 1),
4316 BPF_EXIT_INSN(),
4317 BPF_ALU64_IMM(BPF_LSH, R1, 32),
4318 BPF_ALU64_IMM(BPF_ARSH, R1, 32),
4319 BPF_JMP_IMM(BPF_JEQ, R1, 0xebb90000, 1),
4320 BPF_EXIT_INSN(),
4321 BPF_ALU64_REG(BPF_MUL, R2, R0),
4322 BPF_ALU64_REG(BPF_MUL, R2, R1),
4323 BPF_ALU64_REG(BPF_MUL, R2, R3),
4324 BPF_ALU64_REG(BPF_MUL, R2, R4),
4325 BPF_ALU64_REG(BPF_MUL, R2, R5),
4326 BPF_ALU64_REG(BPF_MUL, R2, R6),
4327 BPF_ALU64_REG(BPF_MUL, R2, R7),
4328 BPF_ALU64_REG(BPF_MUL, R2, R8),
4329 BPF_ALU64_REG(BPF_MUL, R2, R9),
4330 BPF_ALU64_IMM(BPF_MUL, R2, 10),
4331 BPF_ALU64_IMM(BPF_RSH, R2, 32),
4332 BPF_ALU64_REG(BPF_MOV, R0, R2),
4333 BPF_EXIT_INSN(),
4334 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004335 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004336 { },
4337 { { 0, 0x35d97ef2 } }
4338 },
Daniel Borkmann9dd2af82015-12-17 23:51:57 +01004339 { /* Mainly checking JIT here. */
4340 "MOV REG64",
4341 .u.insns_int = {
4342 BPF_LD_IMM64(R0, 0xffffffffffffffffLL),
4343 BPF_MOV64_REG(R1, R0),
4344 BPF_MOV64_REG(R2, R1),
4345 BPF_MOV64_REG(R3, R2),
4346 BPF_MOV64_REG(R4, R3),
4347 BPF_MOV64_REG(R5, R4),
4348 BPF_MOV64_REG(R6, R5),
4349 BPF_MOV64_REG(R7, R6),
4350 BPF_MOV64_REG(R8, R7),
4351 BPF_MOV64_REG(R9, R8),
4352 BPF_ALU64_IMM(BPF_MOV, R0, 0),
4353 BPF_ALU64_IMM(BPF_MOV, R1, 0),
4354 BPF_ALU64_IMM(BPF_MOV, R2, 0),
4355 BPF_ALU64_IMM(BPF_MOV, R3, 0),
4356 BPF_ALU64_IMM(BPF_MOV, R4, 0),
4357 BPF_ALU64_IMM(BPF_MOV, R5, 0),
4358 BPF_ALU64_IMM(BPF_MOV, R6, 0),
4359 BPF_ALU64_IMM(BPF_MOV, R7, 0),
4360 BPF_ALU64_IMM(BPF_MOV, R8, 0),
4361 BPF_ALU64_IMM(BPF_MOV, R9, 0),
4362 BPF_ALU64_REG(BPF_ADD, R0, R0),
4363 BPF_ALU64_REG(BPF_ADD, R0, R1),
4364 BPF_ALU64_REG(BPF_ADD, R0, R2),
4365 BPF_ALU64_REG(BPF_ADD, R0, R3),
4366 BPF_ALU64_REG(BPF_ADD, R0, R4),
4367 BPF_ALU64_REG(BPF_ADD, R0, R5),
4368 BPF_ALU64_REG(BPF_ADD, R0, R6),
4369 BPF_ALU64_REG(BPF_ADD, R0, R7),
4370 BPF_ALU64_REG(BPF_ADD, R0, R8),
4371 BPF_ALU64_REG(BPF_ADD, R0, R9),
4372 BPF_ALU64_IMM(BPF_ADD, R0, 0xfefe),
4373 BPF_EXIT_INSN(),
4374 },
4375 INTERNAL,
4376 { },
4377 { { 0, 0xfefe } }
4378 },
4379 { /* Mainly checking JIT here. */
4380 "MOV REG32",
4381 .u.insns_int = {
4382 BPF_LD_IMM64(R0, 0xffffffffffffffffLL),
4383 BPF_MOV64_REG(R1, R0),
4384 BPF_MOV64_REG(R2, R1),
4385 BPF_MOV64_REG(R3, R2),
4386 BPF_MOV64_REG(R4, R3),
4387 BPF_MOV64_REG(R5, R4),
4388 BPF_MOV64_REG(R6, R5),
4389 BPF_MOV64_REG(R7, R6),
4390 BPF_MOV64_REG(R8, R7),
4391 BPF_MOV64_REG(R9, R8),
4392 BPF_ALU32_IMM(BPF_MOV, R0, 0),
4393 BPF_ALU32_IMM(BPF_MOV, R1, 0),
4394 BPF_ALU32_IMM(BPF_MOV, R2, 0),
4395 BPF_ALU32_IMM(BPF_MOV, R3, 0),
4396 BPF_ALU32_IMM(BPF_MOV, R4, 0),
4397 BPF_ALU32_IMM(BPF_MOV, R5, 0),
4398 BPF_ALU32_IMM(BPF_MOV, R6, 0),
4399 BPF_ALU32_IMM(BPF_MOV, R7, 0),
4400 BPF_ALU32_IMM(BPF_MOV, R8, 0),
4401 BPF_ALU32_IMM(BPF_MOV, R9, 0),
4402 BPF_ALU64_REG(BPF_ADD, R0, R0),
4403 BPF_ALU64_REG(BPF_ADD, R0, R1),
4404 BPF_ALU64_REG(BPF_ADD, R0, R2),
4405 BPF_ALU64_REG(BPF_ADD, R0, R3),
4406 BPF_ALU64_REG(BPF_ADD, R0, R4),
4407 BPF_ALU64_REG(BPF_ADD, R0, R5),
4408 BPF_ALU64_REG(BPF_ADD, R0, R6),
4409 BPF_ALU64_REG(BPF_ADD, R0, R7),
4410 BPF_ALU64_REG(BPF_ADD, R0, R8),
4411 BPF_ALU64_REG(BPF_ADD, R0, R9),
4412 BPF_ALU64_IMM(BPF_ADD, R0, 0xfefe),
4413 BPF_EXIT_INSN(),
4414 },
4415 INTERNAL,
4416 { },
4417 { { 0, 0xfefe } }
4418 },
4419 { /* Mainly checking JIT here. */
4420 "LD IMM64",
4421 .u.insns_int = {
4422 BPF_LD_IMM64(R0, 0xffffffffffffffffLL),
4423 BPF_MOV64_REG(R1, R0),
4424 BPF_MOV64_REG(R2, R1),
4425 BPF_MOV64_REG(R3, R2),
4426 BPF_MOV64_REG(R4, R3),
4427 BPF_MOV64_REG(R5, R4),
4428 BPF_MOV64_REG(R6, R5),
4429 BPF_MOV64_REG(R7, R6),
4430 BPF_MOV64_REG(R8, R7),
4431 BPF_MOV64_REG(R9, R8),
4432 BPF_LD_IMM64(R0, 0x0LL),
4433 BPF_LD_IMM64(R1, 0x0LL),
4434 BPF_LD_IMM64(R2, 0x0LL),
4435 BPF_LD_IMM64(R3, 0x0LL),
4436 BPF_LD_IMM64(R4, 0x0LL),
4437 BPF_LD_IMM64(R5, 0x0LL),
4438 BPF_LD_IMM64(R6, 0x0LL),
4439 BPF_LD_IMM64(R7, 0x0LL),
4440 BPF_LD_IMM64(R8, 0x0LL),
4441 BPF_LD_IMM64(R9, 0x0LL),
4442 BPF_ALU64_REG(BPF_ADD, R0, R0),
4443 BPF_ALU64_REG(BPF_ADD, R0, R1),
4444 BPF_ALU64_REG(BPF_ADD, R0, R2),
4445 BPF_ALU64_REG(BPF_ADD, R0, R3),
4446 BPF_ALU64_REG(BPF_ADD, R0, R4),
4447 BPF_ALU64_REG(BPF_ADD, R0, R5),
4448 BPF_ALU64_REG(BPF_ADD, R0, R6),
4449 BPF_ALU64_REG(BPF_ADD, R0, R7),
4450 BPF_ALU64_REG(BPF_ADD, R0, R8),
4451 BPF_ALU64_REG(BPF_ADD, R0, R9),
4452 BPF_ALU64_IMM(BPF_ADD, R0, 0xfefe),
4453 BPF_EXIT_INSN(),
4454 },
4455 INTERNAL,
4456 { },
4457 { { 0, 0xfefe } }
4458 },
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004459 {
4460 "INT: ALU MIX",
Andrew Mortonece80492014-05-22 10:16:46 -07004461 .u.insns_int = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004462 BPF_ALU64_IMM(BPF_MOV, R0, 11),
4463 BPF_ALU64_IMM(BPF_ADD, R0, -1),
4464 BPF_ALU64_IMM(BPF_MOV, R2, 2),
4465 BPF_ALU64_IMM(BPF_XOR, R2, 3),
4466 BPF_ALU64_REG(BPF_DIV, R0, R2),
4467 BPF_JMP_IMM(BPF_JEQ, R0, 10, 1),
4468 BPF_EXIT_INSN(),
4469 BPF_ALU64_IMM(BPF_MOD, R0, 3),
4470 BPF_JMP_IMM(BPF_JEQ, R0, 1, 1),
4471 BPF_EXIT_INSN(),
4472 BPF_ALU64_IMM(BPF_MOV, R0, -1),
4473 BPF_EXIT_INSN(),
4474 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004475 INTERNAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004476 { },
4477 { { 0, -1 } }
4478 },
4479 {
Alexei Starovoitov72b603e2014-08-25 12:27:02 -07004480 "INT: shifts by register",
4481 .u.insns_int = {
4482 BPF_MOV64_IMM(R0, -1234),
4483 BPF_MOV64_IMM(R1, 1),
4484 BPF_ALU32_REG(BPF_RSH, R0, R1),
4485 BPF_JMP_IMM(BPF_JEQ, R0, 0x7ffffd97, 1),
4486 BPF_EXIT_INSN(),
4487 BPF_MOV64_IMM(R2, 1),
4488 BPF_ALU64_REG(BPF_LSH, R0, R2),
4489 BPF_MOV32_IMM(R4, -1234),
4490 BPF_JMP_REG(BPF_JEQ, R0, R4, 1),
4491 BPF_EXIT_INSN(),
4492 BPF_ALU64_IMM(BPF_AND, R4, 63),
4493 BPF_ALU64_REG(BPF_LSH, R0, R4), /* R0 <= 46 */
4494 BPF_MOV64_IMM(R3, 47),
4495 BPF_ALU64_REG(BPF_ARSH, R0, R3),
4496 BPF_JMP_IMM(BPF_JEQ, R0, -617, 1),
4497 BPF_EXIT_INSN(),
4498 BPF_MOV64_IMM(R2, 1),
4499 BPF_ALU64_REG(BPF_LSH, R4, R2), /* R4 = 46 << 1 */
4500 BPF_JMP_IMM(BPF_JEQ, R4, 92, 1),
4501 BPF_EXIT_INSN(),
4502 BPF_MOV64_IMM(R4, 4),
4503 BPF_ALU64_REG(BPF_LSH, R4, R4), /* R4 = 4 << 4 */
4504 BPF_JMP_IMM(BPF_JEQ, R4, 64, 1),
4505 BPF_EXIT_INSN(),
4506 BPF_MOV64_IMM(R4, 5),
4507 BPF_ALU32_REG(BPF_LSH, R4, R4), /* R4 = 5 << 5 */
4508 BPF_JMP_IMM(BPF_JEQ, R4, 160, 1),
4509 BPF_EXIT_INSN(),
4510 BPF_MOV64_IMM(R0, -1),
4511 BPF_EXIT_INSN(),
4512 },
4513 INTERNAL,
4514 { },
4515 { { 0, -1 } }
4516 },
Johan Almbladh53e33f92021-08-09 11:18:26 +02004517#ifdef CONFIG_32BIT
4518 {
4519 "INT: 32-bit context pointer word order and zero-extension",
4520 .u.insns_int = {
4521 BPF_ALU32_IMM(BPF_MOV, R0, 0),
4522 BPF_JMP32_IMM(BPF_JEQ, R1, 0, 3),
4523 BPF_ALU64_IMM(BPF_RSH, R1, 32),
4524 BPF_JMP32_IMM(BPF_JNE, R1, 0, 1),
4525 BPF_ALU32_IMM(BPF_MOV, R0, 1),
4526 BPF_EXIT_INSN(),
4527 },
4528 INTERNAL,
4529 { },
4530 { { 0, 1 } }
4531 },
4532#endif
Johan Almbladh84024a42021-08-09 11:18:23 +02004533 {
Alexei Starovoitov64a89462014-05-08 14:10:52 -07004534 "check: missing ret",
Andrew Mortonece80492014-05-22 10:16:46 -07004535 .u.insns = {
Alexei Starovoitov64a89462014-05-08 14:10:52 -07004536 BPF_STMT(BPF_LD | BPF_IMM, 1),
4537 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004538 CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL,
Alexei Starovoitov64a89462014-05-08 14:10:52 -07004539 { },
Yonghong Song09584b42018-02-02 22:37:15 -08004540 { },
4541 .fill_helper = NULL,
4542 .expected_errcode = -EINVAL,
Alexei Starovoitov64a89462014-05-08 14:10:52 -07004543 },
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004544 {
4545 "check: div_k_0",
Andrew Mortonece80492014-05-22 10:16:46 -07004546 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004547 BPF_STMT(BPF_ALU | BPF_DIV | BPF_K, 0),
4548 BPF_STMT(BPF_RET | BPF_K, 0)
4549 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004550 CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004551 { },
Yonghong Song09584b42018-02-02 22:37:15 -08004552 { },
4553 .fill_helper = NULL,
4554 .expected_errcode = -EINVAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004555 },
4556 {
4557 "check: unknown insn",
Andrew Mortonece80492014-05-22 10:16:46 -07004558 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004559 /* seccomp insn, rejected in socket filter */
4560 BPF_STMT(BPF_LDX | BPF_W | BPF_ABS, 0),
4561 BPF_STMT(BPF_RET | BPF_K, 0)
4562 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004563 CLASSIC | FLAG_EXPECTED_FAIL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004564 { },
Yonghong Song09584b42018-02-02 22:37:15 -08004565 { },
4566 .fill_helper = NULL,
4567 .expected_errcode = -EINVAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004568 },
4569 {
4570 "check: out of range spill/fill",
Andrew Mortonece80492014-05-22 10:16:46 -07004571 .u.insns = {
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004572 BPF_STMT(BPF_STX, 16),
4573 BPF_STMT(BPF_RET | BPF_K, 0)
4574 },
Daniel Borkmann10f18e02014-05-23 18:44:00 +02004575 CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004576 { },
Yonghong Song09584b42018-02-02 22:37:15 -08004577 { },
4578 .fill_helper = NULL,
4579 .expected_errcode = -EINVAL,
Alexei Starovoitov9def6242014-05-08 14:10:53 -07004580 },
Daniel Borkmann2e8a83c2014-05-23 18:44:01 +02004581 {
4582 "JUMPS + HOLES",
4583 .u.insns = {
4584 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4585 BPF_JUMP(BPF_JMP | BPF_JGE, 0, 13, 15),
4586 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4587 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4588 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4589 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4590 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4591 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4592 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4593 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4594 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4595 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4596 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4597 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4598 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4599 BPF_JUMP(BPF_JMP | BPF_JEQ, 0x90c2894d, 3, 4),
4600 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4601 BPF_JUMP(BPF_JMP | BPF_JEQ, 0x90c2894d, 1, 2),
4602 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4603 BPF_JUMP(BPF_JMP | BPF_JGE, 0, 14, 15),
4604 BPF_JUMP(BPF_JMP | BPF_JGE, 0, 13, 14),
4605 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4606 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4607 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4608 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4609 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4610 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4611 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4612 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4613 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4614 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4615 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4616 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4617 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4618 BPF_JUMP(BPF_JMP | BPF_JEQ, 0x2ac28349, 2, 3),
4619 BPF_JUMP(BPF_JMP | BPF_JEQ, 0x2ac28349, 1, 2),
4620 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4621 BPF_JUMP(BPF_JMP | BPF_JGE, 0, 14, 15),
4622 BPF_JUMP(BPF_JMP | BPF_JGE, 0, 13, 14),
4623 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4624 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4625 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4626 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4627 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4628 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4629 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4630 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4631 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4632 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4633 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4634 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4635 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4636 BPF_JUMP(BPF_JMP | BPF_JEQ, 0x90d2ff41, 2, 3),
4637 BPF_JUMP(BPF_JMP | BPF_JEQ, 0x90d2ff41, 1, 2),
4638 BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0),
4639 BPF_STMT(BPF_RET | BPF_A, 0),
4640 BPF_STMT(BPF_RET | BPF_A, 0),
4641 },
4642 CLASSIC,
Daniel Borkmannce25b682014-05-26 20:17:35 +02004643 { 0x00, 0x1b, 0x21, 0x3c, 0x9d, 0xf8,
4644 0x90, 0xe2, 0xba, 0x0a, 0x56, 0xb4,
4645 0x08, 0x00,
4646 0x45, 0x00, 0x00, 0x28, 0x00, 0x00,
4647 0x20, 0x00, 0x40, 0x11, 0x00, 0x00, /* IP header */
4648 0xc0, 0xa8, 0x33, 0x01,
4649 0xc0, 0xa8, 0x33, 0x02,
4650 0xbb, 0xb6,
4651 0xa9, 0xfa,
4652 0x00, 0x14, 0x00, 0x00,
4653 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc,
4654 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc,
4655 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc,
4656 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc,
4657 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc,
4658 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc,
4659 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc,
4660 0xcc, 0xcc, 0xcc, 0xcc },
Daniel Borkmann2e8a83c2014-05-23 18:44:01 +02004661 { { 88, 0x001b } }
4662 },
4663 {
4664 "check: RET X",
4665 .u.insns = {
4666 BPF_STMT(BPF_RET | BPF_X, 0),
4667 },
4668 CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL,
4669 { },
4670 { },
Yonghong Song09584b42018-02-02 22:37:15 -08004671 .fill_helper = NULL,
4672 .expected_errcode = -EINVAL,
Daniel Borkmann2e8a83c2014-05-23 18:44:01 +02004673 },
4674 {
4675 "check: LDX + RET X",
4676 .u.insns = {
4677 BPF_STMT(BPF_LDX | BPF_IMM, 42),
4678 BPF_STMT(BPF_RET | BPF_X, 0),
4679 },
4680 CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL,
4681 { },
4682 { },
Yonghong Song09584b42018-02-02 22:37:15 -08004683 .fill_helper = NULL,
4684 .expected_errcode = -EINVAL,
Daniel Borkmann2e8a83c2014-05-23 18:44:01 +02004685 },
Daniel Borkmann108cc222014-05-26 20:17:34 +02004686 { /* Mainly checking JIT here. */
Daniel Borkmann9fe13ba2014-05-29 10:22:48 +02004687 "M[]: alt STX + LDX",
Daniel Borkmann108cc222014-05-26 20:17:34 +02004688 .u.insns = {
4689 BPF_STMT(BPF_LDX | BPF_IMM, 100),
4690 BPF_STMT(BPF_STX, 0),
4691 BPF_STMT(BPF_LDX | BPF_MEM, 0),
4692 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4693 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4694 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4695 BPF_STMT(BPF_STX, 1),
4696 BPF_STMT(BPF_LDX | BPF_MEM, 1),
4697 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4698 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4699 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4700 BPF_STMT(BPF_STX, 2),
4701 BPF_STMT(BPF_LDX | BPF_MEM, 2),
4702 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4703 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4704 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4705 BPF_STMT(BPF_STX, 3),
4706 BPF_STMT(BPF_LDX | BPF_MEM, 3),
4707 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4708 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4709 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4710 BPF_STMT(BPF_STX, 4),
4711 BPF_STMT(BPF_LDX | BPF_MEM, 4),
4712 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4713 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4714 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4715 BPF_STMT(BPF_STX, 5),
4716 BPF_STMT(BPF_LDX | BPF_MEM, 5),
4717 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4718 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4719 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4720 BPF_STMT(BPF_STX, 6),
4721 BPF_STMT(BPF_LDX | BPF_MEM, 6),
4722 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4723 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4724 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4725 BPF_STMT(BPF_STX, 7),
4726 BPF_STMT(BPF_LDX | BPF_MEM, 7),
4727 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4728 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4729 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4730 BPF_STMT(BPF_STX, 8),
4731 BPF_STMT(BPF_LDX | BPF_MEM, 8),
4732 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4733 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4734 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4735 BPF_STMT(BPF_STX, 9),
4736 BPF_STMT(BPF_LDX | BPF_MEM, 9),
4737 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4738 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4739 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4740 BPF_STMT(BPF_STX, 10),
4741 BPF_STMT(BPF_LDX | BPF_MEM, 10),
4742 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4743 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4744 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4745 BPF_STMT(BPF_STX, 11),
4746 BPF_STMT(BPF_LDX | BPF_MEM, 11),
4747 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4748 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4749 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4750 BPF_STMT(BPF_STX, 12),
4751 BPF_STMT(BPF_LDX | BPF_MEM, 12),
4752 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4753 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4754 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4755 BPF_STMT(BPF_STX, 13),
4756 BPF_STMT(BPF_LDX | BPF_MEM, 13),
4757 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4758 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4759 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4760 BPF_STMT(BPF_STX, 14),
4761 BPF_STMT(BPF_LDX | BPF_MEM, 14),
4762 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4763 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4764 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4765 BPF_STMT(BPF_STX, 15),
4766 BPF_STMT(BPF_LDX | BPF_MEM, 15),
4767 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4768 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1),
4769 BPF_STMT(BPF_MISC | BPF_TAX, 0),
4770 BPF_STMT(BPF_RET | BPF_A, 0),
4771 },
4772 CLASSIC | FLAG_NO_DATA,
4773 { },
4774 { { 0, 116 } },
4775 },
Daniel Borkmann9fe13ba2014-05-29 10:22:48 +02004776 { /* Mainly checking JIT here. */
4777 "M[]: full STX + full LDX",
4778 .u.insns = {
4779 BPF_STMT(BPF_LDX | BPF_IMM, 0xbadfeedb),
4780 BPF_STMT(BPF_STX, 0),
4781 BPF_STMT(BPF_LDX | BPF_IMM, 0xecabedae),
4782 BPF_STMT(BPF_STX, 1),
4783 BPF_STMT(BPF_LDX | BPF_IMM, 0xafccfeaf),
4784 BPF_STMT(BPF_STX, 2),
4785 BPF_STMT(BPF_LDX | BPF_IMM, 0xbffdcedc),
4786 BPF_STMT(BPF_STX, 3),
4787 BPF_STMT(BPF_LDX | BPF_IMM, 0xfbbbdccb),
4788 BPF_STMT(BPF_STX, 4),
4789 BPF_STMT(BPF_LDX | BPF_IMM, 0xfbabcbda),
4790 BPF_STMT(BPF_STX, 5),
4791 BPF_STMT(BPF_LDX | BPF_IMM, 0xaedecbdb),
4792 BPF_STMT(BPF_STX, 6),
4793 BPF_STMT(BPF_LDX | BPF_IMM, 0xadebbade),
4794 BPF_STMT(BPF_STX, 7),
4795 BPF_STMT(BPF_LDX | BPF_IMM, 0xfcfcfaec),
4796 BPF_STMT(BPF_STX, 8),
4797 BPF_STMT(BPF_LDX | BPF_IMM, 0xbcdddbdc),
4798 BPF_STMT(BPF_STX, 9),
4799 BPF_STMT(BPF_LDX | BPF_IMM, 0xfeefdfac),
4800 BPF_STMT(BPF_STX, 10),
4801 BPF_STMT(BPF_LDX | BPF_IMM, 0xcddcdeea),
4802 BPF_STMT(BPF_STX, 11),
4803 BPF_STMT(BPF_LDX | BPF_IMM, 0xaccfaebb),
4804 BPF_STMT(BPF_STX, 12),
4805 BPF_STMT(BPF_LDX | BPF_IMM, 0xbdcccdcf),
4806 BPF_STMT(BPF_STX, 13),
4807 BPF_STMT(BPF_LDX | BPF_IMM, 0xaaedecde),
4808 BPF_STMT(BPF_STX, 14),
4809 BPF_STMT(BPF_LDX | BPF_IMM, 0xfaeacdad),
4810 BPF_STMT(BPF_STX, 15),
4811 BPF_STMT(BPF_LDX | BPF_MEM, 0),
4812 BPF_STMT(BPF_MISC | BPF_TXA, 0),
4813 BPF_STMT(BPF_LDX | BPF_MEM, 1),
4814 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4815 BPF_STMT(BPF_LDX | BPF_MEM, 2),
4816 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4817 BPF_STMT(BPF_LDX | BPF_MEM, 3),
4818 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4819 BPF_STMT(BPF_LDX | BPF_MEM, 4),
4820 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4821 BPF_STMT(BPF_LDX | BPF_MEM, 5),
4822 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4823 BPF_STMT(BPF_LDX | BPF_MEM, 6),
4824 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4825 BPF_STMT(BPF_LDX | BPF_MEM, 7),
4826 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4827 BPF_STMT(BPF_LDX | BPF_MEM, 8),
4828 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4829 BPF_STMT(BPF_LDX | BPF_MEM, 9),
4830 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4831 BPF_STMT(BPF_LDX | BPF_MEM, 10),
4832 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4833 BPF_STMT(BPF_LDX | BPF_MEM, 11),
4834 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4835 BPF_STMT(BPF_LDX | BPF_MEM, 12),
4836 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4837 BPF_STMT(BPF_LDX | BPF_MEM, 13),
4838 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4839 BPF_STMT(BPF_LDX | BPF_MEM, 14),
4840 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4841 BPF_STMT(BPF_LDX | BPF_MEM, 15),
4842 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
4843 BPF_STMT(BPF_RET | BPF_A, 0),
4844 },
4845 CLASSIC | FLAG_NO_DATA,
4846 { },
4847 { { 0, 0x2a5a5e5 } },
4848 },
Daniel Borkmannd50bc152014-05-29 10:22:49 +02004849 {
4850 "check: SKF_AD_MAX",
4851 .u.insns = {
4852 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
4853 SKF_AD_OFF + SKF_AD_MAX),
4854 BPF_STMT(BPF_RET | BPF_A, 0),
4855 },
4856 CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL,
4857 { },
4858 { },
Yonghong Song09584b42018-02-02 22:37:15 -08004859 .fill_helper = NULL,
4860 .expected_errcode = -EINVAL,
Daniel Borkmannd50bc152014-05-29 10:22:49 +02004861 },
4862 { /* Passes checker but fails during runtime. */
4863 "LD [SKF_AD_OFF-1]",
4864 .u.insns = {
4865 BPF_STMT(BPF_LD | BPF_W | BPF_ABS,
4866 SKF_AD_OFF - 1),
4867 BPF_STMT(BPF_RET | BPF_K, 1),
4868 },
4869 CLASSIC,
4870 { },
4871 { { 1, 0 } },
4872 },
Alexei Starovoitov02ab6952014-09-04 22:17:17 -07004873 {
4874 "load 64-bit immediate",
4875 .u.insns_int = {
Alexei Starovoitov25ee7322014-09-19 13:53:51 -07004876 BPF_LD_IMM64(R1, 0x567800001234LL),
Alexei Starovoitov02ab6952014-09-04 22:17:17 -07004877 BPF_MOV64_REG(R2, R1),
4878 BPF_MOV64_REG(R3, R2),
4879 BPF_ALU64_IMM(BPF_RSH, R2, 32),
4880 BPF_ALU64_IMM(BPF_LSH, R3, 32),
4881 BPF_ALU64_IMM(BPF_RSH, R3, 32),
4882 BPF_ALU64_IMM(BPF_MOV, R0, 0),
4883 BPF_JMP_IMM(BPF_JEQ, R2, 0x5678, 1),
4884 BPF_EXIT_INSN(),
4885 BPF_JMP_IMM(BPF_JEQ, R3, 0x1234, 1),
4886 BPF_EXIT_INSN(),
Xi Wang986ccfd2015-05-09 04:14:30 -04004887 BPF_LD_IMM64(R0, 0x1ffffffffLL),
4888 BPF_ALU64_IMM(BPF_RSH, R0, 32), /* R0 = 1 */
Alexei Starovoitov02ab6952014-09-04 22:17:17 -07004889 BPF_EXIT_INSN(),
4890 },
4891 INTERNAL,
4892 { },
4893 { { 0, 1 } }
4894 },
Michael Holzheucffc6422015-05-11 22:22:44 -07004895 /* BPF_ALU | BPF_MOV | BPF_X */
4896 {
4897 "ALU_MOV_X: dst = 2",
4898 .u.insns_int = {
4899 BPF_ALU32_IMM(BPF_MOV, R1, 2),
4900 BPF_ALU32_REG(BPF_MOV, R0, R1),
4901 BPF_EXIT_INSN(),
4902 },
4903 INTERNAL,
4904 { },
4905 { { 0, 2 } },
4906 },
4907 {
4908 "ALU_MOV_X: dst = 4294967295",
4909 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07004910 BPF_ALU32_IMM(BPF_MOV, R1, 4294967295U),
Michael Holzheucffc6422015-05-11 22:22:44 -07004911 BPF_ALU32_REG(BPF_MOV, R0, R1),
4912 BPF_EXIT_INSN(),
4913 },
4914 INTERNAL,
4915 { },
Michael Holzheu56cbaa42015-05-13 20:40:39 -07004916 { { 0, 4294967295U } },
Michael Holzheucffc6422015-05-11 22:22:44 -07004917 },
4918 {
4919 "ALU64_MOV_X: dst = 2",
4920 .u.insns_int = {
4921 BPF_ALU32_IMM(BPF_MOV, R1, 2),
4922 BPF_ALU64_REG(BPF_MOV, R0, R1),
4923 BPF_EXIT_INSN(),
4924 },
4925 INTERNAL,
4926 { },
4927 { { 0, 2 } },
4928 },
4929 {
4930 "ALU64_MOV_X: dst = 4294967295",
4931 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07004932 BPF_ALU32_IMM(BPF_MOV, R1, 4294967295U),
Michael Holzheucffc6422015-05-11 22:22:44 -07004933 BPF_ALU64_REG(BPF_MOV, R0, R1),
4934 BPF_EXIT_INSN(),
4935 },
4936 INTERNAL,
4937 { },
Michael Holzheu56cbaa42015-05-13 20:40:39 -07004938 { { 0, 4294967295U } },
Michael Holzheucffc6422015-05-11 22:22:44 -07004939 },
4940 /* BPF_ALU | BPF_MOV | BPF_K */
4941 {
4942 "ALU_MOV_K: dst = 2",
4943 .u.insns_int = {
4944 BPF_ALU32_IMM(BPF_MOV, R0, 2),
4945 BPF_EXIT_INSN(),
4946 },
4947 INTERNAL,
4948 { },
4949 { { 0, 2 } },
4950 },
4951 {
4952 "ALU_MOV_K: dst = 4294967295",
4953 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07004954 BPF_ALU32_IMM(BPF_MOV, R0, 4294967295U),
Michael Holzheucffc6422015-05-11 22:22:44 -07004955 BPF_EXIT_INSN(),
4956 },
4957 INTERNAL,
4958 { },
Michael Holzheu56cbaa42015-05-13 20:40:39 -07004959 { { 0, 4294967295U } },
Michael Holzheucffc6422015-05-11 22:22:44 -07004960 },
4961 {
4962 "ALU_MOV_K: 0x0000ffffffff0000 = 0x00000000ffffffff",
4963 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07004964 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
4965 BPF_LD_IMM64(R3, 0x00000000ffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07004966 BPF_ALU32_IMM(BPF_MOV, R2, 0xffffffff),
4967 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
4968 BPF_MOV32_IMM(R0, 2),
4969 BPF_EXIT_INSN(),
4970 BPF_MOV32_IMM(R0, 1),
4971 BPF_EXIT_INSN(),
4972 },
4973 INTERNAL,
4974 { },
4975 { { 0, 0x1 } },
4976 },
4977 {
Johan Almbladh565731a2021-08-09 11:18:17 +02004978 "ALU_MOV_K: small negative",
4979 .u.insns_int = {
4980 BPF_ALU32_IMM(BPF_MOV, R0, -123),
4981 BPF_EXIT_INSN(),
4982 },
4983 INTERNAL,
4984 { },
4985 { { 0, -123 } }
4986 },
4987 {
4988 "ALU_MOV_K: small negative zero extension",
4989 .u.insns_int = {
4990 BPF_ALU32_IMM(BPF_MOV, R0, -123),
4991 BPF_ALU64_IMM(BPF_RSH, R0, 32),
4992 BPF_EXIT_INSN(),
4993 },
4994 INTERNAL,
4995 { },
4996 { { 0, 0 } }
4997 },
4998 {
4999 "ALU_MOV_K: large negative",
5000 .u.insns_int = {
5001 BPF_ALU32_IMM(BPF_MOV, R0, -123456789),
5002 BPF_EXIT_INSN(),
5003 },
5004 INTERNAL,
5005 { },
5006 { { 0, -123456789 } }
5007 },
5008 {
5009 "ALU_MOV_K: large negative zero extension",
5010 .u.insns_int = {
5011 BPF_ALU32_IMM(BPF_MOV, R0, -123456789),
5012 BPF_ALU64_IMM(BPF_RSH, R0, 32),
5013 BPF_EXIT_INSN(),
5014 },
5015 INTERNAL,
5016 { },
5017 { { 0, 0 } }
5018 },
5019 {
Michael Holzheucffc6422015-05-11 22:22:44 -07005020 "ALU64_MOV_K: dst = 2",
5021 .u.insns_int = {
5022 BPF_ALU64_IMM(BPF_MOV, R0, 2),
5023 BPF_EXIT_INSN(),
5024 },
5025 INTERNAL,
5026 { },
5027 { { 0, 2 } },
5028 },
5029 {
5030 "ALU64_MOV_K: dst = 2147483647",
5031 .u.insns_int = {
5032 BPF_ALU64_IMM(BPF_MOV, R0, 2147483647),
5033 BPF_EXIT_INSN(),
5034 },
5035 INTERNAL,
5036 { },
5037 { { 0, 2147483647 } },
5038 },
5039 {
5040 "ALU64_OR_K: dst = 0x0",
5041 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005042 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
Michael Holzheucffc6422015-05-11 22:22:44 -07005043 BPF_LD_IMM64(R3, 0x0),
5044 BPF_ALU64_IMM(BPF_MOV, R2, 0x0),
5045 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5046 BPF_MOV32_IMM(R0, 2),
5047 BPF_EXIT_INSN(),
5048 BPF_MOV32_IMM(R0, 1),
5049 BPF_EXIT_INSN(),
5050 },
5051 INTERNAL,
5052 { },
5053 { { 0, 0x1 } },
5054 },
5055 {
5056 "ALU64_MOV_K: dst = -1",
5057 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005058 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
5059 BPF_LD_IMM64(R3, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07005060 BPF_ALU64_IMM(BPF_MOV, R2, 0xffffffff),
5061 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5062 BPF_MOV32_IMM(R0, 2),
5063 BPF_EXIT_INSN(),
5064 BPF_MOV32_IMM(R0, 1),
5065 BPF_EXIT_INSN(),
5066 },
5067 INTERNAL,
5068 { },
5069 { { 0, 0x1 } },
5070 },
Johan Almbladh565731a2021-08-09 11:18:17 +02005071 {
5072 "ALU64_MOV_K: small negative",
5073 .u.insns_int = {
5074 BPF_ALU64_IMM(BPF_MOV, R0, -123),
5075 BPF_EXIT_INSN(),
5076 },
5077 INTERNAL,
5078 { },
5079 { { 0, -123 } }
5080 },
5081 {
5082 "ALU64_MOV_K: small negative sign extension",
5083 .u.insns_int = {
5084 BPF_ALU64_IMM(BPF_MOV, R0, -123),
5085 BPF_ALU64_IMM(BPF_RSH, R0, 32),
5086 BPF_EXIT_INSN(),
5087 },
5088 INTERNAL,
5089 { },
5090 { { 0, 0xffffffff } }
5091 },
5092 {
5093 "ALU64_MOV_K: large negative",
5094 .u.insns_int = {
5095 BPF_ALU64_IMM(BPF_MOV, R0, -123456789),
5096 BPF_EXIT_INSN(),
5097 },
5098 INTERNAL,
5099 { },
5100 { { 0, -123456789 } }
5101 },
5102 {
5103 "ALU64_MOV_K: large negative sign extension",
5104 .u.insns_int = {
5105 BPF_ALU64_IMM(BPF_MOV, R0, -123456789),
5106 BPF_ALU64_IMM(BPF_RSH, R0, 32),
5107 BPF_EXIT_INSN(),
5108 },
5109 INTERNAL,
5110 { },
5111 { { 0, 0xffffffff } }
5112 },
Michael Holzheucffc6422015-05-11 22:22:44 -07005113 /* BPF_ALU | BPF_ADD | BPF_X */
5114 {
5115 "ALU_ADD_X: 1 + 2 = 3",
5116 .u.insns_int = {
5117 BPF_LD_IMM64(R0, 1),
5118 BPF_ALU32_IMM(BPF_MOV, R1, 2),
5119 BPF_ALU32_REG(BPF_ADD, R0, R1),
5120 BPF_EXIT_INSN(),
5121 },
5122 INTERNAL,
5123 { },
5124 { { 0, 3 } },
5125 },
5126 {
5127 "ALU_ADD_X: 1 + 4294967294 = 4294967295",
5128 .u.insns_int = {
5129 BPF_LD_IMM64(R0, 1),
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005130 BPF_ALU32_IMM(BPF_MOV, R1, 4294967294U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005131 BPF_ALU32_REG(BPF_ADD, R0, R1),
5132 BPF_EXIT_INSN(),
5133 },
5134 INTERNAL,
5135 { },
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005136 { { 0, 4294967295U } },
Michael Holzheucffc6422015-05-11 22:22:44 -07005137 },
5138 {
Naveen N. Raob64b50e2016-04-05 15:32:55 +05305139 "ALU_ADD_X: 2 + 4294967294 = 0",
5140 .u.insns_int = {
5141 BPF_LD_IMM64(R0, 2),
5142 BPF_LD_IMM64(R1, 4294967294U),
5143 BPF_ALU32_REG(BPF_ADD, R0, R1),
5144 BPF_JMP_IMM(BPF_JEQ, R0, 0, 2),
5145 BPF_ALU32_IMM(BPF_MOV, R0, 0),
5146 BPF_EXIT_INSN(),
5147 BPF_ALU32_IMM(BPF_MOV, R0, 1),
5148 BPF_EXIT_INSN(),
5149 },
5150 INTERNAL,
5151 { },
5152 { { 0, 1 } },
5153 },
5154 {
Michael Holzheucffc6422015-05-11 22:22:44 -07005155 "ALU64_ADD_X: 1 + 2 = 3",
5156 .u.insns_int = {
5157 BPF_LD_IMM64(R0, 1),
5158 BPF_ALU32_IMM(BPF_MOV, R1, 2),
5159 BPF_ALU64_REG(BPF_ADD, R0, R1),
5160 BPF_EXIT_INSN(),
5161 },
5162 INTERNAL,
5163 { },
5164 { { 0, 3 } },
5165 },
5166 {
5167 "ALU64_ADD_X: 1 + 4294967294 = 4294967295",
5168 .u.insns_int = {
5169 BPF_LD_IMM64(R0, 1),
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005170 BPF_ALU32_IMM(BPF_MOV, R1, 4294967294U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005171 BPF_ALU64_REG(BPF_ADD, R0, R1),
5172 BPF_EXIT_INSN(),
5173 },
5174 INTERNAL,
5175 { },
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005176 { { 0, 4294967295U } },
Michael Holzheucffc6422015-05-11 22:22:44 -07005177 },
Naveen N. Raob64b50e2016-04-05 15:32:55 +05305178 {
5179 "ALU64_ADD_X: 2 + 4294967294 = 4294967296",
5180 .u.insns_int = {
5181 BPF_LD_IMM64(R0, 2),
5182 BPF_LD_IMM64(R1, 4294967294U),
5183 BPF_LD_IMM64(R2, 4294967296ULL),
5184 BPF_ALU64_REG(BPF_ADD, R0, R1),
5185 BPF_JMP_REG(BPF_JEQ, R0, R2, 2),
5186 BPF_MOV32_IMM(R0, 0),
5187 BPF_EXIT_INSN(),
5188 BPF_MOV32_IMM(R0, 1),
5189 BPF_EXIT_INSN(),
5190 },
5191 INTERNAL,
5192 { },
5193 { { 0, 1 } },
5194 },
Michael Holzheucffc6422015-05-11 22:22:44 -07005195 /* BPF_ALU | BPF_ADD | BPF_K */
5196 {
5197 "ALU_ADD_K: 1 + 2 = 3",
5198 .u.insns_int = {
5199 BPF_LD_IMM64(R0, 1),
5200 BPF_ALU32_IMM(BPF_ADD, R0, 2),
5201 BPF_EXIT_INSN(),
5202 },
5203 INTERNAL,
5204 { },
5205 { { 0, 3 } },
5206 },
5207 {
5208 "ALU_ADD_K: 3 + 0 = 3",
5209 .u.insns_int = {
5210 BPF_LD_IMM64(R0, 3),
5211 BPF_ALU32_IMM(BPF_ADD, R0, 0),
5212 BPF_EXIT_INSN(),
5213 },
5214 INTERNAL,
5215 { },
5216 { { 0, 3 } },
5217 },
5218 {
5219 "ALU_ADD_K: 1 + 4294967294 = 4294967295",
5220 .u.insns_int = {
5221 BPF_LD_IMM64(R0, 1),
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005222 BPF_ALU32_IMM(BPF_ADD, R0, 4294967294U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005223 BPF_EXIT_INSN(),
5224 },
5225 INTERNAL,
5226 { },
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005227 { { 0, 4294967295U } },
Michael Holzheucffc6422015-05-11 22:22:44 -07005228 },
5229 {
Naveen N. Raob64b50e2016-04-05 15:32:55 +05305230 "ALU_ADD_K: 4294967294 + 2 = 0",
5231 .u.insns_int = {
5232 BPF_LD_IMM64(R0, 4294967294U),
5233 BPF_ALU32_IMM(BPF_ADD, R0, 2),
5234 BPF_JMP_IMM(BPF_JEQ, R0, 0, 2),
5235 BPF_ALU32_IMM(BPF_MOV, R0, 0),
5236 BPF_EXIT_INSN(),
5237 BPF_ALU32_IMM(BPF_MOV, R0, 1),
5238 BPF_EXIT_INSN(),
5239 },
5240 INTERNAL,
5241 { },
5242 { { 0, 1 } },
5243 },
5244 {
Michael Holzheucffc6422015-05-11 22:22:44 -07005245 "ALU_ADD_K: 0 + (-1) = 0x00000000ffffffff",
5246 .u.insns_int = {
5247 BPF_LD_IMM64(R2, 0x0),
5248 BPF_LD_IMM64(R3, 0x00000000ffffffff),
5249 BPF_ALU32_IMM(BPF_ADD, R2, 0xffffffff),
5250 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5251 BPF_MOV32_IMM(R0, 2),
5252 BPF_EXIT_INSN(),
5253 BPF_MOV32_IMM(R0, 1),
5254 BPF_EXIT_INSN(),
5255 },
5256 INTERNAL,
5257 { },
5258 { { 0, 0x1 } },
5259 },
5260 {
Naveen N. Rao9c94f6c2016-04-05 15:32:56 +05305261 "ALU_ADD_K: 0 + 0xffff = 0xffff",
5262 .u.insns_int = {
5263 BPF_LD_IMM64(R2, 0x0),
5264 BPF_LD_IMM64(R3, 0xffff),
5265 BPF_ALU32_IMM(BPF_ADD, R2, 0xffff),
5266 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5267 BPF_MOV32_IMM(R0, 2),
5268 BPF_EXIT_INSN(),
5269 BPF_MOV32_IMM(R0, 1),
5270 BPF_EXIT_INSN(),
5271 },
5272 INTERNAL,
5273 { },
5274 { { 0, 0x1 } },
5275 },
5276 {
5277 "ALU_ADD_K: 0 + 0x7fffffff = 0x7fffffff",
5278 .u.insns_int = {
5279 BPF_LD_IMM64(R2, 0x0),
5280 BPF_LD_IMM64(R3, 0x7fffffff),
5281 BPF_ALU32_IMM(BPF_ADD, R2, 0x7fffffff),
5282 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5283 BPF_MOV32_IMM(R0, 2),
5284 BPF_EXIT_INSN(),
5285 BPF_MOV32_IMM(R0, 1),
5286 BPF_EXIT_INSN(),
5287 },
5288 INTERNAL,
5289 { },
5290 { { 0, 0x1 } },
5291 },
5292 {
5293 "ALU_ADD_K: 0 + 0x80000000 = 0x80000000",
5294 .u.insns_int = {
5295 BPF_LD_IMM64(R2, 0x0),
5296 BPF_LD_IMM64(R3, 0x80000000),
5297 BPF_ALU32_IMM(BPF_ADD, R2, 0x80000000),
5298 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5299 BPF_MOV32_IMM(R0, 2),
5300 BPF_EXIT_INSN(),
5301 BPF_MOV32_IMM(R0, 1),
5302 BPF_EXIT_INSN(),
5303 },
5304 INTERNAL,
5305 { },
5306 { { 0, 0x1 } },
5307 },
5308 {
5309 "ALU_ADD_K: 0 + 0x80008000 = 0x80008000",
5310 .u.insns_int = {
5311 BPF_LD_IMM64(R2, 0x0),
5312 BPF_LD_IMM64(R3, 0x80008000),
5313 BPF_ALU32_IMM(BPF_ADD, R2, 0x80008000),
5314 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5315 BPF_MOV32_IMM(R0, 2),
5316 BPF_EXIT_INSN(),
5317 BPF_MOV32_IMM(R0, 1),
5318 BPF_EXIT_INSN(),
5319 },
5320 INTERNAL,
5321 { },
5322 { { 0, 0x1 } },
5323 },
5324 {
Michael Holzheucffc6422015-05-11 22:22:44 -07005325 "ALU64_ADD_K: 1 + 2 = 3",
5326 .u.insns_int = {
5327 BPF_LD_IMM64(R0, 1),
5328 BPF_ALU64_IMM(BPF_ADD, R0, 2),
5329 BPF_EXIT_INSN(),
5330 },
5331 INTERNAL,
5332 { },
5333 { { 0, 3 } },
5334 },
5335 {
5336 "ALU64_ADD_K: 3 + 0 = 3",
5337 .u.insns_int = {
5338 BPF_LD_IMM64(R0, 3),
5339 BPF_ALU64_IMM(BPF_ADD, R0, 0),
5340 BPF_EXIT_INSN(),
5341 },
5342 INTERNAL,
5343 { },
5344 { { 0, 3 } },
5345 },
5346 {
5347 "ALU64_ADD_K: 1 + 2147483646 = 2147483647",
5348 .u.insns_int = {
5349 BPF_LD_IMM64(R0, 1),
5350 BPF_ALU64_IMM(BPF_ADD, R0, 2147483646),
5351 BPF_EXIT_INSN(),
5352 },
5353 INTERNAL,
5354 { },
5355 { { 0, 2147483647 } },
5356 },
5357 {
Naveen N. Raob64b50e2016-04-05 15:32:55 +05305358 "ALU64_ADD_K: 4294967294 + 2 = 4294967296",
5359 .u.insns_int = {
5360 BPF_LD_IMM64(R0, 4294967294U),
5361 BPF_LD_IMM64(R1, 4294967296ULL),
5362 BPF_ALU64_IMM(BPF_ADD, R0, 2),
5363 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
5364 BPF_ALU32_IMM(BPF_MOV, R0, 0),
5365 BPF_EXIT_INSN(),
5366 BPF_ALU32_IMM(BPF_MOV, R0, 1),
5367 BPF_EXIT_INSN(),
5368 },
5369 INTERNAL,
5370 { },
5371 { { 0, 1 } },
5372 },
5373 {
Michael Holzheucffc6422015-05-11 22:22:44 -07005374 "ALU64_ADD_K: 2147483646 + -2147483647 = -1",
5375 .u.insns_int = {
5376 BPF_LD_IMM64(R0, 2147483646),
5377 BPF_ALU64_IMM(BPF_ADD, R0, -2147483647),
5378 BPF_EXIT_INSN(),
5379 },
5380 INTERNAL,
5381 { },
5382 { { 0, -1 } },
5383 },
5384 {
5385 "ALU64_ADD_K: 1 + 0 = 1",
5386 .u.insns_int = {
5387 BPF_LD_IMM64(R2, 0x1),
5388 BPF_LD_IMM64(R3, 0x1),
5389 BPF_ALU64_IMM(BPF_ADD, R2, 0x0),
5390 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5391 BPF_MOV32_IMM(R0, 2),
5392 BPF_EXIT_INSN(),
5393 BPF_MOV32_IMM(R0, 1),
5394 BPF_EXIT_INSN(),
5395 },
5396 INTERNAL,
5397 { },
5398 { { 0, 0x1 } },
5399 },
5400 {
5401 "ALU64_ADD_K: 0 + (-1) = 0xffffffffffffffff",
5402 .u.insns_int = {
5403 BPF_LD_IMM64(R2, 0x0),
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005404 BPF_LD_IMM64(R3, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07005405 BPF_ALU64_IMM(BPF_ADD, R2, 0xffffffff),
5406 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5407 BPF_MOV32_IMM(R0, 2),
5408 BPF_EXIT_INSN(),
5409 BPF_MOV32_IMM(R0, 1),
5410 BPF_EXIT_INSN(),
5411 },
5412 INTERNAL,
5413 { },
5414 { { 0, 0x1 } },
5415 },
Naveen N. Rao9c94f6c2016-04-05 15:32:56 +05305416 {
5417 "ALU64_ADD_K: 0 + 0xffff = 0xffff",
5418 .u.insns_int = {
5419 BPF_LD_IMM64(R2, 0x0),
5420 BPF_LD_IMM64(R3, 0xffff),
5421 BPF_ALU64_IMM(BPF_ADD, R2, 0xffff),
5422 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5423 BPF_MOV32_IMM(R0, 2),
5424 BPF_EXIT_INSN(),
5425 BPF_MOV32_IMM(R0, 1),
5426 BPF_EXIT_INSN(),
5427 },
5428 INTERNAL,
5429 { },
5430 { { 0, 0x1 } },
5431 },
5432 {
5433 "ALU64_ADD_K: 0 + 0x7fffffff = 0x7fffffff",
5434 .u.insns_int = {
5435 BPF_LD_IMM64(R2, 0x0),
5436 BPF_LD_IMM64(R3, 0x7fffffff),
5437 BPF_ALU64_IMM(BPF_ADD, R2, 0x7fffffff),
5438 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5439 BPF_MOV32_IMM(R0, 2),
5440 BPF_EXIT_INSN(),
5441 BPF_MOV32_IMM(R0, 1),
5442 BPF_EXIT_INSN(),
5443 },
5444 INTERNAL,
5445 { },
5446 { { 0, 0x1 } },
5447 },
5448 {
5449 "ALU64_ADD_K: 0 + 0x80000000 = 0xffffffff80000000",
5450 .u.insns_int = {
5451 BPF_LD_IMM64(R2, 0x0),
5452 BPF_LD_IMM64(R3, 0xffffffff80000000LL),
5453 BPF_ALU64_IMM(BPF_ADD, R2, 0x80000000),
5454 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5455 BPF_MOV32_IMM(R0, 2),
5456 BPF_EXIT_INSN(),
5457 BPF_MOV32_IMM(R0, 1),
5458 BPF_EXIT_INSN(),
5459 },
5460 INTERNAL,
5461 { },
5462 { { 0, 0x1 } },
5463 },
5464 {
5465 "ALU_ADD_K: 0 + 0x80008000 = 0xffffffff80008000",
5466 .u.insns_int = {
5467 BPF_LD_IMM64(R2, 0x0),
5468 BPF_LD_IMM64(R3, 0xffffffff80008000LL),
5469 BPF_ALU64_IMM(BPF_ADD, R2, 0x80008000),
5470 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5471 BPF_MOV32_IMM(R0, 2),
5472 BPF_EXIT_INSN(),
5473 BPF_MOV32_IMM(R0, 1),
5474 BPF_EXIT_INSN(),
5475 },
5476 INTERNAL,
5477 { },
5478 { { 0, 0x1 } },
5479 },
Michael Holzheucffc6422015-05-11 22:22:44 -07005480 /* BPF_ALU | BPF_SUB | BPF_X */
5481 {
5482 "ALU_SUB_X: 3 - 1 = 2",
5483 .u.insns_int = {
5484 BPF_LD_IMM64(R0, 3),
5485 BPF_ALU32_IMM(BPF_MOV, R1, 1),
5486 BPF_ALU32_REG(BPF_SUB, R0, R1),
5487 BPF_EXIT_INSN(),
5488 },
5489 INTERNAL,
5490 { },
5491 { { 0, 2 } },
5492 },
5493 {
5494 "ALU_SUB_X: 4294967295 - 4294967294 = 1",
5495 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005496 BPF_LD_IMM64(R0, 4294967295U),
5497 BPF_ALU32_IMM(BPF_MOV, R1, 4294967294U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005498 BPF_ALU32_REG(BPF_SUB, R0, R1),
5499 BPF_EXIT_INSN(),
5500 },
5501 INTERNAL,
5502 { },
5503 { { 0, 1 } },
5504 },
5505 {
5506 "ALU64_SUB_X: 3 - 1 = 2",
5507 .u.insns_int = {
5508 BPF_LD_IMM64(R0, 3),
5509 BPF_ALU32_IMM(BPF_MOV, R1, 1),
5510 BPF_ALU64_REG(BPF_SUB, R0, R1),
5511 BPF_EXIT_INSN(),
5512 },
5513 INTERNAL,
5514 { },
5515 { { 0, 2 } },
5516 },
5517 {
5518 "ALU64_SUB_X: 4294967295 - 4294967294 = 1",
5519 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005520 BPF_LD_IMM64(R0, 4294967295U),
5521 BPF_ALU32_IMM(BPF_MOV, R1, 4294967294U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005522 BPF_ALU64_REG(BPF_SUB, R0, R1),
5523 BPF_EXIT_INSN(),
5524 },
5525 INTERNAL,
5526 { },
5527 { { 0, 1 } },
5528 },
5529 /* BPF_ALU | BPF_SUB | BPF_K */
5530 {
5531 "ALU_SUB_K: 3 - 1 = 2",
5532 .u.insns_int = {
5533 BPF_LD_IMM64(R0, 3),
5534 BPF_ALU32_IMM(BPF_SUB, R0, 1),
5535 BPF_EXIT_INSN(),
5536 },
5537 INTERNAL,
5538 { },
5539 { { 0, 2 } },
5540 },
5541 {
5542 "ALU_SUB_K: 3 - 0 = 3",
5543 .u.insns_int = {
5544 BPF_LD_IMM64(R0, 3),
5545 BPF_ALU32_IMM(BPF_SUB, R0, 0),
5546 BPF_EXIT_INSN(),
5547 },
5548 INTERNAL,
5549 { },
5550 { { 0, 3 } },
5551 },
5552 {
5553 "ALU_SUB_K: 4294967295 - 4294967294 = 1",
5554 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005555 BPF_LD_IMM64(R0, 4294967295U),
5556 BPF_ALU32_IMM(BPF_SUB, R0, 4294967294U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005557 BPF_EXIT_INSN(),
5558 },
5559 INTERNAL,
5560 { },
5561 { { 0, 1 } },
5562 },
5563 {
5564 "ALU64_SUB_K: 3 - 1 = 2",
5565 .u.insns_int = {
5566 BPF_LD_IMM64(R0, 3),
5567 BPF_ALU64_IMM(BPF_SUB, R0, 1),
5568 BPF_EXIT_INSN(),
5569 },
5570 INTERNAL,
5571 { },
5572 { { 0, 2 } },
5573 },
5574 {
5575 "ALU64_SUB_K: 3 - 0 = 3",
5576 .u.insns_int = {
5577 BPF_LD_IMM64(R0, 3),
5578 BPF_ALU64_IMM(BPF_SUB, R0, 0),
5579 BPF_EXIT_INSN(),
5580 },
5581 INTERNAL,
5582 { },
5583 { { 0, 3 } },
5584 },
5585 {
5586 "ALU64_SUB_K: 4294967294 - 4294967295 = -1",
5587 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005588 BPF_LD_IMM64(R0, 4294967294U),
5589 BPF_ALU64_IMM(BPF_SUB, R0, 4294967295U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005590 BPF_EXIT_INSN(),
5591 },
5592 INTERNAL,
5593 { },
5594 { { 0, -1 } },
5595 },
5596 {
5597 "ALU64_ADD_K: 2147483646 - 2147483647 = -1",
5598 .u.insns_int = {
5599 BPF_LD_IMM64(R0, 2147483646),
5600 BPF_ALU64_IMM(BPF_SUB, R0, 2147483647),
5601 BPF_EXIT_INSN(),
5602 },
5603 INTERNAL,
5604 { },
5605 { { 0, -1 } },
5606 },
5607 /* BPF_ALU | BPF_MUL | BPF_X */
5608 {
5609 "ALU_MUL_X: 2 * 3 = 6",
5610 .u.insns_int = {
5611 BPF_LD_IMM64(R0, 2),
5612 BPF_ALU32_IMM(BPF_MOV, R1, 3),
5613 BPF_ALU32_REG(BPF_MUL, R0, R1),
5614 BPF_EXIT_INSN(),
5615 },
5616 INTERNAL,
5617 { },
5618 { { 0, 6 } },
5619 },
5620 {
5621 "ALU_MUL_X: 2 * 0x7FFFFFF8 = 0xFFFFFFF0",
5622 .u.insns_int = {
5623 BPF_LD_IMM64(R0, 2),
5624 BPF_ALU32_IMM(BPF_MOV, R1, 0x7FFFFFF8),
5625 BPF_ALU32_REG(BPF_MUL, R0, R1),
5626 BPF_EXIT_INSN(),
5627 },
5628 INTERNAL,
5629 { },
5630 { { 0, 0xFFFFFFF0 } },
5631 },
5632 {
5633 "ALU_MUL_X: -1 * -1 = 1",
5634 .u.insns_int = {
5635 BPF_LD_IMM64(R0, -1),
5636 BPF_ALU32_IMM(BPF_MOV, R1, -1),
5637 BPF_ALU32_REG(BPF_MUL, R0, R1),
5638 BPF_EXIT_INSN(),
5639 },
5640 INTERNAL,
5641 { },
5642 { { 0, 1 } },
5643 },
5644 {
5645 "ALU64_MUL_X: 2 * 3 = 6",
5646 .u.insns_int = {
5647 BPF_LD_IMM64(R0, 2),
5648 BPF_ALU32_IMM(BPF_MOV, R1, 3),
5649 BPF_ALU64_REG(BPF_MUL, R0, R1),
5650 BPF_EXIT_INSN(),
5651 },
5652 INTERNAL,
5653 { },
5654 { { 0, 6 } },
5655 },
5656 {
5657 "ALU64_MUL_X: 1 * 2147483647 = 2147483647",
5658 .u.insns_int = {
5659 BPF_LD_IMM64(R0, 1),
5660 BPF_ALU32_IMM(BPF_MOV, R1, 2147483647),
5661 BPF_ALU64_REG(BPF_MUL, R0, R1),
5662 BPF_EXIT_INSN(),
5663 },
5664 INTERNAL,
5665 { },
5666 { { 0, 2147483647 } },
5667 },
Johan Almbladhfaa57622021-08-09 11:18:22 +02005668 {
5669 "ALU64_MUL_X: 64x64 multiply, low word",
5670 .u.insns_int = {
5671 BPF_LD_IMM64(R0, 0x0fedcba987654321LL),
5672 BPF_LD_IMM64(R1, 0x123456789abcdef0LL),
5673 BPF_ALU64_REG(BPF_MUL, R0, R1),
5674 BPF_EXIT_INSN(),
5675 },
5676 INTERNAL,
5677 { },
5678 { { 0, 0xe5618cf0 } }
5679 },
5680 {
5681 "ALU64_MUL_X: 64x64 multiply, high word",
5682 .u.insns_int = {
5683 BPF_LD_IMM64(R0, 0x0fedcba987654321LL),
5684 BPF_LD_IMM64(R1, 0x123456789abcdef0LL),
5685 BPF_ALU64_REG(BPF_MUL, R0, R1),
5686 BPF_ALU64_IMM(BPF_RSH, R0, 32),
5687 BPF_EXIT_INSN(),
5688 },
5689 INTERNAL,
5690 { },
5691 { { 0, 0x2236d88f } }
5692 },
Michael Holzheucffc6422015-05-11 22:22:44 -07005693 /* BPF_ALU | BPF_MUL | BPF_K */
5694 {
5695 "ALU_MUL_K: 2 * 3 = 6",
5696 .u.insns_int = {
5697 BPF_LD_IMM64(R0, 2),
5698 BPF_ALU32_IMM(BPF_MUL, R0, 3),
5699 BPF_EXIT_INSN(),
5700 },
5701 INTERNAL,
5702 { },
5703 { { 0, 6 } },
5704 },
5705 {
5706 "ALU_MUL_K: 3 * 1 = 3",
5707 .u.insns_int = {
5708 BPF_LD_IMM64(R0, 3),
5709 BPF_ALU32_IMM(BPF_MUL, R0, 1),
5710 BPF_EXIT_INSN(),
5711 },
5712 INTERNAL,
5713 { },
5714 { { 0, 3 } },
5715 },
5716 {
5717 "ALU_MUL_K: 2 * 0x7FFFFFF8 = 0xFFFFFFF0",
5718 .u.insns_int = {
5719 BPF_LD_IMM64(R0, 2),
5720 BPF_ALU32_IMM(BPF_MUL, R0, 0x7FFFFFF8),
5721 BPF_EXIT_INSN(),
5722 },
5723 INTERNAL,
5724 { },
5725 { { 0, 0xFFFFFFF0 } },
5726 },
5727 {
5728 "ALU_MUL_K: 1 * (-1) = 0x00000000ffffffff",
5729 .u.insns_int = {
5730 BPF_LD_IMM64(R2, 0x1),
5731 BPF_LD_IMM64(R3, 0x00000000ffffffff),
5732 BPF_ALU32_IMM(BPF_MUL, R2, 0xffffffff),
5733 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5734 BPF_MOV32_IMM(R0, 2),
5735 BPF_EXIT_INSN(),
5736 BPF_MOV32_IMM(R0, 1),
5737 BPF_EXIT_INSN(),
5738 },
5739 INTERNAL,
5740 { },
5741 { { 0, 0x1 } },
5742 },
5743 {
5744 "ALU64_MUL_K: 2 * 3 = 6",
5745 .u.insns_int = {
5746 BPF_LD_IMM64(R0, 2),
5747 BPF_ALU64_IMM(BPF_MUL, R0, 3),
5748 BPF_EXIT_INSN(),
5749 },
5750 INTERNAL,
5751 { },
5752 { { 0, 6 } },
5753 },
5754 {
5755 "ALU64_MUL_K: 3 * 1 = 3",
5756 .u.insns_int = {
5757 BPF_LD_IMM64(R0, 3),
5758 BPF_ALU64_IMM(BPF_MUL, R0, 1),
5759 BPF_EXIT_INSN(),
5760 },
5761 INTERNAL,
5762 { },
5763 { { 0, 3 } },
5764 },
5765 {
5766 "ALU64_MUL_K: 1 * 2147483647 = 2147483647",
5767 .u.insns_int = {
5768 BPF_LD_IMM64(R0, 1),
5769 BPF_ALU64_IMM(BPF_MUL, R0, 2147483647),
5770 BPF_EXIT_INSN(),
5771 },
5772 INTERNAL,
5773 { },
5774 { { 0, 2147483647 } },
5775 },
5776 {
5777 "ALU64_MUL_K: 1 * -2147483647 = -2147483647",
5778 .u.insns_int = {
5779 BPF_LD_IMM64(R0, 1),
5780 BPF_ALU64_IMM(BPF_MUL, R0, -2147483647),
5781 BPF_EXIT_INSN(),
5782 },
5783 INTERNAL,
5784 { },
5785 { { 0, -2147483647 } },
5786 },
5787 {
5788 "ALU64_MUL_K: 1 * (-1) = 0xffffffffffffffff",
5789 .u.insns_int = {
5790 BPF_LD_IMM64(R2, 0x1),
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005791 BPF_LD_IMM64(R3, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07005792 BPF_ALU64_IMM(BPF_MUL, R2, 0xffffffff),
5793 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5794 BPF_MOV32_IMM(R0, 2),
5795 BPF_EXIT_INSN(),
5796 BPF_MOV32_IMM(R0, 1),
5797 BPF_EXIT_INSN(),
5798 },
5799 INTERNAL,
5800 { },
5801 { { 0, 0x1 } },
5802 },
Johan Almbladhfaa57622021-08-09 11:18:22 +02005803 {
5804 "ALU64_MUL_K: 64x32 multiply, low word",
5805 .u.insns_int = {
5806 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
5807 BPF_ALU64_IMM(BPF_MUL, R0, 0x12345678),
5808 BPF_EXIT_INSN(),
5809 },
5810 INTERNAL,
5811 { },
5812 { { 0, 0xe242d208 } }
5813 },
5814 {
5815 "ALU64_MUL_K: 64x32 multiply, high word",
5816 .u.insns_int = {
5817 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
5818 BPF_ALU64_IMM(BPF_MUL, R0, 0x12345678),
5819 BPF_ALU64_IMM(BPF_RSH, R0, 32),
5820 BPF_EXIT_INSN(),
5821 },
5822 INTERNAL,
5823 { },
5824 { { 0, 0xc28f5c28 } }
5825 },
Michael Holzheucffc6422015-05-11 22:22:44 -07005826 /* BPF_ALU | BPF_DIV | BPF_X */
5827 {
5828 "ALU_DIV_X: 6 / 2 = 3",
5829 .u.insns_int = {
5830 BPF_LD_IMM64(R0, 6),
5831 BPF_ALU32_IMM(BPF_MOV, R1, 2),
5832 BPF_ALU32_REG(BPF_DIV, R0, R1),
5833 BPF_EXIT_INSN(),
5834 },
5835 INTERNAL,
5836 { },
5837 { { 0, 3 } },
5838 },
5839 {
5840 "ALU_DIV_X: 4294967295 / 4294967295 = 1",
5841 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005842 BPF_LD_IMM64(R0, 4294967295U),
5843 BPF_ALU32_IMM(BPF_MOV, R1, 4294967295U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005844 BPF_ALU32_REG(BPF_DIV, R0, R1),
5845 BPF_EXIT_INSN(),
5846 },
5847 INTERNAL,
5848 { },
5849 { { 0, 1 } },
5850 },
5851 {
5852 "ALU64_DIV_X: 6 / 2 = 3",
5853 .u.insns_int = {
5854 BPF_LD_IMM64(R0, 6),
5855 BPF_ALU32_IMM(BPF_MOV, R1, 2),
5856 BPF_ALU64_REG(BPF_DIV, R0, R1),
5857 BPF_EXIT_INSN(),
5858 },
5859 INTERNAL,
5860 { },
5861 { { 0, 3 } },
5862 },
5863 {
5864 "ALU64_DIV_X: 2147483647 / 2147483647 = 1",
5865 .u.insns_int = {
5866 BPF_LD_IMM64(R0, 2147483647),
5867 BPF_ALU32_IMM(BPF_MOV, R1, 2147483647),
5868 BPF_ALU64_REG(BPF_DIV, R0, R1),
5869 BPF_EXIT_INSN(),
5870 },
5871 INTERNAL,
5872 { },
5873 { { 0, 1 } },
5874 },
5875 {
5876 "ALU64_DIV_X: 0xffffffffffffffff / (-1) = 0x0000000000000001",
5877 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005878 BPF_LD_IMM64(R2, 0xffffffffffffffffLL),
5879 BPF_LD_IMM64(R4, 0xffffffffffffffffLL),
5880 BPF_LD_IMM64(R3, 0x0000000000000001LL),
Michael Holzheucffc6422015-05-11 22:22:44 -07005881 BPF_ALU64_REG(BPF_DIV, R2, R4),
5882 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5883 BPF_MOV32_IMM(R0, 2),
5884 BPF_EXIT_INSN(),
5885 BPF_MOV32_IMM(R0, 1),
5886 BPF_EXIT_INSN(),
5887 },
5888 INTERNAL,
5889 { },
5890 { { 0, 0x1 } },
5891 },
5892 /* BPF_ALU | BPF_DIV | BPF_K */
5893 {
5894 "ALU_DIV_K: 6 / 2 = 3",
5895 .u.insns_int = {
5896 BPF_LD_IMM64(R0, 6),
5897 BPF_ALU32_IMM(BPF_DIV, R0, 2),
5898 BPF_EXIT_INSN(),
5899 },
5900 INTERNAL,
5901 { },
5902 { { 0, 3 } },
5903 },
5904 {
5905 "ALU_DIV_K: 3 / 1 = 3",
5906 .u.insns_int = {
5907 BPF_LD_IMM64(R0, 3),
5908 BPF_ALU32_IMM(BPF_DIV, R0, 1),
5909 BPF_EXIT_INSN(),
5910 },
5911 INTERNAL,
5912 { },
5913 { { 0, 3 } },
5914 },
5915 {
5916 "ALU_DIV_K: 4294967295 / 4294967295 = 1",
5917 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005918 BPF_LD_IMM64(R0, 4294967295U),
5919 BPF_ALU32_IMM(BPF_DIV, R0, 4294967295U),
Michael Holzheucffc6422015-05-11 22:22:44 -07005920 BPF_EXIT_INSN(),
5921 },
5922 INTERNAL,
5923 { },
5924 { { 0, 1 } },
5925 },
5926 {
5927 "ALU_DIV_K: 0xffffffffffffffff / (-1) = 0x1",
5928 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005929 BPF_LD_IMM64(R2, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07005930 BPF_LD_IMM64(R3, 0x1UL),
5931 BPF_ALU32_IMM(BPF_DIV, R2, 0xffffffff),
5932 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5933 BPF_MOV32_IMM(R0, 2),
5934 BPF_EXIT_INSN(),
5935 BPF_MOV32_IMM(R0, 1),
5936 BPF_EXIT_INSN(),
5937 },
5938 INTERNAL,
5939 { },
5940 { { 0, 0x1 } },
5941 },
5942 {
5943 "ALU64_DIV_K: 6 / 2 = 3",
5944 .u.insns_int = {
5945 BPF_LD_IMM64(R0, 6),
5946 BPF_ALU64_IMM(BPF_DIV, R0, 2),
5947 BPF_EXIT_INSN(),
5948 },
5949 INTERNAL,
5950 { },
5951 { { 0, 3 } },
5952 },
5953 {
5954 "ALU64_DIV_K: 3 / 1 = 3",
5955 .u.insns_int = {
5956 BPF_LD_IMM64(R0, 3),
5957 BPF_ALU64_IMM(BPF_DIV, R0, 1),
5958 BPF_EXIT_INSN(),
5959 },
5960 INTERNAL,
5961 { },
5962 { { 0, 3 } },
5963 },
5964 {
5965 "ALU64_DIV_K: 2147483647 / 2147483647 = 1",
5966 .u.insns_int = {
5967 BPF_LD_IMM64(R0, 2147483647),
5968 BPF_ALU64_IMM(BPF_DIV, R0, 2147483647),
5969 BPF_EXIT_INSN(),
5970 },
5971 INTERNAL,
5972 { },
5973 { { 0, 1 } },
5974 },
5975 {
5976 "ALU64_DIV_K: 0xffffffffffffffff / (-1) = 0x0000000000000001",
5977 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07005978 BPF_LD_IMM64(R2, 0xffffffffffffffffLL),
5979 BPF_LD_IMM64(R3, 0x0000000000000001LL),
Michael Holzheucffc6422015-05-11 22:22:44 -07005980 BPF_ALU64_IMM(BPF_DIV, R2, 0xffffffff),
5981 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
5982 BPF_MOV32_IMM(R0, 2),
5983 BPF_EXIT_INSN(),
5984 BPF_MOV32_IMM(R0, 1),
5985 BPF_EXIT_INSN(),
5986 },
5987 INTERNAL,
5988 { },
5989 { { 0, 0x1 } },
5990 },
5991 /* BPF_ALU | BPF_MOD | BPF_X */
5992 {
5993 "ALU_MOD_X: 3 % 2 = 1",
5994 .u.insns_int = {
5995 BPF_LD_IMM64(R0, 3),
5996 BPF_ALU32_IMM(BPF_MOV, R1, 2),
5997 BPF_ALU32_REG(BPF_MOD, R0, R1),
5998 BPF_EXIT_INSN(),
5999 },
6000 INTERNAL,
6001 { },
6002 { { 0, 1 } },
6003 },
6004 {
6005 "ALU_MOD_X: 4294967295 % 4294967293 = 2",
6006 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006007 BPF_LD_IMM64(R0, 4294967295U),
6008 BPF_ALU32_IMM(BPF_MOV, R1, 4294967293U),
Michael Holzheucffc6422015-05-11 22:22:44 -07006009 BPF_ALU32_REG(BPF_MOD, R0, R1),
6010 BPF_EXIT_INSN(),
6011 },
6012 INTERNAL,
6013 { },
6014 { { 0, 2 } },
6015 },
6016 {
6017 "ALU64_MOD_X: 3 % 2 = 1",
6018 .u.insns_int = {
6019 BPF_LD_IMM64(R0, 3),
6020 BPF_ALU32_IMM(BPF_MOV, R1, 2),
6021 BPF_ALU64_REG(BPF_MOD, R0, R1),
6022 BPF_EXIT_INSN(),
6023 },
6024 INTERNAL,
6025 { },
6026 { { 0, 1 } },
6027 },
6028 {
6029 "ALU64_MOD_X: 2147483647 % 2147483645 = 2",
6030 .u.insns_int = {
6031 BPF_LD_IMM64(R0, 2147483647),
6032 BPF_ALU32_IMM(BPF_MOV, R1, 2147483645),
6033 BPF_ALU64_REG(BPF_MOD, R0, R1),
6034 BPF_EXIT_INSN(),
6035 },
6036 INTERNAL,
6037 { },
6038 { { 0, 2 } },
6039 },
6040 /* BPF_ALU | BPF_MOD | BPF_K */
6041 {
6042 "ALU_MOD_K: 3 % 2 = 1",
6043 .u.insns_int = {
6044 BPF_LD_IMM64(R0, 3),
6045 BPF_ALU32_IMM(BPF_MOD, R0, 2),
6046 BPF_EXIT_INSN(),
6047 },
6048 INTERNAL,
6049 { },
6050 { { 0, 1 } },
6051 },
6052 {
6053 "ALU_MOD_K: 3 % 1 = 0",
6054 .u.insns_int = {
6055 BPF_LD_IMM64(R0, 3),
6056 BPF_ALU32_IMM(BPF_MOD, R0, 1),
6057 BPF_EXIT_INSN(),
6058 },
6059 INTERNAL,
6060 { },
6061 { { 0, 0 } },
6062 },
6063 {
6064 "ALU_MOD_K: 4294967295 % 4294967293 = 2",
6065 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006066 BPF_LD_IMM64(R0, 4294967295U),
6067 BPF_ALU32_IMM(BPF_MOD, R0, 4294967293U),
Michael Holzheucffc6422015-05-11 22:22:44 -07006068 BPF_EXIT_INSN(),
6069 },
6070 INTERNAL,
6071 { },
6072 { { 0, 2 } },
6073 },
6074 {
6075 "ALU64_MOD_K: 3 % 2 = 1",
6076 .u.insns_int = {
6077 BPF_LD_IMM64(R0, 3),
6078 BPF_ALU64_IMM(BPF_MOD, R0, 2),
6079 BPF_EXIT_INSN(),
6080 },
6081 INTERNAL,
6082 { },
6083 { { 0, 1 } },
6084 },
6085 {
6086 "ALU64_MOD_K: 3 % 1 = 0",
6087 .u.insns_int = {
6088 BPF_LD_IMM64(R0, 3),
6089 BPF_ALU64_IMM(BPF_MOD, R0, 1),
6090 BPF_EXIT_INSN(),
6091 },
6092 INTERNAL,
6093 { },
6094 { { 0, 0 } },
6095 },
6096 {
6097 "ALU64_MOD_K: 2147483647 % 2147483645 = 2",
6098 .u.insns_int = {
6099 BPF_LD_IMM64(R0, 2147483647),
6100 BPF_ALU64_IMM(BPF_MOD, R0, 2147483645),
6101 BPF_EXIT_INSN(),
6102 },
6103 INTERNAL,
6104 { },
6105 { { 0, 2 } },
6106 },
6107 /* BPF_ALU | BPF_AND | BPF_X */
6108 {
6109 "ALU_AND_X: 3 & 2 = 2",
6110 .u.insns_int = {
6111 BPF_LD_IMM64(R0, 3),
6112 BPF_ALU32_IMM(BPF_MOV, R1, 2),
6113 BPF_ALU32_REG(BPF_AND, R0, R1),
6114 BPF_EXIT_INSN(),
6115 },
6116 INTERNAL,
6117 { },
6118 { { 0, 2 } },
6119 },
6120 {
6121 "ALU_AND_X: 0xffffffff & 0xffffffff = 0xffffffff",
6122 .u.insns_int = {
6123 BPF_LD_IMM64(R0, 0xffffffff),
6124 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
6125 BPF_ALU32_REG(BPF_AND, R0, R1),
6126 BPF_EXIT_INSN(),
6127 },
6128 INTERNAL,
6129 { },
6130 { { 0, 0xffffffff } },
6131 },
6132 {
6133 "ALU64_AND_X: 3 & 2 = 2",
6134 .u.insns_int = {
6135 BPF_LD_IMM64(R0, 3),
6136 BPF_ALU32_IMM(BPF_MOV, R1, 2),
6137 BPF_ALU64_REG(BPF_AND, R0, R1),
6138 BPF_EXIT_INSN(),
6139 },
6140 INTERNAL,
6141 { },
6142 { { 0, 2 } },
6143 },
6144 {
6145 "ALU64_AND_X: 0xffffffff & 0xffffffff = 0xffffffff",
6146 .u.insns_int = {
6147 BPF_LD_IMM64(R0, 0xffffffff),
6148 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
6149 BPF_ALU64_REG(BPF_AND, R0, R1),
6150 BPF_EXIT_INSN(),
6151 },
6152 INTERNAL,
6153 { },
6154 { { 0, 0xffffffff } },
6155 },
6156 /* BPF_ALU | BPF_AND | BPF_K */
6157 {
6158 "ALU_AND_K: 3 & 2 = 2",
6159 .u.insns_int = {
6160 BPF_LD_IMM64(R0, 3),
6161 BPF_ALU32_IMM(BPF_AND, R0, 2),
6162 BPF_EXIT_INSN(),
6163 },
6164 INTERNAL,
6165 { },
6166 { { 0, 2 } },
6167 },
6168 {
6169 "ALU_AND_K: 0xffffffff & 0xffffffff = 0xffffffff",
6170 .u.insns_int = {
6171 BPF_LD_IMM64(R0, 0xffffffff),
6172 BPF_ALU32_IMM(BPF_AND, R0, 0xffffffff),
6173 BPF_EXIT_INSN(),
6174 },
6175 INTERNAL,
6176 { },
6177 { { 0, 0xffffffff } },
6178 },
6179 {
Johan Almbladhba89bcf2021-08-09 11:18:19 +02006180 "ALU_AND_K: Small immediate",
6181 .u.insns_int = {
6182 BPF_ALU32_IMM(BPF_MOV, R0, 0x01020304),
6183 BPF_ALU32_IMM(BPF_AND, R0, 15),
6184 BPF_EXIT_INSN(),
6185 },
6186 INTERNAL,
6187 { },
6188 { { 0, 4 } }
6189 },
6190 {
6191 "ALU_AND_K: Large immediate",
6192 .u.insns_int = {
6193 BPF_ALU32_IMM(BPF_MOV, R0, 0xf1f2f3f4),
6194 BPF_ALU32_IMM(BPF_AND, R0, 0xafbfcfdf),
6195 BPF_EXIT_INSN(),
6196 },
6197 INTERNAL,
6198 { },
6199 { { 0, 0xa1b2c3d4 } }
6200 },
6201 {
6202 "ALU_AND_K: Zero extension",
6203 .u.insns_int = {
6204 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6205 BPF_LD_IMM64(R1, 0x0000000080a0c0e0LL),
6206 BPF_ALU32_IMM(BPF_AND, R0, 0xf0f0f0f0),
6207 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6208 BPF_MOV32_IMM(R0, 2),
6209 BPF_EXIT_INSN(),
6210 BPF_MOV32_IMM(R0, 1),
6211 BPF_EXIT_INSN(),
6212 },
6213 INTERNAL,
6214 { },
6215 { { 0, 1 } }
6216 },
6217 {
Michael Holzheucffc6422015-05-11 22:22:44 -07006218 "ALU64_AND_K: 3 & 2 = 2",
6219 .u.insns_int = {
6220 BPF_LD_IMM64(R0, 3),
6221 BPF_ALU64_IMM(BPF_AND, R0, 2),
6222 BPF_EXIT_INSN(),
6223 },
6224 INTERNAL,
6225 { },
6226 { { 0, 2 } },
6227 },
6228 {
6229 "ALU64_AND_K: 0xffffffff & 0xffffffff = 0xffffffff",
6230 .u.insns_int = {
6231 BPF_LD_IMM64(R0, 0xffffffff),
6232 BPF_ALU64_IMM(BPF_AND, R0, 0xffffffff),
6233 BPF_EXIT_INSN(),
6234 },
6235 INTERNAL,
6236 { },
6237 { { 0, 0xffffffff } },
6238 },
6239 {
Johan Almbladhe92c8132021-08-09 11:18:18 +02006240 "ALU64_AND_K: 0x0000ffffffff0000 & 0x0 = 0x0000000000000000",
Michael Holzheucffc6422015-05-11 22:22:44 -07006241 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006242 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
6243 BPF_LD_IMM64(R3, 0x0000000000000000LL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006244 BPF_ALU64_IMM(BPF_AND, R2, 0x0),
6245 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6246 BPF_MOV32_IMM(R0, 2),
6247 BPF_EXIT_INSN(),
6248 BPF_MOV32_IMM(R0, 1),
6249 BPF_EXIT_INSN(),
6250 },
6251 INTERNAL,
6252 { },
6253 { { 0, 0x1 } },
6254 },
6255 {
Johan Almbladhe92c8132021-08-09 11:18:18 +02006256 "ALU64_AND_K: 0x0000ffffffff0000 & -1 = 0x0000ffffffff0000",
Michael Holzheucffc6422015-05-11 22:22:44 -07006257 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006258 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
6259 BPF_LD_IMM64(R3, 0x0000ffffffff0000LL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006260 BPF_ALU64_IMM(BPF_AND, R2, 0xffffffff),
6261 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6262 BPF_MOV32_IMM(R0, 2),
6263 BPF_EXIT_INSN(),
6264 BPF_MOV32_IMM(R0, 1),
6265 BPF_EXIT_INSN(),
6266 },
6267 INTERNAL,
6268 { },
6269 { { 0, 0x1 } },
6270 },
6271 {
6272 "ALU64_AND_K: 0xffffffffffffffff & -1 = 0xffffffffffffffff",
6273 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006274 BPF_LD_IMM64(R2, 0xffffffffffffffffLL),
6275 BPF_LD_IMM64(R3, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006276 BPF_ALU64_IMM(BPF_AND, R2, 0xffffffff),
6277 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6278 BPF_MOV32_IMM(R0, 2),
6279 BPF_EXIT_INSN(),
6280 BPF_MOV32_IMM(R0, 1),
6281 BPF_EXIT_INSN(),
6282 },
6283 INTERNAL,
6284 { },
6285 { { 0, 0x1 } },
6286 },
Johan Almbladhba89bcf2021-08-09 11:18:19 +02006287 {
6288 "ALU64_AND_K: Sign extension 1",
6289 .u.insns_int = {
6290 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6291 BPF_LD_IMM64(R1, 0x00000000090b0d0fLL),
6292 BPF_ALU64_IMM(BPF_AND, R0, 0x0f0f0f0f),
6293 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6294 BPF_MOV32_IMM(R0, 2),
6295 BPF_EXIT_INSN(),
6296 BPF_MOV32_IMM(R0, 1),
6297 BPF_EXIT_INSN(),
6298 },
6299 INTERNAL,
6300 { },
6301 { { 0, 1 } }
6302 },
6303 {
6304 "ALU64_AND_K: Sign extension 2",
6305 .u.insns_int = {
6306 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6307 BPF_LD_IMM64(R1, 0x0123456780a0c0e0LL),
6308 BPF_ALU64_IMM(BPF_AND, R0, 0xf0f0f0f0),
6309 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6310 BPF_MOV32_IMM(R0, 2),
6311 BPF_EXIT_INSN(),
6312 BPF_MOV32_IMM(R0, 1),
6313 BPF_EXIT_INSN(),
6314 },
6315 INTERNAL,
6316 { },
6317 { { 0, 1 } }
6318 },
Michael Holzheucffc6422015-05-11 22:22:44 -07006319 /* BPF_ALU | BPF_OR | BPF_X */
6320 {
6321 "ALU_OR_X: 1 | 2 = 3",
6322 .u.insns_int = {
6323 BPF_LD_IMM64(R0, 1),
6324 BPF_ALU32_IMM(BPF_MOV, R1, 2),
6325 BPF_ALU32_REG(BPF_OR, R0, R1),
6326 BPF_EXIT_INSN(),
6327 },
6328 INTERNAL,
6329 { },
6330 { { 0, 3 } },
6331 },
6332 {
6333 "ALU_OR_X: 0x0 | 0xffffffff = 0xffffffff",
6334 .u.insns_int = {
6335 BPF_LD_IMM64(R0, 0),
6336 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
6337 BPF_ALU32_REG(BPF_OR, R0, R1),
6338 BPF_EXIT_INSN(),
6339 },
6340 INTERNAL,
6341 { },
6342 { { 0, 0xffffffff } },
6343 },
6344 {
6345 "ALU64_OR_X: 1 | 2 = 3",
6346 .u.insns_int = {
6347 BPF_LD_IMM64(R0, 1),
6348 BPF_ALU32_IMM(BPF_MOV, R1, 2),
6349 BPF_ALU64_REG(BPF_OR, R0, R1),
6350 BPF_EXIT_INSN(),
6351 },
6352 INTERNAL,
6353 { },
6354 { { 0, 3 } },
6355 },
6356 {
6357 "ALU64_OR_X: 0 | 0xffffffff = 0xffffffff",
6358 .u.insns_int = {
6359 BPF_LD_IMM64(R0, 0),
6360 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
6361 BPF_ALU64_REG(BPF_OR, R0, R1),
6362 BPF_EXIT_INSN(),
6363 },
6364 INTERNAL,
6365 { },
6366 { { 0, 0xffffffff } },
6367 },
6368 /* BPF_ALU | BPF_OR | BPF_K */
6369 {
6370 "ALU_OR_K: 1 | 2 = 3",
6371 .u.insns_int = {
6372 BPF_LD_IMM64(R0, 1),
6373 BPF_ALU32_IMM(BPF_OR, R0, 2),
6374 BPF_EXIT_INSN(),
6375 },
6376 INTERNAL,
6377 { },
6378 { { 0, 3 } },
6379 },
6380 {
6381 "ALU_OR_K: 0 & 0xffffffff = 0xffffffff",
6382 .u.insns_int = {
6383 BPF_LD_IMM64(R0, 0),
6384 BPF_ALU32_IMM(BPF_OR, R0, 0xffffffff),
6385 BPF_EXIT_INSN(),
6386 },
6387 INTERNAL,
6388 { },
6389 { { 0, 0xffffffff } },
6390 },
6391 {
Johan Almbladhba89bcf2021-08-09 11:18:19 +02006392 "ALU_OR_K: Small immediate",
6393 .u.insns_int = {
6394 BPF_ALU32_IMM(BPF_MOV, R0, 0x01020304),
6395 BPF_ALU32_IMM(BPF_OR, R0, 1),
6396 BPF_EXIT_INSN(),
6397 },
6398 INTERNAL,
6399 { },
6400 { { 0, 0x01020305 } }
6401 },
6402 {
6403 "ALU_OR_K: Large immediate",
6404 .u.insns_int = {
6405 BPF_ALU32_IMM(BPF_MOV, R0, 0x01020304),
6406 BPF_ALU32_IMM(BPF_OR, R0, 0xa0b0c0d0),
6407 BPF_EXIT_INSN(),
6408 },
6409 INTERNAL,
6410 { },
6411 { { 0, 0xa1b2c3d4 } }
6412 },
6413 {
6414 "ALU_OR_K: Zero extension",
6415 .u.insns_int = {
6416 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6417 BPF_LD_IMM64(R1, 0x00000000f9fbfdffLL),
6418 BPF_ALU32_IMM(BPF_OR, R0, 0xf0f0f0f0),
6419 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6420 BPF_MOV32_IMM(R0, 2),
6421 BPF_EXIT_INSN(),
6422 BPF_MOV32_IMM(R0, 1),
6423 BPF_EXIT_INSN(),
6424 },
6425 INTERNAL,
6426 { },
6427 { { 0, 1 } }
6428 },
6429 {
Michael Holzheucffc6422015-05-11 22:22:44 -07006430 "ALU64_OR_K: 1 | 2 = 3",
6431 .u.insns_int = {
6432 BPF_LD_IMM64(R0, 1),
6433 BPF_ALU64_IMM(BPF_OR, R0, 2),
6434 BPF_EXIT_INSN(),
6435 },
6436 INTERNAL,
6437 { },
6438 { { 0, 3 } },
6439 },
6440 {
6441 "ALU64_OR_K: 0 & 0xffffffff = 0xffffffff",
6442 .u.insns_int = {
6443 BPF_LD_IMM64(R0, 0),
6444 BPF_ALU64_IMM(BPF_OR, R0, 0xffffffff),
6445 BPF_EXIT_INSN(),
6446 },
6447 INTERNAL,
6448 { },
6449 { { 0, 0xffffffff } },
6450 },
6451 {
Johan Almbladhe92c8132021-08-09 11:18:18 +02006452 "ALU64_OR_K: 0x0000ffffffff0000 | 0x0 = 0x0000ffffffff0000",
Michael Holzheucffc6422015-05-11 22:22:44 -07006453 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006454 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
6455 BPF_LD_IMM64(R3, 0x0000ffffffff0000LL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006456 BPF_ALU64_IMM(BPF_OR, R2, 0x0),
6457 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6458 BPF_MOV32_IMM(R0, 2),
6459 BPF_EXIT_INSN(),
6460 BPF_MOV32_IMM(R0, 1),
6461 BPF_EXIT_INSN(),
6462 },
6463 INTERNAL,
6464 { },
6465 { { 0, 0x1 } },
6466 },
6467 {
6468 "ALU64_OR_K: 0x0000ffffffff0000 | -1 = 0xffffffffffffffff",
6469 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006470 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
6471 BPF_LD_IMM64(R3, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006472 BPF_ALU64_IMM(BPF_OR, R2, 0xffffffff),
6473 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6474 BPF_MOV32_IMM(R0, 2),
6475 BPF_EXIT_INSN(),
6476 BPF_MOV32_IMM(R0, 1),
6477 BPF_EXIT_INSN(),
6478 },
6479 INTERNAL,
6480 { },
6481 { { 0, 0x1 } },
6482 },
6483 {
6484 "ALU64_OR_K: 0x000000000000000 | -1 = 0xffffffffffffffff",
6485 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006486 BPF_LD_IMM64(R2, 0x0000000000000000LL),
6487 BPF_LD_IMM64(R3, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006488 BPF_ALU64_IMM(BPF_OR, R2, 0xffffffff),
6489 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6490 BPF_MOV32_IMM(R0, 2),
6491 BPF_EXIT_INSN(),
6492 BPF_MOV32_IMM(R0, 1),
6493 BPF_EXIT_INSN(),
6494 },
6495 INTERNAL,
6496 { },
6497 { { 0, 0x1 } },
6498 },
Johan Almbladhba89bcf2021-08-09 11:18:19 +02006499 {
6500 "ALU64_OR_K: Sign extension 1",
6501 .u.insns_int = {
6502 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6503 BPF_LD_IMM64(R1, 0x012345678fafcfefLL),
6504 BPF_ALU64_IMM(BPF_OR, R0, 0x0f0f0f0f),
6505 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6506 BPF_MOV32_IMM(R0, 2),
6507 BPF_EXIT_INSN(),
6508 BPF_MOV32_IMM(R0, 1),
6509 BPF_EXIT_INSN(),
6510 },
6511 INTERNAL,
6512 { },
6513 { { 0, 1 } }
6514 },
6515 {
6516 "ALU64_OR_K: Sign extension 2",
6517 .u.insns_int = {
6518 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6519 BPF_LD_IMM64(R1, 0xfffffffff9fbfdffLL),
6520 BPF_ALU64_IMM(BPF_OR, R0, 0xf0f0f0f0),
6521 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6522 BPF_MOV32_IMM(R0, 2),
6523 BPF_EXIT_INSN(),
6524 BPF_MOV32_IMM(R0, 1),
6525 BPF_EXIT_INSN(),
6526 },
6527 INTERNAL,
6528 { },
6529 { { 0, 1 } }
6530 },
Michael Holzheucffc6422015-05-11 22:22:44 -07006531 /* BPF_ALU | BPF_XOR | BPF_X */
6532 {
6533 "ALU_XOR_X: 5 ^ 6 = 3",
6534 .u.insns_int = {
6535 BPF_LD_IMM64(R0, 5),
6536 BPF_ALU32_IMM(BPF_MOV, R1, 6),
6537 BPF_ALU32_REG(BPF_XOR, R0, R1),
6538 BPF_EXIT_INSN(),
6539 },
6540 INTERNAL,
6541 { },
6542 { { 0, 3 } },
6543 },
6544 {
6545 "ALU_XOR_X: 0x1 ^ 0xffffffff = 0xfffffffe",
6546 .u.insns_int = {
6547 BPF_LD_IMM64(R0, 1),
6548 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
6549 BPF_ALU32_REG(BPF_XOR, R0, R1),
6550 BPF_EXIT_INSN(),
6551 },
6552 INTERNAL,
6553 { },
6554 { { 0, 0xfffffffe } },
6555 },
6556 {
6557 "ALU64_XOR_X: 5 ^ 6 = 3",
6558 .u.insns_int = {
6559 BPF_LD_IMM64(R0, 5),
6560 BPF_ALU32_IMM(BPF_MOV, R1, 6),
6561 BPF_ALU64_REG(BPF_XOR, R0, R1),
6562 BPF_EXIT_INSN(),
6563 },
6564 INTERNAL,
6565 { },
6566 { { 0, 3 } },
6567 },
6568 {
6569 "ALU64_XOR_X: 1 ^ 0xffffffff = 0xfffffffe",
6570 .u.insns_int = {
6571 BPF_LD_IMM64(R0, 1),
6572 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
6573 BPF_ALU64_REG(BPF_XOR, R0, R1),
6574 BPF_EXIT_INSN(),
6575 },
6576 INTERNAL,
6577 { },
6578 { { 0, 0xfffffffe } },
6579 },
6580 /* BPF_ALU | BPF_XOR | BPF_K */
6581 {
6582 "ALU_XOR_K: 5 ^ 6 = 3",
6583 .u.insns_int = {
6584 BPF_LD_IMM64(R0, 5),
6585 BPF_ALU32_IMM(BPF_XOR, R0, 6),
6586 BPF_EXIT_INSN(),
6587 },
6588 INTERNAL,
6589 { },
6590 { { 0, 3 } },
6591 },
6592 {
6593 "ALU_XOR_K: 1 ^ 0xffffffff = 0xfffffffe",
6594 .u.insns_int = {
6595 BPF_LD_IMM64(R0, 1),
6596 BPF_ALU32_IMM(BPF_XOR, R0, 0xffffffff),
6597 BPF_EXIT_INSN(),
6598 },
6599 INTERNAL,
6600 { },
6601 { { 0, 0xfffffffe } },
6602 },
6603 {
Johan Almbladhba89bcf2021-08-09 11:18:19 +02006604 "ALU_XOR_K: Small immediate",
6605 .u.insns_int = {
6606 BPF_ALU32_IMM(BPF_MOV, R0, 0x01020304),
6607 BPF_ALU32_IMM(BPF_XOR, R0, 15),
6608 BPF_EXIT_INSN(),
6609 },
6610 INTERNAL,
6611 { },
6612 { { 0, 0x0102030b } }
6613 },
6614 {
6615 "ALU_XOR_K: Large immediate",
6616 .u.insns_int = {
6617 BPF_ALU32_IMM(BPF_MOV, R0, 0xf1f2f3f4),
6618 BPF_ALU32_IMM(BPF_XOR, R0, 0xafbfcfdf),
6619 BPF_EXIT_INSN(),
6620 },
6621 INTERNAL,
6622 { },
6623 { { 0, 0x5e4d3c2b } }
6624 },
6625 {
6626 "ALU_XOR_K: Zero extension",
6627 .u.insns_int = {
6628 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6629 BPF_LD_IMM64(R1, 0x00000000795b3d1fLL),
6630 BPF_ALU32_IMM(BPF_XOR, R0, 0xf0f0f0f0),
6631 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6632 BPF_MOV32_IMM(R0, 2),
6633 BPF_EXIT_INSN(),
6634 BPF_MOV32_IMM(R0, 1),
6635 BPF_EXIT_INSN(),
6636 },
6637 INTERNAL,
6638 { },
6639 { { 0, 1 } }
6640 },
6641 {
Michael Holzheucffc6422015-05-11 22:22:44 -07006642 "ALU64_XOR_K: 5 ^ 6 = 3",
6643 .u.insns_int = {
6644 BPF_LD_IMM64(R0, 5),
6645 BPF_ALU64_IMM(BPF_XOR, R0, 6),
6646 BPF_EXIT_INSN(),
6647 },
6648 INTERNAL,
6649 { },
6650 { { 0, 3 } },
6651 },
6652 {
Johan Almbladhe92c8132021-08-09 11:18:18 +02006653 "ALU64_XOR_K: 1 ^ 0xffffffff = 0xfffffffe",
Michael Holzheucffc6422015-05-11 22:22:44 -07006654 .u.insns_int = {
6655 BPF_LD_IMM64(R0, 1),
6656 BPF_ALU64_IMM(BPF_XOR, R0, 0xffffffff),
6657 BPF_EXIT_INSN(),
6658 },
6659 INTERNAL,
6660 { },
6661 { { 0, 0xfffffffe } },
6662 },
6663 {
6664 "ALU64_XOR_K: 0x0000ffffffff0000 ^ 0x0 = 0x0000ffffffff0000",
6665 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006666 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
6667 BPF_LD_IMM64(R3, 0x0000ffffffff0000LL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006668 BPF_ALU64_IMM(BPF_XOR, R2, 0x0),
6669 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6670 BPF_MOV32_IMM(R0, 2),
6671 BPF_EXIT_INSN(),
6672 BPF_MOV32_IMM(R0, 1),
6673 BPF_EXIT_INSN(),
6674 },
6675 INTERNAL,
6676 { },
6677 { { 0, 0x1 } },
6678 },
6679 {
6680 "ALU64_XOR_K: 0x0000ffffffff0000 ^ -1 = 0xffff00000000ffff",
6681 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006682 BPF_LD_IMM64(R2, 0x0000ffffffff0000LL),
6683 BPF_LD_IMM64(R3, 0xffff00000000ffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006684 BPF_ALU64_IMM(BPF_XOR, R2, 0xffffffff),
6685 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6686 BPF_MOV32_IMM(R0, 2),
6687 BPF_EXIT_INSN(),
6688 BPF_MOV32_IMM(R0, 1),
6689 BPF_EXIT_INSN(),
6690 },
6691 INTERNAL,
6692 { },
6693 { { 0, 0x1 } },
6694 },
6695 {
6696 "ALU64_XOR_K: 0x000000000000000 ^ -1 = 0xffffffffffffffff",
6697 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07006698 BPF_LD_IMM64(R2, 0x0000000000000000LL),
6699 BPF_LD_IMM64(R3, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07006700 BPF_ALU64_IMM(BPF_XOR, R2, 0xffffffff),
6701 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
6702 BPF_MOV32_IMM(R0, 2),
6703 BPF_EXIT_INSN(),
6704 BPF_MOV32_IMM(R0, 1),
6705 BPF_EXIT_INSN(),
6706 },
6707 INTERNAL,
6708 { },
6709 { { 0, 0x1 } },
6710 },
Johan Almbladhba89bcf2021-08-09 11:18:19 +02006711 {
6712 "ALU64_XOR_K: Sign extension 1",
6713 .u.insns_int = {
6714 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6715 BPF_LD_IMM64(R1, 0x0123456786a4c2e0LL),
6716 BPF_ALU64_IMM(BPF_XOR, R0, 0x0f0f0f0f),
6717 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6718 BPF_MOV32_IMM(R0, 2),
6719 BPF_EXIT_INSN(),
6720 BPF_MOV32_IMM(R0, 1),
6721 BPF_EXIT_INSN(),
6722 },
6723 INTERNAL,
6724 { },
6725 { { 0, 1 } }
6726 },
6727 {
6728 "ALU64_XOR_K: Sign extension 2",
6729 .u.insns_int = {
6730 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6731 BPF_LD_IMM64(R1, 0xfedcba98795b3d1fLL),
6732 BPF_ALU64_IMM(BPF_XOR, R0, 0xf0f0f0f0),
6733 BPF_JMP_REG(BPF_JEQ, R0, R1, 2),
6734 BPF_MOV32_IMM(R0, 2),
6735 BPF_EXIT_INSN(),
6736 BPF_MOV32_IMM(R0, 1),
6737 BPF_EXIT_INSN(),
6738 },
6739 INTERNAL,
6740 { },
6741 { { 0, 1 } }
6742 },
Michael Holzheucffc6422015-05-11 22:22:44 -07006743 /* BPF_ALU | BPF_LSH | BPF_X */
6744 {
6745 "ALU_LSH_X: 1 << 1 = 2",
6746 .u.insns_int = {
6747 BPF_LD_IMM64(R0, 1),
6748 BPF_ALU32_IMM(BPF_MOV, R1, 1),
6749 BPF_ALU32_REG(BPF_LSH, R0, R1),
6750 BPF_EXIT_INSN(),
6751 },
6752 INTERNAL,
6753 { },
6754 { { 0, 2 } },
6755 },
6756 {
6757 "ALU_LSH_X: 1 << 31 = 0x80000000",
6758 .u.insns_int = {
6759 BPF_LD_IMM64(R0, 1),
6760 BPF_ALU32_IMM(BPF_MOV, R1, 31),
6761 BPF_ALU32_REG(BPF_LSH, R0, R1),
6762 BPF_EXIT_INSN(),
6763 },
6764 INTERNAL,
6765 { },
6766 { { 0, 0x80000000 } },
6767 },
6768 {
Johan Almbladh0f2fca12021-08-09 11:18:20 +02006769 "ALU_LSH_X: 0x12345678 << 12 = 0x45678000",
6770 .u.insns_int = {
6771 BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678),
6772 BPF_ALU32_IMM(BPF_MOV, R1, 12),
6773 BPF_ALU32_REG(BPF_LSH, R0, R1),
6774 BPF_EXIT_INSN(),
6775 },
6776 INTERNAL,
6777 { },
6778 { { 0, 0x45678000 } }
6779 },
6780 {
Michael Holzheucffc6422015-05-11 22:22:44 -07006781 "ALU64_LSH_X: 1 << 1 = 2",
6782 .u.insns_int = {
6783 BPF_LD_IMM64(R0, 1),
6784 BPF_ALU32_IMM(BPF_MOV, R1, 1),
6785 BPF_ALU64_REG(BPF_LSH, R0, R1),
6786 BPF_EXIT_INSN(),
6787 },
6788 INTERNAL,
6789 { },
6790 { { 0, 2 } },
6791 },
6792 {
6793 "ALU64_LSH_X: 1 << 31 = 0x80000000",
6794 .u.insns_int = {
6795 BPF_LD_IMM64(R0, 1),
6796 BPF_ALU32_IMM(BPF_MOV, R1, 31),
6797 BPF_ALU64_REG(BPF_LSH, R0, R1),
6798 BPF_EXIT_INSN(),
6799 },
6800 INTERNAL,
6801 { },
6802 { { 0, 0x80000000 } },
6803 },
Johan Almbladh3b9890e2021-08-09 11:18:21 +02006804 {
6805 "ALU64_LSH_X: Shift < 32, low word",
6806 .u.insns_int = {
6807 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6808 BPF_ALU32_IMM(BPF_MOV, R1, 12),
6809 BPF_ALU64_REG(BPF_LSH, R0, R1),
6810 BPF_EXIT_INSN(),
6811 },
6812 INTERNAL,
6813 { },
6814 { { 0, 0xbcdef000 } }
6815 },
6816 {
6817 "ALU64_LSH_X: Shift < 32, high word",
6818 .u.insns_int = {
6819 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6820 BPF_ALU32_IMM(BPF_MOV, R1, 12),
6821 BPF_ALU64_REG(BPF_LSH, R0, R1),
6822 BPF_ALU64_IMM(BPF_RSH, R0, 32),
6823 BPF_EXIT_INSN(),
6824 },
6825 INTERNAL,
6826 { },
6827 { { 0, 0x3456789a } }
6828 },
6829 {
6830 "ALU64_LSH_X: Shift > 32, low word",
6831 .u.insns_int = {
6832 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6833 BPF_ALU32_IMM(BPF_MOV, R1, 36),
6834 BPF_ALU64_REG(BPF_LSH, R0, R1),
6835 BPF_EXIT_INSN(),
6836 },
6837 INTERNAL,
6838 { },
6839 { { 0, 0 } }
6840 },
6841 {
6842 "ALU64_LSH_X: Shift > 32, high word",
6843 .u.insns_int = {
6844 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6845 BPF_ALU32_IMM(BPF_MOV, R1, 36),
6846 BPF_ALU64_REG(BPF_LSH, R0, R1),
6847 BPF_ALU64_IMM(BPF_RSH, R0, 32),
6848 BPF_EXIT_INSN(),
6849 },
6850 INTERNAL,
6851 { },
6852 { { 0, 0x9abcdef0 } }
6853 },
6854 {
6855 "ALU64_LSH_X: Shift == 32, low word",
6856 .u.insns_int = {
6857 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6858 BPF_ALU32_IMM(BPF_MOV, R1, 32),
6859 BPF_ALU64_REG(BPF_LSH, R0, R1),
6860 BPF_EXIT_INSN(),
6861 },
6862 INTERNAL,
6863 { },
6864 { { 0, 0 } }
6865 },
6866 {
6867 "ALU64_LSH_X: Shift == 32, high word",
6868 .u.insns_int = {
6869 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6870 BPF_ALU32_IMM(BPF_MOV, R1, 32),
6871 BPF_ALU64_REG(BPF_LSH, R0, R1),
6872 BPF_ALU64_IMM(BPF_RSH, R0, 32),
6873 BPF_EXIT_INSN(),
6874 },
6875 INTERNAL,
6876 { },
6877 { { 0, 0x89abcdef } }
6878 },
6879 {
6880 "ALU64_LSH_X: Zero shift, low word",
6881 .u.insns_int = {
6882 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6883 BPF_ALU32_IMM(BPF_MOV, R1, 0),
6884 BPF_ALU64_REG(BPF_LSH, R0, R1),
6885 BPF_EXIT_INSN(),
6886 },
6887 INTERNAL,
6888 { },
6889 { { 0, 0x89abcdef } }
6890 },
6891 {
6892 "ALU64_LSH_X: Zero shift, high word",
6893 .u.insns_int = {
6894 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6895 BPF_ALU32_IMM(BPF_MOV, R1, 0),
6896 BPF_ALU64_REG(BPF_LSH, R0, R1),
6897 BPF_ALU64_IMM(BPF_RSH, R0, 32),
6898 BPF_EXIT_INSN(),
6899 },
6900 INTERNAL,
6901 { },
6902 { { 0, 0x01234567 } }
6903 },
Michael Holzheucffc6422015-05-11 22:22:44 -07006904 /* BPF_ALU | BPF_LSH | BPF_K */
6905 {
6906 "ALU_LSH_K: 1 << 1 = 2",
6907 .u.insns_int = {
6908 BPF_LD_IMM64(R0, 1),
6909 BPF_ALU32_IMM(BPF_LSH, R0, 1),
6910 BPF_EXIT_INSN(),
6911 },
6912 INTERNAL,
6913 { },
6914 { { 0, 2 } },
6915 },
6916 {
6917 "ALU_LSH_K: 1 << 31 = 0x80000000",
6918 .u.insns_int = {
6919 BPF_LD_IMM64(R0, 1),
6920 BPF_ALU32_IMM(BPF_LSH, R0, 31),
6921 BPF_EXIT_INSN(),
6922 },
6923 INTERNAL,
6924 { },
6925 { { 0, 0x80000000 } },
6926 },
6927 {
Johan Almbladh0f2fca12021-08-09 11:18:20 +02006928 "ALU_LSH_K: 0x12345678 << 12 = 0x45678000",
6929 .u.insns_int = {
6930 BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678),
6931 BPF_ALU32_IMM(BPF_LSH, R0, 12),
6932 BPF_EXIT_INSN(),
6933 },
6934 INTERNAL,
6935 { },
6936 { { 0, 0x45678000 } }
6937 },
6938 {
6939 "ALU_LSH_K: 0x12345678 << 0 = 0x12345678",
6940 .u.insns_int = {
6941 BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678),
6942 BPF_ALU32_IMM(BPF_LSH, R0, 0),
6943 BPF_EXIT_INSN(),
6944 },
6945 INTERNAL,
6946 { },
6947 { { 0, 0x12345678 } }
6948 },
6949 {
Michael Holzheucffc6422015-05-11 22:22:44 -07006950 "ALU64_LSH_K: 1 << 1 = 2",
6951 .u.insns_int = {
6952 BPF_LD_IMM64(R0, 1),
6953 BPF_ALU64_IMM(BPF_LSH, R0, 1),
6954 BPF_EXIT_INSN(),
6955 },
6956 INTERNAL,
6957 { },
6958 { { 0, 2 } },
6959 },
6960 {
6961 "ALU64_LSH_K: 1 << 31 = 0x80000000",
6962 .u.insns_int = {
6963 BPF_LD_IMM64(R0, 1),
6964 BPF_ALU64_IMM(BPF_LSH, R0, 31),
6965 BPF_EXIT_INSN(),
6966 },
6967 INTERNAL,
6968 { },
6969 { { 0, 0x80000000 } },
6970 },
Johan Almbladh3b9890e2021-08-09 11:18:21 +02006971 {
6972 "ALU64_LSH_K: Shift < 32, low word",
6973 .u.insns_int = {
6974 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6975 BPF_ALU64_IMM(BPF_LSH, R0, 12),
6976 BPF_EXIT_INSN(),
6977 },
6978 INTERNAL,
6979 { },
6980 { { 0, 0xbcdef000 } }
6981 },
6982 {
6983 "ALU64_LSH_K: Shift < 32, high word",
6984 .u.insns_int = {
6985 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6986 BPF_ALU64_IMM(BPF_LSH, R0, 12),
6987 BPF_ALU64_IMM(BPF_RSH, R0, 32),
6988 BPF_EXIT_INSN(),
6989 },
6990 INTERNAL,
6991 { },
6992 { { 0, 0x3456789a } }
6993 },
6994 {
6995 "ALU64_LSH_K: Shift > 32, low word",
6996 .u.insns_int = {
6997 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
6998 BPF_ALU64_IMM(BPF_LSH, R0, 36),
6999 BPF_EXIT_INSN(),
7000 },
7001 INTERNAL,
7002 { },
7003 { { 0, 0 } }
7004 },
7005 {
7006 "ALU64_LSH_K: Shift > 32, high word",
7007 .u.insns_int = {
7008 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7009 BPF_ALU64_IMM(BPF_LSH, R0, 36),
7010 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7011 BPF_EXIT_INSN(),
7012 },
7013 INTERNAL,
7014 { },
7015 { { 0, 0x9abcdef0 } }
7016 },
7017 {
7018 "ALU64_LSH_K: Shift == 32, low word",
7019 .u.insns_int = {
7020 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7021 BPF_ALU64_IMM(BPF_LSH, R0, 32),
7022 BPF_EXIT_INSN(),
7023 },
7024 INTERNAL,
7025 { },
7026 { { 0, 0 } }
7027 },
7028 {
7029 "ALU64_LSH_K: Shift == 32, high word",
7030 .u.insns_int = {
7031 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7032 BPF_ALU64_IMM(BPF_LSH, R0, 32),
7033 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7034 BPF_EXIT_INSN(),
7035 },
7036 INTERNAL,
7037 { },
7038 { { 0, 0x89abcdef } }
7039 },
7040 {
7041 "ALU64_LSH_K: Zero shift",
7042 .u.insns_int = {
7043 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7044 BPF_ALU64_IMM(BPF_LSH, R0, 0),
7045 BPF_EXIT_INSN(),
7046 },
7047 INTERNAL,
7048 { },
7049 { { 0, 0x89abcdef } }
7050 },
Michael Holzheucffc6422015-05-11 22:22:44 -07007051 /* BPF_ALU | BPF_RSH | BPF_X */
7052 {
7053 "ALU_RSH_X: 2 >> 1 = 1",
7054 .u.insns_int = {
7055 BPF_LD_IMM64(R0, 2),
7056 BPF_ALU32_IMM(BPF_MOV, R1, 1),
7057 BPF_ALU32_REG(BPF_RSH, R0, R1),
7058 BPF_EXIT_INSN(),
7059 },
7060 INTERNAL,
7061 { },
7062 { { 0, 1 } },
7063 },
7064 {
7065 "ALU_RSH_X: 0x80000000 >> 31 = 1",
7066 .u.insns_int = {
7067 BPF_LD_IMM64(R0, 0x80000000),
7068 BPF_ALU32_IMM(BPF_MOV, R1, 31),
7069 BPF_ALU32_REG(BPF_RSH, R0, R1),
7070 BPF_EXIT_INSN(),
7071 },
7072 INTERNAL,
7073 { },
7074 { { 0, 1 } },
7075 },
7076 {
Johan Almbladh0f2fca12021-08-09 11:18:20 +02007077 "ALU_RSH_X: 0x12345678 >> 20 = 0x123",
7078 .u.insns_int = {
7079 BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678),
7080 BPF_ALU32_IMM(BPF_MOV, R1, 20),
7081 BPF_ALU32_REG(BPF_RSH, R0, R1),
7082 BPF_EXIT_INSN(),
7083 },
7084 INTERNAL,
7085 { },
7086 { { 0, 0x123 } }
7087 },
7088 {
Michael Holzheucffc6422015-05-11 22:22:44 -07007089 "ALU64_RSH_X: 2 >> 1 = 1",
7090 .u.insns_int = {
7091 BPF_LD_IMM64(R0, 2),
7092 BPF_ALU32_IMM(BPF_MOV, R1, 1),
7093 BPF_ALU64_REG(BPF_RSH, R0, R1),
7094 BPF_EXIT_INSN(),
7095 },
7096 INTERNAL,
7097 { },
7098 { { 0, 1 } },
7099 },
7100 {
7101 "ALU64_RSH_X: 0x80000000 >> 31 = 1",
7102 .u.insns_int = {
7103 BPF_LD_IMM64(R0, 0x80000000),
7104 BPF_ALU32_IMM(BPF_MOV, R1, 31),
7105 BPF_ALU64_REG(BPF_RSH, R0, R1),
7106 BPF_EXIT_INSN(),
7107 },
7108 INTERNAL,
7109 { },
7110 { { 0, 1 } },
7111 },
Johan Almbladh3b9890e2021-08-09 11:18:21 +02007112 {
7113 "ALU64_RSH_X: Shift < 32, low word",
7114 .u.insns_int = {
7115 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7116 BPF_ALU32_IMM(BPF_MOV, R1, 12),
7117 BPF_ALU64_REG(BPF_RSH, R0, R1),
7118 BPF_EXIT_INSN(),
7119 },
7120 INTERNAL,
7121 { },
7122 { { 0, 0x56789abc } }
7123 },
7124 {
7125 "ALU64_RSH_X: Shift < 32, high word",
7126 .u.insns_int = {
7127 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7128 BPF_ALU32_IMM(BPF_MOV, R1, 12),
7129 BPF_ALU64_REG(BPF_RSH, R0, R1),
7130 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7131 BPF_EXIT_INSN(),
7132 },
7133 INTERNAL,
7134 { },
7135 { { 0, 0x00081234 } }
7136 },
7137 {
7138 "ALU64_RSH_X: Shift > 32, low word",
7139 .u.insns_int = {
7140 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7141 BPF_ALU32_IMM(BPF_MOV, R1, 36),
7142 BPF_ALU64_REG(BPF_RSH, R0, R1),
7143 BPF_EXIT_INSN(),
7144 },
7145 INTERNAL,
7146 { },
7147 { { 0, 0x08123456 } }
7148 },
7149 {
7150 "ALU64_RSH_X: Shift > 32, high word",
7151 .u.insns_int = {
7152 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7153 BPF_ALU32_IMM(BPF_MOV, R1, 36),
7154 BPF_ALU64_REG(BPF_RSH, R0, R1),
7155 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7156 BPF_EXIT_INSN(),
7157 },
7158 INTERNAL,
7159 { },
7160 { { 0, 0 } }
7161 },
7162 {
7163 "ALU64_RSH_X: Shift == 32, low word",
7164 .u.insns_int = {
7165 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7166 BPF_ALU32_IMM(BPF_MOV, R1, 32),
7167 BPF_ALU64_REG(BPF_RSH, R0, R1),
7168 BPF_EXIT_INSN(),
7169 },
7170 INTERNAL,
7171 { },
7172 { { 0, 0x81234567 } }
7173 },
7174 {
7175 "ALU64_RSH_X: Shift == 32, high word",
7176 .u.insns_int = {
7177 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7178 BPF_ALU32_IMM(BPF_MOV, R1, 32),
7179 BPF_ALU64_REG(BPF_RSH, R0, R1),
7180 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7181 BPF_EXIT_INSN(),
7182 },
7183 INTERNAL,
7184 { },
7185 { { 0, 0 } }
7186 },
7187 {
7188 "ALU64_RSH_X: Zero shift, low word",
7189 .u.insns_int = {
7190 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7191 BPF_ALU32_IMM(BPF_MOV, R1, 0),
7192 BPF_ALU64_REG(BPF_RSH, R0, R1),
7193 BPF_EXIT_INSN(),
7194 },
7195 INTERNAL,
7196 { },
7197 { { 0, 0x89abcdef } }
7198 },
7199 {
7200 "ALU64_RSH_X: Zero shift, high word",
7201 .u.insns_int = {
7202 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7203 BPF_ALU32_IMM(BPF_MOV, R1, 0),
7204 BPF_ALU64_REG(BPF_RSH, R0, R1),
7205 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7206 BPF_EXIT_INSN(),
7207 },
7208 INTERNAL,
7209 { },
7210 { { 0, 0x81234567 } }
7211 },
Michael Holzheucffc6422015-05-11 22:22:44 -07007212 /* BPF_ALU | BPF_RSH | BPF_K */
7213 {
7214 "ALU_RSH_K: 2 >> 1 = 1",
7215 .u.insns_int = {
7216 BPF_LD_IMM64(R0, 2),
7217 BPF_ALU32_IMM(BPF_RSH, R0, 1),
7218 BPF_EXIT_INSN(),
7219 },
7220 INTERNAL,
7221 { },
7222 { { 0, 1 } },
7223 },
7224 {
7225 "ALU_RSH_K: 0x80000000 >> 31 = 1",
7226 .u.insns_int = {
7227 BPF_LD_IMM64(R0, 0x80000000),
7228 BPF_ALU32_IMM(BPF_RSH, R0, 31),
7229 BPF_EXIT_INSN(),
7230 },
7231 INTERNAL,
7232 { },
7233 { { 0, 1 } },
7234 },
7235 {
Johan Almbladh0f2fca12021-08-09 11:18:20 +02007236 "ALU_RSH_K: 0x12345678 >> 20 = 0x123",
7237 .u.insns_int = {
7238 BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678),
7239 BPF_ALU32_IMM(BPF_RSH, R0, 20),
7240 BPF_EXIT_INSN(),
7241 },
7242 INTERNAL,
7243 { },
7244 { { 0, 0x123 } }
7245 },
7246 {
7247 "ALU_RSH_K: 0x12345678 >> 0 = 0x12345678",
7248 .u.insns_int = {
7249 BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678),
7250 BPF_ALU32_IMM(BPF_RSH, R0, 0),
7251 BPF_EXIT_INSN(),
7252 },
7253 INTERNAL,
7254 { },
7255 { { 0, 0x12345678 } }
7256 },
7257 {
Michael Holzheucffc6422015-05-11 22:22:44 -07007258 "ALU64_RSH_K: 2 >> 1 = 1",
7259 .u.insns_int = {
7260 BPF_LD_IMM64(R0, 2),
7261 BPF_ALU64_IMM(BPF_RSH, R0, 1),
7262 BPF_EXIT_INSN(),
7263 },
7264 INTERNAL,
7265 { },
7266 { { 0, 1 } },
7267 },
7268 {
7269 "ALU64_RSH_K: 0x80000000 >> 31 = 1",
7270 .u.insns_int = {
7271 BPF_LD_IMM64(R0, 0x80000000),
7272 BPF_ALU64_IMM(BPF_RSH, R0, 31),
7273 BPF_EXIT_INSN(),
7274 },
7275 INTERNAL,
7276 { },
7277 { { 0, 1 } },
7278 },
Johan Almbladh3b9890e2021-08-09 11:18:21 +02007279 {
7280 "ALU64_RSH_K: Shift < 32, low word",
7281 .u.insns_int = {
7282 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7283 BPF_ALU64_IMM(BPF_RSH, R0, 12),
7284 BPF_EXIT_INSN(),
7285 },
7286 INTERNAL,
7287 { },
7288 { { 0, 0x56789abc } }
7289 },
7290 {
7291 "ALU64_RSH_K: Shift < 32, high word",
7292 .u.insns_int = {
7293 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7294 BPF_ALU64_IMM(BPF_RSH, R0, 12),
7295 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7296 BPF_EXIT_INSN(),
7297 },
7298 INTERNAL,
7299 { },
7300 { { 0, 0x00081234 } }
7301 },
7302 {
7303 "ALU64_RSH_K: Shift > 32, low word",
7304 .u.insns_int = {
7305 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7306 BPF_ALU64_IMM(BPF_RSH, R0, 36),
7307 BPF_EXIT_INSN(),
7308 },
7309 INTERNAL,
7310 { },
7311 { { 0, 0x08123456 } }
7312 },
7313 {
7314 "ALU64_RSH_K: Shift > 32, high word",
7315 .u.insns_int = {
7316 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7317 BPF_ALU64_IMM(BPF_RSH, R0, 36),
7318 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7319 BPF_EXIT_INSN(),
7320 },
7321 INTERNAL,
7322 { },
7323 { { 0, 0 } }
7324 },
7325 {
7326 "ALU64_RSH_K: Shift == 32, low word",
7327 .u.insns_int = {
7328 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7329 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7330 BPF_EXIT_INSN(),
7331 },
7332 INTERNAL,
7333 { },
7334 { { 0, 0x81234567 } }
7335 },
7336 {
7337 "ALU64_RSH_K: Shift == 32, high word",
7338 .u.insns_int = {
7339 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7340 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7341 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7342 BPF_EXIT_INSN(),
7343 },
7344 INTERNAL,
7345 { },
7346 { { 0, 0 } }
7347 },
7348 {
7349 "ALU64_RSH_K: Zero shift",
7350 .u.insns_int = {
7351 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7352 BPF_ALU64_IMM(BPF_RSH, R0, 0),
7353 BPF_EXIT_INSN(),
7354 },
7355 INTERNAL,
7356 { },
7357 { { 0, 0x89abcdef } }
7358 },
Michael Holzheucffc6422015-05-11 22:22:44 -07007359 /* BPF_ALU | BPF_ARSH | BPF_X */
7360 {
Johan Almbladh0f2fca12021-08-09 11:18:20 +02007361 "ALU32_ARSH_X: -1234 >> 7 = -10",
7362 .u.insns_int = {
7363 BPF_ALU32_IMM(BPF_MOV, R0, -1234),
7364 BPF_ALU32_IMM(BPF_MOV, R1, 7),
7365 BPF_ALU32_REG(BPF_ARSH, R0, R1),
7366 BPF_EXIT_INSN(),
7367 },
7368 INTERNAL,
7369 { },
7370 { { 0, -10 } }
7371 },
7372 {
Johan Almbladh3b9890e2021-08-09 11:18:21 +02007373 "ALU64_ARSH_X: 0xff00ff0000000000 >> 40 = 0xffffffffffff00ff",
Michael Holzheucffc6422015-05-11 22:22:44 -07007374 .u.insns_int = {
7375 BPF_LD_IMM64(R0, 0xff00ff0000000000LL),
7376 BPF_ALU32_IMM(BPF_MOV, R1, 40),
7377 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7378 BPF_EXIT_INSN(),
7379 },
7380 INTERNAL,
7381 { },
7382 { { 0, 0xffff00ff } },
7383 },
Johan Almbladh3b9890e2021-08-09 11:18:21 +02007384 {
7385 "ALU64_ARSH_X: Shift < 32, low word",
7386 .u.insns_int = {
7387 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7388 BPF_ALU32_IMM(BPF_MOV, R1, 12),
7389 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7390 BPF_EXIT_INSN(),
7391 },
7392 INTERNAL,
7393 { },
7394 { { 0, 0x56789abc } }
7395 },
7396 {
7397 "ALU64_ARSH_X: Shift < 32, high word",
7398 .u.insns_int = {
7399 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7400 BPF_ALU32_IMM(BPF_MOV, R1, 12),
7401 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7402 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7403 BPF_EXIT_INSN(),
7404 },
7405 INTERNAL,
7406 { },
7407 { { 0, 0xfff81234 } }
7408 },
7409 {
7410 "ALU64_ARSH_X: Shift > 32, low word",
7411 .u.insns_int = {
7412 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7413 BPF_ALU32_IMM(BPF_MOV, R1, 36),
7414 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7415 BPF_EXIT_INSN(),
7416 },
7417 INTERNAL,
7418 { },
7419 { { 0, 0xf8123456 } }
7420 },
7421 {
7422 "ALU64_ARSH_X: Shift > 32, high word",
7423 .u.insns_int = {
7424 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7425 BPF_ALU32_IMM(BPF_MOV, R1, 36),
7426 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7427 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7428 BPF_EXIT_INSN(),
7429 },
7430 INTERNAL,
7431 { },
7432 { { 0, -1 } }
7433 },
7434 {
7435 "ALU64_ARSH_X: Shift == 32, low word",
7436 .u.insns_int = {
7437 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7438 BPF_ALU32_IMM(BPF_MOV, R1, 32),
7439 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7440 BPF_EXIT_INSN(),
7441 },
7442 INTERNAL,
7443 { },
7444 { { 0, 0x81234567 } }
7445 },
7446 {
7447 "ALU64_ARSH_X: Shift == 32, high word",
7448 .u.insns_int = {
7449 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7450 BPF_ALU32_IMM(BPF_MOV, R1, 32),
7451 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7452 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7453 BPF_EXIT_INSN(),
7454 },
7455 INTERNAL,
7456 { },
7457 { { 0, -1 } }
7458 },
7459 {
7460 "ALU64_ARSH_X: Zero shift, low word",
7461 .u.insns_int = {
7462 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7463 BPF_ALU32_IMM(BPF_MOV, R1, 0),
7464 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7465 BPF_EXIT_INSN(),
7466 },
7467 INTERNAL,
7468 { },
7469 { { 0, 0x89abcdef } }
7470 },
7471 {
7472 "ALU64_ARSH_X: Zero shift, high word",
7473 .u.insns_int = {
7474 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7475 BPF_ALU32_IMM(BPF_MOV, R1, 0),
7476 BPF_ALU64_REG(BPF_ARSH, R0, R1),
7477 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7478 BPF_EXIT_INSN(),
7479 },
7480 INTERNAL,
7481 { },
7482 { { 0, 0x81234567 } }
7483 },
Michael Holzheucffc6422015-05-11 22:22:44 -07007484 /* BPF_ALU | BPF_ARSH | BPF_K */
7485 {
Johan Almbladh0f2fca12021-08-09 11:18:20 +02007486 "ALU32_ARSH_K: -1234 >> 7 = -10",
7487 .u.insns_int = {
7488 BPF_ALU32_IMM(BPF_MOV, R0, -1234),
7489 BPF_ALU32_IMM(BPF_ARSH, R0, 7),
7490 BPF_EXIT_INSN(),
7491 },
7492 INTERNAL,
7493 { },
7494 { { 0, -10 } }
7495 },
7496 {
7497 "ALU32_ARSH_K: -1234 >> 0 = -1234",
7498 .u.insns_int = {
7499 BPF_ALU32_IMM(BPF_MOV, R0, -1234),
7500 BPF_ALU32_IMM(BPF_ARSH, R0, 0),
7501 BPF_EXIT_INSN(),
7502 },
7503 INTERNAL,
7504 { },
7505 { { 0, -1234 } }
7506 },
7507 {
Johan Almbladh3b9890e2021-08-09 11:18:21 +02007508 "ALU64_ARSH_K: 0xff00ff0000000000 >> 40 = 0xffffffffffff00ff",
Michael Holzheucffc6422015-05-11 22:22:44 -07007509 .u.insns_int = {
7510 BPF_LD_IMM64(R0, 0xff00ff0000000000LL),
7511 BPF_ALU64_IMM(BPF_ARSH, R0, 40),
7512 BPF_EXIT_INSN(),
7513 },
7514 INTERNAL,
7515 { },
7516 { { 0, 0xffff00ff } },
7517 },
Johan Almbladh3b9890e2021-08-09 11:18:21 +02007518 {
7519 "ALU64_ARSH_K: Shift < 32, low word",
7520 .u.insns_int = {
7521 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7522 BPF_ALU64_IMM(BPF_RSH, R0, 12),
7523 BPF_EXIT_INSN(),
7524 },
7525 INTERNAL,
7526 { },
7527 { { 0, 0x56789abc } }
7528 },
7529 {
7530 "ALU64_ARSH_K: Shift < 32, high word",
7531 .u.insns_int = {
7532 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7533 BPF_ALU64_IMM(BPF_ARSH, R0, 12),
7534 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7535 BPF_EXIT_INSN(),
7536 },
7537 INTERNAL,
7538 { },
7539 { { 0, 0xfff81234 } }
7540 },
7541 {
7542 "ALU64_ARSH_K: Shift > 32, low word",
7543 .u.insns_int = {
7544 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7545 BPF_ALU64_IMM(BPF_ARSH, R0, 36),
7546 BPF_EXIT_INSN(),
7547 },
7548 INTERNAL,
7549 { },
7550 { { 0, 0xf8123456 } }
7551 },
7552 {
7553 "ALU64_ARSH_K: Shift > 32, high word",
7554 .u.insns_int = {
7555 BPF_LD_IMM64(R0, 0xf123456789abcdefLL),
7556 BPF_ALU64_IMM(BPF_ARSH, R0, 36),
7557 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7558 BPF_EXIT_INSN(),
7559 },
7560 INTERNAL,
7561 { },
7562 { { 0, -1 } }
7563 },
7564 {
7565 "ALU64_ARSH_K: Shift == 32, low word",
7566 .u.insns_int = {
7567 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7568 BPF_ALU64_IMM(BPF_ARSH, R0, 32),
7569 BPF_EXIT_INSN(),
7570 },
7571 INTERNAL,
7572 { },
7573 { { 0, 0x81234567 } }
7574 },
7575 {
7576 "ALU64_ARSH_K: Shift == 32, high word",
7577 .u.insns_int = {
7578 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7579 BPF_ALU64_IMM(BPF_ARSH, R0, 32),
7580 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7581 BPF_EXIT_INSN(),
7582 },
7583 INTERNAL,
7584 { },
7585 { { 0, -1 } }
7586 },
7587 {
Colin Ian King1bda52f2021-08-15 22:39:50 +01007588 "ALU64_ARSH_K: Zero shift",
Johan Almbladh3b9890e2021-08-09 11:18:21 +02007589 .u.insns_int = {
7590 BPF_LD_IMM64(R0, 0x8123456789abcdefLL),
7591 BPF_ALU64_IMM(BPF_ARSH, R0, 0),
7592 BPF_EXIT_INSN(),
7593 },
7594 INTERNAL,
7595 { },
7596 { { 0, 0x89abcdef } }
7597 },
Michael Holzheucffc6422015-05-11 22:22:44 -07007598 /* BPF_ALU | BPF_NEG */
7599 {
7600 "ALU_NEG: -(3) = -3",
7601 .u.insns_int = {
7602 BPF_ALU32_IMM(BPF_MOV, R0, 3),
7603 BPF_ALU32_IMM(BPF_NEG, R0, 0),
7604 BPF_EXIT_INSN(),
7605 },
7606 INTERNAL,
7607 { },
7608 { { 0, -3 } },
7609 },
7610 {
7611 "ALU_NEG: -(-3) = 3",
7612 .u.insns_int = {
7613 BPF_ALU32_IMM(BPF_MOV, R0, -3),
7614 BPF_ALU32_IMM(BPF_NEG, R0, 0),
7615 BPF_EXIT_INSN(),
7616 },
7617 INTERNAL,
7618 { },
7619 { { 0, 3 } },
7620 },
7621 {
7622 "ALU64_NEG: -(3) = -3",
7623 .u.insns_int = {
7624 BPF_LD_IMM64(R0, 3),
7625 BPF_ALU64_IMM(BPF_NEG, R0, 0),
7626 BPF_EXIT_INSN(),
7627 },
7628 INTERNAL,
7629 { },
7630 { { 0, -3 } },
7631 },
7632 {
7633 "ALU64_NEG: -(-3) = 3",
7634 .u.insns_int = {
7635 BPF_LD_IMM64(R0, -3),
7636 BPF_ALU64_IMM(BPF_NEG, R0, 0),
7637 BPF_EXIT_INSN(),
7638 },
7639 INTERNAL,
7640 { },
7641 { { 0, 3 } },
7642 },
7643 /* BPF_ALU | BPF_END | BPF_FROM_BE */
7644 {
7645 "ALU_END_FROM_BE 16: 0x0123456789abcdef -> 0xcdef",
7646 .u.insns_int = {
7647 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7648 BPF_ENDIAN(BPF_FROM_BE, R0, 16),
7649 BPF_EXIT_INSN(),
7650 },
7651 INTERNAL,
7652 { },
7653 { { 0, cpu_to_be16(0xcdef) } },
7654 },
7655 {
7656 "ALU_END_FROM_BE 32: 0x0123456789abcdef -> 0x89abcdef",
7657 .u.insns_int = {
7658 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7659 BPF_ENDIAN(BPF_FROM_BE, R0, 32),
Xi Wangba29bec2015-07-08 14:00:56 -07007660 BPF_ALU64_REG(BPF_MOV, R1, R0),
7661 BPF_ALU64_IMM(BPF_RSH, R1, 32),
7662 BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */
Michael Holzheucffc6422015-05-11 22:22:44 -07007663 BPF_EXIT_INSN(),
7664 },
7665 INTERNAL,
7666 { },
7667 { { 0, cpu_to_be32(0x89abcdef) } },
7668 },
7669 {
7670 "ALU_END_FROM_BE 64: 0x0123456789abcdef -> 0x89abcdef",
7671 .u.insns_int = {
7672 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7673 BPF_ENDIAN(BPF_FROM_BE, R0, 64),
7674 BPF_EXIT_INSN(),
7675 },
7676 INTERNAL,
7677 { },
7678 { { 0, (u32) cpu_to_be64(0x0123456789abcdefLL) } },
7679 },
Johan Almbladhf536a7c2021-09-14 11:18:40 +02007680 {
7681 "ALU_END_FROM_BE 64: 0x0123456789abcdef >> 32 -> 0x01234567",
7682 .u.insns_int = {
7683 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7684 BPF_ENDIAN(BPF_FROM_BE, R0, 64),
7685 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7686 BPF_EXIT_INSN(),
7687 },
7688 INTERNAL,
7689 { },
7690 { { 0, (u32) (cpu_to_be64(0x0123456789abcdefLL) >> 32) } },
7691 },
7692 /* BPF_ALU | BPF_END | BPF_FROM_BE, reversed */
7693 {
7694 "ALU_END_FROM_BE 16: 0xfedcba9876543210 -> 0x3210",
7695 .u.insns_int = {
7696 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
7697 BPF_ENDIAN(BPF_FROM_BE, R0, 16),
7698 BPF_EXIT_INSN(),
7699 },
7700 INTERNAL,
7701 { },
7702 { { 0, cpu_to_be16(0x3210) } },
7703 },
7704 {
7705 "ALU_END_FROM_BE 32: 0xfedcba9876543210 -> 0x76543210",
7706 .u.insns_int = {
7707 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
7708 BPF_ENDIAN(BPF_FROM_BE, R0, 32),
7709 BPF_ALU64_REG(BPF_MOV, R1, R0),
7710 BPF_ALU64_IMM(BPF_RSH, R1, 32),
7711 BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */
7712 BPF_EXIT_INSN(),
7713 },
7714 INTERNAL,
7715 { },
7716 { { 0, cpu_to_be32(0x76543210) } },
7717 },
7718 {
7719 "ALU_END_FROM_BE 64: 0xfedcba9876543210 -> 0x76543210",
7720 .u.insns_int = {
7721 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
7722 BPF_ENDIAN(BPF_FROM_BE, R0, 64),
7723 BPF_EXIT_INSN(),
7724 },
7725 INTERNAL,
7726 { },
7727 { { 0, (u32) cpu_to_be64(0xfedcba9876543210ULL) } },
7728 },
7729 {
7730 "ALU_END_FROM_BE 64: 0xfedcba9876543210 >> 32 -> 0xfedcba98",
7731 .u.insns_int = {
7732 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
7733 BPF_ENDIAN(BPF_FROM_BE, R0, 64),
7734 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7735 BPF_EXIT_INSN(),
7736 },
7737 INTERNAL,
7738 { },
7739 { { 0, (u32) (cpu_to_be64(0xfedcba9876543210ULL) >> 32) } },
7740 },
Michael Holzheucffc6422015-05-11 22:22:44 -07007741 /* BPF_ALU | BPF_END | BPF_FROM_LE */
7742 {
7743 "ALU_END_FROM_LE 16: 0x0123456789abcdef -> 0xefcd",
7744 .u.insns_int = {
7745 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7746 BPF_ENDIAN(BPF_FROM_LE, R0, 16),
7747 BPF_EXIT_INSN(),
7748 },
7749 INTERNAL,
7750 { },
7751 { { 0, cpu_to_le16(0xcdef) } },
7752 },
7753 {
7754 "ALU_END_FROM_LE 32: 0x0123456789abcdef -> 0xefcdab89",
7755 .u.insns_int = {
7756 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7757 BPF_ENDIAN(BPF_FROM_LE, R0, 32),
Xi Wangba29bec2015-07-08 14:00:56 -07007758 BPF_ALU64_REG(BPF_MOV, R1, R0),
7759 BPF_ALU64_IMM(BPF_RSH, R1, 32),
7760 BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */
Michael Holzheucffc6422015-05-11 22:22:44 -07007761 BPF_EXIT_INSN(),
7762 },
7763 INTERNAL,
7764 { },
7765 { { 0, cpu_to_le32(0x89abcdef) } },
7766 },
7767 {
7768 "ALU_END_FROM_LE 64: 0x0123456789abcdef -> 0x67452301",
7769 .u.insns_int = {
7770 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7771 BPF_ENDIAN(BPF_FROM_LE, R0, 64),
7772 BPF_EXIT_INSN(),
7773 },
7774 INTERNAL,
7775 { },
7776 { { 0, (u32) cpu_to_le64(0x0123456789abcdefLL) } },
7777 },
Johan Almbladhf536a7c2021-09-14 11:18:40 +02007778 {
7779 "ALU_END_FROM_LE 64: 0x0123456789abcdef >> 32 -> 0xefcdab89",
7780 .u.insns_int = {
7781 BPF_LD_IMM64(R0, 0x0123456789abcdefLL),
7782 BPF_ENDIAN(BPF_FROM_LE, R0, 64),
7783 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7784 BPF_EXIT_INSN(),
7785 },
7786 INTERNAL,
7787 { },
7788 { { 0, (u32) (cpu_to_le64(0x0123456789abcdefLL) >> 32) } },
7789 },
7790 /* BPF_ALU | BPF_END | BPF_FROM_LE, reversed */
7791 {
7792 "ALU_END_FROM_LE 16: 0xfedcba9876543210 -> 0x1032",
7793 .u.insns_int = {
7794 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
7795 BPF_ENDIAN(BPF_FROM_LE, R0, 16),
7796 BPF_EXIT_INSN(),
7797 },
7798 INTERNAL,
7799 { },
7800 { { 0, cpu_to_le16(0x3210) } },
7801 },
7802 {
7803 "ALU_END_FROM_LE 32: 0xfedcba9876543210 -> 0x10325476",
7804 .u.insns_int = {
7805 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
7806 BPF_ENDIAN(BPF_FROM_LE, R0, 32),
7807 BPF_ALU64_REG(BPF_MOV, R1, R0),
7808 BPF_ALU64_IMM(BPF_RSH, R1, 32),
7809 BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */
7810 BPF_EXIT_INSN(),
7811 },
7812 INTERNAL,
7813 { },
7814 { { 0, cpu_to_le32(0x76543210) } },
7815 },
7816 {
7817 "ALU_END_FROM_LE 64: 0xfedcba9876543210 -> 0x10325476",
7818 .u.insns_int = {
7819 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
7820 BPF_ENDIAN(BPF_FROM_LE, R0, 64),
7821 BPF_EXIT_INSN(),
7822 },
7823 INTERNAL,
7824 { },
7825 { { 0, (u32) cpu_to_le64(0xfedcba9876543210ULL) } },
7826 },
7827 {
7828 "ALU_END_FROM_LE 64: 0xfedcba9876543210 >> 32 -> 0x98badcfe",
7829 .u.insns_int = {
7830 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
7831 BPF_ENDIAN(BPF_FROM_LE, R0, 64),
7832 BPF_ALU64_IMM(BPF_RSH, R0, 32),
7833 BPF_EXIT_INSN(),
7834 },
7835 INTERNAL,
7836 { },
7837 { { 0, (u32) (cpu_to_le64(0xfedcba9876543210ULL) >> 32) } },
7838 },
Johan Almbladhcaaaa162021-10-01 15:03:39 +02007839 /* BPF_LDX_MEM B/H/W/DW */
7840 {
7841 "BPF_LDX_MEM | BPF_B",
7842 .u.insns_int = {
7843 BPF_LD_IMM64(R1, 0x0102030405060708ULL),
7844 BPF_LD_IMM64(R2, 0x0000000000000008ULL),
7845 BPF_STX_MEM(BPF_DW, R10, R1, -8),
7846#ifdef __BIG_ENDIAN
7847 BPF_LDX_MEM(BPF_B, R0, R10, -1),
7848#else
7849 BPF_LDX_MEM(BPF_B, R0, R10, -8),
7850#endif
7851 BPF_JMP_REG(BPF_JNE, R0, R2, 1),
7852 BPF_ALU64_IMM(BPF_MOV, R0, 0),
7853 BPF_EXIT_INSN(),
7854 },
7855 INTERNAL,
7856 { },
7857 { { 0, 0 } },
7858 .stack_depth = 8,
7859 },
7860 {
7861 "BPF_LDX_MEM | BPF_B, MSB set",
7862 .u.insns_int = {
7863 BPF_LD_IMM64(R1, 0x8182838485868788ULL),
7864 BPF_LD_IMM64(R2, 0x0000000000000088ULL),
7865 BPF_STX_MEM(BPF_DW, R10, R1, -8),
7866#ifdef __BIG_ENDIAN
7867 BPF_LDX_MEM(BPF_B, R0, R10, -1),
7868#else
7869 BPF_LDX_MEM(BPF_B, R0, R10, -8),
7870#endif
7871 BPF_JMP_REG(BPF_JNE, R0, R2, 1),
7872 BPF_ALU64_IMM(BPF_MOV, R0, 0),
7873 BPF_EXIT_INSN(),
7874 },
7875 INTERNAL,
7876 { },
7877 { { 0, 0 } },
7878 .stack_depth = 8,
7879 },
7880 {
7881 "BPF_LDX_MEM | BPF_H",
7882 .u.insns_int = {
7883 BPF_LD_IMM64(R1, 0x0102030405060708ULL),
7884 BPF_LD_IMM64(R2, 0x0000000000000708ULL),
7885 BPF_STX_MEM(BPF_DW, R10, R1, -8),
7886#ifdef __BIG_ENDIAN
7887 BPF_LDX_MEM(BPF_H, R0, R10, -2),
7888#else
7889 BPF_LDX_MEM(BPF_H, R0, R10, -8),
7890#endif
7891 BPF_JMP_REG(BPF_JNE, R0, R2, 1),
7892 BPF_ALU64_IMM(BPF_MOV, R0, 0),
7893 BPF_EXIT_INSN(),
7894 },
7895 INTERNAL,
7896 { },
7897 { { 0, 0 } },
7898 .stack_depth = 8,
7899 },
7900 {
7901 "BPF_LDX_MEM | BPF_H, MSB set",
7902 .u.insns_int = {
7903 BPF_LD_IMM64(R1, 0x8182838485868788ULL),
7904 BPF_LD_IMM64(R2, 0x0000000000008788ULL),
7905 BPF_STX_MEM(BPF_DW, R10, R1, -8),
7906#ifdef __BIG_ENDIAN
7907 BPF_LDX_MEM(BPF_H, R0, R10, -2),
7908#else
7909 BPF_LDX_MEM(BPF_H, R0, R10, -8),
7910#endif
7911 BPF_JMP_REG(BPF_JNE, R0, R2, 1),
7912 BPF_ALU64_IMM(BPF_MOV, R0, 0),
7913 BPF_EXIT_INSN(),
7914 },
7915 INTERNAL,
7916 { },
7917 { { 0, 0 } },
7918 .stack_depth = 8,
7919 },
7920 {
7921 "BPF_LDX_MEM | BPF_W",
7922 .u.insns_int = {
7923 BPF_LD_IMM64(R1, 0x0102030405060708ULL),
7924 BPF_LD_IMM64(R2, 0x0000000005060708ULL),
7925 BPF_STX_MEM(BPF_DW, R10, R1, -8),
7926#ifdef __BIG_ENDIAN
7927 BPF_LDX_MEM(BPF_W, R0, R10, -4),
7928#else
7929 BPF_LDX_MEM(BPF_W, R0, R10, -8),
7930#endif
7931 BPF_JMP_REG(BPF_JNE, R0, R2, 1),
7932 BPF_ALU64_IMM(BPF_MOV, R0, 0),
7933 BPF_EXIT_INSN(),
7934 },
7935 INTERNAL,
7936 { },
7937 { { 0, 0 } },
7938 .stack_depth = 8,
7939 },
7940 {
7941 "BPF_LDX_MEM | BPF_W, MSB set",
7942 .u.insns_int = {
7943 BPF_LD_IMM64(R1, 0x8182838485868788ULL),
7944 BPF_LD_IMM64(R2, 0x0000000085868788ULL),
7945 BPF_STX_MEM(BPF_DW, R10, R1, -8),
7946#ifdef __BIG_ENDIAN
7947 BPF_LDX_MEM(BPF_W, R0, R10, -4),
7948#else
7949 BPF_LDX_MEM(BPF_W, R0, R10, -8),
7950#endif
7951 BPF_JMP_REG(BPF_JNE, R0, R2, 1),
7952 BPF_ALU64_IMM(BPF_MOV, R0, 0),
7953 BPF_EXIT_INSN(),
7954 },
7955 INTERNAL,
7956 { },
7957 { { 0, 0 } },
7958 .stack_depth = 8,
7959 },
7960 /* BPF_STX_MEM B/H/W/DW */
7961 {
7962 "BPF_STX_MEM | BPF_B",
7963 .u.insns_int = {
7964 BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL),
7965 BPF_LD_IMM64(R2, 0x0102030405060708ULL),
7966 BPF_LD_IMM64(R3, 0x8090a0b0c0d0e008ULL),
7967 BPF_STX_MEM(BPF_DW, R10, R1, -8),
7968#ifdef __BIG_ENDIAN
7969 BPF_STX_MEM(BPF_B, R10, R2, -1),
7970#else
7971 BPF_STX_MEM(BPF_B, R10, R2, -8),
7972#endif
7973 BPF_LDX_MEM(BPF_DW, R0, R10, -8),
7974 BPF_JMP_REG(BPF_JNE, R0, R3, 1),
7975 BPF_ALU64_IMM(BPF_MOV, R0, 0),
7976 BPF_EXIT_INSN(),
7977 },
7978 INTERNAL,
7979 { },
7980 { { 0, 0 } },
7981 .stack_depth = 8,
7982 },
7983 {
7984 "BPF_STX_MEM | BPF_B, MSB set",
7985 .u.insns_int = {
7986 BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL),
7987 BPF_LD_IMM64(R2, 0x8182838485868788ULL),
7988 BPF_LD_IMM64(R3, 0x8090a0b0c0d0e088ULL),
7989 BPF_STX_MEM(BPF_DW, R10, R1, -8),
7990#ifdef __BIG_ENDIAN
7991 BPF_STX_MEM(BPF_B, R10, R2, -1),
7992#else
7993 BPF_STX_MEM(BPF_B, R10, R2, -8),
7994#endif
7995 BPF_LDX_MEM(BPF_DW, R0, R10, -8),
7996 BPF_JMP_REG(BPF_JNE, R0, R3, 1),
7997 BPF_ALU64_IMM(BPF_MOV, R0, 0),
7998 BPF_EXIT_INSN(),
7999 },
8000 INTERNAL,
8001 { },
8002 { { 0, 0 } },
8003 .stack_depth = 8,
8004 },
8005 {
8006 "BPF_STX_MEM | BPF_H",
8007 .u.insns_int = {
8008 BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL),
8009 BPF_LD_IMM64(R2, 0x0102030405060708ULL),
8010 BPF_LD_IMM64(R3, 0x8090a0b0c0d00708ULL),
8011 BPF_STX_MEM(BPF_DW, R10, R1, -8),
8012#ifdef __BIG_ENDIAN
8013 BPF_STX_MEM(BPF_H, R10, R2, -2),
8014#else
8015 BPF_STX_MEM(BPF_H, R10, R2, -8),
8016#endif
8017 BPF_LDX_MEM(BPF_DW, R0, R10, -8),
8018 BPF_JMP_REG(BPF_JNE, R0, R3, 1),
8019 BPF_ALU64_IMM(BPF_MOV, R0, 0),
8020 BPF_EXIT_INSN(),
8021 },
8022 INTERNAL,
8023 { },
8024 { { 0, 0 } },
8025 .stack_depth = 8,
8026 },
8027 {
8028 "BPF_STX_MEM | BPF_H, MSB set",
8029 .u.insns_int = {
8030 BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL),
8031 BPF_LD_IMM64(R2, 0x8182838485868788ULL),
8032 BPF_LD_IMM64(R3, 0x8090a0b0c0d08788ULL),
8033 BPF_STX_MEM(BPF_DW, R10, R1, -8),
8034#ifdef __BIG_ENDIAN
8035 BPF_STX_MEM(BPF_H, R10, R2, -2),
8036#else
8037 BPF_STX_MEM(BPF_H, R10, R2, -8),
8038#endif
8039 BPF_LDX_MEM(BPF_DW, R0, R10, -8),
8040 BPF_JMP_REG(BPF_JNE, R0, R3, 1),
8041 BPF_ALU64_IMM(BPF_MOV, R0, 0),
8042 BPF_EXIT_INSN(),
8043 },
8044 INTERNAL,
8045 { },
8046 { { 0, 0 } },
8047 .stack_depth = 8,
8048 },
8049 {
8050 "BPF_STX_MEM | BPF_W",
8051 .u.insns_int = {
8052 BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL),
8053 BPF_LD_IMM64(R2, 0x0102030405060708ULL),
8054 BPF_LD_IMM64(R3, 0x8090a0b005060708ULL),
8055 BPF_STX_MEM(BPF_DW, R10, R1, -8),
8056#ifdef __BIG_ENDIAN
8057 BPF_STX_MEM(BPF_W, R10, R2, -4),
8058#else
8059 BPF_STX_MEM(BPF_W, R10, R2, -8),
8060#endif
8061 BPF_LDX_MEM(BPF_DW, R0, R10, -8),
8062 BPF_JMP_REG(BPF_JNE, R0, R3, 1),
8063 BPF_ALU64_IMM(BPF_MOV, R0, 0),
8064 BPF_EXIT_INSN(),
8065 },
8066 INTERNAL,
8067 { },
8068 { { 0, 0 } },
8069 .stack_depth = 8,
8070 },
8071 {
8072 "BPF_STX_MEM | BPF_W, MSB set",
8073 .u.insns_int = {
8074 BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL),
8075 BPF_LD_IMM64(R2, 0x8182838485868788ULL),
8076 BPF_LD_IMM64(R3, 0x8090a0b085868788ULL),
8077 BPF_STX_MEM(BPF_DW, R10, R1, -8),
8078#ifdef __BIG_ENDIAN
8079 BPF_STX_MEM(BPF_W, R10, R2, -4),
8080#else
8081 BPF_STX_MEM(BPF_W, R10, R2, -8),
8082#endif
8083 BPF_LDX_MEM(BPF_DW, R0, R10, -8),
8084 BPF_JMP_REG(BPF_JNE, R0, R3, 1),
8085 BPF_ALU64_IMM(BPF_MOV, R0, 0),
8086 BPF_EXIT_INSN(),
8087 },
8088 INTERNAL,
8089 { },
8090 { { 0, 0 } },
8091 .stack_depth = 8,
8092 },
Michael Holzheucffc6422015-05-11 22:22:44 -07008093 /* BPF_ST(X) | BPF_MEM | BPF_B/H/W/DW */
8094 {
8095 "ST_MEM_B: Store/Load byte: max negative",
8096 .u.insns_int = {
8097 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8098 BPF_ST_MEM(BPF_B, R10, -40, 0xff),
8099 BPF_LDX_MEM(BPF_B, R0, R10, -40),
8100 BPF_EXIT_INSN(),
8101 },
8102 INTERNAL,
8103 { },
8104 { { 0, 0xff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008105 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008106 },
8107 {
8108 "ST_MEM_B: Store/Load byte: max positive",
8109 .u.insns_int = {
8110 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8111 BPF_ST_MEM(BPF_H, R10, -40, 0x7f),
8112 BPF_LDX_MEM(BPF_H, R0, R10, -40),
8113 BPF_EXIT_INSN(),
8114 },
8115 INTERNAL,
8116 { },
8117 { { 0, 0x7f } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008118 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008119 },
8120 {
8121 "STX_MEM_B: Store/Load byte: max negative",
8122 .u.insns_int = {
8123 BPF_LD_IMM64(R0, 0),
8124 BPF_LD_IMM64(R1, 0xffLL),
8125 BPF_STX_MEM(BPF_B, R10, R1, -40),
8126 BPF_LDX_MEM(BPF_B, R0, R10, -40),
8127 BPF_EXIT_INSN(),
8128 },
8129 INTERNAL,
8130 { },
8131 { { 0, 0xff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008132 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008133 },
8134 {
8135 "ST_MEM_H: Store/Load half word: max negative",
8136 .u.insns_int = {
8137 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8138 BPF_ST_MEM(BPF_H, R10, -40, 0xffff),
8139 BPF_LDX_MEM(BPF_H, R0, R10, -40),
8140 BPF_EXIT_INSN(),
8141 },
8142 INTERNAL,
8143 { },
8144 { { 0, 0xffff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008145 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008146 },
8147 {
8148 "ST_MEM_H: Store/Load half word: max positive",
8149 .u.insns_int = {
8150 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8151 BPF_ST_MEM(BPF_H, R10, -40, 0x7fff),
8152 BPF_LDX_MEM(BPF_H, R0, R10, -40),
8153 BPF_EXIT_INSN(),
8154 },
8155 INTERNAL,
8156 { },
8157 { { 0, 0x7fff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008158 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008159 },
8160 {
8161 "STX_MEM_H: Store/Load half word: max negative",
8162 .u.insns_int = {
8163 BPF_LD_IMM64(R0, 0),
8164 BPF_LD_IMM64(R1, 0xffffLL),
8165 BPF_STX_MEM(BPF_H, R10, R1, -40),
8166 BPF_LDX_MEM(BPF_H, R0, R10, -40),
8167 BPF_EXIT_INSN(),
8168 },
8169 INTERNAL,
8170 { },
8171 { { 0, 0xffff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008172 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008173 },
8174 {
8175 "ST_MEM_W: Store/Load word: max negative",
8176 .u.insns_int = {
8177 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8178 BPF_ST_MEM(BPF_W, R10, -40, 0xffffffff),
8179 BPF_LDX_MEM(BPF_W, R0, R10, -40),
8180 BPF_EXIT_INSN(),
8181 },
8182 INTERNAL,
8183 { },
8184 { { 0, 0xffffffff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008185 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008186 },
8187 {
8188 "ST_MEM_W: Store/Load word: max positive",
8189 .u.insns_int = {
8190 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8191 BPF_ST_MEM(BPF_W, R10, -40, 0x7fffffff),
8192 BPF_LDX_MEM(BPF_W, R0, R10, -40),
8193 BPF_EXIT_INSN(),
8194 },
8195 INTERNAL,
8196 { },
8197 { { 0, 0x7fffffff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008198 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008199 },
8200 {
8201 "STX_MEM_W: Store/Load word: max negative",
8202 .u.insns_int = {
8203 BPF_LD_IMM64(R0, 0),
8204 BPF_LD_IMM64(R1, 0xffffffffLL),
8205 BPF_STX_MEM(BPF_W, R10, R1, -40),
8206 BPF_LDX_MEM(BPF_W, R0, R10, -40),
8207 BPF_EXIT_INSN(),
8208 },
8209 INTERNAL,
8210 { },
8211 { { 0, 0xffffffff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008212 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008213 },
8214 {
8215 "ST_MEM_DW: Store/Load double word: max negative",
8216 .u.insns_int = {
8217 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8218 BPF_ST_MEM(BPF_DW, R10, -40, 0xffffffff),
8219 BPF_LDX_MEM(BPF_DW, R0, R10, -40),
8220 BPF_EXIT_INSN(),
8221 },
8222 INTERNAL,
8223 { },
8224 { { 0, 0xffffffff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008225 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008226 },
8227 {
8228 "ST_MEM_DW: Store/Load double word: max negative 2",
8229 .u.insns_int = {
Michael Holzheu56cbaa42015-05-13 20:40:39 -07008230 BPF_LD_IMM64(R2, 0xffff00000000ffffLL),
8231 BPF_LD_IMM64(R3, 0xffffffffffffffffLL),
Michael Holzheucffc6422015-05-11 22:22:44 -07008232 BPF_ST_MEM(BPF_DW, R10, -40, 0xffffffff),
8233 BPF_LDX_MEM(BPF_DW, R2, R10, -40),
8234 BPF_JMP_REG(BPF_JEQ, R2, R3, 2),
8235 BPF_MOV32_IMM(R0, 2),
8236 BPF_EXIT_INSN(),
8237 BPF_MOV32_IMM(R0, 1),
8238 BPF_EXIT_INSN(),
8239 },
8240 INTERNAL,
8241 { },
8242 { { 0, 0x1 } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008243 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008244 },
8245 {
8246 "ST_MEM_DW: Store/Load double word: max positive",
8247 .u.insns_int = {
8248 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8249 BPF_ST_MEM(BPF_DW, R10, -40, 0x7fffffff),
8250 BPF_LDX_MEM(BPF_DW, R0, R10, -40),
8251 BPF_EXIT_INSN(),
8252 },
8253 INTERNAL,
8254 { },
8255 { { 0, 0x7fffffff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008256 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008257 },
8258 {
8259 "STX_MEM_DW: Store/Load double word: max negative",
8260 .u.insns_int = {
8261 BPF_LD_IMM64(R0, 0),
8262 BPF_LD_IMM64(R1, 0xffffffffffffffffLL),
Johan Almbladhae7f4702021-07-21 12:40:58 +02008263 BPF_STX_MEM(BPF_DW, R10, R1, -40),
8264 BPF_LDX_MEM(BPF_DW, R0, R10, -40),
Michael Holzheucffc6422015-05-11 22:22:44 -07008265 BPF_EXIT_INSN(),
8266 },
8267 INTERNAL,
8268 { },
8269 { { 0, 0xffffffff } },
Alexei Starovoitov105c0362017-05-30 13:31:32 -07008270 .stack_depth = 40,
Michael Holzheucffc6422015-05-11 22:22:44 -07008271 },
Johan Almbladhe5009b42021-08-09 11:18:24 +02008272 {
8273 "STX_MEM_DW: Store double word: first word in memory",
8274 .u.insns_int = {
8275 BPF_LD_IMM64(R0, 0),
8276 BPF_LD_IMM64(R1, 0x0123456789abcdefLL),
8277 BPF_STX_MEM(BPF_DW, R10, R1, -40),
8278 BPF_LDX_MEM(BPF_W, R0, R10, -40),
8279 BPF_EXIT_INSN(),
8280 },
8281 INTERNAL,
8282 { },
8283#ifdef __BIG_ENDIAN
8284 { { 0, 0x01234567 } },
8285#else
8286 { { 0, 0x89abcdef } },
8287#endif
8288 .stack_depth = 40,
8289 },
8290 {
8291 "STX_MEM_DW: Store double word: second word in memory",
8292 .u.insns_int = {
8293 BPF_LD_IMM64(R0, 0),
8294 BPF_LD_IMM64(R1, 0x0123456789abcdefLL),
8295 BPF_STX_MEM(BPF_DW, R10, R1, -40),
8296 BPF_LDX_MEM(BPF_W, R0, R10, -36),
8297 BPF_EXIT_INSN(),
8298 },
8299 INTERNAL,
8300 { },
8301#ifdef __BIG_ENDIAN
8302 { { 0, 0x89abcdef } },
8303#else
8304 { { 0, 0x01234567 } },
8305#endif
8306 .stack_depth = 40,
8307 },
Brendan Jackman91c960b2021-01-14 18:17:44 +00008308 /* BPF_STX | BPF_ATOMIC | BPF_W/DW */
Michael Holzheucffc6422015-05-11 22:22:44 -07008309 {
Daniel Borkmann85f68fe2017-05-01 02:57:20 +02008310 "STX_XADD_W: X + 1 + 1 + 1 + ...",
8311 { },
8312 INTERNAL,
8313 { },
8314 { { 0, 4134 } },
8315 .fill_helper = bpf_fill_stxw,
8316 },
8317 {
Daniel Borkmann85f68fe2017-05-01 02:57:20 +02008318 "STX_XADD_DW: X + 1 + 1 + 1 + ...",
8319 { },
8320 INTERNAL,
8321 { },
8322 { { 0, 4134 } },
8323 .fill_helper = bpf_fill_stxdw,
8324 },
Johan Almbladhe4517b32021-08-09 11:18:27 +02008325 /*
8326 * Exhaustive tests of atomic operation variants.
8327 * Individual tests are expanded from template macros for all
8328 * combinations of ALU operation, word size and fetching.
8329 */
Johan Almbladh89b63462021-10-01 15:03:40 +02008330#define BPF_ATOMIC_POISON(width) ((width) == BPF_W ? (0xbaadf00dULL << 32) : 0)
8331
Johan Almbladhe4517b32021-08-09 11:18:27 +02008332#define BPF_ATOMIC_OP_TEST1(width, op, logic, old, update, result) \
8333{ \
8334 "BPF_ATOMIC | " #width ", " #op ": Test: " \
8335 #old " " #logic " " #update " = " #result, \
8336 .u.insns_int = { \
Johan Almbladh89b63462021-10-01 15:03:40 +02008337 BPF_LD_IMM64(R5, (update) | BPF_ATOMIC_POISON(width)), \
Johan Almbladhe4517b32021-08-09 11:18:27 +02008338 BPF_ST_MEM(width, R10, -40, old), \
8339 BPF_ATOMIC_OP(width, op, R10, R5, -40), \
8340 BPF_LDX_MEM(width, R0, R10, -40), \
Johan Almbladh89b63462021-10-01 15:03:40 +02008341 BPF_ALU64_REG(BPF_MOV, R1, R0), \
8342 BPF_ALU64_IMM(BPF_RSH, R1, 32), \
8343 BPF_ALU64_REG(BPF_OR, R0, R1), \
Johan Almbladhe4517b32021-08-09 11:18:27 +02008344 BPF_EXIT_INSN(), \
8345 }, \
8346 INTERNAL, \
8347 { }, \
8348 { { 0, result } }, \
8349 .stack_depth = 40, \
8350}
8351#define BPF_ATOMIC_OP_TEST2(width, op, logic, old, update, result) \
8352{ \
8353 "BPF_ATOMIC | " #width ", " #op ": Test side effects, r10: " \
8354 #old " " #logic " " #update " = " #result, \
8355 .u.insns_int = { \
8356 BPF_ALU64_REG(BPF_MOV, R1, R10), \
Johan Almbladh89b63462021-10-01 15:03:40 +02008357 BPF_LD_IMM64(R0, (update) | BPF_ATOMIC_POISON(width)), \
Johan Almbladhe4517b32021-08-09 11:18:27 +02008358 BPF_ST_MEM(BPF_W, R10, -40, old), \
8359 BPF_ATOMIC_OP(width, op, R10, R0, -40), \
8360 BPF_ALU64_REG(BPF_MOV, R0, R10), \
8361 BPF_ALU64_REG(BPF_SUB, R0, R1), \
Johan Almbladh89b63462021-10-01 15:03:40 +02008362 BPF_ALU64_REG(BPF_MOV, R1, R0), \
8363 BPF_ALU64_IMM(BPF_RSH, R1, 32), \
8364 BPF_ALU64_REG(BPF_OR, R0, R1), \
Johan Almbladhe4517b32021-08-09 11:18:27 +02008365 BPF_EXIT_INSN(), \
8366 }, \
8367 INTERNAL, \
8368 { }, \
8369 { { 0, 0 } }, \
8370 .stack_depth = 40, \
8371}
8372#define BPF_ATOMIC_OP_TEST3(width, op, logic, old, update, result) \
8373{ \
8374 "BPF_ATOMIC | " #width ", " #op ": Test side effects, r0: " \
8375 #old " " #logic " " #update " = " #result, \
8376 .u.insns_int = { \
8377 BPF_ALU64_REG(BPF_MOV, R0, R10), \
Johan Almbladh89b63462021-10-01 15:03:40 +02008378 BPF_LD_IMM64(R1, (update) | BPF_ATOMIC_POISON(width)), \
Johan Almbladhe4517b32021-08-09 11:18:27 +02008379 BPF_ST_MEM(width, R10, -40, old), \
8380 BPF_ATOMIC_OP(width, op, R10, R1, -40), \
8381 BPF_ALU64_REG(BPF_SUB, R0, R10), \
Johan Almbladh89b63462021-10-01 15:03:40 +02008382 BPF_ALU64_REG(BPF_MOV, R1, R0), \
8383 BPF_ALU64_IMM(BPF_RSH, R1, 32), \
8384 BPF_ALU64_REG(BPF_OR, R0, R1), \
Johan Almbladhe4517b32021-08-09 11:18:27 +02008385 BPF_EXIT_INSN(), \
8386 }, \
8387 INTERNAL, \
8388 { }, \
8389 { { 0, 0 } }, \
8390 .stack_depth = 40, \
8391}
8392#define BPF_ATOMIC_OP_TEST4(width, op, logic, old, update, result) \
8393{ \
8394 "BPF_ATOMIC | " #width ", " #op ": Test fetch: " \
8395 #old " " #logic " " #update " = " #result, \
8396 .u.insns_int = { \
Johan Almbladh89b63462021-10-01 15:03:40 +02008397 BPF_LD_IMM64(R3, (update) | BPF_ATOMIC_POISON(width)), \
Johan Almbladhe4517b32021-08-09 11:18:27 +02008398 BPF_ST_MEM(width, R10, -40, old), \
8399 BPF_ATOMIC_OP(width, op, R10, R3, -40), \
Johan Almbladh89b63462021-10-01 15:03:40 +02008400 BPF_ALU32_REG(BPF_MOV, R0, R3), \
Johan Almbladhe4517b32021-08-09 11:18:27 +02008401 BPF_EXIT_INSN(), \
8402 }, \
8403 INTERNAL, \
8404 { }, \
8405 { { 0, (op) & BPF_FETCH ? old : update } }, \
8406 .stack_depth = 40, \
8407}
8408 /* BPF_ATOMIC | BPF_W: BPF_ADD */
8409 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_ADD, +, 0x12, 0xab, 0xbd),
8410 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_ADD, +, 0x12, 0xab, 0xbd),
8411 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_ADD, +, 0x12, 0xab, 0xbd),
8412 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_ADD, +, 0x12, 0xab, 0xbd),
8413 /* BPF_ATOMIC | BPF_W: BPF_ADD | BPF_FETCH */
8414 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd),
8415 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd),
8416 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd),
8417 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd),
8418 /* BPF_ATOMIC | BPF_DW: BPF_ADD */
8419 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_ADD, +, 0x12, 0xab, 0xbd),
8420 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_ADD, +, 0x12, 0xab, 0xbd),
8421 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_ADD, +, 0x12, 0xab, 0xbd),
8422 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_ADD, +, 0x12, 0xab, 0xbd),
8423 /* BPF_ATOMIC | BPF_DW: BPF_ADD | BPF_FETCH */
8424 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd),
8425 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd),
8426 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd),
8427 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd),
8428 /* BPF_ATOMIC | BPF_W: BPF_AND */
8429 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_AND, &, 0x12, 0xab, 0x02),
8430 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_AND, &, 0x12, 0xab, 0x02),
8431 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_AND, &, 0x12, 0xab, 0x02),
8432 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_AND, &, 0x12, 0xab, 0x02),
8433 /* BPF_ATOMIC | BPF_W: BPF_AND | BPF_FETCH */
8434 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02),
8435 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02),
8436 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02),
8437 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02),
8438 /* BPF_ATOMIC | BPF_DW: BPF_AND */
8439 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_AND, &, 0x12, 0xab, 0x02),
8440 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_AND, &, 0x12, 0xab, 0x02),
8441 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_AND, &, 0x12, 0xab, 0x02),
8442 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_AND, &, 0x12, 0xab, 0x02),
8443 /* BPF_ATOMIC | BPF_DW: BPF_AND | BPF_FETCH */
8444 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02),
8445 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02),
8446 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02),
8447 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02),
8448 /* BPF_ATOMIC | BPF_W: BPF_OR */
8449 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_OR, |, 0x12, 0xab, 0xbb),
8450 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_OR, |, 0x12, 0xab, 0xbb),
8451 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_OR, |, 0x12, 0xab, 0xbb),
8452 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_OR, |, 0x12, 0xab, 0xbb),
8453 /* BPF_ATOMIC | BPF_W: BPF_OR | BPF_FETCH */
8454 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb),
8455 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb),
8456 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb),
8457 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb),
8458 /* BPF_ATOMIC | BPF_DW: BPF_OR */
8459 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_OR, |, 0x12, 0xab, 0xbb),
8460 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_OR, |, 0x12, 0xab, 0xbb),
8461 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_OR, |, 0x12, 0xab, 0xbb),
8462 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_OR, |, 0x12, 0xab, 0xbb),
8463 /* BPF_ATOMIC | BPF_DW: BPF_OR | BPF_FETCH */
8464 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb),
8465 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb),
8466 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb),
8467 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb),
8468 /* BPF_ATOMIC | BPF_W: BPF_XOR */
8469 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_XOR, ^, 0x12, 0xab, 0xb9),
8470 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_XOR, ^, 0x12, 0xab, 0xb9),
8471 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_XOR, ^, 0x12, 0xab, 0xb9),
8472 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_XOR, ^, 0x12, 0xab, 0xb9),
8473 /* BPF_ATOMIC | BPF_W: BPF_XOR | BPF_FETCH */
8474 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9),
8475 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9),
8476 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9),
8477 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9),
8478 /* BPF_ATOMIC | BPF_DW: BPF_XOR */
8479 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_XOR, ^, 0x12, 0xab, 0xb9),
8480 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_XOR, ^, 0x12, 0xab, 0xb9),
8481 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_XOR, ^, 0x12, 0xab, 0xb9),
8482 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_XOR, ^, 0x12, 0xab, 0xb9),
8483 /* BPF_ATOMIC | BPF_DW: BPF_XOR | BPF_FETCH */
8484 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9),
8485 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9),
8486 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9),
8487 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9),
8488 /* BPF_ATOMIC | BPF_W: BPF_XCHG */
8489 BPF_ATOMIC_OP_TEST1(BPF_W, BPF_XCHG, xchg, 0x12, 0xab, 0xab),
8490 BPF_ATOMIC_OP_TEST2(BPF_W, BPF_XCHG, xchg, 0x12, 0xab, 0xab),
8491 BPF_ATOMIC_OP_TEST3(BPF_W, BPF_XCHG, xchg, 0x12, 0xab, 0xab),
8492 BPF_ATOMIC_OP_TEST4(BPF_W, BPF_XCHG, xchg, 0x12, 0xab, 0xab),
8493 /* BPF_ATOMIC | BPF_DW: BPF_XCHG */
8494 BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_XCHG, xchg, 0x12, 0xab, 0xab),
8495 BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_XCHG, xchg, 0x12, 0xab, 0xab),
8496 BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_XCHG, xchg, 0x12, 0xab, 0xab),
8497 BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_XCHG, xchg, 0x12, 0xab, 0xab),
Johan Almbladh89b63462021-10-01 15:03:40 +02008498#undef BPF_ATOMIC_POISON
Johan Almbladhe4517b32021-08-09 11:18:27 +02008499#undef BPF_ATOMIC_OP_TEST1
8500#undef BPF_ATOMIC_OP_TEST2
8501#undef BPF_ATOMIC_OP_TEST3
8502#undef BPF_ATOMIC_OP_TEST4
Johan Almbladh6a3b24c2021-08-09 11:18:28 +02008503 /* BPF_ATOMIC | BPF_W, BPF_CMPXCHG */
8504 {
8505 "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test successful return",
8506 .u.insns_int = {
8507 BPF_ST_MEM(BPF_W, R10, -40, 0x01234567),
8508 BPF_ALU32_IMM(BPF_MOV, R0, 0x01234567),
8509 BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef),
8510 BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40),
8511 BPF_EXIT_INSN(),
8512 },
8513 INTERNAL,
8514 { },
8515 { { 0, 0x01234567 } },
8516 .stack_depth = 40,
8517 },
8518 {
8519 "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test successful store",
8520 .u.insns_int = {
8521 BPF_ST_MEM(BPF_W, R10, -40, 0x01234567),
8522 BPF_ALU32_IMM(BPF_MOV, R0, 0x01234567),
8523 BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef),
8524 BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40),
8525 BPF_LDX_MEM(BPF_W, R0, R10, -40),
8526 BPF_EXIT_INSN(),
8527 },
8528 INTERNAL,
8529 { },
8530 { { 0, 0x89abcdef } },
8531 .stack_depth = 40,
8532 },
8533 {
8534 "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test failure return",
8535 .u.insns_int = {
8536 BPF_ST_MEM(BPF_W, R10, -40, 0x01234567),
8537 BPF_ALU32_IMM(BPF_MOV, R0, 0x76543210),
8538 BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef),
8539 BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40),
8540 BPF_EXIT_INSN(),
8541 },
8542 INTERNAL,
8543 { },
8544 { { 0, 0x01234567 } },
8545 .stack_depth = 40,
8546 },
8547 {
8548 "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test failure store",
8549 .u.insns_int = {
8550 BPF_ST_MEM(BPF_W, R10, -40, 0x01234567),
8551 BPF_ALU32_IMM(BPF_MOV, R0, 0x76543210),
8552 BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef),
8553 BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40),
8554 BPF_LDX_MEM(BPF_W, R0, R10, -40),
8555 BPF_EXIT_INSN(),
8556 },
8557 INTERNAL,
8558 { },
8559 { { 0, 0x01234567 } },
8560 .stack_depth = 40,
8561 },
8562 {
8563 "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test side effects",
8564 .u.insns_int = {
8565 BPF_ST_MEM(BPF_W, R10, -40, 0x01234567),
8566 BPF_ALU32_IMM(BPF_MOV, R0, 0x01234567),
8567 BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef),
8568 BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40),
8569 BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40),
8570 BPF_ALU32_REG(BPF_MOV, R0, R3),
8571 BPF_EXIT_INSN(),
8572 },
8573 INTERNAL,
8574 { },
8575 { { 0, 0x89abcdef } },
8576 .stack_depth = 40,
8577 },
8578 /* BPF_ATOMIC | BPF_DW, BPF_CMPXCHG */
8579 {
8580 "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test successful return",
8581 .u.insns_int = {
8582 BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
Johan Almbladhe42fc3c2021-10-01 15:03:44 +02008583 BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
Johan Almbladh6a3b24c2021-08-09 11:18:28 +02008584 BPF_ALU64_REG(BPF_MOV, R0, R1),
8585 BPF_STX_MEM(BPF_DW, R10, R1, -40),
8586 BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40),
8587 BPF_JMP_REG(BPF_JNE, R0, R1, 1),
8588 BPF_ALU64_REG(BPF_SUB, R0, R1),
8589 BPF_EXIT_INSN(),
8590 },
8591 INTERNAL,
8592 { },
8593 { { 0, 0 } },
8594 .stack_depth = 40,
8595 },
8596 {
8597 "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test successful store",
8598 .u.insns_int = {
8599 BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
Johan Almbladhe42fc3c2021-10-01 15:03:44 +02008600 BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
Johan Almbladh6a3b24c2021-08-09 11:18:28 +02008601 BPF_ALU64_REG(BPF_MOV, R0, R1),
8602 BPF_STX_MEM(BPF_DW, R10, R0, -40),
8603 BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40),
8604 BPF_LDX_MEM(BPF_DW, R0, R10, -40),
8605 BPF_JMP_REG(BPF_JNE, R0, R2, 1),
8606 BPF_ALU64_REG(BPF_SUB, R0, R2),
8607 BPF_EXIT_INSN(),
8608 },
8609 INTERNAL,
8610 { },
8611 { { 0, 0 } },
8612 .stack_depth = 40,
8613 },
8614 {
8615 "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test failure return",
8616 .u.insns_int = {
8617 BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
Johan Almbladhe42fc3c2021-10-01 15:03:44 +02008618 BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
Johan Almbladh6a3b24c2021-08-09 11:18:28 +02008619 BPF_ALU64_REG(BPF_MOV, R0, R1),
8620 BPF_ALU64_IMM(BPF_ADD, R0, 1),
8621 BPF_STX_MEM(BPF_DW, R10, R1, -40),
8622 BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40),
8623 BPF_JMP_REG(BPF_JNE, R0, R1, 1),
8624 BPF_ALU64_REG(BPF_SUB, R0, R1),
8625 BPF_EXIT_INSN(),
8626 },
8627 INTERNAL,
8628 { },
8629 { { 0, 0 } },
8630 .stack_depth = 40,
8631 },
8632 {
8633 "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test failure store",
8634 .u.insns_int = {
8635 BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
Johan Almbladhe42fc3c2021-10-01 15:03:44 +02008636 BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
Johan Almbladh6a3b24c2021-08-09 11:18:28 +02008637 BPF_ALU64_REG(BPF_MOV, R0, R1),
8638 BPF_ALU64_IMM(BPF_ADD, R0, 1),
8639 BPF_STX_MEM(BPF_DW, R10, R1, -40),
8640 BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40),
8641 BPF_LDX_MEM(BPF_DW, R0, R10, -40),
8642 BPF_JMP_REG(BPF_JNE, R0, R1, 1),
8643 BPF_ALU64_REG(BPF_SUB, R0, R1),
8644 BPF_EXIT_INSN(),
8645 },
8646 INTERNAL,
8647 { },
8648 { { 0, 0 } },
8649 .stack_depth = 40,
8650 },
8651 {
8652 "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test side effects",
8653 .u.insns_int = {
8654 BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
Johan Almbladhe42fc3c2021-10-01 15:03:44 +02008655 BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
Johan Almbladh6a3b24c2021-08-09 11:18:28 +02008656 BPF_ALU64_REG(BPF_MOV, R0, R1),
8657 BPF_STX_MEM(BPF_DW, R10, R1, -40),
8658 BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40),
Johan Almbladhe42fc3c2021-10-01 15:03:44 +02008659 BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
Johan Almbladh6a3b24c2021-08-09 11:18:28 +02008660 BPF_JMP_REG(BPF_JNE, R0, R2, 1),
8661 BPF_ALU64_REG(BPF_SUB, R0, R2),
8662 BPF_EXIT_INSN(),
8663 },
8664 INTERNAL,
8665 { },
8666 { { 0, 0 } },
8667 .stack_depth = 40,
8668 },
Johan Almbladhb55dfa82021-08-09 11:18:16 +02008669 /* BPF_JMP32 | BPF_JEQ | BPF_K */
8670 {
8671 "JMP32_JEQ_K: Small immediate",
8672 .u.insns_int = {
8673 BPF_ALU32_IMM(BPF_MOV, R0, 123),
8674 BPF_JMP32_IMM(BPF_JEQ, R0, 321, 1),
8675 BPF_JMP32_IMM(BPF_JEQ, R0, 123, 1),
8676 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8677 BPF_EXIT_INSN(),
8678 },
8679 INTERNAL,
8680 { },
8681 { { 0, 123 } }
8682 },
8683 {
8684 "JMP32_JEQ_K: Large immediate",
8685 .u.insns_int = {
8686 BPF_ALU32_IMM(BPF_MOV, R0, 12345678),
8687 BPF_JMP32_IMM(BPF_JEQ, R0, 12345678 & 0xffff, 1),
8688 BPF_JMP32_IMM(BPF_JEQ, R0, 12345678, 1),
8689 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8690 BPF_EXIT_INSN(),
8691 },
8692 INTERNAL,
8693 { },
8694 { { 0, 12345678 } }
8695 },
8696 {
8697 "JMP32_JEQ_K: negative immediate",
8698 .u.insns_int = {
8699 BPF_ALU32_IMM(BPF_MOV, R0, -123),
8700 BPF_JMP32_IMM(BPF_JEQ, R0, 123, 1),
8701 BPF_JMP32_IMM(BPF_JEQ, R0, -123, 1),
8702 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8703 BPF_EXIT_INSN(),
8704 },
8705 INTERNAL,
8706 { },
8707 { { 0, -123 } }
8708 },
8709 /* BPF_JMP32 | BPF_JEQ | BPF_X */
8710 {
8711 "JMP32_JEQ_X",
8712 .u.insns_int = {
8713 BPF_ALU32_IMM(BPF_MOV, R0, 1234),
8714 BPF_ALU32_IMM(BPF_MOV, R1, 4321),
8715 BPF_JMP32_REG(BPF_JEQ, R0, R1, 2),
8716 BPF_ALU32_IMM(BPF_MOV, R1, 1234),
8717 BPF_JMP32_REG(BPF_JEQ, R0, R1, 1),
8718 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8719 BPF_EXIT_INSN(),
8720 },
8721 INTERNAL,
8722 { },
8723 { { 0, 1234 } }
8724 },
8725 /* BPF_JMP32 | BPF_JNE | BPF_K */
8726 {
8727 "JMP32_JNE_K: Small immediate",
8728 .u.insns_int = {
8729 BPF_ALU32_IMM(BPF_MOV, R0, 123),
8730 BPF_JMP32_IMM(BPF_JNE, R0, 123, 1),
8731 BPF_JMP32_IMM(BPF_JNE, R0, 321, 1),
8732 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8733 BPF_EXIT_INSN(),
8734 },
8735 INTERNAL,
8736 { },
8737 { { 0, 123 } }
8738 },
8739 {
8740 "JMP32_JNE_K: Large immediate",
8741 .u.insns_int = {
8742 BPF_ALU32_IMM(BPF_MOV, R0, 12345678),
8743 BPF_JMP32_IMM(BPF_JNE, R0, 12345678, 1),
8744 BPF_JMP32_IMM(BPF_JNE, R0, 12345678 & 0xffff, 1),
8745 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8746 BPF_EXIT_INSN(),
8747 },
8748 INTERNAL,
8749 { },
8750 { { 0, 12345678 } }
8751 },
8752 {
8753 "JMP32_JNE_K: negative immediate",
8754 .u.insns_int = {
8755 BPF_ALU32_IMM(BPF_MOV, R0, -123),
8756 BPF_JMP32_IMM(BPF_JNE, R0, -123, 1),
8757 BPF_JMP32_IMM(BPF_JNE, R0, 123, 1),
8758 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8759 BPF_EXIT_INSN(),
8760 },
8761 INTERNAL,
8762 { },
8763 { { 0, -123 } }
8764 },
8765 /* BPF_JMP32 | BPF_JNE | BPF_X */
8766 {
8767 "JMP32_JNE_X",
8768 .u.insns_int = {
8769 BPF_ALU32_IMM(BPF_MOV, R0, 1234),
8770 BPF_ALU32_IMM(BPF_MOV, R1, 1234),
8771 BPF_JMP32_REG(BPF_JNE, R0, R1, 2),
8772 BPF_ALU32_IMM(BPF_MOV, R1, 4321),
8773 BPF_JMP32_REG(BPF_JNE, R0, R1, 1),
8774 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8775 BPF_EXIT_INSN(),
8776 },
8777 INTERNAL,
8778 { },
8779 { { 0, 1234 } }
8780 },
8781 /* BPF_JMP32 | BPF_JSET | BPF_K */
8782 {
8783 "JMP32_JSET_K: Small immediate",
8784 .u.insns_int = {
8785 BPF_ALU32_IMM(BPF_MOV, R0, 1),
8786 BPF_JMP32_IMM(BPF_JSET, R0, 2, 1),
8787 BPF_JMP32_IMM(BPF_JSET, R0, 3, 1),
8788 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8789 BPF_EXIT_INSN(),
8790 },
8791 INTERNAL,
8792 { },
8793 { { 0, 1 } }
8794 },
8795 {
8796 "JMP32_JSET_K: Large immediate",
8797 .u.insns_int = {
8798 BPF_ALU32_IMM(BPF_MOV, R0, 0x40000000),
8799 BPF_JMP32_IMM(BPF_JSET, R0, 0x3fffffff, 1),
8800 BPF_JMP32_IMM(BPF_JSET, R0, 0x60000000, 1),
8801 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8802 BPF_EXIT_INSN(),
8803 },
8804 INTERNAL,
8805 { },
8806 { { 0, 0x40000000 } }
8807 },
8808 {
8809 "JMP32_JSET_K: negative immediate",
8810 .u.insns_int = {
8811 BPF_ALU32_IMM(BPF_MOV, R0, -123),
8812 BPF_JMP32_IMM(BPF_JSET, R0, -1, 1),
8813 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8814 BPF_EXIT_INSN(),
8815 },
8816 INTERNAL,
8817 { },
8818 { { 0, -123 } }
8819 },
8820 /* BPF_JMP32 | BPF_JSET | BPF_X */
8821 {
8822 "JMP32_JSET_X",
8823 .u.insns_int = {
8824 BPF_ALU32_IMM(BPF_MOV, R0, 8),
8825 BPF_ALU32_IMM(BPF_MOV, R1, 7),
8826 BPF_JMP32_REG(BPF_JSET, R0, R1, 2),
8827 BPF_ALU32_IMM(BPF_MOV, R1, 8 | 2),
8828 BPF_JMP32_REG(BPF_JNE, R0, R1, 1),
8829 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8830 BPF_EXIT_INSN(),
8831 },
8832 INTERNAL,
8833 { },
8834 { { 0, 8 } }
8835 },
8836 /* BPF_JMP32 | BPF_JGT | BPF_K */
8837 {
8838 "JMP32_JGT_K: Small immediate",
8839 .u.insns_int = {
8840 BPF_ALU32_IMM(BPF_MOV, R0, 123),
8841 BPF_JMP32_IMM(BPF_JGT, R0, 123, 1),
8842 BPF_JMP32_IMM(BPF_JGT, R0, 122, 1),
8843 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8844 BPF_EXIT_INSN(),
8845 },
8846 INTERNAL,
8847 { },
8848 { { 0, 123 } }
8849 },
8850 {
8851 "JMP32_JGT_K: Large immediate",
8852 .u.insns_int = {
8853 BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe),
8854 BPF_JMP32_IMM(BPF_JGT, R0, 0xffffffff, 1),
8855 BPF_JMP32_IMM(BPF_JGT, R0, 0xfffffffd, 1),
8856 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8857 BPF_EXIT_INSN(),
8858 },
8859 INTERNAL,
8860 { },
8861 { { 0, 0xfffffffe } }
8862 },
8863 /* BPF_JMP32 | BPF_JGT | BPF_X */
8864 {
8865 "JMP32_JGT_X",
8866 .u.insns_int = {
8867 BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe),
8868 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
8869 BPF_JMP32_REG(BPF_JGT, R0, R1, 2),
8870 BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffd),
8871 BPF_JMP32_REG(BPF_JGT, R0, R1, 1),
8872 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8873 BPF_EXIT_INSN(),
8874 },
8875 INTERNAL,
8876 { },
8877 { { 0, 0xfffffffe } }
8878 },
8879 /* BPF_JMP32 | BPF_JGE | BPF_K */
8880 {
8881 "JMP32_JGE_K: Small immediate",
8882 .u.insns_int = {
8883 BPF_ALU32_IMM(BPF_MOV, R0, 123),
8884 BPF_JMP32_IMM(BPF_JGE, R0, 124, 1),
8885 BPF_JMP32_IMM(BPF_JGE, R0, 123, 1),
8886 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8887 BPF_EXIT_INSN(),
8888 },
8889 INTERNAL,
8890 { },
8891 { { 0, 123 } }
8892 },
8893 {
8894 "JMP32_JGE_K: Large immediate",
8895 .u.insns_int = {
8896 BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe),
8897 BPF_JMP32_IMM(BPF_JGE, R0, 0xffffffff, 1),
8898 BPF_JMP32_IMM(BPF_JGE, R0, 0xfffffffe, 1),
8899 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8900 BPF_EXIT_INSN(),
8901 },
8902 INTERNAL,
8903 { },
8904 { { 0, 0xfffffffe } }
8905 },
8906 /* BPF_JMP32 | BPF_JGE | BPF_X */
8907 {
8908 "JMP32_JGE_X",
8909 .u.insns_int = {
8910 BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe),
8911 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
8912 BPF_JMP32_REG(BPF_JGE, R0, R1, 2),
8913 BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffe),
8914 BPF_JMP32_REG(BPF_JGE, R0, R1, 1),
8915 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8916 BPF_EXIT_INSN(),
8917 },
8918 INTERNAL,
8919 { },
8920 { { 0, 0xfffffffe } }
8921 },
8922 /* BPF_JMP32 | BPF_JLT | BPF_K */
8923 {
8924 "JMP32_JLT_K: Small immediate",
8925 .u.insns_int = {
8926 BPF_ALU32_IMM(BPF_MOV, R0, 123),
8927 BPF_JMP32_IMM(BPF_JLT, R0, 123, 1),
8928 BPF_JMP32_IMM(BPF_JLT, R0, 124, 1),
8929 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8930 BPF_EXIT_INSN(),
8931 },
8932 INTERNAL,
8933 { },
8934 { { 0, 123 } }
8935 },
8936 {
8937 "JMP32_JLT_K: Large immediate",
8938 .u.insns_int = {
8939 BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe),
8940 BPF_JMP32_IMM(BPF_JLT, R0, 0xfffffffd, 1),
8941 BPF_JMP32_IMM(BPF_JLT, R0, 0xffffffff, 1),
8942 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8943 BPF_EXIT_INSN(),
8944 },
8945 INTERNAL,
8946 { },
8947 { { 0, 0xfffffffe } }
8948 },
8949 /* BPF_JMP32 | BPF_JLT | BPF_X */
8950 {
8951 "JMP32_JLT_X",
8952 .u.insns_int = {
8953 BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe),
8954 BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffd),
8955 BPF_JMP32_REG(BPF_JLT, R0, R1, 2),
8956 BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff),
8957 BPF_JMP32_REG(BPF_JLT, R0, R1, 1),
8958 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8959 BPF_EXIT_INSN(),
8960 },
8961 INTERNAL,
8962 { },
8963 { { 0, 0xfffffffe } }
8964 },
8965 /* BPF_JMP32 | BPF_JLE | BPF_K */
8966 {
8967 "JMP32_JLE_K: Small immediate",
8968 .u.insns_int = {
8969 BPF_ALU32_IMM(BPF_MOV, R0, 123),
8970 BPF_JMP32_IMM(BPF_JLE, R0, 122, 1),
8971 BPF_JMP32_IMM(BPF_JLE, R0, 123, 1),
8972 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8973 BPF_EXIT_INSN(),
8974 },
8975 INTERNAL,
8976 { },
8977 { { 0, 123 } }
8978 },
8979 {
8980 "JMP32_JLE_K: Large immediate",
8981 .u.insns_int = {
8982 BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe),
8983 BPF_JMP32_IMM(BPF_JLE, R0, 0xfffffffd, 1),
8984 BPF_JMP32_IMM(BPF_JLE, R0, 0xfffffffe, 1),
8985 BPF_ALU32_IMM(BPF_MOV, R0, 0),
8986 BPF_EXIT_INSN(),
8987 },
8988 INTERNAL,
8989 { },
8990 { { 0, 0xfffffffe } }
8991 },
8992 /* BPF_JMP32 | BPF_JLE | BPF_X */
8993 {
8994 "JMP32_JLE_X",
8995 .u.insns_int = {
8996 BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe),
8997 BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffd),
8998 BPF_JMP32_REG(BPF_JLE, R0, R1, 2),
8999 BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffe),
9000 BPF_JMP32_REG(BPF_JLE, R0, R1, 1),
9001 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9002 BPF_EXIT_INSN(),
9003 },
9004 INTERNAL,
9005 { },
9006 { { 0, 0xfffffffe } }
9007 },
9008 /* BPF_JMP32 | BPF_JSGT | BPF_K */
9009 {
9010 "JMP32_JSGT_K: Small immediate",
9011 .u.insns_int = {
9012 BPF_ALU32_IMM(BPF_MOV, R0, -123),
9013 BPF_JMP32_IMM(BPF_JSGT, R0, -123, 1),
9014 BPF_JMP32_IMM(BPF_JSGT, R0, -124, 1),
9015 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9016 BPF_EXIT_INSN(),
9017 },
9018 INTERNAL,
9019 { },
9020 { { 0, -123 } }
9021 },
9022 {
9023 "JMP32_JSGT_K: Large immediate",
9024 .u.insns_int = {
9025 BPF_ALU32_IMM(BPF_MOV, R0, -12345678),
9026 BPF_JMP32_IMM(BPF_JSGT, R0, -12345678, 1),
9027 BPF_JMP32_IMM(BPF_JSGT, R0, -12345679, 1),
9028 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9029 BPF_EXIT_INSN(),
9030 },
9031 INTERNAL,
9032 { },
9033 { { 0, -12345678 } }
9034 },
9035 /* BPF_JMP32 | BPF_JSGT | BPF_X */
9036 {
9037 "JMP32_JSGT_X",
9038 .u.insns_int = {
9039 BPF_ALU32_IMM(BPF_MOV, R0, -12345678),
9040 BPF_ALU32_IMM(BPF_MOV, R1, -12345678),
9041 BPF_JMP32_REG(BPF_JSGT, R0, R1, 2),
9042 BPF_ALU32_IMM(BPF_MOV, R1, -12345679),
9043 BPF_JMP32_REG(BPF_JSGT, R0, R1, 1),
9044 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9045 BPF_EXIT_INSN(),
9046 },
9047 INTERNAL,
9048 { },
9049 { { 0, -12345678 } }
9050 },
9051 /* BPF_JMP32 | BPF_JSGE | BPF_K */
9052 {
9053 "JMP32_JSGE_K: Small immediate",
9054 .u.insns_int = {
9055 BPF_ALU32_IMM(BPF_MOV, R0, -123),
9056 BPF_JMP32_IMM(BPF_JSGE, R0, -122, 1),
9057 BPF_JMP32_IMM(BPF_JSGE, R0, -123, 1),
9058 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9059 BPF_EXIT_INSN(),
9060 },
9061 INTERNAL,
9062 { },
9063 { { 0, -123 } }
9064 },
9065 {
9066 "JMP32_JSGE_K: Large immediate",
9067 .u.insns_int = {
9068 BPF_ALU32_IMM(BPF_MOV, R0, -12345678),
9069 BPF_JMP32_IMM(BPF_JSGE, R0, -12345677, 1),
9070 BPF_JMP32_IMM(BPF_JSGE, R0, -12345678, 1),
9071 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9072 BPF_EXIT_INSN(),
9073 },
9074 INTERNAL,
9075 { },
9076 { { 0, -12345678 } }
9077 },
9078 /* BPF_JMP32 | BPF_JSGE | BPF_X */
9079 {
9080 "JMP32_JSGE_X",
9081 .u.insns_int = {
9082 BPF_ALU32_IMM(BPF_MOV, R0, -12345678),
9083 BPF_ALU32_IMM(BPF_MOV, R1, -12345677),
9084 BPF_JMP32_REG(BPF_JSGE, R0, R1, 2),
9085 BPF_ALU32_IMM(BPF_MOV, R1, -12345678),
9086 BPF_JMP32_REG(BPF_JSGE, R0, R1, 1),
9087 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9088 BPF_EXIT_INSN(),
9089 },
9090 INTERNAL,
9091 { },
9092 { { 0, -12345678 } }
9093 },
9094 /* BPF_JMP32 | BPF_JSLT | BPF_K */
9095 {
9096 "JMP32_JSLT_K: Small immediate",
9097 .u.insns_int = {
9098 BPF_ALU32_IMM(BPF_MOV, R0, -123),
9099 BPF_JMP32_IMM(BPF_JSLT, R0, -123, 1),
9100 BPF_JMP32_IMM(BPF_JSLT, R0, -122, 1),
9101 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9102 BPF_EXIT_INSN(),
9103 },
9104 INTERNAL,
9105 { },
9106 { { 0, -123 } }
9107 },
9108 {
9109 "JMP32_JSLT_K: Large immediate",
9110 .u.insns_int = {
9111 BPF_ALU32_IMM(BPF_MOV, R0, -12345678),
9112 BPF_JMP32_IMM(BPF_JSLT, R0, -12345678, 1),
9113 BPF_JMP32_IMM(BPF_JSLT, R0, -12345677, 1),
9114 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9115 BPF_EXIT_INSN(),
9116 },
9117 INTERNAL,
9118 { },
9119 { { 0, -12345678 } }
9120 },
9121 /* BPF_JMP32 | BPF_JSLT | BPF_X */
9122 {
9123 "JMP32_JSLT_X",
9124 .u.insns_int = {
9125 BPF_ALU32_IMM(BPF_MOV, R0, -12345678),
9126 BPF_ALU32_IMM(BPF_MOV, R1, -12345678),
9127 BPF_JMP32_REG(BPF_JSLT, R0, R1, 2),
9128 BPF_ALU32_IMM(BPF_MOV, R1, -12345677),
9129 BPF_JMP32_REG(BPF_JSLT, R0, R1, 1),
9130 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9131 BPF_EXIT_INSN(),
9132 },
9133 INTERNAL,
9134 { },
9135 { { 0, -12345678 } }
9136 },
9137 /* BPF_JMP32 | BPF_JSLE | BPF_K */
9138 {
9139 "JMP32_JSLE_K: Small immediate",
9140 .u.insns_int = {
9141 BPF_ALU32_IMM(BPF_MOV, R0, -123),
9142 BPF_JMP32_IMM(BPF_JSLE, R0, -124, 1),
9143 BPF_JMP32_IMM(BPF_JSLE, R0, -123, 1),
9144 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9145 BPF_EXIT_INSN(),
9146 },
9147 INTERNAL,
9148 { },
9149 { { 0, -123 } }
9150 },
9151 {
9152 "JMP32_JSLE_K: Large immediate",
9153 .u.insns_int = {
9154 BPF_ALU32_IMM(BPF_MOV, R0, -12345678),
9155 BPF_JMP32_IMM(BPF_JSLE, R0, -12345679, 1),
9156 BPF_JMP32_IMM(BPF_JSLE, R0, -12345678, 1),
9157 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9158 BPF_EXIT_INSN(),
9159 },
9160 INTERNAL,
9161 { },
9162 { { 0, -12345678 } }
9163 },
9164 /* BPF_JMP32 | BPF_JSLE | BPF_K */
9165 {
9166 "JMP32_JSLE_X",
9167 .u.insns_int = {
9168 BPF_ALU32_IMM(BPF_MOV, R0, -12345678),
9169 BPF_ALU32_IMM(BPF_MOV, R1, -12345679),
9170 BPF_JMP32_REG(BPF_JSLE, R0, R1, 2),
9171 BPF_ALU32_IMM(BPF_MOV, R1, -12345678),
9172 BPF_JMP32_REG(BPF_JSLE, R0, R1, 1),
9173 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9174 BPF_EXIT_INSN(),
9175 },
9176 INTERNAL,
9177 { },
9178 { { 0, -12345678 } }
9179 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009180 /* BPF_JMP | BPF_EXIT */
9181 {
9182 "JMP_EXIT",
9183 .u.insns_int = {
9184 BPF_ALU32_IMM(BPF_MOV, R0, 0x4711),
9185 BPF_EXIT_INSN(),
9186 BPF_ALU32_IMM(BPF_MOV, R0, 0x4712),
9187 },
9188 INTERNAL,
9189 { },
9190 { { 0, 0x4711 } },
9191 },
9192 /* BPF_JMP | BPF_JA */
9193 {
9194 "JMP_JA: Unconditional jump: if (true) return 1",
9195 .u.insns_int = {
9196 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9197 BPF_JMP_IMM(BPF_JA, 0, 0, 1),
9198 BPF_EXIT_INSN(),
9199 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9200 BPF_EXIT_INSN(),
9201 },
9202 INTERNAL,
9203 { },
9204 { { 0, 1 } },
9205 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009206 /* BPF_JMP | BPF_JSLT | BPF_K */
9207 {
9208 "JMP_JSLT_K: Signed jump: if (-2 < -1) return 1",
9209 .u.insns_int = {
9210 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9211 BPF_LD_IMM64(R1, 0xfffffffffffffffeLL),
9212 BPF_JMP_IMM(BPF_JSLT, R1, -1, 1),
9213 BPF_EXIT_INSN(),
9214 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9215 BPF_EXIT_INSN(),
9216 },
9217 INTERNAL,
9218 { },
9219 { { 0, 1 } },
9220 },
9221 {
9222 "JMP_JSLT_K: Signed jump: if (-1 < -1) return 0",
9223 .u.insns_int = {
9224 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9225 BPF_LD_IMM64(R1, 0xffffffffffffffffLL),
9226 BPF_JMP_IMM(BPF_JSLT, R1, -1, 1),
9227 BPF_EXIT_INSN(),
9228 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9229 BPF_EXIT_INSN(),
9230 },
9231 INTERNAL,
9232 { },
9233 { { 0, 1 } },
9234 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009235 /* BPF_JMP | BPF_JSGT | BPF_K */
9236 {
9237 "JMP_JSGT_K: Signed jump: if (-1 > -2) return 1",
9238 .u.insns_int = {
9239 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9240 BPF_LD_IMM64(R1, 0xffffffffffffffffLL),
9241 BPF_JMP_IMM(BPF_JSGT, R1, -2, 1),
9242 BPF_EXIT_INSN(),
9243 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9244 BPF_EXIT_INSN(),
9245 },
9246 INTERNAL,
9247 { },
9248 { { 0, 1 } },
9249 },
9250 {
9251 "JMP_JSGT_K: Signed jump: if (-1 > -1) return 0",
9252 .u.insns_int = {
9253 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9254 BPF_LD_IMM64(R1, 0xffffffffffffffffLL),
9255 BPF_JMP_IMM(BPF_JSGT, R1, -1, 1),
9256 BPF_EXIT_INSN(),
9257 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9258 BPF_EXIT_INSN(),
9259 },
9260 INTERNAL,
9261 { },
9262 { { 0, 1 } },
9263 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009264 /* BPF_JMP | BPF_JSLE | BPF_K */
9265 {
9266 "JMP_JSLE_K: Signed jump: if (-2 <= -1) return 1",
9267 .u.insns_int = {
9268 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9269 BPF_LD_IMM64(R1, 0xfffffffffffffffeLL),
9270 BPF_JMP_IMM(BPF_JSLE, R1, -1, 1),
9271 BPF_EXIT_INSN(),
9272 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9273 BPF_EXIT_INSN(),
9274 },
9275 INTERNAL,
9276 { },
9277 { { 0, 1 } },
9278 },
9279 {
9280 "JMP_JSLE_K: Signed jump: if (-1 <= -1) return 1",
9281 .u.insns_int = {
9282 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9283 BPF_LD_IMM64(R1, 0xffffffffffffffffLL),
9284 BPF_JMP_IMM(BPF_JSLE, R1, -1, 1),
9285 BPF_EXIT_INSN(),
9286 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9287 BPF_EXIT_INSN(),
9288 },
9289 INTERNAL,
9290 { },
9291 { { 0, 1 } },
9292 },
9293 {
9294 "JMP_JSLE_K: Signed jump: value walk 1",
9295 .u.insns_int = {
9296 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9297 BPF_LD_IMM64(R1, 3),
9298 BPF_JMP_IMM(BPF_JSLE, R1, 0, 6),
9299 BPF_ALU64_IMM(BPF_SUB, R1, 1),
9300 BPF_JMP_IMM(BPF_JSLE, R1, 0, 4),
9301 BPF_ALU64_IMM(BPF_SUB, R1, 1),
9302 BPF_JMP_IMM(BPF_JSLE, R1, 0, 2),
9303 BPF_ALU64_IMM(BPF_SUB, R1, 1),
9304 BPF_JMP_IMM(BPF_JSLE, R1, 0, 1),
9305 BPF_EXIT_INSN(), /* bad exit */
9306 BPF_ALU32_IMM(BPF_MOV, R0, 1), /* good exit */
9307 BPF_EXIT_INSN(),
9308 },
9309 INTERNAL,
9310 { },
9311 { { 0, 1 } },
9312 },
9313 {
9314 "JMP_JSLE_K: Signed jump: value walk 2",
9315 .u.insns_int = {
9316 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9317 BPF_LD_IMM64(R1, 3),
9318 BPF_JMP_IMM(BPF_JSLE, R1, 0, 4),
9319 BPF_ALU64_IMM(BPF_SUB, R1, 2),
9320 BPF_JMP_IMM(BPF_JSLE, R1, 0, 2),
9321 BPF_ALU64_IMM(BPF_SUB, R1, 2),
9322 BPF_JMP_IMM(BPF_JSLE, R1, 0, 1),
9323 BPF_EXIT_INSN(), /* bad exit */
9324 BPF_ALU32_IMM(BPF_MOV, R0, 1), /* good exit */
9325 BPF_EXIT_INSN(),
9326 },
9327 INTERNAL,
9328 { },
9329 { { 0, 1 } },
9330 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009331 /* BPF_JMP | BPF_JSGE | BPF_K */
9332 {
9333 "JMP_JSGE_K: Signed jump: if (-1 >= -2) return 1",
9334 .u.insns_int = {
9335 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9336 BPF_LD_IMM64(R1, 0xffffffffffffffffLL),
9337 BPF_JMP_IMM(BPF_JSGE, R1, -2, 1),
9338 BPF_EXIT_INSN(),
9339 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9340 BPF_EXIT_INSN(),
9341 },
9342 INTERNAL,
9343 { },
9344 { { 0, 1 } },
9345 },
9346 {
9347 "JMP_JSGE_K: Signed jump: if (-1 >= -1) return 1",
9348 .u.insns_int = {
9349 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9350 BPF_LD_IMM64(R1, 0xffffffffffffffffLL),
9351 BPF_JMP_IMM(BPF_JSGE, R1, -1, 1),
9352 BPF_EXIT_INSN(),
9353 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9354 BPF_EXIT_INSN(),
9355 },
9356 INTERNAL,
9357 { },
9358 { { 0, 1 } },
9359 },
David Daney791caeb2017-05-24 16:35:49 -07009360 {
9361 "JMP_JSGE_K: Signed jump: value walk 1",
9362 .u.insns_int = {
9363 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9364 BPF_LD_IMM64(R1, -3),
9365 BPF_JMP_IMM(BPF_JSGE, R1, 0, 6),
9366 BPF_ALU64_IMM(BPF_ADD, R1, 1),
9367 BPF_JMP_IMM(BPF_JSGE, R1, 0, 4),
9368 BPF_ALU64_IMM(BPF_ADD, R1, 1),
9369 BPF_JMP_IMM(BPF_JSGE, R1, 0, 2),
9370 BPF_ALU64_IMM(BPF_ADD, R1, 1),
9371 BPF_JMP_IMM(BPF_JSGE, R1, 0, 1),
9372 BPF_EXIT_INSN(), /* bad exit */
9373 BPF_ALU32_IMM(BPF_MOV, R0, 1), /* good exit */
9374 BPF_EXIT_INSN(),
9375 },
9376 INTERNAL,
9377 { },
9378 { { 0, 1 } },
9379 },
9380 {
9381 "JMP_JSGE_K: Signed jump: value walk 2",
9382 .u.insns_int = {
9383 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9384 BPF_LD_IMM64(R1, -3),
9385 BPF_JMP_IMM(BPF_JSGE, R1, 0, 4),
9386 BPF_ALU64_IMM(BPF_ADD, R1, 2),
9387 BPF_JMP_IMM(BPF_JSGE, R1, 0, 2),
9388 BPF_ALU64_IMM(BPF_ADD, R1, 2),
9389 BPF_JMP_IMM(BPF_JSGE, R1, 0, 1),
9390 BPF_EXIT_INSN(), /* bad exit */
9391 BPF_ALU32_IMM(BPF_MOV, R0, 1), /* good exit */
9392 BPF_EXIT_INSN(),
9393 },
9394 INTERNAL,
9395 { },
9396 { { 0, 1 } },
9397 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009398 /* BPF_JMP | BPF_JGT | BPF_K */
9399 {
9400 "JMP_JGT_K: if (3 > 2) return 1",
9401 .u.insns_int = {
9402 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9403 BPF_LD_IMM64(R1, 3),
9404 BPF_JMP_IMM(BPF_JGT, R1, 2, 1),
9405 BPF_EXIT_INSN(),
9406 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9407 BPF_EXIT_INSN(),
9408 },
9409 INTERNAL,
9410 { },
9411 { { 0, 1 } },
9412 },
Naveen N. Raoc7395d62016-04-05 15:32:54 +05309413 {
9414 "JMP_JGT_K: Unsigned jump: if (-1 > 1) return 1",
9415 .u.insns_int = {
9416 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9417 BPF_LD_IMM64(R1, -1),
9418 BPF_JMP_IMM(BPF_JGT, R1, 1, 1),
9419 BPF_EXIT_INSN(),
9420 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9421 BPF_EXIT_INSN(),
9422 },
9423 INTERNAL,
9424 { },
9425 { { 0, 1 } },
9426 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009427 /* BPF_JMP | BPF_JLT | BPF_K */
9428 {
9429 "JMP_JLT_K: if (2 < 3) return 1",
9430 .u.insns_int = {
9431 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9432 BPF_LD_IMM64(R1, 2),
9433 BPF_JMP_IMM(BPF_JLT, R1, 3, 1),
9434 BPF_EXIT_INSN(),
9435 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9436 BPF_EXIT_INSN(),
9437 },
9438 INTERNAL,
9439 { },
9440 { { 0, 1 } },
9441 },
9442 {
9443 "JMP_JGT_K: Unsigned jump: if (1 < -1) return 1",
9444 .u.insns_int = {
9445 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9446 BPF_LD_IMM64(R1, 1),
9447 BPF_JMP_IMM(BPF_JLT, R1, -1, 1),
9448 BPF_EXIT_INSN(),
9449 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9450 BPF_EXIT_INSN(),
9451 },
9452 INTERNAL,
9453 { },
9454 { { 0, 1 } },
9455 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009456 /* BPF_JMP | BPF_JGE | BPF_K */
9457 {
9458 "JMP_JGE_K: if (3 >= 2) return 1",
9459 .u.insns_int = {
9460 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9461 BPF_LD_IMM64(R1, 3),
9462 BPF_JMP_IMM(BPF_JGE, R1, 2, 1),
9463 BPF_EXIT_INSN(),
9464 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9465 BPF_EXIT_INSN(),
9466 },
9467 INTERNAL,
9468 { },
9469 { { 0, 1 } },
9470 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009471 /* BPF_JMP | BPF_JLE | BPF_K */
9472 {
9473 "JMP_JLE_K: if (2 <= 3) return 1",
9474 .u.insns_int = {
9475 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9476 BPF_LD_IMM64(R1, 2),
9477 BPF_JMP_IMM(BPF_JLE, R1, 3, 1),
9478 BPF_EXIT_INSN(),
9479 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9480 BPF_EXIT_INSN(),
9481 },
9482 INTERNAL,
9483 { },
9484 { { 0, 1 } },
9485 },
Michael Holzheufe593842015-05-22 08:36:40 -07009486 /* BPF_JMP | BPF_JGT | BPF_K jump backwards */
9487 {
9488 "JMP_JGT_K: if (3 > 2) return 1 (jump backwards)",
9489 .u.insns_int = {
9490 BPF_JMP_IMM(BPF_JA, 0, 0, 2), /* goto start */
9491 BPF_ALU32_IMM(BPF_MOV, R0, 1), /* out: */
9492 BPF_EXIT_INSN(),
9493 BPF_ALU32_IMM(BPF_MOV, R0, 0), /* start: */
9494 BPF_LD_IMM64(R1, 3), /* note: this takes 2 insns */
9495 BPF_JMP_IMM(BPF_JGT, R1, 2, -6), /* goto out */
9496 BPF_EXIT_INSN(),
9497 },
9498 INTERNAL,
9499 { },
9500 { { 0, 1 } },
9501 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009502 {
9503 "JMP_JGE_K: if (3 >= 3) return 1",
9504 .u.insns_int = {
9505 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9506 BPF_LD_IMM64(R1, 3),
9507 BPF_JMP_IMM(BPF_JGE, R1, 3, 1),
9508 BPF_EXIT_INSN(),
9509 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9510 BPF_EXIT_INSN(),
9511 },
9512 INTERNAL,
9513 { },
9514 { { 0, 1 } },
9515 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009516 /* BPF_JMP | BPF_JLT | BPF_K jump backwards */
9517 {
9518 "JMP_JGT_K: if (2 < 3) return 1 (jump backwards)",
9519 .u.insns_int = {
9520 BPF_JMP_IMM(BPF_JA, 0, 0, 2), /* goto start */
9521 BPF_ALU32_IMM(BPF_MOV, R0, 1), /* out: */
9522 BPF_EXIT_INSN(),
9523 BPF_ALU32_IMM(BPF_MOV, R0, 0), /* start: */
9524 BPF_LD_IMM64(R1, 2), /* note: this takes 2 insns */
9525 BPF_JMP_IMM(BPF_JLT, R1, 3, -6), /* goto out */
9526 BPF_EXIT_INSN(),
9527 },
9528 INTERNAL,
9529 { },
9530 { { 0, 1 } },
9531 },
9532 {
9533 "JMP_JLE_K: if (3 <= 3) return 1",
9534 .u.insns_int = {
9535 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9536 BPF_LD_IMM64(R1, 3),
9537 BPF_JMP_IMM(BPF_JLE, R1, 3, 1),
9538 BPF_EXIT_INSN(),
9539 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9540 BPF_EXIT_INSN(),
9541 },
9542 INTERNAL,
9543 { },
9544 { { 0, 1 } },
9545 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009546 /* BPF_JMP | BPF_JNE | BPF_K */
9547 {
9548 "JMP_JNE_K: if (3 != 2) return 1",
9549 .u.insns_int = {
9550 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9551 BPF_LD_IMM64(R1, 3),
9552 BPF_JMP_IMM(BPF_JNE, R1, 2, 1),
9553 BPF_EXIT_INSN(),
9554 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9555 BPF_EXIT_INSN(),
9556 },
9557 INTERNAL,
9558 { },
9559 { { 0, 1 } },
9560 },
9561 /* BPF_JMP | BPF_JEQ | BPF_K */
9562 {
9563 "JMP_JEQ_K: if (3 == 3) return 1",
9564 .u.insns_int = {
9565 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9566 BPF_LD_IMM64(R1, 3),
9567 BPF_JMP_IMM(BPF_JEQ, R1, 3, 1),
9568 BPF_EXIT_INSN(),
9569 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9570 BPF_EXIT_INSN(),
9571 },
9572 INTERNAL,
9573 { },
9574 { { 0, 1 } },
9575 },
9576 /* BPF_JMP | BPF_JSET | BPF_K */
9577 {
9578 "JMP_JSET_K: if (0x3 & 0x2) return 1",
9579 .u.insns_int = {
9580 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9581 BPF_LD_IMM64(R1, 3),
Naveen N. Rao9f134c32016-04-05 15:32:53 +05309582 BPF_JMP_IMM(BPF_JSET, R1, 2, 1),
Michael Holzheucffc6422015-05-11 22:22:44 -07009583 BPF_EXIT_INSN(),
9584 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9585 BPF_EXIT_INSN(),
9586 },
9587 INTERNAL,
9588 { },
9589 { { 0, 1 } },
9590 },
9591 {
9592 "JMP_JSET_K: if (0x3 & 0xffffffff) return 1",
9593 .u.insns_int = {
9594 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9595 BPF_LD_IMM64(R1, 3),
Naveen N. Rao9f134c32016-04-05 15:32:53 +05309596 BPF_JMP_IMM(BPF_JSET, R1, 0xffffffff, 1),
Michael Holzheucffc6422015-05-11 22:22:44 -07009597 BPF_EXIT_INSN(),
9598 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9599 BPF_EXIT_INSN(),
9600 },
9601 INTERNAL,
9602 { },
9603 { { 0, 1 } },
9604 },
9605 /* BPF_JMP | BPF_JSGT | BPF_X */
9606 {
9607 "JMP_JSGT_X: Signed jump: if (-1 > -2) return 1",
9608 .u.insns_int = {
9609 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9610 BPF_LD_IMM64(R1, -1),
9611 BPF_LD_IMM64(R2, -2),
9612 BPF_JMP_REG(BPF_JSGT, R1, R2, 1),
9613 BPF_EXIT_INSN(),
9614 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9615 BPF_EXIT_INSN(),
9616 },
9617 INTERNAL,
9618 { },
9619 { { 0, 1 } },
9620 },
9621 {
9622 "JMP_JSGT_X: Signed jump: if (-1 > -1) return 0",
9623 .u.insns_int = {
9624 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9625 BPF_LD_IMM64(R1, -1),
9626 BPF_LD_IMM64(R2, -1),
9627 BPF_JMP_REG(BPF_JSGT, R1, R2, 1),
9628 BPF_EXIT_INSN(),
9629 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9630 BPF_EXIT_INSN(),
9631 },
9632 INTERNAL,
9633 { },
9634 { { 0, 1 } },
9635 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009636 /* BPF_JMP | BPF_JSLT | BPF_X */
9637 {
9638 "JMP_JSLT_X: Signed jump: if (-2 < -1) return 1",
9639 .u.insns_int = {
9640 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9641 BPF_LD_IMM64(R1, -1),
9642 BPF_LD_IMM64(R2, -2),
9643 BPF_JMP_REG(BPF_JSLT, R2, R1, 1),
9644 BPF_EXIT_INSN(),
9645 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9646 BPF_EXIT_INSN(),
9647 },
9648 INTERNAL,
9649 { },
9650 { { 0, 1 } },
9651 },
9652 {
9653 "JMP_JSLT_X: Signed jump: if (-1 < -1) return 0",
9654 .u.insns_int = {
9655 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9656 BPF_LD_IMM64(R1, -1),
9657 BPF_LD_IMM64(R2, -1),
9658 BPF_JMP_REG(BPF_JSLT, R1, R2, 1),
9659 BPF_EXIT_INSN(),
9660 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9661 BPF_EXIT_INSN(),
9662 },
9663 INTERNAL,
9664 { },
9665 { { 0, 1 } },
9666 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009667 /* BPF_JMP | BPF_JSGE | BPF_X */
9668 {
9669 "JMP_JSGE_X: Signed jump: if (-1 >= -2) return 1",
9670 .u.insns_int = {
9671 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9672 BPF_LD_IMM64(R1, -1),
9673 BPF_LD_IMM64(R2, -2),
9674 BPF_JMP_REG(BPF_JSGE, R1, R2, 1),
9675 BPF_EXIT_INSN(),
9676 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9677 BPF_EXIT_INSN(),
9678 },
9679 INTERNAL,
9680 { },
9681 { { 0, 1 } },
9682 },
9683 {
9684 "JMP_JSGE_X: Signed jump: if (-1 >= -1) return 1",
9685 .u.insns_int = {
9686 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9687 BPF_LD_IMM64(R1, -1),
9688 BPF_LD_IMM64(R2, -1),
9689 BPF_JMP_REG(BPF_JSGE, R1, R2, 1),
9690 BPF_EXIT_INSN(),
9691 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9692 BPF_EXIT_INSN(),
9693 },
9694 INTERNAL,
9695 { },
9696 { { 0, 1 } },
9697 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009698 /* BPF_JMP | BPF_JSLE | BPF_X */
9699 {
9700 "JMP_JSLE_X: Signed jump: if (-2 <= -1) return 1",
9701 .u.insns_int = {
9702 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9703 BPF_LD_IMM64(R1, -1),
9704 BPF_LD_IMM64(R2, -2),
9705 BPF_JMP_REG(BPF_JSLE, R2, R1, 1),
9706 BPF_EXIT_INSN(),
9707 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9708 BPF_EXIT_INSN(),
9709 },
9710 INTERNAL,
9711 { },
9712 { { 0, 1 } },
9713 },
9714 {
9715 "JMP_JSLE_X: Signed jump: if (-1 <= -1) return 1",
9716 .u.insns_int = {
9717 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9718 BPF_LD_IMM64(R1, -1),
9719 BPF_LD_IMM64(R2, -1),
9720 BPF_JMP_REG(BPF_JSLE, R1, R2, 1),
9721 BPF_EXIT_INSN(),
9722 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9723 BPF_EXIT_INSN(),
9724 },
9725 INTERNAL,
9726 { },
9727 { { 0, 1 } },
9728 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009729 /* BPF_JMP | BPF_JGT | BPF_X */
9730 {
9731 "JMP_JGT_X: if (3 > 2) return 1",
9732 .u.insns_int = {
9733 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9734 BPF_LD_IMM64(R1, 3),
9735 BPF_LD_IMM64(R2, 2),
9736 BPF_JMP_REG(BPF_JGT, R1, R2, 1),
9737 BPF_EXIT_INSN(),
9738 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9739 BPF_EXIT_INSN(),
9740 },
9741 INTERNAL,
9742 { },
9743 { { 0, 1 } },
9744 },
Naveen N. Raoc7395d62016-04-05 15:32:54 +05309745 {
9746 "JMP_JGT_X: Unsigned jump: if (-1 > 1) return 1",
9747 .u.insns_int = {
9748 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9749 BPF_LD_IMM64(R1, -1),
9750 BPF_LD_IMM64(R2, 1),
9751 BPF_JMP_REG(BPF_JGT, R1, R2, 1),
9752 BPF_EXIT_INSN(),
9753 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9754 BPF_EXIT_INSN(),
9755 },
9756 INTERNAL,
9757 { },
9758 { { 0, 1 } },
9759 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009760 /* BPF_JMP | BPF_JLT | BPF_X */
9761 {
9762 "JMP_JLT_X: if (2 < 3) return 1",
9763 .u.insns_int = {
9764 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9765 BPF_LD_IMM64(R1, 3),
9766 BPF_LD_IMM64(R2, 2),
9767 BPF_JMP_REG(BPF_JLT, R2, R1, 1),
9768 BPF_EXIT_INSN(),
9769 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9770 BPF_EXIT_INSN(),
9771 },
9772 INTERNAL,
9773 { },
9774 { { 0, 1 } },
9775 },
9776 {
9777 "JMP_JLT_X: Unsigned jump: if (1 < -1) return 1",
9778 .u.insns_int = {
9779 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9780 BPF_LD_IMM64(R1, -1),
9781 BPF_LD_IMM64(R2, 1),
9782 BPF_JMP_REG(BPF_JLT, R2, R1, 1),
9783 BPF_EXIT_INSN(),
9784 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9785 BPF_EXIT_INSN(),
9786 },
9787 INTERNAL,
9788 { },
9789 { { 0, 1 } },
9790 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009791 /* BPF_JMP | BPF_JGE | BPF_X */
9792 {
9793 "JMP_JGE_X: if (3 >= 2) return 1",
9794 .u.insns_int = {
9795 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9796 BPF_LD_IMM64(R1, 3),
9797 BPF_LD_IMM64(R2, 2),
9798 BPF_JMP_REG(BPF_JGE, R1, R2, 1),
9799 BPF_EXIT_INSN(),
9800 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9801 BPF_EXIT_INSN(),
9802 },
9803 INTERNAL,
9804 { },
9805 { { 0, 1 } },
9806 },
9807 {
9808 "JMP_JGE_X: if (3 >= 3) return 1",
9809 .u.insns_int = {
9810 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9811 BPF_LD_IMM64(R1, 3),
9812 BPF_LD_IMM64(R2, 3),
9813 BPF_JMP_REG(BPF_JGE, R1, R2, 1),
9814 BPF_EXIT_INSN(),
9815 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9816 BPF_EXIT_INSN(),
9817 },
9818 INTERNAL,
9819 { },
9820 { { 0, 1 } },
9821 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009822 /* BPF_JMP | BPF_JLE | BPF_X */
9823 {
9824 "JMP_JLE_X: if (2 <= 3) return 1",
9825 .u.insns_int = {
9826 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9827 BPF_LD_IMM64(R1, 3),
9828 BPF_LD_IMM64(R2, 2),
9829 BPF_JMP_REG(BPF_JLE, R2, R1, 1),
9830 BPF_EXIT_INSN(),
9831 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9832 BPF_EXIT_INSN(),
9833 },
9834 INTERNAL,
9835 { },
9836 { { 0, 1 } },
9837 },
9838 {
9839 "JMP_JLE_X: if (3 <= 3) return 1",
9840 .u.insns_int = {
9841 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9842 BPF_LD_IMM64(R1, 3),
9843 BPF_LD_IMM64(R2, 3),
9844 BPF_JMP_REG(BPF_JLE, R1, R2, 1),
9845 BPF_EXIT_INSN(),
9846 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9847 BPF_EXIT_INSN(),
9848 },
9849 INTERNAL,
9850 { },
9851 { { 0, 1 } },
9852 },
Daniel Borkmannddc665a2017-05-02 20:34:54 +02009853 {
9854 /* Mainly testing JIT + imm64 here. */
9855 "JMP_JGE_X: ldimm64 test 1",
9856 .u.insns_int = {
9857 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9858 BPF_LD_IMM64(R1, 3),
9859 BPF_LD_IMM64(R2, 2),
9860 BPF_JMP_REG(BPF_JGE, R1, R2, 2),
Geert Uytterhoeven86f8e242017-05-03 13:31:04 +02009861 BPF_LD_IMM64(R0, 0xffffffffffffffffULL),
9862 BPF_LD_IMM64(R0, 0xeeeeeeeeeeeeeeeeULL),
Daniel Borkmannddc665a2017-05-02 20:34:54 +02009863 BPF_EXIT_INSN(),
9864 },
9865 INTERNAL,
9866 { },
9867 { { 0, 0xeeeeeeeeU } },
9868 },
9869 {
9870 "JMP_JGE_X: ldimm64 test 2",
9871 .u.insns_int = {
9872 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9873 BPF_LD_IMM64(R1, 3),
9874 BPF_LD_IMM64(R2, 2),
9875 BPF_JMP_REG(BPF_JGE, R1, R2, 0),
Geert Uytterhoeven86f8e242017-05-03 13:31:04 +02009876 BPF_LD_IMM64(R0, 0xffffffffffffffffULL),
Daniel Borkmannddc665a2017-05-02 20:34:54 +02009877 BPF_EXIT_INSN(),
9878 },
9879 INTERNAL,
9880 { },
9881 { { 0, 0xffffffffU } },
9882 },
9883 {
9884 "JMP_JGE_X: ldimm64 test 3",
9885 .u.insns_int = {
9886 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9887 BPF_LD_IMM64(R1, 3),
9888 BPF_LD_IMM64(R2, 2),
9889 BPF_JMP_REG(BPF_JGE, R1, R2, 4),
Geert Uytterhoeven86f8e242017-05-03 13:31:04 +02009890 BPF_LD_IMM64(R0, 0xffffffffffffffffULL),
9891 BPF_LD_IMM64(R0, 0xeeeeeeeeeeeeeeeeULL),
Daniel Borkmannddc665a2017-05-02 20:34:54 +02009892 BPF_EXIT_INSN(),
9893 },
9894 INTERNAL,
9895 { },
9896 { { 0, 1 } },
9897 },
Daniel Borkmann92b31a92017-08-10 01:39:55 +02009898 {
9899 "JMP_JLE_X: ldimm64 test 1",
9900 .u.insns_int = {
9901 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9902 BPF_LD_IMM64(R1, 3),
9903 BPF_LD_IMM64(R2, 2),
9904 BPF_JMP_REG(BPF_JLE, R2, R1, 2),
9905 BPF_LD_IMM64(R0, 0xffffffffffffffffULL),
9906 BPF_LD_IMM64(R0, 0xeeeeeeeeeeeeeeeeULL),
9907 BPF_EXIT_INSN(),
9908 },
9909 INTERNAL,
9910 { },
9911 { { 0, 0xeeeeeeeeU } },
9912 },
9913 {
9914 "JMP_JLE_X: ldimm64 test 2",
9915 .u.insns_int = {
9916 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9917 BPF_LD_IMM64(R1, 3),
9918 BPF_LD_IMM64(R2, 2),
9919 BPF_JMP_REG(BPF_JLE, R2, R1, 0),
9920 BPF_LD_IMM64(R0, 0xffffffffffffffffULL),
9921 BPF_EXIT_INSN(),
9922 },
9923 INTERNAL,
9924 { },
9925 { { 0, 0xffffffffU } },
9926 },
9927 {
9928 "JMP_JLE_X: ldimm64 test 3",
9929 .u.insns_int = {
9930 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9931 BPF_LD_IMM64(R1, 3),
9932 BPF_LD_IMM64(R2, 2),
9933 BPF_JMP_REG(BPF_JLE, R2, R1, 4),
9934 BPF_LD_IMM64(R0, 0xffffffffffffffffULL),
9935 BPF_LD_IMM64(R0, 0xeeeeeeeeeeeeeeeeULL),
9936 BPF_EXIT_INSN(),
9937 },
9938 INTERNAL,
9939 { },
9940 { { 0, 1 } },
9941 },
Michael Holzheucffc6422015-05-11 22:22:44 -07009942 /* BPF_JMP | BPF_JNE | BPF_X */
9943 {
9944 "JMP_JNE_X: if (3 != 2) return 1",
9945 .u.insns_int = {
9946 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9947 BPF_LD_IMM64(R1, 3),
9948 BPF_LD_IMM64(R2, 2),
9949 BPF_JMP_REG(BPF_JNE, R1, R2, 1),
9950 BPF_EXIT_INSN(),
9951 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9952 BPF_EXIT_INSN(),
9953 },
9954 INTERNAL,
9955 { },
9956 { { 0, 1 } },
9957 },
9958 /* BPF_JMP | BPF_JEQ | BPF_X */
9959 {
9960 "JMP_JEQ_X: if (3 == 3) return 1",
9961 .u.insns_int = {
9962 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9963 BPF_LD_IMM64(R1, 3),
9964 BPF_LD_IMM64(R2, 3),
9965 BPF_JMP_REG(BPF_JEQ, R1, R2, 1),
9966 BPF_EXIT_INSN(),
9967 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9968 BPF_EXIT_INSN(),
9969 },
9970 INTERNAL,
9971 { },
9972 { { 0, 1 } },
9973 },
9974 /* BPF_JMP | BPF_JSET | BPF_X */
9975 {
9976 "JMP_JSET_X: if (0x3 & 0x2) return 1",
9977 .u.insns_int = {
9978 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9979 BPF_LD_IMM64(R1, 3),
9980 BPF_LD_IMM64(R2, 2),
Naveen N. Rao9f134c32016-04-05 15:32:53 +05309981 BPF_JMP_REG(BPF_JSET, R1, R2, 1),
Michael Holzheucffc6422015-05-11 22:22:44 -07009982 BPF_EXIT_INSN(),
9983 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9984 BPF_EXIT_INSN(),
9985 },
9986 INTERNAL,
9987 { },
9988 { { 0, 1 } },
9989 },
9990 {
9991 "JMP_JSET_X: if (0x3 & 0xffffffff) return 1",
9992 .u.insns_int = {
9993 BPF_ALU32_IMM(BPF_MOV, R0, 0),
9994 BPF_LD_IMM64(R1, 3),
9995 BPF_LD_IMM64(R2, 0xffffffff),
Naveen N. Rao9f134c32016-04-05 15:32:53 +05309996 BPF_JMP_REG(BPF_JSET, R1, R2, 1),
Michael Holzheucffc6422015-05-11 22:22:44 -07009997 BPF_EXIT_INSN(),
9998 BPF_ALU32_IMM(BPF_MOV, R0, 1),
9999 BPF_EXIT_INSN(),
10000 },
10001 INTERNAL,
10002 { },
10003 { { 0, 1 } },
10004 },
Daniel Borkmannbde28bc2015-05-26 22:35:43 +020010005 {
10006 "JMP_JA: Jump, gap, jump, ...",
10007 { },
10008 CLASSIC | FLAG_NO_DATA,
10009 { },
10010 { { 0, 0xababcbac } },
10011 .fill_helper = bpf_fill_ja,
10012 },
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020010013 { /* Mainly checking JIT here. */
10014 "BPF_MAXINSNS: Maximum possible literals",
10015 { },
10016 CLASSIC | FLAG_NO_DATA,
10017 { },
10018 { { 0, 0xffffffff } },
10019 .fill_helper = bpf_fill_maxinsns1,
10020 },
10021 { /* Mainly checking JIT here. */
10022 "BPF_MAXINSNS: Single literal",
10023 { },
10024 CLASSIC | FLAG_NO_DATA,
10025 { },
10026 { { 0, 0xfefefefe } },
10027 .fill_helper = bpf_fill_maxinsns2,
10028 },
10029 { /* Mainly checking JIT here. */
10030 "BPF_MAXINSNS: Run/add until end",
10031 { },
10032 CLASSIC | FLAG_NO_DATA,
10033 { },
10034 { { 0, 0x947bf368 } },
10035 .fill_helper = bpf_fill_maxinsns3,
10036 },
10037 {
10038 "BPF_MAXINSNS: Too many instructions",
10039 { },
10040 CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL,
10041 { },
10042 { },
10043 .fill_helper = bpf_fill_maxinsns4,
Yonghong Song09584b42018-02-02 22:37:15 -080010044 .expected_errcode = -EINVAL,
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020010045 },
10046 { /* Mainly checking JIT here. */
10047 "BPF_MAXINSNS: Very long jump",
10048 { },
10049 CLASSIC | FLAG_NO_DATA,
10050 { },
10051 { { 0, 0xabababab } },
10052 .fill_helper = bpf_fill_maxinsns5,
10053 },
10054 { /* Mainly checking JIT here. */
10055 "BPF_MAXINSNS: Ctx heavy transformations",
10056 { },
10057 CLASSIC,
10058 { },
10059 {
Michał Mirosław0c4b2d32018-11-10 19:58:36 +010010060 { 1, SKB_VLAN_PRESENT },
10061 { 10, SKB_VLAN_PRESENT }
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020010062 },
10063 .fill_helper = bpf_fill_maxinsns6,
10064 },
10065 { /* Mainly checking JIT here. */
10066 "BPF_MAXINSNS: Call heavy transformations",
10067 { },
10068 CLASSIC | FLAG_NO_DATA,
10069 { },
10070 { { 1, 0 }, { 10, 0 } },
10071 .fill_helper = bpf_fill_maxinsns7,
10072 },
10073 { /* Mainly checking JIT here. */
10074 "BPF_MAXINSNS: Jump heavy test",
10075 { },
10076 CLASSIC | FLAG_NO_DATA,
10077 { },
10078 { { 0, 0xffffffff } },
10079 .fill_helper = bpf_fill_maxinsns8,
10080 },
Daniel Borkmann3b529602015-05-23 01:10:07 +020010081 { /* Mainly checking JIT here. */
10082 "BPF_MAXINSNS: Very long jump backwards",
10083 { },
10084 INTERNAL | FLAG_NO_DATA,
10085 { },
10086 { { 0, 0xcbababab } },
10087 .fill_helper = bpf_fill_maxinsns9,
10088 },
10089 { /* Mainly checking JIT here. */
10090 "BPF_MAXINSNS: Edge hopping nuthouse",
10091 { },
10092 INTERNAL | FLAG_NO_DATA,
10093 { },
10094 { { 0, 0xabababac } },
10095 .fill_helper = bpf_fill_maxinsns10,
10096 },
Daniel Borkmannbde28bc2015-05-26 22:35:43 +020010097 {
10098 "BPF_MAXINSNS: Jump, gap, jump, ...",
10099 { },
10100 CLASSIC | FLAG_NO_DATA,
10101 { },
10102 { { 0, 0xababcbac } },
10103 .fill_helper = bpf_fill_maxinsns11,
10104 },
Alexei Starovoitov4d9c5c52015-07-20 20:34:19 -070010105 {
Daniel Borkmannbe088152018-06-02 23:06:32 +020010106 "BPF_MAXINSNS: jump over MSH",
10107 { },
10108 CLASSIC | FLAG_EXPECTED_FAIL,
10109 { 0xfa, 0xfb, 0xfc, 0xfd, },
10110 { { 4, 0xabababab } },
10111 .fill_helper = bpf_fill_maxinsns12,
10112 .expected_errcode = -EINVAL,
10113 },
10114 {
10115 "BPF_MAXINSNS: exec all MSH",
10116 { },
10117 CLASSIC,
10118 { 0xfa, 0xfb, 0xfc, 0xfd, },
10119 { { 4, 0xababab83 } },
10120 .fill_helper = bpf_fill_maxinsns13,
10121 },
10122 {
Alexei Starovoitov4d9c5c52015-07-20 20:34:19 -070010123 "BPF_MAXINSNS: ld_abs+get_processor_id",
10124 { },
10125 CLASSIC,
10126 { },
10127 { { 1, 0xbee } },
10128 .fill_helper = bpf_fill_ld_abs_get_processor_id,
10129 },
Nicolas Schichan2cf1ad72015-08-04 15:19:09 +020010130 /*
10131 * LD_IND / LD_ABS on fragmented SKBs
10132 */
10133 {
10134 "LD_IND byte frag",
10135 .u.insns = {
10136 BPF_STMT(BPF_LDX | BPF_IMM, 0x40),
10137 BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x0),
10138 BPF_STMT(BPF_RET | BPF_A, 0x0),
10139 },
10140 CLASSIC | FLAG_SKB_FRAG,
10141 { },
10142 { {0x40, 0x42} },
10143 .frag_data = {
10144 0x42, 0x00, 0x00, 0x00,
10145 0x43, 0x44, 0x00, 0x00,
10146 0x21, 0x07, 0x19, 0x83,
10147 },
10148 },
10149 {
10150 "LD_IND halfword frag",
10151 .u.insns = {
10152 BPF_STMT(BPF_LDX | BPF_IMM, 0x40),
10153 BPF_STMT(BPF_LD | BPF_IND | BPF_H, 0x4),
10154 BPF_STMT(BPF_RET | BPF_A, 0x0),
10155 },
10156 CLASSIC | FLAG_SKB_FRAG,
10157 { },
10158 { {0x40, 0x4344} },
10159 .frag_data = {
10160 0x42, 0x00, 0x00, 0x00,
10161 0x43, 0x44, 0x00, 0x00,
10162 0x21, 0x07, 0x19, 0x83,
10163 },
10164 },
10165 {
10166 "LD_IND word frag",
10167 .u.insns = {
10168 BPF_STMT(BPF_LDX | BPF_IMM, 0x40),
10169 BPF_STMT(BPF_LD | BPF_IND | BPF_W, 0x8),
10170 BPF_STMT(BPF_RET | BPF_A, 0x0),
10171 },
10172 CLASSIC | FLAG_SKB_FRAG,
10173 { },
10174 { {0x40, 0x21071983} },
10175 .frag_data = {
10176 0x42, 0x00, 0x00, 0x00,
10177 0x43, 0x44, 0x00, 0x00,
10178 0x21, 0x07, 0x19, 0x83,
10179 },
10180 },
10181 {
10182 "LD_IND halfword mixed head/frag",
10183 .u.insns = {
10184 BPF_STMT(BPF_LDX | BPF_IMM, 0x40),
10185 BPF_STMT(BPF_LD | BPF_IND | BPF_H, -0x1),
10186 BPF_STMT(BPF_RET | BPF_A, 0x0),
10187 },
10188 CLASSIC | FLAG_SKB_FRAG,
10189 { [0x3e] = 0x25, [0x3f] = 0x05, },
10190 { {0x40, 0x0519} },
10191 .frag_data = { 0x19, 0x82 },
10192 },
10193 {
10194 "LD_IND word mixed head/frag",
10195 .u.insns = {
10196 BPF_STMT(BPF_LDX | BPF_IMM, 0x40),
10197 BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x2),
10198 BPF_STMT(BPF_RET | BPF_A, 0x0),
10199 },
10200 CLASSIC | FLAG_SKB_FRAG,
10201 { [0x3e] = 0x25, [0x3f] = 0x05, },
10202 { {0x40, 0x25051982} },
10203 .frag_data = { 0x19, 0x82 },
10204 },
10205 {
10206 "LD_ABS byte frag",
10207 .u.insns = {
10208 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, 0x40),
10209 BPF_STMT(BPF_RET | BPF_A, 0x0),
10210 },
10211 CLASSIC | FLAG_SKB_FRAG,
10212 { },
10213 { {0x40, 0x42} },
10214 .frag_data = {
10215 0x42, 0x00, 0x00, 0x00,
10216 0x43, 0x44, 0x00, 0x00,
10217 0x21, 0x07, 0x19, 0x83,
10218 },
10219 },
10220 {
10221 "LD_ABS halfword frag",
10222 .u.insns = {
10223 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x44),
10224 BPF_STMT(BPF_RET | BPF_A, 0x0),
10225 },
10226 CLASSIC | FLAG_SKB_FRAG,
10227 { },
10228 { {0x40, 0x4344} },
10229 .frag_data = {
10230 0x42, 0x00, 0x00, 0x00,
10231 0x43, 0x44, 0x00, 0x00,
10232 0x21, 0x07, 0x19, 0x83,
10233 },
10234 },
10235 {
10236 "LD_ABS word frag",
10237 .u.insns = {
10238 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x48),
10239 BPF_STMT(BPF_RET | BPF_A, 0x0),
10240 },
10241 CLASSIC | FLAG_SKB_FRAG,
10242 { },
10243 { {0x40, 0x21071983} },
10244 .frag_data = {
10245 0x42, 0x00, 0x00, 0x00,
10246 0x43, 0x44, 0x00, 0x00,
10247 0x21, 0x07, 0x19, 0x83,
10248 },
10249 },
10250 {
10251 "LD_ABS halfword mixed head/frag",
10252 .u.insns = {
10253 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x3f),
10254 BPF_STMT(BPF_RET | BPF_A, 0x0),
10255 },
10256 CLASSIC | FLAG_SKB_FRAG,
10257 { [0x3e] = 0x25, [0x3f] = 0x05, },
10258 { {0x40, 0x0519} },
10259 .frag_data = { 0x19, 0x82 },
10260 },
10261 {
10262 "LD_ABS word mixed head/frag",
10263 .u.insns = {
10264 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x3e),
10265 BPF_STMT(BPF_RET | BPF_A, 0x0),
10266 },
10267 CLASSIC | FLAG_SKB_FRAG,
10268 { [0x3e] = 0x25, [0x3f] = 0x05, },
10269 { {0x40, 0x25051982} },
10270 .frag_data = { 0x19, 0x82 },
10271 },
Nicolas Schichan08fcb082015-08-04 15:19:11 +020010272 /*
10273 * LD_IND / LD_ABS on non fragmented SKBs
10274 */
10275 {
10276 /*
10277 * this tests that the JIT/interpreter correctly resets X
10278 * before using it in an LD_IND instruction.
10279 */
10280 "LD_IND byte default X",
10281 .u.insns = {
10282 BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x1),
10283 BPF_STMT(BPF_RET | BPF_A, 0x0),
10284 },
10285 CLASSIC,
10286 { [0x1] = 0x42 },
10287 { {0x40, 0x42 } },
10288 },
10289 {
10290 "LD_IND byte positive offset",
10291 .u.insns = {
10292 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10293 BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x1),
10294 BPF_STMT(BPF_RET | BPF_A, 0x0),
10295 },
10296 CLASSIC,
10297 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10298 { {0x40, 0x82 } },
10299 },
10300 {
10301 "LD_IND byte negative offset",
10302 .u.insns = {
10303 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10304 BPF_STMT(BPF_LD | BPF_IND | BPF_B, -0x1),
10305 BPF_STMT(BPF_RET | BPF_A, 0x0),
10306 },
10307 CLASSIC,
10308 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10309 { {0x40, 0x05 } },
10310 },
10311 {
Daniel Borkmann93731ef2018-05-04 01:08:13 +020010312 "LD_IND byte positive offset, all ff",
10313 .u.insns = {
10314 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10315 BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x1),
10316 BPF_STMT(BPF_RET | BPF_A, 0x0),
10317 },
10318 CLASSIC,
10319 { [0x3c] = 0xff, [0x3d] = 0xff, [0x3e] = 0xff, [0x3f] = 0xff },
10320 { {0x40, 0xff } },
10321 },
10322 {
10323 "LD_IND byte positive offset, out of bounds",
10324 .u.insns = {
10325 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10326 BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x1),
10327 BPF_STMT(BPF_RET | BPF_A, 0x0),
10328 },
10329 CLASSIC,
10330 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10331 { {0x3f, 0 }, },
10332 },
10333 {
10334 "LD_IND byte negative offset, out of bounds",
10335 .u.insns = {
10336 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10337 BPF_STMT(BPF_LD | BPF_IND | BPF_B, -0x3f),
10338 BPF_STMT(BPF_RET | BPF_A, 0x0),
10339 },
10340 CLASSIC,
10341 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10342 { {0x3f, 0 } },
10343 },
10344 {
10345 "LD_IND byte negative offset, multiple calls",
10346 .u.insns = {
10347 BPF_STMT(BPF_LDX | BPF_IMM, 0x3b),
10348 BPF_STMT(BPF_LD | BPF_IND | BPF_B, SKF_LL_OFF + 1),
10349 BPF_STMT(BPF_LD | BPF_IND | BPF_B, SKF_LL_OFF + 2),
10350 BPF_STMT(BPF_LD | BPF_IND | BPF_B, SKF_LL_OFF + 3),
10351 BPF_STMT(BPF_LD | BPF_IND | BPF_B, SKF_LL_OFF + 4),
10352 BPF_STMT(BPF_RET | BPF_A, 0x0),
10353 },
10354 CLASSIC,
10355 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10356 { {0x40, 0x82 }, },
10357 },
10358 {
Nicolas Schichan08fcb082015-08-04 15:19:11 +020010359 "LD_IND halfword positive offset",
10360 .u.insns = {
10361 BPF_STMT(BPF_LDX | BPF_IMM, 0x20),
10362 BPF_STMT(BPF_LD | BPF_IND | BPF_H, 0x2),
10363 BPF_STMT(BPF_RET | BPF_A, 0x0),
10364 },
10365 CLASSIC,
10366 {
10367 [0x1c] = 0xaa, [0x1d] = 0x55,
10368 [0x1e] = 0xbb, [0x1f] = 0x66,
10369 [0x20] = 0xcc, [0x21] = 0x77,
10370 [0x22] = 0xdd, [0x23] = 0x88,
10371 },
10372 { {0x40, 0xdd88 } },
10373 },
10374 {
10375 "LD_IND halfword negative offset",
10376 .u.insns = {
10377 BPF_STMT(BPF_LDX | BPF_IMM, 0x20),
10378 BPF_STMT(BPF_LD | BPF_IND | BPF_H, -0x2),
10379 BPF_STMT(BPF_RET | BPF_A, 0x0),
10380 },
10381 CLASSIC,
10382 {
10383 [0x1c] = 0xaa, [0x1d] = 0x55,
10384 [0x1e] = 0xbb, [0x1f] = 0x66,
10385 [0x20] = 0xcc, [0x21] = 0x77,
10386 [0x22] = 0xdd, [0x23] = 0x88,
10387 },
10388 { {0x40, 0xbb66 } },
10389 },
10390 {
10391 "LD_IND halfword unaligned",
10392 .u.insns = {
10393 BPF_STMT(BPF_LDX | BPF_IMM, 0x20),
10394 BPF_STMT(BPF_LD | BPF_IND | BPF_H, -0x1),
10395 BPF_STMT(BPF_RET | BPF_A, 0x0),
10396 },
10397 CLASSIC,
10398 {
10399 [0x1c] = 0xaa, [0x1d] = 0x55,
10400 [0x1e] = 0xbb, [0x1f] = 0x66,
10401 [0x20] = 0xcc, [0x21] = 0x77,
10402 [0x22] = 0xdd, [0x23] = 0x88,
10403 },
10404 { {0x40, 0x66cc } },
10405 },
10406 {
Daniel Borkmann93731ef2018-05-04 01:08:13 +020010407 "LD_IND halfword positive offset, all ff",
10408 .u.insns = {
10409 BPF_STMT(BPF_LDX | BPF_IMM, 0x3d),
10410 BPF_STMT(BPF_LD | BPF_IND | BPF_H, 0x1),
10411 BPF_STMT(BPF_RET | BPF_A, 0x0),
10412 },
10413 CLASSIC,
10414 { [0x3c] = 0xff, [0x3d] = 0xff, [0x3e] = 0xff, [0x3f] = 0xff },
10415 { {0x40, 0xffff } },
10416 },
10417 {
10418 "LD_IND halfword positive offset, out of bounds",
10419 .u.insns = {
10420 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10421 BPF_STMT(BPF_LD | BPF_IND | BPF_H, 0x1),
10422 BPF_STMT(BPF_RET | BPF_A, 0x0),
10423 },
10424 CLASSIC,
10425 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10426 { {0x3f, 0 }, },
10427 },
10428 {
10429 "LD_IND halfword negative offset, out of bounds",
10430 .u.insns = {
10431 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10432 BPF_STMT(BPF_LD | BPF_IND | BPF_H, -0x3f),
10433 BPF_STMT(BPF_RET | BPF_A, 0x0),
10434 },
10435 CLASSIC,
10436 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10437 { {0x3f, 0 } },
10438 },
10439 {
Nicolas Schichan08fcb082015-08-04 15:19:11 +020010440 "LD_IND word positive offset",
10441 .u.insns = {
10442 BPF_STMT(BPF_LDX | BPF_IMM, 0x20),
10443 BPF_STMT(BPF_LD | BPF_IND | BPF_W, 0x4),
10444 BPF_STMT(BPF_RET | BPF_A, 0x0),
10445 },
10446 CLASSIC,
10447 {
10448 [0x1c] = 0xaa, [0x1d] = 0x55,
10449 [0x1e] = 0xbb, [0x1f] = 0x66,
10450 [0x20] = 0xcc, [0x21] = 0x77,
10451 [0x22] = 0xdd, [0x23] = 0x88,
10452 [0x24] = 0xee, [0x25] = 0x99,
10453 [0x26] = 0xff, [0x27] = 0xaa,
10454 },
10455 { {0x40, 0xee99ffaa } },
10456 },
10457 {
10458 "LD_IND word negative offset",
10459 .u.insns = {
10460 BPF_STMT(BPF_LDX | BPF_IMM, 0x20),
10461 BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x4),
10462 BPF_STMT(BPF_RET | BPF_A, 0x0),
10463 },
10464 CLASSIC,
10465 {
10466 [0x1c] = 0xaa, [0x1d] = 0x55,
10467 [0x1e] = 0xbb, [0x1f] = 0x66,
10468 [0x20] = 0xcc, [0x21] = 0x77,
10469 [0x22] = 0xdd, [0x23] = 0x88,
10470 [0x24] = 0xee, [0x25] = 0x99,
10471 [0x26] = 0xff, [0x27] = 0xaa,
10472 },
10473 { {0x40, 0xaa55bb66 } },
10474 },
10475 {
10476 "LD_IND word unaligned (addr & 3 == 2)",
10477 .u.insns = {
10478 BPF_STMT(BPF_LDX | BPF_IMM, 0x20),
10479 BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x2),
10480 BPF_STMT(BPF_RET | BPF_A, 0x0),
10481 },
10482 CLASSIC,
10483 {
10484 [0x1c] = 0xaa, [0x1d] = 0x55,
10485 [0x1e] = 0xbb, [0x1f] = 0x66,
10486 [0x20] = 0xcc, [0x21] = 0x77,
10487 [0x22] = 0xdd, [0x23] = 0x88,
10488 [0x24] = 0xee, [0x25] = 0x99,
10489 [0x26] = 0xff, [0x27] = 0xaa,
10490 },
10491 { {0x40, 0xbb66cc77 } },
10492 },
10493 {
10494 "LD_IND word unaligned (addr & 3 == 1)",
10495 .u.insns = {
10496 BPF_STMT(BPF_LDX | BPF_IMM, 0x20),
10497 BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x3),
10498 BPF_STMT(BPF_RET | BPF_A, 0x0),
10499 },
10500 CLASSIC,
10501 {
10502 [0x1c] = 0xaa, [0x1d] = 0x55,
10503 [0x1e] = 0xbb, [0x1f] = 0x66,
10504 [0x20] = 0xcc, [0x21] = 0x77,
10505 [0x22] = 0xdd, [0x23] = 0x88,
10506 [0x24] = 0xee, [0x25] = 0x99,
10507 [0x26] = 0xff, [0x27] = 0xaa,
10508 },
10509 { {0x40, 0x55bb66cc } },
10510 },
10511 {
10512 "LD_IND word unaligned (addr & 3 == 3)",
10513 .u.insns = {
10514 BPF_STMT(BPF_LDX | BPF_IMM, 0x20),
10515 BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x1),
10516 BPF_STMT(BPF_RET | BPF_A, 0x0),
10517 },
10518 CLASSIC,
10519 {
10520 [0x1c] = 0xaa, [0x1d] = 0x55,
10521 [0x1e] = 0xbb, [0x1f] = 0x66,
10522 [0x20] = 0xcc, [0x21] = 0x77,
10523 [0x22] = 0xdd, [0x23] = 0x88,
10524 [0x24] = 0xee, [0x25] = 0x99,
10525 [0x26] = 0xff, [0x27] = 0xaa,
10526 },
10527 { {0x40, 0x66cc77dd } },
10528 },
10529 {
Daniel Borkmann93731ef2018-05-04 01:08:13 +020010530 "LD_IND word positive offset, all ff",
10531 .u.insns = {
10532 BPF_STMT(BPF_LDX | BPF_IMM, 0x3b),
10533 BPF_STMT(BPF_LD | BPF_IND | BPF_W, 0x1),
10534 BPF_STMT(BPF_RET | BPF_A, 0x0),
10535 },
10536 CLASSIC,
10537 { [0x3c] = 0xff, [0x3d] = 0xff, [0x3e] = 0xff, [0x3f] = 0xff },
10538 { {0x40, 0xffffffff } },
10539 },
10540 {
10541 "LD_IND word positive offset, out of bounds",
10542 .u.insns = {
10543 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10544 BPF_STMT(BPF_LD | BPF_IND | BPF_W, 0x1),
10545 BPF_STMT(BPF_RET | BPF_A, 0x0),
10546 },
10547 CLASSIC,
10548 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10549 { {0x3f, 0 }, },
10550 },
10551 {
10552 "LD_IND word negative offset, out of bounds",
10553 .u.insns = {
10554 BPF_STMT(BPF_LDX | BPF_IMM, 0x3e),
10555 BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x3f),
10556 BPF_STMT(BPF_RET | BPF_A, 0x0),
10557 },
10558 CLASSIC,
10559 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10560 { {0x3f, 0 } },
10561 },
10562 {
Nicolas Schichan08fcb082015-08-04 15:19:11 +020010563 "LD_ABS byte",
10564 .u.insns = {
10565 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, 0x20),
10566 BPF_STMT(BPF_RET | BPF_A, 0x0),
10567 },
10568 CLASSIC,
10569 {
10570 [0x1c] = 0xaa, [0x1d] = 0x55,
10571 [0x1e] = 0xbb, [0x1f] = 0x66,
10572 [0x20] = 0xcc, [0x21] = 0x77,
10573 [0x22] = 0xdd, [0x23] = 0x88,
10574 [0x24] = 0xee, [0x25] = 0x99,
10575 [0x26] = 0xff, [0x27] = 0xaa,
10576 },
10577 { {0x40, 0xcc } },
10578 },
10579 {
Daniel Borkmann93731ef2018-05-04 01:08:13 +020010580 "LD_ABS byte positive offset, all ff",
10581 .u.insns = {
10582 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, 0x3f),
10583 BPF_STMT(BPF_RET | BPF_A, 0x0),
10584 },
10585 CLASSIC,
10586 { [0x3c] = 0xff, [0x3d] = 0xff, [0x3e] = 0xff, [0x3f] = 0xff },
10587 { {0x40, 0xff } },
10588 },
10589 {
10590 "LD_ABS byte positive offset, out of bounds",
10591 .u.insns = {
10592 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, 0x3f),
10593 BPF_STMT(BPF_RET | BPF_A, 0x0),
10594 },
10595 CLASSIC,
10596 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10597 { {0x3f, 0 }, },
10598 },
10599 {
10600 "LD_ABS byte negative offset, out of bounds load",
10601 .u.insns = {
10602 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, -1),
10603 BPF_STMT(BPF_RET | BPF_A, 0x0),
10604 },
10605 CLASSIC | FLAG_EXPECTED_FAIL,
10606 .expected_errcode = -EINVAL,
10607 },
10608 {
10609 "LD_ABS byte negative offset, in bounds",
10610 .u.insns = {
10611 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3f),
10612 BPF_STMT(BPF_RET | BPF_A, 0x0),
10613 },
10614 CLASSIC,
10615 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10616 { {0x40, 0x82 }, },
10617 },
10618 {
10619 "LD_ABS byte negative offset, out of bounds",
10620 .u.insns = {
10621 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3f),
10622 BPF_STMT(BPF_RET | BPF_A, 0x0),
10623 },
10624 CLASSIC,
10625 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10626 { {0x3f, 0 }, },
10627 },
10628 {
10629 "LD_ABS byte negative offset, multiple calls",
10630 .u.insns = {
10631 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3c),
10632 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3d),
10633 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3e),
10634 BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3f),
10635 BPF_STMT(BPF_RET | BPF_A, 0x0),
10636 },
10637 CLASSIC,
10638 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10639 { {0x40, 0x82 }, },
10640 },
10641 {
Nicolas Schichan08fcb082015-08-04 15:19:11 +020010642 "LD_ABS halfword",
10643 .u.insns = {
10644 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x22),
10645 BPF_STMT(BPF_RET | BPF_A, 0x0),
10646 },
10647 CLASSIC,
10648 {
10649 [0x1c] = 0xaa, [0x1d] = 0x55,
10650 [0x1e] = 0xbb, [0x1f] = 0x66,
10651 [0x20] = 0xcc, [0x21] = 0x77,
10652 [0x22] = 0xdd, [0x23] = 0x88,
10653 [0x24] = 0xee, [0x25] = 0x99,
10654 [0x26] = 0xff, [0x27] = 0xaa,
10655 },
10656 { {0x40, 0xdd88 } },
10657 },
10658 {
10659 "LD_ABS halfword unaligned",
10660 .u.insns = {
10661 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x25),
10662 BPF_STMT(BPF_RET | BPF_A, 0x0),
10663 },
10664 CLASSIC,
10665 {
10666 [0x1c] = 0xaa, [0x1d] = 0x55,
10667 [0x1e] = 0xbb, [0x1f] = 0x66,
10668 [0x20] = 0xcc, [0x21] = 0x77,
10669 [0x22] = 0xdd, [0x23] = 0x88,
10670 [0x24] = 0xee, [0x25] = 0x99,
10671 [0x26] = 0xff, [0x27] = 0xaa,
10672 },
10673 { {0x40, 0x99ff } },
10674 },
10675 {
Daniel Borkmann93731ef2018-05-04 01:08:13 +020010676 "LD_ABS halfword positive offset, all ff",
10677 .u.insns = {
10678 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x3e),
10679 BPF_STMT(BPF_RET | BPF_A, 0x0),
10680 },
10681 CLASSIC,
10682 { [0x3c] = 0xff, [0x3d] = 0xff, [0x3e] = 0xff, [0x3f] = 0xff },
10683 { {0x40, 0xffff } },
10684 },
10685 {
10686 "LD_ABS halfword positive offset, out of bounds",
10687 .u.insns = {
10688 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x3f),
10689 BPF_STMT(BPF_RET | BPF_A, 0x0),
10690 },
10691 CLASSIC,
10692 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10693 { {0x3f, 0 }, },
10694 },
10695 {
10696 "LD_ABS halfword negative offset, out of bounds load",
10697 .u.insns = {
10698 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, -1),
10699 BPF_STMT(BPF_RET | BPF_A, 0x0),
10700 },
10701 CLASSIC | FLAG_EXPECTED_FAIL,
10702 .expected_errcode = -EINVAL,
10703 },
10704 {
10705 "LD_ABS halfword negative offset, in bounds",
10706 .u.insns = {
10707 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, SKF_LL_OFF + 0x3e),
10708 BPF_STMT(BPF_RET | BPF_A, 0x0),
10709 },
10710 CLASSIC,
10711 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10712 { {0x40, 0x1982 }, },
10713 },
10714 {
10715 "LD_ABS halfword negative offset, out of bounds",
10716 .u.insns = {
10717 BPF_STMT(BPF_LD | BPF_ABS | BPF_H, SKF_LL_OFF + 0x3e),
10718 BPF_STMT(BPF_RET | BPF_A, 0x0),
10719 },
10720 CLASSIC,
10721 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10722 { {0x3f, 0 }, },
10723 },
10724 {
Nicolas Schichan08fcb082015-08-04 15:19:11 +020010725 "LD_ABS word",
10726 .u.insns = {
10727 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x1c),
10728 BPF_STMT(BPF_RET | BPF_A, 0x0),
10729 },
10730 CLASSIC,
10731 {
10732 [0x1c] = 0xaa, [0x1d] = 0x55,
10733 [0x1e] = 0xbb, [0x1f] = 0x66,
10734 [0x20] = 0xcc, [0x21] = 0x77,
10735 [0x22] = 0xdd, [0x23] = 0x88,
10736 [0x24] = 0xee, [0x25] = 0x99,
10737 [0x26] = 0xff, [0x27] = 0xaa,
10738 },
10739 { {0x40, 0xaa55bb66 } },
10740 },
10741 {
10742 "LD_ABS word unaligned (addr & 3 == 2)",
10743 .u.insns = {
10744 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x22),
10745 BPF_STMT(BPF_RET | BPF_A, 0x0),
10746 },
10747 CLASSIC,
10748 {
10749 [0x1c] = 0xaa, [0x1d] = 0x55,
10750 [0x1e] = 0xbb, [0x1f] = 0x66,
10751 [0x20] = 0xcc, [0x21] = 0x77,
10752 [0x22] = 0xdd, [0x23] = 0x88,
10753 [0x24] = 0xee, [0x25] = 0x99,
10754 [0x26] = 0xff, [0x27] = 0xaa,
10755 },
10756 { {0x40, 0xdd88ee99 } },
10757 },
10758 {
10759 "LD_ABS word unaligned (addr & 3 == 1)",
10760 .u.insns = {
10761 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x21),
10762 BPF_STMT(BPF_RET | BPF_A, 0x0),
10763 },
10764 CLASSIC,
10765 {
10766 [0x1c] = 0xaa, [0x1d] = 0x55,
10767 [0x1e] = 0xbb, [0x1f] = 0x66,
10768 [0x20] = 0xcc, [0x21] = 0x77,
10769 [0x22] = 0xdd, [0x23] = 0x88,
10770 [0x24] = 0xee, [0x25] = 0x99,
10771 [0x26] = 0xff, [0x27] = 0xaa,
10772 },
10773 { {0x40, 0x77dd88ee } },
10774 },
10775 {
10776 "LD_ABS word unaligned (addr & 3 == 3)",
10777 .u.insns = {
10778 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x23),
10779 BPF_STMT(BPF_RET | BPF_A, 0x0),
10780 },
10781 CLASSIC,
10782 {
10783 [0x1c] = 0xaa, [0x1d] = 0x55,
10784 [0x1e] = 0xbb, [0x1f] = 0x66,
10785 [0x20] = 0xcc, [0x21] = 0x77,
10786 [0x22] = 0xdd, [0x23] = 0x88,
10787 [0x24] = 0xee, [0x25] = 0x99,
10788 [0x26] = 0xff, [0x27] = 0xaa,
10789 },
10790 { {0x40, 0x88ee99ff } },
10791 },
Daniel Borkmann93731ef2018-05-04 01:08:13 +020010792 {
10793 "LD_ABS word positive offset, all ff",
10794 .u.insns = {
10795 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x3c),
10796 BPF_STMT(BPF_RET | BPF_A, 0x0),
10797 },
10798 CLASSIC,
10799 { [0x3c] = 0xff, [0x3d] = 0xff, [0x3e] = 0xff, [0x3f] = 0xff },
10800 { {0x40, 0xffffffff } },
10801 },
10802 {
10803 "LD_ABS word positive offset, out of bounds",
10804 .u.insns = {
10805 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x3f),
10806 BPF_STMT(BPF_RET | BPF_A, 0x0),
10807 },
10808 CLASSIC,
10809 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10810 { {0x3f, 0 }, },
10811 },
10812 {
10813 "LD_ABS word negative offset, out of bounds load",
10814 .u.insns = {
10815 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, -1),
10816 BPF_STMT(BPF_RET | BPF_A, 0x0),
10817 },
10818 CLASSIC | FLAG_EXPECTED_FAIL,
10819 .expected_errcode = -EINVAL,
10820 },
10821 {
10822 "LD_ABS word negative offset, in bounds",
10823 .u.insns = {
10824 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, SKF_LL_OFF + 0x3c),
10825 BPF_STMT(BPF_RET | BPF_A, 0x0),
10826 },
10827 CLASSIC,
10828 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10829 { {0x40, 0x25051982 }, },
10830 },
10831 {
10832 "LD_ABS word negative offset, out of bounds",
10833 .u.insns = {
10834 BPF_STMT(BPF_LD | BPF_ABS | BPF_W, SKF_LL_OFF + 0x3c),
10835 BPF_STMT(BPF_RET | BPF_A, 0x0),
10836 },
10837 CLASSIC,
10838 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10839 { {0x3f, 0 }, },
10840 },
10841 {
10842 "LDX_MSH standalone, preserved A",
10843 .u.insns = {
10844 BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa),
10845 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3c),
10846 BPF_STMT(BPF_RET | BPF_A, 0x0),
10847 },
10848 CLASSIC,
10849 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10850 { {0x40, 0xffeebbaa }, },
10851 },
10852 {
10853 "LDX_MSH standalone, preserved A 2",
10854 .u.insns = {
10855 BPF_STMT(BPF_LD | BPF_IMM, 0x175e9d63),
10856 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3c),
10857 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3d),
10858 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3e),
10859 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3f),
10860 BPF_STMT(BPF_RET | BPF_A, 0x0),
10861 },
10862 CLASSIC,
10863 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10864 { {0x40, 0x175e9d63 }, },
10865 },
10866 {
10867 "LDX_MSH standalone, test result 1",
10868 .u.insns = {
10869 BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa),
10870 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3c),
10871 BPF_STMT(BPF_MISC | BPF_TXA, 0),
10872 BPF_STMT(BPF_RET | BPF_A, 0x0),
10873 },
10874 CLASSIC,
10875 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10876 { {0x40, 0x14 }, },
10877 },
10878 {
10879 "LDX_MSH standalone, test result 2",
10880 .u.insns = {
10881 BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa),
10882 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3e),
10883 BPF_STMT(BPF_MISC | BPF_TXA, 0),
10884 BPF_STMT(BPF_RET | BPF_A, 0x0),
10885 },
10886 CLASSIC,
10887 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10888 { {0x40, 0x24 }, },
10889 },
10890 {
10891 "LDX_MSH standalone, negative offset",
10892 .u.insns = {
10893 BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa),
10894 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, -1),
10895 BPF_STMT(BPF_MISC | BPF_TXA, 0),
10896 BPF_STMT(BPF_RET | BPF_A, 0x0),
10897 },
10898 CLASSIC,
10899 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10900 { {0x40, 0 }, },
10901 },
10902 {
10903 "LDX_MSH standalone, negative offset 2",
10904 .u.insns = {
10905 BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa),
10906 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, SKF_LL_OFF + 0x3e),
10907 BPF_STMT(BPF_MISC | BPF_TXA, 0),
10908 BPF_STMT(BPF_RET | BPF_A, 0x0),
10909 },
10910 CLASSIC,
10911 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10912 { {0x40, 0x24 }, },
10913 },
10914 {
10915 "LDX_MSH standalone, out of bounds",
10916 .u.insns = {
10917 BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa),
10918 BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x40),
10919 BPF_STMT(BPF_MISC | BPF_TXA, 0),
10920 BPF_STMT(BPF_RET | BPF_A, 0x0),
10921 },
10922 CLASSIC,
10923 { [0x3c] = 0x25, [0x3d] = 0x05, [0x3e] = 0x19, [0x3f] = 0x82 },
10924 { {0x40, 0 }, },
10925 },
Nicolas Schichan86bf1722015-08-04 15:19:12 +020010926 /*
10927 * verify that the interpreter or JIT correctly sets A and X
10928 * to 0.
10929 */
10930 {
10931 "ADD default X",
10932 .u.insns = {
10933 /*
10934 * A = 0x42
10935 * A = A + X
10936 * ret A
10937 */
10938 BPF_STMT(BPF_LD | BPF_IMM, 0x42),
10939 BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0),
10940 BPF_STMT(BPF_RET | BPF_A, 0x0),
10941 },
10942 CLASSIC | FLAG_NO_DATA,
10943 {},
10944 { {0x1, 0x42 } },
10945 },
10946 {
10947 "ADD default A",
10948 .u.insns = {
10949 /*
10950 * A = A + 0x42
10951 * ret A
10952 */
10953 BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 0x42),
10954 BPF_STMT(BPF_RET | BPF_A, 0x0),
10955 },
10956 CLASSIC | FLAG_NO_DATA,
10957 {},
10958 { {0x1, 0x42 } },
10959 },
10960 {
10961 "SUB default X",
10962 .u.insns = {
10963 /*
10964 * A = 0x66
10965 * A = A - X
10966 * ret A
10967 */
10968 BPF_STMT(BPF_LD | BPF_IMM, 0x66),
10969 BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0),
10970 BPF_STMT(BPF_RET | BPF_A, 0x0),
10971 },
10972 CLASSIC | FLAG_NO_DATA,
10973 {},
10974 { {0x1, 0x66 } },
10975 },
10976 {
10977 "SUB default A",
10978 .u.insns = {
10979 /*
10980 * A = A - -0x66
10981 * ret A
10982 */
10983 BPF_STMT(BPF_ALU | BPF_SUB | BPF_K, -0x66),
10984 BPF_STMT(BPF_RET | BPF_A, 0x0),
10985 },
10986 CLASSIC | FLAG_NO_DATA,
10987 {},
10988 { {0x1, 0x66 } },
10989 },
10990 {
10991 "MUL default X",
10992 .u.insns = {
10993 /*
10994 * A = 0x42
10995 * A = A * X
10996 * ret A
10997 */
10998 BPF_STMT(BPF_LD | BPF_IMM, 0x42),
10999 BPF_STMT(BPF_ALU | BPF_MUL | BPF_X, 0),
11000 BPF_STMT(BPF_RET | BPF_A, 0x0),
11001 },
11002 CLASSIC | FLAG_NO_DATA,
11003 {},
11004 { {0x1, 0x0 } },
11005 },
11006 {
11007 "MUL default A",
11008 .u.insns = {
11009 /*
11010 * A = A * 0x66
11011 * ret A
11012 */
11013 BPF_STMT(BPF_ALU | BPF_MUL | BPF_K, 0x66),
11014 BPF_STMT(BPF_RET | BPF_A, 0x0),
11015 },
11016 CLASSIC | FLAG_NO_DATA,
11017 {},
11018 { {0x1, 0x0 } },
11019 },
11020 {
11021 "DIV default X",
11022 .u.insns = {
11023 /*
11024 * A = 0x42
11025 * A = A / X ; this halt the filter execution if X is 0
11026 * ret 0x42
11027 */
11028 BPF_STMT(BPF_LD | BPF_IMM, 0x42),
11029 BPF_STMT(BPF_ALU | BPF_DIV | BPF_X, 0),
11030 BPF_STMT(BPF_RET | BPF_K, 0x42),
11031 },
11032 CLASSIC | FLAG_NO_DATA,
11033 {},
11034 { {0x1, 0x0 } },
11035 },
11036 {
11037 "DIV default A",
11038 .u.insns = {
11039 /*
11040 * A = A / 1
11041 * ret A
11042 */
11043 BPF_STMT(BPF_ALU | BPF_DIV | BPF_K, 0x1),
11044 BPF_STMT(BPF_RET | BPF_A, 0x0),
11045 },
11046 CLASSIC | FLAG_NO_DATA,
11047 {},
11048 { {0x1, 0x0 } },
11049 },
11050 {
Yang Shid4e4bc12015-11-04 11:36:37 -080011051 "MOD default X",
11052 .u.insns = {
11053 /*
11054 * A = 0x42
11055 * A = A mod X ; this halt the filter execution if X is 0
11056 * ret 0x42
11057 */
11058 BPF_STMT(BPF_LD | BPF_IMM, 0x42),
11059 BPF_STMT(BPF_ALU | BPF_MOD | BPF_X, 0),
11060 BPF_STMT(BPF_RET | BPF_K, 0x42),
11061 },
11062 CLASSIC | FLAG_NO_DATA,
11063 {},
11064 { {0x1, 0x0 } },
11065 },
11066 {
11067 "MOD default A",
11068 .u.insns = {
11069 /*
11070 * A = A mod 1
11071 * ret A
11072 */
11073 BPF_STMT(BPF_ALU | BPF_MOD | BPF_K, 0x1),
11074 BPF_STMT(BPF_RET | BPF_A, 0x0),
11075 },
11076 CLASSIC | FLAG_NO_DATA,
11077 {},
11078 { {0x1, 0x0 } },
11079 },
11080 {
Nicolas Schichan86bf1722015-08-04 15:19:12 +020011081 "JMP EQ default A",
11082 .u.insns = {
11083 /*
11084 * cmp A, 0x0, 0, 1
11085 * ret 0x42
11086 * ret 0x66
11087 */
11088 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x0, 0, 1),
11089 BPF_STMT(BPF_RET | BPF_K, 0x42),
11090 BPF_STMT(BPF_RET | BPF_K, 0x66),
11091 },
11092 CLASSIC | FLAG_NO_DATA,
11093 {},
11094 { {0x1, 0x42 } },
11095 },
11096 {
11097 "JMP EQ default X",
11098 .u.insns = {
11099 /*
11100 * A = 0x0
11101 * cmp A, X, 0, 1
11102 * ret 0x42
11103 * ret 0x66
11104 */
11105 BPF_STMT(BPF_LD | BPF_IMM, 0x0),
11106 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_X, 0x0, 0, 1),
11107 BPF_STMT(BPF_RET | BPF_K, 0x42),
11108 BPF_STMT(BPF_RET | BPF_K, 0x66),
11109 },
11110 CLASSIC | FLAG_NO_DATA,
11111 {},
11112 { {0x1, 0x42 } },
11113 },
Daniel Borkmannfcd1c912018-01-20 01:24:31 +010011114 /* Checking interpreter vs JIT wrt signed extended imms. */
11115 {
11116 "JNE signed compare, test 1",
11117 .u.insns_int = {
11118 BPF_ALU32_IMM(BPF_MOV, R1, 0xfefbbc12),
11119 BPF_ALU32_IMM(BPF_MOV, R3, 0xffff0000),
11120 BPF_MOV64_REG(R2, R1),
11121 BPF_ALU64_REG(BPF_AND, R2, R3),
11122 BPF_ALU32_IMM(BPF_MOV, R0, 1),
11123 BPF_JMP_IMM(BPF_JNE, R2, -17104896, 1),
11124 BPF_ALU32_IMM(BPF_MOV, R0, 2),
11125 BPF_EXIT_INSN(),
11126 },
11127 INTERNAL,
11128 { },
11129 { { 0, 1 } },
11130 },
11131 {
11132 "JNE signed compare, test 2",
11133 .u.insns_int = {
11134 BPF_ALU32_IMM(BPF_MOV, R1, 0xfefbbc12),
11135 BPF_ALU32_IMM(BPF_MOV, R3, 0xffff0000),
11136 BPF_MOV64_REG(R2, R1),
11137 BPF_ALU64_REG(BPF_AND, R2, R3),
11138 BPF_ALU32_IMM(BPF_MOV, R0, 1),
11139 BPF_JMP_IMM(BPF_JNE, R2, 0xfefb0000, 1),
11140 BPF_ALU32_IMM(BPF_MOV, R0, 2),
11141 BPF_EXIT_INSN(),
11142 },
11143 INTERNAL,
11144 { },
11145 { { 0, 1 } },
11146 },
11147 {
11148 "JNE signed compare, test 3",
11149 .u.insns_int = {
11150 BPF_ALU32_IMM(BPF_MOV, R1, 0xfefbbc12),
11151 BPF_ALU32_IMM(BPF_MOV, R3, 0xffff0000),
11152 BPF_ALU32_IMM(BPF_MOV, R4, 0xfefb0000),
11153 BPF_MOV64_REG(R2, R1),
11154 BPF_ALU64_REG(BPF_AND, R2, R3),
11155 BPF_ALU32_IMM(BPF_MOV, R0, 1),
11156 BPF_JMP_REG(BPF_JNE, R2, R4, 1),
11157 BPF_ALU32_IMM(BPF_MOV, R0, 2),
11158 BPF_EXIT_INSN(),
11159 },
11160 INTERNAL,
11161 { },
11162 { { 0, 2 } },
11163 },
11164 {
11165 "JNE signed compare, test 4",
11166 .u.insns_int = {
11167 BPF_LD_IMM64(R1, -17104896),
11168 BPF_ALU32_IMM(BPF_MOV, R0, 1),
11169 BPF_JMP_IMM(BPF_JNE, R1, -17104896, 1),
11170 BPF_ALU32_IMM(BPF_MOV, R0, 2),
11171 BPF_EXIT_INSN(),
11172 },
11173 INTERNAL,
11174 { },
11175 { { 0, 2 } },
11176 },
11177 {
11178 "JNE signed compare, test 5",
11179 .u.insns_int = {
11180 BPF_LD_IMM64(R1, 0xfefb0000),
11181 BPF_ALU32_IMM(BPF_MOV, R0, 1),
11182 BPF_JMP_IMM(BPF_JNE, R1, 0xfefb0000, 1),
11183 BPF_ALU32_IMM(BPF_MOV, R0, 2),
11184 BPF_EXIT_INSN(),
11185 },
11186 INTERNAL,
11187 { },
11188 { { 0, 1 } },
11189 },
11190 {
11191 "JNE signed compare, test 6",
11192 .u.insns_int = {
11193 BPF_LD_IMM64(R1, 0x7efb0000),
11194 BPF_ALU32_IMM(BPF_MOV, R0, 1),
11195 BPF_JMP_IMM(BPF_JNE, R1, 0x7efb0000, 1),
11196 BPF_ALU32_IMM(BPF_MOV, R0, 2),
11197 BPF_EXIT_INSN(),
11198 },
11199 INTERNAL,
11200 { },
11201 { { 0, 2 } },
11202 },
11203 {
11204 "JNE signed compare, test 7",
11205 .u.insns = {
11206 BPF_STMT(BPF_LD | BPF_IMM, 0xffff0000),
11207 BPF_STMT(BPF_MISC | BPF_TAX, 0),
11208 BPF_STMT(BPF_LD | BPF_IMM, 0xfefbbc12),
11209 BPF_STMT(BPF_ALU | BPF_AND | BPF_X, 0),
11210 BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0xfefb0000, 1, 0),
11211 BPF_STMT(BPF_RET | BPF_K, 1),
11212 BPF_STMT(BPF_RET | BPF_K, 2),
11213 },
11214 CLASSIC | FLAG_NO_DATA,
11215 {},
11216 { { 0, 2 } },
11217 },
Johan Almbladh7bceeb92021-10-01 15:03:48 +020011218 /* BPF_LDX_MEM with operand aliasing */
11219 {
11220 "LDX_MEM_B: operand register aliasing",
11221 .u.insns_int = {
11222 BPF_ST_MEM(BPF_B, R10, -8, 123),
11223 BPF_MOV64_REG(R0, R10),
11224 BPF_LDX_MEM(BPF_B, R0, R0, -8),
11225 BPF_EXIT_INSN(),
11226 },
11227 INTERNAL,
11228 { },
11229 { { 0, 123 } },
11230 .stack_depth = 8,
11231 },
11232 {
11233 "LDX_MEM_H: operand register aliasing",
11234 .u.insns_int = {
11235 BPF_ST_MEM(BPF_H, R10, -8, 12345),
11236 BPF_MOV64_REG(R0, R10),
11237 BPF_LDX_MEM(BPF_H, R0, R0, -8),
11238 BPF_EXIT_INSN(),
11239 },
11240 INTERNAL,
11241 { },
11242 { { 0, 12345 } },
11243 .stack_depth = 8,
11244 },
11245 {
11246 "LDX_MEM_W: operand register aliasing",
11247 .u.insns_int = {
11248 BPF_ST_MEM(BPF_W, R10, -8, 123456789),
11249 BPF_MOV64_REG(R0, R10),
11250 BPF_LDX_MEM(BPF_W, R0, R0, -8),
11251 BPF_EXIT_INSN(),
11252 },
11253 INTERNAL,
11254 { },
11255 { { 0, 123456789 } },
11256 .stack_depth = 8,
11257 },
11258 {
11259 "LDX_MEM_DW: operand register aliasing",
11260 .u.insns_int = {
11261 BPF_LD_IMM64(R1, 0x123456789abcdefULL),
11262 BPF_STX_MEM(BPF_DW, R10, R1, -8),
11263 BPF_MOV64_REG(R0, R10),
11264 BPF_LDX_MEM(BPF_DW, R0, R0, -8),
11265 BPF_ALU64_REG(BPF_SUB, R0, R1),
11266 BPF_MOV64_REG(R1, R0),
11267 BPF_ALU64_IMM(BPF_RSH, R1, 32),
11268 BPF_ALU64_REG(BPF_OR, R0, R1),
11269 BPF_EXIT_INSN(),
11270 },
11271 INTERNAL,
11272 { },
11273 { { 0, 0 } },
11274 .stack_depth = 8,
11275 },
Johan Almbladhe2f97972021-10-01 15:03:43 +020011276 /*
11277 * Register (non-)clobbering tests for the case where a JIT implements
11278 * complex ALU or ATOMIC operations via function calls. If so, the
11279 * function call must be transparent to the eBPF registers. The JIT
11280 * must therefore save and restore relevant registers across the call.
11281 * The following tests check that the eBPF registers retain their
11282 * values after such an operation. Mainly intended for complex ALU
11283 * and atomic operation, but we run it for all. You never know...
11284 *
11285 * Note that each operations should be tested twice with different
11286 * destinations, to check preservation for all registers.
11287 */
11288#define BPF_TEST_CLOBBER_ALU(alu, op, dst, src) \
11289 { \
11290 #alu "_" #op " to " #dst ": no clobbering", \
11291 .u.insns_int = { \
11292 BPF_ALU64_IMM(BPF_MOV, R0, R0), \
11293 BPF_ALU64_IMM(BPF_MOV, R1, R1), \
11294 BPF_ALU64_IMM(BPF_MOV, R2, R2), \
11295 BPF_ALU64_IMM(BPF_MOV, R3, R3), \
11296 BPF_ALU64_IMM(BPF_MOV, R4, R4), \
11297 BPF_ALU64_IMM(BPF_MOV, R5, R5), \
11298 BPF_ALU64_IMM(BPF_MOV, R6, R6), \
11299 BPF_ALU64_IMM(BPF_MOV, R7, R7), \
11300 BPF_ALU64_IMM(BPF_MOV, R8, R8), \
11301 BPF_ALU64_IMM(BPF_MOV, R9, R9), \
11302 BPF_##alu(BPF_ ##op, dst, src), \
11303 BPF_ALU32_IMM(BPF_MOV, dst, dst), \
11304 BPF_JMP_IMM(BPF_JNE, R0, R0, 10), \
11305 BPF_JMP_IMM(BPF_JNE, R1, R1, 9), \
11306 BPF_JMP_IMM(BPF_JNE, R2, R2, 8), \
11307 BPF_JMP_IMM(BPF_JNE, R3, R3, 7), \
11308 BPF_JMP_IMM(BPF_JNE, R4, R4, 6), \
11309 BPF_JMP_IMM(BPF_JNE, R5, R5, 5), \
11310 BPF_JMP_IMM(BPF_JNE, R6, R6, 4), \
11311 BPF_JMP_IMM(BPF_JNE, R7, R7, 3), \
11312 BPF_JMP_IMM(BPF_JNE, R8, R8, 2), \
11313 BPF_JMP_IMM(BPF_JNE, R9, R9, 1), \
11314 BPF_ALU64_IMM(BPF_MOV, R0, 1), \
11315 BPF_EXIT_INSN(), \
11316 }, \
11317 INTERNAL, \
11318 { }, \
11319 { { 0, 1 } } \
11320 }
11321 /* ALU64 operations, register clobbering */
11322 BPF_TEST_CLOBBER_ALU(ALU64_IMM, AND, R8, 123456789),
11323 BPF_TEST_CLOBBER_ALU(ALU64_IMM, AND, R9, 123456789),
11324 BPF_TEST_CLOBBER_ALU(ALU64_IMM, OR, R8, 123456789),
11325 BPF_TEST_CLOBBER_ALU(ALU64_IMM, OR, R9, 123456789),
11326 BPF_TEST_CLOBBER_ALU(ALU64_IMM, XOR, R8, 123456789),
11327 BPF_TEST_CLOBBER_ALU(ALU64_IMM, XOR, R9, 123456789),
11328 BPF_TEST_CLOBBER_ALU(ALU64_IMM, LSH, R8, 12),
11329 BPF_TEST_CLOBBER_ALU(ALU64_IMM, LSH, R9, 12),
11330 BPF_TEST_CLOBBER_ALU(ALU64_IMM, RSH, R8, 12),
11331 BPF_TEST_CLOBBER_ALU(ALU64_IMM, RSH, R9, 12),
11332 BPF_TEST_CLOBBER_ALU(ALU64_IMM, ARSH, R8, 12),
11333 BPF_TEST_CLOBBER_ALU(ALU64_IMM, ARSH, R9, 12),
11334 BPF_TEST_CLOBBER_ALU(ALU64_IMM, ADD, R8, 123456789),
11335 BPF_TEST_CLOBBER_ALU(ALU64_IMM, ADD, R9, 123456789),
11336 BPF_TEST_CLOBBER_ALU(ALU64_IMM, SUB, R8, 123456789),
11337 BPF_TEST_CLOBBER_ALU(ALU64_IMM, SUB, R9, 123456789),
11338 BPF_TEST_CLOBBER_ALU(ALU64_IMM, MUL, R8, 123456789),
11339 BPF_TEST_CLOBBER_ALU(ALU64_IMM, MUL, R9, 123456789),
11340 BPF_TEST_CLOBBER_ALU(ALU64_IMM, DIV, R8, 123456789),
11341 BPF_TEST_CLOBBER_ALU(ALU64_IMM, DIV, R9, 123456789),
11342 BPF_TEST_CLOBBER_ALU(ALU64_IMM, MOD, R8, 123456789),
11343 BPF_TEST_CLOBBER_ALU(ALU64_IMM, MOD, R9, 123456789),
11344 /* ALU32 immediate operations, register clobbering */
11345 BPF_TEST_CLOBBER_ALU(ALU32_IMM, AND, R8, 123456789),
11346 BPF_TEST_CLOBBER_ALU(ALU32_IMM, AND, R9, 123456789),
11347 BPF_TEST_CLOBBER_ALU(ALU32_IMM, OR, R8, 123456789),
11348 BPF_TEST_CLOBBER_ALU(ALU32_IMM, OR, R9, 123456789),
11349 BPF_TEST_CLOBBER_ALU(ALU32_IMM, XOR, R8, 123456789),
11350 BPF_TEST_CLOBBER_ALU(ALU32_IMM, XOR, R9, 123456789),
11351 BPF_TEST_CLOBBER_ALU(ALU32_IMM, LSH, R8, 12),
11352 BPF_TEST_CLOBBER_ALU(ALU32_IMM, LSH, R9, 12),
11353 BPF_TEST_CLOBBER_ALU(ALU32_IMM, RSH, R8, 12),
11354 BPF_TEST_CLOBBER_ALU(ALU32_IMM, RSH, R9, 12),
11355 BPF_TEST_CLOBBER_ALU(ALU32_IMM, ARSH, R8, 12),
11356 BPF_TEST_CLOBBER_ALU(ALU32_IMM, ARSH, R9, 12),
11357 BPF_TEST_CLOBBER_ALU(ALU32_IMM, ADD, R8, 123456789),
11358 BPF_TEST_CLOBBER_ALU(ALU32_IMM, ADD, R9, 123456789),
11359 BPF_TEST_CLOBBER_ALU(ALU32_IMM, SUB, R8, 123456789),
11360 BPF_TEST_CLOBBER_ALU(ALU32_IMM, SUB, R9, 123456789),
11361 BPF_TEST_CLOBBER_ALU(ALU32_IMM, MUL, R8, 123456789),
11362 BPF_TEST_CLOBBER_ALU(ALU32_IMM, MUL, R9, 123456789),
11363 BPF_TEST_CLOBBER_ALU(ALU32_IMM, DIV, R8, 123456789),
11364 BPF_TEST_CLOBBER_ALU(ALU32_IMM, DIV, R9, 123456789),
11365 BPF_TEST_CLOBBER_ALU(ALU32_IMM, MOD, R8, 123456789),
11366 BPF_TEST_CLOBBER_ALU(ALU32_IMM, MOD, R9, 123456789),
11367 /* ALU64 register operations, register clobbering */
11368 BPF_TEST_CLOBBER_ALU(ALU64_REG, AND, R8, R1),
11369 BPF_TEST_CLOBBER_ALU(ALU64_REG, AND, R9, R1),
11370 BPF_TEST_CLOBBER_ALU(ALU64_REG, OR, R8, R1),
11371 BPF_TEST_CLOBBER_ALU(ALU64_REG, OR, R9, R1),
11372 BPF_TEST_CLOBBER_ALU(ALU64_REG, XOR, R8, R1),
11373 BPF_TEST_CLOBBER_ALU(ALU64_REG, XOR, R9, R1),
11374 BPF_TEST_CLOBBER_ALU(ALU64_REG, LSH, R8, R1),
11375 BPF_TEST_CLOBBER_ALU(ALU64_REG, LSH, R9, R1),
11376 BPF_TEST_CLOBBER_ALU(ALU64_REG, RSH, R8, R1),
11377 BPF_TEST_CLOBBER_ALU(ALU64_REG, RSH, R9, R1),
11378 BPF_TEST_CLOBBER_ALU(ALU64_REG, ARSH, R8, R1),
11379 BPF_TEST_CLOBBER_ALU(ALU64_REG, ARSH, R9, R1),
11380 BPF_TEST_CLOBBER_ALU(ALU64_REG, ADD, R8, R1),
11381 BPF_TEST_CLOBBER_ALU(ALU64_REG, ADD, R9, R1),
11382 BPF_TEST_CLOBBER_ALU(ALU64_REG, SUB, R8, R1),
11383 BPF_TEST_CLOBBER_ALU(ALU64_REG, SUB, R9, R1),
11384 BPF_TEST_CLOBBER_ALU(ALU64_REG, MUL, R8, R1),
11385 BPF_TEST_CLOBBER_ALU(ALU64_REG, MUL, R9, R1),
11386 BPF_TEST_CLOBBER_ALU(ALU64_REG, DIV, R8, R1),
11387 BPF_TEST_CLOBBER_ALU(ALU64_REG, DIV, R9, R1),
11388 BPF_TEST_CLOBBER_ALU(ALU64_REG, MOD, R8, R1),
11389 BPF_TEST_CLOBBER_ALU(ALU64_REG, MOD, R9, R1),
11390 /* ALU32 register operations, register clobbering */
11391 BPF_TEST_CLOBBER_ALU(ALU32_REG, AND, R8, R1),
11392 BPF_TEST_CLOBBER_ALU(ALU32_REG, AND, R9, R1),
11393 BPF_TEST_CLOBBER_ALU(ALU32_REG, OR, R8, R1),
11394 BPF_TEST_CLOBBER_ALU(ALU32_REG, OR, R9, R1),
11395 BPF_TEST_CLOBBER_ALU(ALU32_REG, XOR, R8, R1),
11396 BPF_TEST_CLOBBER_ALU(ALU32_REG, XOR, R9, R1),
11397 BPF_TEST_CLOBBER_ALU(ALU32_REG, LSH, R8, R1),
11398 BPF_TEST_CLOBBER_ALU(ALU32_REG, LSH, R9, R1),
11399 BPF_TEST_CLOBBER_ALU(ALU32_REG, RSH, R8, R1),
11400 BPF_TEST_CLOBBER_ALU(ALU32_REG, RSH, R9, R1),
11401 BPF_TEST_CLOBBER_ALU(ALU32_REG, ARSH, R8, R1),
11402 BPF_TEST_CLOBBER_ALU(ALU32_REG, ARSH, R9, R1),
11403 BPF_TEST_CLOBBER_ALU(ALU32_REG, ADD, R8, R1),
11404 BPF_TEST_CLOBBER_ALU(ALU32_REG, ADD, R9, R1),
11405 BPF_TEST_CLOBBER_ALU(ALU32_REG, SUB, R8, R1),
11406 BPF_TEST_CLOBBER_ALU(ALU32_REG, SUB, R9, R1),
11407 BPF_TEST_CLOBBER_ALU(ALU32_REG, MUL, R8, R1),
11408 BPF_TEST_CLOBBER_ALU(ALU32_REG, MUL, R9, R1),
11409 BPF_TEST_CLOBBER_ALU(ALU32_REG, DIV, R8, R1),
11410 BPF_TEST_CLOBBER_ALU(ALU32_REG, DIV, R9, R1),
11411 BPF_TEST_CLOBBER_ALU(ALU32_REG, MOD, R8, R1),
11412 BPF_TEST_CLOBBER_ALU(ALU32_REG, MOD, R9, R1),
11413#undef BPF_TEST_CLOBBER_ALU
11414#define BPF_TEST_CLOBBER_ATOMIC(width, op) \
11415 { \
11416 "Atomic_" #width " " #op ": no clobbering", \
11417 .u.insns_int = { \
11418 BPF_ALU64_IMM(BPF_MOV, R0, 0), \
11419 BPF_ALU64_IMM(BPF_MOV, R1, 1), \
11420 BPF_ALU64_IMM(BPF_MOV, R2, 2), \
11421 BPF_ALU64_IMM(BPF_MOV, R3, 3), \
11422 BPF_ALU64_IMM(BPF_MOV, R4, 4), \
11423 BPF_ALU64_IMM(BPF_MOV, R5, 5), \
11424 BPF_ALU64_IMM(BPF_MOV, R6, 6), \
11425 BPF_ALU64_IMM(BPF_MOV, R7, 7), \
11426 BPF_ALU64_IMM(BPF_MOV, R8, 8), \
11427 BPF_ALU64_IMM(BPF_MOV, R9, 9), \
11428 BPF_ST_MEM(width, R10, -8, \
11429 (op) == BPF_CMPXCHG ? 0 : \
11430 (op) & BPF_FETCH ? 1 : 0), \
11431 BPF_ATOMIC_OP(width, op, R10, R1, -8), \
11432 BPF_JMP_IMM(BPF_JNE, R0, 0, 10), \
11433 BPF_JMP_IMM(BPF_JNE, R1, 1, 9), \
11434 BPF_JMP_IMM(BPF_JNE, R2, 2, 8), \
11435 BPF_JMP_IMM(BPF_JNE, R3, 3, 7), \
11436 BPF_JMP_IMM(BPF_JNE, R4, 4, 6), \
11437 BPF_JMP_IMM(BPF_JNE, R5, 5, 5), \
11438 BPF_JMP_IMM(BPF_JNE, R6, 6, 4), \
11439 BPF_JMP_IMM(BPF_JNE, R7, 7, 3), \
11440 BPF_JMP_IMM(BPF_JNE, R8, 8, 2), \
11441 BPF_JMP_IMM(BPF_JNE, R9, 9, 1), \
11442 BPF_ALU64_IMM(BPF_MOV, R0, 1), \
11443 BPF_EXIT_INSN(), \
11444 }, \
11445 INTERNAL, \
11446 { }, \
11447 { { 0, 1 } }, \
11448 .stack_depth = 8, \
11449 }
11450 /* 64-bit atomic operations, register clobbering */
11451 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_ADD),
11452 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_AND),
11453 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_OR),
11454 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_XOR),
11455 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_ADD | BPF_FETCH),
11456 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_AND | BPF_FETCH),
11457 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_OR | BPF_FETCH),
11458 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_XOR | BPF_FETCH),
11459 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_XCHG),
11460 BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_CMPXCHG),
11461 /* 32-bit atomic operations, register clobbering */
11462 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_ADD),
11463 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_AND),
11464 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_OR),
11465 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_XOR),
11466 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_ADD | BPF_FETCH),
11467 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_AND | BPF_FETCH),
11468 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_OR | BPF_FETCH),
11469 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_XOR | BPF_FETCH),
11470 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_XCHG),
11471 BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_CMPXCHG),
11472#undef BPF_TEST_CLOBBER_ATOMIC
Johan Almbladh0bbaa022021-10-01 15:03:42 +020011473 /* Checking that ALU32 src is not zero extended in place */
11474#define BPF_ALU32_SRC_ZEXT(op) \
11475 { \
11476 "ALU32_" #op "_X: src preserved in zext", \
11477 .u.insns_int = { \
11478 BPF_LD_IMM64(R1, 0x0123456789acbdefULL),\
11479 BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),\
11480 BPF_ALU64_REG(BPF_MOV, R0, R1), \
11481 BPF_ALU32_REG(BPF_##op, R2, R1), \
11482 BPF_ALU64_REG(BPF_SUB, R0, R1), \
11483 BPF_ALU64_REG(BPF_MOV, R1, R0), \
11484 BPF_ALU64_IMM(BPF_RSH, R1, 32), \
11485 BPF_ALU64_REG(BPF_OR, R0, R1), \
11486 BPF_EXIT_INSN(), \
11487 }, \
11488 INTERNAL, \
11489 { }, \
11490 { { 0, 0 } }, \
11491 }
11492 BPF_ALU32_SRC_ZEXT(MOV),
11493 BPF_ALU32_SRC_ZEXT(AND),
11494 BPF_ALU32_SRC_ZEXT(OR),
11495 BPF_ALU32_SRC_ZEXT(XOR),
11496 BPF_ALU32_SRC_ZEXT(ADD),
11497 BPF_ALU32_SRC_ZEXT(SUB),
11498 BPF_ALU32_SRC_ZEXT(MUL),
11499 BPF_ALU32_SRC_ZEXT(DIV),
11500 BPF_ALU32_SRC_ZEXT(MOD),
11501#undef BPF_ALU32_SRC_ZEXT
11502 /* Checking that ATOMIC32 src is not zero extended in place */
11503#define BPF_ATOMIC32_SRC_ZEXT(op) \
11504 { \
11505 "ATOMIC_W_" #op ": src preserved in zext", \
11506 .u.insns_int = { \
11507 BPF_LD_IMM64(R0, 0x0123456789acbdefULL), \
11508 BPF_ALU64_REG(BPF_MOV, R1, R0), \
11509 BPF_ST_MEM(BPF_W, R10, -4, 0), \
11510 BPF_ATOMIC_OP(BPF_W, BPF_##op, R10, R1, -4), \
11511 BPF_ALU64_REG(BPF_SUB, R0, R1), \
11512 BPF_ALU64_REG(BPF_MOV, R1, R0), \
11513 BPF_ALU64_IMM(BPF_RSH, R1, 32), \
11514 BPF_ALU64_REG(BPF_OR, R0, R1), \
11515 BPF_EXIT_INSN(), \
11516 }, \
11517 INTERNAL, \
11518 { }, \
11519 { { 0, 0 } }, \
11520 .stack_depth = 8, \
11521 }
11522 BPF_ATOMIC32_SRC_ZEXT(ADD),
11523 BPF_ATOMIC32_SRC_ZEXT(AND),
11524 BPF_ATOMIC32_SRC_ZEXT(OR),
11525 BPF_ATOMIC32_SRC_ZEXT(XOR),
11526#undef BPF_ATOMIC32_SRC_ZEXT
11527 /* Checking that CMPXCHG32 src is not zero extended in place */
11528 {
11529 "ATOMIC_W_CMPXCHG: src preserved in zext",
11530 .u.insns_int = {
11531 BPF_LD_IMM64(R1, 0x0123456789acbdefULL),
11532 BPF_ALU64_REG(BPF_MOV, R2, R1),
11533 BPF_ALU64_REG(BPF_MOV, R0, 0),
11534 BPF_ST_MEM(BPF_W, R10, -4, 0),
11535 BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R1, -4),
11536 BPF_ALU64_REG(BPF_SUB, R1, R2),
11537 BPF_ALU64_REG(BPF_MOV, R2, R1),
11538 BPF_ALU64_IMM(BPF_RSH, R2, 32),
11539 BPF_ALU64_REG(BPF_OR, R1, R2),
11540 BPF_ALU64_REG(BPF_MOV, R0, R1),
11541 BPF_EXIT_INSN(),
11542 },
11543 INTERNAL,
11544 { },
11545 { { 0, 0 } },
11546 .stack_depth = 8,
11547 },
11548 /* Checking that JMP32 immediate src is not zero extended in place */
11549#define BPF_JMP32_IMM_ZEXT(op) \
11550 { \
11551 "JMP32_" #op "_K: operand preserved in zext", \
11552 .u.insns_int = { \
11553 BPF_LD_IMM64(R0, 0x0123456789acbdefULL),\
11554 BPF_ALU64_REG(BPF_MOV, R1, R0), \
11555 BPF_JMP32_IMM(BPF_##op, R0, 1234, 1), \
11556 BPF_JMP_A(0), /* Nop */ \
11557 BPF_ALU64_REG(BPF_SUB, R0, R1), \
11558 BPF_ALU64_REG(BPF_MOV, R1, R0), \
11559 BPF_ALU64_IMM(BPF_RSH, R1, 32), \
11560 BPF_ALU64_REG(BPF_OR, R0, R1), \
11561 BPF_EXIT_INSN(), \
11562 }, \
11563 INTERNAL, \
11564 { }, \
11565 { { 0, 0 } }, \
11566 }
11567 BPF_JMP32_IMM_ZEXT(JEQ),
11568 BPF_JMP32_IMM_ZEXT(JNE),
11569 BPF_JMP32_IMM_ZEXT(JSET),
11570 BPF_JMP32_IMM_ZEXT(JGT),
11571 BPF_JMP32_IMM_ZEXT(JGE),
11572 BPF_JMP32_IMM_ZEXT(JLT),
11573 BPF_JMP32_IMM_ZEXT(JLE),
11574 BPF_JMP32_IMM_ZEXT(JSGT),
11575 BPF_JMP32_IMM_ZEXT(JSGE),
11576 BPF_JMP32_IMM_ZEXT(JSGT),
11577 BPF_JMP32_IMM_ZEXT(JSLT),
11578 BPF_JMP32_IMM_ZEXT(JSLE),
11579#undef BPF_JMP2_IMM_ZEXT
11580 /* Checking that JMP32 dst & src are not zero extended in place */
11581#define BPF_JMP32_REG_ZEXT(op) \
11582 { \
11583 "JMP32_" #op "_X: operands preserved in zext", \
11584 .u.insns_int = { \
11585 BPF_LD_IMM64(R0, 0x0123456789acbdefULL),\
11586 BPF_LD_IMM64(R1, 0xfedcba9876543210ULL),\
11587 BPF_ALU64_REG(BPF_MOV, R2, R0), \
11588 BPF_ALU64_REG(BPF_MOV, R3, R1), \
11589 BPF_JMP32_IMM(BPF_##op, R0, R1, 1), \
11590 BPF_JMP_A(0), /* Nop */ \
11591 BPF_ALU64_REG(BPF_SUB, R0, R2), \
11592 BPF_ALU64_REG(BPF_SUB, R1, R3), \
11593 BPF_ALU64_REG(BPF_OR, R0, R1), \
11594 BPF_ALU64_REG(BPF_MOV, R1, R0), \
11595 BPF_ALU64_IMM(BPF_RSH, R1, 32), \
11596 BPF_ALU64_REG(BPF_OR, R0, R1), \
11597 BPF_EXIT_INSN(), \
11598 }, \
11599 INTERNAL, \
11600 { }, \
11601 { { 0, 0 } }, \
11602 }
11603 BPF_JMP32_REG_ZEXT(JEQ),
11604 BPF_JMP32_REG_ZEXT(JNE),
11605 BPF_JMP32_REG_ZEXT(JSET),
11606 BPF_JMP32_REG_ZEXT(JGT),
11607 BPF_JMP32_REG_ZEXT(JGE),
11608 BPF_JMP32_REG_ZEXT(JLT),
11609 BPF_JMP32_REG_ZEXT(JLE),
11610 BPF_JMP32_REG_ZEXT(JSGT),
11611 BPF_JMP32_REG_ZEXT(JSGE),
11612 BPF_JMP32_REG_ZEXT(JSGT),
11613 BPF_JMP32_REG_ZEXT(JSLT),
11614 BPF_JMP32_REG_ZEXT(JSLE),
11615#undef BPF_JMP2_REG_ZEXT
Johan Almbladhdaed6082021-10-01 15:03:45 +020011616 /* ALU64 K register combinations */
11617 {
11618 "ALU64_MOV_K: registers",
11619 { },
11620 INTERNAL,
11621 { },
11622 { { 0, 1 } },
11623 .fill_helper = bpf_fill_alu64_mov_imm_regs,
11624 },
11625 {
11626 "ALU64_AND_K: registers",
11627 { },
11628 INTERNAL,
11629 { },
11630 { { 0, 1 } },
11631 .fill_helper = bpf_fill_alu64_and_imm_regs,
11632 },
11633 {
11634 "ALU64_OR_K: registers",
11635 { },
11636 INTERNAL,
11637 { },
11638 { { 0, 1 } },
11639 .fill_helper = bpf_fill_alu64_or_imm_regs,
11640 },
11641 {
11642 "ALU64_XOR_K: registers",
11643 { },
11644 INTERNAL,
11645 { },
11646 { { 0, 1 } },
11647 .fill_helper = bpf_fill_alu64_xor_imm_regs,
11648 },
11649 {
11650 "ALU64_LSH_K: registers",
11651 { },
11652 INTERNAL,
11653 { },
11654 { { 0, 1 } },
11655 .fill_helper = bpf_fill_alu64_lsh_imm_regs,
11656 },
11657 {
11658 "ALU64_RSH_K: registers",
11659 { },
11660 INTERNAL,
11661 { },
11662 { { 0, 1 } },
11663 .fill_helper = bpf_fill_alu64_rsh_imm_regs,
11664 },
11665 {
11666 "ALU64_ARSH_K: registers",
11667 { },
11668 INTERNAL,
11669 { },
11670 { { 0, 1 } },
11671 .fill_helper = bpf_fill_alu64_arsh_imm_regs,
11672 },
11673 {
11674 "ALU64_ADD_K: registers",
11675 { },
11676 INTERNAL,
11677 { },
11678 { { 0, 1 } },
11679 .fill_helper = bpf_fill_alu64_add_imm_regs,
11680 },
11681 {
11682 "ALU64_SUB_K: registers",
11683 { },
11684 INTERNAL,
11685 { },
11686 { { 0, 1 } },
11687 .fill_helper = bpf_fill_alu64_sub_imm_regs,
11688 },
11689 {
11690 "ALU64_MUL_K: registers",
11691 { },
11692 INTERNAL,
11693 { },
11694 { { 0, 1 } },
11695 .fill_helper = bpf_fill_alu64_mul_imm_regs,
11696 },
11697 {
11698 "ALU64_DIV_K: registers",
11699 { },
11700 INTERNAL,
11701 { },
11702 { { 0, 1 } },
11703 .fill_helper = bpf_fill_alu64_div_imm_regs,
11704 },
11705 {
11706 "ALU64_MOD_K: registers",
11707 { },
11708 INTERNAL,
11709 { },
11710 { { 0, 1 } },
11711 .fill_helper = bpf_fill_alu64_mod_imm_regs,
11712 },
11713 /* ALU32 K registers */
11714 {
11715 "ALU32_MOV_K: registers",
11716 { },
11717 INTERNAL,
11718 { },
11719 { { 0, 1 } },
11720 .fill_helper = bpf_fill_alu32_mov_imm_regs,
11721 },
11722 {
11723 "ALU32_AND_K: registers",
11724 { },
11725 INTERNAL,
11726 { },
11727 { { 0, 1 } },
11728 .fill_helper = bpf_fill_alu32_and_imm_regs,
11729 },
11730 {
11731 "ALU32_OR_K: registers",
11732 { },
11733 INTERNAL,
11734 { },
11735 { { 0, 1 } },
11736 .fill_helper = bpf_fill_alu32_or_imm_regs,
11737 },
11738 {
11739 "ALU32_XOR_K: registers",
11740 { },
11741 INTERNAL,
11742 { },
11743 { { 0, 1 } },
11744 .fill_helper = bpf_fill_alu32_xor_imm_regs,
11745 },
11746 {
11747 "ALU32_LSH_K: registers",
11748 { },
11749 INTERNAL,
11750 { },
11751 { { 0, 1 } },
11752 .fill_helper = bpf_fill_alu32_lsh_imm_regs,
11753 },
11754 {
11755 "ALU32_RSH_K: registers",
11756 { },
11757 INTERNAL,
11758 { },
11759 { { 0, 1 } },
11760 .fill_helper = bpf_fill_alu32_rsh_imm_regs,
11761 },
11762 {
11763 "ALU32_ARSH_K: registers",
11764 { },
11765 INTERNAL,
11766 { },
11767 { { 0, 1 } },
11768 .fill_helper = bpf_fill_alu32_arsh_imm_regs,
11769 },
11770 {
11771 "ALU32_ADD_K: registers",
11772 { },
11773 INTERNAL,
11774 { },
11775 { { 0, 1 } },
11776 .fill_helper = bpf_fill_alu32_add_imm_regs,
11777 },
11778 {
11779 "ALU32_SUB_K: registers",
11780 { },
11781 INTERNAL,
11782 { },
11783 { { 0, 1 } },
11784 .fill_helper = bpf_fill_alu32_sub_imm_regs,
11785 },
11786 {
11787 "ALU32_MUL_K: registers",
11788 { },
11789 INTERNAL,
11790 { },
11791 { { 0, 1 } },
11792 .fill_helper = bpf_fill_alu32_mul_imm_regs,
11793 },
11794 {
11795 "ALU32_DIV_K: registers",
11796 { },
11797 INTERNAL,
11798 { },
11799 { { 0, 1 } },
11800 .fill_helper = bpf_fill_alu32_div_imm_regs,
11801 },
11802 {
11803 "ALU32_MOD_K: registers",
11804 { },
11805 INTERNAL,
11806 { },
11807 { { 0, 1 } },
11808 .fill_helper = bpf_fill_alu32_mod_imm_regs,
11809 },
11810 /* ALU64 X register combinations */
11811 {
11812 "ALU64_MOV_X: register combinations",
11813 { },
11814 INTERNAL,
11815 { },
11816 { { 0, 1 } },
11817 .fill_helper = bpf_fill_alu64_mov_reg_pairs,
11818 },
11819 {
11820 "ALU64_AND_X: register combinations",
11821 { },
11822 INTERNAL,
11823 { },
11824 { { 0, 1 } },
11825 .fill_helper = bpf_fill_alu64_and_reg_pairs,
11826 },
11827 {
11828 "ALU64_OR_X: register combinations",
11829 { },
11830 INTERNAL,
11831 { },
11832 { { 0, 1 } },
11833 .fill_helper = bpf_fill_alu64_or_reg_pairs,
11834 },
11835 {
11836 "ALU64_XOR_X: register combinations",
11837 { },
11838 INTERNAL,
11839 { },
11840 { { 0, 1 } },
11841 .fill_helper = bpf_fill_alu64_xor_reg_pairs,
11842 },
11843 {
11844 "ALU64_LSH_X: register combinations",
11845 { },
11846 INTERNAL,
11847 { },
11848 { { 0, 1 } },
11849 .fill_helper = bpf_fill_alu64_lsh_reg_pairs,
11850 },
11851 {
11852 "ALU64_RSH_X: register combinations",
11853 { },
11854 INTERNAL,
11855 { },
11856 { { 0, 1 } },
11857 .fill_helper = bpf_fill_alu64_rsh_reg_pairs,
11858 },
11859 {
11860 "ALU64_ARSH_X: register combinations",
11861 { },
11862 INTERNAL,
11863 { },
11864 { { 0, 1 } },
11865 .fill_helper = bpf_fill_alu64_arsh_reg_pairs,
11866 },
11867 {
11868 "ALU64_ADD_X: register combinations",
11869 { },
11870 INTERNAL,
11871 { },
11872 { { 0, 1 } },
11873 .fill_helper = bpf_fill_alu64_add_reg_pairs,
11874 },
11875 {
11876 "ALU64_SUB_X: register combinations",
11877 { },
11878 INTERNAL,
11879 { },
11880 { { 0, 1 } },
11881 .fill_helper = bpf_fill_alu64_sub_reg_pairs,
11882 },
11883 {
11884 "ALU64_MUL_X: register combinations",
11885 { },
11886 INTERNAL,
11887 { },
11888 { { 0, 1 } },
11889 .fill_helper = bpf_fill_alu64_mul_reg_pairs,
11890 },
11891 {
11892 "ALU64_DIV_X: register combinations",
11893 { },
11894 INTERNAL,
11895 { },
11896 { { 0, 1 } },
11897 .fill_helper = bpf_fill_alu64_div_reg_pairs,
11898 },
11899 {
11900 "ALU64_MOD_X: register combinations",
11901 { },
11902 INTERNAL,
11903 { },
11904 { { 0, 1 } },
11905 .fill_helper = bpf_fill_alu64_mod_reg_pairs,
11906 },
11907 /* ALU32 X register combinations */
11908 {
11909 "ALU32_MOV_X: register combinations",
11910 { },
11911 INTERNAL,
11912 { },
11913 { { 0, 1 } },
11914 .fill_helper = bpf_fill_alu32_mov_reg_pairs,
11915 },
11916 {
11917 "ALU32_AND_X: register combinations",
11918 { },
11919 INTERNAL,
11920 { },
11921 { { 0, 1 } },
11922 .fill_helper = bpf_fill_alu32_and_reg_pairs,
11923 },
11924 {
11925 "ALU32_OR_X: register combinations",
11926 { },
11927 INTERNAL,
11928 { },
11929 { { 0, 1 } },
11930 .fill_helper = bpf_fill_alu32_or_reg_pairs,
11931 },
11932 {
11933 "ALU32_XOR_X: register combinations",
11934 { },
11935 INTERNAL,
11936 { },
11937 { { 0, 1 } },
11938 .fill_helper = bpf_fill_alu32_xor_reg_pairs,
11939 },
11940 {
11941 "ALU32_LSH_X: register combinations",
11942 { },
11943 INTERNAL,
11944 { },
11945 { { 0, 1 } },
11946 .fill_helper = bpf_fill_alu32_lsh_reg_pairs,
11947 },
11948 {
11949 "ALU32_RSH_X: register combinations",
11950 { },
11951 INTERNAL,
11952 { },
11953 { { 0, 1 } },
11954 .fill_helper = bpf_fill_alu32_rsh_reg_pairs,
11955 },
11956 {
11957 "ALU32_ARSH_X: register combinations",
11958 { },
11959 INTERNAL,
11960 { },
11961 { { 0, 1 } },
11962 .fill_helper = bpf_fill_alu32_arsh_reg_pairs,
11963 },
11964 {
11965 "ALU32_ADD_X: register combinations",
11966 { },
11967 INTERNAL,
11968 { },
11969 { { 0, 1 } },
11970 .fill_helper = bpf_fill_alu32_add_reg_pairs,
11971 },
11972 {
11973 "ALU32_SUB_X: register combinations",
11974 { },
11975 INTERNAL,
11976 { },
11977 { { 0, 1 } },
11978 .fill_helper = bpf_fill_alu32_sub_reg_pairs,
11979 },
11980 {
11981 "ALU32_MUL_X: register combinations",
11982 { },
11983 INTERNAL,
11984 { },
11985 { { 0, 1 } },
11986 .fill_helper = bpf_fill_alu32_mul_reg_pairs,
11987 },
11988 {
11989 "ALU32_DIV_X: register combinations",
11990 { },
11991 INTERNAL,
11992 { },
11993 { { 0, 1 } },
11994 .fill_helper = bpf_fill_alu32_div_reg_pairs,
11995 },
11996 {
11997 "ALU32_MOD_X register combinations",
11998 { },
11999 INTERNAL,
12000 { },
12001 { { 0, 1 } },
12002 .fill_helper = bpf_fill_alu32_mod_reg_pairs,
12003 },
Johan Almbladh68c956f2021-09-14 11:18:31 +020012004 /* Exhaustive test of ALU64 shift operations */
12005 {
12006 "ALU64_LSH_K: all shift values",
12007 { },
12008 INTERNAL | FLAG_NO_DATA,
12009 { },
12010 { { 0, 1 } },
Johan Almbladh68813602021-10-01 15:03:47 +020012011 .fill_helper = bpf_fill_alu64_lsh_imm,
Johan Almbladh68c956f2021-09-14 11:18:31 +020012012 },
12013 {
12014 "ALU64_RSH_K: all shift values",
12015 { },
12016 INTERNAL | FLAG_NO_DATA,
12017 { },
12018 { { 0, 1 } },
Johan Almbladh68813602021-10-01 15:03:47 +020012019 .fill_helper = bpf_fill_alu64_rsh_imm,
Johan Almbladh68c956f2021-09-14 11:18:31 +020012020 },
12021 {
12022 "ALU64_ARSH_K: all shift values",
12023 { },
12024 INTERNAL | FLAG_NO_DATA,
12025 { },
12026 { { 0, 1 } },
Johan Almbladh68813602021-10-01 15:03:47 +020012027 .fill_helper = bpf_fill_alu64_arsh_imm,
Johan Almbladh68c956f2021-09-14 11:18:31 +020012028 },
12029 {
12030 "ALU64_LSH_X: all shift values",
12031 { },
12032 INTERNAL | FLAG_NO_DATA,
12033 { },
12034 { { 0, 1 } },
Johan Almbladh68813602021-10-01 15:03:47 +020012035 .fill_helper = bpf_fill_alu64_lsh_reg,
Johan Almbladh68c956f2021-09-14 11:18:31 +020012036 },
12037 {
12038 "ALU64_RSH_X: all shift values",
12039 { },
12040 INTERNAL | FLAG_NO_DATA,
12041 { },
12042 { { 0, 1 } },
Johan Almbladh68813602021-10-01 15:03:47 +020012043 .fill_helper = bpf_fill_alu64_rsh_reg,
Johan Almbladh68c956f2021-09-14 11:18:31 +020012044 },
12045 {
12046 "ALU64_ARSH_X: all shift values",
12047 { },
12048 INTERNAL | FLAG_NO_DATA,
12049 { },
12050 { { 0, 1 } },
Johan Almbladh68813602021-10-01 15:03:47 +020012051 .fill_helper = bpf_fill_alu64_arsh_reg,
Johan Almbladh68c956f2021-09-14 11:18:31 +020012052 },
12053 /* Exhaustive test of ALU32 shift operations */
12054 {
12055 "ALU32_LSH_K: all shift values",
12056 { },
12057 INTERNAL | FLAG_NO_DATA,
12058 { },
12059 { { 0, 1 } },
12060 .fill_helper = bpf_fill_alu32_lsh_imm,
12061 },
12062 {
12063 "ALU32_RSH_K: all shift values",
12064 { },
12065 INTERNAL | FLAG_NO_DATA,
12066 { },
12067 { { 0, 1 } },
12068 .fill_helper = bpf_fill_alu32_rsh_imm,
12069 },
12070 {
12071 "ALU32_ARSH_K: all shift values",
12072 { },
12073 INTERNAL | FLAG_NO_DATA,
12074 { },
12075 { { 0, 1 } },
12076 .fill_helper = bpf_fill_alu32_arsh_imm,
12077 },
12078 {
12079 "ALU32_LSH_X: all shift values",
12080 { },
12081 INTERNAL | FLAG_NO_DATA,
12082 { },
12083 { { 0, 1 } },
12084 .fill_helper = bpf_fill_alu32_lsh_reg,
12085 },
12086 {
12087 "ALU32_RSH_X: all shift values",
12088 { },
12089 INTERNAL | FLAG_NO_DATA,
12090 { },
12091 { { 0, 1 } },
12092 .fill_helper = bpf_fill_alu32_rsh_reg,
12093 },
12094 {
12095 "ALU32_ARSH_X: all shift values",
12096 { },
12097 INTERNAL | FLAG_NO_DATA,
12098 { },
12099 { { 0, 1 } },
12100 .fill_helper = bpf_fill_alu32_arsh_reg,
12101 },
Johan Almbladh68813602021-10-01 15:03:47 +020012102 /*
12103 * Exhaustive test of ALU64 shift operations when
12104 * source and destination register are the same.
12105 */
12106 {
12107 "ALU64_LSH_X: all shift values with the same register",
12108 { },
12109 INTERNAL | FLAG_NO_DATA,
12110 { },
12111 { { 0, 1 } },
12112 .fill_helper = bpf_fill_alu64_lsh_same_reg,
12113 },
12114 {
12115 "ALU64_RSH_X: all shift values with the same register",
12116 { },
12117 INTERNAL | FLAG_NO_DATA,
12118 { },
12119 { { 0, 1 } },
12120 .fill_helper = bpf_fill_alu64_rsh_same_reg,
12121 },
12122 {
12123 "ALU64_ARSH_X: all shift values with the same register",
12124 { },
12125 INTERNAL | FLAG_NO_DATA,
12126 { },
12127 { { 0, 1 } },
12128 .fill_helper = bpf_fill_alu64_arsh_same_reg,
12129 },
12130 /*
12131 * Exhaustive test of ALU32 shift operations when
12132 * source and destination register are the same.
12133 */
12134 {
12135 "ALU32_LSH_X: all shift values with the same register",
12136 { },
12137 INTERNAL | FLAG_NO_DATA,
12138 { },
12139 { { 0, 1 } },
12140 .fill_helper = bpf_fill_alu32_lsh_same_reg,
12141 },
12142 {
12143 "ALU32_RSH_X: all shift values with the same register",
12144 { },
12145 INTERNAL | FLAG_NO_DATA,
12146 { },
12147 { { 0, 1 } },
12148 .fill_helper = bpf_fill_alu32_rsh_same_reg,
12149 },
12150 {
12151 "ALU32_ARSH_X: all shift values with the same register",
12152 { },
12153 INTERNAL | FLAG_NO_DATA,
12154 { },
12155 { { 0, 1 } },
12156 .fill_helper = bpf_fill_alu32_arsh_same_reg,
12157 },
Johan Almbladh9298e632021-09-14 11:18:32 +020012158 /* ALU64 immediate magnitudes */
12159 {
12160 "ALU64_MOV_K: all immediate value magnitudes",
12161 { },
12162 INTERNAL | FLAG_NO_DATA,
12163 { },
12164 { { 0, 1 } },
12165 .fill_helper = bpf_fill_alu64_mov_imm,
12166 .nr_testruns = NR_PATTERN_RUNS,
12167 },
12168 {
12169 "ALU64_AND_K: all immediate value magnitudes",
12170 { },
12171 INTERNAL | FLAG_NO_DATA,
12172 { },
12173 { { 0, 1 } },
12174 .fill_helper = bpf_fill_alu64_and_imm,
12175 .nr_testruns = NR_PATTERN_RUNS,
12176 },
12177 {
12178 "ALU64_OR_K: all immediate value magnitudes",
12179 { },
12180 INTERNAL | FLAG_NO_DATA,
12181 { },
12182 { { 0, 1 } },
12183 .fill_helper = bpf_fill_alu64_or_imm,
12184 .nr_testruns = NR_PATTERN_RUNS,
12185 },
12186 {
12187 "ALU64_XOR_K: all immediate value magnitudes",
12188 { },
12189 INTERNAL | FLAG_NO_DATA,
12190 { },
12191 { { 0, 1 } },
12192 .fill_helper = bpf_fill_alu64_xor_imm,
12193 .nr_testruns = NR_PATTERN_RUNS,
12194 },
12195 {
12196 "ALU64_ADD_K: all immediate value magnitudes",
12197 { },
12198 INTERNAL | FLAG_NO_DATA,
12199 { },
12200 { { 0, 1 } },
12201 .fill_helper = bpf_fill_alu64_add_imm,
12202 .nr_testruns = NR_PATTERN_RUNS,
12203 },
12204 {
12205 "ALU64_SUB_K: all immediate value magnitudes",
12206 { },
12207 INTERNAL | FLAG_NO_DATA,
12208 { },
12209 { { 0, 1 } },
12210 .fill_helper = bpf_fill_alu64_sub_imm,
12211 .nr_testruns = NR_PATTERN_RUNS,
12212 },
12213 {
12214 "ALU64_MUL_K: all immediate value magnitudes",
12215 { },
12216 INTERNAL | FLAG_NO_DATA,
12217 { },
12218 { { 0, 1 } },
12219 .fill_helper = bpf_fill_alu64_mul_imm,
12220 .nr_testruns = NR_PATTERN_RUNS,
12221 },
12222 {
12223 "ALU64_DIV_K: all immediate value magnitudes",
12224 { },
12225 INTERNAL | FLAG_NO_DATA,
12226 { },
12227 { { 0, 1 } },
12228 .fill_helper = bpf_fill_alu64_div_imm,
12229 .nr_testruns = NR_PATTERN_RUNS,
12230 },
12231 {
12232 "ALU64_MOD_K: all immediate value magnitudes",
12233 { },
12234 INTERNAL | FLAG_NO_DATA,
12235 { },
12236 { { 0, 1 } },
12237 .fill_helper = bpf_fill_alu64_mod_imm,
12238 .nr_testruns = NR_PATTERN_RUNS,
12239 },
12240 /* ALU32 immediate magnitudes */
12241 {
12242 "ALU32_MOV_K: all immediate value magnitudes",
12243 { },
12244 INTERNAL | FLAG_NO_DATA,
12245 { },
12246 { { 0, 1 } },
12247 .fill_helper = bpf_fill_alu32_mov_imm,
12248 .nr_testruns = NR_PATTERN_RUNS,
12249 },
12250 {
12251 "ALU32_AND_K: all immediate value magnitudes",
12252 { },
12253 INTERNAL | FLAG_NO_DATA,
12254 { },
12255 { { 0, 1 } },
12256 .fill_helper = bpf_fill_alu32_and_imm,
12257 .nr_testruns = NR_PATTERN_RUNS,
12258 },
12259 {
12260 "ALU32_OR_K: all immediate value magnitudes",
12261 { },
12262 INTERNAL | FLAG_NO_DATA,
12263 { },
12264 { { 0, 1 } },
12265 .fill_helper = bpf_fill_alu32_or_imm,
12266 .nr_testruns = NR_PATTERN_RUNS,
12267 },
12268 {
12269 "ALU32_XOR_K: all immediate value magnitudes",
12270 { },
12271 INTERNAL | FLAG_NO_DATA,
12272 { },
12273 { { 0, 1 } },
12274 .fill_helper = bpf_fill_alu32_xor_imm,
12275 .nr_testruns = NR_PATTERN_RUNS,
12276 },
12277 {
12278 "ALU32_ADD_K: all immediate value magnitudes",
12279 { },
12280 INTERNAL | FLAG_NO_DATA,
12281 { },
12282 { { 0, 1 } },
12283 .fill_helper = bpf_fill_alu32_add_imm,
12284 .nr_testruns = NR_PATTERN_RUNS,
12285 },
12286 {
12287 "ALU32_SUB_K: all immediate value magnitudes",
12288 { },
12289 INTERNAL | FLAG_NO_DATA,
12290 { },
12291 { { 0, 1 } },
12292 .fill_helper = bpf_fill_alu32_sub_imm,
12293 .nr_testruns = NR_PATTERN_RUNS,
12294 },
12295 {
12296 "ALU32_MUL_K: all immediate value magnitudes",
12297 { },
12298 INTERNAL | FLAG_NO_DATA,
12299 { },
12300 { { 0, 1 } },
12301 .fill_helper = bpf_fill_alu32_mul_imm,
12302 .nr_testruns = NR_PATTERN_RUNS,
12303 },
12304 {
12305 "ALU32_DIV_K: all immediate value magnitudes",
12306 { },
12307 INTERNAL | FLAG_NO_DATA,
12308 { },
12309 { { 0, 1 } },
12310 .fill_helper = bpf_fill_alu32_div_imm,
12311 .nr_testruns = NR_PATTERN_RUNS,
12312 },
12313 {
12314 "ALU32_MOD_K: all immediate value magnitudes",
12315 { },
12316 INTERNAL | FLAG_NO_DATA,
12317 { },
12318 { { 0, 1 } },
12319 .fill_helper = bpf_fill_alu32_mod_imm,
Johan Almbladha5a36542021-09-14 11:18:33 +020012320 .nr_testruns = NR_PATTERN_RUNS,
Johan Almbladh9298e632021-09-14 11:18:32 +020012321 },
12322 /* ALU64 register magnitudes */
12323 {
12324 "ALU64_MOV_X: all register value magnitudes",
12325 { },
12326 INTERNAL | FLAG_NO_DATA,
12327 { },
12328 { { 0, 1 } },
12329 .fill_helper = bpf_fill_alu64_mov_reg,
12330 .nr_testruns = NR_PATTERN_RUNS,
12331 },
12332 {
12333 "ALU64_AND_X: all register value magnitudes",
12334 { },
12335 INTERNAL | FLAG_NO_DATA,
12336 { },
12337 { { 0, 1 } },
12338 .fill_helper = bpf_fill_alu64_and_reg,
12339 .nr_testruns = NR_PATTERN_RUNS,
12340 },
12341 {
12342 "ALU64_OR_X: all register value magnitudes",
12343 { },
12344 INTERNAL | FLAG_NO_DATA,
12345 { },
12346 { { 0, 1 } },
12347 .fill_helper = bpf_fill_alu64_or_reg,
12348 .nr_testruns = NR_PATTERN_RUNS,
12349 },
12350 {
12351 "ALU64_XOR_X: all register value magnitudes",
12352 { },
12353 INTERNAL | FLAG_NO_DATA,
12354 { },
12355 { { 0, 1 } },
12356 .fill_helper = bpf_fill_alu64_xor_reg,
12357 .nr_testruns = NR_PATTERN_RUNS,
12358 },
12359 {
12360 "ALU64_ADD_X: all register value magnitudes",
12361 { },
12362 INTERNAL | FLAG_NO_DATA,
12363 { },
12364 { { 0, 1 } },
12365 .fill_helper = bpf_fill_alu64_add_reg,
12366 .nr_testruns = NR_PATTERN_RUNS,
12367 },
12368 {
12369 "ALU64_SUB_X: all register value magnitudes",
12370 { },
12371 INTERNAL | FLAG_NO_DATA,
12372 { },
12373 { { 0, 1 } },
12374 .fill_helper = bpf_fill_alu64_sub_reg,
12375 .nr_testruns = NR_PATTERN_RUNS,
12376 },
12377 {
12378 "ALU64_MUL_X: all register value magnitudes",
12379 { },
12380 INTERNAL | FLAG_NO_DATA,
12381 { },
12382 { { 0, 1 } },
12383 .fill_helper = bpf_fill_alu64_mul_reg,
12384 .nr_testruns = NR_PATTERN_RUNS,
12385 },
12386 {
12387 "ALU64_DIV_X: all register value magnitudes",
12388 { },
12389 INTERNAL | FLAG_NO_DATA,
12390 { },
12391 { { 0, 1 } },
12392 .fill_helper = bpf_fill_alu64_div_reg,
12393 .nr_testruns = NR_PATTERN_RUNS,
12394 },
12395 {
12396 "ALU64_MOD_X: all register value magnitudes",
12397 { },
12398 INTERNAL | FLAG_NO_DATA,
12399 { },
12400 { { 0, 1 } },
12401 .fill_helper = bpf_fill_alu64_mod_reg,
12402 .nr_testruns = NR_PATTERN_RUNS,
12403 },
12404 /* ALU32 register magnitudes */
12405 {
12406 "ALU32_MOV_X: all register value magnitudes",
12407 { },
12408 INTERNAL | FLAG_NO_DATA,
12409 { },
12410 { { 0, 1 } },
12411 .fill_helper = bpf_fill_alu32_mov_reg,
12412 .nr_testruns = NR_PATTERN_RUNS,
12413 },
12414 {
12415 "ALU32_AND_X: all register value magnitudes",
12416 { },
12417 INTERNAL | FLAG_NO_DATA,
12418 { },
12419 { { 0, 1 } },
12420 .fill_helper = bpf_fill_alu32_and_reg,
12421 .nr_testruns = NR_PATTERN_RUNS,
12422 },
12423 {
12424 "ALU32_OR_X: all register value magnitudes",
12425 { },
12426 INTERNAL | FLAG_NO_DATA,
12427 { },
12428 { { 0, 1 } },
12429 .fill_helper = bpf_fill_alu32_or_reg,
12430 .nr_testruns = NR_PATTERN_RUNS,
12431 },
12432 {
12433 "ALU32_XOR_X: all register value magnitudes",
12434 { },
12435 INTERNAL | FLAG_NO_DATA,
12436 { },
12437 { { 0, 1 } },
12438 .fill_helper = bpf_fill_alu32_xor_reg,
12439 .nr_testruns = NR_PATTERN_RUNS,
12440 },
12441 {
12442 "ALU32_ADD_X: all register value magnitudes",
12443 { },
12444 INTERNAL | FLAG_NO_DATA,
12445 { },
12446 { { 0, 1 } },
12447 .fill_helper = bpf_fill_alu32_add_reg,
12448 .nr_testruns = NR_PATTERN_RUNS,
12449 },
12450 {
12451 "ALU32_SUB_X: all register value magnitudes",
12452 { },
12453 INTERNAL | FLAG_NO_DATA,
12454 { },
12455 { { 0, 1 } },
12456 .fill_helper = bpf_fill_alu32_sub_reg,
12457 .nr_testruns = NR_PATTERN_RUNS,
12458 },
12459 {
12460 "ALU32_MUL_X: all register value magnitudes",
12461 { },
12462 INTERNAL | FLAG_NO_DATA,
12463 { },
12464 { { 0, 1 } },
12465 .fill_helper = bpf_fill_alu32_mul_reg,
12466 .nr_testruns = NR_PATTERN_RUNS,
12467 },
12468 {
12469 "ALU32_DIV_X: all register value magnitudes",
12470 { },
12471 INTERNAL | FLAG_NO_DATA,
12472 { },
12473 { { 0, 1 } },
12474 .fill_helper = bpf_fill_alu32_div_reg,
12475 .nr_testruns = NR_PATTERN_RUNS,
12476 },
12477 {
12478 "ALU32_MOD_X: all register value magnitudes",
12479 { },
12480 INTERNAL | FLAG_NO_DATA,
12481 { },
12482 { { 0, 1 } },
12483 .fill_helper = bpf_fill_alu32_mod_reg,
12484 .nr_testruns = NR_PATTERN_RUNS,
12485 },
Johan Almbladh0eb4ef82021-10-07 16:30:06 +020012486 /* LD_IMM64 immediate magnitudes and byte patterns */
Johan Almbladh2e807612021-09-14 11:18:35 +020012487 {
12488 "LD_IMM64: all immediate value magnitudes",
12489 { },
12490 INTERNAL | FLAG_NO_DATA,
12491 { },
12492 { { 0, 1 } },
Johan Almbladh0eb4ef82021-10-07 16:30:06 +020012493 .fill_helper = bpf_fill_ld_imm64_magn,
12494 },
12495 {
12496 "LD_IMM64: checker byte patterns",
12497 { },
12498 INTERNAL | FLAG_NO_DATA,
12499 { },
12500 { { 0, 1 } },
12501 .fill_helper = bpf_fill_ld_imm64_checker,
12502 },
12503 {
12504 "LD_IMM64: random positive and zero byte patterns",
12505 { },
12506 INTERNAL | FLAG_NO_DATA,
12507 { },
12508 { { 0, 1 } },
12509 .fill_helper = bpf_fill_ld_imm64_pos_zero,
12510 },
12511 {
12512 "LD_IMM64: random negative and zero byte patterns",
12513 { },
12514 INTERNAL | FLAG_NO_DATA,
12515 { },
12516 { { 0, 1 } },
12517 .fill_helper = bpf_fill_ld_imm64_neg_zero,
12518 },
12519 {
12520 "LD_IMM64: random positive and negative byte patterns",
12521 { },
12522 INTERNAL | FLAG_NO_DATA,
12523 { },
12524 { { 0, 1 } },
12525 .fill_helper = bpf_fill_ld_imm64_pos_neg,
Johan Almbladh2e807612021-09-14 11:18:35 +020012526 },
Johan Almbladh6fae2e82021-10-01 15:03:46 +020012527 /* 64-bit ATOMIC register combinations */
12528 {
12529 "ATOMIC_DW_ADD: register combinations",
12530 { },
12531 INTERNAL,
12532 { },
12533 { { 0, 1 } },
12534 .fill_helper = bpf_fill_atomic64_add_reg_pairs,
12535 .stack_depth = 8,
12536 },
12537 {
12538 "ATOMIC_DW_AND: register combinations",
12539 { },
12540 INTERNAL,
12541 { },
12542 { { 0, 1 } },
12543 .fill_helper = bpf_fill_atomic64_and_reg_pairs,
12544 .stack_depth = 8,
12545 },
12546 {
12547 "ATOMIC_DW_OR: register combinations",
12548 { },
12549 INTERNAL,
12550 { },
12551 { { 0, 1 } },
12552 .fill_helper = bpf_fill_atomic64_or_reg_pairs,
12553 .stack_depth = 8,
12554 },
12555 {
12556 "ATOMIC_DW_XOR: register combinations",
12557 { },
12558 INTERNAL,
12559 { },
12560 { { 0, 1 } },
12561 .fill_helper = bpf_fill_atomic64_xor_reg_pairs,
12562 .stack_depth = 8,
12563 },
12564 {
12565 "ATOMIC_DW_ADD_FETCH: register combinations",
12566 { },
12567 INTERNAL,
12568 { },
12569 { { 0, 1 } },
12570 .fill_helper = bpf_fill_atomic64_add_fetch_reg_pairs,
12571 .stack_depth = 8,
12572 },
12573 {
12574 "ATOMIC_DW_AND_FETCH: register combinations",
12575 { },
12576 INTERNAL,
12577 { },
12578 { { 0, 1 } },
12579 .fill_helper = bpf_fill_atomic64_and_fetch_reg_pairs,
12580 .stack_depth = 8,
12581 },
12582 {
12583 "ATOMIC_DW_OR_FETCH: register combinations",
12584 { },
12585 INTERNAL,
12586 { },
12587 { { 0, 1 } },
12588 .fill_helper = bpf_fill_atomic64_or_fetch_reg_pairs,
12589 .stack_depth = 8,
12590 },
12591 {
12592 "ATOMIC_DW_XOR_FETCH: register combinations",
12593 { },
12594 INTERNAL,
12595 { },
12596 { { 0, 1 } },
12597 .fill_helper = bpf_fill_atomic64_xor_fetch_reg_pairs,
12598 .stack_depth = 8,
12599 },
12600 {
12601 "ATOMIC_DW_XCHG: register combinations",
12602 { },
12603 INTERNAL,
12604 { },
12605 { { 0, 1 } },
12606 .fill_helper = bpf_fill_atomic64_xchg_reg_pairs,
12607 .stack_depth = 8,
12608 },
12609 {
12610 "ATOMIC_DW_CMPXCHG: register combinations",
12611 { },
12612 INTERNAL,
12613 { },
12614 { { 0, 1 } },
12615 .fill_helper = bpf_fill_atomic64_cmpxchg_reg_pairs,
12616 .stack_depth = 8,
12617 },
12618 /* 32-bit ATOMIC register combinations */
12619 {
12620 "ATOMIC_W_ADD: register combinations",
12621 { },
12622 INTERNAL,
12623 { },
12624 { { 0, 1 } },
12625 .fill_helper = bpf_fill_atomic32_add_reg_pairs,
12626 .stack_depth = 8,
12627 },
12628 {
12629 "ATOMIC_W_AND: register combinations",
12630 { },
12631 INTERNAL,
12632 { },
12633 { { 0, 1 } },
12634 .fill_helper = bpf_fill_atomic32_and_reg_pairs,
12635 .stack_depth = 8,
12636 },
12637 {
12638 "ATOMIC_W_OR: register combinations",
12639 { },
12640 INTERNAL,
12641 { },
12642 { { 0, 1 } },
12643 .fill_helper = bpf_fill_atomic32_or_reg_pairs,
12644 .stack_depth = 8,
12645 },
12646 {
12647 "ATOMIC_W_XOR: register combinations",
12648 { },
12649 INTERNAL,
12650 { },
12651 { { 0, 1 } },
12652 .fill_helper = bpf_fill_atomic32_xor_reg_pairs,
12653 .stack_depth = 8,
12654 },
12655 {
12656 "ATOMIC_W_ADD_FETCH: register combinations",
12657 { },
12658 INTERNAL,
12659 { },
12660 { { 0, 1 } },
12661 .fill_helper = bpf_fill_atomic32_add_fetch_reg_pairs,
12662 .stack_depth = 8,
12663 },
12664 {
12665 "ATOMIC_W_AND_FETCH: register combinations",
12666 { },
12667 INTERNAL,
12668 { },
12669 { { 0, 1 } },
12670 .fill_helper = bpf_fill_atomic32_and_fetch_reg_pairs,
12671 .stack_depth = 8,
12672 },
12673 {
12674 "ATOMIC_W_OR_FETCH: register combinations",
12675 { },
12676 INTERNAL,
12677 { },
12678 { { 0, 1 } },
12679 .fill_helper = bpf_fill_atomic32_or_fetch_reg_pairs,
12680 .stack_depth = 8,
12681 },
12682 {
12683 "ATOMIC_W_XOR_FETCH: register combinations",
12684 { },
12685 INTERNAL,
12686 { },
12687 { { 0, 1 } },
12688 .fill_helper = bpf_fill_atomic32_xor_fetch_reg_pairs,
12689 .stack_depth = 8,
12690 },
12691 {
12692 "ATOMIC_W_XCHG: register combinations",
12693 { },
12694 INTERNAL,
12695 { },
12696 { { 0, 1 } },
12697 .fill_helper = bpf_fill_atomic32_xchg_reg_pairs,
12698 .stack_depth = 8,
12699 },
12700 {
12701 "ATOMIC_W_CMPXCHG: register combinations",
12702 { },
12703 INTERNAL,
12704 { },
12705 { { 0, 1 } },
12706 .fill_helper = bpf_fill_atomic32_cmpxchg_reg_pairs,
12707 .stack_depth = 8,
12708 },
Johan Almbladhf68e8ef2021-10-01 15:03:41 +020012709 /* 64-bit ATOMIC magnitudes */
12710 {
12711 "ATOMIC_DW_ADD: all operand magnitudes",
12712 { },
12713 INTERNAL | FLAG_NO_DATA,
12714 { },
12715 { { 0, 1 } },
12716 .fill_helper = bpf_fill_atomic64_add,
12717 .stack_depth = 8,
12718 .nr_testruns = NR_PATTERN_RUNS,
12719 },
12720 {
12721 "ATOMIC_DW_AND: all operand magnitudes",
12722 { },
12723 INTERNAL | FLAG_NO_DATA,
12724 { },
12725 { { 0, 1 } },
12726 .fill_helper = bpf_fill_atomic64_and,
12727 .stack_depth = 8,
12728 .nr_testruns = NR_PATTERN_RUNS,
12729 },
12730 {
12731 "ATOMIC_DW_OR: all operand magnitudes",
12732 { },
12733 INTERNAL | FLAG_NO_DATA,
12734 { },
12735 { { 0, 1 } },
12736 .fill_helper = bpf_fill_atomic64_or,
12737 .stack_depth = 8,
12738 .nr_testruns = NR_PATTERN_RUNS,
12739 },
12740 {
12741 "ATOMIC_DW_XOR: all operand magnitudes",
12742 { },
12743 INTERNAL | FLAG_NO_DATA,
12744 { },
12745 { { 0, 1 } },
12746 .fill_helper = bpf_fill_atomic64_xor,
12747 .stack_depth = 8,
12748 .nr_testruns = NR_PATTERN_RUNS,
12749 },
12750 {
12751 "ATOMIC_DW_ADD_FETCH: all operand magnitudes",
12752 { },
12753 INTERNAL | FLAG_NO_DATA,
12754 { },
12755 { { 0, 1 } },
12756 .fill_helper = bpf_fill_atomic64_add_fetch,
12757 .stack_depth = 8,
12758 .nr_testruns = NR_PATTERN_RUNS,
12759 },
12760 {
12761 "ATOMIC_DW_AND_FETCH: all operand magnitudes",
12762 { },
12763 INTERNAL | FLAG_NO_DATA,
12764 { },
12765 { { 0, 1 } },
12766 .fill_helper = bpf_fill_atomic64_and_fetch,
12767 .stack_depth = 8,
12768 .nr_testruns = NR_PATTERN_RUNS,
12769 },
12770 {
12771 "ATOMIC_DW_OR_FETCH: all operand magnitudes",
12772 { },
12773 INTERNAL | FLAG_NO_DATA,
12774 { },
12775 { { 0, 1 } },
12776 .fill_helper = bpf_fill_atomic64_or_fetch,
12777 .stack_depth = 8,
12778 .nr_testruns = NR_PATTERN_RUNS,
12779 },
12780 {
12781 "ATOMIC_DW_XOR_FETCH: all operand magnitudes",
12782 { },
12783 INTERNAL | FLAG_NO_DATA,
12784 { },
12785 { { 0, 1 } },
12786 .fill_helper = bpf_fill_atomic64_xor_fetch,
12787 .stack_depth = 8,
12788 .nr_testruns = NR_PATTERN_RUNS,
12789 },
12790 {
12791 "ATOMIC_DW_XCHG: all operand magnitudes",
12792 { },
12793 INTERNAL | FLAG_NO_DATA,
12794 { },
12795 { { 0, 1 } },
12796 .fill_helper = bpf_fill_atomic64_xchg,
12797 .stack_depth = 8,
12798 .nr_testruns = NR_PATTERN_RUNS,
12799 },
12800 {
12801 "ATOMIC_DW_CMPXCHG: all operand magnitudes",
12802 { },
12803 INTERNAL | FLAG_NO_DATA,
12804 { },
12805 { { 0, 1 } },
12806 .fill_helper = bpf_fill_cmpxchg64,
12807 .stack_depth = 8,
12808 .nr_testruns = NR_PATTERN_RUNS,
12809 },
12810 /* 64-bit atomic magnitudes */
12811 {
12812 "ATOMIC_W_ADD: all operand magnitudes",
12813 { },
12814 INTERNAL | FLAG_NO_DATA,
12815 { },
12816 { { 0, 1 } },
12817 .fill_helper = bpf_fill_atomic32_add,
12818 .stack_depth = 8,
12819 .nr_testruns = NR_PATTERN_RUNS,
12820 },
12821 {
12822 "ATOMIC_W_AND: all operand magnitudes",
12823 { },
12824 INTERNAL | FLAG_NO_DATA,
12825 { },
12826 { { 0, 1 } },
12827 .fill_helper = bpf_fill_atomic32_and,
12828 .stack_depth = 8,
12829 .nr_testruns = NR_PATTERN_RUNS,
12830 },
12831 {
12832 "ATOMIC_W_OR: all operand magnitudes",
12833 { },
12834 INTERNAL | FLAG_NO_DATA,
12835 { },
12836 { { 0, 1 } },
12837 .fill_helper = bpf_fill_atomic32_or,
12838 .stack_depth = 8,
12839 .nr_testruns = NR_PATTERN_RUNS,
12840 },
12841 {
12842 "ATOMIC_W_XOR: all operand magnitudes",
12843 { },
12844 INTERNAL | FLAG_NO_DATA,
12845 { },
12846 { { 0, 1 } },
12847 .fill_helper = bpf_fill_atomic32_xor,
12848 .stack_depth = 8,
12849 .nr_testruns = NR_PATTERN_RUNS,
12850 },
12851 {
12852 "ATOMIC_W_ADD_FETCH: all operand magnitudes",
12853 { },
12854 INTERNAL | FLAG_NO_DATA,
12855 { },
12856 { { 0, 1 } },
12857 .fill_helper = bpf_fill_atomic32_add_fetch,
12858 .stack_depth = 8,
12859 .nr_testruns = NR_PATTERN_RUNS,
12860 },
12861 {
12862 "ATOMIC_W_AND_FETCH: all operand magnitudes",
12863 { },
12864 INTERNAL | FLAG_NO_DATA,
12865 { },
12866 { { 0, 1 } },
12867 .fill_helper = bpf_fill_atomic32_and_fetch,
12868 .stack_depth = 8,
12869 .nr_testruns = NR_PATTERN_RUNS,
12870 },
12871 {
12872 "ATOMIC_W_OR_FETCH: all operand magnitudes",
12873 { },
12874 INTERNAL | FLAG_NO_DATA,
12875 { },
12876 { { 0, 1 } },
12877 .fill_helper = bpf_fill_atomic32_or_fetch,
12878 .stack_depth = 8,
12879 .nr_testruns = NR_PATTERN_RUNS,
12880 },
12881 {
12882 "ATOMIC_W_XOR_FETCH: all operand magnitudes",
12883 { },
12884 INTERNAL | FLAG_NO_DATA,
12885 { },
12886 { { 0, 1 } },
12887 .fill_helper = bpf_fill_atomic32_xor_fetch,
12888 .stack_depth = 8,
12889 .nr_testruns = NR_PATTERN_RUNS,
12890 },
12891 {
12892 "ATOMIC_W_XCHG: all operand magnitudes",
12893 { },
12894 INTERNAL | FLAG_NO_DATA,
12895 { },
12896 { { 0, 1 } },
12897 .fill_helper = bpf_fill_atomic32_xchg,
12898 .stack_depth = 8,
12899 .nr_testruns = NR_PATTERN_RUNS,
12900 },
12901 {
12902 "ATOMIC_W_CMPXCHG: all operand magnitudes",
12903 { },
12904 INTERNAL | FLAG_NO_DATA,
12905 { },
12906 { { 0, 1 } },
12907 .fill_helper = bpf_fill_cmpxchg32,
12908 .stack_depth = 8,
12909 .nr_testruns = NR_PATTERN_RUNS,
12910 },
Johan Almbladha5a36542021-09-14 11:18:33 +020012911 /* JMP immediate magnitudes */
12912 {
12913 "JMP_JSET_K: all immediate value magnitudes",
12914 { },
12915 INTERNAL | FLAG_NO_DATA,
12916 { },
12917 { { 0, 1 } },
12918 .fill_helper = bpf_fill_jmp_jset_imm,
12919 .nr_testruns = NR_PATTERN_RUNS,
12920 },
12921 {
12922 "JMP_JEQ_K: all immediate value magnitudes",
12923 { },
12924 INTERNAL | FLAG_NO_DATA,
12925 { },
12926 { { 0, 1 } },
12927 .fill_helper = bpf_fill_jmp_jeq_imm,
12928 .nr_testruns = NR_PATTERN_RUNS,
12929 },
12930 {
12931 "JMP_JNE_K: all immediate value magnitudes",
12932 { },
12933 INTERNAL | FLAG_NO_DATA,
12934 { },
12935 { { 0, 1 } },
12936 .fill_helper = bpf_fill_jmp_jne_imm,
12937 .nr_testruns = NR_PATTERN_RUNS,
12938 },
12939 {
12940 "JMP_JGT_K: all immediate value magnitudes",
12941 { },
12942 INTERNAL | FLAG_NO_DATA,
12943 { },
12944 { { 0, 1 } },
12945 .fill_helper = bpf_fill_jmp_jgt_imm,
12946 .nr_testruns = NR_PATTERN_RUNS,
12947 },
12948 {
12949 "JMP_JGE_K: all immediate value magnitudes",
12950 { },
12951 INTERNAL | FLAG_NO_DATA,
12952 { },
12953 { { 0, 1 } },
12954 .fill_helper = bpf_fill_jmp_jge_imm,
12955 .nr_testruns = NR_PATTERN_RUNS,
12956 },
12957 {
12958 "JMP_JLT_K: all immediate value magnitudes",
12959 { },
12960 INTERNAL | FLAG_NO_DATA,
12961 { },
12962 { { 0, 1 } },
12963 .fill_helper = bpf_fill_jmp_jlt_imm,
12964 .nr_testruns = NR_PATTERN_RUNS,
12965 },
12966 {
12967 "JMP_JLE_K: all immediate value magnitudes",
12968 { },
12969 INTERNAL | FLAG_NO_DATA,
12970 { },
12971 { { 0, 1 } },
12972 .fill_helper = bpf_fill_jmp_jle_imm,
12973 .nr_testruns = NR_PATTERN_RUNS,
12974 },
12975 {
12976 "JMP_JSGT_K: all immediate value magnitudes",
12977 { },
12978 INTERNAL | FLAG_NO_DATA,
12979 { },
12980 { { 0, 1 } },
12981 .fill_helper = bpf_fill_jmp_jsgt_imm,
12982 .nr_testruns = NR_PATTERN_RUNS,
12983 },
12984 {
12985 "JMP_JSGE_K: all immediate value magnitudes",
12986 { },
12987 INTERNAL | FLAG_NO_DATA,
12988 { },
12989 { { 0, 1 } },
12990 .fill_helper = bpf_fill_jmp_jsge_imm,
12991 .nr_testruns = NR_PATTERN_RUNS,
12992 },
12993 {
12994 "JMP_JSLT_K: all immediate value magnitudes",
12995 { },
12996 INTERNAL | FLAG_NO_DATA,
12997 { },
12998 { { 0, 1 } },
12999 .fill_helper = bpf_fill_jmp_jslt_imm,
13000 .nr_testruns = NR_PATTERN_RUNS,
13001 },
13002 {
13003 "JMP_JSLE_K: all immediate value magnitudes",
13004 { },
13005 INTERNAL | FLAG_NO_DATA,
13006 { },
13007 { { 0, 1 } },
13008 .fill_helper = bpf_fill_jmp_jsle_imm,
13009 .nr_testruns = NR_PATTERN_RUNS,
13010 },
13011 /* JMP register magnitudes */
13012 {
13013 "JMP_JSET_X: all register value magnitudes",
13014 { },
13015 INTERNAL | FLAG_NO_DATA,
13016 { },
13017 { { 0, 1 } },
13018 .fill_helper = bpf_fill_jmp_jset_reg,
13019 .nr_testruns = NR_PATTERN_RUNS,
13020 },
13021 {
13022 "JMP_JEQ_X: all register value magnitudes",
13023 { },
13024 INTERNAL | FLAG_NO_DATA,
13025 { },
13026 { { 0, 1 } },
13027 .fill_helper = bpf_fill_jmp_jeq_reg,
13028 .nr_testruns = NR_PATTERN_RUNS,
13029 },
13030 {
13031 "JMP_JNE_X: all register value magnitudes",
13032 { },
13033 INTERNAL | FLAG_NO_DATA,
13034 { },
13035 { { 0, 1 } },
13036 .fill_helper = bpf_fill_jmp_jne_reg,
13037 .nr_testruns = NR_PATTERN_RUNS,
13038 },
13039 {
13040 "JMP_JGT_X: all register value magnitudes",
13041 { },
13042 INTERNAL | FLAG_NO_DATA,
13043 { },
13044 { { 0, 1 } },
13045 .fill_helper = bpf_fill_jmp_jgt_reg,
13046 .nr_testruns = NR_PATTERN_RUNS,
13047 },
13048 {
13049 "JMP_JGE_X: all register value magnitudes",
13050 { },
13051 INTERNAL | FLAG_NO_DATA,
13052 { },
13053 { { 0, 1 } },
13054 .fill_helper = bpf_fill_jmp_jge_reg,
13055 .nr_testruns = NR_PATTERN_RUNS,
13056 },
13057 {
13058 "JMP_JLT_X: all register value magnitudes",
13059 { },
13060 INTERNAL | FLAG_NO_DATA,
13061 { },
13062 { { 0, 1 } },
13063 .fill_helper = bpf_fill_jmp_jlt_reg,
13064 .nr_testruns = NR_PATTERN_RUNS,
13065 },
13066 {
13067 "JMP_JLE_X: all register value magnitudes",
13068 { },
13069 INTERNAL | FLAG_NO_DATA,
13070 { },
13071 { { 0, 1 } },
13072 .fill_helper = bpf_fill_jmp_jle_reg,
13073 .nr_testruns = NR_PATTERN_RUNS,
13074 },
13075 {
13076 "JMP_JSGT_X: all register value magnitudes",
13077 { },
13078 INTERNAL | FLAG_NO_DATA,
13079 { },
13080 { { 0, 1 } },
13081 .fill_helper = bpf_fill_jmp_jsgt_reg,
13082 .nr_testruns = NR_PATTERN_RUNS,
13083 },
13084 {
13085 "JMP_JSGE_X: all register value magnitudes",
13086 { },
13087 INTERNAL | FLAG_NO_DATA,
13088 { },
13089 { { 0, 1 } },
13090 .fill_helper = bpf_fill_jmp_jsge_reg,
13091 .nr_testruns = NR_PATTERN_RUNS,
13092 },
13093 {
13094 "JMP_JSLT_X: all register value magnitudes",
13095 { },
13096 INTERNAL | FLAG_NO_DATA,
13097 { },
13098 { { 0, 1 } },
13099 .fill_helper = bpf_fill_jmp_jslt_reg,
13100 .nr_testruns = NR_PATTERN_RUNS,
13101 },
13102 {
13103 "JMP_JSLE_X: all register value magnitudes",
13104 { },
13105 INTERNAL | FLAG_NO_DATA,
13106 { },
13107 { { 0, 1 } },
13108 .fill_helper = bpf_fill_jmp_jsle_reg,
13109 .nr_testruns = NR_PATTERN_RUNS,
13110 },
13111 /* JMP32 immediate magnitudes */
13112 {
13113 "JMP32_JSET_K: all immediate value magnitudes",
13114 { },
13115 INTERNAL | FLAG_NO_DATA,
13116 { },
13117 { { 0, 1 } },
13118 .fill_helper = bpf_fill_jmp32_jset_imm,
13119 .nr_testruns = NR_PATTERN_RUNS,
13120 },
13121 {
13122 "JMP32_JEQ_K: all immediate value magnitudes",
13123 { },
13124 INTERNAL | FLAG_NO_DATA,
13125 { },
13126 { { 0, 1 } },
13127 .fill_helper = bpf_fill_jmp32_jeq_imm,
13128 .nr_testruns = NR_PATTERN_RUNS,
13129 },
13130 {
13131 "JMP32_JNE_K: all immediate value magnitudes",
13132 { },
13133 INTERNAL | FLAG_NO_DATA,
13134 { },
13135 { { 0, 1 } },
13136 .fill_helper = bpf_fill_jmp32_jne_imm,
13137 .nr_testruns = NR_PATTERN_RUNS,
13138 },
13139 {
13140 "JMP32_JGT_K: all immediate value magnitudes",
13141 { },
13142 INTERNAL | FLAG_NO_DATA,
13143 { },
13144 { { 0, 1 } },
13145 .fill_helper = bpf_fill_jmp32_jgt_imm,
13146 .nr_testruns = NR_PATTERN_RUNS,
13147 },
13148 {
13149 "JMP32_JGE_K: all immediate value magnitudes",
13150 { },
13151 INTERNAL | FLAG_NO_DATA,
13152 { },
13153 { { 0, 1 } },
13154 .fill_helper = bpf_fill_jmp32_jge_imm,
13155 .nr_testruns = NR_PATTERN_RUNS,
13156 },
13157 {
13158 "JMP32_JLT_K: all immediate value magnitudes",
13159 { },
13160 INTERNAL | FLAG_NO_DATA,
13161 { },
13162 { { 0, 1 } },
13163 .fill_helper = bpf_fill_jmp32_jlt_imm,
13164 .nr_testruns = NR_PATTERN_RUNS,
13165 },
13166 {
13167 "JMP32_JLE_K: all immediate value magnitudes",
13168 { },
13169 INTERNAL | FLAG_NO_DATA,
13170 { },
13171 { { 0, 1 } },
13172 .fill_helper = bpf_fill_jmp32_jle_imm,
13173 .nr_testruns = NR_PATTERN_RUNS,
13174 },
13175 {
13176 "JMP32_JSGT_K: all immediate value magnitudes",
13177 { },
13178 INTERNAL | FLAG_NO_DATA,
13179 { },
13180 { { 0, 1 } },
13181 .fill_helper = bpf_fill_jmp32_jsgt_imm,
13182 .nr_testruns = NR_PATTERN_RUNS,
13183 },
13184 {
13185 "JMP32_JSGE_K: all immediate value magnitudes",
13186 { },
13187 INTERNAL | FLAG_NO_DATA,
13188 { },
13189 { { 0, 1 } },
13190 .fill_helper = bpf_fill_jmp32_jsge_imm,
13191 .nr_testruns = NR_PATTERN_RUNS,
13192 },
13193 {
13194 "JMP32_JSLT_K: all immediate value magnitudes",
13195 { },
13196 INTERNAL | FLAG_NO_DATA,
13197 { },
13198 { { 0, 1 } },
13199 .fill_helper = bpf_fill_jmp32_jslt_imm,
13200 .nr_testruns = NR_PATTERN_RUNS,
13201 },
13202 {
13203 "JMP32_JSLE_K: all immediate value magnitudes",
13204 { },
13205 INTERNAL | FLAG_NO_DATA,
13206 { },
13207 { { 0, 1 } },
13208 .fill_helper = bpf_fill_jmp32_jsle_imm,
13209 .nr_testruns = NR_PATTERN_RUNS,
13210 },
13211 /* JMP32 register magnitudes */
13212 {
13213 "JMP32_JSET_X: all register value magnitudes",
13214 { },
13215 INTERNAL | FLAG_NO_DATA,
13216 { },
13217 { { 0, 1 } },
13218 .fill_helper = bpf_fill_jmp32_jset_reg,
13219 .nr_testruns = NR_PATTERN_RUNS,
13220 },
13221 {
13222 "JMP32_JEQ_X: all register value magnitudes",
13223 { },
13224 INTERNAL | FLAG_NO_DATA,
13225 { },
13226 { { 0, 1 } },
13227 .fill_helper = bpf_fill_jmp32_jeq_reg,
13228 .nr_testruns = NR_PATTERN_RUNS,
13229 },
13230 {
13231 "JMP32_JNE_X: all register value magnitudes",
13232 { },
13233 INTERNAL | FLAG_NO_DATA,
13234 { },
13235 { { 0, 1 } },
13236 .fill_helper = bpf_fill_jmp32_jne_reg,
13237 .nr_testruns = NR_PATTERN_RUNS,
13238 },
13239 {
13240 "JMP32_JGT_X: all register value magnitudes",
13241 { },
13242 INTERNAL | FLAG_NO_DATA,
13243 { },
13244 { { 0, 1 } },
13245 .fill_helper = bpf_fill_jmp32_jgt_reg,
13246 .nr_testruns = NR_PATTERN_RUNS,
13247 },
13248 {
13249 "JMP32_JGE_X: all register value magnitudes",
13250 { },
13251 INTERNAL | FLAG_NO_DATA,
13252 { },
13253 { { 0, 1 } },
13254 .fill_helper = bpf_fill_jmp32_jge_reg,
13255 .nr_testruns = NR_PATTERN_RUNS,
13256 },
13257 {
13258 "JMP32_JLT_X: all register value magnitudes",
13259 { },
13260 INTERNAL | FLAG_NO_DATA,
13261 { },
13262 { { 0, 1 } },
13263 .fill_helper = bpf_fill_jmp32_jlt_reg,
13264 .nr_testruns = NR_PATTERN_RUNS,
13265 },
13266 {
13267 "JMP32_JLE_X: all register value magnitudes",
13268 { },
13269 INTERNAL | FLAG_NO_DATA,
13270 { },
13271 { { 0, 1 } },
13272 .fill_helper = bpf_fill_jmp32_jle_reg,
13273 .nr_testruns = NR_PATTERN_RUNS,
13274 },
13275 {
13276 "JMP32_JSGT_X: all register value magnitudes",
13277 { },
13278 INTERNAL | FLAG_NO_DATA,
13279 { },
13280 { { 0, 1 } },
13281 .fill_helper = bpf_fill_jmp32_jsgt_reg,
13282 .nr_testruns = NR_PATTERN_RUNS,
13283 },
13284 {
13285 "JMP32_JSGE_X: all register value magnitudes",
13286 { },
13287 INTERNAL | FLAG_NO_DATA,
13288 { },
13289 { { 0, 1 } },
13290 .fill_helper = bpf_fill_jmp32_jsge_reg,
13291 .nr_testruns = NR_PATTERN_RUNS,
13292 },
13293 {
13294 "JMP32_JSLT_X: all register value magnitudes",
13295 { },
13296 INTERNAL | FLAG_NO_DATA,
13297 { },
13298 { { 0, 1 } },
13299 .fill_helper = bpf_fill_jmp32_jslt_reg,
13300 .nr_testruns = NR_PATTERN_RUNS,
13301 },
13302 {
13303 "JMP32_JSLE_X: all register value magnitudes",
13304 { },
13305 INTERNAL | FLAG_NO_DATA,
13306 { },
13307 { { 0, 1 } },
13308 .fill_helper = bpf_fill_jmp32_jsle_reg,
13309 .nr_testruns = NR_PATTERN_RUNS,
13310 },
Johan Almbladhc4df4552021-09-14 11:18:38 +020013311 /* Conditional jumps with constant decision */
13312 {
13313 "JMP_JSET_K: imm = 0 -> never taken",
13314 .u.insns_int = {
13315 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13316 BPF_JMP_IMM(BPF_JSET, R1, 0, 1),
13317 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13318 BPF_EXIT_INSN(),
13319 },
13320 INTERNAL | FLAG_NO_DATA,
13321 { },
13322 { { 0, 0 } },
13323 },
13324 {
13325 "JMP_JLT_K: imm = 0 -> never taken",
13326 .u.insns_int = {
13327 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13328 BPF_JMP_IMM(BPF_JLT, R1, 0, 1),
13329 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13330 BPF_EXIT_INSN(),
13331 },
13332 INTERNAL | FLAG_NO_DATA,
13333 { },
13334 { { 0, 0 } },
13335 },
13336 {
13337 "JMP_JGE_K: imm = 0 -> always taken",
13338 .u.insns_int = {
13339 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13340 BPF_JMP_IMM(BPF_JGE, R1, 0, 1),
13341 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13342 BPF_EXIT_INSN(),
13343 },
13344 INTERNAL | FLAG_NO_DATA,
13345 { },
13346 { { 0, 1 } },
13347 },
13348 {
13349 "JMP_JGT_K: imm = 0xffffffff -> never taken",
13350 .u.insns_int = {
13351 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13352 BPF_JMP_IMM(BPF_JGT, R1, U32_MAX, 1),
13353 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13354 BPF_EXIT_INSN(),
13355 },
13356 INTERNAL | FLAG_NO_DATA,
13357 { },
13358 { { 0, 0 } },
13359 },
13360 {
13361 "JMP_JLE_K: imm = 0xffffffff -> always taken",
13362 .u.insns_int = {
13363 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13364 BPF_JMP_IMM(BPF_JLE, R1, U32_MAX, 1),
13365 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13366 BPF_EXIT_INSN(),
13367 },
13368 INTERNAL | FLAG_NO_DATA,
13369 { },
13370 { { 0, 1 } },
13371 },
13372 {
13373 "JMP32_JSGT_K: imm = 0x7fffffff -> never taken",
13374 .u.insns_int = {
13375 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13376 BPF_JMP32_IMM(BPF_JSGT, R1, S32_MAX, 1),
13377 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13378 BPF_EXIT_INSN(),
13379 },
13380 INTERNAL | FLAG_NO_DATA,
13381 { },
13382 { { 0, 0 } },
13383 },
13384 {
13385 "JMP32_JSGE_K: imm = -0x80000000 -> always taken",
13386 .u.insns_int = {
13387 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13388 BPF_JMP32_IMM(BPF_JSGE, R1, S32_MIN, 1),
13389 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13390 BPF_EXIT_INSN(),
13391 },
13392 INTERNAL | FLAG_NO_DATA,
13393 { },
13394 { { 0, 1 } },
13395 },
13396 {
13397 "JMP32_JSLT_K: imm = -0x80000000 -> never taken",
13398 .u.insns_int = {
13399 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13400 BPF_JMP32_IMM(BPF_JSLT, R1, S32_MIN, 1),
13401 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13402 BPF_EXIT_INSN(),
13403 },
13404 INTERNAL | FLAG_NO_DATA,
13405 { },
13406 { { 0, 0 } },
13407 },
13408 {
13409 "JMP32_JSLE_K: imm = 0x7fffffff -> always taken",
13410 .u.insns_int = {
13411 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13412 BPF_JMP32_IMM(BPF_JSLE, R1, S32_MAX, 1),
13413 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13414 BPF_EXIT_INSN(),
13415 },
13416 INTERNAL | FLAG_NO_DATA,
13417 { },
13418 { { 0, 1 } },
13419 },
13420 {
13421 "JMP_JEQ_X: dst = src -> always taken",
13422 .u.insns_int = {
13423 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13424 BPF_JMP_REG(BPF_JEQ, R1, R1, 1),
13425 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13426 BPF_EXIT_INSN(),
13427 },
13428 INTERNAL | FLAG_NO_DATA,
13429 { },
13430 { { 0, 1 } },
13431 },
13432 {
13433 "JMP_JGE_X: dst = src -> always taken",
13434 .u.insns_int = {
13435 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13436 BPF_JMP_REG(BPF_JGE, R1, R1, 1),
13437 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13438 BPF_EXIT_INSN(),
13439 },
13440 INTERNAL | FLAG_NO_DATA,
13441 { },
13442 { { 0, 1 } },
13443 },
13444 {
13445 "JMP_JLE_X: dst = src -> always taken",
13446 .u.insns_int = {
13447 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13448 BPF_JMP_REG(BPF_JLE, R1, R1, 1),
13449 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13450 BPF_EXIT_INSN(),
13451 },
13452 INTERNAL | FLAG_NO_DATA,
13453 { },
13454 { { 0, 1 } },
13455 },
13456 {
13457 "JMP_JSGE_X: dst = src -> always taken",
13458 .u.insns_int = {
13459 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13460 BPF_JMP_REG(BPF_JSGE, R1, R1, 1),
13461 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13462 BPF_EXIT_INSN(),
13463 },
13464 INTERNAL | FLAG_NO_DATA,
13465 { },
13466 { { 0, 1 } },
13467 },
13468 {
13469 "JMP_JSLE_X: dst = src -> always taken",
13470 .u.insns_int = {
13471 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13472 BPF_JMP_REG(BPF_JSLE, R1, R1, 1),
13473 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13474 BPF_EXIT_INSN(),
13475 },
13476 INTERNAL | FLAG_NO_DATA,
13477 { },
13478 { { 0, 1 } },
13479 },
13480 {
13481 "JMP_JNE_X: dst = src -> never taken",
13482 .u.insns_int = {
13483 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13484 BPF_JMP_REG(BPF_JNE, R1, R1, 1),
13485 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13486 BPF_EXIT_INSN(),
13487 },
13488 INTERNAL | FLAG_NO_DATA,
13489 { },
13490 { { 0, 0 } },
13491 },
13492 {
13493 "JMP_JGT_X: dst = src -> never taken",
13494 .u.insns_int = {
13495 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13496 BPF_JMP_REG(BPF_JGT, R1, R1, 1),
13497 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13498 BPF_EXIT_INSN(),
13499 },
13500 INTERNAL | FLAG_NO_DATA,
13501 { },
13502 { { 0, 0 } },
13503 },
13504 {
13505 "JMP_JLT_X: dst = src -> never taken",
13506 .u.insns_int = {
13507 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13508 BPF_JMP_REG(BPF_JLT, R1, R1, 1),
13509 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13510 BPF_EXIT_INSN(),
13511 },
13512 INTERNAL | FLAG_NO_DATA,
13513 { },
13514 { { 0, 0 } },
13515 },
13516 {
13517 "JMP_JSGT_X: dst = src -> never taken",
13518 .u.insns_int = {
13519 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13520 BPF_JMP_REG(BPF_JSGT, R1, R1, 1),
13521 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13522 BPF_EXIT_INSN(),
13523 },
13524 INTERNAL | FLAG_NO_DATA,
13525 { },
13526 { { 0, 0 } },
13527 },
13528 {
13529 "JMP_JSLT_X: dst = src -> never taken",
13530 .u.insns_int = {
13531 BPF_ALU64_IMM(BPF_MOV, R0, 1),
13532 BPF_JMP_REG(BPF_JSLT, R1, R1, 1),
13533 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13534 BPF_EXIT_INSN(),
13535 },
13536 INTERNAL | FLAG_NO_DATA,
13537 { },
13538 { { 0, 0 } },
13539 },
Johan Almbladhd4ff9ee2021-09-14 11:18:37 +020013540 /* Short relative jumps */
13541 {
13542 "Short relative jump: offset=0",
13543 .u.insns_int = {
13544 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13545 BPF_JMP_IMM(BPF_JEQ, R0, 0, 0),
13546 BPF_EXIT_INSN(),
13547 BPF_ALU32_IMM(BPF_MOV, R0, -1),
13548 },
13549 INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT,
13550 { },
13551 { { 0, 0 } },
13552 },
13553 {
13554 "Short relative jump: offset=1",
13555 .u.insns_int = {
13556 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13557 BPF_JMP_IMM(BPF_JEQ, R0, 0, 1),
13558 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13559 BPF_EXIT_INSN(),
13560 BPF_ALU32_IMM(BPF_MOV, R0, -1),
13561 },
13562 INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT,
13563 { },
13564 { { 0, 0 } },
13565 },
13566 {
13567 "Short relative jump: offset=2",
13568 .u.insns_int = {
13569 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13570 BPF_JMP_IMM(BPF_JEQ, R0, 0, 2),
13571 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13572 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13573 BPF_EXIT_INSN(),
13574 BPF_ALU32_IMM(BPF_MOV, R0, -1),
13575 },
13576 INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT,
13577 { },
13578 { { 0, 0 } },
13579 },
13580 {
13581 "Short relative jump: offset=3",
13582 .u.insns_int = {
13583 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13584 BPF_JMP_IMM(BPF_JEQ, R0, 0, 3),
13585 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13586 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13587 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13588 BPF_EXIT_INSN(),
13589 BPF_ALU32_IMM(BPF_MOV, R0, -1),
13590 },
13591 INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT,
13592 { },
13593 { { 0, 0 } },
13594 },
13595 {
13596 "Short relative jump: offset=4",
13597 .u.insns_int = {
13598 BPF_ALU64_IMM(BPF_MOV, R0, 0),
13599 BPF_JMP_IMM(BPF_JEQ, R0, 0, 4),
13600 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13601 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13602 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13603 BPF_ALU32_IMM(BPF_ADD, R0, 1),
13604 BPF_EXIT_INSN(),
13605 BPF_ALU32_IMM(BPF_MOV, R0, -1),
13606 },
13607 INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT,
13608 { },
13609 { { 0, 0 } },
13610 },
Johan Almbladhf1517eb2021-09-14 11:18:39 +020013611 /* Conditional branch conversions */
13612 {
13613 "Long conditional jump: taken at runtime",
13614 { },
13615 INTERNAL | FLAG_NO_DATA,
13616 { },
13617 { { 0, 1 } },
13618 .fill_helper = bpf_fill_max_jmp_taken,
13619 },
13620 {
13621 "Long conditional jump: not taken at runtime",
13622 { },
13623 INTERNAL | FLAG_NO_DATA,
13624 { },
13625 { { 0, 2 } },
13626 .fill_helper = bpf_fill_max_jmp_not_taken,
13627 },
13628 {
13629 "Long conditional jump: always taken, known at JIT time",
13630 { },
13631 INTERNAL | FLAG_NO_DATA,
13632 { },
13633 { { 0, 1 } },
13634 .fill_helper = bpf_fill_max_jmp_always_taken,
13635 },
13636 {
13637 "Long conditional jump: never taken, known at JIT time",
13638 { },
13639 INTERNAL | FLAG_NO_DATA,
13640 { },
13641 { { 0, 2 } },
13642 .fill_helper = bpf_fill_max_jmp_never_taken,
13643 },
Johan Almbladha7d2e752021-09-14 11:18:34 +020013644 /* Staggered jump sequences, immediate */
13645 {
13646 "Staggered jumps: JMP_JA",
13647 { },
13648 INTERNAL | FLAG_NO_DATA,
13649 { },
13650 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13651 .fill_helper = bpf_fill_staggered_ja,
13652 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13653 },
13654 {
13655 "Staggered jumps: JMP_JEQ_K",
13656 { },
13657 INTERNAL | FLAG_NO_DATA,
13658 { },
13659 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13660 .fill_helper = bpf_fill_staggered_jeq_imm,
13661 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13662 },
13663 {
13664 "Staggered jumps: JMP_JNE_K",
13665 { },
13666 INTERNAL | FLAG_NO_DATA,
13667 { },
13668 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13669 .fill_helper = bpf_fill_staggered_jne_imm,
13670 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13671 },
13672 {
13673 "Staggered jumps: JMP_JSET_K",
13674 { },
13675 INTERNAL | FLAG_NO_DATA,
13676 { },
13677 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13678 .fill_helper = bpf_fill_staggered_jset_imm,
13679 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13680 },
13681 {
13682 "Staggered jumps: JMP_JGT_K",
13683 { },
13684 INTERNAL | FLAG_NO_DATA,
13685 { },
13686 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13687 .fill_helper = bpf_fill_staggered_jgt_imm,
13688 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13689 },
13690 {
13691 "Staggered jumps: JMP_JGE_K",
13692 { },
13693 INTERNAL | FLAG_NO_DATA,
13694 { },
13695 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13696 .fill_helper = bpf_fill_staggered_jge_imm,
13697 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13698 },
13699 {
13700 "Staggered jumps: JMP_JLT_K",
13701 { },
13702 INTERNAL | FLAG_NO_DATA,
13703 { },
13704 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13705 .fill_helper = bpf_fill_staggered_jlt_imm,
13706 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13707 },
13708 {
13709 "Staggered jumps: JMP_JLE_K",
13710 { },
13711 INTERNAL | FLAG_NO_DATA,
13712 { },
13713 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13714 .fill_helper = bpf_fill_staggered_jle_imm,
13715 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13716 },
13717 {
13718 "Staggered jumps: JMP_JSGT_K",
13719 { },
13720 INTERNAL | FLAG_NO_DATA,
13721 { },
13722 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13723 .fill_helper = bpf_fill_staggered_jsgt_imm,
13724 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13725 },
13726 {
13727 "Staggered jumps: JMP_JSGE_K",
13728 { },
13729 INTERNAL | FLAG_NO_DATA,
13730 { },
13731 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13732 .fill_helper = bpf_fill_staggered_jsge_imm,
13733 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13734 },
13735 {
13736 "Staggered jumps: JMP_JSLT_K",
13737 { },
13738 INTERNAL | FLAG_NO_DATA,
13739 { },
13740 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13741 .fill_helper = bpf_fill_staggered_jslt_imm,
13742 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13743 },
13744 {
13745 "Staggered jumps: JMP_JSLE_K",
13746 { },
13747 INTERNAL | FLAG_NO_DATA,
13748 { },
13749 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13750 .fill_helper = bpf_fill_staggered_jsle_imm,
13751 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13752 },
13753 /* Staggered jump sequences, register */
13754 {
13755 "Staggered jumps: JMP_JEQ_X",
13756 { },
13757 INTERNAL | FLAG_NO_DATA,
13758 { },
13759 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13760 .fill_helper = bpf_fill_staggered_jeq_reg,
13761 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13762 },
13763 {
13764 "Staggered jumps: JMP_JNE_X",
13765 { },
13766 INTERNAL | FLAG_NO_DATA,
13767 { },
13768 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13769 .fill_helper = bpf_fill_staggered_jne_reg,
13770 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13771 },
13772 {
13773 "Staggered jumps: JMP_JSET_X",
13774 { },
13775 INTERNAL | FLAG_NO_DATA,
13776 { },
13777 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13778 .fill_helper = bpf_fill_staggered_jset_reg,
13779 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13780 },
13781 {
13782 "Staggered jumps: JMP_JGT_X",
13783 { },
13784 INTERNAL | FLAG_NO_DATA,
13785 { },
13786 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13787 .fill_helper = bpf_fill_staggered_jgt_reg,
13788 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13789 },
13790 {
13791 "Staggered jumps: JMP_JGE_X",
13792 { },
13793 INTERNAL | FLAG_NO_DATA,
13794 { },
13795 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13796 .fill_helper = bpf_fill_staggered_jge_reg,
13797 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13798 },
13799 {
13800 "Staggered jumps: JMP_JLT_X",
13801 { },
13802 INTERNAL | FLAG_NO_DATA,
13803 { },
13804 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13805 .fill_helper = bpf_fill_staggered_jlt_reg,
13806 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13807 },
13808 {
13809 "Staggered jumps: JMP_JLE_X",
13810 { },
13811 INTERNAL | FLAG_NO_DATA,
13812 { },
13813 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13814 .fill_helper = bpf_fill_staggered_jle_reg,
13815 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13816 },
13817 {
13818 "Staggered jumps: JMP_JSGT_X",
13819 { },
13820 INTERNAL | FLAG_NO_DATA,
13821 { },
13822 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13823 .fill_helper = bpf_fill_staggered_jsgt_reg,
13824 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13825 },
13826 {
13827 "Staggered jumps: JMP_JSGE_X",
13828 { },
13829 INTERNAL | FLAG_NO_DATA,
13830 { },
13831 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13832 .fill_helper = bpf_fill_staggered_jsge_reg,
13833 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13834 },
13835 {
13836 "Staggered jumps: JMP_JSLT_X",
13837 { },
13838 INTERNAL | FLAG_NO_DATA,
13839 { },
13840 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13841 .fill_helper = bpf_fill_staggered_jslt_reg,
13842 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13843 },
13844 {
13845 "Staggered jumps: JMP_JSLE_X",
13846 { },
13847 INTERNAL | FLAG_NO_DATA,
13848 { },
13849 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13850 .fill_helper = bpf_fill_staggered_jsle_reg,
13851 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13852 },
13853 /* Staggered jump sequences, JMP32 immediate */
13854 {
13855 "Staggered jumps: JMP32_JEQ_K",
13856 { },
13857 INTERNAL | FLAG_NO_DATA,
13858 { },
13859 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13860 .fill_helper = bpf_fill_staggered_jeq32_imm,
13861 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13862 },
13863 {
13864 "Staggered jumps: JMP32_JNE_K",
13865 { },
13866 INTERNAL | FLAG_NO_DATA,
13867 { },
13868 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13869 .fill_helper = bpf_fill_staggered_jne32_imm,
13870 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13871 },
13872 {
13873 "Staggered jumps: JMP32_JSET_K",
13874 { },
13875 INTERNAL | FLAG_NO_DATA,
13876 { },
13877 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13878 .fill_helper = bpf_fill_staggered_jset32_imm,
13879 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13880 },
13881 {
13882 "Staggered jumps: JMP32_JGT_K",
13883 { },
13884 INTERNAL | FLAG_NO_DATA,
13885 { },
13886 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13887 .fill_helper = bpf_fill_staggered_jgt32_imm,
13888 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13889 },
13890 {
13891 "Staggered jumps: JMP32_JGE_K",
13892 { },
13893 INTERNAL | FLAG_NO_DATA,
13894 { },
13895 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13896 .fill_helper = bpf_fill_staggered_jge32_imm,
13897 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13898 },
13899 {
13900 "Staggered jumps: JMP32_JLT_K",
13901 { },
13902 INTERNAL | FLAG_NO_DATA,
13903 { },
13904 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13905 .fill_helper = bpf_fill_staggered_jlt32_imm,
13906 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13907 },
13908 {
13909 "Staggered jumps: JMP32_JLE_K",
13910 { },
13911 INTERNAL | FLAG_NO_DATA,
13912 { },
13913 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13914 .fill_helper = bpf_fill_staggered_jle32_imm,
13915 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13916 },
13917 {
13918 "Staggered jumps: JMP32_JSGT_K",
13919 { },
13920 INTERNAL | FLAG_NO_DATA,
13921 { },
13922 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13923 .fill_helper = bpf_fill_staggered_jsgt32_imm,
13924 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13925 },
13926 {
13927 "Staggered jumps: JMP32_JSGE_K",
13928 { },
13929 INTERNAL | FLAG_NO_DATA,
13930 { },
13931 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13932 .fill_helper = bpf_fill_staggered_jsge32_imm,
13933 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13934 },
13935 {
13936 "Staggered jumps: JMP32_JSLT_K",
13937 { },
13938 INTERNAL | FLAG_NO_DATA,
13939 { },
13940 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13941 .fill_helper = bpf_fill_staggered_jslt32_imm,
13942 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13943 },
13944 {
13945 "Staggered jumps: JMP32_JSLE_K",
13946 { },
13947 INTERNAL | FLAG_NO_DATA,
13948 { },
13949 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13950 .fill_helper = bpf_fill_staggered_jsle32_imm,
13951 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13952 },
13953 /* Staggered jump sequences, JMP32 register */
13954 {
13955 "Staggered jumps: JMP32_JEQ_X",
13956 { },
13957 INTERNAL | FLAG_NO_DATA,
13958 { },
13959 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13960 .fill_helper = bpf_fill_staggered_jeq32_reg,
13961 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13962 },
13963 {
13964 "Staggered jumps: JMP32_JNE_X",
13965 { },
13966 INTERNAL | FLAG_NO_DATA,
13967 { },
13968 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13969 .fill_helper = bpf_fill_staggered_jne32_reg,
13970 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13971 },
13972 {
13973 "Staggered jumps: JMP32_JSET_X",
13974 { },
13975 INTERNAL | FLAG_NO_DATA,
13976 { },
13977 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13978 .fill_helper = bpf_fill_staggered_jset32_reg,
13979 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13980 },
13981 {
13982 "Staggered jumps: JMP32_JGT_X",
13983 { },
13984 INTERNAL | FLAG_NO_DATA,
13985 { },
13986 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13987 .fill_helper = bpf_fill_staggered_jgt32_reg,
13988 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13989 },
13990 {
13991 "Staggered jumps: JMP32_JGE_X",
13992 { },
13993 INTERNAL | FLAG_NO_DATA,
13994 { },
13995 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
13996 .fill_helper = bpf_fill_staggered_jge32_reg,
13997 .nr_testruns = NR_STAGGERED_JMP_RUNS,
13998 },
13999 {
14000 "Staggered jumps: JMP32_JLT_X",
14001 { },
14002 INTERNAL | FLAG_NO_DATA,
14003 { },
14004 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
14005 .fill_helper = bpf_fill_staggered_jlt32_reg,
14006 .nr_testruns = NR_STAGGERED_JMP_RUNS,
14007 },
14008 {
14009 "Staggered jumps: JMP32_JLE_X",
14010 { },
14011 INTERNAL | FLAG_NO_DATA,
14012 { },
14013 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
14014 .fill_helper = bpf_fill_staggered_jle32_reg,
14015 .nr_testruns = NR_STAGGERED_JMP_RUNS,
14016 },
14017 {
14018 "Staggered jumps: JMP32_JSGT_X",
14019 { },
14020 INTERNAL | FLAG_NO_DATA,
14021 { },
14022 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
14023 .fill_helper = bpf_fill_staggered_jsgt32_reg,
14024 .nr_testruns = NR_STAGGERED_JMP_RUNS,
14025 },
14026 {
14027 "Staggered jumps: JMP32_JSGE_X",
14028 { },
14029 INTERNAL | FLAG_NO_DATA,
14030 { },
14031 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
14032 .fill_helper = bpf_fill_staggered_jsge32_reg,
14033 .nr_testruns = NR_STAGGERED_JMP_RUNS,
14034 },
14035 {
14036 "Staggered jumps: JMP32_JSLT_X",
14037 { },
14038 INTERNAL | FLAG_NO_DATA,
14039 { },
14040 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
14041 .fill_helper = bpf_fill_staggered_jslt32_reg,
14042 .nr_testruns = NR_STAGGERED_JMP_RUNS,
14043 },
14044 {
14045 "Staggered jumps: JMP32_JSLE_X",
14046 { },
14047 INTERNAL | FLAG_NO_DATA,
14048 { },
14049 { { 0, MAX_STAGGERED_JMP_SIZE + 1 } },
14050 .fill_helper = bpf_fill_staggered_jsle32_reg,
14051 .nr_testruns = NR_STAGGERED_JMP_RUNS,
14052 },
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014053};
14054
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014055static struct net_device dev;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014056
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014057static struct sk_buff *populate_skb(char *buf, int size)
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014058{
14059 struct sk_buff *skb;
14060
14061 if (size >= MAX_DATA)
14062 return NULL;
14063
14064 skb = alloc_skb(MAX_DATA, GFP_KERNEL);
14065 if (!skb)
14066 return NULL;
14067
yuan linyude77b962017-06-18 22:48:17 +080014068 __skb_put_data(skb, buf, size);
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014069
14070 /* Initialize a fake skb with test pattern. */
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014071 skb_reset_mac_header(skb);
14072 skb->protocol = htons(ETH_P_IP);
14073 skb->pkt_type = SKB_TYPE;
14074 skb->mark = SKB_MARK;
14075 skb->hash = SKB_HASH;
14076 skb->queue_mapping = SKB_QUEUE_MAP;
14077 skb->vlan_tci = SKB_VLAN_TCI;
Michał Mirosław0c4b2d32018-11-10 19:58:36 +010014078 skb->vlan_present = SKB_VLAN_PRESENT;
Jakub Kicinski5c0ca3f2016-09-12 13:04:57 +010014079 skb->vlan_proto = htons(ETH_P_IP);
Song Liu100811932018-09-27 09:34:41 -070014080 dev_net_set(&dev, &init_net);
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014081 skb->dev = &dev;
14082 skb->dev->ifindex = SKB_DEV_IFINDEX;
14083 skb->dev->type = SKB_DEV_TYPE;
14084 skb_set_network_header(skb, min(size, ETH_HLEN));
14085
14086 return skb;
14087}
14088
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014089static void *generate_test_data(struct bpf_test *test, int sub)
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014090{
Nicolas Schichanbac142a2015-08-04 15:19:08 +020014091 struct sk_buff *skb;
14092 struct page *page;
14093
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014094 if (test->aux & FLAG_NO_DATA)
14095 return NULL;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014096
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014097 /* Test case expects an skb, so populate one. Various
14098 * subtests generate skbs of different sizes based on
14099 * the same data.
14100 */
Nicolas Schichanbac142a2015-08-04 15:19:08 +020014101 skb = populate_skb(test->data, test->test[sub].data_size);
14102 if (!skb)
14103 return NULL;
14104
14105 if (test->aux & FLAG_SKB_FRAG) {
14106 /*
14107 * when the test requires a fragmented skb, add a
14108 * single fragment to the skb, filled with
14109 * test->frag_data.
14110 */
14111 void *ptr;
14112
14113 page = alloc_page(GFP_KERNEL);
14114
14115 if (!page)
14116 goto err_kfree_skb;
14117
14118 ptr = kmap(page);
14119 if (!ptr)
14120 goto err_free_page;
14121 memcpy(ptr, test->frag_data, MAX_DATA);
14122 kunmap(page);
14123 skb_add_rx_frag(skb, 0, page, 0, MAX_DATA, MAX_DATA);
14124 }
14125
14126 return skb;
14127
14128err_free_page:
14129 __free_page(page);
14130err_kfree_skb:
14131 kfree_skb(skb);
14132 return NULL;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014133}
14134
14135static void release_test_data(const struct bpf_test *test, void *data)
14136{
14137 if (test->aux & FLAG_NO_DATA)
14138 return;
14139
14140 kfree_skb(data);
14141}
14142
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020014143static int filter_length(int which)
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014144{
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020014145 struct sock_filter *fp;
14146 int len;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014147
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020014148 if (tests[which].fill_helper)
14149 return tests[which].u.ptr.len;
14150
14151 fp = tests[which].u.insns;
Chema Gonzaleze9d94502014-05-30 10:15:12 -070014152 for (len = MAX_INSNS - 1; len > 0; --len)
14153 if (fp[len].code != 0 || fp[len].k != 0)
14154 break;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014155
Chema Gonzaleze9d94502014-05-30 10:15:12 -070014156 return len + 1;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014157}
14158
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020014159static void *filter_pointer(int which)
14160{
14161 if (tests[which].fill_helper)
14162 return tests[which].u.ptr.insns;
14163 else
14164 return tests[which].u.insns;
14165}
14166
Alexei Starovoitov7ae457c2014-07-30 20:34:16 -070014167static struct bpf_prog *generate_filter(int which, int *err)
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014168{
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014169 __u8 test_type = tests[which].aux & TEST_TYPE_MASK;
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020014170 unsigned int flen = filter_length(which);
14171 void *fptr = filter_pointer(which);
14172 struct sock_fprog_kern fprog;
14173 struct bpf_prog *fp;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014174
14175 switch (test_type) {
14176 case CLASSIC:
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020014177 fprog.filter = fptr;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014178 fprog.len = flen;
14179
Alexei Starovoitov7ae457c2014-07-30 20:34:16 -070014180 *err = bpf_prog_create(&fp, &fprog);
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014181 if (tests[which].aux & FLAG_EXPECTED_FAIL) {
Yonghong Song09584b42018-02-02 22:37:15 -080014182 if (*err == tests[which].expected_errcode) {
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014183 pr_cont("PASS\n");
14184 /* Verifier rejected filter as expected. */
14185 *err = 0;
14186 return NULL;
14187 } else {
14188 pr_cont("UNEXPECTED_PASS\n");
14189 /* Verifier didn't reject the test that's
14190 * bad enough, just return!
14191 */
14192 *err = -EINVAL;
14193 return NULL;
14194 }
14195 }
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014196 if (*err) {
Alexei Starovoitov290af862018-01-09 10:04:29 -080014197 pr_cont("FAIL to prog_create err=%d len=%d\n",
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014198 *err, fprog.len);
14199 return NULL;
14200 }
14201 break;
14202
14203 case INTERNAL:
Daniel Borkmann60a3b222014-09-02 22:53:44 +020014204 fp = bpf_prog_alloc(bpf_prog_size(flen), 0);
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014205 if (fp == NULL) {
14206 pr_cont("UNEXPECTED_FAIL no memory left\n");
14207 *err = -ENOMEM;
14208 return NULL;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014209 }
14210
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014211 fp->len = flen;
Daniel Borkmann4962fa12015-07-30 12:42:46 +020014212 /* Type doesn't really matter here as long as it's not unspec. */
14213 fp->type = BPF_PROG_TYPE_SOCKET_FILTER;
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020014214 memcpy(fp->insnsi, fptr, fp->len * sizeof(struct bpf_insn));
Alexei Starovoitov105c0362017-05-30 13:31:32 -070014215 fp->aux->stack_depth = tests[which].stack_depth;
Johan Almbladh27cc6da2021-09-14 11:18:36 +020014216 fp->aux->verifier_zext = !!(tests[which].aux &
14217 FLAG_VERIFIER_ZEXT);
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014218
Daniel Borkmannd1c55ab2016-05-13 19:08:31 +020014219 /* We cannot error here as we don't need type compatibility
14220 * checks.
14221 */
14222 fp = bpf_prog_select_runtime(fp, err);
Alexei Starovoitov290af862018-01-09 10:04:29 -080014223 if (*err) {
14224 pr_cont("FAIL to select_runtime err=%d\n", *err);
14225 return NULL;
14226 }
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014227 break;
14228 }
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014229
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014230 *err = 0;
14231 return fp;
14232}
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014233
Alexei Starovoitov7ae457c2014-07-30 20:34:16 -070014234static void release_filter(struct bpf_prog *fp, int which)
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014235{
14236 __u8 test_type = tests[which].aux & TEST_TYPE_MASK;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014237
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014238 switch (test_type) {
14239 case CLASSIC:
Alexei Starovoitov7ae457c2014-07-30 20:34:16 -070014240 bpf_prog_destroy(fp);
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014241 break;
14242 case INTERNAL:
Alexei Starovoitov7ae457c2014-07-30 20:34:16 -070014243 bpf_prog_free(fp);
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014244 break;
14245 }
14246}
14247
Alexei Starovoitov7ae457c2014-07-30 20:34:16 -070014248static int __run_one(const struct bpf_prog *fp, const void *data,
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014249 int runs, u64 *duration)
14250{
14251 u64 start, finish;
Alexei Starovoitov25ee7322014-09-19 13:53:51 -070014252 int ret = 0, i;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014253
David Miller6eac7792020-02-24 15:01:44 +010014254 migrate_disable();
Alexei Starovoitov4d9c5c52015-07-20 20:34:19 -070014255 start = ktime_get_ns();
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014256
14257 for (i = 0; i < runs; i++)
Andrii Nakryikofb7dd8b2021-08-15 00:05:54 -070014258 ret = bpf_prog_run(fp, data);
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014259
Alexei Starovoitov4d9c5c52015-07-20 20:34:19 -070014260 finish = ktime_get_ns();
David Miller6eac7792020-02-24 15:01:44 +010014261 migrate_enable();
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014262
Alexei Starovoitov4d9c5c52015-07-20 20:34:19 -070014263 *duration = finish - start;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014264 do_div(*duration, runs);
14265
14266 return ret;
14267}
14268
Alexei Starovoitov7ae457c2014-07-30 20:34:16 -070014269static int run_one(const struct bpf_prog *fp, struct bpf_test *test)
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014270{
14271 int err_cnt = 0, i, runs = MAX_TESTRUNS;
14272
Johan Almbladhc2a228d2021-09-14 11:18:29 +020014273 if (test->nr_testruns)
14274 runs = min(test->nr_testruns, MAX_TESTRUNS);
14275
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014276 for (i = 0; i < MAX_SUBTESTS; i++) {
14277 void *data;
14278 u64 duration;
14279 u32 ret;
14280
Johan Almbladh2b7e9f22021-07-21 12:38:22 +020014281 /*
14282 * NOTE: Several sub-tests may be present, in which case
14283 * a zero {data_size, result} tuple indicates the end of
14284 * the sub-test array. The first test is always run,
14285 * even if both data_size and result happen to be zero.
14286 */
14287 if (i > 0 &&
14288 test->test[i].data_size == 0 &&
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014289 test->test[i].result == 0)
14290 break;
14291
14292 data = generate_test_data(test, i);
Nicolas Schichane34684f2015-08-04 15:19:07 +020014293 if (!data && !(test->aux & FLAG_NO_DATA)) {
14294 pr_cont("data generation failed ");
14295 err_cnt++;
14296 break;
14297 }
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014298 ret = __run_one(fp, data, runs, &duration);
14299 release_test_data(test, data);
14300
14301 if (ret == test->test[i].result) {
14302 pr_cont("%lld ", duration);
14303 } else {
14304 pr_cont("ret %d != %d ", ret,
14305 test->test[i].result);
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014306 err_cnt++;
14307 }
14308 }
14309
14310 return err_cnt;
14311}
14312
Nicolas Schichand2648d42015-08-04 15:19:10 +020014313static char test_name[64];
14314module_param_string(test_name, test_name, sizeof(test_name), 0);
14315
14316static int test_id = -1;
14317module_param(test_id, int, 0);
14318
Tiezhu Yangb066abb2021-10-28 09:25:21 +080014319static int test_range[2] = { 0, INT_MAX };
Nicolas Schichand2648d42015-08-04 15:19:10 +020014320module_param_array(test_range, int, NULL, 0);
14321
Nicolas Schichand2648d42015-08-04 15:19:10 +020014322static bool exclude_test(int test_id)
14323{
14324 return test_id < test_range[0] || test_id > test_range[1];
14325}
14326
Yonghong Song76db8082018-03-21 16:31:04 -070014327static __init struct sk_buff *build_test_skb(void)
14328{
14329 u32 headroom = NET_SKB_PAD + NET_IP_ALIGN + ETH_HLEN;
14330 struct sk_buff *skb[2];
14331 struct page *page[2];
14332 int i, data_size = 8;
14333
14334 for (i = 0; i < 2; i++) {
14335 page[i] = alloc_page(GFP_KERNEL);
14336 if (!page[i]) {
14337 if (i == 0)
14338 goto err_page0;
14339 else
14340 goto err_page1;
14341 }
14342
14343 /* this will set skb[i]->head_frag */
14344 skb[i] = dev_alloc_skb(headroom + data_size);
14345 if (!skb[i]) {
14346 if (i == 0)
14347 goto err_skb0;
14348 else
14349 goto err_skb1;
14350 }
14351
14352 skb_reserve(skb[i], headroom);
14353 skb_put(skb[i], data_size);
14354 skb[i]->protocol = htons(ETH_P_IP);
14355 skb_reset_network_header(skb[i]);
14356 skb_set_mac_header(skb[i], -ETH_HLEN);
14357
14358 skb_add_rx_frag(skb[i], 0, page[i], 0, 64, 64);
14359 // skb_headlen(skb[i]): 8, skb[i]->head_frag = 1
14360 }
14361
14362 /* setup shinfo */
14363 skb_shinfo(skb[0])->gso_size = 1448;
14364 skb_shinfo(skb[0])->gso_type = SKB_GSO_TCPV4;
14365 skb_shinfo(skb[0])->gso_type |= SKB_GSO_DODGY;
14366 skb_shinfo(skb[0])->gso_segs = 0;
14367 skb_shinfo(skb[0])->frag_list = skb[1];
Vadim Fedorenko3384c7c2021-09-10 01:04:09 +030014368 skb_shinfo(skb[0])->hwtstamps.hwtstamp = 1000;
Yonghong Song76db8082018-03-21 16:31:04 -070014369
14370 /* adjust skb[0]'s len */
14371 skb[0]->len += skb[1]->len;
14372 skb[0]->data_len += skb[1]->data_len;
14373 skb[0]->truesize += skb[1]->truesize;
14374
14375 return skb[0];
14376
14377err_skb1:
14378 __free_page(page[1]);
14379err_page1:
14380 kfree_skb(skb[0]);
14381err_skb0:
14382 __free_page(page[0]);
14383err_page0:
14384 return NULL;
14385}
14386
Shmulik Ladkanicf204a72019-10-25 16:42:23 +030014387static __init struct sk_buff *build_test_skb_linear_no_head_frag(void)
14388{
14389 unsigned int alloc_size = 2000;
14390 unsigned int headroom = 102, doffset = 72, data_size = 1308;
14391 struct sk_buff *skb[2];
14392 int i;
14393
14394 /* skbs linked in a frag_list, both with linear data, with head_frag=0
14395 * (data allocated by kmalloc), both have tcp data of 1308 bytes
14396 * (total payload is 2616 bytes).
14397 * Data offset is 72 bytes (40 ipv6 hdr, 32 tcp hdr). Some headroom.
14398 */
14399 for (i = 0; i < 2; i++) {
14400 skb[i] = alloc_skb(alloc_size, GFP_KERNEL);
14401 if (!skb[i]) {
14402 if (i == 0)
14403 goto err_skb0;
14404 else
14405 goto err_skb1;
14406 }
14407
14408 skb[i]->protocol = htons(ETH_P_IPV6);
14409 skb_reserve(skb[i], headroom);
14410 skb_put(skb[i], doffset + data_size);
14411 skb_reset_network_header(skb[i]);
14412 if (i == 0)
14413 skb_reset_mac_header(skb[i]);
14414 else
14415 skb_set_mac_header(skb[i], -ETH_HLEN);
14416 __skb_pull(skb[i], doffset);
14417 }
14418
14419 /* setup shinfo.
14420 * mimic bpf_skb_proto_4_to_6, which resets gso_segs and assigns a
14421 * reduced gso_size.
14422 */
14423 skb_shinfo(skb[0])->gso_size = 1288;
14424 skb_shinfo(skb[0])->gso_type = SKB_GSO_TCPV6 | SKB_GSO_DODGY;
14425 skb_shinfo(skb[0])->gso_segs = 0;
14426 skb_shinfo(skb[0])->frag_list = skb[1];
14427
14428 /* adjust skb[0]'s len */
14429 skb[0]->len += skb[1]->len;
14430 skb[0]->data_len += skb[1]->len;
14431 skb[0]->truesize += skb[1]->truesize;
14432
14433 return skb[0];
14434
14435err_skb1:
14436 kfree_skb(skb[0]);
14437err_skb0:
14438 return NULL;
14439}
14440
Shmulik Ladkaniaf21c712019-10-25 16:42:22 +030014441struct skb_segment_test {
14442 const char *descr;
14443 struct sk_buff *(*build_skb)(void);
Yonghong Song76db8082018-03-21 16:31:04 -070014444 netdev_features_t features;
Shmulik Ladkaniaf21c712019-10-25 16:42:22 +030014445};
14446
14447static struct skb_segment_test skb_segment_tests[] __initconst = {
14448 {
14449 .descr = "gso_with_rx_frags",
14450 .build_skb = build_test_skb,
14451 .features = NETIF_F_SG | NETIF_F_GSO_PARTIAL | NETIF_F_IP_CSUM |
14452 NETIF_F_IPV6_CSUM | NETIF_F_RXCSUM
Shmulik Ladkanicf204a72019-10-25 16:42:23 +030014453 },
14454 {
14455 .descr = "gso_linear_no_head_frag",
14456 .build_skb = build_test_skb_linear_no_head_frag,
14457 .features = NETIF_F_SG | NETIF_F_FRAGLIST |
14458 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_GSO |
14459 NETIF_F_LLTX_BIT | NETIF_F_GRO |
14460 NETIF_F_IPV6_CSUM | NETIF_F_RXCSUM |
14461 NETIF_F_HW_VLAN_STAG_TX_BIT
Shmulik Ladkaniaf21c712019-10-25 16:42:22 +030014462 }
14463};
14464
14465static __init int test_skb_segment_single(const struct skb_segment_test *test)
14466{
Yonghong Song76db8082018-03-21 16:31:04 -070014467 struct sk_buff *skb, *segs;
14468 int ret = -1;
14469
Shmulik Ladkaniaf21c712019-10-25 16:42:22 +030014470 skb = test->build_skb();
Yonghong Song76db8082018-03-21 16:31:04 -070014471 if (!skb) {
14472 pr_info("%s: failed to build_test_skb", __func__);
14473 goto done;
14474 }
14475
Shmulik Ladkaniaf21c712019-10-25 16:42:22 +030014476 segs = skb_segment(skb, test->features);
Dan Carpenter99fe29d2018-03-28 14:48:36 +030014477 if (!IS_ERR(segs)) {
Yonghong Song76db8082018-03-21 16:31:04 -070014478 kfree_skb_list(segs);
14479 ret = 0;
Yonghong Song76db8082018-03-21 16:31:04 -070014480 }
14481 kfree_skb(skb);
14482done:
14483 return ret;
14484}
14485
Shmulik Ladkaniaf21c712019-10-25 16:42:22 +030014486static __init int test_skb_segment(void)
14487{
14488 int i, err_cnt = 0, pass_cnt = 0;
14489
14490 for (i = 0; i < ARRAY_SIZE(skb_segment_tests); i++) {
14491 const struct skb_segment_test *test = &skb_segment_tests[i];
14492
Tiezhu Yangb066abb2021-10-28 09:25:21 +080014493 cond_resched();
14494 if (exclude_test(i))
14495 continue;
14496
Shmulik Ladkaniaf21c712019-10-25 16:42:22 +030014497 pr_info("#%d %s ", i, test->descr);
14498
14499 if (test_skb_segment_single(test)) {
14500 pr_cont("FAIL\n");
14501 err_cnt++;
14502 } else {
14503 pr_cont("PASS\n");
14504 pass_cnt++;
14505 }
14506 }
14507
14508 pr_info("%s: Summary: %d PASSED, %d FAILED\n", __func__,
14509 pass_cnt, err_cnt);
14510 return err_cnt ? -EINVAL : 0;
14511}
14512
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014513static __init int test_bpf(void)
14514{
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014515 int i, err_cnt = 0, pass_cnt = 0;
Daniel Borkmann327941f2015-04-30 16:17:27 +020014516 int jit_cnt = 0, run_cnt = 0;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014517
14518 for (i = 0; i < ARRAY_SIZE(tests); i++) {
Alexei Starovoitov7ae457c2014-07-30 20:34:16 -070014519 struct bpf_prog *fp;
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014520 int err;
14521
Eric Dumazetd40bc962018-02-26 10:52:46 -080014522 cond_resched();
Nicolas Schichand2648d42015-08-04 15:19:10 +020014523 if (exclude_test(i))
14524 continue;
14525
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014526 pr_info("#%d %s ", i, tests[i].descr);
14527
Johan Almbladh4bc35412021-09-14 11:18:30 +020014528 if (tests[i].fill_helper &&
14529 tests[i].fill_helper(&tests[i]) < 0) {
14530 pr_cont("FAIL to prog_fill\n");
14531 continue;
14532 }
14533
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014534 fp = generate_filter(i, &err);
Johan Almbladh4bc35412021-09-14 11:18:30 +020014535
14536 if (tests[i].fill_helper) {
14537 kfree(tests[i].u.ptr.insns);
14538 tests[i].u.ptr.insns = NULL;
14539 }
14540
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014541 if (fp == NULL) {
14542 if (err == 0) {
14543 pass_cnt++;
14544 continue;
14545 }
Alexei Starovoitov290af862018-01-09 10:04:29 -080014546 err_cnt++;
14547 continue;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014548 }
Daniel Borkmann327941f2015-04-30 16:17:27 +020014549
14550 pr_cont("jited:%u ", fp->jited);
14551
14552 run_cnt++;
14553 if (fp->jited)
14554 jit_cnt++;
14555
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014556 err = run_one(fp, &tests[i]);
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014557 release_filter(fp, i);
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014558
14559 if (err) {
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014560 pr_cont("FAIL (%d times)\n", err);
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014561 err_cnt++;
14562 } else {
14563 pr_cont("PASS\n");
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014564 pass_cnt++;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014565 }
14566 }
14567
Daniel Borkmann327941f2015-04-30 16:17:27 +020014568 pr_info("Summary: %d PASSED, %d FAILED, [%d/%d JIT'ed]\n",
14569 pass_cnt, err_cnt, jit_cnt, run_cnt);
14570
Daniel Borkmann10f18e02014-05-23 18:44:00 +020014571 return err_cnt ? -EINVAL : 0;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070014572}
14573
Johan Almbladh874be052021-08-09 11:18:29 +020014574struct tail_call_test {
14575 const char *descr;
14576 struct bpf_insn insns[MAX_INSNS];
Johan Almbladh18935a72021-09-14 11:18:41 +020014577 int flags;
Johan Almbladh874be052021-08-09 11:18:29 +020014578 int result;
14579 int stack_depth;
14580};
14581
Johan Almbladh18935a72021-09-14 11:18:41 +020014582/* Flags that can be passed to tail call test cases */
14583#define FLAG_NEED_STATE BIT(0)
14584#define FLAG_RESULT_IN_STATE BIT(1)
14585
Johan Almbladh874be052021-08-09 11:18:29 +020014586/*
14587 * Magic marker used in test snippets for tail calls below.
14588 * BPF_LD/MOV to R2 and R2 with this immediate value is replaced
14589 * with the proper values by the test runner.
14590 */
14591#define TAIL_CALL_MARKER 0x7a11ca11
14592
14593/* Special offset to indicate a NULL call target */
14594#define TAIL_CALL_NULL 0x7fff
14595
14596/* Special offset to indicate an out-of-range index */
14597#define TAIL_CALL_INVALID 0x7ffe
14598
14599#define TAIL_CALL(offset) \
14600 BPF_LD_IMM64(R2, TAIL_CALL_MARKER), \
14601 BPF_RAW_INSN(BPF_ALU | BPF_MOV | BPF_K, R3, 0, \
14602 offset, TAIL_CALL_MARKER), \
14603 BPF_JMP_IMM(BPF_TAIL_CALL, 0, 0, 0)
14604
14605/*
Johan Almbladh29eef852021-09-14 11:18:42 +020014606 * A test function to be called from a BPF program, clobbering a lot of
14607 * CPU registers in the process. A JITed BPF program calling this function
14608 * must save and restore any caller-saved registers it uses for internal
14609 * state, for example the current tail call count.
14610 */
14611BPF_CALL_1(bpf_test_func, u64, arg)
14612{
14613 char buf[64];
14614 long a = 0;
14615 long b = 1;
14616 long c = 2;
14617 long d = 3;
14618 long e = 4;
14619 long f = 5;
14620 long g = 6;
14621 long h = 7;
14622
14623 return snprintf(buf, sizeof(buf),
14624 "%ld %lu %lx %ld %lu %lx %ld %lu %x",
14625 a, b, c, d, e, f, g, h, (int)arg);
14626}
14627#define BPF_FUNC_test_func __BPF_FUNC_MAX_ID
14628
14629/*
Johan Almbladh874be052021-08-09 11:18:29 +020014630 * Tail call tests. Each test case may call any other test in the table,
14631 * including itself, specified as a relative index offset from the calling
14632 * test. The index TAIL_CALL_NULL can be used to specify a NULL target
14633 * function to test the JIT error path. Similarly, the index TAIL_CALL_INVALID
14634 * results in a target index that is out of range.
14635 */
14636static struct tail_call_test tail_call_tests[] = {
14637 {
14638 "Tail call leaf",
14639 .insns = {
14640 BPF_ALU64_REG(BPF_MOV, R0, R1),
14641 BPF_ALU64_IMM(BPF_ADD, R0, 1),
14642 BPF_EXIT_INSN(),
14643 },
14644 .result = 1,
14645 },
14646 {
14647 "Tail call 2",
14648 .insns = {
14649 BPF_ALU64_IMM(BPF_ADD, R1, 2),
14650 TAIL_CALL(-1),
14651 BPF_ALU64_IMM(BPF_MOV, R0, -1),
14652 BPF_EXIT_INSN(),
14653 },
14654 .result = 3,
14655 },
14656 {
14657 "Tail call 3",
14658 .insns = {
14659 BPF_ALU64_IMM(BPF_ADD, R1, 3),
14660 TAIL_CALL(-1),
14661 BPF_ALU64_IMM(BPF_MOV, R0, -1),
14662 BPF_EXIT_INSN(),
14663 },
14664 .result = 6,
14665 },
14666 {
14667 "Tail call 4",
14668 .insns = {
14669 BPF_ALU64_IMM(BPF_ADD, R1, 4),
14670 TAIL_CALL(-1),
14671 BPF_ALU64_IMM(BPF_MOV, R0, -1),
14672 BPF_EXIT_INSN(),
14673 },
14674 .result = 10,
14675 },
14676 {
14677 "Tail call error path, max count reached",
14678 .insns = {
Johan Almbladh18935a72021-09-14 11:18:41 +020014679 BPF_LDX_MEM(BPF_W, R2, R1, 0),
14680 BPF_ALU64_IMM(BPF_ADD, R2, 1),
14681 BPF_STX_MEM(BPF_W, R1, R2, 0),
Johan Almbladh874be052021-08-09 11:18:29 +020014682 TAIL_CALL(0),
14683 BPF_EXIT_INSN(),
14684 },
Johan Almbladh18935a72021-09-14 11:18:41 +020014685 .flags = FLAG_NEED_STATE | FLAG_RESULT_IN_STATE,
Tiezhu Yangebf7f6f2021-11-05 09:30:00 +080014686 .result = (MAX_TAIL_CALL_CNT + 1) * MAX_TESTRUNS,
Johan Almbladh874be052021-08-09 11:18:29 +020014687 },
14688 {
Johan Almbladh29eef852021-09-14 11:18:42 +020014689 "Tail call count preserved across function calls",
14690 .insns = {
14691 BPF_LDX_MEM(BPF_W, R2, R1, 0),
14692 BPF_ALU64_IMM(BPF_ADD, R2, 1),
14693 BPF_STX_MEM(BPF_W, R1, R2, 0),
14694 BPF_STX_MEM(BPF_DW, R10, R1, -8),
14695 BPF_CALL_REL(BPF_FUNC_get_numa_node_id),
14696 BPF_CALL_REL(BPF_FUNC_ktime_get_ns),
14697 BPF_CALL_REL(BPF_FUNC_ktime_get_boot_ns),
14698 BPF_CALL_REL(BPF_FUNC_ktime_get_coarse_ns),
14699 BPF_CALL_REL(BPF_FUNC_jiffies64),
14700 BPF_CALL_REL(BPF_FUNC_test_func),
14701 BPF_LDX_MEM(BPF_DW, R1, R10, -8),
14702 BPF_ALU32_REG(BPF_MOV, R0, R1),
14703 TAIL_CALL(0),
14704 BPF_EXIT_INSN(),
14705 },
14706 .stack_depth = 8,
14707 .flags = FLAG_NEED_STATE | FLAG_RESULT_IN_STATE,
Tiezhu Yangebf7f6f2021-11-05 09:30:00 +080014708 .result = (MAX_TAIL_CALL_CNT + 1) * MAX_TESTRUNS,
Johan Almbladh29eef852021-09-14 11:18:42 +020014709 },
14710 {
Johan Almbladh874be052021-08-09 11:18:29 +020014711 "Tail call error path, NULL target",
14712 .insns = {
Johan Almbladh18935a72021-09-14 11:18:41 +020014713 BPF_LDX_MEM(BPF_W, R2, R1, 0),
14714 BPF_ALU64_IMM(BPF_ADD, R2, 1),
14715 BPF_STX_MEM(BPF_W, R1, R2, 0),
Johan Almbladh874be052021-08-09 11:18:29 +020014716 TAIL_CALL(TAIL_CALL_NULL),
Johan Almbladh874be052021-08-09 11:18:29 +020014717 BPF_EXIT_INSN(),
14718 },
Johan Almbladh18935a72021-09-14 11:18:41 +020014719 .flags = FLAG_NEED_STATE | FLAG_RESULT_IN_STATE,
14720 .result = MAX_TESTRUNS,
Johan Almbladh874be052021-08-09 11:18:29 +020014721 },
14722 {
14723 "Tail call error path, index out of range",
14724 .insns = {
Johan Almbladh18935a72021-09-14 11:18:41 +020014725 BPF_LDX_MEM(BPF_W, R2, R1, 0),
14726 BPF_ALU64_IMM(BPF_ADD, R2, 1),
14727 BPF_STX_MEM(BPF_W, R1, R2, 0),
Johan Almbladh874be052021-08-09 11:18:29 +020014728 TAIL_CALL(TAIL_CALL_INVALID),
Johan Almbladh874be052021-08-09 11:18:29 +020014729 BPF_EXIT_INSN(),
14730 },
Johan Almbladh18935a72021-09-14 11:18:41 +020014731 .flags = FLAG_NEED_STATE | FLAG_RESULT_IN_STATE,
14732 .result = MAX_TESTRUNS,
Johan Almbladh874be052021-08-09 11:18:29 +020014733 },
14734};
14735
14736static void __init destroy_tail_call_tests(struct bpf_array *progs)
14737{
14738 int i;
14739
14740 for (i = 0; i < ARRAY_SIZE(tail_call_tests); i++)
14741 if (progs->ptrs[i])
14742 bpf_prog_free(progs->ptrs[i]);
14743 kfree(progs);
14744}
14745
14746static __init int prepare_tail_call_tests(struct bpf_array **pprogs)
14747{
14748 int ntests = ARRAY_SIZE(tail_call_tests);
14749 struct bpf_array *progs;
14750 int which, err;
14751
14752 /* Allocate the table of programs to be used for tall calls */
14753 progs = kzalloc(sizeof(*progs) + (ntests + 1) * sizeof(progs->ptrs[0]),
14754 GFP_KERNEL);
14755 if (!progs)
14756 goto out_nomem;
14757
14758 /* Create all eBPF programs and populate the table */
14759 for (which = 0; which < ntests; which++) {
14760 struct tail_call_test *test = &tail_call_tests[which];
14761 struct bpf_prog *fp;
14762 int len, i;
14763
14764 /* Compute the number of program instructions */
14765 for (len = 0; len < MAX_INSNS; len++) {
14766 struct bpf_insn *insn = &test->insns[len];
14767
14768 if (len < MAX_INSNS - 1 &&
14769 insn->code == (BPF_LD | BPF_DW | BPF_IMM))
14770 len++;
14771 if (insn->code == 0)
14772 break;
14773 }
14774
14775 /* Allocate and initialize the program */
14776 fp = bpf_prog_alloc(bpf_prog_size(len), 0);
14777 if (!fp)
14778 goto out_nomem;
14779
14780 fp->len = len;
14781 fp->type = BPF_PROG_TYPE_SOCKET_FILTER;
14782 fp->aux->stack_depth = test->stack_depth;
14783 memcpy(fp->insnsi, test->insns, len * sizeof(struct bpf_insn));
14784
14785 /* Relocate runtime tail call offsets and addresses */
14786 for (i = 0; i < len; i++) {
14787 struct bpf_insn *insn = &fp->insnsi[i];
Johan Almbladh29eef852021-09-14 11:18:42 +020014788 long addr = 0;
Johan Almbladh874be052021-08-09 11:18:29 +020014789
14790 switch (insn->code) {
14791 case BPF_LD | BPF_DW | BPF_IMM:
Johan Almbladh29eef852021-09-14 11:18:42 +020014792 if (insn->imm != TAIL_CALL_MARKER)
14793 break;
Johan Almbladh874be052021-08-09 11:18:29 +020014794 insn[0].imm = (u32)(long)progs;
14795 insn[1].imm = ((u64)(long)progs) >> 32;
14796 break;
14797
14798 case BPF_ALU | BPF_MOV | BPF_K:
Johan Almbladh29eef852021-09-14 11:18:42 +020014799 if (insn->imm != TAIL_CALL_MARKER)
14800 break;
Johan Almbladh874be052021-08-09 11:18:29 +020014801 if (insn->off == TAIL_CALL_NULL)
14802 insn->imm = ntests;
14803 else if (insn->off == TAIL_CALL_INVALID)
14804 insn->imm = ntests + 1;
14805 else
14806 insn->imm = which + insn->off;
14807 insn->off = 0;
Johan Almbladh29eef852021-09-14 11:18:42 +020014808 break;
14809
14810 case BPF_JMP | BPF_CALL:
14811 if (insn->src_reg != BPF_PSEUDO_CALL)
14812 break;
14813 switch (insn->imm) {
14814 case BPF_FUNC_get_numa_node_id:
14815 addr = (long)&numa_node_id;
14816 break;
14817 case BPF_FUNC_ktime_get_ns:
14818 addr = (long)&ktime_get_ns;
14819 break;
14820 case BPF_FUNC_ktime_get_boot_ns:
14821 addr = (long)&ktime_get_boot_fast_ns;
14822 break;
14823 case BPF_FUNC_ktime_get_coarse_ns:
14824 addr = (long)&ktime_get_coarse_ns;
14825 break;
14826 case BPF_FUNC_jiffies64:
14827 addr = (long)&get_jiffies_64;
14828 break;
14829 case BPF_FUNC_test_func:
14830 addr = (long)&bpf_test_func;
14831 break;
14832 default:
14833 err = -EFAULT;
14834 goto out_err;
14835 }
Kees Cook3d717fa2021-09-28 16:09:45 -070014836 *insn = BPF_EMIT_CALL(addr);
Johan Almbladh29eef852021-09-14 11:18:42 +020014837 if ((long)__bpf_call_base + insn->imm != addr)
14838 *insn = BPF_JMP_A(0); /* Skip: NOP */
14839 break;
Johan Almbladh874be052021-08-09 11:18:29 +020014840 }
14841 }
14842
14843 fp = bpf_prog_select_runtime(fp, &err);
14844 if (err)
14845 goto out_err;
14846
14847 progs->ptrs[which] = fp;
14848 }
14849
14850 /* The last entry contains a NULL program pointer */
14851 progs->map.max_entries = ntests + 1;
14852 *pprogs = progs;
14853 return 0;
14854
14855out_nomem:
14856 err = -ENOMEM;
14857
14858out_err:
14859 if (progs)
14860 destroy_tail_call_tests(progs);
14861 return err;
14862}
14863
14864static __init int test_tail_calls(struct bpf_array *progs)
14865{
14866 int i, err_cnt = 0, pass_cnt = 0;
14867 int jit_cnt = 0, run_cnt = 0;
14868
14869 for (i = 0; i < ARRAY_SIZE(tail_call_tests); i++) {
14870 struct tail_call_test *test = &tail_call_tests[i];
14871 struct bpf_prog *fp = progs->ptrs[i];
Johan Almbladh18935a72021-09-14 11:18:41 +020014872 int *data = NULL;
14873 int state = 0;
Johan Almbladh874be052021-08-09 11:18:29 +020014874 u64 duration;
14875 int ret;
14876
14877 cond_resched();
Tiezhu Yangb066abb2021-10-28 09:25:21 +080014878 if (exclude_test(i))
14879 continue;
Johan Almbladh874be052021-08-09 11:18:29 +020014880
14881 pr_info("#%d %s ", i, test->descr);
14882 if (!fp) {
14883 err_cnt++;
14884 continue;
14885 }
14886 pr_cont("jited:%u ", fp->jited);
14887
14888 run_cnt++;
14889 if (fp->jited)
14890 jit_cnt++;
14891
Johan Almbladh18935a72021-09-14 11:18:41 +020014892 if (test->flags & FLAG_NEED_STATE)
14893 data = &state;
14894 ret = __run_one(fp, data, MAX_TESTRUNS, &duration);
14895 if (test->flags & FLAG_RESULT_IN_STATE)
14896 ret = state;
Johan Almbladh874be052021-08-09 11:18:29 +020014897 if (ret == test->result) {
14898 pr_cont("%lld PASS", duration);
14899 pass_cnt++;
14900 } else {
14901 pr_cont("ret %d != %d FAIL", ret, test->result);
14902 err_cnt++;
14903 }
14904 }
14905
14906 pr_info("%s: Summary: %d PASSED, %d FAILED, [%d/%d JIT'ed]\n",
14907 __func__, pass_cnt, err_cnt, jit_cnt, run_cnt);
14908
14909 return err_cnt ? -EINVAL : 0;
14910}
14911
Tiezhu Yangb066abb2021-10-28 09:25:21 +080014912static char test_suite[32];
14913module_param_string(test_suite, test_suite, sizeof(test_suite), 0);
14914
14915static __init int find_test_index(const char *test_name)
14916{
14917 int i;
14918
14919 if (!strcmp(test_suite, "test_bpf")) {
14920 for (i = 0; i < ARRAY_SIZE(tests); i++) {
14921 if (!strcmp(tests[i].descr, test_name))
14922 return i;
14923 }
14924 }
14925
14926 if (!strcmp(test_suite, "test_tail_calls")) {
14927 for (i = 0; i < ARRAY_SIZE(tail_call_tests); i++) {
14928 if (!strcmp(tail_call_tests[i].descr, test_name))
14929 return i;
14930 }
14931 }
14932
14933 if (!strcmp(test_suite, "test_skb_segment")) {
14934 for (i = 0; i < ARRAY_SIZE(skb_segment_tests); i++) {
14935 if (!strcmp(skb_segment_tests[i].descr, test_name))
14936 return i;
14937 }
14938 }
14939
14940 return -1;
14941}
14942
14943static __init int prepare_test_range(void)
14944{
14945 int valid_range;
14946
14947 if (!strcmp(test_suite, "test_bpf"))
14948 valid_range = ARRAY_SIZE(tests);
14949 else if (!strcmp(test_suite, "test_tail_calls"))
14950 valid_range = ARRAY_SIZE(tail_call_tests);
14951 else if (!strcmp(test_suite, "test_skb_segment"))
14952 valid_range = ARRAY_SIZE(skb_segment_tests);
14953 else
14954 return 0;
14955
14956 if (test_id >= 0) {
14957 /*
14958 * if a test_id was specified, use test_range to
14959 * cover only that test.
14960 */
14961 if (test_id >= valid_range) {
14962 pr_err("test_bpf: invalid test_id specified for '%s' suite.\n",
14963 test_suite);
14964 return -EINVAL;
14965 }
14966
14967 test_range[0] = test_id;
14968 test_range[1] = test_id;
14969 } else if (*test_name) {
14970 /*
14971 * if a test_name was specified, find it and setup
14972 * test_range to cover only that test.
14973 */
14974 int idx = find_test_index(test_name);
14975
14976 if (idx < 0) {
14977 pr_err("test_bpf: no test named '%s' found for '%s' suite.\n",
14978 test_name, test_suite);
14979 return -EINVAL;
14980 }
14981 test_range[0] = idx;
14982 test_range[1] = idx;
14983 } else if (test_range[0] != 0 || test_range[1] != INT_MAX) {
14984 /*
14985 * check that the supplied test_range is valid.
14986 */
14987 if (test_range[0] < 0 || test_range[1] >= valid_range) {
14988 pr_err("test_bpf: test_range is out of bound for '%s' suite.\n",
14989 test_suite);
14990 return -EINVAL;
14991 }
14992
14993 if (test_range[1] < test_range[0]) {
14994 pr_err("test_bpf: test_range is ending before it starts.\n");
14995 return -EINVAL;
14996 }
14997 }
14998
14999 return 0;
15000}
15001
Alexei Starovoitov64a89462014-05-08 14:10:52 -070015002static int __init test_bpf_init(void)
15003{
Johan Almbladh874be052021-08-09 11:18:29 +020015004 struct bpf_array *progs = NULL;
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020015005 int ret;
15006
Tiezhu Yangb066abb2021-10-28 09:25:21 +080015007 if (strlen(test_suite) &&
15008 strcmp(test_suite, "test_bpf") &&
15009 strcmp(test_suite, "test_tail_calls") &&
15010 strcmp(test_suite, "test_skb_segment")) {
15011 pr_err("test_bpf: invalid test_suite '%s' specified.\n", test_suite);
15012 return -EINVAL;
15013 }
15014
15015 /*
15016 * if test_suite is not specified, but test_id, test_name or test_range
15017 * is specified, set 'test_bpf' as the default test suite.
15018 */
15019 if (!strlen(test_suite) &&
15020 (test_id != -1 || strlen(test_name) ||
15021 (test_range[0] != 0 || test_range[1] != INT_MAX))) {
15022 pr_info("test_bpf: set 'test_bpf' as the default test_suite.\n");
15023 strscpy(test_suite, "test_bpf", sizeof(test_suite));
15024 }
15025
15026 ret = prepare_test_range();
Daniel Borkmanna4afd37b2015-05-13 13:12:43 +020015027 if (ret < 0)
15028 return ret;
15029
Tiezhu Yangb066abb2021-10-28 09:25:21 +080015030 if (!strlen(test_suite) || !strcmp(test_suite, "test_bpf")) {
15031 ret = test_bpf();
15032 if (ret)
15033 return ret;
15034 }
Yonghong Song76db8082018-03-21 16:31:04 -070015035
Tiezhu Yangb066abb2021-10-28 09:25:21 +080015036 if (!strlen(test_suite) || !strcmp(test_suite, "test_tail_calls")) {
15037 ret = prepare_tail_call_tests(&progs);
15038 if (ret)
15039 return ret;
15040 ret = test_tail_calls(progs);
15041 destroy_tail_call_tests(progs);
15042 if (ret)
15043 return ret;
15044 }
Johan Almbladh874be052021-08-09 11:18:29 +020015045
Tiezhu Yangb066abb2021-10-28 09:25:21 +080015046 if (!strlen(test_suite) || !strcmp(test_suite, "test_skb_segment"))
15047 return test_skb_segment();
15048
15049 return 0;
Alexei Starovoitov64a89462014-05-08 14:10:52 -070015050}
15051
15052static void __exit test_bpf_exit(void)
15053{
15054}
15055
15056module_init(test_bpf_init);
15057module_exit(test_bpf_exit);
Daniel Borkmann10f18e02014-05-23 18:44:00 +020015058
Alexei Starovoitov64a89462014-05-08 14:10:52 -070015059MODULE_LICENSE("GPL");