|  | // SPDX-License-Identifier: GPL-2.0-only | 
|  | /* | 
|  | * Testsuite for BPF interpreter and BPF JIT compiler | 
|  | * | 
|  | * Copyright (c) 2011-2014 PLUMgrid, http://plumgrid.com | 
|  | */ | 
|  |  | 
|  | #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt | 
|  |  | 
|  | #include <linux/init.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/filter.h> | 
|  | #include <linux/bpf.h> | 
|  | #include <linux/skbuff.h> | 
|  | #include <linux/netdevice.h> | 
|  | #include <linux/if_vlan.h> | 
|  | #include <linux/prandom.h> | 
|  | #include <linux/highmem.h> | 
|  | #include <linux/sched.h> | 
|  |  | 
|  | /* General test specific settings */ | 
|  | #define MAX_SUBTESTS	3 | 
|  | #define MAX_TESTRUNS	1000 | 
|  | #define MAX_DATA	128 | 
|  | #define MAX_INSNS	512 | 
|  | #define MAX_K		0xffffFFFF | 
|  |  | 
|  | /* Few constants used to init test 'skb' */ | 
|  | #define SKB_TYPE	3 | 
|  | #define SKB_MARK	0x1234aaaa | 
|  | #define SKB_HASH	0x1234aaab | 
|  | #define SKB_QUEUE_MAP	123 | 
|  | #define SKB_VLAN_TCI	0xffff | 
|  | #define SKB_VLAN_PRESENT	1 | 
|  | #define SKB_DEV_IFINDEX	577 | 
|  | #define SKB_DEV_TYPE	588 | 
|  |  | 
|  | /* Redefine REGs to make tests less verbose */ | 
|  | #define R0		BPF_REG_0 | 
|  | #define R1		BPF_REG_1 | 
|  | #define R2		BPF_REG_2 | 
|  | #define R3		BPF_REG_3 | 
|  | #define R4		BPF_REG_4 | 
|  | #define R5		BPF_REG_5 | 
|  | #define R6		BPF_REG_6 | 
|  | #define R7		BPF_REG_7 | 
|  | #define R8		BPF_REG_8 | 
|  | #define R9		BPF_REG_9 | 
|  | #define R10		BPF_REG_10 | 
|  |  | 
|  | /* Flags that can be passed to test cases */ | 
|  | #define FLAG_NO_DATA		BIT(0) | 
|  | #define FLAG_EXPECTED_FAIL	BIT(1) | 
|  | #define FLAG_SKB_FRAG		BIT(2) | 
|  | #define FLAG_VERIFIER_ZEXT	BIT(3) | 
|  | #define FLAG_LARGE_MEM		BIT(4) | 
|  |  | 
|  | enum { | 
|  | CLASSIC  = BIT(6),	/* Old BPF instructions only. */ | 
|  | INTERNAL = BIT(7),	/* Extended instruction set.  */ | 
|  | }; | 
|  |  | 
|  | #define TEST_TYPE_MASK		(CLASSIC | INTERNAL) | 
|  |  | 
|  | struct bpf_test { | 
|  | const char *descr; | 
|  | union { | 
|  | struct sock_filter insns[MAX_INSNS]; | 
|  | struct bpf_insn insns_int[MAX_INSNS]; | 
|  | struct { | 
|  | void *insns; | 
|  | unsigned int len; | 
|  | } ptr; | 
|  | } u; | 
|  | __u8 aux; | 
|  | __u8 data[MAX_DATA]; | 
|  | struct { | 
|  | int data_size; | 
|  | __u32 result; | 
|  | } test[MAX_SUBTESTS]; | 
|  | int (*fill_helper)(struct bpf_test *self); | 
|  | int expected_errcode; /* used when FLAG_EXPECTED_FAIL is set in the aux */ | 
|  | __u8 frag_data[MAX_DATA]; | 
|  | int stack_depth; /* for eBPF only, since tests don't call verifier */ | 
|  | int nr_testruns; /* Custom run count, defaults to MAX_TESTRUNS if 0 */ | 
|  | }; | 
|  |  | 
|  | /* Large test cases need separate allocation and fill handler. */ | 
|  |  | 
|  | static int bpf_fill_maxinsns1(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | __u32 k = ~0; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | for (i = 0; i < len; i++, k--) | 
|  | insn[i] = __BPF_STMT(BPF_RET | BPF_K, k); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns2(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | for (i = 0; i < len; i++) | 
|  | insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns3(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | struct rnd_state rnd; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | prandom_seed_state(&rnd, 3141592653589793238ULL); | 
|  |  | 
|  | for (i = 0; i < len - 1; i++) { | 
|  | __u32 k = prandom_u32_state(&rnd); | 
|  |  | 
|  | insn[i] = __BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, k); | 
|  | } | 
|  |  | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns4(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS + 1; | 
|  | struct sock_filter *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | for (i = 0; i < len; i++) | 
|  | insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns5(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[0] = __BPF_JUMP(BPF_JMP | BPF_JA, len - 2, 0, 0); | 
|  |  | 
|  | for (i = 1; i < len - 1; i++) | 
|  | insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); | 
|  |  | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xabababab); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns6(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | for (i = 0; i < len - 1; i++) | 
|  | insn[i] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + | 
|  | SKF_AD_VLAN_TAG_PRESENT); | 
|  |  | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns7(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | for (i = 0; i < len - 4; i++) | 
|  | insn[i] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + | 
|  | SKF_AD_CPU); | 
|  |  | 
|  | insn[len - 4] = __BPF_STMT(BPF_MISC | BPF_TAX, 0); | 
|  | insn[len - 3] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + | 
|  | SKF_AD_CPU); | 
|  | insn[len - 2] = __BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0); | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns8(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | int i, jmp_off = len - 3; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[0] = __BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff); | 
|  |  | 
|  | for (i = 1; i < len - 1; i++) | 
|  | insn[i] = __BPF_JUMP(BPF_JMP | BPF_JGT, 0xffffffff, jmp_off--, 0); | 
|  |  | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns9(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct bpf_insn *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[0] = BPF_JMP_IMM(BPF_JA, 0, 0, len - 2); | 
|  | insn[1] = BPF_ALU32_IMM(BPF_MOV, R0, 0xcbababab); | 
|  | insn[2] = BPF_EXIT_INSN(); | 
|  |  | 
|  | for (i = 3; i < len - 2; i++) | 
|  | insn[i] = BPF_ALU32_IMM(BPF_MOV, R0, 0xfefefefe); | 
|  |  | 
|  | insn[len - 2] = BPF_EXIT_INSN(); | 
|  | insn[len - 1] = BPF_JMP_IMM(BPF_JA, 0, 0, -(len - 1)); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns10(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS, hlen = len - 2; | 
|  | struct bpf_insn *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | for (i = 0; i < hlen / 2; i++) | 
|  | insn[i] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen - 2 - 2 * i); | 
|  | for (i = hlen - 1; i > hlen / 2; i--) | 
|  | insn[i] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen - 1 - 2 * i); | 
|  |  | 
|  | insn[hlen / 2] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen / 2 - 1); | 
|  | insn[hlen]     = BPF_ALU32_IMM(BPF_MOV, R0, 0xabababac); | 
|  | insn[hlen + 1] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_ja(struct bpf_test *self, unsigned int len, | 
|  | unsigned int plen) | 
|  | { | 
|  | struct sock_filter *insn; | 
|  | unsigned int rlen; | 
|  | int i, j; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | rlen = (len % plen) - 1; | 
|  |  | 
|  | for (i = 0; i + plen < len; i += plen) | 
|  | for (j = 0; j < plen; j++) | 
|  | insn[i + j] = __BPF_JUMP(BPF_JMP | BPF_JA, | 
|  | plen - 1 - j, 0, 0); | 
|  | for (j = 0; j < rlen; j++) | 
|  | insn[i + j] = __BPF_JUMP(BPF_JMP | BPF_JA, rlen - 1 - j, | 
|  | 0, 0); | 
|  |  | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xababcbac); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns11(struct bpf_test *self) | 
|  | { | 
|  | /* Hits 70 passes on x86_64 and triggers NOPs padding. */ | 
|  | return __bpf_fill_ja(self, BPF_MAXINSNS, 68); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns12(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | int i = 0; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[0] = __BPF_JUMP(BPF_JMP | BPF_JA, len - 2, 0, 0); | 
|  |  | 
|  | for (i = 1; i < len - 1; i++) | 
|  | insn[i] = __BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0); | 
|  |  | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xabababab); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_maxinsns13(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | int i = 0; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | for (i = 0; i < len - 3; i++) | 
|  | insn[i] = __BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0); | 
|  |  | 
|  | insn[len - 3] = __BPF_STMT(BPF_LD | BPF_IMM, 0xabababab); | 
|  | insn[len - 2] = __BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0); | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_ja(struct bpf_test *self) | 
|  | { | 
|  | /* Hits exactly 11 passes on x86_64 JIT. */ | 
|  | return __bpf_fill_ja(self, 12, 9); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_ld_abs_get_processor_id(struct bpf_test *self) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct sock_filter *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | for (i = 0; i < len - 1; i += 2) { | 
|  | insn[i] = __BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 0); | 
|  | insn[i + 1] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_CPU); | 
|  | } | 
|  |  | 
|  | insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xbee); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_stxdw(struct bpf_test *self, int size) | 
|  | { | 
|  | unsigned int len = BPF_MAXINSNS; | 
|  | struct bpf_insn *insn; | 
|  | int i; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[0] = BPF_ALU32_IMM(BPF_MOV, R0, 1); | 
|  | insn[1] = BPF_ST_MEM(size, R10, -40, 42); | 
|  |  | 
|  | for (i = 2; i < len - 2; i++) | 
|  | insn[i] = BPF_STX_XADD(size, R10, R0, -40); | 
|  |  | 
|  | insn[len - 2] = BPF_LDX_MEM(size, R0, R10, -40); | 
|  | insn[len - 1] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  | self->stack_depth = 40; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_stxw(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_stxdw(self, BPF_W); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_stxdw(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_stxdw(self, BPF_DW); | 
|  | } | 
|  |  | 
|  | static int __bpf_ld_imm64(struct bpf_insn insns[2], u8 reg, s64 imm64) | 
|  | { | 
|  | struct bpf_insn tmp[] = {BPF_LD_IMM64(reg, imm64)}; | 
|  |  | 
|  | memcpy(insns, tmp, sizeof(tmp)); | 
|  | return 2; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Branch conversion tests. Complex operations can expand to a lot | 
|  | * of instructions when JITed. This in turn may cause jump offsets | 
|  | * to overflow the field size of the native instruction, triggering | 
|  | * a branch conversion mechanism in some JITs. | 
|  | */ | 
|  | static int __bpf_fill_max_jmp(struct bpf_test *self, int jmp, int imm, bool alu32) | 
|  | { | 
|  | struct bpf_insn *insns; | 
|  | int len = S16_MAX + 5; | 
|  | int i; | 
|  |  | 
|  | insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); | 
|  | if (!insns) | 
|  | return -ENOMEM; | 
|  |  | 
|  | i = __bpf_ld_imm64(insns, R1, 0x0123456789abcdefULL); | 
|  | insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1); | 
|  | insns[i++] = BPF_JMP_IMM(jmp, R0, imm, S16_MAX); | 
|  | insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 2); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | while (i < len - 1) { | 
|  | static const int ops[] = { | 
|  | BPF_LSH, BPF_RSH, BPF_ARSH, BPF_ADD, | 
|  | BPF_SUB, BPF_MUL, BPF_DIV, BPF_MOD, | 
|  | }; | 
|  | int op = ops[(i >> 1) % ARRAY_SIZE(ops)]; | 
|  |  | 
|  | if ((i & 1) || alu32) | 
|  | insns[i++] = BPF_ALU32_REG(op, R0, R1); | 
|  | else | 
|  | insns[i++] = BPF_ALU64_REG(op, R0, R1); | 
|  | } | 
|  |  | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  | self->u.ptr.insns = insns; | 
|  | self->u.ptr.len = len; | 
|  | BUG_ON(i != len); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Branch taken by runtime decision */ | 
|  | static int bpf_fill_max_jmp_taken_32(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_max_jmp(self, BPF_JEQ, 1, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_max_jmp_taken(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_max_jmp(self, BPF_JEQ, 1, false); | 
|  | } | 
|  |  | 
|  | /* Branch not taken by runtime decision */ | 
|  | static int bpf_fill_max_jmp_not_taken_32(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_max_jmp(self, BPF_JEQ, 0, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_max_jmp_not_taken(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_max_jmp(self, BPF_JEQ, 0, false); | 
|  | } | 
|  |  | 
|  | /* Branch always taken, known at JIT time */ | 
|  | static int bpf_fill_max_jmp_always_taken_32(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_max_jmp(self, BPF_JGE, 0, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_max_jmp_always_taken(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_max_jmp(self, BPF_JGE, 0, false); | 
|  | } | 
|  |  | 
|  | /* Branch never taken, known at JIT time */ | 
|  | static int bpf_fill_max_jmp_never_taken_32(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_max_jmp(self, BPF_JLT, 0, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_max_jmp_never_taken(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_max_jmp(self, BPF_JLT, 0, false); | 
|  | } | 
|  |  | 
|  | /* ALU result computation used in tests */ | 
|  | static bool __bpf_alu_result(u64 *res, u64 v1, u64 v2, u8 op) | 
|  | { | 
|  | *res = 0; | 
|  | switch (op) { | 
|  | case BPF_MOV: | 
|  | *res = v2; | 
|  | break; | 
|  | case BPF_AND: | 
|  | *res = v1 & v2; | 
|  | break; | 
|  | case BPF_OR: | 
|  | *res = v1 | v2; | 
|  | break; | 
|  | case BPF_XOR: | 
|  | *res = v1 ^ v2; | 
|  | break; | 
|  | case BPF_LSH: | 
|  | *res = v1 << v2; | 
|  | break; | 
|  | case BPF_RSH: | 
|  | *res = v1 >> v2; | 
|  | break; | 
|  | case BPF_ARSH: | 
|  | *res = v1 >> v2; | 
|  | if (v2 > 0 && v1 > S64_MAX) | 
|  | *res |= ~0ULL << (64 - v2); | 
|  | break; | 
|  | case BPF_ADD: | 
|  | *res = v1 + v2; | 
|  | break; | 
|  | case BPF_SUB: | 
|  | *res = v1 - v2; | 
|  | break; | 
|  | case BPF_MUL: | 
|  | *res = v1 * v2; | 
|  | break; | 
|  | case BPF_DIV: | 
|  | if (v2 == 0) | 
|  | return false; | 
|  | *res = div64_u64(v1, v2); | 
|  | break; | 
|  | case BPF_MOD: | 
|  | if (v2 == 0) | 
|  | return false; | 
|  | div64_u64_rem(v1, v2, res); | 
|  | break; | 
|  | } | 
|  | return true; | 
|  | } | 
|  |  | 
|  | /* Test an ALU shift operation for all valid shift values */ | 
|  | static int __bpf_fill_alu_shift(struct bpf_test *self, u8 op, | 
|  | u8 mode, bool alu32) | 
|  | { | 
|  | static const s64 regs[] = { | 
|  | 0x0123456789abcdefLL, /* dword > 0, word < 0 */ | 
|  | 0xfedcba9876543210LL, /* dword < 0, word > 0 */ | 
|  | 0xfedcba0198765432LL, /* dword < 0, word < 0 */ | 
|  | 0x0123458967abcdefLL, /* dword > 0, word > 0 */ | 
|  | }; | 
|  | int bits = alu32 ? 32 : 64; | 
|  | int len = (2 + 7 * bits) * ARRAY_SIZE(regs) + 3; | 
|  | struct bpf_insn *insn; | 
|  | int imm, k; | 
|  | int i = 0; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0); | 
|  |  | 
|  | for (k = 0; k < ARRAY_SIZE(regs); k++) { | 
|  | s64 reg = regs[k]; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insn[i], R3, reg); | 
|  |  | 
|  | for (imm = 0; imm < bits; imm++) { | 
|  | u64 val; | 
|  |  | 
|  | /* Perform operation */ | 
|  | insn[i++] = BPF_ALU64_REG(BPF_MOV, R1, R3); | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R2, imm); | 
|  | if (alu32) { | 
|  | if (mode == BPF_K) | 
|  | insn[i++] = BPF_ALU32_IMM(op, R1, imm); | 
|  | else | 
|  | insn[i++] = BPF_ALU32_REG(op, R1, R2); | 
|  |  | 
|  | if (op == BPF_ARSH) | 
|  | reg = (s32)reg; | 
|  | else | 
|  | reg = (u32)reg; | 
|  | __bpf_alu_result(&val, reg, imm, op); | 
|  | val = (u32)val; | 
|  | } else { | 
|  | if (mode == BPF_K) | 
|  | insn[i++] = BPF_ALU64_IMM(op, R1, imm); | 
|  | else | 
|  | insn[i++] = BPF_ALU64_REG(op, R1, R2); | 
|  | __bpf_alu_result(&val, reg, imm, op); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * When debugging a JIT that fails this test, one | 
|  | * can write the immediate value to R0 here to find | 
|  | * out which operand values that fail. | 
|  | */ | 
|  |  | 
|  | /* Load reference and check the result */ | 
|  | i += __bpf_ld_imm64(&insn[i], R4, val); | 
|  | insn[i++] = BPF_JMP_REG(BPF_JEQ, R1, R4, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  | } | 
|  |  | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  | BUG_ON(i != len); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_lsh_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_LSH, BPF_K, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_rsh_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_RSH, BPF_K, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_arsh_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_ARSH, BPF_K, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_lsh_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_LSH, BPF_X, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_rsh_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_RSH, BPF_X, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_arsh_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_ARSH, BPF_X, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_lsh_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_LSH, BPF_K, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_rsh_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_RSH, BPF_K, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_arsh_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_ARSH, BPF_K, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_lsh_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_LSH, BPF_X, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_rsh_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_RSH, BPF_X, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_arsh_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift(self, BPF_ARSH, BPF_X, true); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Test an ALU register shift operation for all valid shift values | 
|  | * for the case when the source and destination are the same. | 
|  | */ | 
|  | static int __bpf_fill_alu_shift_same_reg(struct bpf_test *self, u8 op, | 
|  | bool alu32) | 
|  | { | 
|  | int bits = alu32 ? 32 : 64; | 
|  | int len = 3 + 6 * bits; | 
|  | struct bpf_insn *insn; | 
|  | int i = 0; | 
|  | u64 val; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0); | 
|  |  | 
|  | for (val = 0; val < bits; val++) { | 
|  | u64 res; | 
|  |  | 
|  | /* Perform operation */ | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R1, val); | 
|  | if (alu32) | 
|  | insn[i++] = BPF_ALU32_REG(op, R1, R1); | 
|  | else | 
|  | insn[i++] = BPF_ALU64_REG(op, R1, R1); | 
|  |  | 
|  | /* Compute the reference result */ | 
|  | __bpf_alu_result(&res, val, val, op); | 
|  | if (alu32) | 
|  | res = (u32)res; | 
|  | i += __bpf_ld_imm64(&insn[i], R2, res); | 
|  |  | 
|  | /* Check the actual result */ | 
|  | insn[i++] = BPF_JMP_REG(BPF_JEQ, R1, R2, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  | BUG_ON(i != len); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_lsh_same_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift_same_reg(self, BPF_LSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_rsh_same_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift_same_reg(self, BPF_RSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_arsh_same_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift_same_reg(self, BPF_ARSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_lsh_same_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift_same_reg(self, BPF_LSH, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_rsh_same_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift_same_reg(self, BPF_RSH, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_arsh_same_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_shift_same_reg(self, BPF_ARSH, true); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Common operand pattern generator for exhaustive power-of-two magnitudes | 
|  | * tests. The block size parameters can be adjusted to increase/reduce the | 
|  | * number of combinatons tested and thereby execution speed and memory | 
|  | * footprint. | 
|  | */ | 
|  |  | 
|  | static inline s64 value(int msb, int delta, int sign) | 
|  | { | 
|  | return sign * (1LL << msb) + delta; | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_pattern(struct bpf_test *self, void *arg, | 
|  | int dbits, int sbits, int block1, int block2, | 
|  | int (*emit)(struct bpf_test*, void*, | 
|  | struct bpf_insn*, s64, s64)) | 
|  | { | 
|  | static const int sgn[][2] = {{1, 1}, {1, -1}, {-1, 1}, {-1, -1}}; | 
|  | struct bpf_insn *insns; | 
|  | int di, si, bt, db, sb; | 
|  | int count, len, k; | 
|  | int extra = 1 + 2; | 
|  | int i = 0; | 
|  |  | 
|  | /* Total number of iterations for the two pattern */ | 
|  | count = (dbits - 1) * (sbits - 1) * block1 * block1 * ARRAY_SIZE(sgn); | 
|  | count += (max(dbits, sbits) - 1) * block2 * block2 * ARRAY_SIZE(sgn); | 
|  |  | 
|  | /* Compute the maximum number of insns and allocate the buffer */ | 
|  | len = extra + count * (*emit)(self, arg, NULL, 0, 0); | 
|  | insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); | 
|  | if (!insns) | 
|  | return -ENOMEM; | 
|  |  | 
|  | /* Add head instruction(s) */ | 
|  | insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0); | 
|  |  | 
|  | /* | 
|  | * Pattern 1: all combinations of power-of-two magnitudes and sign, | 
|  | * and with a block of contiguous values around each magnitude. | 
|  | */ | 
|  | for (di = 0; di < dbits - 1; di++)                 /* Dst magnitudes */ | 
|  | for (si = 0; si < sbits - 1; si++)         /* Src magnitudes */ | 
|  | for (k = 0; k < ARRAY_SIZE(sgn); k++) /* Sign combos */ | 
|  | for (db = -(block1 / 2); | 
|  | db < (block1 + 1) / 2; db++) | 
|  | for (sb = -(block1 / 2); | 
|  | sb < (block1 + 1) / 2; sb++) { | 
|  | s64 dst, src; | 
|  |  | 
|  | dst = value(di, db, sgn[k][0]); | 
|  | src = value(si, sb, sgn[k][1]); | 
|  | i += (*emit)(self, arg, | 
|  | &insns[i], | 
|  | dst, src); | 
|  | } | 
|  | /* | 
|  | * Pattern 2: all combinations for a larger block of values | 
|  | * for each power-of-two magnitude and sign, where the magnitude is | 
|  | * the same for both operands. | 
|  | */ | 
|  | for (bt = 0; bt < max(dbits, sbits) - 1; bt++)        /* Magnitude   */ | 
|  | for (k = 0; k < ARRAY_SIZE(sgn); k++)         /* Sign combos */ | 
|  | for (db = -(block2 / 2); db < (block2 + 1) / 2; db++) | 
|  | for (sb = -(block2 / 2); | 
|  | sb < (block2 + 1) / 2; sb++) { | 
|  | s64 dst, src; | 
|  |  | 
|  | dst = value(bt % dbits, db, sgn[k][0]); | 
|  | src = value(bt % sbits, sb, sgn[k][1]); | 
|  | i += (*emit)(self, arg, &insns[i], | 
|  | dst, src); | 
|  | } | 
|  |  | 
|  | /* Append tail instructions */ | 
|  | insns[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  | BUG_ON(i > len); | 
|  |  | 
|  | self->u.ptr.insns = insns; | 
|  | self->u.ptr.len = i; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Block size parameters used in pattern tests below. une as needed to | 
|  | * increase/reduce the number combinations tested, see following examples. | 
|  | *        block   values per operand MSB | 
|  | * ---------------------------------------- | 
|  | *           0     none | 
|  | *           1     (1 << MSB) | 
|  | *           2     (1 << MSB) + [-1, 0] | 
|  | *           3     (1 << MSB) + [-1, 0, 1] | 
|  | */ | 
|  | #define PATTERN_BLOCK1 1 | 
|  | #define PATTERN_BLOCK2 5 | 
|  |  | 
|  | /* Number of test runs for a pattern test */ | 
|  | #define NR_PATTERN_RUNS 1 | 
|  |  | 
|  | /* | 
|  | * Exhaustive tests of ALU operations for all combinations of power-of-two | 
|  | * magnitudes of the operands, both for positive and negative values. The | 
|  | * test is designed to verify e.g. the ALU and ALU64 operations for JITs that | 
|  | * emit different code depending on the magnitude of the immediate value. | 
|  | */ | 
|  | static int __bpf_emit_alu64_imm(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 imm) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  | int i = 0; | 
|  | u64 res; | 
|  |  | 
|  | if (!insns) | 
|  | return 7; | 
|  |  | 
|  | if (__bpf_alu_result(&res, dst, (s32)imm, op)) { | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R3, res); | 
|  | insns[i++] = BPF_ALU64_IMM(op, R1, imm); | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_alu32_imm(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 imm) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  | int i = 0; | 
|  | u64 res; | 
|  |  | 
|  | if (!insns) | 
|  | return 7; | 
|  |  | 
|  | if (__bpf_alu_result(&res, (u32)dst, (u32)imm, op)) { | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R3, (u32)res); | 
|  | insns[i++] = BPF_ALU32_IMM(op, R1, imm); | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_alu64_reg(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 src) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  | int i = 0; | 
|  | u64 res; | 
|  |  | 
|  | if (!insns) | 
|  | return 9; | 
|  |  | 
|  | if (__bpf_alu_result(&res, dst, src, op)) { | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R2, src); | 
|  | i += __bpf_ld_imm64(&insns[i], R3, res); | 
|  | insns[i++] = BPF_ALU64_REG(op, R1, R2); | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_alu32_reg(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 src) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  | int i = 0; | 
|  | u64 res; | 
|  |  | 
|  | if (!insns) | 
|  | return 9; | 
|  |  | 
|  | if (__bpf_alu_result(&res, (u32)dst, (u32)src, op)) { | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R2, src); | 
|  | i += __bpf_ld_imm64(&insns[i], R3, (u32)res); | 
|  | insns[i++] = BPF_ALU32_REG(op, R1, R2); | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_alu64_imm(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 32, | 
|  | PATTERN_BLOCK1, PATTERN_BLOCK2, | 
|  | &__bpf_emit_alu64_imm); | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_alu32_imm(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 32, | 
|  | PATTERN_BLOCK1, PATTERN_BLOCK2, | 
|  | &__bpf_emit_alu32_imm); | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_alu64_reg(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 64, | 
|  | PATTERN_BLOCK1, PATTERN_BLOCK2, | 
|  | &__bpf_emit_alu64_reg); | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_alu32_reg(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 64, | 
|  | PATTERN_BLOCK1, PATTERN_BLOCK2, | 
|  | &__bpf_emit_alu32_reg); | 
|  | } | 
|  |  | 
|  | /* ALU64 immediate operations */ | 
|  | static int bpf_fill_alu64_mov_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_MOV); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_and_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_AND); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_or_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_OR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_xor_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_XOR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_add_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_ADD); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_sub_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_SUB); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_mul_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_MUL); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_div_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_DIV); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_mod_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_imm(self, BPF_MOD); | 
|  | } | 
|  |  | 
|  | /* ALU32 immediate operations */ | 
|  | static int bpf_fill_alu32_mov_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_MOV); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_and_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_AND); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_or_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_OR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_xor_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_XOR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_add_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_ADD); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_sub_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_SUB); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_mul_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_MUL); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_div_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_DIV); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_mod_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_imm(self, BPF_MOD); | 
|  | } | 
|  |  | 
|  | /* ALU64 register operations */ | 
|  | static int bpf_fill_alu64_mov_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_MOV); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_and_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_AND); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_or_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_OR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_xor_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_XOR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_add_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_ADD); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_sub_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_SUB); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_mul_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_MUL); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_div_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_DIV); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_mod_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu64_reg(self, BPF_MOD); | 
|  | } | 
|  |  | 
|  | /* ALU32 register operations */ | 
|  | static int bpf_fill_alu32_mov_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_MOV); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_and_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_AND); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_or_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_OR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_xor_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_XOR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_add_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_ADD); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_sub_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_SUB); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_mul_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_MUL); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_div_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_DIV); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_mod_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu32_reg(self, BPF_MOD); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Test JITs that implement complex ALU operations as function | 
|  | * calls, and must re-arrange operands for argument passing. | 
|  | */ | 
|  | static int __bpf_fill_alu_imm_regs(struct bpf_test *self, u8 op, bool alu32) | 
|  | { | 
|  | int len = 2 + 10 * 10; | 
|  | struct bpf_insn *insns; | 
|  | u64 dst, res; | 
|  | int i = 0; | 
|  | u32 imm; | 
|  | int rd; | 
|  |  | 
|  | insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); | 
|  | if (!insns) | 
|  | return -ENOMEM; | 
|  |  | 
|  | /* Operand and result values according to operation */ | 
|  | if (alu32) | 
|  | dst = 0x76543210U; | 
|  | else | 
|  | dst = 0x7edcba9876543210ULL; | 
|  | imm = 0x01234567U; | 
|  |  | 
|  | if (op == BPF_LSH || op == BPF_RSH || op == BPF_ARSH) | 
|  | imm &= 31; | 
|  |  | 
|  | __bpf_alu_result(&res, dst, imm, op); | 
|  |  | 
|  | if (alu32) | 
|  | res = (u32)res; | 
|  |  | 
|  | /* Check all operand registers */ | 
|  | for (rd = R0; rd <= R9; rd++) { | 
|  | i += __bpf_ld_imm64(&insns[i], rd, dst); | 
|  |  | 
|  | if (alu32) | 
|  | insns[i++] = BPF_ALU32_IMM(op, rd, imm); | 
|  | else | 
|  | insns[i++] = BPF_ALU64_IMM(op, rd, imm); | 
|  |  | 
|  | insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, res, 2); | 
|  | insns[i++] = BPF_MOV64_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_ALU64_IMM(BPF_RSH, rd, 32); | 
|  | insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, res >> 32, 2); | 
|  | insns[i++] = BPF_MOV64_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | insns[i++] = BPF_MOV64_IMM(R0, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insns; | 
|  | self->u.ptr.len = len; | 
|  | BUG_ON(i != len); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* ALU64 K registers */ | 
|  | static int bpf_fill_alu64_mov_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_MOV, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_and_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_AND, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_or_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_OR, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_xor_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_XOR, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_lsh_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_LSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_rsh_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_RSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_arsh_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_ARSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_add_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_ADD, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_sub_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_SUB, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_mul_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_MUL, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_div_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_DIV, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_mod_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_MOD, false); | 
|  | } | 
|  |  | 
|  | /* ALU32 K registers */ | 
|  | static int bpf_fill_alu32_mov_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_MOV, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_and_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_AND, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_or_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_OR, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_xor_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_XOR, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_lsh_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_LSH, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_rsh_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_RSH, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_arsh_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_ARSH, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_add_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_ADD, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_sub_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_SUB, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_mul_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_MUL, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_div_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_DIV, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_mod_imm_regs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_imm_regs(self, BPF_MOD, true); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Test JITs that implement complex ALU operations as function | 
|  | * calls, and must re-arrange operands for argument passing. | 
|  | */ | 
|  | static int __bpf_fill_alu_reg_pairs(struct bpf_test *self, u8 op, bool alu32) | 
|  | { | 
|  | int len = 2 + 10 * 10 * 12; | 
|  | u64 dst, src, res, same; | 
|  | struct bpf_insn *insns; | 
|  | int rd, rs; | 
|  | int i = 0; | 
|  |  | 
|  | insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); | 
|  | if (!insns) | 
|  | return -ENOMEM; | 
|  |  | 
|  | /* Operand and result values according to operation */ | 
|  | if (alu32) { | 
|  | dst = 0x76543210U; | 
|  | src = 0x01234567U; | 
|  | } else { | 
|  | dst = 0x7edcba9876543210ULL; | 
|  | src = 0x0123456789abcdefULL; | 
|  | } | 
|  |  | 
|  | if (op == BPF_LSH || op == BPF_RSH || op == BPF_ARSH) | 
|  | src &= 31; | 
|  |  | 
|  | __bpf_alu_result(&res, dst, src, op); | 
|  | __bpf_alu_result(&same, src, src, op); | 
|  |  | 
|  | if (alu32) { | 
|  | res = (u32)res; | 
|  | same = (u32)same; | 
|  | } | 
|  |  | 
|  | /* Check all combinations of operand registers */ | 
|  | for (rd = R0; rd <= R9; rd++) { | 
|  | for (rs = R0; rs <= R9; rs++) { | 
|  | u64 val = rd == rs ? same : res; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], rd, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], rs, src); | 
|  |  | 
|  | if (alu32) | 
|  | insns[i++] = BPF_ALU32_REG(op, rd, rs); | 
|  | else | 
|  | insns[i++] = BPF_ALU64_REG(op, rd, rs); | 
|  |  | 
|  | insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, val, 2); | 
|  | insns[i++] = BPF_MOV64_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_ALU64_IMM(BPF_RSH, rd, 32); | 
|  | insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, val >> 32, 2); | 
|  | insns[i++] = BPF_MOV64_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  | } | 
|  |  | 
|  | insns[i++] = BPF_MOV64_IMM(R0, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insns; | 
|  | self->u.ptr.len = len; | 
|  | BUG_ON(i != len); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* ALU64 X register combinations */ | 
|  | static int bpf_fill_alu64_mov_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_MOV, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_and_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_AND, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_or_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_OR, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_xor_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_XOR, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_lsh_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_LSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_rsh_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_RSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_arsh_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_ARSH, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_add_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_ADD, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_sub_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_SUB, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_mul_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_MUL, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_div_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_DIV, false); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu64_mod_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_MOD, false); | 
|  | } | 
|  |  | 
|  | /* ALU32 X register combinations */ | 
|  | static int bpf_fill_alu32_mov_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_MOV, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_and_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_AND, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_or_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_OR, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_xor_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_XOR, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_lsh_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_LSH, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_rsh_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_RSH, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_arsh_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_ARSH, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_add_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_ADD, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_sub_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_SUB, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_mul_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_MUL, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_div_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_DIV, true); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_alu32_mod_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_alu_reg_pairs(self, BPF_MOD, true); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Exhaustive tests of atomic operations for all power-of-two operand | 
|  | * magnitudes, both for positive and negative values. | 
|  | */ | 
|  |  | 
|  | static int __bpf_emit_atomic64(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 src) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  | u64 keep, fetch, res; | 
|  | int i = 0; | 
|  |  | 
|  | if (!insns) | 
|  | return 21; | 
|  |  | 
|  | switch (op) { | 
|  | case BPF_XCHG: | 
|  | res = src; | 
|  | break; | 
|  | default: | 
|  | __bpf_alu_result(&res, dst, src, BPF_OP(op)); | 
|  | } | 
|  |  | 
|  | keep = 0x0123456789abcdefULL; | 
|  | if (op & BPF_FETCH) | 
|  | fetch = dst; | 
|  | else | 
|  | fetch = src; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], R0, keep); | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R2, src); | 
|  | i += __bpf_ld_imm64(&insns[i], R3, res); | 
|  | i += __bpf_ld_imm64(&insns[i], R4, fetch); | 
|  | i += __bpf_ld_imm64(&insns[i], R5, keep); | 
|  |  | 
|  | insns[i++] = BPF_STX_MEM(BPF_DW, R10, R1, -8); | 
|  | insns[i++] = BPF_ATOMIC_OP(BPF_DW, op, R10, R2, -8); | 
|  | insns[i++] = BPF_LDX_MEM(BPF_DW, R1, R10, -8); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R4, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R5, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_atomic32(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 src) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  | u64 keep, fetch, res; | 
|  | int i = 0; | 
|  |  | 
|  | if (!insns) | 
|  | return 21; | 
|  |  | 
|  | switch (op) { | 
|  | case BPF_XCHG: | 
|  | res = src; | 
|  | break; | 
|  | default: | 
|  | __bpf_alu_result(&res, (u32)dst, (u32)src, BPF_OP(op)); | 
|  | } | 
|  |  | 
|  | keep = 0x0123456789abcdefULL; | 
|  | if (op & BPF_FETCH) | 
|  | fetch = (u32)dst; | 
|  | else | 
|  | fetch = src; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], R0, keep); | 
|  | i += __bpf_ld_imm64(&insns[i], R1, (u32)dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R2, src); | 
|  | i += __bpf_ld_imm64(&insns[i], R3, (u32)res); | 
|  | i += __bpf_ld_imm64(&insns[i], R4, fetch); | 
|  | i += __bpf_ld_imm64(&insns[i], R5, keep); | 
|  |  | 
|  | insns[i++] = BPF_STX_MEM(BPF_W, R10, R1, -4); | 
|  | insns[i++] = BPF_ATOMIC_OP(BPF_W, op, R10, R2, -4); | 
|  | insns[i++] = BPF_LDX_MEM(BPF_W, R1, R10, -4); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R4, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R5, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_cmpxchg64(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 src) | 
|  | { | 
|  | int i = 0; | 
|  |  | 
|  | if (!insns) | 
|  | return 23; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], R0, ~dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R2, src); | 
|  |  | 
|  | /* Result unsuccessful */ | 
|  | insns[i++] = BPF_STX_MEM(BPF_DW, R10, R1, -8); | 
|  | insns[i++] = BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -8); | 
|  | insns[i++] = BPF_LDX_MEM(BPF_DW, R3, R10, -8); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R1, R3, 2); | 
|  | insns[i++] = BPF_MOV64_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R3, 2); | 
|  | insns[i++] = BPF_MOV64_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | /* Result successful */ | 
|  | insns[i++] = BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -8); | 
|  | insns[i++] = BPF_LDX_MEM(BPF_DW, R3, R10, -8); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R2, R3, 2); | 
|  | insns[i++] = BPF_MOV64_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R1, 2); | 
|  | insns[i++] = BPF_MOV64_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_cmpxchg32(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 src) | 
|  | { | 
|  | int i = 0; | 
|  |  | 
|  | if (!insns) | 
|  | return 27; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], R0, ~dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R1, (u32)dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R2, src); | 
|  |  | 
|  | /* Result unsuccessful */ | 
|  | insns[i++] = BPF_STX_MEM(BPF_W, R10, R1, -4); | 
|  | insns[i++] = BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R2, -4); | 
|  | insns[i++] = BPF_ZEXT_REG(R0); /* Zext always inserted by verifier */ | 
|  | insns[i++] = BPF_LDX_MEM(BPF_W, R3, R10, -4); | 
|  |  | 
|  | insns[i++] = BPF_JMP32_REG(BPF_JEQ, R1, R3, 2); | 
|  | insns[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R3, 2); | 
|  | insns[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | /* Result successful */ | 
|  | i += __bpf_ld_imm64(&insns[i], R0, dst); | 
|  | insns[i++] = BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R2, -4); | 
|  | insns[i++] = BPF_ZEXT_REG(R0); /* Zext always inserted by verifier */ | 
|  | insns[i++] = BPF_LDX_MEM(BPF_W, R3, R10, -4); | 
|  |  | 
|  | insns[i++] = BPF_JMP32_REG(BPF_JEQ, R2, R3, 2); | 
|  | insns[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | insns[i++] = BPF_JMP_REG(BPF_JEQ, R0, R1, 2); | 
|  | insns[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_atomic64(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 64, | 
|  | 0, PATTERN_BLOCK2, | 
|  | &__bpf_emit_atomic64); | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_atomic32(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 64, | 
|  | 0, PATTERN_BLOCK2, | 
|  | &__bpf_emit_atomic32); | 
|  | } | 
|  |  | 
|  | /* 64-bit atomic operations */ | 
|  | static int bpf_fill_atomic64_add(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_ADD); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_and(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_AND); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_or(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_OR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_xor(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_XOR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_add_fetch(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_ADD | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_and_fetch(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_AND | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_or_fetch(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_OR | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_xor_fetch(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_XOR | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_xchg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic64(self, BPF_XCHG); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_cmpxchg64(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_pattern(self, NULL, 64, 64, 0, PATTERN_BLOCK2, | 
|  | &__bpf_emit_cmpxchg64); | 
|  | } | 
|  |  | 
|  | /* 32-bit atomic operations */ | 
|  | static int bpf_fill_atomic32_add(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_ADD); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_and(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_AND); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_or(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_OR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_xor(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_XOR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_add_fetch(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_ADD | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_and_fetch(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_AND | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_or_fetch(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_OR | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_xor_fetch(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_XOR | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_xchg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic32(self, BPF_XCHG); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_cmpxchg32(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_pattern(self, NULL, 64, 64, 0, PATTERN_BLOCK2, | 
|  | &__bpf_emit_cmpxchg32); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Test JITs that implement ATOMIC operations as function calls or | 
|  | * other primitives, and must re-arrange operands for argument passing. | 
|  | */ | 
|  | static int __bpf_fill_atomic_reg_pairs(struct bpf_test *self, u8 width, u8 op) | 
|  | { | 
|  | struct bpf_insn *insn; | 
|  | int len = 2 + 34 * 10 * 10; | 
|  | u64 mem, upd, res; | 
|  | int rd, rs, i = 0; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | /* Operand and memory values */ | 
|  | if (width == BPF_DW) { | 
|  | mem = 0x0123456789abcdefULL; | 
|  | upd = 0xfedcba9876543210ULL; | 
|  | } else { /* BPF_W */ | 
|  | mem = 0x01234567U; | 
|  | upd = 0x76543210U; | 
|  | } | 
|  |  | 
|  | /* Memory updated according to operation */ | 
|  | switch (op) { | 
|  | case BPF_XCHG: | 
|  | res = upd; | 
|  | break; | 
|  | case BPF_CMPXCHG: | 
|  | res = mem; | 
|  | break; | 
|  | default: | 
|  | __bpf_alu_result(&res, mem, upd, BPF_OP(op)); | 
|  | } | 
|  |  | 
|  | /* Test all operand registers */ | 
|  | for (rd = R0; rd <= R9; rd++) { | 
|  | for (rs = R0; rs <= R9; rs++) { | 
|  | u64 cmp, src; | 
|  |  | 
|  | /* Initialize value in memory */ | 
|  | i += __bpf_ld_imm64(&insn[i], R0, mem); | 
|  | insn[i++] = BPF_STX_MEM(width, R10, R0, -8); | 
|  |  | 
|  | /* Initialize registers in order */ | 
|  | i += __bpf_ld_imm64(&insn[i], R0, ~mem); | 
|  | i += __bpf_ld_imm64(&insn[i], rs, upd); | 
|  | insn[i++] = BPF_MOV64_REG(rd, R10); | 
|  |  | 
|  | /* Perform atomic operation */ | 
|  | insn[i++] = BPF_ATOMIC_OP(width, op, rd, rs, -8); | 
|  | if (op == BPF_CMPXCHG && width == BPF_W) | 
|  | insn[i++] = BPF_ZEXT_REG(R0); | 
|  |  | 
|  | /* Check R0 register value */ | 
|  | if (op == BPF_CMPXCHG) | 
|  | cmp = mem;  /* Expect value from memory */ | 
|  | else if (R0 == rd || R0 == rs) | 
|  | cmp = 0;    /* Aliased, checked below */ | 
|  | else | 
|  | cmp = ~mem; /* Expect value to be preserved */ | 
|  | if (cmp) { | 
|  | insn[i++] = BPF_JMP32_IMM(BPF_JEQ, R0, | 
|  | (u32)cmp, 2); | 
|  | insn[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_RSH, R0, 32); | 
|  | insn[i++] = BPF_JMP32_IMM(BPF_JEQ, R0, | 
|  | cmp >> 32, 2); | 
|  | insn[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | /* Check source register value */ | 
|  | if (rs == R0 && op == BPF_CMPXCHG) | 
|  | src = 0;   /* Aliased with R0, checked above */ | 
|  | else if (rs == rd && (op == BPF_CMPXCHG || | 
|  | !(op & BPF_FETCH))) | 
|  | src = 0;   /* Aliased with rd, checked below */ | 
|  | else if (op == BPF_CMPXCHG) | 
|  | src = upd; /* Expect value to be preserved */ | 
|  | else if (op & BPF_FETCH) | 
|  | src = mem; /* Expect fetched value from mem */ | 
|  | else /* no fetch */ | 
|  | src = upd; /* Expect value to be preserved */ | 
|  | if (src) { | 
|  | insn[i++] = BPF_JMP32_IMM(BPF_JEQ, rs, | 
|  | (u32)src, 2); | 
|  | insn[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_RSH, rs, 32); | 
|  | insn[i++] = BPF_JMP32_IMM(BPF_JEQ, rs, | 
|  | src >> 32, 2); | 
|  | insn[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | /* Check destination register value */ | 
|  | if (!(rd == R0 && op == BPF_CMPXCHG) && | 
|  | !(rd == rs && (op & BPF_FETCH))) { | 
|  | insn[i++] = BPF_JMP_REG(BPF_JEQ, rd, R10, 2); | 
|  | insn[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | /* Check value in memory */ | 
|  | if (rs != rd) {                  /* No aliasing */ | 
|  | i += __bpf_ld_imm64(&insn[i], R1, res); | 
|  | } else if (op == BPF_XCHG) {     /* Aliased, XCHG */ | 
|  | insn[i++] = BPF_MOV64_REG(R1, R10); | 
|  | } else if (op == BPF_CMPXCHG) {  /* Aliased, CMPXCHG */ | 
|  | i += __bpf_ld_imm64(&insn[i], R1, mem); | 
|  | } else {                        /* Aliased, ALU oper */ | 
|  | i += __bpf_ld_imm64(&insn[i], R1, mem); | 
|  | insn[i++] = BPF_ALU64_REG(BPF_OP(op), R1, R10); | 
|  | } | 
|  |  | 
|  | insn[i++] = BPF_LDX_MEM(width, R0, R10, -8); | 
|  | if (width == BPF_DW) | 
|  | insn[i++] = BPF_JMP_REG(BPF_JEQ, R0, R1, 2); | 
|  | else /* width == BPF_W */ | 
|  | insn[i++] = BPF_JMP32_REG(BPF_JEQ, R0, R1, 2); | 
|  | insn[i++] = BPF_MOV32_IMM(R0, __LINE__); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  | } | 
|  |  | 
|  | insn[i++] = BPF_MOV64_IMM(R0, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = i; | 
|  | BUG_ON(i > len); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* 64-bit atomic register tests */ | 
|  | static int bpf_fill_atomic64_add_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_ADD); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_and_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_AND); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_or_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_OR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_xor_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_XOR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_add_fetch_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_ADD | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_and_fetch_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_AND | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_or_fetch_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_OR | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_xor_fetch_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_XOR | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_xchg_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_XCHG); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic64_cmpxchg_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_DW, BPF_CMPXCHG); | 
|  | } | 
|  |  | 
|  | /* 32-bit atomic register tests */ | 
|  | static int bpf_fill_atomic32_add_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_ADD); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_and_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_AND); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_or_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_OR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_xor_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_XOR); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_add_fetch_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_ADD | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_and_fetch_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_AND | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_or_fetch_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_OR | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_xor_fetch_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_XOR | BPF_FETCH); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_xchg_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_XCHG); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_atomic32_cmpxchg_reg_pairs(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_atomic_reg_pairs(self, BPF_W, BPF_CMPXCHG); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Test the two-instruction 64-bit immediate load operation for all | 
|  | * power-of-two magnitudes of the immediate operand. For each MSB, a block | 
|  | * of immediate values centered around the power-of-two MSB are tested, | 
|  | * both for positive and negative values. The test is designed to verify | 
|  | * the operation for JITs that emit different code depending on the magnitude | 
|  | * of the immediate value. This is often the case if the native instruction | 
|  | * immediate field width is narrower than 32 bits. | 
|  | */ | 
|  | static int bpf_fill_ld_imm64_magn(struct bpf_test *self) | 
|  | { | 
|  | int block = 64; /* Increase for more tests per MSB position */ | 
|  | int len = 3 + 8 * 63 * block * 2; | 
|  | struct bpf_insn *insn; | 
|  | int bit, adj, sign; | 
|  | int i = 0; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0); | 
|  |  | 
|  | for (bit = 0; bit <= 62; bit++) { | 
|  | for (adj = -block / 2; adj < block / 2; adj++) { | 
|  | for (sign = -1; sign <= 1; sign += 2) { | 
|  | s64 imm = sign * ((1LL << bit) + adj); | 
|  |  | 
|  | /* Perform operation */ | 
|  | i += __bpf_ld_imm64(&insn[i], R1, imm); | 
|  |  | 
|  | /* Load reference */ | 
|  | insn[i++] = BPF_ALU32_IMM(BPF_MOV, R2, imm); | 
|  | insn[i++] = BPF_ALU32_IMM(BPF_MOV, R3, | 
|  | (u32)(imm >> 32)); | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_LSH, R3, 32); | 
|  | insn[i++] = BPF_ALU64_REG(BPF_OR, R2, R3); | 
|  |  | 
|  | /* Check result */ | 
|  | insn[i++] = BPF_JMP_REG(BPF_JEQ, R1, R2, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  | BUG_ON(i != len); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Test the two-instruction 64-bit immediate load operation for different | 
|  | * combinations of bytes. Each byte in the 64-bit word is constructed as | 
|  | * (base & mask) | (rand() & ~mask), where rand() is a deterministic LCG. | 
|  | * All patterns (base1, mask1) and (base2, mask2) bytes are tested. | 
|  | */ | 
|  | static int __bpf_fill_ld_imm64_bytes(struct bpf_test *self, | 
|  | u8 base1, u8 mask1, | 
|  | u8 base2, u8 mask2) | 
|  | { | 
|  | struct bpf_insn *insn; | 
|  | int len = 3 + 8 * BIT(8); | 
|  | int pattern, index; | 
|  | u32 rand = 1; | 
|  | int i = 0; | 
|  |  | 
|  | insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); | 
|  | if (!insn) | 
|  | return -ENOMEM; | 
|  |  | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 0); | 
|  |  | 
|  | for (pattern = 0; pattern < BIT(8); pattern++) { | 
|  | u64 imm = 0; | 
|  |  | 
|  | for (index = 0; index < 8; index++) { | 
|  | int byte; | 
|  |  | 
|  | if (pattern & BIT(index)) | 
|  | byte = (base1 & mask1) | (rand & ~mask1); | 
|  | else | 
|  | byte = (base2 & mask2) | (rand & ~mask2); | 
|  | imm = (imm << 8) | byte; | 
|  | } | 
|  |  | 
|  | /* Update our LCG */ | 
|  | rand = rand * 1664525 + 1013904223; | 
|  |  | 
|  | /* Perform operation */ | 
|  | i += __bpf_ld_imm64(&insn[i], R1, imm); | 
|  |  | 
|  | /* Load reference */ | 
|  | insn[i++] = BPF_ALU32_IMM(BPF_MOV, R2, imm); | 
|  | insn[i++] = BPF_ALU32_IMM(BPF_MOV, R3, (u32)(imm >> 32)); | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_LSH, R3, 32); | 
|  | insn[i++] = BPF_ALU64_REG(BPF_OR, R2, R3); | 
|  |  | 
|  | /* Check result */ | 
|  | insn[i++] = BPF_JMP_REG(BPF_JEQ, R1, R2, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  | } | 
|  |  | 
|  | insn[i++] = BPF_ALU64_IMM(BPF_MOV, R0, 1); | 
|  | insn[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insn; | 
|  | self->u.ptr.len = len; | 
|  | BUG_ON(i != len); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int bpf_fill_ld_imm64_checker(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_ld_imm64_bytes(self, 0, 0xff, 0xff, 0xff); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_ld_imm64_pos_neg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_ld_imm64_bytes(self, 1, 0x81, 0x80, 0x80); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_ld_imm64_pos_zero(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_ld_imm64_bytes(self, 1, 0x81, 0, 0xff); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_ld_imm64_neg_zero(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_ld_imm64_bytes(self, 0x80, 0x80, 0, 0xff); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Exhaustive tests of JMP operations for all combinations of power-of-two | 
|  | * magnitudes of the operands, both for positive and negative values. The | 
|  | * test is designed to verify e.g. the JMP and JMP32 operations for JITs that | 
|  | * emit different code depending on the magnitude of the immediate value. | 
|  | */ | 
|  |  | 
|  | static bool __bpf_match_jmp_cond(s64 v1, s64 v2, u8 op) | 
|  | { | 
|  | switch (op) { | 
|  | case BPF_JSET: | 
|  | return !!(v1 & v2); | 
|  | case BPF_JEQ: | 
|  | return v1 == v2; | 
|  | case BPF_JNE: | 
|  | return v1 != v2; | 
|  | case BPF_JGT: | 
|  | return (u64)v1 > (u64)v2; | 
|  | case BPF_JGE: | 
|  | return (u64)v1 >= (u64)v2; | 
|  | case BPF_JLT: | 
|  | return (u64)v1 < (u64)v2; | 
|  | case BPF_JLE: | 
|  | return (u64)v1 <= (u64)v2; | 
|  | case BPF_JSGT: | 
|  | return v1 > v2; | 
|  | case BPF_JSGE: | 
|  | return v1 >= v2; | 
|  | case BPF_JSLT: | 
|  | return v1 < v2; | 
|  | case BPF_JSLE: | 
|  | return v1 <= v2; | 
|  | } | 
|  | return false; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_jmp_imm(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 imm) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  |  | 
|  | if (insns) { | 
|  | bool match = __bpf_match_jmp_cond(dst, (s32)imm, op); | 
|  | int i = 0; | 
|  |  | 
|  | insns[i++] = BPF_ALU32_IMM(BPF_MOV, R0, match); | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | insns[i++] = BPF_JMP_IMM(op, R1, imm, 1); | 
|  | if (!match) | 
|  | insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | return 5 + 1; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_jmp32_imm(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 imm) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  |  | 
|  | if (insns) { | 
|  | bool match = __bpf_match_jmp_cond((s32)dst, (s32)imm, op); | 
|  | int i = 0; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | insns[i++] = BPF_JMP32_IMM(op, R1, imm, 1); | 
|  | if (!match) | 
|  | insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | return 5; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_jmp_reg(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 src) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  |  | 
|  | if (insns) { | 
|  | bool match = __bpf_match_jmp_cond(dst, src, op); | 
|  | int i = 0; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R2, src); | 
|  | insns[i++] = BPF_JMP_REG(op, R1, R2, 1); | 
|  | if (!match) | 
|  | insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | return 7; | 
|  | } | 
|  |  | 
|  | static int __bpf_emit_jmp32_reg(struct bpf_test *self, void *arg, | 
|  | struct bpf_insn *insns, s64 dst, s64 src) | 
|  | { | 
|  | int op = *(int *)arg; | 
|  |  | 
|  | if (insns) { | 
|  | bool match = __bpf_match_jmp_cond((s32)dst, (s32)src, op); | 
|  | int i = 0; | 
|  |  | 
|  | i += __bpf_ld_imm64(&insns[i], R1, dst); | 
|  | i += __bpf_ld_imm64(&insns[i], R2, src); | 
|  | insns[i++] = BPF_JMP32_REG(op, R1, R2, 1); | 
|  | if (!match) | 
|  | insns[i++] = BPF_JMP_IMM(BPF_JA, 0, 0, 1); | 
|  | insns[i++] = BPF_EXIT_INSN(); | 
|  |  | 
|  | return i; | 
|  | } | 
|  |  | 
|  | return 7; | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_jmp_imm(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 32, | 
|  | PATTERN_BLOCK1, PATTERN_BLOCK2, | 
|  | &__bpf_emit_jmp_imm); | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_jmp32_imm(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 32, | 
|  | PATTERN_BLOCK1, PATTERN_BLOCK2, | 
|  | &__bpf_emit_jmp32_imm); | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_jmp_reg(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 64, | 
|  | PATTERN_BLOCK1, PATTERN_BLOCK2, | 
|  | &__bpf_emit_jmp_reg); | 
|  | } | 
|  |  | 
|  | static int __bpf_fill_jmp32_reg(struct bpf_test *self, int op) | 
|  | { | 
|  | return __bpf_fill_pattern(self, &op, 64, 64, | 
|  | PATTERN_BLOCK1, PATTERN_BLOCK2, | 
|  | &__bpf_emit_jmp32_reg); | 
|  | } | 
|  |  | 
|  | /* JMP immediate tests */ | 
|  | static int bpf_fill_jmp_jset_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JSET); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jeq_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JEQ); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jne_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JNE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jgt_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JGT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jge_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JGE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jlt_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JLT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jle_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JLE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jsgt_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JSGT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jsge_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JSGE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jslt_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JSLT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jsle_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_imm(self, BPF_JSLE); | 
|  | } | 
|  |  | 
|  | /* JMP32 immediate tests */ | 
|  | static int bpf_fill_jmp32_jset_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JSET); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jeq_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JEQ); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jne_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JNE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jgt_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JGT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jge_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JGE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jlt_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JLT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jle_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JLE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jsgt_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JSGT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jsge_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JSGE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jslt_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JSLT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jsle_imm(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_imm(self, BPF_JSLE); | 
|  | } | 
|  |  | 
|  | /* JMP register tests */ | 
|  | static int bpf_fill_jmp_jset_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JSET); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jeq_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JEQ); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jne_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JNE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jgt_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JGT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jge_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JGE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jlt_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JLT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jle_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JLE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jsgt_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JSGT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jsge_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JSGE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jslt_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JSLT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp_jsle_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp_reg(self, BPF_JSLE); | 
|  | } | 
|  |  | 
|  | /* JMP32 register tests */ | 
|  | static int bpf_fill_jmp32_jset_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JSET); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jeq_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JEQ); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jne_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JNE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jgt_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JGT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jge_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JGE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jlt_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JLT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jle_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JLE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jsgt_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JSGT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jsge_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JSGE); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jslt_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JSLT); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_jmp32_jsle_reg(struct bpf_test *self) | 
|  | { | 
|  | return __bpf_fill_jmp32_reg(self, BPF_JSLE); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Set up a sequence of staggered jumps, forwards and backwards with | 
|  | * increasing offset. This tests the conversion of relative jumps to | 
|  | * JITed native jumps. On some architectures, for example MIPS, a large | 
|  | * PC-relative jump offset may overflow the immediate field of the native | 
|  | * conditional branch instruction, triggering a conversion to use an | 
|  | * absolute jump instead. Since this changes the jump offsets, another | 
|  | * offset computation pass is necessary, and that may in turn trigger | 
|  | * another branch conversion. This jump sequence is particularly nasty | 
|  | * in that regard. | 
|  | * | 
|  | * The sequence generation is parameterized by size and jump type. | 
|  | * The size must be even, and the expected result is always size + 1. | 
|  | * Below is an example with size=8 and result=9. | 
|  | * | 
|  | *                     ________________________Start | 
|  | *                     R0 = 0 | 
|  | *                     R1 = r1 | 
|  | *                     R2 = r2 | 
|  | *            ,------- JMP +4 * 3______________Preamble: 4 insns | 
|  | * ,----------|-ind 0- if R0 != 7 JMP 8 * 3 + 1 <--------------------. | 
|  | * |          |        R0 = 8                                        | | 
|  | * |          |        JMP +7 * 3               ------------------------. | 
|  | * | ,--------|-----1- if R0 != 5 JMP 7 * 3 + 1 <--------------.     |  | | 
|  | * | |        |        R0 = 6                                  |     |  | | 
|  | * | |        |        JMP +5 * 3               ------------------.  |  | | 
|  | * | | ,------|-----2- if R0 != 3 JMP 6 * 3 + 1 <--------.     |  |  |  | | 
|  | * | | |      |        R0 = 4                            |     |  |  |  | | 
|  | * | | |      |        JMP +3 * 3               ------------.  |  |  |  | | 
|  | * | | | ,----|-----3- if R0 != 1 JMP 5 * 3 + 1 <--.     |  |  |  |  |  | | 
|  | * | | | |    |        R0 = 2                      |     |  |  |  |  |  | | 
|  | * | | | |    |        JMP +1 * 3               ------.  |  |  |  |  |  | | 
|  | * | | | | ,--t=====4> if R0 != 0 JMP 4 * 3 + 1    1  2  3  4  5  6  7  8 loc | 
|  | * | | | | |           R0 = 1                     -1 +2 -3 +4 -5 +6 -7 +8 off | 
|  | * | | | | |           JMP -2 * 3               ---'  |  |  |  |  |  |  | | 
|  | * | | | | | ,------5- if R0 != 2 JMP 3 * 3 + 1 <-----'  |  |  |  |  |  | | 
|  | * | | | | | |         R0 = 3                            |  |  |  |  |  | | 
|  | * | | | | | |         JMP -4 * 3               ---------'  |  |  |  |  | | 
|  | * | | | | | | ,----6- if R0 != 4 JMP 2 * 3 + 1 <-----------'  |  |  |  | | 
|  | * | | | | | | |       R0 = 5                                  |  |  |  | | 
|  | * | | | | | | |       JMP -6 * 3               ---------------'  |  |  | | 
|  | * | | | | | | | ,--7- if R0 != 6 JMP 1 * 3 + 1 <-----------------'  |  | | 
|  | * | | | | | | | |     R0 = 7                                        |  | | 
|  | * | | Error | | |     JMP -8 * 3               ---------------------'  | | 
|  | * | | paths | | | ,8- if R0 != 8 JMP 0 * 3 + 1 <-----------------------' | 
|  | * | | | | | | | | |   R0 = 9__________________Sequence: 3 * size - 1 insns | 
|  | * `-+-+-+-+-+-+-+-+-> EXIT____________________Return: 1 insn | 
|  | * | 
|  | */ | 
|  |  | 
|  | /* The maximum size parameter */ | 
|  | #define MAX_STAGGERED_JMP_SIZE ((0x7fff / 3) & ~1) | 
|  |  | 
|  | /* We use a reduced number of iterations to get a reasonable execution time */ | 
|  | #define NR_STAGGERED_JMP_RUNS 10 | 
|  |  | 
|  | static int __bpf_fill_staggered_jumps(struct bpf_test *self, | 
|  | const struct bpf_insn *jmp, | 
|  | u64 r1, u64 r2) | 
|  | { | 
|  | int size = self->test[0].result - 1; | 
|  | int len = 4 + 3 * (size + 1); | 
|  | struct bpf_insn *insns; | 
|  | int off, ind; | 
|  |  | 
|  | insns = kmalloc_array(len, sizeof(*insns), GFP_KERNEL); | 
|  | if (!insns) | 
|  | return -ENOMEM; | 
|  |  | 
|  | /* Preamble */ | 
|  | insns[0] = BPF_ALU64_IMM(BPF_MOV, R0, 0); | 
|  | insns[1] = BPF_ALU64_IMM(BPF_MOV, R1, r1); | 
|  | insns[2] = BPF_ALU64_IMM(BPF_MOV, R2, r2); | 
|  | insns[3] = BPF_JMP_IMM(BPF_JA, 0, 0, 3 * size / 2); | 
|  |  | 
|  | /* Sequence */ | 
|  | for (ind = 0, off = size; ind <= size; ind++, off -= 2) { | 
|  | struct bpf_insn *ins = &insns[4 + 3 * ind]; | 
|  | int loc; | 
|  |  | 
|  | if (off == 0) | 
|  | off--; | 
|  |  | 
|  | loc = abs(off); | 
|  | ins[0] = BPF_JMP_IMM(BPF_JNE, R0, loc - 1, | 
|  | 3 * (size - ind) + 1); | 
|  | ins[1] = BPF_ALU64_IMM(BPF_MOV, R0, loc); | 
|  | ins[2] = *jmp; | 
|  | ins[2].off = 3 * (off - 1); | 
|  | } | 
|  |  | 
|  | /* Return */ | 
|  | insns[len - 1] = BPF_EXIT_INSN(); | 
|  |  | 
|  | self->u.ptr.insns = insns; | 
|  | self->u.ptr.len = len; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* 64-bit unconditional jump */ | 
|  | static int bpf_fill_staggered_ja(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JA, 0, 0, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0, 0); | 
|  | } | 
|  |  | 
|  | /* 64-bit immediate jumps */ | 
|  | static int bpf_fill_staggered_jeq_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JEQ, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jne_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JNE, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 4321, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jset_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSET, R1, 0x82, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0x86, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jgt_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JGT, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0x80000000, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jge_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JGE, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jlt_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JLT, R1, 0x80000000, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jle_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JLE, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsgt_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSGT, R1, -2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -1, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsge_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSGE, R1, -2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -2, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jslt_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSLT, R1, -1, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -2, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsle_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_IMM(BPF_JSLE, R1, -1, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -1, 0); | 
|  | } | 
|  |  | 
|  | /* 64-bit register jumps */ | 
|  | static int bpf_fill_staggered_jeq_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JEQ, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jne_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JNE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 4321, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jset_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JSET, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0x86, 0x82); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jgt_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JGT, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0x80000000, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jge_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JGE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jlt_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JLT, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0x80000000); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jle_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JLE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsgt_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JSGT, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -1, -2); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsge_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JSGE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -2, -2); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jslt_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JSLT, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -2, -1); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsle_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP_REG(BPF_JSLE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -1, -1); | 
|  | } | 
|  |  | 
|  | /* 32-bit immediate jumps */ | 
|  | static int bpf_fill_staggered_jeq32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JEQ, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jne32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JNE, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 4321, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jset32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSET, R1, 0x82, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0x86, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jgt32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JGT, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0x80000000, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jge32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JGE, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jlt32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JLT, R1, 0x80000000, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jle32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JLE, R1, 1234, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsgt32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSGT, R1, -2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -1, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsge32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSGE, R1, -2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -2, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jslt32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSLT, R1, -1, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -2, 0); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsle32_imm(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JSLE, R1, -1, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -1, 0); | 
|  | } | 
|  |  | 
|  | /* 32-bit register jumps */ | 
|  | static int bpf_fill_staggered_jeq32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JEQ, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jne32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JNE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 4321, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jset32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSET, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0x86, 0x82); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jgt32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JGT, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 0x80000000, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jge32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JGE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jlt32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JLT, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 0x80000000); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jle32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JLE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, 1234, 1234); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsgt32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSGT, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -1, -2); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsge32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSGE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -2, -2); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jslt32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSLT, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -2, -1); | 
|  | } | 
|  |  | 
|  | static int bpf_fill_staggered_jsle32_reg(struct bpf_test *self) | 
|  | { | 
|  | struct bpf_insn jmp = BPF_JMP32_REG(BPF_JSLE, R1, R2, 0); | 
|  |  | 
|  | return __bpf_fill_staggered_jumps(self, &jmp, -1, -1); | 
|  | } | 
|  |  | 
|  |  | 
|  | static struct bpf_test tests[] = { | 
|  | { | 
|  | "TAX", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_NEG, 0), /* A == -3 */ | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), /* X == len - 3 */ | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_IND, 1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 10, 20, 30, 40, 50 }, | 
|  | { { 2, 10 }, { 3, 20 }, { 4, 30 } }, | 
|  | }, | 
|  | { | 
|  | "TXA", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) /* A == len * 2 */ | 
|  | }, | 
|  | CLASSIC, | 
|  | { 10, 20, 30, 40, 50 }, | 
|  | { { 1, 2 }, { 3, 6 }, { 4, 8 } }, | 
|  | }, | 
|  | { | 
|  | "ADD_SUB_MUL_K", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 1), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 2), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 3), | 
|  | BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 0xffffffff), | 
|  | BPF_STMT(BPF_ALU | BPF_MUL | BPF_K, 3), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xfffffffd } } | 
|  | }, | 
|  | { | 
|  | "DIV_MOD_KX", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 8), | 
|  | BPF_STMT(BPF_ALU | BPF_DIV | BPF_K, 2), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff), | 
|  | BPF_STMT(BPF_ALU | BPF_DIV | BPF_X, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff), | 
|  | BPF_STMT(BPF_ALU | BPF_DIV | BPF_K, 0x70000000), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff), | 
|  | BPF_STMT(BPF_ALU | BPF_MOD | BPF_X, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff), | 
|  | BPF_STMT(BPF_ALU | BPF_MOD | BPF_K, 0x70000000), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0x20000000 } } | 
|  | }, | 
|  | { | 
|  | "AND_OR_LSH_K", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xff), | 
|  | BPF_STMT(BPF_ALU | BPF_AND | BPF_K, 0xf0), | 
|  | BPF_STMT(BPF_ALU | BPF_LSH | BPF_K, 27), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xf), | 
|  | BPF_STMT(BPF_ALU | BPF_OR | BPF_K, 0xf0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0x800000ff }, { 1, 0x800000ff } }, | 
|  | }, | 
|  | { | 
|  | "LD_IMM_0", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0), /* ld #0 */ | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, 1 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_IND, MAX_K), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, 0 }, { 10, 0 }, { 60, 0 } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, 1000), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, 0 }, { 10, 0 }, { 60, 0 } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS_LL", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, SKF_LL_OFF), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, SKF_LL_OFF + 1), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 1, 2, 3 }, | 
|  | { { 1, 0 }, { 2, 3 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND_LL", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, SKF_LL_OFF - 1), | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_IND, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 1, 2, 3, 0xff }, | 
|  | { { 1, 1 }, { 3, 3 }, { 4, 0xff } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS_NET", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, SKF_NET_OFF), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, SKF_NET_OFF + 1), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 3 }, | 
|  | { { 15, 0 }, { 16, 3 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND_NET", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, SKF_NET_OFF - 15), | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_IND, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 3 }, | 
|  | { { 14, 0 }, { 15, 1 }, { 17, 3 } }, | 
|  | }, | 
|  | { | 
|  | "LD_PKTTYPE", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PKTTYPE), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, SKB_TYPE, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PKTTYPE), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, SKB_TYPE, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PKTTYPE), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, SKB_TYPE, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, 3 }, { 10, 3 } }, | 
|  | }, | 
|  | { | 
|  | "LD_MARK", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_MARK), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, SKB_MARK}, { 10, SKB_MARK} }, | 
|  | }, | 
|  | { | 
|  | "LD_RXHASH", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_RXHASH), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, SKB_HASH}, { 10, SKB_HASH} }, | 
|  | }, | 
|  | { | 
|  | "LD_QUEUE", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_QUEUE), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, SKB_QUEUE_MAP }, { 10, SKB_QUEUE_MAP } }, | 
|  | }, | 
|  | { | 
|  | "LD_PROTOCOL", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 1), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 20, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PROTOCOL), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 30, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 10, 20, 30 }, | 
|  | { { 10, ETH_P_IP }, { 100, ETH_P_IP } }, | 
|  | }, | 
|  | { | 
|  | "LD_VLAN_TAG", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_VLAN_TAG), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { | 
|  | { 1, SKB_VLAN_TCI }, | 
|  | { 10, SKB_VLAN_TCI } | 
|  | }, | 
|  | }, | 
|  | { | 
|  | "LD_VLAN_TAG_PRESENT", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_VLAN_TAG_PRESENT), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { | 
|  | { 1, SKB_VLAN_PRESENT }, | 
|  | { 10, SKB_VLAN_PRESENT } | 
|  | }, | 
|  | }, | 
|  | { | 
|  | "LD_IFINDEX", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_IFINDEX), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, SKB_DEV_IFINDEX }, { 10, SKB_DEV_IFINDEX } }, | 
|  | }, | 
|  | { | 
|  | "LD_HATYPE", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_HATYPE), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, SKB_DEV_TYPE }, { 10, SKB_DEV_TYPE } }, | 
|  | }, | 
|  | { | 
|  | "LD_CPU", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_CPU), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_CPU), | 
|  | BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, 0 }, { 10, 0 } }, | 
|  | }, | 
|  | { | 
|  | "LD_NLATTR", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 3), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | #ifdef __BIG_ENDIAN | 
|  | { 0xff, 0xff, 0, 4, 0, 2, 0, 4, 0, 3 }, | 
|  | #else | 
|  | { 0xff, 0xff, 4, 0, 2, 0, 4, 0, 3, 0 }, | 
|  | #endif | 
|  | { { 4, 0 }, { 20, 6 } }, | 
|  | }, | 
|  | { | 
|  | "LD_NLATTR_NEST", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 3), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR_NEST), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR_NEST), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR_NEST), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR_NEST), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR_NEST), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR_NEST), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR_NEST), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_NLATTR_NEST), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | #ifdef __BIG_ENDIAN | 
|  | { 0xff, 0xff, 0, 12, 0, 1, 0, 4, 0, 2, 0, 4, 0, 3 }, | 
|  | #else | 
|  | { 0xff, 0xff, 12, 0, 1, 0, 4, 0, 2, 0, 4, 0, 3, 0 }, | 
|  | #endif | 
|  | { { 4, 0 }, { 20, 10 } }, | 
|  | }, | 
|  | { | 
|  | "LD_PAYLOAD_OFF", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PAY_OFFSET), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PAY_OFFSET), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PAY_OFFSET), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PAY_OFFSET), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_PAY_OFFSET), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | /* 00:00:00:00:00:00 > 00:00:00:00:00:00, ethtype IPv4 (0x0800), | 
|  | * length 98: 127.0.0.1 > 127.0.0.1: ICMP echo request, | 
|  | * id 9737, seq 1, length 64 | 
|  | */ | 
|  | { 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, | 
|  | 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, | 
|  | 0x08, 0x00, | 
|  | 0x45, 0x00, 0x00, 0x54, 0xac, 0x8b, 0x40, 0x00, 0x40, | 
|  | 0x01, 0x90, 0x1b, 0x7f, 0x00, 0x00, 0x01 }, | 
|  | { { 30, 0 }, { 100, 42 } }, | 
|  | }, | 
|  | { | 
|  | "LD_ANC_XOR", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 10), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 300), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_ALU_XOR_X), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 4, 0xA ^ 300 }, { 20, 0xA ^ 300 } }, | 
|  | }, | 
|  | { | 
|  | "SPILL_FILL", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 2), | 
|  | BPF_STMT(BPF_ALU | BPF_RSH, 1), | 
|  | BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0), | 
|  | BPF_STMT(BPF_ST, 1), /* M1 = 1 ^ len */ | 
|  | BPF_STMT(BPF_ALU | BPF_XOR | BPF_K, 0x80000000), | 
|  | BPF_STMT(BPF_ST, 2), /* M2 = 1 ^ len ^ 0x80000000 */ | 
|  | BPF_STMT(BPF_STX, 15), /* M3 = len */ | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 1), | 
|  | BPF_STMT(BPF_LD | BPF_MEM, 2), | 
|  | BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 15), | 
|  | BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, 0x80000001 }, { 2, 0x80000002 }, { 60, 0x80000000 ^ 60 } } | 
|  | }, | 
|  | { | 
|  | "JEQ", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_X, 0, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, MAX_K) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 3, 3, 3, 3, 3 }, | 
|  | { { 1, 0 }, { 3, 1 }, { 4, MAX_K } }, | 
|  | }, | 
|  | { | 
|  | "JGT", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGT | BPF_X, 0, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, MAX_K) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 4, 4, 4, 3, 3 }, | 
|  | { { 2, 0 }, { 3, 1 }, { 4, MAX_K } }, | 
|  | }, | 
|  | { | 
|  | "JGE (jt 0), test 1", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE | BPF_X, 0, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, MAX_K) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 4, 4, 4, 3, 3 }, | 
|  | { { 2, 0 }, { 3, 1 }, { 4, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JGE (jt 0), test 2", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 2), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE | BPF_X, 0, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, MAX_K) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 4, 4, 5, 3, 3 }, | 
|  | { { 4, 1 }, { 5, 1 }, { 6, MAX_K } }, | 
|  | }, | 
|  | { | 
|  | "JGE", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_IND, MAX_K), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 1, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 10), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 2, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 20), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 3, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 30), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 4, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 40), | 
|  | BPF_STMT(BPF_RET | BPF_K, MAX_K) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 1, 2, 3, 4, 5 }, | 
|  | { { 1, 20 }, { 3, 40 }, { 5, MAX_K } }, | 
|  | }, | 
|  | { | 
|  | "JSET", | 
|  | .u.insns = { | 
|  | BPF_JUMP(BPF_JMP | BPF_JA, 0, 0, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JA, 1, 1, 1), | 
|  | BPF_JUMP(BPF_JMP | BPF_JA, 0, 0, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JA, 0, 0, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_LEN, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_SUB | BPF_K, 4), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_IND, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 1, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 10), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0x80000000, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 20), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 30), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 30), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 30), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 30), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0xffffff, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 30), | 
|  | BPF_STMT(BPF_RET | BPF_K, MAX_K) | 
|  | }, | 
|  | CLASSIC, | 
|  | { 0, 0xAA, 0x55, 1 }, | 
|  | { { 4, 10 }, { 5, 20 }, { 6, MAX_K } }, | 
|  | }, | 
|  | { | 
|  | "tcpdump port 22", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 12), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x86dd, 0, 8), /* IPv6 */ | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 20), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x84, 2, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x6, 1, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x11, 0, 17), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 54), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 14, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 56), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 12, 13), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x0800, 0, 12), /* IPv4 */ | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 23), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x84, 2, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x6, 1, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x11, 0, 8), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 20), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0x1fff, 6, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 14), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_IND, 14), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 2, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_IND, 16), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0xffff), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0), | 
|  | }, | 
|  | CLASSIC, | 
|  | /* 3c:07:54:43:e5:76 > 10:bf:48:d6:43:d6, ethertype IPv4(0x0800) | 
|  | * length 114: 10.1.1.149.49700 > 10.1.2.10.22: Flags [P.], | 
|  | * seq 1305692979:1305693027, ack 3650467037, win 65535, | 
|  | * options [nop,nop,TS val 2502645400 ecr 3971138], length 48 | 
|  | */ | 
|  | { 0x10, 0xbf, 0x48, 0xd6, 0x43, 0xd6, | 
|  | 0x3c, 0x07, 0x54, 0x43, 0xe5, 0x76, | 
|  | 0x08, 0x00, | 
|  | 0x45, 0x10, 0x00, 0x64, 0x75, 0xb5, | 
|  | 0x40, 0x00, 0x40, 0x06, 0xad, 0x2e, /* IP header */ | 
|  | 0x0a, 0x01, 0x01, 0x95, /* ip src */ | 
|  | 0x0a, 0x01, 0x02, 0x0a, /* ip dst */ | 
|  | 0xc2, 0x24, | 
|  | 0x00, 0x16 /* dst port */ }, | 
|  | { { 10, 0 }, { 30, 0 }, { 100, 65535 } }, | 
|  | }, | 
|  | { | 
|  | "tcpdump complex", | 
|  | .u.insns = { | 
|  | /* tcpdump -nei eth0 'tcp port 22 and (((ip[2:2] - | 
|  | * ((ip[0]&0xf)<<2)) - ((tcp[12]&0xf0)>>2)) != 0) and | 
|  | * (len > 115 or len < 30000000000)' -d | 
|  | */ | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 12), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x86dd, 30, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x800, 0, 29), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 23), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x6, 0, 27), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 20), | 
|  | BPF_JUMP(BPF_JMP | BPF_JSET | BPF_K, 0x1fff, 25, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 14), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_IND, 14), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 2, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_IND, 16), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 22, 0, 20), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 16), | 
|  | BPF_STMT(BPF_ST, 1), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 14), | 
|  | BPF_STMT(BPF_ALU | BPF_AND | BPF_K, 0xf), | 
|  | BPF_STMT(BPF_ALU | BPF_LSH | BPF_K, 2), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0x5), /* libpcap emits K on TAX */ | 
|  | BPF_STMT(BPF_LD | BPF_MEM, 1), | 
|  | BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0), | 
|  | BPF_STMT(BPF_ST, 5), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 14), | 
|  | BPF_STMT(BPF_LD | BPF_B | BPF_IND, 26), | 
|  | BPF_STMT(BPF_ALU | BPF_AND | BPF_K, 0xf0), | 
|  | BPF_STMT(BPF_ALU | BPF_RSH | BPF_K, 2), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0x9), /* libpcap emits K on TAX */ | 
|  | BPF_STMT(BPF_LD | BPF_MEM, 5), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_X, 0, 4, 0), | 
|  | BPF_STMT(BPF_LD | BPF_LEN, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGT | BPF_K, 0x73, 1, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE | BPF_K, 0xfc23ac00, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0xffff), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { 0x10, 0xbf, 0x48, 0xd6, 0x43, 0xd6, | 
|  | 0x3c, 0x07, 0x54, 0x43, 0xe5, 0x76, | 
|  | 0x08, 0x00, | 
|  | 0x45, 0x10, 0x00, 0x64, 0x75, 0xb5, | 
|  | 0x40, 0x00, 0x40, 0x06, 0xad, 0x2e, /* IP header */ | 
|  | 0x0a, 0x01, 0x01, 0x95, /* ip src */ | 
|  | 0x0a, 0x01, 0x02, 0x0a, /* ip dst */ | 
|  | 0xc2, 0x24, | 
|  | 0x00, 0x16 /* dst port */ }, | 
|  | { { 10, 0 }, { 30, 0 }, { 100, 65535 } }, | 
|  | }, | 
|  | { | 
|  | "RET_A", | 
|  | .u.insns = { | 
|  | /* check that uninitialized X and A contain zeros */ | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0) | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { {1, 0}, {2, 0} }, | 
|  | }, | 
|  | { | 
|  | "INT: ADD trivial", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 2), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 3), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R2), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, -1), | 
|  | BPF_ALU64_IMM(BPF_MUL, R1, 3), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffd } } | 
|  | }, | 
|  | { | 
|  | "INT: MUL_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, -1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 3), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R2), | 
|  | BPF_JMP_IMM(BPF_JEQ, R1, 0xfffffffd, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "INT: MUL_X2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R2, 3), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R2), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 8), | 
|  | BPF_JMP_IMM(BPF_JEQ, R1, 0x2ffffff, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "INT: MUL32_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, -1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R2, 3), | 
|  | BPF_ALU32_REG(BPF_MUL, R1, R2), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 8), | 
|  | BPF_JMP_IMM(BPF_JEQ, R1, 0xffffff, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | /* Have to test all register combinations, since | 
|  | * JITing of different registers will produce | 
|  | * different asm code. | 
|  | */ | 
|  | "INT: ADD 64-bit", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 2), | 
|  | BPF_ALU64_IMM(BPF_MOV, R3, 3), | 
|  | BPF_ALU64_IMM(BPF_MOV, R4, 4), | 
|  | BPF_ALU64_IMM(BPF_MOV, R5, 5), | 
|  | BPF_ALU64_IMM(BPF_MOV, R6, 6), | 
|  | BPF_ALU64_IMM(BPF_MOV, R7, 7), | 
|  | BPF_ALU64_IMM(BPF_MOV, R8, 8), | 
|  | BPF_ALU64_IMM(BPF_MOV, R9, 9), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R3, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R4, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R5, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R6, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R7, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R8, 20), | 
|  | BPF_ALU64_IMM(BPF_ADD, R9, 20), | 
|  | BPF_ALU64_IMM(BPF_SUB, R0, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R1, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R2, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R3, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R4, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R5, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R6, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R7, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R8, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R9, 10), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R9), /* R0 == 155 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 155, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R1, R9), /* R1 == 456 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R1, 456, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R2, R9), /* R2 == 1358 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R2, 1358, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R3, R9), /* R3 == 4063 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R3, 4063, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R4, R9), /* R4 == 12177 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R4, 12177, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R5, R9), /* R5 == 36518 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R5, 36518, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R6, R9), /* R6 == 109540 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R6, 109540, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R7, R9), /* R7 == 328605 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R7, 328605, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R8, R9), /* R8 == 985799 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R8, 985799, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R9, R9), /* R9 == 2957380 */ | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R9), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2957380 } } | 
|  | }, | 
|  | { | 
|  | "INT: ADD 32-bit", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 20), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R2, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R4, 4), | 
|  | BPF_ALU32_IMM(BPF_MOV, R5, 5), | 
|  | BPF_ALU32_IMM(BPF_MOV, R6, 6), | 
|  | BPF_ALU32_IMM(BPF_MOV, R7, 7), | 
|  | BPF_ALU32_IMM(BPF_MOV, R8, 8), | 
|  | BPF_ALU32_IMM(BPF_MOV, R9, 9), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 10), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 10), | 
|  | BPF_ALU64_IMM(BPF_ADD, R3, 10), | 
|  | BPF_ALU64_IMM(BPF_ADD, R4, 10), | 
|  | BPF_ALU64_IMM(BPF_ADD, R5, 10), | 
|  | BPF_ALU64_IMM(BPF_ADD, R6, 10), | 
|  | BPF_ALU64_IMM(BPF_ADD, R7, 10), | 
|  | BPF_ALU64_IMM(BPF_ADD, R8, 10), | 
|  | BPF_ALU64_IMM(BPF_ADD, R9, 10), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R9), /* R0 == 155 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 155, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R1, R9), /* R1 == 456 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R1, 456, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R2, R9), /* R2 == 1358 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R2, 1358, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R3, R9), /* R3 == 4063 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R3, 4063, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R4, R9), /* R4 == 12177 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R4, 12177, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R5, R9), /* R5 == 36518 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R5, 36518, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R6, R9), /* R6 == 109540 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R6, 109540, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R7, R9), /* R7 == 328605 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R7, 328605, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R8, R9), /* R8 == 985799 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R8, 985799, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R0), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R1), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R2), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R3), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R4), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R5), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R6), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R7), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R8), | 
|  | BPF_ALU32_REG(BPF_ADD, R9, R9), /* R9 == 2957380 */ | 
|  | BPF_ALU32_REG(BPF_MOV, R0, R9), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2957380 } } | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "INT: SUB", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 2), | 
|  | BPF_ALU64_IMM(BPF_MOV, R3, 3), | 
|  | BPF_ALU64_IMM(BPF_MOV, R4, 4), | 
|  | BPF_ALU64_IMM(BPF_MOV, R5, 5), | 
|  | BPF_ALU64_IMM(BPF_MOV, R6, 6), | 
|  | BPF_ALU64_IMM(BPF_MOV, R7, 7), | 
|  | BPF_ALU64_IMM(BPF_MOV, R8, 8), | 
|  | BPF_ALU64_IMM(BPF_MOV, R9, 9), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R0, 10), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, -55, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R1, 10), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R2, 10), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R3, 10), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R4, 10), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R5, 10), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R6, 10), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R7, 10), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R9), | 
|  | BPF_ALU64_IMM(BPF_SUB, R8, 10), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R0), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R8), | 
|  | BPF_ALU64_IMM(BPF_SUB, R9, 10), | 
|  | BPF_ALU64_IMM(BPF_SUB, R0, 10), | 
|  | BPF_ALU64_IMM(BPF_NEG, R0, 0), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R2), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R3), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R4), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R5), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R6), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R7), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R8), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R9), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 11 } } | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "INT: XOR", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R0), | 
|  | BPF_ALU64_REG(BPF_XOR, R1, R1), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 10), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, -1), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R1), | 
|  | BPF_ALU64_REG(BPF_XOR, R2, R2), | 
|  | BPF_JMP_REG(BPF_JEQ, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R2, R2), | 
|  | BPF_ALU64_REG(BPF_XOR, R3, R3), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 10), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, -1), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R3, R3), | 
|  | BPF_ALU64_REG(BPF_XOR, R4, R4), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R5, -1), | 
|  | BPF_JMP_REG(BPF_JEQ, R3, R4, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R4, R4), | 
|  | BPF_ALU64_REG(BPF_XOR, R5, R5), | 
|  | BPF_ALU64_IMM(BPF_MOV, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R7, -1), | 
|  | BPF_JMP_REG(BPF_JEQ, R5, R4, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_IMM(BPF_MOV, R5, 1), | 
|  | BPF_ALU64_REG(BPF_SUB, R5, R5), | 
|  | BPF_ALU64_REG(BPF_XOR, R6, R6), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R8, -1), | 
|  | BPF_JMP_REG(BPF_JEQ, R5, R6, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R6, R6), | 
|  | BPF_ALU64_REG(BPF_XOR, R7, R7), | 
|  | BPF_JMP_REG(BPF_JEQ, R7, R6, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R7, R7), | 
|  | BPF_ALU64_REG(BPF_XOR, R8, R8), | 
|  | BPF_JMP_REG(BPF_JEQ, R7, R8, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R8, R8), | 
|  | BPF_ALU64_REG(BPF_XOR, R9, R9), | 
|  | BPF_JMP_REG(BPF_JEQ, R9, R8, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R9, R9), | 
|  | BPF_ALU64_REG(BPF_XOR, R0, R0), | 
|  | BPF_JMP_REG(BPF_JEQ, R9, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R1), | 
|  | BPF_ALU64_REG(BPF_XOR, R0, R0), | 
|  | BPF_JMP_REG(BPF_JEQ, R9, R0, 2), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "INT: MUL", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 11), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 2), | 
|  | BPF_ALU64_IMM(BPF_MOV, R3, 3), | 
|  | BPF_ALU64_IMM(BPF_MOV, R4, 4), | 
|  | BPF_ALU64_IMM(BPF_MOV, R5, 5), | 
|  | BPF_ALU64_IMM(BPF_MOV, R6, 6), | 
|  | BPF_ALU64_IMM(BPF_MOV, R7, 7), | 
|  | BPF_ALU64_IMM(BPF_MOV, R8, 8), | 
|  | BPF_ALU64_IMM(BPF_MOV, R9, 9), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R0), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R1), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R2), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R3), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R4), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R5), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R6), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R7), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R8), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R9), | 
|  | BPF_ALU64_IMM(BPF_MUL, R0, 10), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 439084800, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R0), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R2), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R3), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R4), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R5), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R6), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R7), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R8), | 
|  | BPF_ALU64_REG(BPF_MUL, R1, R9), | 
|  | BPF_ALU64_IMM(BPF_MUL, R1, 10), | 
|  | BPF_ALU64_REG(BPF_MOV, R2, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R2, 32), | 
|  | BPF_JMP_IMM(BPF_JEQ, R2, 0x5a924, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_IMM(BPF_LSH, R1, 32), | 
|  | BPF_ALU64_IMM(BPF_ARSH, R1, 32), | 
|  | BPF_JMP_IMM(BPF_JEQ, R1, 0xebb90000, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R0), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R1), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R3), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R4), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R5), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R6), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R7), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R8), | 
|  | BPF_ALU64_REG(BPF_MUL, R2, R9), | 
|  | BPF_ALU64_IMM(BPF_MUL, R2, 10), | 
|  | BPF_ALU64_IMM(BPF_RSH, R2, 32), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x35d97ef2 } } | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "MOV REG64", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffLL), | 
|  | BPF_MOV64_REG(R1, R0), | 
|  | BPF_MOV64_REG(R2, R1), | 
|  | BPF_MOV64_REG(R3, R2), | 
|  | BPF_MOV64_REG(R4, R3), | 
|  | BPF_MOV64_REG(R5, R4), | 
|  | BPF_MOV64_REG(R6, R5), | 
|  | BPF_MOV64_REG(R7, R6), | 
|  | BPF_MOV64_REG(R8, R7), | 
|  | BPF_MOV64_REG(R9, R8), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R3, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R4, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R5, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R6, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R7, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R8, 0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R9, 0), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R9), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 0xfefe), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfefe } } | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "MOV REG32", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffLL), | 
|  | BPF_MOV64_REG(R1, R0), | 
|  | BPF_MOV64_REG(R2, R1), | 
|  | BPF_MOV64_REG(R3, R2), | 
|  | BPF_MOV64_REG(R4, R3), | 
|  | BPF_MOV64_REG(R5, R4), | 
|  | BPF_MOV64_REG(R6, R5), | 
|  | BPF_MOV64_REG(R7, R6), | 
|  | BPF_MOV64_REG(R8, R7), | 
|  | BPF_MOV64_REG(R9, R8), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R2, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R4, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R5, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R6, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R7, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R8, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R9, 0), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R9), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 0xfefe), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfefe } } | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "LD IMM64", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffLL), | 
|  | BPF_MOV64_REG(R1, R0), | 
|  | BPF_MOV64_REG(R2, R1), | 
|  | BPF_MOV64_REG(R3, R2), | 
|  | BPF_MOV64_REG(R4, R3), | 
|  | BPF_MOV64_REG(R5, R4), | 
|  | BPF_MOV64_REG(R6, R5), | 
|  | BPF_MOV64_REG(R7, R6), | 
|  | BPF_MOV64_REG(R8, R7), | 
|  | BPF_MOV64_REG(R9, R8), | 
|  | BPF_LD_IMM64(R0, 0x0LL), | 
|  | BPF_LD_IMM64(R1, 0x0LL), | 
|  | BPF_LD_IMM64(R2, 0x0LL), | 
|  | BPF_LD_IMM64(R3, 0x0LL), | 
|  | BPF_LD_IMM64(R4, 0x0LL), | 
|  | BPF_LD_IMM64(R5, 0x0LL), | 
|  | BPF_LD_IMM64(R6, 0x0LL), | 
|  | BPF_LD_IMM64(R7, 0x0LL), | 
|  | BPF_LD_IMM64(R8, 0x0LL), | 
|  | BPF_LD_IMM64(R9, 0x0LL), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R0), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R1), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R2), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R3), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R4), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R5), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R6), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R7), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R8), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R9), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 0xfefe), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfefe } } | 
|  | }, | 
|  | { | 
|  | "INT: ALU MIX", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 11), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, -1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 2), | 
|  | BPF_ALU64_IMM(BPF_XOR, R2, 3), | 
|  | BPF_ALU64_REG(BPF_DIV, R0, R2), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 10, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_IMM(BPF_MOD, R0, 3), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } } | 
|  | }, | 
|  | { | 
|  | "INT: shifts by register", | 
|  | .u.insns_int = { | 
|  | BPF_MOV64_IMM(R0, -1234), | 
|  | BPF_MOV64_IMM(R1, 1), | 
|  | BPF_ALU32_REG(BPF_RSH, R0, R1), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 0x7ffffd97, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV64_IMM(R2, 1), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R2), | 
|  | BPF_MOV32_IMM(R4, -1234), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R4, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU64_IMM(BPF_AND, R4, 63), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R4), /* R0 <= 46 */ | 
|  | BPF_MOV64_IMM(R3, 47), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R3), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, -617, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV64_IMM(R2, 1), | 
|  | BPF_ALU64_REG(BPF_LSH, R4, R2), /* R4 = 46 << 1 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R4, 92, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV64_IMM(R4, 4), | 
|  | BPF_ALU64_REG(BPF_LSH, R4, R4), /* R4 = 4 << 4 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R4, 64, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV64_IMM(R4, 5), | 
|  | BPF_ALU32_REG(BPF_LSH, R4, R4), /* R4 = 5 << 5 */ | 
|  | BPF_JMP_IMM(BPF_JEQ, R4, 160, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV64_IMM(R0, -1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } } | 
|  | }, | 
|  | #ifdef CONFIG_32BIT | 
|  | { | 
|  | "INT: 32-bit context pointer word order and zero-extension", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP32_IMM(BPF_JEQ, R1, 0, 3), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32), | 
|  | BPF_JMP32_IMM(BPF_JNE, R1, 0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | #endif | 
|  | { | 
|  | "check: missing ret", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 1), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL, | 
|  | { }, | 
|  | { }, | 
|  | .fill_helper = NULL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "check: div_k_0", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_ALU | BPF_DIV | BPF_K, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0) | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL, | 
|  | { }, | 
|  | { }, | 
|  | .fill_helper = NULL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "check: unknown insn", | 
|  | .u.insns = { | 
|  | /* seccomp insn, rejected in socket filter */ | 
|  | BPF_STMT(BPF_LDX | BPF_W | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0) | 
|  | }, | 
|  | CLASSIC | FLAG_EXPECTED_FAIL, | 
|  | { }, | 
|  | { }, | 
|  | .fill_helper = NULL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "check: out of range spill/fill", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_STX, 16), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0) | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL, | 
|  | { }, | 
|  | { }, | 
|  | .fill_helper = NULL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "JUMPS + HOLES", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE, 0, 13, 15), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ, 0x90c2894d, 3, 4), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ, 0x90c2894d, 1, 2), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE, 0, 14, 15), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE, 0, 13, 14), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ, 0x2ac28349, 2, 3), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ, 0x2ac28349, 1, 2), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE, 0, 14, 15), | 
|  | BPF_JUMP(BPF_JMP | BPF_JGE, 0, 13, 14), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ, 0x90d2ff41, 2, 3), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ, 0x90d2ff41, 1, 2), | 
|  | BPF_STMT(BPF_LD | BPF_H | BPF_ABS, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { 0x00, 0x1b, 0x21, 0x3c, 0x9d, 0xf8, | 
|  | 0x90, 0xe2, 0xba, 0x0a, 0x56, 0xb4, | 
|  | 0x08, 0x00, | 
|  | 0x45, 0x00, 0x00, 0x28, 0x00, 0x00, | 
|  | 0x20, 0x00, 0x40, 0x11, 0x00, 0x00, /* IP header */ | 
|  | 0xc0, 0xa8, 0x33, 0x01, | 
|  | 0xc0, 0xa8, 0x33, 0x02, | 
|  | 0xbb, 0xb6, | 
|  | 0xa9, 0xfa, | 
|  | 0x00, 0x14, 0x00, 0x00, | 
|  | 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, | 
|  | 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, | 
|  | 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, | 
|  | 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, | 
|  | 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, | 
|  | 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, | 
|  | 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, | 
|  | 0xcc, 0xcc, 0xcc, 0xcc }, | 
|  | { { 88, 0x001b } } | 
|  | }, | 
|  | { | 
|  | "check: RET X", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_RET | BPF_X, 0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL, | 
|  | { }, | 
|  | { }, | 
|  | .fill_helper = NULL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "check: LDX + RET X", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 42), | 
|  | BPF_STMT(BPF_RET | BPF_X, 0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL, | 
|  | { }, | 
|  | { }, | 
|  | .fill_helper = NULL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "M[]: alt STX + LDX", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 100), | 
|  | BPF_STMT(BPF_STX, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 1), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 2), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 2), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 3), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 3), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 4), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 4), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 5), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 5), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 6), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 6), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 7), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 7), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 8), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 8), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 9), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 9), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 10), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 10), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 11), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 11), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 12), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 12), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 13), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 13), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 14), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 14), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_STX, 15), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 15), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 1), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 116 } }, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "M[]: full STX + full LDX", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xbadfeedb), | 
|  | BPF_STMT(BPF_STX, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xecabedae), | 
|  | BPF_STMT(BPF_STX, 1), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xafccfeaf), | 
|  | BPF_STMT(BPF_STX, 2), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xbffdcedc), | 
|  | BPF_STMT(BPF_STX, 3), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xfbbbdccb), | 
|  | BPF_STMT(BPF_STX, 4), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xfbabcbda), | 
|  | BPF_STMT(BPF_STX, 5), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xaedecbdb), | 
|  | BPF_STMT(BPF_STX, 6), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xadebbade), | 
|  | BPF_STMT(BPF_STX, 7), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xfcfcfaec), | 
|  | BPF_STMT(BPF_STX, 8), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xbcdddbdc), | 
|  | BPF_STMT(BPF_STX, 9), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xfeefdfac), | 
|  | BPF_STMT(BPF_STX, 10), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xcddcdeea), | 
|  | BPF_STMT(BPF_STX, 11), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xaccfaebb), | 
|  | BPF_STMT(BPF_STX, 12), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xbdcccdcf), | 
|  | BPF_STMT(BPF_STX, 13), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xaaedecde), | 
|  | BPF_STMT(BPF_STX, 14), | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0xfaeacdad), | 
|  | BPF_STMT(BPF_STX, 15), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 0), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 1), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 2), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 3), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 4), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 5), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 6), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 7), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 8), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 9), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 10), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 11), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 12), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 13), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 14), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_LDX | BPF_MEM, 15), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0x2a5a5e5 } }, | 
|  | }, | 
|  | { | 
|  | "check: SKF_AD_MAX", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF + SKF_AD_MAX), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL, | 
|  | { }, | 
|  | { }, | 
|  | .fill_helper = NULL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | {	/* Passes checker but fails during runtime. */ | 
|  | "LD [SKF_AD_OFF-1]", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_W | BPF_ABS, | 
|  | SKF_AD_OFF - 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, 0 } }, | 
|  | }, | 
|  | { | 
|  | "load 64-bit immediate", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x567800001234LL), | 
|  | BPF_MOV64_REG(R2, R1), | 
|  | BPF_MOV64_REG(R3, R2), | 
|  | BPF_ALU64_IMM(BPF_RSH, R2, 32), | 
|  | BPF_ALU64_IMM(BPF_LSH, R3, 32), | 
|  | BPF_ALU64_IMM(BPF_RSH, R3, 32), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP_IMM(BPF_JEQ, R2, 0x5678, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_JMP_IMM(BPF_JEQ, R3, 0x1234, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_LD_IMM64(R0, 0x1ffffffffLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), /* R0 = 1 */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_MOV | BPF_X */ | 
|  | { | 
|  | "ALU_MOV_X: dst = 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU32_REG(BPF_MOV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MOV_X: dst = 4294967295", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4294967295U), | 
|  | BPF_ALU32_REG(BPF_MOV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4294967295U } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_X: dst = 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_X: dst = 4294967295", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4294967295U), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4294967295U } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_MOV | BPF_K */ | 
|  | { | 
|  | "ALU_MOV_K: dst = 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MOV_K: dst = 4294967295", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 4294967295U), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4294967295U } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MOV_K: 0x0000ffffffff0000 = 0x00000000ffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0x00000000ffffffffLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MOV_K: small negative", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | { | 
|  | "ALU_MOV_K: small negative zero extension", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU_MOV_K: large negative", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123456789), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123456789 } } | 
|  | }, | 
|  | { | 
|  | "ALU_MOV_K: large negative zero extension", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123456789), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_K: dst = 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_K: dst = 2147483647", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 2147483647), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2147483647 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: dst = 0x0", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0x0), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 0x0), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_K: dst = -1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0xffffffffffffffffLL), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_K: small negative", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -123), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_K: small negative sign extension", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -123), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } } | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_K: large negative", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -123456789), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123456789 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_MOV_K: large negative sign extension", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -123456789), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } } | 
|  | }, | 
|  | /* MOVSX32 */ | 
|  | { | 
|  | "ALU_MOVSX | BPF_B", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x00000000ffffffefLL), | 
|  | BPF_LD_IMM64(R3, 0xdeadbeefdeadbeefLL), | 
|  | BPF_MOVSX32_REG(R1, R3, 8), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MOVSX | BPF_H", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x00000000ffffbeefLL), | 
|  | BPF_LD_IMM64(R3, 0xdeadbeefdeadbeefLL), | 
|  | BPF_MOVSX32_REG(R1, R3, 16), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | /* MOVSX64 REG */ | 
|  | { | 
|  | "ALU64_MOVSX | BPF_B", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0xffffffffffffffefLL), | 
|  | BPF_LD_IMM64(R3, 0xdeadbeefdeadbeefLL), | 
|  | BPF_MOVSX64_REG(R1, R3, 8), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOVSX | BPF_H", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0xffffffffffffbeefLL), | 
|  | BPF_LD_IMM64(R3, 0xdeadbeefdeadbeefLL), | 
|  | BPF_MOVSX64_REG(R1, R3, 16), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOVSX | BPF_W", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0xffffffffdeadbeefLL), | 
|  | BPF_LD_IMM64(R3, 0xdeadbeefdeadbeefLL), | 
|  | BPF_MOVSX64_REG(R1, R3, 32), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_ADD | BPF_X */ | 
|  | { | 
|  | "ALU_ADD_X: 1 + 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_X: 1 + 4294967294 = 4294967295", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4294967294U), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4294967295U } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_X: 2 + 4294967294 = 0", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_LD_IMM64(R1, 4294967294U), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 0, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_X: 1 + 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_X: 1 + 4294967294 = 4294967295", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4294967294U), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4294967295U } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_X: 2 + 4294967294 = 4294967296", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_LD_IMM64(R1, 4294967294U), | 
|  | BPF_LD_IMM64(R2, 4294967296ULL), | 
|  | BPF_ALU64_REG(BPF_ADD, R0, R1), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R2, 2), | 
|  | BPF_MOV32_IMM(R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_ADD | BPF_K */ | 
|  | { | 
|  | "ALU_ADD_K: 1 + 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 3 + 0 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 1 + 4294967294 = 4294967295", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 4294967294U), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4294967295U } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 4294967294 + 2 = 0", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967294U), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 2), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 0, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 0 + (-1) = 0x00000000ffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0x00000000ffffffff), | 
|  | BPF_ALU32_IMM(BPF_ADD, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 0 + 0xffff = 0xffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0xffff), | 
|  | BPF_ALU32_IMM(BPF_ADD, R2, 0xffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 0 + 0x7fffffff = 0x7fffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0x7fffffff), | 
|  | BPF_ALU32_IMM(BPF_ADD, R2, 0x7fffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 0 + 0x80000000 = 0x80000000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0x80000000), | 
|  | BPF_ALU32_IMM(BPF_ADD, R2, 0x80000000), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 0 + 0x80008000 = 0x80008000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0x80008000), | 
|  | BPF_ALU32_IMM(BPF_ADD, R2, 0x80008000), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 1 + 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 3 + 0 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 1 + 2147483646 = 2147483647", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 2147483646), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2147483647 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 4294967294 + 2 = 4294967296", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967294U), | 
|  | BPF_LD_IMM64(R1, 4294967296ULL), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 2), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 2147483646 + -2147483647 = -1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2147483646), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, -2147483647), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 1 + 0 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x1), | 
|  | BPF_LD_IMM64(R3, 0x1), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 0x0), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 0 + (-1) = 0xffffffffffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0xffffffffffffffffLL), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 0 + 0xffff = 0xffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0xffff), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 0xffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 0 + 0x7fffffff = 0x7fffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0x7fffffff), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 0x7fffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 0 + 0x80000000 = 0xffffffff80000000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0xffffffff80000000LL), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 0x80000000), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_ADD_K: 0 + 0x80008000 = 0xffffffff80008000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0), | 
|  | BPF_LD_IMM64(R3, 0xffffffff80008000LL), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 0x80008000), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_SUB | BPF_X */ | 
|  | { | 
|  | "ALU_SUB_X: 3 - 1 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU32_REG(BPF_SUB, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_SUB_X: 4294967295 - 4294967294 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967295U), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4294967294U), | 
|  | BPF_ALU32_REG(BPF_SUB, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_X: 3 - 1 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_X: 4294967295 - 4294967294 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967295U), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4294967294U), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_SUB | BPF_K */ | 
|  | { | 
|  | "ALU_SUB_K: 3 - 1 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_SUB, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_SUB_K: 3 - 0 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_SUB, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_SUB_K: 4294967295 - 4294967294 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967295U), | 
|  | BPF_ALU32_IMM(BPF_SUB, R0, 4294967294U), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_K: 3 - 1 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_SUB, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_K: 3 - 0 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_SUB, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_K: 4294967294 - 4294967295 = -1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967294U), | 
|  | BPF_ALU64_IMM(BPF_SUB, R0, 4294967295U), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: 2147483646 - 2147483647 = -1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2147483646), | 
|  | BPF_ALU64_IMM(BPF_SUB, R0, 2147483647), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_MUL | BPF_X */ | 
|  | { | 
|  | "ALU_MUL_X: 2 * 3 = 6", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 3), | 
|  | BPF_ALU32_REG(BPF_MUL, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 6 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MUL_X: 2 * 0x7FFFFFF8 = 0xFFFFFFF0", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0x7FFFFFF8), | 
|  | BPF_ALU32_REG(BPF_MUL, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xFFFFFFF0 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MUL_X: -1 * -1 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -1), | 
|  | BPF_ALU32_REG(BPF_MUL, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_X: 2 * 3 = 6", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 3), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 6 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_X: 1 * 2147483647 = 2147483647", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2147483647), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2147483647 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_X: 64x64 multiply, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0fedcba987654321LL), | 
|  | BPF_LD_IMM64(R1, 0x123456789abcdef0LL), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xe5618cf0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_X: 64x64 multiply, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0fedcba987654321LL), | 
|  | BPF_LD_IMM64(R1, 0x123456789abcdef0LL), | 
|  | BPF_ALU64_REG(BPF_MUL, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x2236d88f } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_MUL | BPF_K */ | 
|  | { | 
|  | "ALU_MUL_K: 2 * 3 = 6", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU32_IMM(BPF_MUL, R0, 3), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 6 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MUL_K: 3 * 1 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MUL, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MUL_K: 2 * 0x7FFFFFF8 = 0xFFFFFFF0", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU32_IMM(BPF_MUL, R0, 0x7FFFFFF8), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xFFFFFFF0 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MUL_K: 1 * (-1) = 0x00000000ffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x1), | 
|  | BPF_LD_IMM64(R3, 0x00000000ffffffff), | 
|  | BPF_ALU32_IMM(BPF_MUL, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: 2 * 3 = 6", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU64_IMM(BPF_MUL, R0, 3), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 6 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: 3 * 1 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_MUL, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: 1 * 2147483647 = 2147483647", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU64_IMM(BPF_MUL, R0, 2147483647), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2147483647 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: 1 * -2147483647 = -2147483647", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU64_IMM(BPF_MUL, R0, -2147483647), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -2147483647 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: 1 * (-1) = 0xffffffffffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x1), | 
|  | BPF_LD_IMM64(R3, 0xffffffffffffffffLL), | 
|  | BPF_ALU64_IMM(BPF_MUL, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: 64x32 multiply, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_MUL, R0, 0x12345678), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xe242d208 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: 64x32 multiply, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_MUL, R0, 0x12345678), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xc28f5c28 } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_DIV | BPF_X */ | 
|  | { | 
|  | "ALU_DIV_X: 6 / 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 6), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU32_REG(BPF_DIV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_DIV_X: 4294967295 / 4294967295 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967295U), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4294967295U), | 
|  | BPF_ALU32_REG(BPF_DIV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_X: 6 / 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 6), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU64_REG(BPF_DIV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_X: 2147483647 / 2147483647 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2147483647), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2147483647), | 
|  | BPF_ALU64_REG(BPF_DIV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_X: 0xffffffffffffffff / (-1) = 0x0000000000000001", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0xffffffffffffffffLL), | 
|  | BPF_LD_IMM64(R4, 0xffffffffffffffffLL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000000001LL), | 
|  | BPF_ALU64_REG(BPF_DIV, R2, R4), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_DIV | BPF_K */ | 
|  | { | 
|  | "ALU_DIV_K: 6 / 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 6), | 
|  | BPF_ALU32_IMM(BPF_DIV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_DIV_K: 3 / 1 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_DIV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_DIV_K: 4294967295 / 4294967295 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967295U), | 
|  | BPF_ALU32_IMM(BPF_DIV, R0, 4294967295U), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_DIV_K: 0xffffffffffffffff / (-1) = 0x1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0xffffffffffffffffLL), | 
|  | BPF_LD_IMM64(R3, 0x1UL), | 
|  | BPF_ALU32_IMM(BPF_DIV, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_K: 6 / 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 6), | 
|  | BPF_ALU64_IMM(BPF_DIV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_K: 3 / 1 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_DIV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_K: 2147483647 / 2147483647 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2147483647), | 
|  | BPF_ALU64_IMM(BPF_DIV, R0, 2147483647), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_K: 0xffffffffffffffff / (-1) = 0x0000000000000001", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0xffffffffffffffffLL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000000001LL), | 
|  | BPF_ALU64_IMM(BPF_DIV, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_MOD | BPF_X */ | 
|  | { | 
|  | "ALU_MOD_X: 3 % 2 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU32_REG(BPF_MOD, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MOD_X: 4294967295 % 4294967293 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967295U), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4294967293U), | 
|  | BPF_ALU32_REG(BPF_MOD, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_X: 3 % 2 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU64_REG(BPF_MOD, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_X: 2147483647 % 2147483645 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2147483647), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2147483645), | 
|  | BPF_ALU64_REG(BPF_MOD, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_MOD | BPF_K */ | 
|  | { | 
|  | "ALU_MOD_K: 3 % 2 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MOD, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MOD_K: 3 % 1 = 0", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MOD, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_MOD_K: 4294967295 % 4294967293 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 4294967295U), | 
|  | BPF_ALU32_IMM(BPF_MOD, R0, 4294967293U), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_K: 3 % 2 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_MOD, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_K: 3 % 1 = 0", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_MOD, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_K: 2147483647 % 2147483645 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2147483647), | 
|  | BPF_ALU64_IMM(BPF_MOD, R0, 2147483645), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_DIV | BPF_X off=1 (SDIV) */ | 
|  | { | 
|  | "ALU_SDIV_X: -6 / 2 = -3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -6), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU32_REG_OFF(BPF_DIV, R0, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -3 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_DIV | BPF_K off=1 (SDIV) */ | 
|  | { | 
|  | "ALU_SDIV_K: -6 / 2 = -3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -6), | 
|  | BPF_ALU32_IMM_OFF(BPF_DIV, R0, 2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -3 } }, | 
|  | }, | 
|  | /* BPF_ALU64 | BPF_DIV | BPF_X off=1 (SDIV64) */ | 
|  | { | 
|  | "ALU64_SDIV_X: -6 / 2 = -3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -6), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU64_REG_OFF(BPF_DIV, R0, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -3 } }, | 
|  | }, | 
|  | /* BPF_ALU64 | BPF_DIV | BPF_K off=1 (SDIV64) */ | 
|  | { | 
|  | "ALU64_SDIV_K: -6 / 2 = -3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -6), | 
|  | BPF_ALU64_IMM_OFF(BPF_DIV, R0, 2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -3 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_MOD | BPF_X off=1 (SMOD) */ | 
|  | { | 
|  | "ALU_SMOD_X: -7 % 2 = -1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -7), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU32_REG_OFF(BPF_MOD, R0, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_MOD | BPF_K off=1 (SMOD) */ | 
|  | { | 
|  | "ALU_SMOD_K: -7 % 2 = -1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -7), | 
|  | BPF_ALU32_IMM_OFF(BPF_MOD, R0, 2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } }, | 
|  | }, | 
|  | /* BPF_ALU64 | BPF_MOD | BPF_X off=1 (SMOD64) */ | 
|  | { | 
|  | "ALU64_SMOD_X: -7 % 2 = -1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -7), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU64_REG_OFF(BPF_MOD, R0, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } }, | 
|  | }, | 
|  | /* BPF_ALU64 | BPF_MOD | BPF_K off=1 (SMOD64) */ | 
|  | { | 
|  | "ALU64_SMOD_K: -7 % 2 = -1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -7), | 
|  | BPF_ALU64_IMM_OFF(BPF_MOD, R0, 2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_AND | BPF_X */ | 
|  | { | 
|  | "ALU_AND_X: 3 & 2 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU32_REG(BPF_AND, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_AND_X: 0xffffffff & 0xffffffff = 0xffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xffffffff), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_ALU32_REG(BPF_AND, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_X: 3 & 2 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU64_REG(BPF_AND, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_X: 0xffffffff & 0xffffffff = 0xffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xffffffff), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_ALU64_REG(BPF_AND, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_AND | BPF_K */ | 
|  | { | 
|  | "ALU_AND_K: 3 & 2 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU32_IMM(BPF_AND, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_AND_K: 0xffffffff & 0xffffffff = 0xffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xffffffff), | 
|  | BPF_ALU32_IMM(BPF_AND, R0, 0xffffffff), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | }, | 
|  | { | 
|  | "ALU_AND_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x01020304), | 
|  | BPF_ALU32_IMM(BPF_AND, R0, 15), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4 } } | 
|  | }, | 
|  | { | 
|  | "ALU_AND_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xf1f2f3f4), | 
|  | BPF_ALU32_IMM(BPF_AND, R0, 0xafbfcfdf), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xa1b2c3d4 } } | 
|  | }, | 
|  | { | 
|  | "ALU_AND_K: Zero extension", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0x0000000080a0c0e0LL), | 
|  | BPF_ALU32_IMM(BPF_AND, R0, 0xf0f0f0f0), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: 3 & 2 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_AND, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: 0xffffffff & 0xffffffff = 0xffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xffffffff), | 
|  | BPF_ALU64_IMM(BPF_AND, R0, 0xffffffff), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: 0x0000ffffffff0000 & 0x0 = 0x0000000000000000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000000000LL), | 
|  | BPF_ALU64_IMM(BPF_AND, R2, 0x0), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: 0x0000ffffffff0000 & -1 = 0x0000ffffffff0000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0x0000ffffffff0000LL), | 
|  | BPF_ALU64_IMM(BPF_AND, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: 0xffffffffffffffff & -1 = 0xffffffffffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0xffffffffffffffffLL), | 
|  | BPF_LD_IMM64(R3, 0xffffffffffffffffLL), | 
|  | BPF_ALU64_IMM(BPF_AND, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: Sign extension 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0x00000000090b0d0fLL), | 
|  | BPF_ALU64_IMM(BPF_AND, R0, 0x0f0f0f0f), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: Sign extension 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0x0123456780a0c0e0LL), | 
|  | BPF_ALU64_IMM(BPF_AND, R0, 0xf0f0f0f0), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_OR | BPF_X */ | 
|  | { | 
|  | "ALU_OR_X: 1 | 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU32_REG(BPF_OR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_OR_X: 0x0 | 0xffffffff = 0xffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_ALU32_REG(BPF_OR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_X: 1 | 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 2), | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_X: 0 | 0xffffffff = 0xffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_OR | BPF_K */ | 
|  | { | 
|  | "ALU_OR_K: 1 | 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_OR, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_OR_K: 0 & 0xffffffff = 0xffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_ALU32_IMM(BPF_OR, R0, 0xffffffff), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | }, | 
|  | { | 
|  | "ALU_OR_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x01020304), | 
|  | BPF_ALU32_IMM(BPF_OR, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x01020305 } } | 
|  | }, | 
|  | { | 
|  | "ALU_OR_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x01020304), | 
|  | BPF_ALU32_IMM(BPF_OR, R0, 0xa0b0c0d0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xa1b2c3d4 } } | 
|  | }, | 
|  | { | 
|  | "ALU_OR_K: Zero extension", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0x00000000f9fbfdffLL), | 
|  | BPF_ALU32_IMM(BPF_OR, R0, 0xf0f0f0f0), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: 1 | 2 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU64_IMM(BPF_OR, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: 0 & 0xffffffff = 0xffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_ALU64_IMM(BPF_OR, R0, 0xffffffff), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: 0x0000ffffffff0000 | 0x0 = 0x0000ffffffff0000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0x0000ffffffff0000LL), | 
|  | BPF_ALU64_IMM(BPF_OR, R2, 0x0), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: 0x0000ffffffff0000 | -1 = 0xffffffffffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0xffffffffffffffffLL), | 
|  | BPF_ALU64_IMM(BPF_OR, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: 0x000000000000000 | -1 = 0xffffffffffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000000000000000LL), | 
|  | BPF_LD_IMM64(R3, 0xffffffffffffffffLL), | 
|  | BPF_ALU64_IMM(BPF_OR, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: Sign extension 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0x012345678fafcfefLL), | 
|  | BPF_ALU64_IMM(BPF_OR, R0, 0x0f0f0f0f), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: Sign extension 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0xfffffffff9fbfdffLL), | 
|  | BPF_ALU64_IMM(BPF_OR, R0, 0xf0f0f0f0), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_XOR | BPF_X */ | 
|  | { | 
|  | "ALU_XOR_X: 5 ^ 6 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 5), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 6), | 
|  | BPF_ALU32_REG(BPF_XOR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_XOR_X: 0x1 ^ 0xffffffff = 0xfffffffe", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_ALU32_REG(BPF_XOR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_X: 5 ^ 6 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 5), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 6), | 
|  | BPF_ALU64_REG(BPF_XOR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_X: 1 ^ 0xffffffff = 0xfffffffe", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_ALU64_REG(BPF_XOR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_XOR | BPF_K */ | 
|  | { | 
|  | "ALU_XOR_K: 5 ^ 6 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 5), | 
|  | BPF_ALU32_IMM(BPF_XOR, R0, 6), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_XOR_K: 1 ^ 0xffffffff = 0xfffffffe", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_XOR, R0, 0xffffffff), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } }, | 
|  | }, | 
|  | { | 
|  | "ALU_XOR_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x01020304), | 
|  | BPF_ALU32_IMM(BPF_XOR, R0, 15), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x0102030b } } | 
|  | }, | 
|  | { | 
|  | "ALU_XOR_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xf1f2f3f4), | 
|  | BPF_ALU32_IMM(BPF_XOR, R0, 0xafbfcfdf), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x5e4d3c2b } } | 
|  | }, | 
|  | { | 
|  | "ALU_XOR_K: Zero extension", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0x00000000795b3d1fLL), | 
|  | BPF_ALU32_IMM(BPF_XOR, R0, 0xf0f0f0f0), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: 5 ^ 6 = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 5), | 
|  | BPF_ALU64_IMM(BPF_XOR, R0, 6), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: 1 ^ 0xffffffff = 0xfffffffe", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU64_IMM(BPF_XOR, R0, 0xffffffff), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: 0x0000ffffffff0000 ^ 0x0 = 0x0000ffffffff0000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0x0000ffffffff0000LL), | 
|  | BPF_ALU64_IMM(BPF_XOR, R2, 0x0), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: 0x0000ffffffff0000 ^ -1 = 0xffff00000000ffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000ffffffff0000LL), | 
|  | BPF_LD_IMM64(R3, 0xffff00000000ffffLL), | 
|  | BPF_ALU64_IMM(BPF_XOR, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: 0x000000000000000 ^ -1 = 0xffffffffffffffff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x0000000000000000LL), | 
|  | BPF_LD_IMM64(R3, 0xffffffffffffffffLL), | 
|  | BPF_ALU64_IMM(BPF_XOR, R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: Sign extension 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0x0123456786a4c2e0LL), | 
|  | BPF_ALU64_IMM(BPF_XOR, R0, 0x0f0f0f0f), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: Sign extension 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_LD_IMM64(R1, 0xfedcba98795b3d1fLL), | 
|  | BPF_ALU64_IMM(BPF_XOR, R0, 0xf0f0f0f0), | 
|  | BPF_JMP_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_LSH | BPF_X */ | 
|  | { | 
|  | "ALU_LSH_X: 1 << 1 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU32_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_LSH_X: 1 << 31 = 0x80000000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 31), | 
|  | BPF_ALU32_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x80000000 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_LSH_X: 0x12345678 << 12 = 0x45678000", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 12), | 
|  | BPF_ALU32_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x45678000 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: 1 << 1 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: 1 << 31 = 0x80000000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 31), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x80000000 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: Shift < 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 12), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xbcdef000 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: Shift < 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 12), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x3456789a } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: Shift > 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 36), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: Shift > 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 36), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x9abcdef0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: Shift == 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 32), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: Shift == 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 32), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: Zero shift, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: Zero shift, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0), | 
|  | BPF_ALU64_REG(BPF_LSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x01234567 } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_LSH | BPF_K */ | 
|  | { | 
|  | "ALU_LSH_K: 1 << 1 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_LSH, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_LSH_K: 1 << 31 = 0x80000000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU32_IMM(BPF_LSH, R0, 31), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x80000000 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_LSH_K: 0x12345678 << 12 = 0x45678000", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678), | 
|  | BPF_ALU32_IMM(BPF_LSH, R0, 12), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x45678000 } } | 
|  | }, | 
|  | { | 
|  | "ALU_LSH_K: 0x12345678 << 0 = 0x12345678", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678), | 
|  | BPF_ALU32_IMM(BPF_LSH, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x12345678 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: 1 << 1 = 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: 1 << 31 = 0x80000000", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 1), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 31), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x80000000 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: Shift < 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 12), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xbcdef000 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: Shift < 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 12), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x3456789a } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: Shift > 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 36), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: Shift > 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 36), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x9abcdef0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: Shift == 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: Shift == 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 32), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } } | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: Zero shift", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_LSH, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_RSH | BPF_X */ | 
|  | { | 
|  | "ALU_RSH_X: 2 >> 1 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU32_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_RSH_X: 0x80000000 >> 31 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x80000000), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 31), | 
|  | BPF_ALU32_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_RSH_X: 0x12345678 >> 20 = 0x123", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 20), | 
|  | BPF_ALU32_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x123 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: 2 >> 1 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: 0x80000000 >> 31 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x80000000), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 31), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: Shift < 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 12), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x56789abc } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: Shift < 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 12), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x00081234 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: Shift > 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 36), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x08123456 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: Shift > 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 36), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: Shift == 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 32), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x81234567 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: Shift == 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 32), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: Zero shift, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: Zero shift, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0), | 
|  | BPF_ALU64_REG(BPF_RSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x81234567 } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_RSH | BPF_K */ | 
|  | { | 
|  | "ALU_RSH_K: 2 >> 1 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU32_IMM(BPF_RSH, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_RSH_K: 0x80000000 >> 31 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x80000000), | 
|  | BPF_ALU32_IMM(BPF_RSH, R0, 31), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_RSH_K: 0x12345678 >> 20 = 0x123", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678), | 
|  | BPF_ALU32_IMM(BPF_RSH, R0, 20), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x123 } } | 
|  | }, | 
|  | { | 
|  | "ALU_RSH_K: 0x12345678 >> 0 = 0x12345678", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x12345678), | 
|  | BPF_ALU32_IMM(BPF_RSH, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x12345678 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: 2 >> 1 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 2), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: 0x80000000 >> 31 = 1", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x80000000), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 31), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: Shift < 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 12), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x56789abc } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: Shift < 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 12), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x00081234 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: Shift > 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 36), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x08123456 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: Shift > 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 36), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: Shift == 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x81234567 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: Shift == 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: Zero shift", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_ARSH | BPF_X */ | 
|  | { | 
|  | "ALU32_ARSH_X: -1234 >> 7 = -10", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1234), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 7), | 
|  | BPF_ALU32_REG(BPF_ARSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -10 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: 0xff00ff0000000000 >> 40 = 0xffffffffffff00ff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xff00ff0000000000LL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 40), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffff00ff } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: Shift < 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 12), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x56789abc } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: Shift < 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 12), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfff81234 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: Shift > 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 36), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xf8123456 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: Shift > 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 36), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: Shift == 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 32), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x81234567 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: Shift == 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 32), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: Zero shift, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: Zero shift, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0), | 
|  | BPF_ALU64_REG(BPF_ARSH, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x81234567 } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_ARSH | BPF_K */ | 
|  | { | 
|  | "ALU32_ARSH_K: -1234 >> 7 = -10", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1234), | 
|  | BPF_ALU32_IMM(BPF_ARSH, R0, 7), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -10 } } | 
|  | }, | 
|  | { | 
|  | "ALU32_ARSH_K: -1234 >> 0 = -1234", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1234), | 
|  | BPF_ALU32_IMM(BPF_ARSH, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1234 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: 0xff00ff0000000000 >> 40 = 0xffffffffffff00ff", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xff00ff0000000000LL), | 
|  | BPF_ALU64_IMM(BPF_ARSH, R0, 40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffff00ff } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: Shift < 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 12), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x56789abc } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: Shift < 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_ARSH, R0, 12), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfff81234 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: Shift > 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_ARSH, R0, 36), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xf8123456 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: Shift > 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xf123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_ARSH, R0, 36), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: Shift == 32, low word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_ARSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x81234567 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: Shift == 32, high word", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_ARSH, R0, 32), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -1 } } | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: Zero shift", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x8123456789abcdefLL), | 
|  | BPF_ALU64_IMM(BPF_ARSH, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } } | 
|  | }, | 
|  | /* BPF_ALU | BPF_NEG */ | 
|  | { | 
|  | "ALU_NEG: -(3) = -3", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 3), | 
|  | BPF_ALU32_IMM(BPF_NEG, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU_NEG: -(-3) = 3", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -3), | 
|  | BPF_ALU32_IMM(BPF_NEG, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_NEG: -(3) = -3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 3), | 
|  | BPF_ALU64_IMM(BPF_NEG, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -3 } }, | 
|  | }, | 
|  | { | 
|  | "ALU64_NEG: -(-3) = 3", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, -3), | 
|  | BPF_ALU64_IMM(BPF_NEG, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 3 } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_END | BPF_FROM_BE */ | 
|  | { | 
|  | "ALU_END_FROM_BE 16: 0x0123456789abcdef -> 0xcdef", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ENDIAN(BPF_FROM_BE, R0, 16), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0,  cpu_to_be16(0xcdef) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_BE 32: 0x0123456789abcdef -> 0x89abcdef", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ENDIAN(BPF_FROM_BE, R0, 32), | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, cpu_to_be32(0x89abcdef) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_BE 64: 0x0123456789abcdef -> 0x89abcdef", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ENDIAN(BPF_FROM_BE, R0, 64), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, (u32) cpu_to_be64(0x0123456789abcdefLL) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_BE 64: 0x0123456789abcdef >> 32 -> 0x01234567", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ENDIAN(BPF_FROM_BE, R0, 64), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, (u32) (cpu_to_be64(0x0123456789abcdefLL) >> 32) } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_END | BPF_FROM_BE, reversed */ | 
|  | { | 
|  | "ALU_END_FROM_BE 16: 0xfedcba9876543210 -> 0x3210", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_ENDIAN(BPF_FROM_BE, R0, 16), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0,  cpu_to_be16(0x3210) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_BE 32: 0xfedcba9876543210 -> 0x76543210", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_ENDIAN(BPF_FROM_BE, R0, 32), | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, cpu_to_be32(0x76543210) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_BE 64: 0xfedcba9876543210 -> 0x76543210", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_ENDIAN(BPF_FROM_BE, R0, 64), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, (u32) cpu_to_be64(0xfedcba9876543210ULL) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_BE 64: 0xfedcba9876543210 >> 32 -> 0xfedcba98", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_ENDIAN(BPF_FROM_BE, R0, 64), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, (u32) (cpu_to_be64(0xfedcba9876543210ULL) >> 32) } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_END | BPF_FROM_LE */ | 
|  | { | 
|  | "ALU_END_FROM_LE 16: 0x0123456789abcdef -> 0xefcd", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ENDIAN(BPF_FROM_LE, R0, 16), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, cpu_to_le16(0xcdef) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_LE 32: 0x0123456789abcdef -> 0xefcdab89", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ENDIAN(BPF_FROM_LE, R0, 32), | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, cpu_to_le32(0x89abcdef) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_LE 64: 0x0123456789abcdef -> 0x67452301", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ENDIAN(BPF_FROM_LE, R0, 64), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, (u32) cpu_to_le64(0x0123456789abcdefLL) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_LE 64: 0x0123456789abcdef >> 32 -> 0xefcdab89", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_ENDIAN(BPF_FROM_LE, R0, 64), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, (u32) (cpu_to_le64(0x0123456789abcdefLL) >> 32) } }, | 
|  | }, | 
|  | /* BPF_ALU | BPF_END | BPF_FROM_LE, reversed */ | 
|  | { | 
|  | "ALU_END_FROM_LE 16: 0xfedcba9876543210 -> 0x1032", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_ENDIAN(BPF_FROM_LE, R0, 16), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0,  cpu_to_le16(0x3210) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_LE 32: 0xfedcba9876543210 -> 0x10325476", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_ENDIAN(BPF_FROM_LE, R0, 32), | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, cpu_to_le32(0x76543210) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_LE 64: 0xfedcba9876543210 -> 0x10325476", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_ENDIAN(BPF_FROM_LE, R0, 64), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, (u32) cpu_to_le64(0xfedcba9876543210ULL) } }, | 
|  | }, | 
|  | { | 
|  | "ALU_END_FROM_LE 64: 0xfedcba9876543210 >> 32 -> 0x98badcfe", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_ENDIAN(BPF_FROM_LE, R0, 64), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, (u32) (cpu_to_le64(0xfedcba9876543210ULL) >> 32) } }, | 
|  | }, | 
|  | /* BSWAP */ | 
|  | { | 
|  | "BSWAP 16: 0x0123456789abcdef -> 0xefcd", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_BSWAP(R0, 16), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xefcd } }, | 
|  | }, | 
|  | { | 
|  | "BSWAP 32: 0x0123456789abcdef -> 0xefcdab89", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_BSWAP(R0, 32), | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xefcdab89 } }, | 
|  | }, | 
|  | { | 
|  | "BSWAP 64: 0x0123456789abcdef -> 0x67452301", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_BSWAP(R0, 64), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x67452301 } }, | 
|  | }, | 
|  | { | 
|  | "BSWAP 64: 0x0123456789abcdef >> 32 -> 0xefcdab89", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0x0123456789abcdefLL), | 
|  | BPF_BSWAP(R0, 64), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xefcdab89 } }, | 
|  | }, | 
|  | /* BSWAP, reversed */ | 
|  | { | 
|  | "BSWAP 16: 0xfedcba9876543210 -> 0x1032", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_BSWAP(R0, 16), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1032 } }, | 
|  | }, | 
|  | { | 
|  | "BSWAP 32: 0xfedcba9876543210 -> 0x10325476", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_BSWAP(R0, 32), | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32), | 
|  | BPF_ALU32_REG(BPF_ADD, R0, R1), /* R1 = 0 */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x10325476 } }, | 
|  | }, | 
|  | { | 
|  | "BSWAP 64: 0xfedcba9876543210 -> 0x98badcfe", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_BSWAP(R0, 64), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x98badcfe } }, | 
|  | }, | 
|  | { | 
|  | "BSWAP 64: 0xfedcba9876543210 >> 32 -> 0x10325476", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_BSWAP(R0, 64), | 
|  | BPF_ALU64_IMM(BPF_RSH, R0, 32), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x10325476 } }, | 
|  | }, | 
|  | /* BPF_LDX_MEM B/H/W/DW */ | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_B, base", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0102030405060708ULL), | 
|  | BPF_LD_IMM64(R2, 0x0000000000000008ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEM(BPF_B, R0, R10, -1), | 
|  | #else | 
|  | BPF_LDX_MEM(BPF_B, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_B, MSB set", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R2, 0x0000000000000088ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEM(BPF_B, R0, R10, -1), | 
|  | #else | 
|  | BPF_LDX_MEM(BPF_B, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_B, negative offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000000088ULL), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 512), | 
|  | BPF_STX_MEM(BPF_B, R1, R2, -256), | 
|  | BPF_LDX_MEM(BPF_B, R0, R1, -256), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 512, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_B, small positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000000088ULL), | 
|  | BPF_STX_MEM(BPF_B, R1, R2, 256), | 
|  | BPF_LDX_MEM(BPF_B, R0, R1, 256), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 512, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_B, large positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000000088ULL), | 
|  | BPF_STX_MEM(BPF_B, R1, R2, 4096), | 
|  | BPF_LDX_MEM(BPF_B, R0, R1, 4096), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 4096 + 16, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_H, base", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0102030405060708ULL), | 
|  | BPF_LD_IMM64(R2, 0x0000000000000708ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEM(BPF_H, R0, R10, -2), | 
|  | #else | 
|  | BPF_LDX_MEM(BPF_H, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_H, MSB set", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R2, 0x0000000000008788ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEM(BPF_H, R0, R10, -2), | 
|  | #else | 
|  | BPF_LDX_MEM(BPF_H, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_H, negative offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000008788ULL), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 512), | 
|  | BPF_STX_MEM(BPF_H, R1, R2, -256), | 
|  | BPF_LDX_MEM(BPF_H, R0, R1, -256), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 512, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_H, small positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000008788ULL), | 
|  | BPF_STX_MEM(BPF_H, R1, R2, 256), | 
|  | BPF_LDX_MEM(BPF_H, R0, R1, 256), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 512, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_H, large positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000008788ULL), | 
|  | BPF_STX_MEM(BPF_H, R1, R2, 8192), | 
|  | BPF_LDX_MEM(BPF_H, R0, R1, 8192), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 8192 + 16, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_H, unaligned positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000000008788ULL), | 
|  | BPF_STX_MEM(BPF_H, R1, R2, 13), | 
|  | BPF_LDX_MEM(BPF_H, R0, R1, 13), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 32, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_W, base", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0102030405060708ULL), | 
|  | BPF_LD_IMM64(R2, 0x0000000005060708ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -4), | 
|  | #else | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_W, MSB set", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R2, 0x0000000085868788ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -4), | 
|  | #else | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_W, negative offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000085868788ULL), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 512), | 
|  | BPF_STX_MEM(BPF_W, R1, R2, -256), | 
|  | BPF_LDX_MEM(BPF_W, R0, R1, -256), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 512, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_W, small positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000085868788ULL), | 
|  | BPF_STX_MEM(BPF_W, R1, R2, 256), | 
|  | BPF_LDX_MEM(BPF_W, R0, R1, 256), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 512, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_W, large positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000085868788ULL), | 
|  | BPF_STX_MEM(BPF_W, R1, R2, 16384), | 
|  | BPF_LDX_MEM(BPF_W, R0, R1, 16384), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 16384 + 16, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_W, unaligned positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x0000000085868788ULL), | 
|  | BPF_STX_MEM(BPF_W, R1, R2, 13), | 
|  | BPF_LDX_MEM(BPF_W, R0, R1, 13), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 32, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_DW, base", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0102030405060708ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_DW, MSB set", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8182838485868788ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_DW, negative offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 512), | 
|  | BPF_STX_MEM(BPF_DW, R1, R2, -256), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R1, -256), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 512, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_DW, small positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_STX_MEM(BPF_DW, R1, R2, 256), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R1, 256), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 512, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_DW, large positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_STX_MEM(BPF_DW, R1, R2, 32760), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R1, 32760), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 32768, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEM | BPF_DW, unaligned positive offset", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_STX_MEM(BPF_DW, R1, R2, 13), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R1, 13), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_LARGE_MEM, | 
|  | { }, | 
|  | { { 32, 0 } }, | 
|  | .stack_depth = 0, | 
|  | }, | 
|  | /* BPF_LDX_MEMSX B/H/W */ | 
|  | { | 
|  | "BPF_LDX_MEMSX | BPF_B", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0xdead0000000000f0ULL), | 
|  | BPF_LD_IMM64(R2, 0xfffffffffffffff0ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEMSX(BPF_B, R0, R10, -1), | 
|  | #else | 
|  | BPF_LDX_MEMSX(BPF_B, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEMSX | BPF_H", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0xdead00000000f123ULL), | 
|  | BPF_LD_IMM64(R2, 0xfffffffffffff123ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEMSX(BPF_H, R0, R10, -2), | 
|  | #else | 
|  | BPF_LDX_MEMSX(BPF_H, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_LDX_MEMSX | BPF_W", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x00000000deadbeefULL), | 
|  | BPF_LD_IMM64(R2, 0xffffffffdeadbeefULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_LDX_MEMSX(BPF_W, R0, R10, -4), | 
|  | #else | 
|  | BPF_LDX_MEMSX(BPF_W, R0, R10, -8), | 
|  | #endif | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | /* BPF_STX_MEM B/H/W/DW */ | 
|  | { | 
|  | "BPF_STX_MEM | BPF_B", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL), | 
|  | BPF_LD_IMM64(R2, 0x0102030405060708ULL), | 
|  | BPF_LD_IMM64(R3, 0x8090a0b0c0d0e008ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_STX_MEM(BPF_B, R10, R2, -1), | 
|  | #else | 
|  | BPF_STX_MEM(BPF_B, R10, R2, -8), | 
|  | #endif | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_STX_MEM | BPF_B, MSB set", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL), | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x8090a0b0c0d0e088ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_STX_MEM(BPF_B, R10, R2, -1), | 
|  | #else | 
|  | BPF_STX_MEM(BPF_B, R10, R2, -8), | 
|  | #endif | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_STX_MEM | BPF_H", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL), | 
|  | BPF_LD_IMM64(R2, 0x0102030405060708ULL), | 
|  | BPF_LD_IMM64(R3, 0x8090a0b0c0d00708ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_STX_MEM(BPF_H, R10, R2, -2), | 
|  | #else | 
|  | BPF_STX_MEM(BPF_H, R10, R2, -8), | 
|  | #endif | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_STX_MEM | BPF_H, MSB set", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL), | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x8090a0b0c0d08788ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_STX_MEM(BPF_H, R10, R2, -2), | 
|  | #else | 
|  | BPF_STX_MEM(BPF_H, R10, R2, -8), | 
|  | #endif | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_STX_MEM | BPF_W", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL), | 
|  | BPF_LD_IMM64(R2, 0x0102030405060708ULL), | 
|  | BPF_LD_IMM64(R3, 0x8090a0b005060708ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_STX_MEM(BPF_W, R10, R2, -4), | 
|  | #else | 
|  | BPF_STX_MEM(BPF_W, R10, R2, -8), | 
|  | #endif | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "BPF_STX_MEM | BPF_W, MSB set", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x8090a0b0c0d0e0f0ULL), | 
|  | BPF_LD_IMM64(R2, 0x8182838485868788ULL), | 
|  | BPF_LD_IMM64(R3, 0x8090a0b085868788ULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | #ifdef __BIG_ENDIAN | 
|  | BPF_STX_MEM(BPF_W, R10, R2, -4), | 
|  | #else | 
|  | BPF_STX_MEM(BPF_W, R10, R2, -8), | 
|  | #endif | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R3, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | /* BPF_ST(X) | BPF_MEM | BPF_B/H/W/DW */ | 
|  | { | 
|  | "ST_MEM_B: Store/Load byte: max negative", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_ST_MEM(BPF_B, R10, -40, 0xff), | 
|  | BPF_LDX_MEM(BPF_B, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "ST_MEM_B: Store/Load byte: max positive", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_ST_MEM(BPF_H, R10, -40, 0x7f), | 
|  | BPF_LDX_MEM(BPF_H, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x7f } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "STX_MEM_B: Store/Load byte: max negative", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xffLL), | 
|  | BPF_STX_MEM(BPF_B, R10, R1, -40), | 
|  | BPF_LDX_MEM(BPF_B, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "ST_MEM_H: Store/Load half word: max negative", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_ST_MEM(BPF_H, R10, -40, 0xffff), | 
|  | BPF_LDX_MEM(BPF_H, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "ST_MEM_H: Store/Load half word: max positive", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_ST_MEM(BPF_H, R10, -40, 0x7fff), | 
|  | BPF_LDX_MEM(BPF_H, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x7fff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "STX_MEM_H: Store/Load half word: max negative", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xffffLL), | 
|  | BPF_STX_MEM(BPF_H, R10, R1, -40), | 
|  | BPF_LDX_MEM(BPF_H, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "ST_MEM_W: Store/Load word: max negative", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_ST_MEM(BPF_W, R10, -40, 0xffffffff), | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "ST_MEM_W: Store/Load word: max positive", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_ST_MEM(BPF_W, R10, -40, 0x7fffffff), | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x7fffffff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "STX_MEM_W: Store/Load word: max negative", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xffffffffLL), | 
|  | BPF_STX_MEM(BPF_W, R10, R1, -40), | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "ST_MEM_DW: Store/Load double word: max negative", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_ST_MEM(BPF_DW, R10, -40, 0xffffffff), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "ST_MEM_DW: Store/Load double word: max negative 2", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R2, 0xffff00000000ffffLL), | 
|  | BPF_LD_IMM64(R3, 0xffffffffffffffffLL), | 
|  | BPF_ST_MEM(BPF_DW, R10, -40, 0xffffffff), | 
|  | BPF_LDX_MEM(BPF_DW, R2, R10, -40), | 
|  | BPF_JMP_REG(BPF_JEQ, R2, R3, 2), | 
|  | BPF_MOV32_IMM(R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_MOV32_IMM(R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x1 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "ST_MEM_DW: Store/Load double word: max positive", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_ST_MEM(BPF_DW, R10, -40, 0x7fffffff), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x7fffffff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "STX_MEM_DW: Store/Load double word: max negative", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xffffffffffffffffLL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -40), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "STX_MEM_DW: Store double word: first word in memory", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_LD_IMM64(R1, 0x0123456789abcdefLL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -40), | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | #ifdef __BIG_ENDIAN | 
|  | { { 0, 0x01234567 } }, | 
|  | #else | 
|  | { { 0, 0x89abcdef } }, | 
|  | #endif | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "STX_MEM_DW: Store double word: second word in memory", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R0, 0), | 
|  | BPF_LD_IMM64(R1, 0x0123456789abcdefLL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -40), | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -36), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | #ifdef __BIG_ENDIAN | 
|  | { { 0, 0x89abcdef } }, | 
|  | #else | 
|  | { { 0, 0x01234567 } }, | 
|  | #endif | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | /* BPF_STX | BPF_ATOMIC | BPF_W/DW */ | 
|  | { | 
|  | "STX_XADD_W: X + 1 + 1 + 1 + ...", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4134 } }, | 
|  | .fill_helper = bpf_fill_stxw, | 
|  | }, | 
|  | { | 
|  | "STX_XADD_DW: X + 1 + 1 + 1 + ...", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 4134 } }, | 
|  | .fill_helper = bpf_fill_stxdw, | 
|  | }, | 
|  | /* | 
|  | * Exhaustive tests of atomic operation variants. | 
|  | * Individual tests are expanded from template macros for all | 
|  | * combinations of ALU operation, word size and fetching. | 
|  | */ | 
|  | #define BPF_ATOMIC_POISON(width) ((width) == BPF_W ? (0xbaadf00dULL << 32) : 0) | 
|  |  | 
|  | #define BPF_ATOMIC_OP_TEST1(width, op, logic, old, update, result)	\ | 
|  | {									\ | 
|  | "BPF_ATOMIC | " #width ", " #op ": Test: "			\ | 
|  | #old " " #logic " " #update " = " #result,		\ | 
|  | .u.insns_int = {						\ | 
|  | BPF_LD_IMM64(R5, (update) | BPF_ATOMIC_POISON(width)),	\ | 
|  | BPF_ST_MEM(width, R10, -40, old),			\ | 
|  | BPF_ATOMIC_OP(width, op, R10, R5, -40),			\ | 
|  | BPF_LDX_MEM(width, R0, R10, -40),			\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),				\ | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32),				\ | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1),				\ | 
|  | BPF_EXIT_INSN(),					\ | 
|  | },								\ | 
|  | INTERNAL,							\ | 
|  | { },								\ | 
|  | { { 0, result } },						\ | 
|  | .stack_depth = 40,						\ | 
|  | } | 
|  | #define BPF_ATOMIC_OP_TEST2(width, op, logic, old, update, result)	\ | 
|  | {									\ | 
|  | "BPF_ATOMIC | " #width ", " #op ": Test side effects, r10: "	\ | 
|  | #old " " #logic " " #update " = " #result,		\ | 
|  | .u.insns_int = {						\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R10),			\ | 
|  | BPF_LD_IMM64(R0, (update) | BPF_ATOMIC_POISON(width)),	\ | 
|  | BPF_ST_MEM(BPF_W, R10, -40, old),			\ | 
|  | BPF_ATOMIC_OP(width, op, R10, R0, -40),			\ | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R10),			\ | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1),				\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),				\ | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32),				\ | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1),				\ | 
|  | BPF_EXIT_INSN(),					\ | 
|  | },								\ | 
|  | INTERNAL,							\ | 
|  | { },								\ | 
|  | { { 0, 0 } },							\ | 
|  | .stack_depth = 40,						\ | 
|  | } | 
|  | #define BPF_ATOMIC_OP_TEST3(width, op, logic, old, update, result)	\ | 
|  | {									\ | 
|  | "BPF_ATOMIC | " #width ", " #op ": Test side effects, r0: "	\ | 
|  | #old " " #logic " " #update " = " #result,		\ | 
|  | .u.insns_int = {						\ | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R10),			\ | 
|  | BPF_LD_IMM64(R1, (update) | BPF_ATOMIC_POISON(width)),	\ | 
|  | BPF_ST_MEM(width, R10, -40, old),			\ | 
|  | BPF_ATOMIC_OP(width, op, R10, R1, -40),			\ | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R10),			\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),				\ | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32),				\ | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1),				\ | 
|  | BPF_EXIT_INSN(),					\ | 
|  | },								\ | 
|  | INTERNAL,                                                       \ | 
|  | { },                                                            \ | 
|  | { { 0, 0 } },                                                   \ | 
|  | .stack_depth = 40,                                              \ | 
|  | } | 
|  | #define BPF_ATOMIC_OP_TEST4(width, op, logic, old, update, result)	\ | 
|  | {									\ | 
|  | "BPF_ATOMIC | " #width ", " #op ": Test fetch: "		\ | 
|  | #old " " #logic " " #update " = " #result,		\ | 
|  | .u.insns_int = {						\ | 
|  | BPF_LD_IMM64(R3, (update) | BPF_ATOMIC_POISON(width)),	\ | 
|  | BPF_ST_MEM(width, R10, -40, old),			\ | 
|  | BPF_ATOMIC_OP(width, op, R10, R3, -40),			\ | 
|  | BPF_ALU32_REG(BPF_MOV, R0, R3),                         \ | 
|  | BPF_EXIT_INSN(),					\ | 
|  | },								\ | 
|  | INTERNAL,                                                       \ | 
|  | { },                                                            \ | 
|  | { { 0, (op) & BPF_FETCH ? old : update } },			\ | 
|  | .stack_depth = 40,                                              \ | 
|  | } | 
|  | /* BPF_ATOMIC | BPF_W: BPF_ADD */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_ADD, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_ADD, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_ADD, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_ADD, +, 0x12, 0xab, 0xbd), | 
|  | /* BPF_ATOMIC | BPF_W: BPF_ADD | BPF_FETCH */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_ADD */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_ADD, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_ADD, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_ADD, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_ADD, +, 0x12, 0xab, 0xbd), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_ADD | BPF_FETCH */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_ADD | BPF_FETCH, +, 0x12, 0xab, 0xbd), | 
|  | /* BPF_ATOMIC | BPF_W: BPF_AND */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_AND, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_AND, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_AND, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_AND, &, 0x12, 0xab, 0x02), | 
|  | /* BPF_ATOMIC | BPF_W: BPF_AND | BPF_FETCH */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_AND */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_AND, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_AND, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_AND, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_AND, &, 0x12, 0xab, 0x02), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_AND | BPF_FETCH */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_AND | BPF_FETCH, &, 0x12, 0xab, 0x02), | 
|  | /* BPF_ATOMIC | BPF_W: BPF_OR */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_OR, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_OR, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_OR, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_OR, |, 0x12, 0xab, 0xbb), | 
|  | /* BPF_ATOMIC | BPF_W: BPF_OR | BPF_FETCH */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_OR */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_OR, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_OR, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_OR, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_OR, |, 0x12, 0xab, 0xbb), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_OR | BPF_FETCH */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_OR | BPF_FETCH, |, 0x12, 0xab, 0xbb), | 
|  | /* BPF_ATOMIC | BPF_W: BPF_XOR */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_XOR, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_XOR, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_XOR, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_XOR, ^, 0x12, 0xab, 0xb9), | 
|  | /* BPF_ATOMIC | BPF_W: BPF_XOR | BPF_FETCH */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_XOR */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_XOR, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_XOR, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_XOR, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_XOR, ^, 0x12, 0xab, 0xb9), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_XOR | BPF_FETCH */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_XOR | BPF_FETCH, ^, 0x12, 0xab, 0xb9), | 
|  | /* BPF_ATOMIC | BPF_W: BPF_XCHG */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_W, BPF_XCHG, xchg, 0x12, 0xab, 0xab), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_W, BPF_XCHG, xchg, 0x12, 0xab, 0xab), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_W, BPF_XCHG, xchg, 0x12, 0xab, 0xab), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_W, BPF_XCHG, xchg, 0x12, 0xab, 0xab), | 
|  | /* BPF_ATOMIC | BPF_DW: BPF_XCHG */ | 
|  | BPF_ATOMIC_OP_TEST1(BPF_DW, BPF_XCHG, xchg, 0x12, 0xab, 0xab), | 
|  | BPF_ATOMIC_OP_TEST2(BPF_DW, BPF_XCHG, xchg, 0x12, 0xab, 0xab), | 
|  | BPF_ATOMIC_OP_TEST3(BPF_DW, BPF_XCHG, xchg, 0x12, 0xab, 0xab), | 
|  | BPF_ATOMIC_OP_TEST4(BPF_DW, BPF_XCHG, xchg, 0x12, 0xab, 0xab), | 
|  | #undef BPF_ATOMIC_POISON | 
|  | #undef BPF_ATOMIC_OP_TEST1 | 
|  | #undef BPF_ATOMIC_OP_TEST2 | 
|  | #undef BPF_ATOMIC_OP_TEST3 | 
|  | #undef BPF_ATOMIC_OP_TEST4 | 
|  | /* BPF_ATOMIC | BPF_W, BPF_CMPXCHG */ | 
|  | { | 
|  | "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test successful return", | 
|  | .u.insns_int = { | 
|  | BPF_ST_MEM(BPF_W, R10, -40, 0x01234567), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x01234567), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef), | 
|  | BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x01234567 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test successful store", | 
|  | .u.insns_int = { | 
|  | BPF_ST_MEM(BPF_W, R10, -40, 0x01234567), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x01234567), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef), | 
|  | BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40), | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test failure return", | 
|  | .u.insns_int = { | 
|  | BPF_ST_MEM(BPF_W, R10, -40, 0x01234567), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x76543210), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef), | 
|  | BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x01234567 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test failure store", | 
|  | .u.insns_int = { | 
|  | BPF_ST_MEM(BPF_W, R10, -40, 0x01234567), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x76543210), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef), | 
|  | BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40), | 
|  | BPF_LDX_MEM(BPF_W, R0, R10, -40), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x01234567 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "BPF_ATOMIC | BPF_W, BPF_CMPXCHG: Test side effects", | 
|  | .u.insns_int = { | 
|  | BPF_ST_MEM(BPF_W, R10, -40, 0x01234567), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x01234567), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0x89abcdef), | 
|  | BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40), | 
|  | BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R3, -40), | 
|  | BPF_ALU32_REG(BPF_MOV, R0, R3), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x89abcdef } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | /* BPF_ATOMIC | BPF_DW, BPF_CMPXCHG */ | 
|  | { | 
|  | "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test successful return", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0123456789abcdefULL), | 
|  | BPF_LD_IMM64(R2, 0xfedcba9876543210ULL), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -40), | 
|  | BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R1, 1), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test successful store", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0123456789abcdefULL), | 
|  | BPF_LD_IMM64(R2, 0xfedcba9876543210ULL), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_STX_MEM(BPF_DW, R10, R0, -40), | 
|  | BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -40), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test failure return", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0123456789abcdefULL), | 
|  | BPF_LD_IMM64(R2, 0xfedcba9876543210ULL), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 1), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -40), | 
|  | BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R1, 1), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test failure store", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0123456789abcdefULL), | 
|  | BPF_LD_IMM64(R2, 0xfedcba9876543210ULL), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 1), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -40), | 
|  | BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R10, -40), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R1, 1), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | { | 
|  | "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test side effects", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0123456789abcdefULL), | 
|  | BPF_LD_IMM64(R2, 0xfedcba9876543210ULL), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -40), | 
|  | BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40), | 
|  | BPF_LD_IMM64(R0, 0xfedcba9876543210ULL), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 40, | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JEQ | BPF_K */ | 
|  | { | 
|  | "JMP32_JEQ_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 123), | 
|  | BPF_JMP32_IMM(BPF_JEQ, R0, 321, 1), | 
|  | BPF_JMP32_IMM(BPF_JEQ, R0, 123, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JEQ_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 12345678), | 
|  | BPF_JMP32_IMM(BPF_JEQ, R0, 12345678 & 0xffff, 1), | 
|  | BPF_JMP32_IMM(BPF_JEQ, R0, 12345678, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 12345678 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JEQ_K: negative immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_JMP32_IMM(BPF_JEQ, R0,  123, 1), | 
|  | BPF_JMP32_IMM(BPF_JEQ, R0, -123, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JEQ | BPF_X */ | 
|  | { | 
|  | "JMP32_JEQ_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1234), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4321), | 
|  | BPF_JMP32_REG(BPF_JEQ, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1234), | 
|  | BPF_JMP32_REG(BPF_JEQ, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1234 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JNE | BPF_K */ | 
|  | { | 
|  | "JMP32_JNE_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 123), | 
|  | BPF_JMP32_IMM(BPF_JNE, R0, 123, 1), | 
|  | BPF_JMP32_IMM(BPF_JNE, R0, 321, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JNE_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 12345678), | 
|  | BPF_JMP32_IMM(BPF_JNE, R0, 12345678, 1), | 
|  | BPF_JMP32_IMM(BPF_JNE, R0, 12345678 & 0xffff, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 12345678 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JNE_K: negative immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_JMP32_IMM(BPF_JNE, R0, -123, 1), | 
|  | BPF_JMP32_IMM(BPF_JNE, R0,  123, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JNE | BPF_X */ | 
|  | { | 
|  | "JMP32_JNE_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1234), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 1234), | 
|  | BPF_JMP32_REG(BPF_JNE, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 4321), | 
|  | BPF_JMP32_REG(BPF_JNE, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1234 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSET | BPF_K */ | 
|  | { | 
|  | "JMP32_JSET_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP32_IMM(BPF_JSET, R0, 2, 1), | 
|  | BPF_JMP32_IMM(BPF_JSET, R0, 3, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JSET_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x40000000), | 
|  | BPF_JMP32_IMM(BPF_JSET, R0, 0x3fffffff, 1), | 
|  | BPF_JMP32_IMM(BPF_JSET, R0, 0x60000000, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x40000000 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JSET_K: negative immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_JMP32_IMM(BPF_JSET, R0, -1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSET | BPF_X */ | 
|  | { | 
|  | "JMP32_JSET_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 8), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 7), | 
|  | BPF_JMP32_REG(BPF_JSET, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 8 | 2), | 
|  | BPF_JMP32_REG(BPF_JNE, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 8 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JGT | BPF_K */ | 
|  | { | 
|  | "JMP32_JGT_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 123), | 
|  | BPF_JMP32_IMM(BPF_JGT, R0, 123, 1), | 
|  | BPF_JMP32_IMM(BPF_JGT, R0, 122, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JGT_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe), | 
|  | BPF_JMP32_IMM(BPF_JGT, R0, 0xffffffff, 1), | 
|  | BPF_JMP32_IMM(BPF_JGT, R0, 0xfffffffd, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JGT | BPF_X */ | 
|  | { | 
|  | "JMP32_JGT_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_JMP32_REG(BPF_JGT, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffd), | 
|  | BPF_JMP32_REG(BPF_JGT, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JGE | BPF_K */ | 
|  | { | 
|  | "JMP32_JGE_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 123), | 
|  | BPF_JMP32_IMM(BPF_JGE, R0, 124, 1), | 
|  | BPF_JMP32_IMM(BPF_JGE, R0, 123, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JGE_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe), | 
|  | BPF_JMP32_IMM(BPF_JGE, R0, 0xffffffff, 1), | 
|  | BPF_JMP32_IMM(BPF_JGE, R0, 0xfffffffe, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JGE | BPF_X */ | 
|  | { | 
|  | "JMP32_JGE_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_JMP32_REG(BPF_JGE, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffe), | 
|  | BPF_JMP32_REG(BPF_JGE, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JLT | BPF_K */ | 
|  | { | 
|  | "JMP32_JLT_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 123), | 
|  | BPF_JMP32_IMM(BPF_JLT, R0, 123, 1), | 
|  | BPF_JMP32_IMM(BPF_JLT, R0, 124, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JLT_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe), | 
|  | BPF_JMP32_IMM(BPF_JLT, R0, 0xfffffffd, 1), | 
|  | BPF_JMP32_IMM(BPF_JLT, R0, 0xffffffff, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JLT | BPF_X */ | 
|  | { | 
|  | "JMP32_JLT_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffd), | 
|  | BPF_JMP32_REG(BPF_JLT, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xffffffff), | 
|  | BPF_JMP32_REG(BPF_JLT, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JLE | BPF_K */ | 
|  | { | 
|  | "JMP32_JLE_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 123), | 
|  | BPF_JMP32_IMM(BPF_JLE, R0, 122, 1), | 
|  | BPF_JMP32_IMM(BPF_JLE, R0, 123, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JLE_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe), | 
|  | BPF_JMP32_IMM(BPF_JLE, R0, 0xfffffffd, 1), | 
|  | BPF_JMP32_IMM(BPF_JLE, R0, 0xfffffffe, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JLE | BPF_X */ | 
|  | { | 
|  | "JMP32_JLE_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0xfffffffe), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffd), | 
|  | BPF_JMP32_REG(BPF_JLE, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xfffffffe), | 
|  | BPF_JMP32_REG(BPF_JLE, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xfffffffe } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSGT | BPF_K */ | 
|  | { | 
|  | "JMP32_JSGT_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_JMP32_IMM(BPF_JSGT, R0, -123, 1), | 
|  | BPF_JMP32_IMM(BPF_JSGT, R0, -124, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JSGT_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -12345678), | 
|  | BPF_JMP32_IMM(BPF_JSGT, R0, -12345678, 1), | 
|  | BPF_JMP32_IMM(BPF_JSGT, R0, -12345679, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -12345678 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSGT | BPF_X */ | 
|  | { | 
|  | "JMP32_JSGT_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -12345678), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -12345678), | 
|  | BPF_JMP32_REG(BPF_JSGT, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -12345679), | 
|  | BPF_JMP32_REG(BPF_JSGT, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -12345678 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSGE | BPF_K */ | 
|  | { | 
|  | "JMP32_JSGE_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_JMP32_IMM(BPF_JSGE, R0, -122, 1), | 
|  | BPF_JMP32_IMM(BPF_JSGE, R0, -123, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JSGE_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -12345678), | 
|  | BPF_JMP32_IMM(BPF_JSGE, R0, -12345677, 1), | 
|  | BPF_JMP32_IMM(BPF_JSGE, R0, -12345678, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -12345678 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSGE | BPF_X */ | 
|  | { | 
|  | "JMP32_JSGE_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -12345678), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -12345677), | 
|  | BPF_JMP32_REG(BPF_JSGE, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -12345678), | 
|  | BPF_JMP32_REG(BPF_JSGE, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -12345678 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSLT | BPF_K */ | 
|  | { | 
|  | "JMP32_JSLT_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_JMP32_IMM(BPF_JSLT, R0, -123, 1), | 
|  | BPF_JMP32_IMM(BPF_JSLT, R0, -122, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JSLT_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -12345678), | 
|  | BPF_JMP32_IMM(BPF_JSLT, R0, -12345678, 1), | 
|  | BPF_JMP32_IMM(BPF_JSLT, R0, -12345677, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -12345678 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSLT | BPF_X */ | 
|  | { | 
|  | "JMP32_JSLT_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -12345678), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -12345678), | 
|  | BPF_JMP32_REG(BPF_JSLT, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -12345677), | 
|  | BPF_JMP32_REG(BPF_JSLT, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -12345678 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSLE | BPF_K */ | 
|  | { | 
|  | "JMP32_JSLE_K: Small immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -123), | 
|  | BPF_JMP32_IMM(BPF_JSLE, R0, -124, 1), | 
|  | BPF_JMP32_IMM(BPF_JSLE, R0, -123, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -123 } } | 
|  | }, | 
|  | { | 
|  | "JMP32_JSLE_K: Large immediate", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -12345678), | 
|  | BPF_JMP32_IMM(BPF_JSLE, R0, -12345679, 1), | 
|  | BPF_JMP32_IMM(BPF_JSLE, R0, -12345678, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -12345678 } } | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JSLE | BPF_K */ | 
|  | { | 
|  | "JMP32_JSLE_X", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -12345678), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -12345679), | 
|  | BPF_JMP32_REG(BPF_JSLE, R0, R1, 2), | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, -12345678), | 
|  | BPF_JMP32_REG(BPF_JSLE, R0, R1, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, -12345678 } } | 
|  | }, | 
|  | /* BPF_JMP | BPF_EXIT */ | 
|  | { | 
|  | "JMP_EXIT", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x4711), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0x4712), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0x4711 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JA */ | 
|  | { | 
|  | "JMP_JA: Unconditional jump: if (true) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP_IMM(BPF_JA, 0, 0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP32 | BPF_JA */ | 
|  | { | 
|  | "JMP32_JA: Unconditional jump: if (true) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP32_IMM(BPF_JA, 0, 1, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSLT | BPF_K */ | 
|  | { | 
|  | "JMP_JSLT_K: Signed jump: if (-2 < -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xfffffffffffffffeLL), | 
|  | BPF_JMP_IMM(BPF_JSLT, R1, -1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLT_K: Signed jump: if (-1 < -1) return 0", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_LD_IMM64(R1, 0xffffffffffffffffLL), | 
|  | BPF_JMP_IMM(BPF_JSLT, R1, -1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSGT | BPF_K */ | 
|  | { | 
|  | "JMP_JSGT_K: Signed jump: if (-1 > -2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xffffffffffffffffLL), | 
|  | BPF_JMP_IMM(BPF_JSGT, R1, -2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGT_K: Signed jump: if (-1 > -1) return 0", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_LD_IMM64(R1, 0xffffffffffffffffLL), | 
|  | BPF_JMP_IMM(BPF_JSGT, R1, -1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSLE | BPF_K */ | 
|  | { | 
|  | "JMP_JSLE_K: Signed jump: if (-2 <= -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xfffffffffffffffeLL), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, -1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLE_K: Signed jump: if (-1 <= -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xffffffffffffffffLL), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, -1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLE_K: Signed jump: value walk 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, 0, 6), | 
|  | BPF_ALU64_IMM(BPF_SUB, R1, 1), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, 0, 4), | 
|  | BPF_ALU64_IMM(BPF_SUB, R1, 1), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, 0, 2), | 
|  | BPF_ALU64_IMM(BPF_SUB, R1, 1), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, 0, 1), | 
|  | BPF_EXIT_INSN(),		/* bad exit */ | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1),	/* good exit */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLE_K: Signed jump: value walk 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, 0, 4), | 
|  | BPF_ALU64_IMM(BPF_SUB, R1, 2), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, 0, 2), | 
|  | BPF_ALU64_IMM(BPF_SUB, R1, 2), | 
|  | BPF_JMP_IMM(BPF_JSLE, R1, 0, 1), | 
|  | BPF_EXIT_INSN(),		/* bad exit */ | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1),	/* good exit */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSGE | BPF_K */ | 
|  | { | 
|  | "JMP_JSGE_K: Signed jump: if (-1 >= -2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xffffffffffffffffLL), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, -2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGE_K: Signed jump: if (-1 >= -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 0xffffffffffffffffLL), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, -1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGE_K: Signed jump: value walk 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -3), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, 0, 6), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 1), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, 0, 4), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 1), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, 0, 2), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 1), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, 0, 1), | 
|  | BPF_EXIT_INSN(),		/* bad exit */ | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1),	/* good exit */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGE_K: Signed jump: value walk 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -3), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, 0, 4), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 2), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, 0, 2), | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 2), | 
|  | BPF_JMP_IMM(BPF_JSGE, R1, 0, 1), | 
|  | BPF_EXIT_INSN(),		/* bad exit */ | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1),	/* good exit */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JGT | BPF_K */ | 
|  | { | 
|  | "JMP_JGT_K: if (3 > 2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JGT, R1, 2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGT_K: Unsigned jump: if (-1 > 1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_JMP_IMM(BPF_JGT, R1, 1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JLT | BPF_K */ | 
|  | { | 
|  | "JMP_JLT_K: if (2 < 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 2), | 
|  | BPF_JMP_IMM(BPF_JLT, R1, 3, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGT_K: Unsigned jump: if (1 < -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 1), | 
|  | BPF_JMP_IMM(BPF_JLT, R1, -1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JGE | BPF_K */ | 
|  | { | 
|  | "JMP_JGE_K: if (3 >= 2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JGE, R1, 2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JLE | BPF_K */ | 
|  | { | 
|  | "JMP_JLE_K: if (2 <= 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 2), | 
|  | BPF_JMP_IMM(BPF_JLE, R1, 3, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JGT | BPF_K jump backwards */ | 
|  | { | 
|  | "JMP_JGT_K: if (3 > 2) return 1 (jump backwards)", | 
|  | .u.insns_int = { | 
|  | BPF_JMP_IMM(BPF_JA, 0, 0, 2), /* goto start */ | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), /* out: */ | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), /* start: */ | 
|  | BPF_LD_IMM64(R1, 3), /* note: this takes 2 insns */ | 
|  | BPF_JMP_IMM(BPF_JGT, R1, 2, -6), /* goto out */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGE_K: if (3 >= 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JGE, R1, 3, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JLT | BPF_K jump backwards */ | 
|  | { | 
|  | "JMP_JGT_K: if (2 < 3) return 1 (jump backwards)", | 
|  | .u.insns_int = { | 
|  | BPF_JMP_IMM(BPF_JA, 0, 0, 2), /* goto start */ | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), /* out: */ | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), /* start: */ | 
|  | BPF_LD_IMM64(R1, 2), /* note: this takes 2 insns */ | 
|  | BPF_JMP_IMM(BPF_JLT, R1, 3, -6), /* goto out */ | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_K: if (3 <= 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JLE, R1, 3, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JNE | BPF_K */ | 
|  | { | 
|  | "JMP_JNE_K: if (3 != 2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JNE, R1, 2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JEQ | BPF_K */ | 
|  | { | 
|  | "JMP_JEQ_K: if (3 == 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JEQ, R1, 3, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSET | BPF_K */ | 
|  | { | 
|  | "JMP_JSET_K: if (0x3 & 0x2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JSET, R1, 2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSET_K: if (0x3 & 0xffffffff) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_JMP_IMM(BPF_JSET, R1, 0xffffffff, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSGT | BPF_X */ | 
|  | { | 
|  | "JMP_JSGT_X: Signed jump: if (-1 > -2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, -2), | 
|  | BPF_JMP_REG(BPF_JSGT, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGT_X: Signed jump: if (-1 > -1) return 0", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, -1), | 
|  | BPF_JMP_REG(BPF_JSGT, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSLT | BPF_X */ | 
|  | { | 
|  | "JMP_JSLT_X: Signed jump: if (-2 < -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, -2), | 
|  | BPF_JMP_REG(BPF_JSLT, R2, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLT_X: Signed jump: if (-1 < -1) return 0", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, -1), | 
|  | BPF_JMP_REG(BPF_JSLT, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSGE | BPF_X */ | 
|  | { | 
|  | "JMP_JSGE_X: Signed jump: if (-1 >= -2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, -2), | 
|  | BPF_JMP_REG(BPF_JSGE, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGE_X: Signed jump: if (-1 >= -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, -1), | 
|  | BPF_JMP_REG(BPF_JSGE, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSLE | BPF_X */ | 
|  | { | 
|  | "JMP_JSLE_X: Signed jump: if (-2 <= -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, -2), | 
|  | BPF_JMP_REG(BPF_JSLE, R2, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLE_X: Signed jump: if (-1 <= -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, -1), | 
|  | BPF_JMP_REG(BPF_JSLE, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JGT | BPF_X */ | 
|  | { | 
|  | "JMP_JGT_X: if (3 > 2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JGT, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGT_X: Unsigned jump: if (-1 > 1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, 1), | 
|  | BPF_JMP_REG(BPF_JGT, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JLT | BPF_X */ | 
|  | { | 
|  | "JMP_JLT_X: if (2 < 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JLT, R2, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLT_X: Unsigned jump: if (1 < -1) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, -1), | 
|  | BPF_LD_IMM64(R2, 1), | 
|  | BPF_JMP_REG(BPF_JLT, R2, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JGE | BPF_X */ | 
|  | { | 
|  | "JMP_JGE_X: if (3 >= 2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JGE, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGE_X: if (3 >= 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 3), | 
|  | BPF_JMP_REG(BPF_JGE, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JLE | BPF_X */ | 
|  | { | 
|  | "JMP_JLE_X: if (2 <= 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JLE, R2, R1, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_X: if (3 <= 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 3), | 
|  | BPF_JMP_REG(BPF_JLE, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | /* Mainly testing JIT + imm64 here. */ | 
|  | "JMP_JGE_X: ldimm64 test 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JGE, R1, R2, 2), | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffULL), | 
|  | BPF_LD_IMM64(R0, 0xeeeeeeeeeeeeeeeeULL), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xeeeeeeeeU } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGE_X: ldimm64 test 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JGE, R1, R2, 0), | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffULL), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffffU } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGE_X: ldimm64 test 3", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JGE, R1, R2, 4), | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffULL), | 
|  | BPF_LD_IMM64(R0, 0xeeeeeeeeeeeeeeeeULL), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_X: ldimm64 test 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JLE, R2, R1, 2), | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffULL), | 
|  | BPF_LD_IMM64(R0, 0xeeeeeeeeeeeeeeeeULL), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xeeeeeeeeU } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_X: ldimm64 test 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JLE, R2, R1, 0), | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffULL), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0xffffffffU } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_X: ldimm64 test 3", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JLE, R2, R1, 4), | 
|  | BPF_LD_IMM64(R0, 0xffffffffffffffffULL), | 
|  | BPF_LD_IMM64(R0, 0xeeeeeeeeeeeeeeeeULL), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JNE | BPF_X */ | 
|  | { | 
|  | "JMP_JNE_X: if (3 != 2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JNE, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JEQ | BPF_X */ | 
|  | { | 
|  | "JMP_JEQ_X: if (3 == 3) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 3), | 
|  | BPF_JMP_REG(BPF_JEQ, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | /* BPF_JMP | BPF_JSET | BPF_X */ | 
|  | { | 
|  | "JMP_JSET_X: if (0x3 & 0x2) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 2), | 
|  | BPF_JMP_REG(BPF_JSET, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSET_X: if (0x3 & 0xffffffff) return 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 0), | 
|  | BPF_LD_IMM64(R1, 3), | 
|  | BPF_LD_IMM64(R2, 0xffffffff), | 
|  | BPF_JMP_REG(BPF_JSET, R1, R2, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JA: Jump, gap, jump, ...", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xababcbac } }, | 
|  | .fill_helper = bpf_fill_ja, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Maximum possible literals", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | .fill_helper = bpf_fill_maxinsns1, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Single literal", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xfefefefe } }, | 
|  | .fill_helper = bpf_fill_maxinsns2, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Run/add until end", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0x947bf368 } }, | 
|  | .fill_helper = bpf_fill_maxinsns3, | 
|  | }, | 
|  | { | 
|  | "BPF_MAXINSNS: Too many instructions", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA | FLAG_EXPECTED_FAIL, | 
|  | { }, | 
|  | { }, | 
|  | .fill_helper = bpf_fill_maxinsns4, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Very long jump", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xabababab } }, | 
|  | .fill_helper = bpf_fill_maxinsns5, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Ctx heavy transformations", | 
|  | { }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { | 
|  | {  1, SKB_VLAN_PRESENT }, | 
|  | { 10, SKB_VLAN_PRESENT } | 
|  | }, | 
|  | .fill_helper = bpf_fill_maxinsns6, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Call heavy transformations", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 1, 0 }, { 10, 0 } }, | 
|  | .fill_helper = bpf_fill_maxinsns7, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Jump heavy test", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xffffffff } }, | 
|  | .fill_helper = bpf_fill_maxinsns8, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Very long jump backwards", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xcbababab } }, | 
|  | .fill_helper = bpf_fill_maxinsns9, | 
|  | }, | 
|  | {	/* Mainly checking JIT here. */ | 
|  | "BPF_MAXINSNS: Edge hopping nuthouse", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xabababac } }, | 
|  | .fill_helper = bpf_fill_maxinsns10, | 
|  | }, | 
|  | { | 
|  | "BPF_MAXINSNS: Jump, gap, jump, ...", | 
|  | { }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0xababcbac } }, | 
|  | .fill_helper = bpf_fill_maxinsns11, | 
|  | }, | 
|  | { | 
|  | "BPF_MAXINSNS: jump over MSH", | 
|  | { }, | 
|  | CLASSIC | FLAG_EXPECTED_FAIL, | 
|  | { 0xfa, 0xfb, 0xfc, 0xfd, }, | 
|  | { { 4, 0xabababab } }, | 
|  | .fill_helper = bpf_fill_maxinsns12, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "BPF_MAXINSNS: exec all MSH", | 
|  | { }, | 
|  | CLASSIC, | 
|  | { 0xfa, 0xfb, 0xfc, 0xfd, }, | 
|  | { { 4, 0xababab83 } }, | 
|  | .fill_helper = bpf_fill_maxinsns13, | 
|  | }, | 
|  | { | 
|  | "BPF_MAXINSNS: ld_abs+get_processor_id", | 
|  | { }, | 
|  | CLASSIC, | 
|  | { }, | 
|  | { { 1, 0xbee } }, | 
|  | .fill_helper = bpf_fill_ld_abs_get_processor_id, | 
|  | }, | 
|  | /* | 
|  | * LD_IND / LD_ABS on fragmented SKBs | 
|  | */ | 
|  | { | 
|  | "LD_IND byte frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x40), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { }, | 
|  | { {0x40, 0x42} }, | 
|  | .frag_data = { | 
|  | 0x42, 0x00, 0x00, 0x00, | 
|  | 0x43, 0x44, 0x00, 0x00, | 
|  | 0x21, 0x07, 0x19, 0x83, | 
|  | }, | 
|  | }, | 
|  | { | 
|  | "LD_IND halfword frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x40), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_H, 0x4), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { }, | 
|  | { {0x40, 0x4344} }, | 
|  | .frag_data = { | 
|  | 0x42, 0x00, 0x00, 0x00, | 
|  | 0x43, 0x44, 0x00, 0x00, | 
|  | 0x21, 0x07, 0x19, 0x83, | 
|  | }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x40), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, 0x8), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { }, | 
|  | { {0x40, 0x21071983} }, | 
|  | .frag_data = { | 
|  | 0x42, 0x00, 0x00, 0x00, | 
|  | 0x43, 0x44, 0x00, 0x00, | 
|  | 0x21, 0x07, 0x19, 0x83, | 
|  | }, | 
|  | }, | 
|  | { | 
|  | "LD_IND halfword mixed head/frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x40), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_H, -0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { [0x3e] = 0x25, [0x3f] = 0x05, }, | 
|  | { {0x40, 0x0519} }, | 
|  | .frag_data = { 0x19, 0x82 }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word mixed head/frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x40), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x2), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { [0x3e] = 0x25, [0x3f] = 0x05, }, | 
|  | { {0x40, 0x25051982} }, | 
|  | .frag_data = { 0x19, 0x82 }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS byte frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, 0x40), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { }, | 
|  | { {0x40, 0x42} }, | 
|  | .frag_data = { | 
|  | 0x42, 0x00, 0x00, 0x00, | 
|  | 0x43, 0x44, 0x00, 0x00, | 
|  | 0x21, 0x07, 0x19, 0x83, | 
|  | }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x44), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { }, | 
|  | { {0x40, 0x4344} }, | 
|  | .frag_data = { | 
|  | 0x42, 0x00, 0x00, 0x00, | 
|  | 0x43, 0x44, 0x00, 0x00, | 
|  | 0x21, 0x07, 0x19, 0x83, | 
|  | }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x48), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { }, | 
|  | { {0x40, 0x21071983} }, | 
|  | .frag_data = { | 
|  | 0x42, 0x00, 0x00, 0x00, | 
|  | 0x43, 0x44, 0x00, 0x00, | 
|  | 0x21, 0x07, 0x19, 0x83, | 
|  | }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword mixed head/frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { [0x3e] = 0x25, [0x3f] = 0x05, }, | 
|  | { {0x40, 0x0519} }, | 
|  | .frag_data = { 0x19, 0x82 }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word mixed head/frag", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x3e), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_SKB_FRAG, | 
|  | { [0x3e] = 0x25, [0x3f] = 0x05, }, | 
|  | { {0x40, 0x25051982} }, | 
|  | .frag_data = { 0x19, 0x82 }, | 
|  | }, | 
|  | /* | 
|  | * LD_IND / LD_ABS on non fragmented SKBs | 
|  | */ | 
|  | { | 
|  | /* | 
|  | * this tests that the JIT/interpreter correctly resets X | 
|  | * before using it in an LD_IND instruction. | 
|  | */ | 
|  | "LD_IND byte default X", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x1] = 0x42 }, | 
|  | { {0x40, 0x42 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND byte positive offset", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x82 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND byte negative offset", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, -0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x05 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND byte positive offset, all ff", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0xff, [0x3d] = 0xff,  [0x3e] = 0xff, [0x3f] = 0xff }, | 
|  | { {0x40, 0xff } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND byte positive offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_IND byte negative offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, -0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND byte negative offset, multiple calls", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3b), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, SKF_LL_OFF + 1), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, SKF_LL_OFF + 2), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, SKF_LL_OFF + 3), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_B, SKF_LL_OFF + 4), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x82 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_IND halfword positive offset", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x20), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_H, 0x2), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | }, | 
|  | { {0x40, 0xdd88 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND halfword negative offset", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x20), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_H, -0x2), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | }, | 
|  | { {0x40, 0xbb66 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND halfword unaligned", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x20), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_H, -0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | }, | 
|  | { {0x40, 0x66cc } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND halfword positive offset, all ff", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3d), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_H, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0xff, [0x3d] = 0xff,  [0x3e] = 0xff, [0x3f] = 0xff }, | 
|  | { {0x40, 0xffff } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND halfword positive offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_H, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_IND halfword negative offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_H, -0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word positive offset", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x20), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, 0x4), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0xee99ffaa } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word negative offset", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x20), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x4), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0xaa55bb66 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word unaligned (addr & 3 == 2)", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x20), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x2), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0xbb66cc77 } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word unaligned (addr & 3 == 1)", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x20), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x3), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0x55bb66cc } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word unaligned (addr & 3 == 3)", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x20), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0x66cc77dd } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word positive offset, all ff", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3b), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0xff, [0x3d] = 0xff,  [0x3e] = 0xff, [0x3f] = 0xff }, | 
|  | { {0x40, 0xffffffff } }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word positive offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_IND word negative offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LDX | BPF_IMM, 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_IND | BPF_W, -0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS byte", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, 0x20), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0xcc } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS byte positive offset, all ff", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0xff, [0x3d] = 0xff,  [0x3e] = 0xff, [0x3f] = 0xff }, | 
|  | { {0x40, 0xff } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS byte positive offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS byte negative offset, out of bounds load", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, -1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_EXPECTED_FAIL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "LD_ABS byte negative offset, in bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x82 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS byte negative offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS byte negative offset, multiple calls", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3c), | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3d), | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3e), | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_B, SKF_LL_OFF + 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x82 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x22), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0xdd88 } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword unaligned", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x25), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0x99ff } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword positive offset, all ff", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x3e), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0xff, [0x3d] = 0xff,  [0x3e] = 0xff, [0x3f] = 0xff }, | 
|  | { {0x40, 0xffff } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword positive offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword negative offset, out of bounds load", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, -1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_EXPECTED_FAIL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword negative offset, in bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, SKF_LL_OFF + 0x3e), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x1982 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS halfword negative offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_H, SKF_LL_OFF + 0x3e), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x1c), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0xaa55bb66 } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word unaligned (addr & 3 == 2)", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x22), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0xdd88ee99 } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word unaligned (addr & 3 == 1)", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x21), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0x77dd88ee } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word unaligned (addr & 3 == 3)", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x23), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { | 
|  | [0x1c] = 0xaa, [0x1d] = 0x55, | 
|  | [0x1e] = 0xbb, [0x1f] = 0x66, | 
|  | [0x20] = 0xcc, [0x21] = 0x77, | 
|  | [0x22] = 0xdd, [0x23] = 0x88, | 
|  | [0x24] = 0xee, [0x25] = 0x99, | 
|  | [0x26] = 0xff, [0x27] = 0xaa, | 
|  | }, | 
|  | { {0x40, 0x88ee99ff } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word positive offset, all ff", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x3c), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0xff, [0x3d] = 0xff,  [0x3e] = 0xff, [0x3f] = 0xff }, | 
|  | { {0x40, 0xffffffff } }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word positive offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word negative offset, out of bounds load", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, -1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_EXPECTED_FAIL, | 
|  | .expected_errcode = -EINVAL, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word negative offset, in bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, SKF_LL_OFF + 0x3c), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x25051982 }, }, | 
|  | }, | 
|  | { | 
|  | "LD_ABS word negative offset, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_ABS | BPF_W, SKF_LL_OFF + 0x3c), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x3f, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LDX_MSH standalone, preserved A", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3c), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0xffeebbaa }, }, | 
|  | }, | 
|  | { | 
|  | "LDX_MSH standalone, preserved A 2", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0x175e9d63), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3c), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3d), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3e), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3f), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x175e9d63 }, }, | 
|  | }, | 
|  | { | 
|  | "LDX_MSH standalone, test result 1", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3c), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x14 }, }, | 
|  | }, | 
|  | { | 
|  | "LDX_MSH standalone, test result 2", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x3e), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x24 }, }, | 
|  | }, | 
|  | { | 
|  | "LDX_MSH standalone, negative offset", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, -1), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0 }, }, | 
|  | }, | 
|  | { | 
|  | "LDX_MSH standalone, negative offset 2", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, SKF_LL_OFF + 0x3e), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0x24 }, }, | 
|  | }, | 
|  | { | 
|  | "LDX_MSH standalone, out of bounds", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffeebbaa), | 
|  | BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0x40), | 
|  | BPF_STMT(BPF_MISC | BPF_TXA, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC, | 
|  | { [0x3c] = 0x25, [0x3d] = 0x05,  [0x3e] = 0x19, [0x3f] = 0x82 }, | 
|  | { {0x40, 0 }, }, | 
|  | }, | 
|  | /* | 
|  | * verify that the interpreter or JIT correctly sets A and X | 
|  | * to 0. | 
|  | */ | 
|  | { | 
|  | "ADD default X", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = 0x42 | 
|  | * A = A + X | 
|  | * ret A | 
|  | */ | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0x42), | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x42 } }, | 
|  | }, | 
|  | { | 
|  | "ADD default A", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = A + 0x42 | 
|  | * ret A | 
|  | */ | 
|  | BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, 0x42), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x42 } }, | 
|  | }, | 
|  | { | 
|  | "SUB default X", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = 0x66 | 
|  | * A = A - X | 
|  | * ret A | 
|  | */ | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0x66), | 
|  | BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x66 } }, | 
|  | }, | 
|  | { | 
|  | "SUB default A", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = A - -0x66 | 
|  | * ret A | 
|  | */ | 
|  | BPF_STMT(BPF_ALU | BPF_SUB | BPF_K, -0x66), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x66 } }, | 
|  | }, | 
|  | { | 
|  | "MUL default X", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = 0x42 | 
|  | * A = A * X | 
|  | * ret A | 
|  | */ | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0x42), | 
|  | BPF_STMT(BPF_ALU | BPF_MUL | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x0 } }, | 
|  | }, | 
|  | { | 
|  | "MUL default A", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = A * 0x66 | 
|  | * ret A | 
|  | */ | 
|  | BPF_STMT(BPF_ALU | BPF_MUL | BPF_K, 0x66), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x0 } }, | 
|  | }, | 
|  | { | 
|  | "DIV default X", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = 0x42 | 
|  | * A = A / X ; this halt the filter execution if X is 0 | 
|  | * ret 0x42 | 
|  | */ | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0x42), | 
|  | BPF_STMT(BPF_ALU | BPF_DIV | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0x42), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x0 } }, | 
|  | }, | 
|  | { | 
|  | "DIV default A", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = A / 1 | 
|  | * ret A | 
|  | */ | 
|  | BPF_STMT(BPF_ALU | BPF_DIV | BPF_K, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x0 } }, | 
|  | }, | 
|  | { | 
|  | "MOD default X", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = 0x42 | 
|  | * A = A mod X ; this halt the filter execution if X is 0 | 
|  | * ret 0x42 | 
|  | */ | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0x42), | 
|  | BPF_STMT(BPF_ALU | BPF_MOD | BPF_X, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0x42), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x0 } }, | 
|  | }, | 
|  | { | 
|  | "MOD default A", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = A mod 1 | 
|  | * ret A | 
|  | */ | 
|  | BPF_STMT(BPF_ALU | BPF_MOD | BPF_K, 0x1), | 
|  | BPF_STMT(BPF_RET | BPF_A, 0x0), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP EQ default A", | 
|  | .u.insns = { | 
|  | /* | 
|  | * cmp A, 0x0, 0, 1 | 
|  | * ret 0x42 | 
|  | * ret 0x66 | 
|  | */ | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0x0, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0x42), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0x66), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x42 } }, | 
|  | }, | 
|  | { | 
|  | "JMP EQ default X", | 
|  | .u.insns = { | 
|  | /* | 
|  | * A = 0x0 | 
|  | * cmp A, X, 0, 1 | 
|  | * ret 0x42 | 
|  | * ret 0x66 | 
|  | */ | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0x0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_X, 0x0, 0, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0x42), | 
|  | BPF_STMT(BPF_RET | BPF_K, 0x66), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { {0x1, 0x42 } }, | 
|  | }, | 
|  | /* Checking interpreter vs JIT wrt signed extended imms. */ | 
|  | { | 
|  | "JNE signed compare, test 1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xfefbbc12), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0xffff0000), | 
|  | BPF_MOV64_REG(R2, R1), | 
|  | BPF_ALU64_REG(BPF_AND, R2, R3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JNE, R2, -17104896, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JNE signed compare, test 2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xfefbbc12), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0xffff0000), | 
|  | BPF_MOV64_REG(R2, R1), | 
|  | BPF_ALU64_REG(BPF_AND, R2, R3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JNE, R2, 0xfefb0000, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JNE signed compare, test 3", | 
|  | .u.insns_int = { | 
|  | BPF_ALU32_IMM(BPF_MOV, R1, 0xfefbbc12), | 
|  | BPF_ALU32_IMM(BPF_MOV, R3, 0xffff0000), | 
|  | BPF_ALU32_IMM(BPF_MOV, R4, 0xfefb0000), | 
|  | BPF_MOV64_REG(R2, R1), | 
|  | BPF_ALU64_REG(BPF_AND, R2, R3), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JNE, R2, R4, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "JNE signed compare, test 4", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, -17104896), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JNE, R1, -17104896, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "JNE signed compare, test 5", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0xfefb0000), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JNE, R1, 0xfefb0000, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JNE signed compare, test 6", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x7efb0000), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JNE, R1, 0x7efb0000, 1), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, 2), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | { | 
|  | "JNE signed compare, test 7", | 
|  | .u.insns = { | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xffff0000), | 
|  | BPF_STMT(BPF_MISC | BPF_TAX, 0), | 
|  | BPF_STMT(BPF_LD | BPF_IMM, 0xfefbbc12), | 
|  | BPF_STMT(BPF_ALU | BPF_AND | BPF_X, 0), | 
|  | BPF_JUMP(BPF_JMP | BPF_JEQ | BPF_K, 0xfefb0000, 1, 0), | 
|  | BPF_STMT(BPF_RET | BPF_K, 1), | 
|  | BPF_STMT(BPF_RET | BPF_K, 2), | 
|  | }, | 
|  | CLASSIC | FLAG_NO_DATA, | 
|  | {}, | 
|  | { { 0, 2 } }, | 
|  | }, | 
|  | /* BPF_LDX_MEM with operand aliasing */ | 
|  | { | 
|  | "LDX_MEM_B: operand register aliasing", | 
|  | .u.insns_int = { | 
|  | BPF_ST_MEM(BPF_B, R10, -8, 123), | 
|  | BPF_MOV64_REG(R0, R10), | 
|  | BPF_LDX_MEM(BPF_B, R0, R0, -8), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 123 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "LDX_MEM_H: operand register aliasing", | 
|  | .u.insns_int = { | 
|  | BPF_ST_MEM(BPF_H, R10, -8, 12345), | 
|  | BPF_MOV64_REG(R0, R10), | 
|  | BPF_LDX_MEM(BPF_H, R0, R0, -8), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 12345 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "LDX_MEM_W: operand register aliasing", | 
|  | .u.insns_int = { | 
|  | BPF_ST_MEM(BPF_W, R10, -8, 123456789), | 
|  | BPF_MOV64_REG(R0, R10), | 
|  | BPF_LDX_MEM(BPF_W, R0, R0, -8), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 123456789 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "LDX_MEM_DW: operand register aliasing", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x123456789abcdefULL), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | BPF_MOV64_REG(R0, R10), | 
|  | BPF_LDX_MEM(BPF_DW, R0, R0, -8), | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1), | 
|  | BPF_MOV64_REG(R1, R0), | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32), | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | /* | 
|  | * Register (non-)clobbering tests for the case where a JIT implements | 
|  | * complex ALU or ATOMIC operations via function calls. If so, the | 
|  | * function call must be transparent to the eBPF registers. The JIT | 
|  | * must therefore save and restore relevant registers across the call. | 
|  | * The following tests check that the eBPF registers retain their | 
|  | * values after such an operation. Mainly intended for complex ALU | 
|  | * and atomic operation, but we run it for all. You never know... | 
|  | * | 
|  | * Note that each operations should be tested twice with different | 
|  | * destinations, to check preservation for all registers. | 
|  | */ | 
|  | #define BPF_TEST_CLOBBER_ALU(alu, op, dst, src)			\ | 
|  | {							\ | 
|  | #alu "_" #op " to " #dst ": no clobbering",	\ | 
|  | .u.insns_int = {				\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, R0),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, R1),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, R2),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R3, R3),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R4, R4),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R5, R5),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R6, R6),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R7, R7),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R8, R8),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R9, R9),		\ | 
|  | BPF_##alu(BPF_ ##op, dst, src),		\ | 
|  | BPF_ALU32_IMM(BPF_MOV, dst, dst),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R0, R0, 10),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R1, R1, 9),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R2, R2, 8),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R3, R3, 7),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R4, R4, 6),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R5, R5, 5),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R6, R6, 4),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R7, R7, 3),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R8, R8, 2),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R9, R9, 1),	\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1),		\ | 
|  | BPF_EXIT_INSN(),			\ | 
|  | },						\ | 
|  | INTERNAL,					\ | 
|  | { },						\ | 
|  | { { 0, 1 } }					\ | 
|  | } | 
|  | /* ALU64 operations, register clobbering */ | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, AND, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, AND, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, OR, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, OR, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, XOR, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, XOR, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, LSH, R8, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, LSH, R9, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, RSH, R8, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, RSH, R9, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, ARSH, R8, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, ARSH, R9, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, ADD, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, ADD, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, SUB, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, SUB, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, MUL, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, MUL, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, DIV, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, DIV, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, MOD, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_IMM, MOD, R9, 123456789), | 
|  | /* ALU32 immediate operations, register clobbering */ | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, AND, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, AND, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, OR, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, OR, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, XOR, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, XOR, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, LSH, R8, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, LSH, R9, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, RSH, R8, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, RSH, R9, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, ARSH, R8, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, ARSH, R9, 12), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, ADD, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, ADD, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, SUB, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, SUB, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, MUL, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, MUL, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, DIV, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, DIV, R9, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, MOD, R8, 123456789), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_IMM, MOD, R9, 123456789), | 
|  | /* ALU64 register operations, register clobbering */ | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, AND, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, AND, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, OR, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, OR, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, XOR, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, XOR, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, LSH, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, LSH, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, RSH, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, RSH, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, ARSH, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, ARSH, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, ADD, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, ADD, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, SUB, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, SUB, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, MUL, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, MUL, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, DIV, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, DIV, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, MOD, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU64_REG, MOD, R9, R1), | 
|  | /* ALU32 register operations, register clobbering */ | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, AND, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, AND, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, OR, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, OR, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, XOR, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, XOR, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, LSH, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, LSH, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, RSH, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, RSH, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, ARSH, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, ARSH, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, ADD, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, ADD, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, SUB, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, SUB, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, MUL, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, MUL, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, DIV, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, DIV, R9, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, MOD, R8, R1), | 
|  | BPF_TEST_CLOBBER_ALU(ALU32_REG, MOD, R9, R1), | 
|  | #undef BPF_TEST_CLOBBER_ALU | 
|  | #define BPF_TEST_CLOBBER_ATOMIC(width, op)			\ | 
|  | {							\ | 
|  | "Atomic_" #width " " #op ": no clobbering",	\ | 
|  | .u.insns_int = {				\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, 1),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 2),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R3, 3),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R4, 4),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R5, 5),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R6, 6),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R7, 7),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R8, 8),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R9, 9),		\ | 
|  | BPF_ST_MEM(width, R10, -8,		\ | 
|  | (op) == BPF_CMPXCHG ? 0 :	\ | 
|  | (op) & BPF_FETCH ? 1 : 0),	\ | 
|  | BPF_ATOMIC_OP(width, op, R10, R1, -8),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R0, 0, 10),	\ | 
|  | BPF_JMP_IMM(BPF_JNE, R1, 1, 9),		\ | 
|  | BPF_JMP_IMM(BPF_JNE, R2, 2, 8),		\ | 
|  | BPF_JMP_IMM(BPF_JNE, R3, 3, 7),		\ | 
|  | BPF_JMP_IMM(BPF_JNE, R4, 4, 6),		\ | 
|  | BPF_JMP_IMM(BPF_JNE, R5, 5, 5),		\ | 
|  | BPF_JMP_IMM(BPF_JNE, R6, 6, 4),		\ | 
|  | BPF_JMP_IMM(BPF_JNE, R7, 7, 3),		\ | 
|  | BPF_JMP_IMM(BPF_JNE, R8, 8, 2),		\ | 
|  | BPF_JMP_IMM(BPF_JNE, R9, 9, 1),		\ | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1),		\ | 
|  | BPF_EXIT_INSN(),			\ | 
|  | },						\ | 
|  | INTERNAL,					\ | 
|  | { },						\ | 
|  | { { 0, 1 } },					\ | 
|  | .stack_depth = 8,				\ | 
|  | } | 
|  | /* 64-bit atomic operations, register clobbering */ | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_ADD), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_AND), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_OR), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_XOR), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_ADD | BPF_FETCH), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_AND | BPF_FETCH), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_OR | BPF_FETCH), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_XOR | BPF_FETCH), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_XCHG), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_DW, BPF_CMPXCHG), | 
|  | /* 32-bit atomic operations, register clobbering */ | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_ADD), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_AND), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_OR), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_XOR), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_ADD | BPF_FETCH), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_AND | BPF_FETCH), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_OR | BPF_FETCH), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_XOR | BPF_FETCH), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_XCHG), | 
|  | BPF_TEST_CLOBBER_ATOMIC(BPF_W, BPF_CMPXCHG), | 
|  | #undef BPF_TEST_CLOBBER_ATOMIC | 
|  | /* Checking that ALU32 src is not zero extended in place */ | 
|  | #define BPF_ALU32_SRC_ZEXT(op)					\ | 
|  | {							\ | 
|  | "ALU32_" #op "_X: src preserved in zext",	\ | 
|  | .u.insns_int = {				\ | 
|  | BPF_LD_IMM64(R1, 0x0123456789acbdefULL),\ | 
|  | BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),\ | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1),		\ | 
|  | BPF_ALU32_REG(BPF_##op, R2, R1),	\ | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1),		\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),		\ | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32),		\ | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1),		\ | 
|  | BPF_EXIT_INSN(),			\ | 
|  | },						\ | 
|  | INTERNAL,					\ | 
|  | { },						\ | 
|  | { { 0, 0 } },					\ | 
|  | } | 
|  | BPF_ALU32_SRC_ZEXT(MOV), | 
|  | BPF_ALU32_SRC_ZEXT(AND), | 
|  | BPF_ALU32_SRC_ZEXT(OR), | 
|  | BPF_ALU32_SRC_ZEXT(XOR), | 
|  | BPF_ALU32_SRC_ZEXT(ADD), | 
|  | BPF_ALU32_SRC_ZEXT(SUB), | 
|  | BPF_ALU32_SRC_ZEXT(MUL), | 
|  | BPF_ALU32_SRC_ZEXT(DIV), | 
|  | BPF_ALU32_SRC_ZEXT(MOD), | 
|  | #undef BPF_ALU32_SRC_ZEXT | 
|  | /* Checking that ATOMIC32 src is not zero extended in place */ | 
|  | #define BPF_ATOMIC32_SRC_ZEXT(op)					\ | 
|  | {								\ | 
|  | "ATOMIC_W_" #op ": src preserved in zext",		\ | 
|  | .u.insns_int = {					\ | 
|  | BPF_LD_IMM64(R0, 0x0123456789acbdefULL),	\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),			\ | 
|  | BPF_ST_MEM(BPF_W, R10, -4, 0),			\ | 
|  | BPF_ATOMIC_OP(BPF_W, BPF_##op, R10, R1, -4),	\ | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1),			\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),			\ | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32),			\ | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1),			\ | 
|  | BPF_EXIT_INSN(),				\ | 
|  | },							\ | 
|  | INTERNAL,						\ | 
|  | { },							\ | 
|  | { { 0, 0 } },						\ | 
|  | .stack_depth = 8,					\ | 
|  | } | 
|  | BPF_ATOMIC32_SRC_ZEXT(ADD), | 
|  | BPF_ATOMIC32_SRC_ZEXT(AND), | 
|  | BPF_ATOMIC32_SRC_ZEXT(OR), | 
|  | BPF_ATOMIC32_SRC_ZEXT(XOR), | 
|  | #undef BPF_ATOMIC32_SRC_ZEXT | 
|  | /* Checking that CMPXCHG32 src is not zero extended in place */ | 
|  | { | 
|  | "ATOMIC_W_CMPXCHG: src preserved in zext", | 
|  | .u.insns_int = { | 
|  | BPF_LD_IMM64(R1, 0x0123456789acbdefULL), | 
|  | BPF_ALU64_REG(BPF_MOV, R2, R1), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, 0), | 
|  | BPF_ST_MEM(BPF_W, R10, -4, 0), | 
|  | BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, R10, R1, -4), | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R2), | 
|  | BPF_ALU64_REG(BPF_MOV, R2, R1), | 
|  | BPF_ALU64_IMM(BPF_RSH, R2, 32), | 
|  | BPF_ALU64_REG(BPF_OR, R1, R2), | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | /* Checking that JMP32 immediate src is not zero extended in place */ | 
|  | #define BPF_JMP32_IMM_ZEXT(op)					\ | 
|  | {							\ | 
|  | "JMP32_" #op "_K: operand preserved in zext",	\ | 
|  | .u.insns_int = {				\ | 
|  | BPF_LD_IMM64(R0, 0x0123456789acbdefULL),\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),		\ | 
|  | BPF_JMP32_IMM(BPF_##op, R0, 1234, 1),	\ | 
|  | BPF_JMP_A(0), /* Nop */			\ | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R1),		\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),		\ | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32),		\ | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1),		\ | 
|  | BPF_EXIT_INSN(),			\ | 
|  | },						\ | 
|  | INTERNAL,					\ | 
|  | { },						\ | 
|  | { { 0, 0 } },					\ | 
|  | } | 
|  | BPF_JMP32_IMM_ZEXT(JEQ), | 
|  | BPF_JMP32_IMM_ZEXT(JNE), | 
|  | BPF_JMP32_IMM_ZEXT(JSET), | 
|  | BPF_JMP32_IMM_ZEXT(JGT), | 
|  | BPF_JMP32_IMM_ZEXT(JGE), | 
|  | BPF_JMP32_IMM_ZEXT(JLT), | 
|  | BPF_JMP32_IMM_ZEXT(JLE), | 
|  | BPF_JMP32_IMM_ZEXT(JSGT), | 
|  | BPF_JMP32_IMM_ZEXT(JSGE), | 
|  | BPF_JMP32_IMM_ZEXT(JSLT), | 
|  | BPF_JMP32_IMM_ZEXT(JSLE), | 
|  | #undef BPF_JMP2_IMM_ZEXT | 
|  | /* Checking that JMP32 dst & src are not zero extended in place */ | 
|  | #define BPF_JMP32_REG_ZEXT(op)					\ | 
|  | {							\ | 
|  | "JMP32_" #op "_X: operands preserved in zext",	\ | 
|  | .u.insns_int = {				\ | 
|  | BPF_LD_IMM64(R0, 0x0123456789acbdefULL),\ | 
|  | BPF_LD_IMM64(R1, 0xfedcba9876543210ULL),\ | 
|  | BPF_ALU64_REG(BPF_MOV, R2, R0),		\ | 
|  | BPF_ALU64_REG(BPF_MOV, R3, R1),		\ | 
|  | BPF_JMP32_IMM(BPF_##op, R0, R1, 1),	\ | 
|  | BPF_JMP_A(0), /* Nop */			\ | 
|  | BPF_ALU64_REG(BPF_SUB, R0, R2),		\ | 
|  | BPF_ALU64_REG(BPF_SUB, R1, R3),		\ | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1),		\ | 
|  | BPF_ALU64_REG(BPF_MOV, R1, R0),		\ | 
|  | BPF_ALU64_IMM(BPF_RSH, R1, 32),		\ | 
|  | BPF_ALU64_REG(BPF_OR, R0, R1),		\ | 
|  | BPF_EXIT_INSN(),			\ | 
|  | },						\ | 
|  | INTERNAL,					\ | 
|  | { },						\ | 
|  | { { 0, 0 } },					\ | 
|  | } | 
|  | BPF_JMP32_REG_ZEXT(JEQ), | 
|  | BPF_JMP32_REG_ZEXT(JNE), | 
|  | BPF_JMP32_REG_ZEXT(JSET), | 
|  | BPF_JMP32_REG_ZEXT(JGT), | 
|  | BPF_JMP32_REG_ZEXT(JGE), | 
|  | BPF_JMP32_REG_ZEXT(JLT), | 
|  | BPF_JMP32_REG_ZEXT(JLE), | 
|  | BPF_JMP32_REG_ZEXT(JSGT), | 
|  | BPF_JMP32_REG_ZEXT(JSGE), | 
|  | BPF_JMP32_REG_ZEXT(JSLT), | 
|  | BPF_JMP32_REG_ZEXT(JSLE), | 
|  | #undef BPF_JMP2_REG_ZEXT | 
|  | /* ALU64 K register combinations */ | 
|  | { | 
|  | "ALU64_MOV_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mov_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_and_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_or_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_xor_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_lsh_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_rsh_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_arsh_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_add_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_sub_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mul_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_div_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mod_imm_regs, | 
|  | }, | 
|  | /* ALU32 K registers */ | 
|  | { | 
|  | "ALU32_MOV_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mov_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_AND_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_and_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_OR_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_or_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_XOR_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_xor_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_LSH_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_lsh_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_RSH_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_rsh_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_ARSH_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_arsh_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_ADD_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_add_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_SUB_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_sub_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_MUL_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mul_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_DIV_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_div_imm_regs, | 
|  | }, | 
|  | { | 
|  | "ALU32_MOD_K: registers", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mod_imm_regs, | 
|  | }, | 
|  | /* ALU64 X register combinations */ | 
|  | { | 
|  | "ALU64_MOV_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mov_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_and_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_or_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_xor_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_lsh_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_rsh_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_arsh_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_add_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_sub_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mul_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_div_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mod_reg_pairs, | 
|  | }, | 
|  | /* ALU32 X register combinations */ | 
|  | { | 
|  | "ALU32_MOV_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mov_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_AND_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_and_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_OR_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_or_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_XOR_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_xor_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_LSH_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_lsh_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_RSH_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_rsh_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_ARSH_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_arsh_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_ADD_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_add_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_SUB_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_sub_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_MUL_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mul_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_DIV_X: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_div_reg_pairs, | 
|  | }, | 
|  | { | 
|  | "ALU32_MOD_X register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mod_reg_pairs, | 
|  | }, | 
|  | /* Exhaustive test of ALU64 shift operations */ | 
|  | { | 
|  | "ALU64_LSH_K: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_lsh_imm, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_K: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_rsh_imm, | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_K: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_arsh_imm, | 
|  | }, | 
|  | { | 
|  | "ALU64_LSH_X: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_lsh_reg, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_rsh_reg, | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_arsh_reg, | 
|  | }, | 
|  | /* Exhaustive test of ALU32 shift operations */ | 
|  | { | 
|  | "ALU32_LSH_K: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_lsh_imm, | 
|  | }, | 
|  | { | 
|  | "ALU32_RSH_K: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_rsh_imm, | 
|  | }, | 
|  | { | 
|  | "ALU32_ARSH_K: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_arsh_imm, | 
|  | }, | 
|  | { | 
|  | "ALU32_LSH_X: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_lsh_reg, | 
|  | }, | 
|  | { | 
|  | "ALU32_RSH_X: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_rsh_reg, | 
|  | }, | 
|  | { | 
|  | "ALU32_ARSH_X: all shift values", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_arsh_reg, | 
|  | }, | 
|  | /* | 
|  | * Exhaustive test of ALU64 shift operations when | 
|  | * source and destination register are the same. | 
|  | */ | 
|  | { | 
|  | "ALU64_LSH_X: all shift values with the same register", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_lsh_same_reg, | 
|  | }, | 
|  | { | 
|  | "ALU64_RSH_X: all shift values with the same register", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_rsh_same_reg, | 
|  | }, | 
|  | { | 
|  | "ALU64_ARSH_X: all shift values with the same register", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_arsh_same_reg, | 
|  | }, | 
|  | /* | 
|  | * Exhaustive test of ALU32 shift operations when | 
|  | * source and destination register are the same. | 
|  | */ | 
|  | { | 
|  | "ALU32_LSH_X: all shift values with the same register", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_lsh_same_reg, | 
|  | }, | 
|  | { | 
|  | "ALU32_RSH_X: all shift values with the same register", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_rsh_same_reg, | 
|  | }, | 
|  | { | 
|  | "ALU32_ARSH_X: all shift values with the same register", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_arsh_same_reg, | 
|  | }, | 
|  | /* ALU64 immediate magnitudes */ | 
|  | { | 
|  | "ALU64_MOV_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mov_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_and_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_or_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_xor_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_add_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_sub_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mul_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_div_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mod_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* ALU32 immediate magnitudes */ | 
|  | { | 
|  | "ALU32_MOV_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mov_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_AND_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_and_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_OR_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_or_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_XOR_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_xor_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_ADD_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_add_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_SUB_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_sub_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_MUL_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mul_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_DIV_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_div_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_MOD_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mod_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* ALU64 register magnitudes */ | 
|  | { | 
|  | "ALU64_MOV_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mov_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_AND_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_and_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_OR_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_or_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_XOR_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_xor_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_ADD_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_add_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_SUB_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_sub_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_MUL_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mul_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_DIV_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_div_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU64_MOD_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu64_mod_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* ALU32 register magnitudes */ | 
|  | { | 
|  | "ALU32_MOV_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mov_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_AND_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_and_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_OR_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_or_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_XOR_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_xor_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_ADD_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_add_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_SUB_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_sub_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_MUL_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mul_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_DIV_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_div_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ALU32_MOD_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_alu32_mod_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* LD_IMM64 immediate magnitudes and byte patterns */ | 
|  | { | 
|  | "LD_IMM64: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_ld_imm64_magn, | 
|  | }, | 
|  | { | 
|  | "LD_IMM64: checker byte patterns", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_ld_imm64_checker, | 
|  | }, | 
|  | { | 
|  | "LD_IMM64: random positive and zero byte patterns", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_ld_imm64_pos_zero, | 
|  | }, | 
|  | { | 
|  | "LD_IMM64: random negative and zero byte patterns", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_ld_imm64_neg_zero, | 
|  | }, | 
|  | { | 
|  | "LD_IMM64: random positive and negative byte patterns", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_ld_imm64_pos_neg, | 
|  | }, | 
|  | /* 64-bit ATOMIC register combinations */ | 
|  | { | 
|  | "ATOMIC_DW_ADD: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_add_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_AND: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_and_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_OR: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_or_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_XOR: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_xor_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_ADD_FETCH: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_add_fetch_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_AND_FETCH: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_and_fetch_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_OR_FETCH: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_or_fetch_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_XOR_FETCH: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_xor_fetch_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_XCHG: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_xchg_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_CMPXCHG: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_cmpxchg_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | /* 32-bit ATOMIC register combinations */ | 
|  | { | 
|  | "ATOMIC_W_ADD: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_add_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_AND: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_and_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_OR: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_or_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_XOR: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_xor_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_ADD_FETCH: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_add_fetch_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_AND_FETCH: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_and_fetch_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_OR_FETCH: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_or_fetch_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_XOR_FETCH: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_xor_fetch_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_XCHG: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_xchg_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_CMPXCHG: register combinations", | 
|  | { }, | 
|  | INTERNAL, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_cmpxchg_reg_pairs, | 
|  | .stack_depth = 8, | 
|  | }, | 
|  | /* 64-bit ATOMIC magnitudes */ | 
|  | { | 
|  | "ATOMIC_DW_ADD: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_add, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_AND: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_and, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_OR: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_or, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_XOR: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_xor, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_ADD_FETCH: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_add_fetch, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_AND_FETCH: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_and_fetch, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_OR_FETCH: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_or_fetch, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_XOR_FETCH: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_xor_fetch, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_XCHG: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic64_xchg, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_DW_CMPXCHG: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_cmpxchg64, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* 32-bit atomic magnitudes */ | 
|  | { | 
|  | "ATOMIC_W_ADD: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_add, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_AND: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_and, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_OR: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_or, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_XOR: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_xor, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_ADD_FETCH: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_add_fetch, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_AND_FETCH: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_and_fetch, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_OR_FETCH: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_or_fetch, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_XOR_FETCH: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_xor_fetch, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_XCHG: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_atomic32_xchg, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "ATOMIC_W_CMPXCHG: all operand magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_cmpxchg32, | 
|  | .stack_depth = 8, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* JMP immediate magnitudes */ | 
|  | { | 
|  | "JMP_JSET_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jset_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JEQ_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jeq_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JNE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jne_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JGT_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jgt_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JGE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jge_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JLT_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jlt_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jle_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGT_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jsgt_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jsge_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLT_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jslt_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jsle_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* JMP register magnitudes */ | 
|  | { | 
|  | "JMP_JSET_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jset_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JEQ_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jeq_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JNE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jne_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JGT_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jgt_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JGE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jge_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JLT_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jlt_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jle_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGT_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jsgt_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jsge_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLT_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jslt_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp_jsle_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* JMP32 immediate magnitudes */ | 
|  | { | 
|  | "JMP32_JSET_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jset_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JEQ_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jeq_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JNE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jne_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JGT_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jgt_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JGE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jge_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JLT_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jlt_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JLE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jle_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSGT_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jsgt_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSGE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jsge_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSLT_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jslt_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSLE_K: all immediate value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jsle_imm, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* JMP32 register magnitudes */ | 
|  | { | 
|  | "JMP32_JSET_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jset_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JEQ_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jeq_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JNE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jne_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JGT_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jgt_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JGE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jge_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JLT_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jlt_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JLE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jle_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSGT_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jsgt_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSGE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jsge_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSLT_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jslt_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSLE_X: all register value magnitudes", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_jmp32_jsle_reg, | 
|  | .nr_testruns = NR_PATTERN_RUNS, | 
|  | }, | 
|  | /* Conditional jumps with constant decision */ | 
|  | { | 
|  | "JMP_JSET_K: imm = 0 -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JSET, R1, 0, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLT_K: imm = 0 -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JLT, R1, 0, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGE_K: imm = 0 -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JGE, R1, 0, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGT_K: imm = 0xffffffff -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JGT, R1, U32_MAX, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_K: imm = 0xffffffff -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_IMM(BPF_JLE, R1, U32_MAX, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSGT_K: imm = 0x7fffffff -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP32_IMM(BPF_JSGT, R1, S32_MAX, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSGE_K: imm = -0x80000000 -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP32_IMM(BPF_JSGE, R1, S32_MIN, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSLT_K: imm = -0x80000000 -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP32_IMM(BPF_JSLT, R1, S32_MIN, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP32_JSLE_K: imm = 0x7fffffff -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP32_IMM(BPF_JSLE, R1, S32_MAX, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JEQ_X: dst = src -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JEQ, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGE_X: dst = src -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JGE, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLE_X: dst = src -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JLE, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGE_X: dst = src -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JSGE, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLE_X: dst = src -> always taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JSLE, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JNE_X: dst = src -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JNE, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JGT_X: dst = src -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JGT, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JLT_X: dst = src -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JLT, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSGT_X: dst = src -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JSGT, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "JMP_JSLT_X: dst = src -> never taken", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 1), | 
|  | BPF_JMP_REG(BPF_JSLT, R1, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | /* Short relative jumps */ | 
|  | { | 
|  | "Short relative jump: offset=0", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "Short relative jump: offset=1", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "Short relative jump: offset=2", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 0, 2), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "Short relative jump: offset=3", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 0, 3), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | { | 
|  | "Short relative jump: offset=4", | 
|  | .u.insns_int = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_JMP_IMM(BPF_JEQ, R0, 0, 4), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_ALU32_IMM(BPF_ADD, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | BPF_ALU32_IMM(BPF_MOV, R0, -1), | 
|  | }, | 
|  | INTERNAL | FLAG_NO_DATA | FLAG_VERIFIER_ZEXT, | 
|  | { }, | 
|  | { { 0, 0 } }, | 
|  | }, | 
|  | /* Conditional branch conversions */ | 
|  | { | 
|  | "Long conditional jump: taken at runtime (32 bits)", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_max_jmp_taken_32, | 
|  | }, | 
|  | { | 
|  | "Long conditional jump: not taken at runtime (32 bits)", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | .fill_helper = bpf_fill_max_jmp_not_taken_32, | 
|  | }, | 
|  | { | 
|  | "Long conditional jump: always taken, known at JIT time (32 bits)", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_max_jmp_always_taken_32, | 
|  | }, | 
|  | { | 
|  | "Long conditional jump: never taken, known at JIT time (32 bits)", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | .fill_helper = bpf_fill_max_jmp_never_taken_32, | 
|  | }, | 
|  | { | 
|  | "Long conditional jump: taken at runtime", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_max_jmp_taken, | 
|  | }, | 
|  | { | 
|  | "Long conditional jump: not taken at runtime", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | .fill_helper = bpf_fill_max_jmp_not_taken, | 
|  | }, | 
|  | { | 
|  | "Long conditional jump: always taken, known at JIT time", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 1 } }, | 
|  | .fill_helper = bpf_fill_max_jmp_always_taken, | 
|  | }, | 
|  | { | 
|  | "Long conditional jump: never taken, known at JIT time", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, 2 } }, | 
|  | .fill_helper = bpf_fill_max_jmp_never_taken, | 
|  | }, | 
|  | /* Staggered jump sequences, immediate */ | 
|  | { | 
|  | "Staggered jumps: JMP_JA", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_ja, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JEQ_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jeq_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JNE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jne_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSET_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jset_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JGT_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jgt_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JGE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jge_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JLT_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jlt_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JLE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jle_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSGT_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsgt_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSGE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsge_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSLT_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jslt_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSLE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsle_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | /* Staggered jump sequences, register */ | 
|  | { | 
|  | "Staggered jumps: JMP_JEQ_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jeq_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JNE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jne_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSET_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jset_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JGT_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jgt_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JGE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jge_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JLT_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jlt_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JLE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jle_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSGT_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsgt_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSGE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsge_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSLT_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jslt_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP_JSLE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsle_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | /* Staggered jump sequences, JMP32 immediate */ | 
|  | { | 
|  | "Staggered jumps: JMP32_JEQ_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jeq32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JNE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jne32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSET_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jset32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JGT_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jgt32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JGE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jge32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JLT_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jlt32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JLE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jle32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSGT_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsgt32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSGE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsge32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSLT_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jslt32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSLE_K", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsle32_imm, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | /* Staggered jump sequences, JMP32 register */ | 
|  | { | 
|  | "Staggered jumps: JMP32_JEQ_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jeq32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JNE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jne32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSET_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jset32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JGT_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jgt32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JGE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jge32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JLT_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jlt32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JLE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jle32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSGT_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsgt32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSGE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsge32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSLT_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jslt32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | { | 
|  | "Staggered jumps: JMP32_JSLE_X", | 
|  | { }, | 
|  | INTERNAL | FLAG_NO_DATA, | 
|  | { }, | 
|  | { { 0, MAX_STAGGERED_JMP_SIZE + 1 } }, | 
|  | .fill_helper = bpf_fill_staggered_jsle32_reg, | 
|  | .nr_testruns = NR_STAGGERED_JMP_RUNS, | 
|  | }, | 
|  | }; | 
|  |  | 
|  | static struct net_device dev; | 
|  |  | 
|  | static struct sk_buff *populate_skb(char *buf, int size) | 
|  | { | 
|  | struct sk_buff *skb; | 
|  |  | 
|  | if (size >= MAX_DATA) | 
|  | return NULL; | 
|  |  | 
|  | skb = alloc_skb(MAX_DATA, GFP_KERNEL); | 
|  | if (!skb) | 
|  | return NULL; | 
|  |  | 
|  | __skb_put_data(skb, buf, size); | 
|  |  | 
|  | /* Initialize a fake skb with test pattern. */ | 
|  | skb_reset_mac_header(skb); | 
|  | skb->protocol = htons(ETH_P_IP); | 
|  | skb->pkt_type = SKB_TYPE; | 
|  | skb->mark = SKB_MARK; | 
|  | skb->hash = SKB_HASH; | 
|  | skb->queue_mapping = SKB_QUEUE_MAP; | 
|  | skb->vlan_tci = SKB_VLAN_TCI; | 
|  | skb->vlan_proto = htons(ETH_P_IP); | 
|  | dev_net_set(&dev, &init_net); | 
|  | skb->dev = &dev; | 
|  | skb->dev->ifindex = SKB_DEV_IFINDEX; | 
|  | skb->dev->type = SKB_DEV_TYPE; | 
|  | skb_set_network_header(skb, min(size, ETH_HLEN)); | 
|  |  | 
|  | return skb; | 
|  | } | 
|  |  | 
|  | static void *generate_test_data(struct bpf_test *test, int sub) | 
|  | { | 
|  | struct sk_buff *skb; | 
|  | struct page *page; | 
|  |  | 
|  | if (test->aux & FLAG_NO_DATA) | 
|  | return NULL; | 
|  |  | 
|  | if (test->aux & FLAG_LARGE_MEM) | 
|  | return kmalloc(test->test[sub].data_size, GFP_KERNEL); | 
|  |  | 
|  | /* Test case expects an skb, so populate one. Various | 
|  | * subtests generate skbs of different sizes based on | 
|  | * the same data. | 
|  | */ | 
|  | skb = populate_skb(test->data, test->test[sub].data_size); | 
|  | if (!skb) | 
|  | return NULL; | 
|  |  | 
|  | if (test->aux & FLAG_SKB_FRAG) { | 
|  | /* | 
|  | * when the test requires a fragmented skb, add a | 
|  | * single fragment to the skb, filled with | 
|  | * test->frag_data. | 
|  | */ | 
|  | page = alloc_page(GFP_KERNEL); | 
|  | if (!page) | 
|  | goto err_kfree_skb; | 
|  |  | 
|  | memcpy(page_address(page), test->frag_data, MAX_DATA); | 
|  | skb_add_rx_frag(skb, 0, page, 0, MAX_DATA, MAX_DATA); | 
|  | } | 
|  |  | 
|  | return skb; | 
|  | err_kfree_skb: | 
|  | kfree_skb(skb); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static void release_test_data(const struct bpf_test *test, void *data) | 
|  | { | 
|  | if (test->aux & FLAG_NO_DATA) | 
|  | return; | 
|  |  | 
|  | if (test->aux & FLAG_LARGE_MEM) | 
|  | kfree(data); | 
|  | else | 
|  | kfree_skb(data); | 
|  | } | 
|  |  | 
|  | static int filter_length(int which) | 
|  | { | 
|  | struct sock_filter *fp; | 
|  | int len; | 
|  |  | 
|  | if (tests[which].fill_helper) | 
|  | return tests[which].u.ptr.len; | 
|  |  | 
|  | fp = tests[which].u.insns; | 
|  | for (len = MAX_INSNS - 1; len > 0; --len) | 
|  | if (fp[len].code != 0 || fp[len].k != 0) | 
|  | break; | 
|  |  | 
|  | return len + 1; | 
|  | } | 
|  |  | 
|  | static void *filter_pointer(int which) | 
|  | { | 
|  | if (tests[which].fill_helper) | 
|  | return tests[which].u.ptr.insns; | 
|  | else | 
|  | return tests[which].u.insns; | 
|  | } | 
|  |  | 
|  | static struct bpf_prog *generate_filter(int which, int *err) | 
|  | { | 
|  | __u8 test_type = tests[which].aux & TEST_TYPE_MASK; | 
|  | unsigned int flen = filter_length(which); | 
|  | void *fptr = filter_pointer(which); | 
|  | struct sock_fprog_kern fprog; | 
|  | struct bpf_prog *fp; | 
|  |  | 
|  | switch (test_type) { | 
|  | case CLASSIC: | 
|  | fprog.filter = fptr; | 
|  | fprog.len = flen; | 
|  |  | 
|  | *err = bpf_prog_create(&fp, &fprog); | 
|  | if (tests[which].aux & FLAG_EXPECTED_FAIL) { | 
|  | if (*err == tests[which].expected_errcode) { | 
|  | pr_cont("PASS\n"); | 
|  | /* Verifier rejected filter as expected. */ | 
|  | *err = 0; | 
|  | return NULL; | 
|  | } else { | 
|  | pr_cont("UNEXPECTED_PASS\n"); | 
|  | /* Verifier didn't reject the test that's | 
|  | * bad enough, just return! | 
|  | */ | 
|  | *err = -EINVAL; | 
|  | return NULL; | 
|  | } | 
|  | } | 
|  | if (*err) { | 
|  | pr_cont("FAIL to prog_create err=%d len=%d\n", | 
|  | *err, fprog.len); | 
|  | return NULL; | 
|  | } | 
|  | break; | 
|  |  | 
|  | case INTERNAL: | 
|  | fp = bpf_prog_alloc(bpf_prog_size(flen), 0); | 
|  | if (fp == NULL) { | 
|  | pr_cont("UNEXPECTED_FAIL no memory left\n"); | 
|  | *err = -ENOMEM; | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | fp->len = flen; | 
|  | /* Type doesn't really matter here as long as it's not unspec. */ | 
|  | fp->type = BPF_PROG_TYPE_SOCKET_FILTER; | 
|  | memcpy(fp->insnsi, fptr, fp->len * sizeof(struct bpf_insn)); | 
|  | fp->aux->stack_depth = tests[which].stack_depth; | 
|  | fp->aux->verifier_zext = !!(tests[which].aux & | 
|  | FLAG_VERIFIER_ZEXT); | 
|  |  | 
|  | /* We cannot error here as we don't need type compatibility | 
|  | * checks. | 
|  | */ | 
|  | fp = bpf_prog_select_runtime(fp, err); | 
|  | if (*err) { | 
|  | pr_cont("FAIL to select_runtime err=%d\n", *err); | 
|  | return NULL; | 
|  | } | 
|  | break; | 
|  | } | 
|  |  | 
|  | *err = 0; | 
|  | return fp; | 
|  | } | 
|  |  | 
|  | static void release_filter(struct bpf_prog *fp, int which) | 
|  | { | 
|  | __u8 test_type = tests[which].aux & TEST_TYPE_MASK; | 
|  |  | 
|  | switch (test_type) { | 
|  | case CLASSIC: | 
|  | bpf_prog_destroy(fp); | 
|  | break; | 
|  | case INTERNAL: | 
|  | bpf_prog_free(fp); | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | static int __run_one(const struct bpf_prog *fp, const void *data, | 
|  | int runs, u64 *duration) | 
|  | { | 
|  | u64 start, finish; | 
|  | int ret = 0, i; | 
|  |  | 
|  | migrate_disable(); | 
|  | start = ktime_get_ns(); | 
|  |  | 
|  | for (i = 0; i < runs; i++) | 
|  | ret = bpf_prog_run(fp, data); | 
|  |  | 
|  | finish = ktime_get_ns(); | 
|  | migrate_enable(); | 
|  |  | 
|  | *duration = finish - start; | 
|  | do_div(*duration, runs); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int run_one(const struct bpf_prog *fp, struct bpf_test *test) | 
|  | { | 
|  | int err_cnt = 0, i, runs = MAX_TESTRUNS; | 
|  |  | 
|  | if (test->nr_testruns) | 
|  | runs = min(test->nr_testruns, MAX_TESTRUNS); | 
|  |  | 
|  | for (i = 0; i < MAX_SUBTESTS; i++) { | 
|  | void *data; | 
|  | u64 duration; | 
|  | u32 ret; | 
|  |  | 
|  | /* | 
|  | * NOTE: Several sub-tests may be present, in which case | 
|  | * a zero {data_size, result} tuple indicates the end of | 
|  | * the sub-test array. The first test is always run, | 
|  | * even if both data_size and result happen to be zero. | 
|  | */ | 
|  | if (i > 0 && | 
|  | test->test[i].data_size == 0 && | 
|  | test->test[i].result == 0) | 
|  | break; | 
|  |  | 
|  | data = generate_test_data(test, i); | 
|  | if (!data && !(test->aux & FLAG_NO_DATA)) { | 
|  | pr_cont("data generation failed "); | 
|  | err_cnt++; | 
|  | break; | 
|  | } | 
|  | ret = __run_one(fp, data, runs, &duration); | 
|  | release_test_data(test, data); | 
|  |  | 
|  | if (ret == test->test[i].result) { | 
|  | pr_cont("%lld ", duration); | 
|  | } else { | 
|  | s32 res = test->test[i].result; | 
|  |  | 
|  | pr_cont("ret %d != %d (%#x != %#x)", | 
|  | ret, res, ret, res); | 
|  | err_cnt++; | 
|  | } | 
|  | } | 
|  |  | 
|  | return err_cnt; | 
|  | } | 
|  |  | 
|  | static char test_name[64]; | 
|  | module_param_string(test_name, test_name, sizeof(test_name), 0); | 
|  |  | 
|  | static int test_id = -1; | 
|  | module_param(test_id, int, 0); | 
|  |  | 
|  | static int test_range[2] = { 0, INT_MAX }; | 
|  | module_param_array(test_range, int, NULL, 0); | 
|  |  | 
|  | static bool exclude_test(int test_id) | 
|  | { | 
|  | return test_id < test_range[0] || test_id > test_range[1]; | 
|  | } | 
|  |  | 
|  | static __init struct sk_buff *build_test_skb(void) | 
|  | { | 
|  | u32 headroom = NET_SKB_PAD + NET_IP_ALIGN + ETH_HLEN; | 
|  | struct sk_buff *skb[2]; | 
|  | struct page *page[2]; | 
|  | int i, data_size = 8; | 
|  |  | 
|  | for (i = 0; i < 2; i++) { | 
|  | page[i] = alloc_page(GFP_KERNEL); | 
|  | if (!page[i]) { | 
|  | if (i == 0) | 
|  | goto err_page0; | 
|  | else | 
|  | goto err_page1; | 
|  | } | 
|  |  | 
|  | /* this will set skb[i]->head_frag */ | 
|  | skb[i] = dev_alloc_skb(headroom + data_size); | 
|  | if (!skb[i]) { | 
|  | if (i == 0) | 
|  | goto err_skb0; | 
|  | else | 
|  | goto err_skb1; | 
|  | } | 
|  |  | 
|  | skb_reserve(skb[i], headroom); | 
|  | skb_put(skb[i], data_size); | 
|  | skb[i]->protocol = htons(ETH_P_IP); | 
|  | skb_reset_network_header(skb[i]); | 
|  | skb_set_mac_header(skb[i], -ETH_HLEN); | 
|  |  | 
|  | skb_add_rx_frag(skb[i], 0, page[i], 0, 64, 64); | 
|  | // skb_headlen(skb[i]): 8, skb[i]->head_frag = 1 | 
|  | } | 
|  |  | 
|  | /* setup shinfo */ | 
|  | skb_shinfo(skb[0])->gso_size = 1448; | 
|  | skb_shinfo(skb[0])->gso_type = SKB_GSO_TCPV4; | 
|  | skb_shinfo(skb[0])->gso_type |= SKB_GSO_DODGY; | 
|  | skb_shinfo(skb[0])->gso_segs = 0; | 
|  | skb_shinfo(skb[0])->frag_list = skb[1]; | 
|  | skb_shinfo(skb[0])->hwtstamps.hwtstamp = 1000; | 
|  |  | 
|  | /* adjust skb[0]'s len */ | 
|  | skb[0]->len += skb[1]->len; | 
|  | skb[0]->data_len += skb[1]->data_len; | 
|  | skb[0]->truesize += skb[1]->truesize; | 
|  |  | 
|  | return skb[0]; | 
|  |  | 
|  | err_skb1: | 
|  | __free_page(page[1]); | 
|  | err_page1: | 
|  | kfree_skb(skb[0]); | 
|  | err_skb0: | 
|  | __free_page(page[0]); | 
|  | err_page0: | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static __init struct sk_buff *build_test_skb_linear_no_head_frag(void) | 
|  | { | 
|  | unsigned int alloc_size = 2000; | 
|  | unsigned int headroom = 102, doffset = 72, data_size = 1308; | 
|  | struct sk_buff *skb[2]; | 
|  | int i; | 
|  |  | 
|  | /* skbs linked in a frag_list, both with linear data, with head_frag=0 | 
|  | * (data allocated by kmalloc), both have tcp data of 1308 bytes | 
|  | * (total payload is 2616 bytes). | 
|  | * Data offset is 72 bytes (40 ipv6 hdr, 32 tcp hdr). Some headroom. | 
|  | */ | 
|  | for (i = 0; i < 2; i++) { | 
|  | skb[i] = alloc_skb(alloc_size, GFP_KERNEL); | 
|  | if (!skb[i]) { | 
|  | if (i == 0) | 
|  | goto err_skb0; | 
|  | else | 
|  | goto err_skb1; | 
|  | } | 
|  |  | 
|  | skb[i]->protocol = htons(ETH_P_IPV6); | 
|  | skb_reserve(skb[i], headroom); | 
|  | skb_put(skb[i], doffset + data_size); | 
|  | skb_reset_network_header(skb[i]); | 
|  | if (i == 0) | 
|  | skb_reset_mac_header(skb[i]); | 
|  | else | 
|  | skb_set_mac_header(skb[i], -ETH_HLEN); | 
|  | __skb_pull(skb[i], doffset); | 
|  | } | 
|  |  | 
|  | /* setup shinfo. | 
|  | * mimic bpf_skb_proto_4_to_6, which resets gso_segs and assigns a | 
|  | * reduced gso_size. | 
|  | */ | 
|  | skb_shinfo(skb[0])->gso_size = 1288; | 
|  | skb_shinfo(skb[0])->gso_type = SKB_GSO_TCPV6 | SKB_GSO_DODGY; | 
|  | skb_shinfo(skb[0])->gso_segs = 0; | 
|  | skb_shinfo(skb[0])->frag_list = skb[1]; | 
|  |  | 
|  | /* adjust skb[0]'s len */ | 
|  | skb[0]->len += skb[1]->len; | 
|  | skb[0]->data_len += skb[1]->len; | 
|  | skb[0]->truesize += skb[1]->truesize; | 
|  |  | 
|  | return skb[0]; | 
|  |  | 
|  | err_skb1: | 
|  | kfree_skb(skb[0]); | 
|  | err_skb0: | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | struct skb_segment_test { | 
|  | const char *descr; | 
|  | struct sk_buff *(*build_skb)(void); | 
|  | netdev_features_t features; | 
|  | }; | 
|  |  | 
|  | static struct skb_segment_test skb_segment_tests[] __initconst = { | 
|  | { | 
|  | .descr = "gso_with_rx_frags", | 
|  | .build_skb = build_test_skb, | 
|  | .features = NETIF_F_SG | NETIF_F_GSO_PARTIAL | NETIF_F_IP_CSUM | | 
|  | NETIF_F_IPV6_CSUM | NETIF_F_RXCSUM | 
|  | }, | 
|  | { | 
|  | .descr = "gso_linear_no_head_frag", | 
|  | .build_skb = build_test_skb_linear_no_head_frag, | 
|  | .features = NETIF_F_SG | NETIF_F_FRAGLIST | | 
|  | NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_GSO | | 
|  | NETIF_F_GRO | NETIF_F_IPV6_CSUM | NETIF_F_RXCSUM | | 
|  | NETIF_F_HW_VLAN_STAG_TX | 
|  | } | 
|  | }; | 
|  |  | 
|  | static __init int test_skb_segment_single(const struct skb_segment_test *test) | 
|  | { | 
|  | struct sk_buff *skb, *segs; | 
|  | int ret = -1; | 
|  |  | 
|  | skb = test->build_skb(); | 
|  | if (!skb) { | 
|  | pr_info("%s: failed to build_test_skb", __func__); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | segs = skb_segment(skb, test->features); | 
|  | if (!IS_ERR(segs)) { | 
|  | kfree_skb_list(segs); | 
|  | ret = 0; | 
|  | } | 
|  | kfree_skb(skb); | 
|  | done: | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static __init int test_skb_segment(void) | 
|  | { | 
|  | int i, err_cnt = 0, pass_cnt = 0; | 
|  |  | 
|  | for (i = 0; i < ARRAY_SIZE(skb_segment_tests); i++) { | 
|  | const struct skb_segment_test *test = &skb_segment_tests[i]; | 
|  |  | 
|  | cond_resched(); | 
|  | if (exclude_test(i)) | 
|  | continue; | 
|  |  | 
|  | pr_info("#%d %s ", i, test->descr); | 
|  |  | 
|  | if (test_skb_segment_single(test)) { | 
|  | pr_cont("FAIL\n"); | 
|  | err_cnt++; | 
|  | } else { | 
|  | pr_cont("PASS\n"); | 
|  | pass_cnt++; | 
|  | } | 
|  | } | 
|  |  | 
|  | pr_info("%s: Summary: %d PASSED, %d FAILED\n", __func__, | 
|  | pass_cnt, err_cnt); | 
|  | return err_cnt ? -EINVAL : 0; | 
|  | } | 
|  |  | 
|  | static __init int test_bpf(void) | 
|  | { | 
|  | int i, err_cnt = 0, pass_cnt = 0; | 
|  | int jit_cnt = 0, run_cnt = 0; | 
|  |  | 
|  | for (i = 0; i < ARRAY_SIZE(tests); i++) { | 
|  | struct bpf_prog *fp; | 
|  | int err; | 
|  |  | 
|  | cond_resched(); | 
|  | if (exclude_test(i)) | 
|  | continue; | 
|  |  | 
|  | pr_info("#%d %s ", i, tests[i].descr); | 
|  |  | 
|  | if (tests[i].fill_helper && | 
|  | tests[i].fill_helper(&tests[i]) < 0) { | 
|  | pr_cont("FAIL to prog_fill\n"); | 
|  | continue; | 
|  | } | 
|  |  | 
|  | fp = generate_filter(i, &err); | 
|  |  | 
|  | if (tests[i].fill_helper) { | 
|  | kfree(tests[i].u.ptr.insns); | 
|  | tests[i].u.ptr.insns = NULL; | 
|  | } | 
|  |  | 
|  | if (fp == NULL) { | 
|  | if (err == 0) { | 
|  | pass_cnt++; | 
|  | continue; | 
|  | } | 
|  | err_cnt++; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | pr_cont("jited:%u ", fp->jited); | 
|  |  | 
|  | run_cnt++; | 
|  | if (fp->jited) | 
|  | jit_cnt++; | 
|  |  | 
|  | err = run_one(fp, &tests[i]); | 
|  | release_filter(fp, i); | 
|  |  | 
|  | if (err) { | 
|  | pr_cont("FAIL (%d times)\n", err); | 
|  | err_cnt++; | 
|  | } else { | 
|  | pr_cont("PASS\n"); | 
|  | pass_cnt++; | 
|  | } | 
|  | } | 
|  |  | 
|  | pr_info("Summary: %d PASSED, %d FAILED, [%d/%d JIT'ed]\n", | 
|  | pass_cnt, err_cnt, jit_cnt, run_cnt); | 
|  |  | 
|  | return err_cnt ? -EINVAL : 0; | 
|  | } | 
|  |  | 
|  | struct tail_call_test { | 
|  | const char *descr; | 
|  | struct bpf_insn insns[MAX_INSNS]; | 
|  | int flags; | 
|  | int result; | 
|  | int stack_depth; | 
|  | bool has_tail_call; | 
|  | }; | 
|  |  | 
|  | /* Flags that can be passed to tail call test cases */ | 
|  | #define FLAG_NEED_STATE		BIT(0) | 
|  | #define FLAG_RESULT_IN_STATE	BIT(1) | 
|  |  | 
|  | /* | 
|  | * Magic marker used in test snippets for tail calls below. | 
|  | * BPF_LD/MOV to R2 and R2 with this immediate value is replaced | 
|  | * with the proper values by the test runner. | 
|  | */ | 
|  | #define TAIL_CALL_MARKER 0x7a11ca11 | 
|  |  | 
|  | /* Special offset to indicate a NULL call target */ | 
|  | #define TAIL_CALL_NULL 0x7fff | 
|  |  | 
|  | /* Special offset to indicate an out-of-range index */ | 
|  | #define TAIL_CALL_INVALID 0x7ffe | 
|  |  | 
|  | #define TAIL_CALL(offset)			       \ | 
|  | BPF_LD_IMM64(R2, TAIL_CALL_MARKER),	       \ | 
|  | BPF_RAW_INSN(BPF_ALU | BPF_MOV | BPF_K, R3, 0, \ | 
|  | offset, TAIL_CALL_MARKER),	       \ | 
|  | BPF_JMP_IMM(BPF_TAIL_CALL, 0, 0, 0) | 
|  |  | 
|  | /* | 
|  | * A test function to be called from a BPF program, clobbering a lot of | 
|  | * CPU registers in the process. A JITed BPF program calling this function | 
|  | * must save and restore any caller-saved registers it uses for internal | 
|  | * state, for example the current tail call count. | 
|  | */ | 
|  | BPF_CALL_1(bpf_test_func, u64, arg) | 
|  | { | 
|  | char buf[64]; | 
|  | long a = 0; | 
|  | long b = 1; | 
|  | long c = 2; | 
|  | long d = 3; | 
|  | long e = 4; | 
|  | long f = 5; | 
|  | long g = 6; | 
|  | long h = 7; | 
|  |  | 
|  | return snprintf(buf, sizeof(buf), | 
|  | "%ld %lu %lx %ld %lu %lx %ld %lu %x", | 
|  | a, b, c, d, e, f, g, h, (int)arg); | 
|  | } | 
|  | #define BPF_FUNC_test_func __BPF_FUNC_MAX_ID | 
|  |  | 
|  | /* | 
|  | * Tail call tests. Each test case may call any other test in the table, | 
|  | * including itself, specified as a relative index offset from the calling | 
|  | * test. The index TAIL_CALL_NULL can be used to specify a NULL target | 
|  | * function to test the JIT error path. Similarly, the index TAIL_CALL_INVALID | 
|  | * results in a target index that is out of range. | 
|  | */ | 
|  | static struct tail_call_test tail_call_tests[] = { | 
|  | { | 
|  | "Tail call leaf", | 
|  | .insns = { | 
|  | BPF_ALU64_REG(BPF_MOV, R0, R1), | 
|  | BPF_ALU64_IMM(BPF_ADD, R0, 1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .result = 1, | 
|  | }, | 
|  | { | 
|  | "Tail call 2", | 
|  | .insns = { | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 2), | 
|  | TAIL_CALL(-1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .result = 3, | 
|  | .has_tail_call = true, | 
|  | }, | 
|  | { | 
|  | "Tail call 3", | 
|  | .insns = { | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 3), | 
|  | TAIL_CALL(-1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .result = 6, | 
|  | .has_tail_call = true, | 
|  | }, | 
|  | { | 
|  | "Tail call 4", | 
|  | .insns = { | 
|  | BPF_ALU64_IMM(BPF_ADD, R1, 4), | 
|  | TAIL_CALL(-1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .result = 10, | 
|  | .has_tail_call = true, | 
|  | }, | 
|  | { | 
|  | "Tail call load/store leaf", | 
|  | .insns = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R1, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R2, 2), | 
|  | BPF_ALU64_REG(BPF_MOV, R3, BPF_REG_FP), | 
|  | BPF_STX_MEM(BPF_DW, R3, R1, -8), | 
|  | BPF_STX_MEM(BPF_DW, R3, R2, -16), | 
|  | BPF_LDX_MEM(BPF_DW, R0, BPF_REG_FP, -8), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R1, 3), | 
|  | BPF_LDX_MEM(BPF_DW, R0, BPF_REG_FP, -16), | 
|  | BPF_JMP_REG(BPF_JNE, R0, R2, 1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .result = 0, | 
|  | .stack_depth = 32, | 
|  | }, | 
|  | { | 
|  | "Tail call load/store", | 
|  | .insns = { | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, 3), | 
|  | BPF_STX_MEM(BPF_DW, BPF_REG_FP, R0, -8), | 
|  | TAIL_CALL(-1), | 
|  | BPF_ALU64_IMM(BPF_MOV, R0, -1), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .result = 0, | 
|  | .stack_depth = 16, | 
|  | .has_tail_call = true, | 
|  | }, | 
|  | { | 
|  | "Tail call error path, max count reached", | 
|  | .insns = { | 
|  | BPF_LDX_MEM(BPF_W, R2, R1, 0), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 1), | 
|  | BPF_STX_MEM(BPF_W, R1, R2, 0), | 
|  | TAIL_CALL(0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .flags = FLAG_NEED_STATE | FLAG_RESULT_IN_STATE, | 
|  | .result = (MAX_TAIL_CALL_CNT + 1) * MAX_TESTRUNS, | 
|  | .has_tail_call = true, | 
|  | }, | 
|  | { | 
|  | "Tail call count preserved across function calls", | 
|  | .insns = { | 
|  | BPF_LDX_MEM(BPF_W, R2, R1, 0), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 1), | 
|  | BPF_STX_MEM(BPF_W, R1, R2, 0), | 
|  | BPF_STX_MEM(BPF_DW, R10, R1, -8), | 
|  | BPF_CALL_REL(BPF_FUNC_get_numa_node_id), | 
|  | BPF_CALL_REL(BPF_FUNC_ktime_get_ns), | 
|  | BPF_CALL_REL(BPF_FUNC_ktime_get_boot_ns), | 
|  | BPF_CALL_REL(BPF_FUNC_ktime_get_coarse_ns), | 
|  | BPF_CALL_REL(BPF_FUNC_jiffies64), | 
|  | BPF_CALL_REL(BPF_FUNC_test_func), | 
|  | BPF_LDX_MEM(BPF_DW, R1, R10, -8), | 
|  | BPF_ALU32_REG(BPF_MOV, R0, R1), | 
|  | TAIL_CALL(0), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .stack_depth = 8, | 
|  | .flags = FLAG_NEED_STATE | FLAG_RESULT_IN_STATE, | 
|  | .result = (MAX_TAIL_CALL_CNT + 1) * MAX_TESTRUNS, | 
|  | .has_tail_call = true, | 
|  | }, | 
|  | { | 
|  | "Tail call error path, NULL target", | 
|  | .insns = { | 
|  | BPF_LDX_MEM(BPF_W, R2, R1, 0), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 1), | 
|  | BPF_STX_MEM(BPF_W, R1, R2, 0), | 
|  | TAIL_CALL(TAIL_CALL_NULL), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .flags = FLAG_NEED_STATE | FLAG_RESULT_IN_STATE, | 
|  | .result = MAX_TESTRUNS, | 
|  | .has_tail_call = true, | 
|  | }, | 
|  | { | 
|  | "Tail call error path, index out of range", | 
|  | .insns = { | 
|  | BPF_LDX_MEM(BPF_W, R2, R1, 0), | 
|  | BPF_ALU64_IMM(BPF_ADD, R2, 1), | 
|  | BPF_STX_MEM(BPF_W, R1, R2, 0), | 
|  | TAIL_CALL(TAIL_CALL_INVALID), | 
|  | BPF_EXIT_INSN(), | 
|  | }, | 
|  | .flags = FLAG_NEED_STATE | FLAG_RESULT_IN_STATE, | 
|  | .result = MAX_TESTRUNS, | 
|  | .has_tail_call = true, | 
|  | }, | 
|  | }; | 
|  |  | 
|  | static void __init destroy_tail_call_tests(struct bpf_array *progs) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | for (i = 0; i < ARRAY_SIZE(tail_call_tests); i++) | 
|  | if (progs->ptrs[i]) | 
|  | bpf_prog_free(progs->ptrs[i]); | 
|  | kfree(progs); | 
|  | } | 
|  |  | 
|  | static __init int prepare_tail_call_tests(struct bpf_array **pprogs) | 
|  | { | 
|  | int ntests = ARRAY_SIZE(tail_call_tests); | 
|  | struct bpf_array *progs; | 
|  | int which, err; | 
|  |  | 
|  | /* Allocate the table of programs to be used for tail calls */ | 
|  | progs = kzalloc(struct_size(progs, ptrs, ntests + 1), GFP_KERNEL); | 
|  | if (!progs) | 
|  | goto out_nomem; | 
|  |  | 
|  | /* Create all eBPF programs and populate the table */ | 
|  | for (which = 0; which < ntests; which++) { | 
|  | struct tail_call_test *test = &tail_call_tests[which]; | 
|  | struct bpf_prog *fp; | 
|  | int len, i; | 
|  |  | 
|  | /* Compute the number of program instructions */ | 
|  | for (len = 0; len < MAX_INSNS; len++) { | 
|  | struct bpf_insn *insn = &test->insns[len]; | 
|  |  | 
|  | if (len < MAX_INSNS - 1 && | 
|  | insn->code == (BPF_LD | BPF_DW | BPF_IMM)) | 
|  | len++; | 
|  | if (insn->code == 0) | 
|  | break; | 
|  | } | 
|  |  | 
|  | /* Allocate and initialize the program */ | 
|  | fp = bpf_prog_alloc(bpf_prog_size(len), 0); | 
|  | if (!fp) | 
|  | goto out_nomem; | 
|  |  | 
|  | fp->len = len; | 
|  | fp->type = BPF_PROG_TYPE_SOCKET_FILTER; | 
|  | fp->aux->stack_depth = test->stack_depth; | 
|  | fp->aux->tail_call_reachable = test->has_tail_call; | 
|  | memcpy(fp->insnsi, test->insns, len * sizeof(struct bpf_insn)); | 
|  |  | 
|  | /* Relocate runtime tail call offsets and addresses */ | 
|  | for (i = 0; i < len; i++) { | 
|  | struct bpf_insn *insn = &fp->insnsi[i]; | 
|  | long addr = 0; | 
|  |  | 
|  | switch (insn->code) { | 
|  | case BPF_LD | BPF_DW | BPF_IMM: | 
|  | if (insn->imm != TAIL_CALL_MARKER) | 
|  | break; | 
|  | insn[0].imm = (u32)(long)progs; | 
|  | insn[1].imm = ((u64)(long)progs) >> 32; | 
|  | break; | 
|  |  | 
|  | case BPF_ALU | BPF_MOV | BPF_K: | 
|  | if (insn->imm != TAIL_CALL_MARKER) | 
|  | break; | 
|  | if (insn->off == TAIL_CALL_NULL) | 
|  | insn->imm = ntests; | 
|  | else if (insn->off == TAIL_CALL_INVALID) | 
|  | insn->imm = ntests + 1; | 
|  | else | 
|  | insn->imm = which + insn->off; | 
|  | insn->off = 0; | 
|  | break; | 
|  |  | 
|  | case BPF_JMP | BPF_CALL: | 
|  | if (insn->src_reg != BPF_PSEUDO_CALL) | 
|  | break; | 
|  | switch (insn->imm) { | 
|  | case BPF_FUNC_get_numa_node_id: | 
|  | addr = (long)&numa_node_id; | 
|  | break; | 
|  | case BPF_FUNC_ktime_get_ns: | 
|  | addr = (long)&ktime_get_ns; | 
|  | break; | 
|  | case BPF_FUNC_ktime_get_boot_ns: | 
|  | addr = (long)&ktime_get_boot_fast_ns; | 
|  | break; | 
|  | case BPF_FUNC_ktime_get_coarse_ns: | 
|  | addr = (long)&ktime_get_coarse_ns; | 
|  | break; | 
|  | case BPF_FUNC_jiffies64: | 
|  | addr = (long)&get_jiffies_64; | 
|  | break; | 
|  | case BPF_FUNC_test_func: | 
|  | addr = (long)&bpf_test_func; | 
|  | break; | 
|  | default: | 
|  | err = -EFAULT; | 
|  | goto out_err; | 
|  | } | 
|  | *insn = BPF_EMIT_CALL(addr); | 
|  | if ((long)__bpf_call_base + insn->imm != addr) | 
|  | *insn = BPF_JMP_A(0); /* Skip: NOP */ | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | fp = bpf_prog_select_runtime(fp, &err); | 
|  | if (err) | 
|  | goto out_err; | 
|  |  | 
|  | progs->ptrs[which] = fp; | 
|  | } | 
|  |  | 
|  | /* The last entry contains a NULL program pointer */ | 
|  | progs->map.max_entries = ntests + 1; | 
|  | *pprogs = progs; | 
|  | return 0; | 
|  |  | 
|  | out_nomem: | 
|  | err = -ENOMEM; | 
|  |  | 
|  | out_err: | 
|  | if (progs) | 
|  | destroy_tail_call_tests(progs); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | static __init int test_tail_calls(struct bpf_array *progs) | 
|  | { | 
|  | int i, err_cnt = 0, pass_cnt = 0; | 
|  | int jit_cnt = 0, run_cnt = 0; | 
|  |  | 
|  | for (i = 0; i < ARRAY_SIZE(tail_call_tests); i++) { | 
|  | struct tail_call_test *test = &tail_call_tests[i]; | 
|  | struct bpf_prog *fp = progs->ptrs[i]; | 
|  | int *data = NULL; | 
|  | int state = 0; | 
|  | u64 duration; | 
|  | int ret; | 
|  |  | 
|  | cond_resched(); | 
|  | if (exclude_test(i)) | 
|  | continue; | 
|  |  | 
|  | pr_info("#%d %s ", i, test->descr); | 
|  | if (!fp) { | 
|  | err_cnt++; | 
|  | continue; | 
|  | } | 
|  | pr_cont("jited:%u ", fp->jited); | 
|  |  | 
|  | run_cnt++; | 
|  | if (fp->jited) | 
|  | jit_cnt++; | 
|  |  | 
|  | if (test->flags & FLAG_NEED_STATE) | 
|  | data = &state; | 
|  | ret = __run_one(fp, data, MAX_TESTRUNS, &duration); | 
|  | if (test->flags & FLAG_RESULT_IN_STATE) | 
|  | ret = state; | 
|  | if (ret == test->result) { | 
|  | pr_cont("%lld PASS", duration); | 
|  | pass_cnt++; | 
|  | } else { | 
|  | pr_cont("ret %d != %d FAIL", ret, test->result); | 
|  | err_cnt++; | 
|  | } | 
|  | } | 
|  |  | 
|  | pr_info("%s: Summary: %d PASSED, %d FAILED, [%d/%d JIT'ed]\n", | 
|  | __func__, pass_cnt, err_cnt, jit_cnt, run_cnt); | 
|  |  | 
|  | return err_cnt ? -EINVAL : 0; | 
|  | } | 
|  |  | 
|  | static char test_suite[32]; | 
|  | module_param_string(test_suite, test_suite, sizeof(test_suite), 0); | 
|  |  | 
|  | static __init int find_test_index(const char *test_name) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | if (!strcmp(test_suite, "test_bpf")) { | 
|  | for (i = 0; i < ARRAY_SIZE(tests); i++) { | 
|  | if (!strcmp(tests[i].descr, test_name)) | 
|  | return i; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (!strcmp(test_suite, "test_tail_calls")) { | 
|  | for (i = 0; i < ARRAY_SIZE(tail_call_tests); i++) { | 
|  | if (!strcmp(tail_call_tests[i].descr, test_name)) | 
|  | return i; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (!strcmp(test_suite, "test_skb_segment")) { | 
|  | for (i = 0; i < ARRAY_SIZE(skb_segment_tests); i++) { | 
|  | if (!strcmp(skb_segment_tests[i].descr, test_name)) | 
|  | return i; | 
|  | } | 
|  | } | 
|  |  | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | static __init int prepare_test_range(void) | 
|  | { | 
|  | int valid_range; | 
|  |  | 
|  | if (!strcmp(test_suite, "test_bpf")) | 
|  | valid_range = ARRAY_SIZE(tests); | 
|  | else if (!strcmp(test_suite, "test_tail_calls")) | 
|  | valid_range = ARRAY_SIZE(tail_call_tests); | 
|  | else if (!strcmp(test_suite, "test_skb_segment")) | 
|  | valid_range = ARRAY_SIZE(skb_segment_tests); | 
|  | else | 
|  | return 0; | 
|  |  | 
|  | if (test_id >= 0) { | 
|  | /* | 
|  | * if a test_id was specified, use test_range to | 
|  | * cover only that test. | 
|  | */ | 
|  | if (test_id >= valid_range) { | 
|  | pr_err("test_bpf: invalid test_id specified for '%s' suite.\n", | 
|  | test_suite); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | test_range[0] = test_id; | 
|  | test_range[1] = test_id; | 
|  | } else if (*test_name) { | 
|  | /* | 
|  | * if a test_name was specified, find it and setup | 
|  | * test_range to cover only that test. | 
|  | */ | 
|  | int idx = find_test_index(test_name); | 
|  |  | 
|  | if (idx < 0) { | 
|  | pr_err("test_bpf: no test named '%s' found for '%s' suite.\n", | 
|  | test_name, test_suite); | 
|  | return -EINVAL; | 
|  | } | 
|  | test_range[0] = idx; | 
|  | test_range[1] = idx; | 
|  | } else if (test_range[0] != 0 || test_range[1] != INT_MAX) { | 
|  | /* | 
|  | * check that the supplied test_range is valid. | 
|  | */ | 
|  | if (test_range[0] < 0 || test_range[1] >= valid_range) { | 
|  | pr_err("test_bpf: test_range is out of bound for '%s' suite.\n", | 
|  | test_suite); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | if (test_range[1] < test_range[0]) { | 
|  | pr_err("test_bpf: test_range is ending before it starts.\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int __init test_bpf_init(void) | 
|  | { | 
|  | struct bpf_array *progs = NULL; | 
|  | int ret; | 
|  |  | 
|  | if (strlen(test_suite) && | 
|  | strcmp(test_suite, "test_bpf") && | 
|  | strcmp(test_suite, "test_tail_calls") && | 
|  | strcmp(test_suite, "test_skb_segment")) { | 
|  | pr_err("test_bpf: invalid test_suite '%s' specified.\n", test_suite); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * if test_suite is not specified, but test_id, test_name or test_range | 
|  | * is specified, set 'test_bpf' as the default test suite. | 
|  | */ | 
|  | if (!strlen(test_suite) && | 
|  | (test_id != -1 || strlen(test_name) || | 
|  | (test_range[0] != 0 || test_range[1] != INT_MAX))) { | 
|  | pr_info("test_bpf: set 'test_bpf' as the default test_suite.\n"); | 
|  | strscpy(test_suite, "test_bpf", sizeof(test_suite)); | 
|  | } | 
|  |  | 
|  | ret = prepare_test_range(); | 
|  | if (ret < 0) | 
|  | return ret; | 
|  |  | 
|  | if (!strlen(test_suite) || !strcmp(test_suite, "test_bpf")) { | 
|  | ret = test_bpf(); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (!strlen(test_suite) || !strcmp(test_suite, "test_tail_calls")) { | 
|  | ret = prepare_tail_call_tests(&progs); | 
|  | if (ret) | 
|  | return ret; | 
|  | ret = test_tail_calls(progs); | 
|  | destroy_tail_call_tests(progs); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (!strlen(test_suite) || !strcmp(test_suite, "test_skb_segment")) | 
|  | return test_skb_segment(); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void __exit test_bpf_exit(void) | 
|  | { | 
|  | } | 
|  |  | 
|  | module_init(test_bpf_init); | 
|  | module_exit(test_bpf_exit); | 
|  |  | 
|  | MODULE_DESCRIPTION("Testsuite for BPF interpreter and BPF JIT compiler"); | 
|  | MODULE_LICENSE("GPL"); |