369 lines
		
	
	
		
			7.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			369 lines
		
	
	
		
			7.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| #include <vmlinux.h>
 | |
| #include <bpf/bpf_tracing.h>
 | |
| #include <bpf/bpf_helpers.h>
 | |
| #include <bpf/bpf_core_read.h>
 | |
| #include <bpf/bpf_endian.h>
 | |
| #include "bpf_misc.h"
 | |
| #include "bpf_experimental.h"
 | |
| 
 | |
| #ifndef ETH_P_IP
 | |
| #define ETH_P_IP 0x0800
 | |
| #endif
 | |
| 
 | |
| struct {
 | |
| 	__uint(type, BPF_MAP_TYPE_PROG_ARRAY);
 | |
| 	__uint(max_entries, 4);
 | |
| 	__uint(key_size, sizeof(__u32));
 | |
| 	__uint(value_size, sizeof(__u32));
 | |
| } jmp_table SEC(".maps");
 | |
| 
 | |
| static __noinline int static_func(u64 i)
 | |
| {
 | |
| 	bpf_throw(32);
 | |
| 	return i;
 | |
| }
 | |
| 
 | |
| __noinline int global2static_simple(u64 i)
 | |
| {
 | |
| 	static_func(i + 2);
 | |
| 	return i - 1;
 | |
| }
 | |
| 
 | |
| __noinline int global2static(u64 i)
 | |
| {
 | |
| 	if (i == ETH_P_IP)
 | |
| 		bpf_throw(16);
 | |
| 	return static_func(i);
 | |
| }
 | |
| 
 | |
| static __noinline int static2global(u64 i)
 | |
| {
 | |
| 	return global2static(i) + i;
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_throw_always_1(struct __sk_buff *ctx)
 | |
| {
 | |
| 	bpf_throw(64);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* In this case, the global func will never be seen executing after call to
 | |
|  * static subprog, hence verifier will DCE the remaining instructions. Ensure we
 | |
|  * are resilient to that.
 | |
|  */
 | |
| SEC("tc")
 | |
| int exception_throw_always_2(struct __sk_buff *ctx)
 | |
| {
 | |
| 	return global2static_simple(ctx->protocol);
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_throw_unwind_1(struct __sk_buff *ctx)
 | |
| {
 | |
| 	return static2global(bpf_ntohs(ctx->protocol));
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_throw_unwind_2(struct __sk_buff *ctx)
 | |
| {
 | |
| 	return static2global(bpf_ntohs(ctx->protocol) - 1);
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_throw_default(struct __sk_buff *ctx)
 | |
| {
 | |
| 	bpf_throw(0);
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_throw_default_value(struct __sk_buff *ctx)
 | |
| {
 | |
| 	bpf_throw(5);
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_tail_call_target(struct __sk_buff *ctx)
 | |
| {
 | |
| 	bpf_throw(16);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static __noinline
 | |
| int exception_tail_call_subprog(struct __sk_buff *ctx)
 | |
| {
 | |
| 	volatile int ret = 10;
 | |
| 
 | |
| 	bpf_tail_call_static(ctx, &jmp_table, 0);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_tail_call(struct __sk_buff *ctx) {
 | |
| 	volatile int ret = 0;
 | |
| 
 | |
| 	ret = exception_tail_call_subprog(ctx);
 | |
| 	return ret + 8;
 | |
| }
 | |
| 
 | |
| __noinline int exception_ext_global(struct __sk_buff *ctx)
 | |
| {
 | |
| 	volatile int ret = 0;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static __noinline int exception_ext_static(struct __sk_buff *ctx)
 | |
| {
 | |
| 	return exception_ext_global(ctx);
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_ext(struct __sk_buff *ctx)
 | |
| {
 | |
| 	return exception_ext_static(ctx);
 | |
| }
 | |
| 
 | |
| __noinline int exception_cb_mod_global(u64 cookie)
 | |
| {
 | |
| 	volatile int ret = 0;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /* Example of how the exception callback supplied during verification can still
 | |
|  * introduce extensions by calling to dummy global functions, and alter runtime
 | |
|  * behavior.
 | |
|  *
 | |
|  * Right now we don't allow freplace attachment to exception callback itself,
 | |
|  * but if the need arises this restriction is technically feasible to relax in
 | |
|  * the future.
 | |
|  */
 | |
| __noinline int exception_cb_mod(u64 cookie)
 | |
| {
 | |
| 	return exception_cb_mod_global(cookie) + cookie + 10;
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| __exception_cb(exception_cb_mod)
 | |
| int exception_ext_mod_cb_runtime(struct __sk_buff *ctx)
 | |
| {
 | |
| 	bpf_throw(25);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline static int subprog(struct __sk_buff *ctx)
 | |
| {
 | |
| 	return bpf_ktime_get_ns();
 | |
| }
 | |
| 
 | |
| __noinline static int throwing_subprog(struct __sk_buff *ctx)
 | |
| {
 | |
| 	if (ctx->tstamp)
 | |
| 		bpf_throw(0);
 | |
| 	return bpf_ktime_get_ns();
 | |
| }
 | |
| 
 | |
| __noinline int global_subprog(struct __sk_buff *ctx)
 | |
| {
 | |
| 	return bpf_ktime_get_ns();
 | |
| }
 | |
| 
 | |
| __noinline int throwing_global_subprog(struct __sk_buff *ctx)
 | |
| {
 | |
| 	if (ctx->tstamp)
 | |
| 		bpf_throw(0);
 | |
| 	return bpf_ktime_get_ns();
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_throw_subprog(struct __sk_buff *ctx)
 | |
| {
 | |
| 	switch (ctx->protocol) {
 | |
| 	case 1:
 | |
| 		return subprog(ctx);
 | |
| 	case 2:
 | |
| 		return global_subprog(ctx);
 | |
| 	case 3:
 | |
| 		return throwing_subprog(ctx);
 | |
| 	case 4:
 | |
| 		return throwing_global_subprog(ctx);
 | |
| 	default:
 | |
| 		break;
 | |
| 	}
 | |
| 	bpf_throw(1);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_nz_gfunc(u64 c)
 | |
| {
 | |
| 	volatile u64 cookie = c;
 | |
| 
 | |
| 	bpf_assert(cookie != 0);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_zero_gfunc(u64 c)
 | |
| {
 | |
| 	volatile u64 cookie = c;
 | |
| 
 | |
| 	bpf_assert(bpf_cmp_unlikely(cookie, ==, 0));
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_neg_gfunc(s64 c)
 | |
| {
 | |
| 	volatile s64 cookie = c;
 | |
| 
 | |
| 	bpf_assert(bpf_cmp_unlikely(cookie, <, 0));
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_pos_gfunc(s64 c)
 | |
| {
 | |
| 	volatile s64 cookie = c;
 | |
| 
 | |
| 	bpf_assert(bpf_cmp_unlikely(cookie, >, 0));
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_negeq_gfunc(s64 c)
 | |
| {
 | |
| 	volatile s64 cookie = c;
 | |
| 
 | |
| 	bpf_assert(bpf_cmp_unlikely(cookie, <=, -1));
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_poseq_gfunc(s64 c)
 | |
| {
 | |
| 	volatile s64 cookie = c;
 | |
| 
 | |
| 	bpf_assert(bpf_cmp_unlikely(cookie, >=, 1));
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_nz_gfunc_with(u64 c)
 | |
| {
 | |
| 	volatile u64 cookie = c;
 | |
| 
 | |
| 	bpf_assert_with(cookie != 0, cookie + 100);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_zero_gfunc_with(u64 c)
 | |
| {
 | |
| 	volatile u64 cookie = c;
 | |
| 
 | |
| 	bpf_assert_with(bpf_cmp_unlikely(cookie, ==, 0), cookie + 100);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_neg_gfunc_with(s64 c)
 | |
| {
 | |
| 	volatile s64 cookie = c;
 | |
| 
 | |
| 	bpf_assert_with(bpf_cmp_unlikely(cookie, <, 0), cookie + 100);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_pos_gfunc_with(s64 c)
 | |
| {
 | |
| 	volatile s64 cookie = c;
 | |
| 
 | |
| 	bpf_assert_with(bpf_cmp_unlikely(cookie, >, 0), cookie + 100);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_negeq_gfunc_with(s64 c)
 | |
| {
 | |
| 	volatile s64 cookie = c;
 | |
| 
 | |
| 	bpf_assert_with(bpf_cmp_unlikely(cookie, <=, -1), cookie + 100);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __noinline int assert_poseq_gfunc_with(s64 c)
 | |
| {
 | |
| 	volatile s64 cookie = c;
 | |
| 
 | |
| 	bpf_assert_with(bpf_cmp_unlikely(cookie, >=, 1), cookie + 100);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| #define check_assert(name, cookie, tag)				\
 | |
| SEC("tc")							\
 | |
| int exception##tag##name(struct __sk_buff *ctx)			\
 | |
| {								\
 | |
| 	return name(cookie) + 1;				\
 | |
| }
 | |
| 
 | |
| check_assert(assert_nz_gfunc, 5, _);
 | |
| check_assert(assert_zero_gfunc, 0, _);
 | |
| check_assert(assert_neg_gfunc, -100, _);
 | |
| check_assert(assert_pos_gfunc, 100, _);
 | |
| check_assert(assert_negeq_gfunc, -1, _);
 | |
| check_assert(assert_poseq_gfunc, 1, _);
 | |
| 
 | |
| check_assert(assert_nz_gfunc_with, 5, _);
 | |
| check_assert(assert_zero_gfunc_with, 0, _);
 | |
| check_assert(assert_neg_gfunc_with, -100, _);
 | |
| check_assert(assert_pos_gfunc_with, 100, _);
 | |
| check_assert(assert_negeq_gfunc_with, -1, _);
 | |
| check_assert(assert_poseq_gfunc_with, 1, _);
 | |
| 
 | |
| check_assert(assert_nz_gfunc, 0, _bad_);
 | |
| check_assert(assert_zero_gfunc, 5, _bad_);
 | |
| check_assert(assert_neg_gfunc, 100, _bad_);
 | |
| check_assert(assert_pos_gfunc, -100, _bad_);
 | |
| check_assert(assert_negeq_gfunc, 1, _bad_);
 | |
| check_assert(assert_poseq_gfunc, -1, _bad_);
 | |
| 
 | |
| check_assert(assert_nz_gfunc_with, 0, _bad_);
 | |
| check_assert(assert_zero_gfunc_with, 5, _bad_);
 | |
| check_assert(assert_neg_gfunc_with, 100, _bad_);
 | |
| check_assert(assert_pos_gfunc_with, -100, _bad_);
 | |
| check_assert(assert_negeq_gfunc_with, 1, _bad_);
 | |
| check_assert(assert_poseq_gfunc_with, -1, _bad_);
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_assert_range(struct __sk_buff *ctx)
 | |
| {
 | |
| 	u64 time = bpf_ktime_get_ns();
 | |
| 
 | |
| 	bpf_assert_range(time, 0, ~0ULL);
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_assert_range_with(struct __sk_buff *ctx)
 | |
| {
 | |
| 	u64 time = bpf_ktime_get_ns();
 | |
| 
 | |
| 	bpf_assert_range_with(time, 0, ~0ULL, 10);
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_bad_assert_range(struct __sk_buff *ctx)
 | |
| {
 | |
| 	u64 time = bpf_ktime_get_ns();
 | |
| 
 | |
| 	bpf_assert_range(time, -100, 100);
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| SEC("tc")
 | |
| int exception_bad_assert_range_with(struct __sk_buff *ctx)
 | |
| {
 | |
| 	u64 time = bpf_ktime_get_ns();
 | |
| 
 | |
| 	bpf_assert_range_with(time, -1000, 1000, 10);
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| char _license[] SEC("license") = "GPL";
 |