~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/tools/testing/selftests/bpf/progs/exceptions.c

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0
  2 #include <vmlinux.h>
  3 #include <bpf/bpf_tracing.h>
  4 #include <bpf/bpf_helpers.h>
  5 #include <bpf/bpf_core_read.h>
  6 #include <bpf/bpf_endian.h>
  7 #include "bpf_misc.h"
  8 #include "bpf_experimental.h"
  9 
 10 #ifndef ETH_P_IP
 11 #define ETH_P_IP 0x0800
 12 #endif
 13 
 14 struct {
 15         __uint(type, BPF_MAP_TYPE_PROG_ARRAY);
 16         __uint(max_entries, 4);
 17         __uint(key_size, sizeof(__u32));
 18         __uint(value_size, sizeof(__u32));
 19 } jmp_table SEC(".maps");
 20 
 21 static __noinline int static_func(u64 i)
 22 {
 23         bpf_throw(32);
 24         return i;
 25 }
 26 
 27 __noinline int global2static_simple(u64 i)
 28 {
 29         static_func(i + 2);
 30         return i - 1;
 31 }
 32 
 33 __noinline int global2static(u64 i)
 34 {
 35         if (i == ETH_P_IP)
 36                 bpf_throw(16);
 37         return static_func(i);
 38 }
 39 
 40 static __noinline int static2global(u64 i)
 41 {
 42         return global2static(i) + i;
 43 }
 44 
 45 SEC("tc")
 46 int exception_throw_always_1(struct __sk_buff *ctx)
 47 {
 48         bpf_throw(64);
 49         return 0;
 50 }
 51 
 52 /* In this case, the global func will never be seen executing after call to
 53  * static subprog, hence verifier will DCE the remaining instructions. Ensure we
 54  * are resilient to that.
 55  */
 56 SEC("tc")
 57 int exception_throw_always_2(struct __sk_buff *ctx)
 58 {
 59         return global2static_simple(ctx->protocol);
 60 }
 61 
 62 SEC("tc")
 63 int exception_throw_unwind_1(struct __sk_buff *ctx)
 64 {
 65         return static2global(bpf_ntohs(ctx->protocol));
 66 }
 67 
 68 SEC("tc")
 69 int exception_throw_unwind_2(struct __sk_buff *ctx)
 70 {
 71         return static2global(bpf_ntohs(ctx->protocol) - 1);
 72 }
 73 
 74 SEC("tc")
 75 int exception_throw_default(struct __sk_buff *ctx)
 76 {
 77         bpf_throw(0);
 78         return 1;
 79 }
 80 
 81 SEC("tc")
 82 int exception_throw_default_value(struct __sk_buff *ctx)
 83 {
 84         bpf_throw(5);
 85         return 1;
 86 }
 87 
 88 SEC("tc")
 89 int exception_tail_call_target(struct __sk_buff *ctx)
 90 {
 91         bpf_throw(16);
 92         return 0;
 93 }
 94 
 95 static __noinline
 96 int exception_tail_call_subprog(struct __sk_buff *ctx)
 97 {
 98         volatile int ret = 10;
 99 
100         bpf_tail_call_static(ctx, &jmp_table, 0);
101         return ret;
102 }
103 
104 SEC("tc")
105 int exception_tail_call(struct __sk_buff *ctx) {
106         volatile int ret = 0;
107 
108         ret = exception_tail_call_subprog(ctx);
109         return ret + 8;
110 }
111 
112 __noinline int exception_ext_global(struct __sk_buff *ctx)
113 {
114         volatile int ret = 0;
115 
116         return ret;
117 }
118 
119 static __noinline int exception_ext_static(struct __sk_buff *ctx)
120 {
121         return exception_ext_global(ctx);
122 }
123 
124 SEC("tc")
125 int exception_ext(struct __sk_buff *ctx)
126 {
127         return exception_ext_static(ctx);
128 }
129 
130 __noinline int exception_cb_mod_global(u64 cookie)
131 {
132         volatile int ret = 0;
133 
134         return ret;
135 }
136 
137 /* Example of how the exception callback supplied during verification can still
138  * introduce extensions by calling to dummy global functions, and alter runtime
139  * behavior.
140  *
141  * Right now we don't allow freplace attachment to exception callback itself,
142  * but if the need arises this restriction is technically feasible to relax in
143  * the future.
144  */
145 __noinline int exception_cb_mod(u64 cookie)
146 {
147         return exception_cb_mod_global(cookie) + cookie + 10;
148 }
149 
150 SEC("tc")
151 __exception_cb(exception_cb_mod)
152 int exception_ext_mod_cb_runtime(struct __sk_buff *ctx)
153 {
154         bpf_throw(25);
155         return 0;
156 }
157 
158 __noinline static int subprog(struct __sk_buff *ctx)
159 {
160         return bpf_ktime_get_ns();
161 }
162 
163 __noinline static int throwing_subprog(struct __sk_buff *ctx)
164 {
165         if (ctx->tstamp)
166                 bpf_throw(0);
167         return bpf_ktime_get_ns();
168 }
169 
170 __noinline int global_subprog(struct __sk_buff *ctx)
171 {
172         return bpf_ktime_get_ns();
173 }
174 
175 __noinline int throwing_global_subprog(struct __sk_buff *ctx)
176 {
177         if (ctx->tstamp)
178                 bpf_throw(0);
179         return bpf_ktime_get_ns();
180 }
181 
182 SEC("tc")
183 int exception_throw_subprog(struct __sk_buff *ctx)
184 {
185         switch (ctx->protocol) {
186         case 1:
187                 return subprog(ctx);
188         case 2:
189                 return global_subprog(ctx);
190         case 3:
191                 return throwing_subprog(ctx);
192         case 4:
193                 return throwing_global_subprog(ctx);
194         default:
195                 break;
196         }
197         bpf_throw(1);
198         return 0;
199 }
200 
201 __noinline int assert_nz_gfunc(u64 c)
202 {
203         volatile u64 cookie = c;
204 
205         bpf_assert(cookie != 0);
206         return 0;
207 }
208 
209 __noinline int assert_zero_gfunc(u64 c)
210 {
211         volatile u64 cookie = c;
212 
213         bpf_assert(bpf_cmp_unlikely(cookie, ==, 0));
214         return 0;
215 }
216 
217 __noinline int assert_neg_gfunc(s64 c)
218 {
219         volatile s64 cookie = c;
220 
221         bpf_assert(bpf_cmp_unlikely(cookie, <, 0));
222         return 0;
223 }
224 
225 __noinline int assert_pos_gfunc(s64 c)
226 {
227         volatile s64 cookie = c;
228 
229         bpf_assert(bpf_cmp_unlikely(cookie, >, 0));
230         return 0;
231 }
232 
233 __noinline int assert_negeq_gfunc(s64 c)
234 {
235         volatile s64 cookie = c;
236 
237         bpf_assert(bpf_cmp_unlikely(cookie, <=, -1));
238         return 0;
239 }
240 
241 __noinline int assert_poseq_gfunc(s64 c)
242 {
243         volatile s64 cookie = c;
244 
245         bpf_assert(bpf_cmp_unlikely(cookie, >=, 1));
246         return 0;
247 }
248 
249 __noinline int assert_nz_gfunc_with(u64 c)
250 {
251         volatile u64 cookie = c;
252 
253         bpf_assert_with(cookie != 0, cookie + 100);
254         return 0;
255 }
256 
257 __noinline int assert_zero_gfunc_with(u64 c)
258 {
259         volatile u64 cookie = c;
260 
261         bpf_assert_with(bpf_cmp_unlikely(cookie, ==, 0), cookie + 100);
262         return 0;
263 }
264 
265 __noinline int assert_neg_gfunc_with(s64 c)
266 {
267         volatile s64 cookie = c;
268 
269         bpf_assert_with(bpf_cmp_unlikely(cookie, <, 0), cookie + 100);
270         return 0;
271 }
272 
273 __noinline int assert_pos_gfunc_with(s64 c)
274 {
275         volatile s64 cookie = c;
276 
277         bpf_assert_with(bpf_cmp_unlikely(cookie, >, 0), cookie + 100);
278         return 0;
279 }
280 
281 __noinline int assert_negeq_gfunc_with(s64 c)
282 {
283         volatile s64 cookie = c;
284 
285         bpf_assert_with(bpf_cmp_unlikely(cookie, <=, -1), cookie + 100);
286         return 0;
287 }
288 
289 __noinline int assert_poseq_gfunc_with(s64 c)
290 {
291         volatile s64 cookie = c;
292 
293         bpf_assert_with(bpf_cmp_unlikely(cookie, >=, 1), cookie + 100);
294         return 0;
295 }
296 
297 #define check_assert(name, cookie, tag)                         \
298 SEC("tc")                                                       \
299 int exception##tag##name(struct __sk_buff *ctx)                 \
300 {                                                               \
301         return name(cookie) + 1;                                \
302 }
303 
304 check_assert(assert_nz_gfunc, 5, _);
305 check_assert(assert_zero_gfunc, 0, _);
306 check_assert(assert_neg_gfunc, -100, _);
307 check_assert(assert_pos_gfunc, 100, _);
308 check_assert(assert_negeq_gfunc, -1, _);
309 check_assert(assert_poseq_gfunc, 1, _);
310 
311 check_assert(assert_nz_gfunc_with, 5, _);
312 check_assert(assert_zero_gfunc_with, 0, _);
313 check_assert(assert_neg_gfunc_with, -100, _);
314 check_assert(assert_pos_gfunc_with, 100, _);
315 check_assert(assert_negeq_gfunc_with, -1, _);
316 check_assert(assert_poseq_gfunc_with, 1, _);
317 
318 check_assert(assert_nz_gfunc, 0, _bad_);
319 check_assert(assert_zero_gfunc, 5, _bad_);
320 check_assert(assert_neg_gfunc, 100, _bad_);
321 check_assert(assert_pos_gfunc, -100, _bad_);
322 check_assert(assert_negeq_gfunc, 1, _bad_);
323 check_assert(assert_poseq_gfunc, -1, _bad_);
324 
325 check_assert(assert_nz_gfunc_with, 0, _bad_);
326 check_assert(assert_zero_gfunc_with, 5, _bad_);
327 check_assert(assert_neg_gfunc_with, 100, _bad_);
328 check_assert(assert_pos_gfunc_with, -100, _bad_);
329 check_assert(assert_negeq_gfunc_with, 1, _bad_);
330 check_assert(assert_poseq_gfunc_with, -1, _bad_);
331 
332 SEC("tc")
333 int exception_assert_range(struct __sk_buff *ctx)
334 {
335         u64 time = bpf_ktime_get_ns();
336 
337         bpf_assert_range(time, 0, ~0ULL);
338         return 1;
339 }
340 
341 SEC("tc")
342 int exception_assert_range_with(struct __sk_buff *ctx)
343 {
344         u64 time = bpf_ktime_get_ns();
345 
346         bpf_assert_range_with(time, 0, ~0ULL, 10);
347         return 1;
348 }
349 
350 SEC("tc")
351 int exception_bad_assert_range(struct __sk_buff *ctx)
352 {
353         u64 time = bpf_ktime_get_ns();
354 
355         bpf_assert_range(time, -100, 100);
356         return 1;
357 }
358 
359 SEC("tc")
360 int exception_bad_assert_range_with(struct __sk_buff *ctx)
361 {
362         u64 time = bpf_ktime_get_ns();
363 
364         bpf_assert_range_with(time, -1000, 1000, 10);
365         return 1;
366 }
367 
368 char _license[] SEC("license") = "GPL";
369 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php