1 // SPDX-License-Identifier: GPL-2.0 2 /* Converted from tools/testing/selftests/bpf/verifier/ref_tracking.c */ 3 4 #include <linux/bpf.h> 5 #include <bpf/bpf_helpers.h> 6 #include "../../../include/linux/filter.h" 7 #include "bpf_misc.h" 8 9 #define BPF_SK_LOOKUP(func) \ 10 /* struct bpf_sock_tuple tuple = {} */ \ 11 "r2 = 0;" \ 12 "*(u32*)(r10 - 8) = r2;" \ 13 "*(u64*)(r10 - 16) = r2;" \ 14 "*(u64*)(r10 - 24) = r2;" \ 15 "*(u64*)(r10 - 32) = r2;" \ 16 "*(u64*)(r10 - 40) = r2;" \ 17 "*(u64*)(r10 - 48) = r2;" \ 18 /* sk = func(ctx, &tuple, sizeof tuple, 0, 0) */ \ 19 "r2 = r10;" \ 20 "r2 += -48;" \ 21 "r3 = %[sizeof_bpf_sock_tuple];"\ 22 "r4 = 0;" \ 23 "r5 = 0;" \ 24 "call %[" #func "];" 25 26 struct bpf_key {} __attribute__((preserve_access_index)); 27 28 extern void bpf_key_put(struct bpf_key *key) __ksym; 29 extern struct bpf_key *bpf_lookup_system_key(__u64 id) __ksym; 30 extern struct bpf_key *bpf_lookup_user_key(__u32 serial, __u64 flags) __ksym; 31 32 /* BTF FUNC records are not generated for kfuncs referenced 33 * from inline assembly. These records are necessary for 34 * libbpf to link the program. The function below is a hack 35 * to ensure that BTF FUNC records are generated. 36 */ 37 void __kfunc_btf_root(void) 38 { 39 bpf_key_put(0); 40 bpf_lookup_system_key(0); 41 bpf_lookup_user_key(0, 0); 42 } 43 44 #define MAX_ENTRIES 11 45 46 struct test_val { 47 unsigned int index; 48 int foo[MAX_ENTRIES]; 49 }; 50 51 struct { 52 __uint(type, BPF_MAP_TYPE_ARRAY); 53 __uint(max_entries, 1); 54 __type(key, int); 55 __type(value, struct test_val); 56 } map_array_48b SEC(".maps"); 57 58 struct { 59 __uint(type, BPF_MAP_TYPE_RINGBUF); 60 __uint(max_entries, 4096); 61 } map_ringbuf SEC(".maps"); 62 63 void dummy_prog_42_tc(void); 64 void dummy_prog_24_tc(void); 65 void dummy_prog_loop1_tc(void); 66 67 struct { 68 __uint(type, BPF_MAP_TYPE_PROG_ARRAY); 69 __uint(max_entries, 4); 70 __uint(key_size, sizeof(int)); 71 __array(values, void (void)); 72 } map_prog1_tc SEC(".maps") = { 73 .values = { 74 [0] = (void *)&dummy_prog_42_tc, 75 [1] = (void *)&dummy_prog_loop1_tc, 76 [2] = (void *)&dummy_prog_24_tc, 77 }, 78 }; 79 80 SEC("tc") 81 __auxiliary 82 __naked void dummy_prog_42_tc(void) 83 { 84 asm volatile ("r0 = 42; exit;"); 85 } 86 87 SEC("tc") 88 __auxiliary 89 __naked void dummy_prog_24_tc(void) 90 { 91 asm volatile ("r0 = 24; exit;"); 92 } 93 94 SEC("tc") 95 __auxiliary 96 __naked void dummy_prog_loop1_tc(void) 97 { 98 asm volatile (" \ 99 r3 = 1; \ 100 r2 = %[map_prog1_tc] ll; \ 101 call %[bpf_tail_call]; \ 102 r0 = 41; \ 103 exit; \ 104 " : 105 : __imm(bpf_tail_call), 106 __imm_addr(map_prog1_tc) 107 : __clobber_all); 108 } 109 110 SEC("tc") 111 __description("reference tracking: leak potential reference") 112 __failure __msg("Unreleased reference") 113 __naked void reference_tracking_leak_potential_reference(void) 114 { 115 asm volatile ( 116 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 117 " r6 = r0; /* leak reference */ \ 118 exit; \ 119 " : 120 : __imm(bpf_sk_lookup_tcp), 121 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 122 : __clobber_all); 123 } 124 125 SEC("tc") 126 __description("reference tracking: leak potential reference to sock_common") 127 __failure __msg("Unreleased reference") 128 __naked void potential_reference_to_sock_common_1(void) 129 { 130 asm volatile ( 131 BPF_SK_LOOKUP(bpf_skc_lookup_tcp) 132 " r6 = r0; /* leak reference */ \ 133 exit; \ 134 " : 135 : __imm(bpf_skc_lookup_tcp), 136 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 137 : __clobber_all); 138 } 139 140 SEC("tc") 141 __description("reference tracking: leak potential reference on stack") 142 __failure __msg("Unreleased reference") 143 __naked void leak_potential_reference_on_stack(void) 144 { 145 asm volatile ( 146 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 147 " r4 = r10; \ 148 r4 += -8; \ 149 *(u64*)(r4 + 0) = r0; \ 150 r0 = 0; \ 151 exit; \ 152 " : 153 : __imm(bpf_sk_lookup_tcp), 154 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 155 : __clobber_all); 156 } 157 158 SEC("tc") 159 __description("reference tracking: leak potential reference on stack 2") 160 __failure __msg("Unreleased reference") 161 __naked void potential_reference_on_stack_2(void) 162 { 163 asm volatile ( 164 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 165 " r4 = r10; \ 166 r4 += -8; \ 167 *(u64*)(r4 + 0) = r0; \ 168 r0 = 0; \ 169 r1 = 0; \ 170 *(u64*)(r4 + 0) = r1; \ 171 exit; \ 172 " : 173 : __imm(bpf_sk_lookup_tcp), 174 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 175 : __clobber_all); 176 } 177 178 SEC("tc") 179 __description("reference tracking: zero potential reference") 180 __failure __msg("Unreleased reference") 181 __naked void reference_tracking_zero_potential_reference(void) 182 { 183 asm volatile ( 184 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 185 " r0 = 0; /* leak reference */ \ 186 exit; \ 187 " : 188 : __imm(bpf_sk_lookup_tcp), 189 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 190 : __clobber_all); 191 } 192 193 SEC("tc") 194 __description("reference tracking: zero potential reference to sock_common") 195 __failure __msg("Unreleased reference") 196 __naked void potential_reference_to_sock_common_2(void) 197 { 198 asm volatile ( 199 BPF_SK_LOOKUP(bpf_skc_lookup_tcp) 200 " r0 = 0; /* leak reference */ \ 201 exit; \ 202 " : 203 : __imm(bpf_skc_lookup_tcp), 204 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 205 : __clobber_all); 206 } 207 208 SEC("tc") 209 __description("reference tracking: copy and zero potential references") 210 __failure __msg("Unreleased reference") 211 __naked void copy_and_zero_potential_references(void) 212 { 213 asm volatile ( 214 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 215 " r7 = r0; \ 216 r0 = 0; \ 217 r7 = 0; /* leak reference */ \ 218 exit; \ 219 " : 220 : __imm(bpf_sk_lookup_tcp), 221 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 222 : __clobber_all); 223 } 224 225 SEC("lsm.s/bpf") 226 __description("reference tracking: acquire/release user key reference") 227 __success 228 __naked void acquire_release_user_key_reference(void) 229 { 230 asm volatile (" \ 231 r1 = -3; \ 232 r2 = 0; \ 233 call %[bpf_lookup_user_key]; \ 234 if r0 == 0 goto l0_%=; \ 235 r1 = r0; \ 236 call %[bpf_key_put]; \ 237 l0_%=: r0 = 0; \ 238 exit; \ 239 " : 240 : __imm(bpf_key_put), 241 __imm(bpf_lookup_user_key) 242 : __clobber_all); 243 } 244 245 SEC("lsm.s/bpf") 246 __description("reference tracking: acquire/release system key reference") 247 __success 248 __naked void acquire_release_system_key_reference(void) 249 { 250 asm volatile (" \ 251 r1 = 1; \ 252 call %[bpf_lookup_system_key]; \ 253 if r0 == 0 goto l0_%=; \ 254 r1 = r0; \ 255 call %[bpf_key_put]; \ 256 l0_%=: r0 = 0; \ 257 exit; \ 258 " : 259 : __imm(bpf_key_put), 260 __imm(bpf_lookup_system_key) 261 : __clobber_all); 262 } 263 264 SEC("lsm.s/bpf") 265 __description("reference tracking: release user key reference without check") 266 __failure __msg("Possibly NULL pointer passed to trusted arg0") 267 __naked void user_key_reference_without_check(void) 268 { 269 asm volatile (" \ 270 r1 = -3; \ 271 r2 = 0; \ 272 call %[bpf_lookup_user_key]; \ 273 r1 = r0; \ 274 call %[bpf_key_put]; \ 275 r0 = 0; \ 276 exit; \ 277 " : 278 : __imm(bpf_key_put), 279 __imm(bpf_lookup_user_key) 280 : __clobber_all); 281 } 282 283 SEC("lsm.s/bpf") 284 __description("reference tracking: release system key reference without check") 285 __failure __msg("Possibly NULL pointer passed to trusted arg0") 286 __naked void system_key_reference_without_check(void) 287 { 288 asm volatile (" \ 289 r1 = 1; \ 290 call %[bpf_lookup_system_key]; \ 291 r1 = r0; \ 292 call %[bpf_key_put]; \ 293 r0 = 0; \ 294 exit; \ 295 " : 296 : __imm(bpf_key_put), 297 __imm(bpf_lookup_system_key) 298 : __clobber_all); 299 } 300 301 SEC("lsm.s/bpf") 302 __description("reference tracking: release with NULL key pointer") 303 __failure __msg("Possibly NULL pointer passed to trusted arg0") 304 __naked void release_with_null_key_pointer(void) 305 { 306 asm volatile (" \ 307 r1 = 0; \ 308 call %[bpf_key_put]; \ 309 r0 = 0; \ 310 exit; \ 311 " : 312 : __imm(bpf_key_put) 313 : __clobber_all); 314 } 315 316 SEC("lsm.s/bpf") 317 __description("reference tracking: leak potential reference to user key") 318 __failure __msg("Unreleased reference") 319 __naked void potential_reference_to_user_key(void) 320 { 321 asm volatile (" \ 322 r1 = -3; \ 323 r2 = 0; \ 324 call %[bpf_lookup_user_key]; \ 325 exit; \ 326 " : 327 : __imm(bpf_lookup_user_key) 328 : __clobber_all); 329 } 330 331 SEC("lsm.s/bpf") 332 __description("reference tracking: leak potential reference to system key") 333 __failure __msg("Unreleased reference") 334 __naked void potential_reference_to_system_key(void) 335 { 336 asm volatile (" \ 337 r1 = 1; \ 338 call %[bpf_lookup_system_key]; \ 339 exit; \ 340 " : 341 : __imm(bpf_lookup_system_key) 342 : __clobber_all); 343 } 344 345 SEC("tc") 346 __description("reference tracking: release reference without check") 347 __failure __msg("type=sock_or_null expected=sock") 348 __naked void tracking_release_reference_without_check(void) 349 { 350 asm volatile ( 351 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 352 " /* reference in r0 may be NULL */ \ 353 r1 = r0; \ 354 r2 = 0; \ 355 call %[bpf_sk_release]; \ 356 exit; \ 357 " : 358 : __imm(bpf_sk_lookup_tcp), 359 __imm(bpf_sk_release), 360 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 361 : __clobber_all); 362 } 363 364 SEC("tc") 365 __description("reference tracking: release reference to sock_common without check") 366 __failure __msg("type=sock_common_or_null expected=sock") 367 __naked void to_sock_common_without_check(void) 368 { 369 asm volatile ( 370 BPF_SK_LOOKUP(bpf_skc_lookup_tcp) 371 " /* reference in r0 may be NULL */ \ 372 r1 = r0; \ 373 r2 = 0; \ 374 call %[bpf_sk_release]; \ 375 exit; \ 376 " : 377 : __imm(bpf_sk_release), 378 __imm(bpf_skc_lookup_tcp), 379 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 380 : __clobber_all); 381 } 382 383 SEC("tc") 384 __description("reference tracking: release reference") 385 __success __retval(0) 386 __naked void reference_tracking_release_reference(void) 387 { 388 asm volatile ( 389 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 390 " r1 = r0; \ 391 if r0 == 0 goto l0_%=; \ 392 call %[bpf_sk_release]; \ 393 l0_%=: exit; \ 394 " : 395 : __imm(bpf_sk_lookup_tcp), 396 __imm(bpf_sk_release), 397 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 398 : __clobber_all); 399 } 400 401 SEC("tc") 402 __description("reference tracking: release reference to sock_common") 403 __success __retval(0) 404 __naked void release_reference_to_sock_common(void) 405 { 406 asm volatile ( 407 BPF_SK_LOOKUP(bpf_skc_lookup_tcp) 408 " r1 = r0; \ 409 if r0 == 0 goto l0_%=; \ 410 call %[bpf_sk_release]; \ 411 l0_%=: exit; \ 412 " : 413 : __imm(bpf_sk_release), 414 __imm(bpf_skc_lookup_tcp), 415 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 416 : __clobber_all); 417 } 418 419 SEC("tc") 420 __description("reference tracking: release reference 2") 421 __success __retval(0) 422 __naked void reference_tracking_release_reference_2(void) 423 { 424 asm volatile ( 425 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 426 " r1 = r0; \ 427 if r0 != 0 goto l0_%=; \ 428 exit; \ 429 l0_%=: call %[bpf_sk_release]; \ 430 exit; \ 431 " : 432 : __imm(bpf_sk_lookup_tcp), 433 __imm(bpf_sk_release), 434 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 435 : __clobber_all); 436 } 437 438 SEC("tc") 439 __description("reference tracking: release reference twice") 440 __failure __msg("type=scalar expected=sock") 441 __naked void reference_tracking_release_reference_twice(void) 442 { 443 asm volatile ( 444 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 445 " r1 = r0; \ 446 r6 = r0; \ 447 if r0 == 0 goto l0_%=; \ 448 call %[bpf_sk_release]; \ 449 l0_%=: r1 = r6; \ 450 call %[bpf_sk_release]; \ 451 exit; \ 452 " : 453 : __imm(bpf_sk_lookup_tcp), 454 __imm(bpf_sk_release), 455 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 456 : __clobber_all); 457 } 458 459 SEC("tc") 460 __description("reference tracking: release reference twice inside branch") 461 __failure __msg("type=scalar expected=sock") 462 __naked void release_reference_twice_inside_branch(void) 463 { 464 asm volatile ( 465 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 466 " r1 = r0; \ 467 r6 = r0; \ 468 if r0 == 0 goto l0_%=; /* goto end */ \ 469 call %[bpf_sk_release]; \ 470 r1 = r6; \ 471 call %[bpf_sk_release]; \ 472 l0_%=: exit; \ 473 " : 474 : __imm(bpf_sk_lookup_tcp), 475 __imm(bpf_sk_release), 476 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 477 : __clobber_all); 478 } 479 480 SEC("tc") 481 __description("reference tracking: alloc, check, free in one subbranch") 482 __failure __msg("Unreleased reference") 483 __flag(BPF_F_ANY_ALIGNMENT) 484 __naked void check_free_in_one_subbranch(void) 485 { 486 asm volatile (" \ 487 r2 = *(u32*)(r1 + %[__sk_buff_data]); \ 488 r3 = *(u32*)(r1 + %[__sk_buff_data_end]); \ 489 r0 = r2; \ 490 r0 += 16; \ 491 /* if (offsetof(skb, mark) > data_len) exit; */ \ 492 if r0 <= r3 goto l0_%=; \ 493 exit; \ 494 l0_%=: r6 = *(u32*)(r2 + %[__sk_buff_mark]); \ 495 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 496 " if r6 == 0 goto l1_%=; /* mark == 0? */\ 497 /* Leak reference in R0 */ \ 498 exit; \ 499 l1_%=: if r0 == 0 goto l2_%=; /* sk NULL? */ \ 500 r1 = r0; \ 501 call %[bpf_sk_release]; \ 502 l2_%=: exit; \ 503 " : 504 : __imm(bpf_sk_lookup_tcp), 505 __imm(bpf_sk_release), 506 __imm_const(__sk_buff_data, offsetof(struct __sk_buff, data)), 507 __imm_const(__sk_buff_data_end, offsetof(struct __sk_buff, data_end)), 508 __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), 509 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 510 : __clobber_all); 511 } 512 513 SEC("tc") 514 __description("reference tracking: alloc, check, free in both subbranches") 515 __success __retval(0) __flag(BPF_F_ANY_ALIGNMENT) 516 __naked void check_free_in_both_subbranches(void) 517 { 518 asm volatile (" \ 519 r2 = *(u32*)(r1 + %[__sk_buff_data]); \ 520 r3 = *(u32*)(r1 + %[__sk_buff_data_end]); \ 521 r0 = r2; \ 522 r0 += 16; \ 523 /* if (offsetof(skb, mark) > data_len) exit; */ \ 524 if r0 <= r3 goto l0_%=; \ 525 exit; \ 526 l0_%=: r6 = *(u32*)(r2 + %[__sk_buff_mark]); \ 527 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 528 " if r6 == 0 goto l1_%=; /* mark == 0? */\ 529 if r0 == 0 goto l2_%=; /* sk NULL? */ \ 530 r1 = r0; \ 531 call %[bpf_sk_release]; \ 532 l2_%=: exit; \ 533 l1_%=: if r0 == 0 goto l3_%=; /* sk NULL? */ \ 534 r1 = r0; \ 535 call %[bpf_sk_release]; \ 536 l3_%=: exit; \ 537 " : 538 : __imm(bpf_sk_lookup_tcp), 539 __imm(bpf_sk_release), 540 __imm_const(__sk_buff_data, offsetof(struct __sk_buff, data)), 541 __imm_const(__sk_buff_data_end, offsetof(struct __sk_buff, data_end)), 542 __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), 543 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 544 : __clobber_all); 545 } 546 547 SEC("tc") 548 __description("reference tracking in call: free reference in subprog") 549 __success __retval(0) 550 __naked void call_free_reference_in_subprog(void) 551 { 552 asm volatile ( 553 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 554 " r1 = r0; /* unchecked reference */ \ 555 call call_free_reference_in_subprog__1; \ 556 r0 = 0; \ 557 exit; \ 558 " : 559 : __imm(bpf_sk_lookup_tcp), 560 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 561 : __clobber_all); 562 } 563 564 static __naked __noinline __attribute__((used)) 565 void call_free_reference_in_subprog__1(void) 566 { 567 asm volatile (" \ 568 /* subprog 1 */ \ 569 r2 = r1; \ 570 if r2 == 0 goto l0_%=; \ 571 call %[bpf_sk_release]; \ 572 l0_%=: exit; \ 573 " : 574 : __imm(bpf_sk_release) 575 : __clobber_all); 576 } 577 578 SEC("tc") 579 __description("reference tracking in call: free reference in subprog and outside") 580 __failure __msg("type=scalar expected=sock") 581 __naked void reference_in_subprog_and_outside(void) 582 { 583 asm volatile ( 584 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 585 " r1 = r0; /* unchecked reference */ \ 586 r6 = r0; \ 587 call reference_in_subprog_and_outside__1; \ 588 r1 = r6; \ 589 call %[bpf_sk_release]; \ 590 exit; \ 591 " : 592 : __imm(bpf_sk_lookup_tcp), 593 __imm(bpf_sk_release), 594 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 595 : __clobber_all); 596 } 597 598 static __naked __noinline __attribute__((used)) 599 void reference_in_subprog_and_outside__1(void) 600 { 601 asm volatile (" \ 602 /* subprog 1 */ \ 603 r2 = r1; \ 604 if r2 == 0 goto l0_%=; \ 605 call %[bpf_sk_release]; \ 606 l0_%=: exit; \ 607 " : 608 : __imm(bpf_sk_release) 609 : __clobber_all); 610 } 611 612 SEC("tc") 613 __description("reference tracking in call: alloc & leak reference in subprog") 614 __failure __msg("Unreleased reference") 615 __naked void alloc_leak_reference_in_subprog(void) 616 { 617 asm volatile (" \ 618 r4 = r10; \ 619 r4 += -8; \ 620 call alloc_leak_reference_in_subprog__1; \ 621 r1 = r0; \ 622 r0 = 0; \ 623 exit; \ 624 " ::: __clobber_all); 625 } 626 627 static __naked __noinline __attribute__((used)) 628 void alloc_leak_reference_in_subprog__1(void) 629 { 630 asm volatile (" \ 631 /* subprog 1 */ \ 632 r6 = r4; \ 633 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 634 " /* spill unchecked sk_ptr into stack of caller */\ 635 *(u64*)(r6 + 0) = r0; \ 636 r1 = r0; \ 637 exit; \ 638 " : 639 : __imm(bpf_sk_lookup_tcp), 640 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 641 : __clobber_all); 642 } 643 644 SEC("tc") 645 __description("reference tracking in call: alloc in subprog, release outside") 646 __success __retval(POINTER_VALUE) 647 __naked void alloc_in_subprog_release_outside(void) 648 { 649 asm volatile (" \ 650 r4 = r10; \ 651 call alloc_in_subprog_release_outside__1; \ 652 r1 = r0; \ 653 if r0 == 0 goto l0_%=; \ 654 call %[bpf_sk_release]; \ 655 l0_%=: exit; \ 656 " : 657 : __imm(bpf_sk_release) 658 : __clobber_all); 659 } 660 661 static __naked __noinline __attribute__((used)) 662 void alloc_in_subprog_release_outside__1(void) 663 { 664 asm volatile (" \ 665 /* subprog 1 */ \ 666 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 667 " exit; /* return sk */ \ 668 " : 669 : __imm(bpf_sk_lookup_tcp), 670 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 671 : __clobber_all); 672 } 673 674 SEC("tc") 675 __description("reference tracking in call: sk_ptr leak into caller stack") 676 __failure __msg("Unreleased reference") 677 __naked void ptr_leak_into_caller_stack(void) 678 { 679 asm volatile (" \ 680 r4 = r10; \ 681 r4 += -8; \ 682 call ptr_leak_into_caller_stack__1; \ 683 r0 = 0; \ 684 exit; \ 685 " ::: __clobber_all); 686 } 687 688 static __naked __noinline __attribute__((used)) 689 void ptr_leak_into_caller_stack__1(void) 690 { 691 asm volatile (" \ 692 /* subprog 1 */ \ 693 r5 = r10; \ 694 r5 += -8; \ 695 *(u64*)(r5 + 0) = r4; \ 696 call ptr_leak_into_caller_stack__2; \ 697 /* spill unchecked sk_ptr into stack of caller */\ 698 r5 = r10; \ 699 r5 += -8; \ 700 r4 = *(u64*)(r5 + 0); \ 701 *(u64*)(r4 + 0) = r0; \ 702 exit; \ 703 " ::: __clobber_all); 704 } 705 706 static __naked __noinline __attribute__((used)) 707 void ptr_leak_into_caller_stack__2(void) 708 { 709 asm volatile (" \ 710 /* subprog 2 */ \ 711 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 712 " exit; \ 713 " : 714 : __imm(bpf_sk_lookup_tcp), 715 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 716 : __clobber_all); 717 } 718 719 SEC("tc") 720 __description("reference tracking in call: sk_ptr spill into caller stack") 721 __success __retval(0) 722 __naked void ptr_spill_into_caller_stack(void) 723 { 724 asm volatile (" \ 725 r4 = r10; \ 726 r4 += -8; \ 727 call ptr_spill_into_caller_stack__1; \ 728 r0 = 0; \ 729 exit; \ 730 " ::: __clobber_all); 731 } 732 733 static __naked __noinline __attribute__((used)) 734 void ptr_spill_into_caller_stack__1(void) 735 { 736 asm volatile (" \ 737 /* subprog 1 */ \ 738 r5 = r10; \ 739 r5 += -8; \ 740 *(u64*)(r5 + 0) = r4; \ 741 call ptr_spill_into_caller_stack__2; \ 742 /* spill unchecked sk_ptr into stack of caller */\ 743 r5 = r10; \ 744 r5 += -8; \ 745 r4 = *(u64*)(r5 + 0); \ 746 *(u64*)(r4 + 0) = r0; \ 747 if r0 == 0 goto l0_%=; \ 748 /* now the sk_ptr is verified, free the reference */\ 749 r1 = *(u64*)(r4 + 0); \ 750 call %[bpf_sk_release]; \ 751 l0_%=: exit; \ 752 " : 753 : __imm(bpf_sk_release) 754 : __clobber_all); 755 } 756 757 static __naked __noinline __attribute__((used)) 758 void ptr_spill_into_caller_stack__2(void) 759 { 760 asm volatile (" \ 761 /* subprog 2 */ \ 762 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 763 " exit; \ 764 " : 765 : __imm(bpf_sk_lookup_tcp), 766 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 767 : __clobber_all); 768 } 769 770 SEC("tc") 771 __description("reference tracking: allow LD_ABS") 772 __success __retval(0) 773 __naked void reference_tracking_allow_ld_abs(void) 774 { 775 asm volatile (" \ 776 r6 = r1; \ 777 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 778 " r1 = r0; \ 779 if r0 == 0 goto l0_%=; \ 780 call %[bpf_sk_release]; \ 781 l0_%=: r0 = *(u8*)skb[0]; \ 782 r0 = *(u16*)skb[0]; \ 783 r0 = *(u32*)skb[0]; \ 784 exit; \ 785 " : 786 : __imm(bpf_sk_lookup_tcp), 787 __imm(bpf_sk_release), 788 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 789 : __clobber_all); 790 } 791 792 SEC("tc") 793 __description("reference tracking: forbid LD_ABS while holding reference") 794 __failure __msg("BPF_LD_[ABS|IND] cannot be mixed with socket references") 795 __naked void ld_abs_while_holding_reference(void) 796 { 797 asm volatile (" \ 798 r6 = r1; \ 799 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 800 " r0 = *(u8*)skb[0]; \ 801 r0 = *(u16*)skb[0]; \ 802 r0 = *(u32*)skb[0]; \ 803 r1 = r0; \ 804 if r0 == 0 goto l0_%=; \ 805 call %[bpf_sk_release]; \ 806 l0_%=: exit; \ 807 " : 808 : __imm(bpf_sk_lookup_tcp), 809 __imm(bpf_sk_release), 810 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 811 : __clobber_all); 812 } 813 814 SEC("tc") 815 __description("reference tracking: allow LD_IND") 816 __success __retval(1) 817 __naked void reference_tracking_allow_ld_ind(void) 818 { 819 asm volatile (" \ 820 r6 = r1; \ 821 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 822 " r1 = r0; \ 823 if r0 == 0 goto l0_%=; \ 824 call %[bpf_sk_release]; \ 825 l0_%=: r7 = 1; \ 826 .8byte %[ld_ind]; \ 827 r0 = r7; \ 828 exit; \ 829 " : 830 : __imm(bpf_sk_lookup_tcp), 831 __imm(bpf_sk_release), 832 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)), 833 __imm_insn(ld_ind, BPF_LD_IND(BPF_W, BPF_REG_7, -0x200000)) 834 : __clobber_all); 835 } 836 837 SEC("tc") 838 __description("reference tracking: forbid LD_IND while holding reference") 839 __failure __msg("BPF_LD_[ABS|IND] cannot be mixed with socket references") 840 __naked void ld_ind_while_holding_reference(void) 841 { 842 asm volatile (" \ 843 r6 = r1; \ 844 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 845 " r4 = r0; \ 846 r7 = 1; \ 847 .8byte %[ld_ind]; \ 848 r0 = r7; \ 849 r1 = r4; \ 850 if r1 == 0 goto l0_%=; \ 851 call %[bpf_sk_release]; \ 852 l0_%=: exit; \ 853 " : 854 : __imm(bpf_sk_lookup_tcp), 855 __imm(bpf_sk_release), 856 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)), 857 __imm_insn(ld_ind, BPF_LD_IND(BPF_W, BPF_REG_7, -0x200000)) 858 : __clobber_all); 859 } 860 861 SEC("tc") 862 __description("reference tracking: check reference or tail call") 863 __success __retval(0) 864 __naked void check_reference_or_tail_call(void) 865 { 866 asm volatile (" \ 867 r7 = r1; \ 868 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 869 " /* if (sk) bpf_sk_release() */ \ 870 r1 = r0; \ 871 if r1 != 0 goto l0_%=; \ 872 /* bpf_tail_call() */ \ 873 r3 = 3; \ 874 r2 = %[map_prog1_tc] ll; \ 875 r1 = r7; \ 876 call %[bpf_tail_call]; \ 877 r0 = 0; \ 878 exit; \ 879 l0_%=: call %[bpf_sk_release]; \ 880 exit; \ 881 " : 882 : __imm(bpf_sk_lookup_tcp), 883 __imm(bpf_sk_release), 884 __imm(bpf_tail_call), 885 __imm_addr(map_prog1_tc), 886 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 887 : __clobber_all); 888 } 889 890 SEC("tc") 891 __description("reference tracking: release reference then tail call") 892 __success __retval(0) 893 __naked void release_reference_then_tail_call(void) 894 { 895 asm volatile (" \ 896 r7 = r1; \ 897 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 898 " /* if (sk) bpf_sk_release() */ \ 899 r1 = r0; \ 900 if r1 == 0 goto l0_%=; \ 901 call %[bpf_sk_release]; \ 902 l0_%=: /* bpf_tail_call() */ \ 903 r3 = 3; \ 904 r2 = %[map_prog1_tc] ll; \ 905 r1 = r7; \ 906 call %[bpf_tail_call]; \ 907 r0 = 0; \ 908 exit; \ 909 " : 910 : __imm(bpf_sk_lookup_tcp), 911 __imm(bpf_sk_release), 912 __imm(bpf_tail_call), 913 __imm_addr(map_prog1_tc), 914 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 915 : __clobber_all); 916 } 917 918 SEC("tc") 919 __description("reference tracking: leak possible reference over tail call") 920 __failure __msg("tail_call would lead to reference leak") 921 __naked void possible_reference_over_tail_call(void) 922 { 923 asm volatile (" \ 924 r7 = r1; \ 925 /* Look up socket and store in REG_6 */ \ 926 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 927 " /* bpf_tail_call() */ \ 928 r6 = r0; \ 929 r3 = 3; \ 930 r2 = %[map_prog1_tc] ll; \ 931 r1 = r7; \ 932 call %[bpf_tail_call]; \ 933 r0 = 0; \ 934 /* if (sk) bpf_sk_release() */ \ 935 r1 = r6; \ 936 if r1 == 0 goto l0_%=; \ 937 call %[bpf_sk_release]; \ 938 l0_%=: exit; \ 939 " : 940 : __imm(bpf_sk_lookup_tcp), 941 __imm(bpf_sk_release), 942 __imm(bpf_tail_call), 943 __imm_addr(map_prog1_tc), 944 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 945 : __clobber_all); 946 } 947 948 SEC("tc") 949 __description("reference tracking: leak checked reference over tail call") 950 __failure __msg("tail_call would lead to reference leak") 951 __naked void checked_reference_over_tail_call(void) 952 { 953 asm volatile (" \ 954 r7 = r1; \ 955 /* Look up socket and store in REG_6 */ \ 956 " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 957 " r6 = r0; \ 958 /* if (!sk) goto end */ \ 959 if r0 == 0 goto l0_%=; \ 960 /* bpf_tail_call() */ \ 961 r3 = 0; \ 962 r2 = %[map_prog1_tc] ll; \ 963 r1 = r7; \ 964 call %[bpf_tail_call]; \ 965 r0 = 0; \ 966 r1 = r6; \ 967 l0_%=: call %[bpf_sk_release]; \ 968 exit; \ 969 " : 970 : __imm(bpf_sk_lookup_tcp), 971 __imm(bpf_sk_release), 972 __imm(bpf_tail_call), 973 __imm_addr(map_prog1_tc), 974 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 975 : __clobber_all); 976 } 977 978 SEC("tc") 979 __description("reference tracking: mangle and release sock_or_null") 980 __failure __msg("R1 pointer arithmetic on sock_or_null prohibited") 981 __naked void and_release_sock_or_null(void) 982 { 983 asm volatile ( 984 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 985 " r1 = r0; \ 986 r1 += 5; \ 987 if r0 == 0 goto l0_%=; \ 988 call %[bpf_sk_release]; \ 989 l0_%=: exit; \ 990 " : 991 : __imm(bpf_sk_lookup_tcp), 992 __imm(bpf_sk_release), 993 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 994 : __clobber_all); 995 } 996 997 SEC("tc") 998 __description("reference tracking: mangle and release sock") 999 __failure __msg("R1 pointer arithmetic on sock prohibited") 1000 __naked void tracking_mangle_and_release_sock(void) 1001 { 1002 asm volatile ( 1003 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1004 " r1 = r0; \ 1005 if r0 == 0 goto l0_%=; \ 1006 r1 += 5; \ 1007 call %[bpf_sk_release]; \ 1008 l0_%=: exit; \ 1009 " : 1010 : __imm(bpf_sk_lookup_tcp), 1011 __imm(bpf_sk_release), 1012 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1013 : __clobber_all); 1014 } 1015 1016 SEC("tc") 1017 __description("reference tracking: access member") 1018 __success __retval(0) 1019 __naked void reference_tracking_access_member(void) 1020 { 1021 asm volatile ( 1022 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1023 " r6 = r0; \ 1024 if r0 == 0 goto l0_%=; \ 1025 r2 = *(u32*)(r0 + 4); \ 1026 r1 = r6; \ 1027 call %[bpf_sk_release]; \ 1028 l0_%=: exit; \ 1029 " : 1030 : __imm(bpf_sk_lookup_tcp), 1031 __imm(bpf_sk_release), 1032 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1033 : __clobber_all); 1034 } 1035 1036 SEC("tc") 1037 __description("reference tracking: write to member") 1038 __failure __msg("cannot write into sock") 1039 __naked void reference_tracking_write_to_member(void) 1040 { 1041 asm volatile ( 1042 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1043 " r6 = r0; \ 1044 if r0 == 0 goto l0_%=; \ 1045 r1 = r6; \ 1046 r2 = 42 ll; \ 1047 *(u32*)(r1 + %[bpf_sock_mark]) = r2; \ 1048 r1 = r6; \ 1049 l0_%=: call %[bpf_sk_release]; \ 1050 r0 = 0 ll; \ 1051 exit; \ 1052 " : 1053 : __imm(bpf_sk_lookup_tcp), 1054 __imm(bpf_sk_release), 1055 __imm_const(bpf_sock_mark, offsetof(struct bpf_sock, mark)), 1056 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1057 : __clobber_all); 1058 } 1059 1060 SEC("tc") 1061 __description("reference tracking: invalid 64-bit access of member") 1062 __failure __msg("invalid sock access off=0 size=8") 1063 __naked void _64_bit_access_of_member(void) 1064 { 1065 asm volatile ( 1066 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1067 " r6 = r0; \ 1068 if r0 == 0 goto l0_%=; \ 1069 r2 = *(u64*)(r0 + 0); \ 1070 r1 = r6; \ 1071 call %[bpf_sk_release]; \ 1072 l0_%=: exit; \ 1073 " : 1074 : __imm(bpf_sk_lookup_tcp), 1075 __imm(bpf_sk_release), 1076 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1077 : __clobber_all); 1078 } 1079 1080 SEC("tc") 1081 __description("reference tracking: access after release") 1082 __failure __msg("!read_ok") 1083 __naked void reference_tracking_access_after_release(void) 1084 { 1085 asm volatile ( 1086 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1087 " r1 = r0; \ 1088 if r0 == 0 goto l0_%=; \ 1089 call %[bpf_sk_release]; \ 1090 r2 = *(u32*)(r1 + 0); \ 1091 l0_%=: exit; \ 1092 " : 1093 : __imm(bpf_sk_lookup_tcp), 1094 __imm(bpf_sk_release), 1095 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1096 : __clobber_all); 1097 } 1098 1099 SEC("tc") 1100 __description("reference tracking: direct access for lookup") 1101 __success __retval(0) 1102 __naked void tracking_direct_access_for_lookup(void) 1103 { 1104 asm volatile (" \ 1105 /* Check that the packet is at least 64B long */\ 1106 r2 = *(u32*)(r1 + %[__sk_buff_data]); \ 1107 r3 = *(u32*)(r1 + %[__sk_buff_data_end]); \ 1108 r0 = r2; \ 1109 r0 += 64; \ 1110 if r0 > r3 goto l0_%=; \ 1111 /* sk = sk_lookup_tcp(ctx, skb->data, ...) */ \ 1112 r3 = %[sizeof_bpf_sock_tuple]; \ 1113 r4 = 0; \ 1114 r5 = 0; \ 1115 call %[bpf_sk_lookup_tcp]; \ 1116 r6 = r0; \ 1117 if r0 == 0 goto l0_%=; \ 1118 r2 = *(u32*)(r0 + 4); \ 1119 r1 = r6; \ 1120 call %[bpf_sk_release]; \ 1121 l0_%=: exit; \ 1122 " : 1123 : __imm(bpf_sk_lookup_tcp), 1124 __imm(bpf_sk_release), 1125 __imm_const(__sk_buff_data, offsetof(struct __sk_buff, data)), 1126 __imm_const(__sk_buff_data_end, offsetof(struct __sk_buff, data_end)), 1127 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1128 : __clobber_all); 1129 } 1130 1131 SEC("tc") 1132 __description("reference tracking: use ptr from bpf_tcp_sock() after release") 1133 __failure __msg("invalid mem access") 1134 __flag(BPF_F_ANY_ALIGNMENT) 1135 __naked void bpf_tcp_sock_after_release(void) 1136 { 1137 asm volatile ( 1138 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1139 " if r0 != 0 goto l0_%=; \ 1140 exit; \ 1141 l0_%=: r6 = r0; \ 1142 r1 = r0; \ 1143 call %[bpf_tcp_sock]; \ 1144 if r0 != 0 goto l1_%=; \ 1145 r1 = r6; \ 1146 call %[bpf_sk_release]; \ 1147 exit; \ 1148 l1_%=: r7 = r0; \ 1149 r1 = r6; \ 1150 call %[bpf_sk_release]; \ 1151 r0 = *(u32*)(r7 + %[bpf_tcp_sock_snd_cwnd]); \ 1152 exit; \ 1153 " : 1154 : __imm(bpf_sk_lookup_tcp), 1155 __imm(bpf_sk_release), 1156 __imm(bpf_tcp_sock), 1157 __imm_const(bpf_tcp_sock_snd_cwnd, offsetof(struct bpf_tcp_sock, snd_cwnd)), 1158 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1159 : __clobber_all); 1160 } 1161 1162 SEC("tc") 1163 __description("reference tracking: use ptr from bpf_sk_fullsock() after release") 1164 __failure __msg("invalid mem access") 1165 __flag(BPF_F_ANY_ALIGNMENT) 1166 __naked void bpf_sk_fullsock_after_release(void) 1167 { 1168 asm volatile ( 1169 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1170 " if r0 != 0 goto l0_%=; \ 1171 exit; \ 1172 l0_%=: r6 = r0; \ 1173 r1 = r0; \ 1174 call %[bpf_sk_fullsock]; \ 1175 if r0 != 0 goto l1_%=; \ 1176 r1 = r6; \ 1177 call %[bpf_sk_release]; \ 1178 exit; \ 1179 l1_%=: r7 = r0; \ 1180 r1 = r6; \ 1181 call %[bpf_sk_release]; \ 1182 r0 = *(u32*)(r7 + %[bpf_sock_type]); \ 1183 exit; \ 1184 " : 1185 : __imm(bpf_sk_fullsock), 1186 __imm(bpf_sk_lookup_tcp), 1187 __imm(bpf_sk_release), 1188 __imm_const(bpf_sock_type, offsetof(struct bpf_sock, type)), 1189 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1190 : __clobber_all); 1191 } 1192 1193 SEC("tc") 1194 __description("reference tracking: use ptr from bpf_sk_fullsock(tp) after release") 1195 __failure __msg("invalid mem access") 1196 __flag(BPF_F_ANY_ALIGNMENT) 1197 __naked void sk_fullsock_tp_after_release(void) 1198 { 1199 asm volatile ( 1200 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1201 " if r0 != 0 goto l0_%=; \ 1202 exit; \ 1203 l0_%=: r6 = r0; \ 1204 r1 = r0; \ 1205 call %[bpf_tcp_sock]; \ 1206 if r0 != 0 goto l1_%=; \ 1207 r1 = r6; \ 1208 call %[bpf_sk_release]; \ 1209 exit; \ 1210 l1_%=: r1 = r0; \ 1211 call %[bpf_sk_fullsock]; \ 1212 r1 = r6; \ 1213 r6 = r0; \ 1214 call %[bpf_sk_release]; \ 1215 if r6 != 0 goto l2_%=; \ 1216 exit; \ 1217 l2_%=: r0 = *(u32*)(r6 + %[bpf_sock_type]); \ 1218 exit; \ 1219 " : 1220 : __imm(bpf_sk_fullsock), 1221 __imm(bpf_sk_lookup_tcp), 1222 __imm(bpf_sk_release), 1223 __imm(bpf_tcp_sock), 1224 __imm_const(bpf_sock_type, offsetof(struct bpf_sock, type)), 1225 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1226 : __clobber_all); 1227 } 1228 1229 SEC("tc") 1230 __description("reference tracking: use sk after bpf_sk_release(tp)") 1231 __failure __msg("invalid mem access") 1232 __flag(BPF_F_ANY_ALIGNMENT) 1233 __naked void after_bpf_sk_release_tp(void) 1234 { 1235 asm volatile ( 1236 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1237 " if r0 != 0 goto l0_%=; \ 1238 exit; \ 1239 l0_%=: r6 = r0; \ 1240 r1 = r0; \ 1241 call %[bpf_tcp_sock]; \ 1242 if r0 != 0 goto l1_%=; \ 1243 r1 = r6; \ 1244 call %[bpf_sk_release]; \ 1245 exit; \ 1246 l1_%=: r1 = r0; \ 1247 call %[bpf_sk_release]; \ 1248 r0 = *(u32*)(r6 + %[bpf_sock_type]); \ 1249 exit; \ 1250 " : 1251 : __imm(bpf_sk_lookup_tcp), 1252 __imm(bpf_sk_release), 1253 __imm(bpf_tcp_sock), 1254 __imm_const(bpf_sock_type, offsetof(struct bpf_sock, type)), 1255 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1256 : __clobber_all); 1257 } 1258 1259 SEC("tc") 1260 __description("reference tracking: use ptr from bpf_get_listener_sock() after bpf_sk_release(sk)") 1261 __success __retval(0) 1262 __naked void after_bpf_sk_release_sk(void) 1263 { 1264 asm volatile ( 1265 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1266 " if r0 != 0 goto l0_%=; \ 1267 exit; \ 1268 l0_%=: r6 = r0; \ 1269 r1 = r0; \ 1270 call %[bpf_get_listener_sock]; \ 1271 if r0 != 0 goto l1_%=; \ 1272 r1 = r6; \ 1273 call %[bpf_sk_release]; \ 1274 exit; \ 1275 l1_%=: r1 = r6; \ 1276 r6 = r0; \ 1277 call %[bpf_sk_release]; \ 1278 r0 = *(u32*)(r6 + %[bpf_sock_src_port]); \ 1279 exit; \ 1280 " : 1281 : __imm(bpf_get_listener_sock), 1282 __imm(bpf_sk_lookup_tcp), 1283 __imm(bpf_sk_release), 1284 __imm_const(bpf_sock_src_port, offsetof(struct bpf_sock, src_port)), 1285 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1286 : __clobber_all); 1287 } 1288 1289 SEC("tc") 1290 __description("reference tracking: bpf_sk_release(listen_sk)") 1291 __failure __msg("R1 must be referenced when passed to release function") 1292 __naked void bpf_sk_release_listen_sk(void) 1293 { 1294 asm volatile ( 1295 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1296 " if r0 != 0 goto l0_%=; \ 1297 exit; \ 1298 l0_%=: r6 = r0; \ 1299 r1 = r0; \ 1300 call %[bpf_get_listener_sock]; \ 1301 if r0 != 0 goto l1_%=; \ 1302 r1 = r6; \ 1303 call %[bpf_sk_release]; \ 1304 exit; \ 1305 l1_%=: r1 = r0; \ 1306 call %[bpf_sk_release]; \ 1307 r0 = *(u32*)(r6 + %[bpf_sock_type]); \ 1308 r1 = r6; \ 1309 call %[bpf_sk_release]; \ 1310 exit; \ 1311 " : 1312 : __imm(bpf_get_listener_sock), 1313 __imm(bpf_sk_lookup_tcp), 1314 __imm(bpf_sk_release), 1315 __imm_const(bpf_sock_type, offsetof(struct bpf_sock, type)), 1316 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1317 : __clobber_all); 1318 } 1319 1320 /* !bpf_sk_fullsock(sk) is checked but !bpf_tcp_sock(sk) is not checked */ 1321 SEC("tc") 1322 __description("reference tracking: tp->snd_cwnd after bpf_sk_fullsock(sk) and bpf_tcp_sock(sk)") 1323 __failure __msg("invalid mem access") 1324 __naked void and_bpf_tcp_sock_sk(void) 1325 { 1326 asm volatile ( 1327 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1328 " if r0 != 0 goto l0_%=; \ 1329 exit; \ 1330 l0_%=: r6 = r0; \ 1331 r1 = r0; \ 1332 call %[bpf_sk_fullsock]; \ 1333 r7 = r0; \ 1334 r1 = r6; \ 1335 call %[bpf_tcp_sock]; \ 1336 r8 = r0; \ 1337 if r7 != 0 goto l1_%=; \ 1338 r1 = r6; \ 1339 call %[bpf_sk_release]; \ 1340 exit; \ 1341 l1_%=: r0 = *(u32*)(r8 + %[bpf_tcp_sock_snd_cwnd]); \ 1342 r1 = r6; \ 1343 call %[bpf_sk_release]; \ 1344 exit; \ 1345 " : 1346 : __imm(bpf_sk_fullsock), 1347 __imm(bpf_sk_lookup_tcp), 1348 __imm(bpf_sk_release), 1349 __imm(bpf_tcp_sock), 1350 __imm_const(bpf_tcp_sock_snd_cwnd, offsetof(struct bpf_tcp_sock, snd_cwnd)), 1351 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1352 : __clobber_all); 1353 } 1354 1355 SEC("tc") 1356 __description("reference tracking: branch tracking valid pointer null comparison") 1357 __success __retval(0) 1358 __naked void tracking_valid_pointer_null_comparison(void) 1359 { 1360 asm volatile ( 1361 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1362 " r6 = r0; \ 1363 r3 = 1; \ 1364 if r6 != 0 goto l0_%=; \ 1365 r3 = 0; \ 1366 l0_%=: if r6 == 0 goto l1_%=; \ 1367 r1 = r6; \ 1368 call %[bpf_sk_release]; \ 1369 l1_%=: exit; \ 1370 " : 1371 : __imm(bpf_sk_lookup_tcp), 1372 __imm(bpf_sk_release), 1373 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1374 : __clobber_all); 1375 } 1376 1377 SEC("tc") 1378 __description("reference tracking: branch tracking valid pointer value comparison") 1379 __failure __msg("Unreleased reference") 1380 __naked void tracking_valid_pointer_value_comparison(void) 1381 { 1382 asm volatile ( 1383 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1384 " r6 = r0; \ 1385 r3 = 1; \ 1386 if r6 == 0 goto l0_%=; \ 1387 r3 = 0; \ 1388 if r6 == 1234 goto l0_%=; \ 1389 r1 = r6; \ 1390 call %[bpf_sk_release]; \ 1391 l0_%=: exit; \ 1392 " : 1393 : __imm(bpf_sk_lookup_tcp), 1394 __imm(bpf_sk_release), 1395 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1396 : __clobber_all); 1397 } 1398 1399 SEC("tc") 1400 __description("reference tracking: bpf_sk_release(btf_tcp_sock)") 1401 __success 1402 __retval(0) 1403 __naked void sk_release_btf_tcp_sock(void) 1404 { 1405 asm volatile ( 1406 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1407 " if r0 != 0 goto l0_%=; \ 1408 exit; \ 1409 l0_%=: r6 = r0; \ 1410 r1 = r0; \ 1411 call %[bpf_skc_to_tcp_sock]; \ 1412 if r0 != 0 goto l1_%=; \ 1413 r1 = r6; \ 1414 call %[bpf_sk_release]; \ 1415 exit; \ 1416 l1_%=: r1 = r0; \ 1417 call %[bpf_sk_release]; \ 1418 exit; \ 1419 " : 1420 : __imm(bpf_sk_lookup_tcp), 1421 __imm(bpf_sk_release), 1422 __imm(bpf_skc_to_tcp_sock), 1423 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1424 : __clobber_all); 1425 } 1426 1427 SEC("tc") 1428 __description("reference tracking: use ptr from bpf_skc_to_tcp_sock() after release") 1429 __failure __msg("invalid mem access") 1430 __naked void to_tcp_sock_after_release(void) 1431 { 1432 asm volatile ( 1433 BPF_SK_LOOKUP(bpf_sk_lookup_tcp) 1434 " if r0 != 0 goto l0_%=; \ 1435 exit; \ 1436 l0_%=: r6 = r0; \ 1437 r1 = r0; \ 1438 call %[bpf_skc_to_tcp_sock]; \ 1439 if r0 != 0 goto l1_%=; \ 1440 r1 = r6; \ 1441 call %[bpf_sk_release]; \ 1442 exit; \ 1443 l1_%=: r7 = r0; \ 1444 r1 = r6; \ 1445 call %[bpf_sk_release]; \ 1446 r0 = *(u8*)(r7 + 0); \ 1447 exit; \ 1448 " : 1449 : __imm(bpf_sk_lookup_tcp), 1450 __imm(bpf_sk_release), 1451 __imm(bpf_skc_to_tcp_sock), 1452 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) 1453 : __clobber_all); 1454 } 1455 1456 SEC("socket") 1457 __description("reference tracking: try to leak released ptr reg") 1458 __success __failure_unpriv __msg_unpriv("R8 !read_ok") 1459 __retval(0) 1460 __naked void to_leak_released_ptr_reg(void) 1461 { 1462 asm volatile (" \ 1463 r0 = 0; \ 1464 *(u32*)(r10 - 4) = r0; \ 1465 r2 = r10; \ 1466 r2 += -4; \ 1467 r1 = %[map_array_48b] ll; \ 1468 call %[bpf_map_lookup_elem]; \ 1469 if r0 != 0 goto l0_%=; \ 1470 exit; \ 1471 l0_%=: r9 = r0; \ 1472 r0 = 0; \ 1473 r1 = %[map_ringbuf] ll; \ 1474 r2 = 8; \ 1475 r3 = 0; \ 1476 call %[bpf_ringbuf_reserve]; \ 1477 if r0 != 0 goto l1_%=; \ 1478 exit; \ 1479 l1_%=: r8 = r0; \ 1480 r1 = r8; \ 1481 r2 = 0; \ 1482 call %[bpf_ringbuf_discard]; \ 1483 r0 = 0; \ 1484 *(u64*)(r9 + 0) = r8; \ 1485 exit; \ 1486 " : 1487 : __imm(bpf_map_lookup_elem), 1488 __imm(bpf_ringbuf_discard), 1489 __imm(bpf_ringbuf_reserve), 1490 __imm_addr(map_array_48b), 1491 __imm_addr(map_ringbuf) 1492 : __clobber_all); 1493 } 1494 1495 char _license[] SEC("license") = "GPL"; 1496
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.