| /OK3568_Linux_fs/kernel/tools/testing/selftests/bpf/prog_tests/ |
| H A D | bpf_iter.c | 31 struct bpf_iter_test_kern3 *skel; in test_btf_id_or_null() local 33 skel = bpf_iter_test_kern3__open_and_load(); in test_btf_id_or_null() 34 if (CHECK(skel, "bpf_iter_test_kern3__open_and_load", in test_btf_id_or_null() 36 bpf_iter_test_kern3__destroy(skel); in test_btf_id_or_null() 68 struct bpf_iter_ipv6_route *skel; in test_ipv6_route() local 70 skel = bpf_iter_ipv6_route__open_and_load(); in test_ipv6_route() 71 if (CHECK(!skel, "bpf_iter_ipv6_route__open_and_load", in test_ipv6_route() 75 do_dummy_read(skel->progs.dump_ipv6_route); in test_ipv6_route() 77 bpf_iter_ipv6_route__destroy(skel); in test_ipv6_route() 82 struct bpf_iter_netlink *skel; in test_netlink() local [all …]
|
| H A D | ringbuf.c | 61 static struct test_ringbuf *skel; variable 66 skel->bss->dropped = 0; in trigger_samples() 67 skel->bss->total = 0; in trigger_samples() 68 skel->bss->discarded = 0; in trigger_samples() 71 skel->bss->value = 333; in trigger_samples() 73 skel->bss->value = 777; in trigger_samples() 91 skel = test_ringbuf__open_and_load(); in test_ringbuf() 92 if (CHECK(!skel, "skel_open_load", "skeleton open&load failed\n")) in test_ringbuf() 96 skel->bss->pid = getpid(); in test_ringbuf() 98 ringbuf = ring_buffer__new(bpf_map__fd(skel->maps.ringbuf), in test_ringbuf() [all …]
|
| H A D | btf_skc_cls_ingress.c | 19 static struct test_btf_skc_cls_ingress *skel; variable 77 memset(&skel->bss->srv_sa6, 0, sizeof(skel->bss->srv_sa6)); in reset_test() 78 skel->bss->listen_tp_sport = 0; in reset_test() 79 skel->bss->req_sk_sport = 0; in reset_test() 80 skel->bss->recv_cookie = 0; in reset_test() 81 skel->bss->gen_cookie = 0; in reset_test() 82 skel->bss->linum = 0; in reset_test() 87 if (skel->bss->linum) in print_err_line() 88 printf("bpf prog error at line %u\n", skel->bss->linum); in print_err_line() 108 memcpy(&skel->bss->srv_sa6, &srv_sa6, sizeof(srv_sa6)); in test_conn() [all …]
|
| H A D | sk_lookup.c | 516 static void query_lookup_prog(struct test_sk_lookup *skel) in query_lookup_prog() argument 533 link[0] = attach_lookup_prog(skel->progs.lookup_pass); in query_lookup_prog() 536 link[1] = attach_lookup_prog(skel->progs.lookup_pass); in query_lookup_prog() 539 link[2] = attach_lookup_prog(skel->progs.lookup_drop); in query_lookup_prog() 672 static void test_redirect_lookup(struct test_sk_lookup *skel) in test_redirect_lookup() argument 677 .lookup_prog = skel->progs.redir_port, in test_redirect_lookup() 678 .sock_map = skel->maps.redir_map, in test_redirect_lookup() 685 .lookup_prog = skel->progs.redir_ip4, in test_redirect_lookup() 686 .sock_map = skel->maps.redir_map, in test_redirect_lookup() 693 .lookup_prog = skel->progs.select_sock_a, in test_redirect_lookup() [all …]
|
| H A D | attach_probe.c | 31 struct test_attach_probe* skel; in test_attach_probe() local 41 skel = test_attach_probe__open_and_load(); in test_attach_probe() 42 if (CHECK(!skel, "skel_open", "failed to open skeleton\n")) in test_attach_probe() 44 if (CHECK(!skel->bss, "check_bss", ".bss wasn't mmap()-ed\n")) in test_attach_probe() 47 kprobe_link = bpf_program__attach_kprobe(skel->progs.handle_kprobe, in test_attach_probe() 53 skel->links.handle_kprobe = kprobe_link; in test_attach_probe() 55 kretprobe_link = bpf_program__attach_kprobe(skel->progs.handle_kretprobe, in test_attach_probe() 61 skel->links.handle_kretprobe = kretprobe_link; in test_attach_probe() 63 uprobe_link = bpf_program__attach_uprobe(skel->progs.handle_uprobe, in test_attach_probe() 71 skel->links.handle_uprobe = uprobe_link; in test_attach_probe() [all …]
|
| H A D | ringbuf_multi.c | 43 struct test_ringbuf_multi *skel; in test_ringbuf_multi() local 47 skel = test_ringbuf_multi__open_and_load(); in test_ringbuf_multi() 48 if (CHECK(!skel, "skel_open_load", "skeleton open&load failed\n")) in test_ringbuf_multi() 52 skel->bss->pid = getpid(); in test_ringbuf_multi() 54 ringbuf = ring_buffer__new(bpf_map__fd(skel->maps.ringbuf1), in test_ringbuf_multi() 59 err = ring_buffer__add(ringbuf, bpf_map__fd(skel->maps.ringbuf2), in test_ringbuf_multi() 64 err = test_ringbuf_multi__attach(skel); in test_ringbuf_multi() 69 skel->bss->target_ring = 0; in test_ringbuf_multi() 70 skel->bss->value = 333; in test_ringbuf_multi() 74 skel->bss->target_ring = 1; in test_ringbuf_multi() [all …]
|
| H A D | btf_map_in_map.c | 60 struct test_btf_map_in_map *skel; in test_lookup_update() local 63 skel = test_btf_map_in_map__open_and_load(); in test_lookup_update() 64 if (CHECK(!skel, "skel_open", "failed to open&load skeleton\n")) in test_lookup_update() 67 err = test_btf_map_in_map__attach(skel); in test_lookup_update() 71 map1_fd = bpf_map__fd(skel->maps.inner_map1); in test_lookup_update() 72 map2_fd = bpf_map__fd(skel->maps.inner_map2); in test_lookup_update() 73 map3_fd = bpf_map__fd(skel->maps.inner_map3); in test_lookup_update() 74 map4_fd = bpf_map__fd(skel->maps.inner_map4); in test_lookup_update() 75 map5_fd = bpf_map__fd(skel->maps.inner_map5); in test_lookup_update() 76 outer_arr_dyn_fd = bpf_map__fd(skel->maps.outer_arr_dyn); in test_lookup_update() [all …]
|
| H A D | map_init.c | 41 struct test_map_init *skel; in setup() local 44 skel = test_map_init__open(); in setup() 45 if (!ASSERT_OK_PTR(skel, "skel_open")) in setup() 48 err = bpf_map__set_type(skel->maps.hashmap1, map_type); in setup() 52 err = bpf_map__set_max_entries(skel->maps.hashmap1, map_sz); in setup() 56 err = test_map_init__load(skel); in setup() 60 *map_fd = bpf_map__fd(skel->maps.hashmap1); in setup() 68 return skel; in setup() 73 test_map_init__destroy(skel); in setup() 78 static int prog_run_insert_elem(struct test_map_init *skel, map_key_t key, in prog_run_insert_elem() argument [all …]
|
| H A D | udp_limit.c | 12 struct udp_limit *skel; in test_udp_limit() local 20 skel = udp_limit__open_and_load(); in test_udp_limit() 21 if (CHECK(!skel, "skel-load", "errno %d", errno)) in test_udp_limit() 24 skel->links.sock = bpf_program__attach_cgroup(skel->progs.sock, cgroup_fd); in test_udp_limit() 25 skel->links.sock_release = bpf_program__attach_cgroup(skel->progs.sock_release, cgroup_fd); in test_udp_limit() 26 if (CHECK(IS_ERR(skel->links.sock) || IS_ERR(skel->links.sock_release), in test_udp_limit() 28 PTR_ERR(skel->links.sock), in test_udp_limit() 29 PTR_ERR(skel->links.sock_release))) in test_udp_limit() 58 if (CHECK(skel->bss->invocations != 4, "bss-invocations", in test_udp_limit() 59 "invocations=%d", skel->bss->invocations)) in test_udp_limit() [all …]
|
| H A D | probe_read_user_str.c | 9 static int test_one_str(struct test_probe_read_user_str *skel, const char *str, in test_one_str() argument 20 skel->bss->user_ptr = buf; in test_one_str() 26 if (CHECK(skel->bss->ret < 0, "prog_ret", "prog returned: %ld\n", in test_one_str() 27 skel->bss->ret)) in test_one_str() 31 err = memcmp(skel->bss->buf, str, len); in test_one_str() 37 err = memcmp(skel->bss->buf + len, buf, sizeof(buf) - len); in test_one_str() 46 struct test_probe_read_user_str *skel; in test_probe_read_user_str() local 49 skel = test_probe_read_user_str__open_and_load(); in test_probe_read_user_str() 50 if (CHECK(!skel, "test_probe_read_user_str__open_and_load", in test_probe_read_user_str() 55 skel->bss->pid = getpid(); in test_probe_read_user_str() [all …]
|
| H A D | sockmap_basic.c | 111 struct test_skmsg_load_helpers *skel; in test_skmsg_helpers() local 114 skel = test_skmsg_load_helpers__open_and_load(); in test_skmsg_helpers() 115 if (CHECK_FAIL(!skel)) { in test_skmsg_helpers() 120 verdict = bpf_program__fd(skel->progs.prog_msg_verdict); in test_skmsg_helpers() 121 map = bpf_map__fd(skel->maps.sock_map); in test_skmsg_helpers() 135 test_skmsg_load_helpers__destroy(skel); in test_skmsg_helpers() 142 struct test_sockmap_update *skel; in test_sockmap_update() local 152 skel = test_sockmap_update__open_and_load(); in test_sockmap_update() 153 if (CHECK(!skel, "open_and_load", "cannot load skeleton\n")) in test_sockmap_update() 156 prog = bpf_program__fd(skel->progs.copy_sock_map); in test_sockmap_update() [all …]
|
| H A D | get_stackid_cannot_attach.c | 21 struct test_stacktrace_build_id *skel; in test_get_stackid_cannot_attach() local 25 skel = test_stacktrace_build_id__open(); in test_get_stackid_cannot_attach() 26 if (CHECK(!skel, "skel_open", "skeleton open failed\n")) in test_get_stackid_cannot_attach() 30 bpf_program__set_perf_event(skel->progs.oncpu); in test_get_stackid_cannot_attach() 32 err = test_stacktrace_build_id__load(skel); in test_get_stackid_cannot_attach() 49 skel->links.oncpu = bpf_program__attach_perf_event(skel->progs.oncpu, in test_get_stackid_cannot_attach() 51 CHECK(!IS_ERR(skel->links.oncpu), "attach_perf_event_no_callchain", in test_get_stackid_cannot_attach() 66 skel->links.oncpu = bpf_program__attach_perf_event(skel->progs.oncpu, in test_get_stackid_cannot_attach() 68 CHECK(IS_ERR(skel->links.oncpu), "attach_perf_event_callchain", in test_get_stackid_cannot_attach() 69 "err: %ld\n", PTR_ERR(skel->links.oncpu)); in test_get_stackid_cannot_attach() [all …]
|
| H A D | subprogs.c | 12 struct test_subprogs *skel; in test_subprogs() local 16 skel = test_subprogs__open_and_load(); in test_subprogs() 17 if (CHECK(!skel, "skel_open", "failed to open skeleton\n")) in test_subprogs() 20 err = test_subprogs__attach(skel); in test_subprogs() 26 CHECK(skel->bss->res1 != 12, "res1", "got %d, exp %d\n", skel->bss->res1, 12); in test_subprogs() 27 CHECK(skel->bss->res2 != 17, "res2", "got %d, exp %d\n", skel->bss->res2, 17); in test_subprogs() 28 CHECK(skel->bss->res3 != 19, "res3", "got %d, exp %d\n", skel->bss->res3, 19); in test_subprogs() 29 CHECK(skel->bss->res4 != 36, "res4", "got %d, exp %d\n", skel->bss->res4, 36); in test_subprogs() 36 test_subprogs__destroy(skel); in test_subprogs()
|
| H A D | perf_branches.c | 10 static void check_good_sample(struct test_perf_branches *skel) in check_good_sample() argument 12 int written_global = skel->bss->written_global_out; in check_good_sample() 13 int required_size = skel->bss->required_size_out; in check_good_sample() 14 int written_stack = skel->bss->written_stack_out; in check_good_sample() 18 if (CHECK(!skel->bss->valid, "output not valid", in check_good_sample() 41 static void check_bad_sample(struct test_perf_branches *skel) in check_bad_sample() argument 43 int written_global = skel->bss->written_global_out; in check_bad_sample() 44 int required_size = skel->bss->required_size_out; in check_bad_sample() 45 int written_stack = skel->bss->written_stack_out; in check_bad_sample() 48 if (CHECK(!skel->bss->valid, "output not valid", in check_bad_sample() [all …]
|
| H A D | tcp_hdr_options.c | 29 static struct test_tcp_hdr_options *skel; variable 251 if (CHECK(expected_inherit_cb_flags != skel->bss->inherit_cb_flags, in check_hdr_and_close_fds() 253 skel->bss->inherit_cb_flags, expected_inherit_cb_flags)) in check_hdr_and_close_fds() 264 if (check_hdr_opt(&exp_passive_estab_in, &skel->bss->passive_estab_in, in check_hdr_and_close_fds() 268 if (check_hdr_opt(&exp_active_estab_in, &skel->bss->active_estab_in, in check_hdr_and_close_fds() 272 if (check_hdr_opt(&exp_passive_fin_in, &skel->bss->passive_fin_in, in check_hdr_and_close_fds() 276 check_hdr_opt(&exp_active_fin_in, &skel->bss->active_fin_in, in check_hdr_and_close_fds() 286 skel->bss->active_syn_out = exp_passive_estab_in; in prepare_out() 287 skel->bss->passive_synack_out = exp_active_estab_in; in prepare_out() 289 skel->bss->active_fin_out = exp_passive_fin_in; in prepare_out() [all …]
|
| H A D | autoload.c | 11 struct test_autoload* skel; in test_autoload() local 13 skel = test_autoload__open_and_load(); in test_autoload() 15 if (CHECK(skel, "skel_open_and_load", "unexpected success\n")) in test_autoload() 18 skel = test_autoload__open(); in test_autoload() 19 if (CHECK(!skel, "skel_open", "failed to open skeleton\n")) in test_autoload() 23 bpf_program__set_autoload(skel->progs.prog3, false); in test_autoload() 25 err = test_autoload__load(skel); in test_autoload() 29 err = test_autoload__attach(skel); in test_autoload() 35 CHECK(!skel->bss->prog1_called, "prog1", "not called\n"); in test_autoload() 36 CHECK(!skel->bss->prog2_called, "prog2", "not called\n"); in test_autoload() [all …]
|
| H A D | core_autosize.c | 56 struct test_core_autosize* skel = NULL; in test_core_autosize() local 128 skel = test_core_autosize__open(); in test_core_autosize() 129 if (!ASSERT_OK_PTR(skel, "skel_open")) in test_core_autosize() 133 prog = bpf_object__find_program_by_name(skel->obj, "handle_signed"); in test_core_autosize() 138 load_attr.obj = skel->obj; in test_core_autosize() 144 prog = bpf_object__find_program_by_name(skel->obj, "handle_samesize"); in test_core_autosize() 147 skel->links.handle_samesize = bpf_program__attach(prog); in test_core_autosize() 148 if (!ASSERT_OK_PTR(skel->links.handle_samesize, "prog_attach")) in test_core_autosize() 151 prog = bpf_object__find_program_by_name(skel->obj, "handle_downsize"); in test_core_autosize() 154 skel->links.handle_downsize = bpf_program__attach(prog); in test_core_autosize() [all …]
|
| H A D | test_lsm.c | 57 struct lsm *skel = NULL; in test_test_lsm() local 61 skel = lsm__open_and_load(); in test_test_lsm() 62 if (CHECK(!skel, "skel_load", "lsm skeleton failed\n")) in test_test_lsm() 65 err = lsm__attach(skel); in test_test_lsm() 69 err = exec_cmd(&skel->bss->monitored_pid); in test_test_lsm() 73 CHECK(skel->bss->bprm_count != 1, "bprm_count", "bprm_count = %d\n", in test_test_lsm() 74 skel->bss->bprm_count); in test_test_lsm() 76 skel->bss->monitored_pid = getpid(); in test_test_lsm() 83 CHECK(skel->bss->mprotect_count != 1, "mprotect_count", in test_test_lsm() 84 "mprotect_count = %d\n", skel->bss->mprotect_count); in test_test_lsm() [all …]
|
| H A D | perf_event_stackmap.c | 69 struct perf_event_stackmap *skel; in test_perf_event_stackmap() local 74 skel = perf_event_stackmap__open(); in test_perf_event_stackmap() 76 if (CHECK(!skel, "skel_open", "skeleton open failed\n")) in test_perf_event_stackmap() 79 err = perf_event_stackmap__load(skel); in test_perf_event_stackmap() 98 skel->links.oncpu = bpf_program__attach_perf_event(skel->progs.oncpu, in test_perf_event_stackmap() 100 if (CHECK(IS_ERR(skel->links.oncpu), "attach_perf_event", in test_perf_event_stackmap() 101 "err %ld\n", PTR_ERR(skel->links.oncpu))) { in test_perf_event_stackmap() 109 CHECK(skel->data->stackid_kernel != 2, "get_stackid_kernel", "failed\n"); in test_perf_event_stackmap() 110 CHECK(skel->data->stackid_user != 2, "get_stackid_user", "failed\n"); in test_perf_event_stackmap() 111 CHECK(skel->data->stack_kernel != 2, "get_stack_kernel", "failed\n"); in test_perf_event_stackmap() [all …]
|
| H A D | xdp_devmap_attach.c | 13 struct test_xdp_with_devmap_helpers *skel; in test_xdp_with_devmap_helpers() local 23 skel = test_xdp_with_devmap_helpers__open_and_load(); in test_xdp_with_devmap_helpers() 24 if (CHECK_FAIL(!skel)) { in test_xdp_with_devmap_helpers() 32 dm_fd = bpf_program__fd(skel->progs.xdp_redir_prog); in test_xdp_with_devmap_helpers() 37 dm_fd = bpf_program__fd(skel->progs.xdp_dummy_dm); in test_xdp_with_devmap_helpers() 38 map_fd = bpf_map__fd(skel->maps.dm_ports); in test_xdp_with_devmap_helpers() 59 val.bpf_prog.fd = bpf_program__fd(skel->progs.xdp_dummy_prog); in test_xdp_with_devmap_helpers() 65 test_xdp_with_devmap_helpers__destroy(skel); in test_xdp_with_devmap_helpers() 70 struct test_xdp_devmap_helpers *skel; in test_neg_xdp_devmap_helpers() local 73 skel = test_xdp_devmap_helpers__open_and_load(); in test_neg_xdp_devmap_helpers() [all …]
|
| H A D | stacktrace_build_id_nmi.c | 23 struct test_stacktrace_build_id *skel; in test_stacktrace_build_id_nmi() local 40 skel = test_stacktrace_build_id__open(); in test_stacktrace_build_id_nmi() 41 if (CHECK(!skel, "skel_open", "skeleton open failed\n")) in test_stacktrace_build_id_nmi() 45 bpf_program__set_perf_event(skel->progs.oncpu); in test_stacktrace_build_id_nmi() 47 err = test_stacktrace_build_id__load(skel); in test_stacktrace_build_id_nmi() 63 skel->links.oncpu = bpf_program__attach_perf_event(skel->progs.oncpu, in test_stacktrace_build_id_nmi() 65 if (CHECK(IS_ERR(skel->links.oncpu), "attach_perf_event", in test_stacktrace_build_id_nmi() 66 "err %ld\n", PTR_ERR(skel->links.oncpu))) { in test_stacktrace_build_id_nmi() 72 control_map_fd = bpf_map__fd(skel->maps.control_map); in test_stacktrace_build_id_nmi() 73 stackid_hmap_fd = bpf_map__fd(skel->maps.stackid_hmap); in test_stacktrace_build_id_nmi() [all …]
|
| H A D | cgroup_link.c | 11 static struct test_cgroup_link *skel = NULL; variable 15 skel->bss->calls = 0; in ping_and_check() 16 skel->bss->alt_calls = 0; in ping_and_check() 18 if (CHECK(skel->bss->calls != exp_calls, "call_cnt", in ping_and_check() 19 "exp %d, got %d\n", exp_calls, skel->bss->calls)) in ping_and_check() 21 if (CHECK(skel->bss->alt_calls != exp_alt_calls, "alt_call_cnt", in ping_and_check() 22 "exp %d, got %d\n", exp_alt_calls, skel->bss->alt_calls)) in ping_and_check() 46 skel = test_cgroup_link__open_and_load(); in test_cgroup_link() 47 if (CHECK(!skel, "skel_open_load", "failed to open/load skeleton\n")) in test_cgroup_link() 49 prog_fd = bpf_program__fd(skel->progs.egress); in test_cgroup_link() [all …]
|
| H A D | modify_return.c | 16 struct modify_return *skel = NULL; in run_test() local 22 skel = modify_return__open_and_load(); in run_test() 23 if (CHECK(!skel, "skel_load", "modify_return skeleton failed\n")) in run_test() 26 err = modify_return__attach(skel); in run_test() 30 skel->bss->input_retval = input_retval; in run_test() 31 prog_fd = bpf_program__fd(skel->progs.fmod_ret_test); in run_test() 45 CHECK(skel->bss->fentry_result != 1, "modify_return", in run_test() 47 CHECK(skel->bss->fexit_result != 1, "modify_return", in run_test() 49 CHECK(skel->bss->fmod_ret_result != 1, "modify_return", in run_test() 53 modify_return__destroy(skel); in run_test()
|
| /OK3568_Linux_fs/kernel/samples/v4l/ |
| H A D | v4l2-pci-skeleton.c | 133 struct skeleton *skel = dev_id; in skeleton_irq() local 140 spin_lock(&skel->qlock); in skeleton_irq() 142 spin_unlock(&skel->qlock); in skeleton_irq() 144 new_buf->vb.sequence = skel->sequence++; in skeleton_irq() 145 new_buf->vb.field = skel->field; in skeleton_irq() 146 if (skel->format.field == V4L2_FIELD_ALTERNATE) { in skeleton_irq() 147 if (skel->field == V4L2_FIELD_BOTTOM) in skeleton_irq() 148 skel->field = V4L2_FIELD_TOP; in skeleton_irq() 149 else if (skel->field == V4L2_FIELD_TOP) in skeleton_irq() 150 skel->field = V4L2_FIELD_BOTTOM; in skeleton_irq() [all …]
|
| /OK3568_Linux_fs/kernel/tools/testing/selftests/bpf/benchs/ |
| H A D | bench_ringbufs.c | 123 struct ringbuf_bench *skel; member 132 res->drops = atomic_swap(&ctx->skel->bss->dropped, 0); in ringbuf_libbpf_measure() 137 struct ringbuf_bench *skel; in ringbuf_setup_skeleton() local 141 skel = ringbuf_bench__open(); in ringbuf_setup_skeleton() 142 if (!skel) { in ringbuf_setup_skeleton() 147 skel->rodata->batch_cnt = args.batch_cnt; in ringbuf_setup_skeleton() 148 skel->rodata->use_output = args.ringbuf_use_output ? 1 : 0; in ringbuf_setup_skeleton() 152 skel->rodata->wakeup_data_size = args.sample_rate * 16; in ringbuf_setup_skeleton() 154 bpf_map__resize(skel->maps.ringbuf, args.ringbuf_sz); in ringbuf_setup_skeleton() 156 if (ringbuf_bench__load(skel)) { in ringbuf_setup_skeleton() [all …]
|