Back to home page

OSCL-LXR

 
 

    


0001 // SPDX-License-Identifier: GPL-2.0
0002 #include <test_progs.h>
0003 #include <network_helpers.h>
0004 #include <net/if.h>
0005 #include "test_xdp.skel.h"
0006 #include "test_xdp_bpf2bpf.skel.h"
0007 
0008 struct meta {
0009     int ifindex;
0010     int pkt_len;
0011 };
0012 
0013 struct test_ctx_s {
0014     bool passed;
0015     int pkt_size;
0016 };
0017 
0018 struct test_ctx_s test_ctx;
0019 
0020 static void on_sample(void *ctx, int cpu, void *data, __u32 size)
0021 {
0022     struct meta *meta = (struct meta *)data;
0023     struct ipv4_packet *trace_pkt_v4 = data + sizeof(*meta);
0024     unsigned char *raw_pkt = data + sizeof(*meta);
0025     struct test_ctx_s *tst_ctx = ctx;
0026 
0027     ASSERT_GE(size, sizeof(pkt_v4) + sizeof(*meta), "check_size");
0028     ASSERT_EQ(meta->ifindex, if_nametoindex("lo"), "check_meta_ifindex");
0029     ASSERT_EQ(meta->pkt_len, tst_ctx->pkt_size, "check_meta_pkt_len");
0030     ASSERT_EQ(memcmp(trace_pkt_v4, &pkt_v4, sizeof(pkt_v4)), 0,
0031           "check_packet_content");
0032 
0033     if (meta->pkt_len > sizeof(pkt_v4)) {
0034         for (int i = 0; i < meta->pkt_len - sizeof(pkt_v4); i++)
0035             ASSERT_EQ(raw_pkt[i + sizeof(pkt_v4)], (unsigned char)i,
0036                   "check_packet_content");
0037     }
0038 
0039     tst_ctx->passed = true;
0040 }
0041 
0042 #define BUF_SZ  9000
0043 
0044 static void run_xdp_bpf2bpf_pkt_size(int pkt_fd, struct perf_buffer *pb,
0045                      struct test_xdp_bpf2bpf *ftrace_skel,
0046                      int pkt_size)
0047 {
0048     __u8 *buf, *buf_in;
0049     int err;
0050     LIBBPF_OPTS(bpf_test_run_opts, topts);
0051 
0052     if (!ASSERT_LE(pkt_size, BUF_SZ, "pkt_size") ||
0053         !ASSERT_GE(pkt_size, sizeof(pkt_v4), "pkt_size"))
0054         return;
0055 
0056     buf_in = malloc(BUF_SZ);
0057     if (!ASSERT_OK_PTR(buf_in, "buf_in malloc()"))
0058         return;
0059 
0060     buf = malloc(BUF_SZ);
0061     if (!ASSERT_OK_PTR(buf, "buf malloc()")) {
0062         free(buf_in);
0063         return;
0064     }
0065 
0066     test_ctx.passed = false;
0067     test_ctx.pkt_size = pkt_size;
0068 
0069     memcpy(buf_in, &pkt_v4, sizeof(pkt_v4));
0070     if (pkt_size > sizeof(pkt_v4)) {
0071         for (int i = 0; i < (pkt_size - sizeof(pkt_v4)); i++)
0072             buf_in[i + sizeof(pkt_v4)] = i;
0073     }
0074 
0075     /* Run test program */
0076     topts.data_in = buf_in;
0077     topts.data_size_in = pkt_size;
0078     topts.data_out = buf;
0079     topts.data_size_out = BUF_SZ;
0080 
0081     err = bpf_prog_test_run_opts(pkt_fd, &topts);
0082 
0083     ASSERT_OK(err, "ipv4");
0084     ASSERT_EQ(topts.retval, XDP_PASS, "ipv4 retval");
0085     ASSERT_EQ(topts.data_size_out, pkt_size, "ipv4 size");
0086 
0087     /* Make sure bpf_xdp_output() was triggered and it sent the expected
0088      * data to the perf ring buffer.
0089      */
0090     err = perf_buffer__poll(pb, 100);
0091 
0092     ASSERT_GE(err, 0, "perf_buffer__poll");
0093     ASSERT_TRUE(test_ctx.passed, "test passed");
0094     /* Verify test results */
0095     ASSERT_EQ(ftrace_skel->bss->test_result_fentry, if_nametoindex("lo"),
0096           "fentry result");
0097     ASSERT_EQ(ftrace_skel->bss->test_result_fexit, XDP_PASS, "fexit result");
0098 
0099     free(buf);
0100     free(buf_in);
0101 }
0102 
0103 void test_xdp_bpf2bpf(void)
0104 {
0105     int err, pkt_fd, map_fd;
0106     int pkt_sizes[] = {sizeof(pkt_v4), 1024, 4100, 8200};
0107     struct iptnl_info value4 = {.family = AF_INET6};
0108     struct test_xdp *pkt_skel = NULL;
0109     struct test_xdp_bpf2bpf *ftrace_skel = NULL;
0110     struct vip key4 = {.protocol = 6, .family = AF_INET};
0111     struct bpf_program *prog;
0112     struct perf_buffer *pb = NULL;
0113 
0114     /* Load XDP program to introspect */
0115     pkt_skel = test_xdp__open_and_load();
0116     if (!ASSERT_OK_PTR(pkt_skel, "test_xdp__open_and_load"))
0117         return;
0118 
0119     pkt_fd = bpf_program__fd(pkt_skel->progs._xdp_tx_iptunnel);
0120 
0121     map_fd = bpf_map__fd(pkt_skel->maps.vip2tnl);
0122     bpf_map_update_elem(map_fd, &key4, &value4, 0);
0123 
0124     /* Load trace program */
0125     ftrace_skel = test_xdp_bpf2bpf__open();
0126     if (!ASSERT_OK_PTR(ftrace_skel, "test_xdp_bpf2bpf__open"))
0127         goto out;
0128 
0129     /* Demonstrate the bpf_program__set_attach_target() API rather than
0130      * the load with options, i.e. opts.attach_prog_fd.
0131      */
0132     prog = ftrace_skel->progs.trace_on_entry;
0133     bpf_program__set_expected_attach_type(prog, BPF_TRACE_FENTRY);
0134     bpf_program__set_attach_target(prog, pkt_fd, "_xdp_tx_iptunnel");
0135 
0136     prog = ftrace_skel->progs.trace_on_exit;
0137     bpf_program__set_expected_attach_type(prog, BPF_TRACE_FEXIT);
0138     bpf_program__set_attach_target(prog, pkt_fd, "_xdp_tx_iptunnel");
0139 
0140     err = test_xdp_bpf2bpf__load(ftrace_skel);
0141     if (!ASSERT_OK(err, "test_xdp_bpf2bpf__load"))
0142         goto out;
0143 
0144     err = test_xdp_bpf2bpf__attach(ftrace_skel);
0145     if (!ASSERT_OK(err, "test_xdp_bpf2bpf__attach"))
0146         goto out;
0147 
0148     /* Set up perf buffer */
0149     pb = perf_buffer__new(bpf_map__fd(ftrace_skel->maps.perf_buf_map), 8,
0150                   on_sample, NULL, &test_ctx, NULL);
0151     if (!ASSERT_OK_PTR(pb, "perf_buf__new"))
0152         goto out;
0153 
0154     for (int i = 0; i < ARRAY_SIZE(pkt_sizes); i++)
0155         run_xdp_bpf2bpf_pkt_size(pkt_fd, pb, ftrace_skel,
0156                      pkt_sizes[i]);
0157 out:
0158     perf_buffer__free(pb);
0159     test_xdp__destroy(pkt_skel);
0160     test_xdp_bpf2bpf__destroy(ftrace_skel);
0161 }