123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266 |
- // SPDX-License-Identifier: GPL-2.0
- /* GPLv2, Copyright(c) 2017 Jesper Dangaard Brouer, Red Hat, Inc. */
- #include "xdp_sample.bpf.h"
- #include <bpf/bpf_tracing.h>
- #include <bpf/bpf_core_read.h>
- #include <bpf/bpf_helpers.h>
- array_map rx_cnt SEC(".maps");
- array_map redir_err_cnt SEC(".maps");
- array_map cpumap_enqueue_cnt SEC(".maps");
- array_map cpumap_kthread_cnt SEC(".maps");
- array_map exception_cnt SEC(".maps");
- array_map devmap_xmit_cnt SEC(".maps");
- struct {
- __uint(type, BPF_MAP_TYPE_PERCPU_HASH);
- __uint(max_entries, 32 * 32);
- __type(key, u64);
- __type(value, struct datarec);
- } devmap_xmit_cnt_multi SEC(".maps");
- const volatile int nr_cpus = 0;
- /* These can be set before loading so that redundant comparisons can be DCE'd by
- * the verifier, and only actual matches are tried after loading tp_btf program.
- * This allows sample to filter tracepoint stats based on net_device.
- */
- const volatile int from_match[32] = {};
- const volatile int to_match[32] = {};
- int cpumap_map_id = 0;
- /* Find if b is part of set a, but if a is empty set then evaluate to true */
- #define IN_SET(a, b) \
- ({ \
- bool __res = !(a)[0]; \
- for (int i = 0; i < ARRAY_SIZE(a) && (a)[i]; i++) { \
- __res = (a)[i] == (b); \
- if (__res) \
- break; \
- } \
- __res; \
- })
- static __always_inline __u32 xdp_get_err_key(int err)
- {
- switch (err) {
- case 0:
- return 0;
- case -EINVAL:
- return 2;
- case -ENETDOWN:
- return 3;
- case -EMSGSIZE:
- return 4;
- case -EOPNOTSUPP:
- return 5;
- case -ENOSPC:
- return 6;
- default:
- return 1;
- }
- }
- static __always_inline int xdp_redirect_collect_stat(int from, int err)
- {
- u32 cpu = bpf_get_smp_processor_id();
- u32 key = XDP_REDIRECT_ERROR;
- struct datarec *rec;
- u32 idx;
- if (!IN_SET(from_match, from))
- return 0;
- key = xdp_get_err_key(err);
- idx = key * nr_cpus + cpu;
- rec = bpf_map_lookup_elem(&redir_err_cnt, &idx);
- if (!rec)
- return 0;
- if (key)
- NO_TEAR_INC(rec->dropped);
- else
- NO_TEAR_INC(rec->processed);
- return 0; /* Indicate event was filtered (no further processing)*/
- /*
- * Returning 1 here would allow e.g. a perf-record tracepoint
- * to see and record these events, but it doesn't work well
- * in-practice as stopping perf-record also unload this
- * bpf_prog. Plus, there is additional overhead of doing so.
- */
- }
- SEC("tp_btf/xdp_redirect_err")
- int BPF_PROG(tp_xdp_redirect_err, const struct net_device *dev,
- const struct bpf_prog *xdp, const void *tgt, int err,
- const struct bpf_map *map, u32 index)
- {
- return xdp_redirect_collect_stat(dev->ifindex, err);
- }
- SEC("tp_btf/xdp_redirect_map_err")
- int BPF_PROG(tp_xdp_redirect_map_err, const struct net_device *dev,
- const struct bpf_prog *xdp, const void *tgt, int err,
- const struct bpf_map *map, u32 index)
- {
- return xdp_redirect_collect_stat(dev->ifindex, err);
- }
- SEC("tp_btf/xdp_redirect")
- int BPF_PROG(tp_xdp_redirect, const struct net_device *dev,
- const struct bpf_prog *xdp, const void *tgt, int err,
- const struct bpf_map *map, u32 index)
- {
- return xdp_redirect_collect_stat(dev->ifindex, err);
- }
- SEC("tp_btf/xdp_redirect_map")
- int BPF_PROG(tp_xdp_redirect_map, const struct net_device *dev,
- const struct bpf_prog *xdp, const void *tgt, int err,
- const struct bpf_map *map, u32 index)
- {
- return xdp_redirect_collect_stat(dev->ifindex, err);
- }
- SEC("tp_btf/xdp_cpumap_enqueue")
- int BPF_PROG(tp_xdp_cpumap_enqueue, int map_id, unsigned int processed,
- unsigned int drops, int to_cpu)
- {
- u32 cpu = bpf_get_smp_processor_id();
- struct datarec *rec;
- u32 idx;
- if (cpumap_map_id && cpumap_map_id != map_id)
- return 0;
- idx = to_cpu * nr_cpus + cpu;
- rec = bpf_map_lookup_elem(&cpumap_enqueue_cnt, &idx);
- if (!rec)
- return 0;
- NO_TEAR_ADD(rec->processed, processed);
- NO_TEAR_ADD(rec->dropped, drops);
- /* Record bulk events, then userspace can calc average bulk size */
- if (processed > 0)
- NO_TEAR_INC(rec->issue);
- /* Inception: It's possible to detect overload situations, via
- * this tracepoint. This can be used for creating a feedback
- * loop to XDP, which can take appropriate actions to mitigate
- * this overload situation.
- */
- return 0;
- }
- SEC("tp_btf/xdp_cpumap_kthread")
- int BPF_PROG(tp_xdp_cpumap_kthread, int map_id, unsigned int processed,
- unsigned int drops, int sched, struct xdp_cpumap_stats *xdp_stats)
- {
- struct datarec *rec;
- u32 cpu;
- if (cpumap_map_id && cpumap_map_id != map_id)
- return 0;
- cpu = bpf_get_smp_processor_id();
- rec = bpf_map_lookup_elem(&cpumap_kthread_cnt, &cpu);
- if (!rec)
- return 0;
- NO_TEAR_ADD(rec->processed, processed);
- NO_TEAR_ADD(rec->dropped, drops);
- NO_TEAR_ADD(rec->xdp_pass, xdp_stats->pass);
- NO_TEAR_ADD(rec->xdp_drop, xdp_stats->drop);
- NO_TEAR_ADD(rec->xdp_redirect, xdp_stats->redirect);
- /* Count times kthread yielded CPU via schedule call */
- if (sched)
- NO_TEAR_INC(rec->issue);
- return 0;
- }
- SEC("tp_btf/xdp_exception")
- int BPF_PROG(tp_xdp_exception, const struct net_device *dev,
- const struct bpf_prog *xdp, u32 act)
- {
- u32 cpu = bpf_get_smp_processor_id();
- struct datarec *rec;
- u32 key = act, idx;
- if (!IN_SET(from_match, dev->ifindex))
- return 0;
- if (!IN_SET(to_match, dev->ifindex))
- return 0;
- if (key > XDP_REDIRECT)
- key = XDP_REDIRECT + 1;
- idx = key * nr_cpus + cpu;
- rec = bpf_map_lookup_elem(&exception_cnt, &idx);
- if (!rec)
- return 0;
- NO_TEAR_INC(rec->dropped);
- return 0;
- }
- SEC("tp_btf/xdp_devmap_xmit")
- int BPF_PROG(tp_xdp_devmap_xmit, const struct net_device *from_dev,
- const struct net_device *to_dev, int sent, int drops, int err)
- {
- struct datarec *rec;
- int idx_in, idx_out;
- u32 cpu;
- idx_in = from_dev->ifindex;
- idx_out = to_dev->ifindex;
- if (!IN_SET(from_match, idx_in))
- return 0;
- if (!IN_SET(to_match, idx_out))
- return 0;
- cpu = bpf_get_smp_processor_id();
- rec = bpf_map_lookup_elem(&devmap_xmit_cnt, &cpu);
- if (!rec)
- return 0;
- NO_TEAR_ADD(rec->processed, sent);
- NO_TEAR_ADD(rec->dropped, drops);
- /* Record bulk events, then userspace can calc average bulk size */
- NO_TEAR_INC(rec->info);
- /* Record error cases, where no frame were sent */
- /* Catch API error of drv ndo_xdp_xmit sent more than count */
- if (err || drops < 0)
- NO_TEAR_INC(rec->issue);
- return 0;
- }
- SEC("tp_btf/xdp_devmap_xmit")
- int BPF_PROG(tp_xdp_devmap_xmit_multi, const struct net_device *from_dev,
- const struct net_device *to_dev, int sent, int drops, int err)
- {
- struct datarec empty = {};
- struct datarec *rec;
- int idx_in, idx_out;
- u64 idx;
- idx_in = from_dev->ifindex;
- idx_out = to_dev->ifindex;
- idx = idx_in;
- idx = idx << 32 | idx_out;
- if (!IN_SET(from_match, idx_in))
- return 0;
- if (!IN_SET(to_match, idx_out))
- return 0;
- bpf_map_update_elem(&devmap_xmit_cnt_multi, &idx, &empty, BPF_NOEXIST);
- rec = bpf_map_lookup_elem(&devmap_xmit_cnt_multi, &idx);
- if (!rec)
- return 0;
- NO_TEAR_ADD(rec->processed, sent);
- NO_TEAR_ADD(rec->dropped, drops);
- NO_TEAR_INC(rec->info);
- if (err || drops < 0)
- NO_TEAR_INC(rec->issue);
- return 0;
- }
|