123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292 |
- // SPDX-License-Identifier: GPL-2.0-only
- /* Copyright (c) 2019, Microsoft Corporation.
- *
- * Author:
- * Haiyang Zhang <[email protected]>
- */
- #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
- #include <linux/netdevice.h>
- #include <linux/etherdevice.h>
- #include <linux/ethtool.h>
- #include <linux/netpoll.h>
- #include <linux/bpf.h>
- #include <linux/bpf_trace.h>
- #include <linux/kernel.h>
- #include <net/xdp.h>
- #include <linux/mutex.h>
- #include <linux/rtnetlink.h>
- #include "hyperv_net.h"
- u32 netvsc_run_xdp(struct net_device *ndev, struct netvsc_channel *nvchan,
- struct xdp_buff *xdp)
- {
- struct netvsc_stats_rx *rx_stats = &nvchan->rx_stats;
- void *data = nvchan->rsc.data[0];
- u32 len = nvchan->rsc.len[0];
- struct page *page = NULL;
- struct bpf_prog *prog;
- u32 act = XDP_PASS;
- bool drop = true;
- xdp->data_hard_start = NULL;
- rcu_read_lock();
- prog = rcu_dereference(nvchan->bpf_prog);
- if (!prog)
- goto out;
- /* Ensure that the below memcpy() won't overflow the page buffer. */
- if (len > ndev->mtu + ETH_HLEN) {
- act = XDP_DROP;
- goto out;
- }
- /* allocate page buffer for data */
- page = alloc_page(GFP_ATOMIC);
- if (!page) {
- act = XDP_DROP;
- goto out;
- }
- xdp_init_buff(xdp, PAGE_SIZE, &nvchan->xdp_rxq);
- xdp_prepare_buff(xdp, page_address(page), NETVSC_XDP_HDRM, len, false);
- memcpy(xdp->data, data, len);
- act = bpf_prog_run_xdp(prog, xdp);
- switch (act) {
- case XDP_PASS:
- case XDP_TX:
- drop = false;
- break;
- case XDP_DROP:
- break;
- case XDP_REDIRECT:
- if (!xdp_do_redirect(ndev, xdp, prog)) {
- nvchan->xdp_flush = true;
- drop = false;
- u64_stats_update_begin(&rx_stats->syncp);
- rx_stats->xdp_redirect++;
- rx_stats->packets++;
- rx_stats->bytes += nvchan->rsc.pktlen;
- u64_stats_update_end(&rx_stats->syncp);
- break;
- } else {
- u64_stats_update_begin(&rx_stats->syncp);
- rx_stats->xdp_drop++;
- u64_stats_update_end(&rx_stats->syncp);
- }
- fallthrough;
- case XDP_ABORTED:
- trace_xdp_exception(ndev, prog, act);
- break;
- default:
- bpf_warn_invalid_xdp_action(ndev, prog, act);
- }
- out:
- rcu_read_unlock();
- if (page && drop) {
- __free_page(page);
- xdp->data_hard_start = NULL;
- }
- return act;
- }
- unsigned int netvsc_xdp_fraglen(unsigned int len)
- {
- return SKB_DATA_ALIGN(len) +
- SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
- }
- struct bpf_prog *netvsc_xdp_get(struct netvsc_device *nvdev)
- {
- return rtnl_dereference(nvdev->chan_table[0].bpf_prog);
- }
- int netvsc_xdp_set(struct net_device *dev, struct bpf_prog *prog,
- struct netlink_ext_ack *extack,
- struct netvsc_device *nvdev)
- {
- struct bpf_prog *old_prog;
- int buf_max, i;
- old_prog = netvsc_xdp_get(nvdev);
- if (!old_prog && !prog)
- return 0;
- buf_max = NETVSC_XDP_HDRM + netvsc_xdp_fraglen(dev->mtu + ETH_HLEN);
- if (prog && buf_max > PAGE_SIZE) {
- netdev_err(dev, "XDP: mtu:%u too large, buf_max:%u\n",
- dev->mtu, buf_max);
- NL_SET_ERR_MSG_MOD(extack, "XDP: mtu too large");
- return -EOPNOTSUPP;
- }
- if (prog && (dev->features & NETIF_F_LRO)) {
- netdev_err(dev, "XDP: not support LRO\n");
- NL_SET_ERR_MSG_MOD(extack, "XDP: not support LRO");
- return -EOPNOTSUPP;
- }
- if (prog)
- bpf_prog_add(prog, nvdev->num_chn - 1);
- for (i = 0; i < nvdev->num_chn; i++)
- rcu_assign_pointer(nvdev->chan_table[i].bpf_prog, prog);
- if (old_prog)
- for (i = 0; i < nvdev->num_chn; i++)
- bpf_prog_put(old_prog);
- return 0;
- }
- int netvsc_vf_setxdp(struct net_device *vf_netdev, struct bpf_prog *prog)
- {
- struct netdev_bpf xdp;
- int ret;
- ASSERT_RTNL();
- if (!vf_netdev)
- return 0;
- if (!vf_netdev->netdev_ops->ndo_bpf)
- return 0;
- memset(&xdp, 0, sizeof(xdp));
- if (prog)
- bpf_prog_inc(prog);
- xdp.command = XDP_SETUP_PROG;
- xdp.prog = prog;
- ret = vf_netdev->netdev_ops->ndo_bpf(vf_netdev, &xdp);
- if (ret && prog)
- bpf_prog_put(prog);
- return ret;
- }
- int netvsc_bpf(struct net_device *dev, struct netdev_bpf *bpf)
- {
- struct net_device_context *ndevctx = netdev_priv(dev);
- struct netvsc_device *nvdev = rtnl_dereference(ndevctx->nvdev);
- struct net_device *vf_netdev = rtnl_dereference(ndevctx->vf_netdev);
- struct netlink_ext_ack *extack = bpf->extack;
- int ret;
- if (!nvdev || nvdev->destroy) {
- return -ENODEV;
- }
- switch (bpf->command) {
- case XDP_SETUP_PROG:
- ret = netvsc_xdp_set(dev, bpf->prog, extack, nvdev);
- if (ret)
- return ret;
- ret = netvsc_vf_setxdp(vf_netdev, bpf->prog);
- if (ret) {
- netdev_err(dev, "vf_setxdp failed:%d\n", ret);
- NL_SET_ERR_MSG_MOD(extack, "vf_setxdp failed");
- netvsc_xdp_set(dev, NULL, extack, nvdev);
- }
- return ret;
- default:
- return -EINVAL;
- }
- }
- static int netvsc_ndoxdp_xmit_fm(struct net_device *ndev,
- struct xdp_frame *frame, u16 q_idx)
- {
- struct sk_buff *skb;
- skb = xdp_build_skb_from_frame(frame, ndev);
- if (unlikely(!skb))
- return -ENOMEM;
- netvsc_get_hash(skb, netdev_priv(ndev));
- skb_record_rx_queue(skb, q_idx);
- netvsc_xdp_xmit(skb, ndev);
- return 0;
- }
- int netvsc_ndoxdp_xmit(struct net_device *ndev, int n,
- struct xdp_frame **frames, u32 flags)
- {
- struct net_device_context *ndev_ctx = netdev_priv(ndev);
- const struct net_device_ops *vf_ops;
- struct netvsc_stats_tx *tx_stats;
- struct netvsc_device *nvsc_dev;
- struct net_device *vf_netdev;
- int i, count = 0;
- u16 q_idx;
- /* Don't transmit if netvsc_device is gone */
- nvsc_dev = rcu_dereference_bh(ndev_ctx->nvdev);
- if (unlikely(!nvsc_dev || nvsc_dev->destroy))
- return 0;
- /* If VF is present and up then redirect packets to it.
- * Skip the VF if it is marked down or has no carrier.
- * If netpoll is in uses, then VF can not be used either.
- */
- vf_netdev = rcu_dereference_bh(ndev_ctx->vf_netdev);
- if (vf_netdev && netif_running(vf_netdev) &&
- netif_carrier_ok(vf_netdev) && !netpoll_tx_running(ndev) &&
- vf_netdev->netdev_ops->ndo_xdp_xmit &&
- ndev_ctx->data_path_is_vf) {
- vf_ops = vf_netdev->netdev_ops;
- return vf_ops->ndo_xdp_xmit(vf_netdev, n, frames, flags);
- }
- q_idx = smp_processor_id() % ndev->real_num_tx_queues;
- for (i = 0; i < n; i++) {
- if (netvsc_ndoxdp_xmit_fm(ndev, frames[i], q_idx))
- break;
- count++;
- }
- tx_stats = &nvsc_dev->chan_table[q_idx].tx_stats;
- u64_stats_update_begin(&tx_stats->syncp);
- tx_stats->xdp_xmit += count;
- u64_stats_update_end(&tx_stats->syncp);
- return count;
- }
|