bareudp.c 20 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805
  1. // SPDX-License-Identifier: GPL-2.0
  2. /* Bareudp: UDP tunnel encasulation for different Payload types like
  3. * MPLS, NSH, IP, etc.
  4. * Copyright (c) 2019 Nokia, Inc.
  5. * Authors: Martin Varghese, <[email protected]>
  6. */
  7. #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  8. #include <linux/kernel.h>
  9. #include <linux/module.h>
  10. #include <linux/etherdevice.h>
  11. #include <linux/hash.h>
  12. #include <net/dst_metadata.h>
  13. #include <net/gro_cells.h>
  14. #include <net/rtnetlink.h>
  15. #include <net/protocol.h>
  16. #include <net/ip6_tunnel.h>
  17. #include <net/ip_tunnels.h>
  18. #include <net/udp_tunnel.h>
  19. #include <net/bareudp.h>
  20. #define BAREUDP_BASE_HLEN sizeof(struct udphdr)
  21. #define BAREUDP_IPV4_HLEN (sizeof(struct iphdr) + \
  22. sizeof(struct udphdr))
  23. #define BAREUDP_IPV6_HLEN (sizeof(struct ipv6hdr) + \
  24. sizeof(struct udphdr))
  25. static bool log_ecn_error = true;
  26. module_param(log_ecn_error, bool, 0644);
  27. MODULE_PARM_DESC(log_ecn_error, "Log packets received with corrupted ECN");
  28. /* per-network namespace private data for this module */
  29. static unsigned int bareudp_net_id;
  30. struct bareudp_net {
  31. struct list_head bareudp_list;
  32. };
  33. struct bareudp_conf {
  34. __be16 ethertype;
  35. __be16 port;
  36. u16 sport_min;
  37. bool multi_proto_mode;
  38. };
  39. /* Pseudo network device */
  40. struct bareudp_dev {
  41. struct net *net; /* netns for packet i/o */
  42. struct net_device *dev; /* netdev for bareudp tunnel */
  43. __be16 ethertype;
  44. __be16 port;
  45. u16 sport_min;
  46. bool multi_proto_mode;
  47. struct socket __rcu *sock;
  48. struct list_head next; /* bareudp node on namespace list */
  49. struct gro_cells gro_cells;
  50. };
  51. static int bareudp_udp_encap_recv(struct sock *sk, struct sk_buff *skb)
  52. {
  53. struct metadata_dst *tun_dst = NULL;
  54. struct bareudp_dev *bareudp;
  55. unsigned short family;
  56. unsigned int len;
  57. __be16 proto;
  58. void *oiph;
  59. int err;
  60. bareudp = rcu_dereference_sk_user_data(sk);
  61. if (!bareudp)
  62. goto drop;
  63. if (skb->protocol == htons(ETH_P_IP))
  64. family = AF_INET;
  65. else
  66. family = AF_INET6;
  67. if (bareudp->ethertype == htons(ETH_P_IP)) {
  68. __u8 ipversion;
  69. if (skb_copy_bits(skb, BAREUDP_BASE_HLEN, &ipversion,
  70. sizeof(ipversion))) {
  71. bareudp->dev->stats.rx_dropped++;
  72. goto drop;
  73. }
  74. ipversion >>= 4;
  75. if (ipversion == 4) {
  76. proto = htons(ETH_P_IP);
  77. } else if (ipversion == 6 && bareudp->multi_proto_mode) {
  78. proto = htons(ETH_P_IPV6);
  79. } else {
  80. bareudp->dev->stats.rx_dropped++;
  81. goto drop;
  82. }
  83. } else if (bareudp->ethertype == htons(ETH_P_MPLS_UC)) {
  84. struct iphdr *tunnel_hdr;
  85. tunnel_hdr = (struct iphdr *)skb_network_header(skb);
  86. if (tunnel_hdr->version == 4) {
  87. if (!ipv4_is_multicast(tunnel_hdr->daddr)) {
  88. proto = bareudp->ethertype;
  89. } else if (bareudp->multi_proto_mode &&
  90. ipv4_is_multicast(tunnel_hdr->daddr)) {
  91. proto = htons(ETH_P_MPLS_MC);
  92. } else {
  93. bareudp->dev->stats.rx_dropped++;
  94. goto drop;
  95. }
  96. } else {
  97. int addr_type;
  98. struct ipv6hdr *tunnel_hdr_v6;
  99. tunnel_hdr_v6 = (struct ipv6hdr *)skb_network_header(skb);
  100. addr_type =
  101. ipv6_addr_type((struct in6_addr *)&tunnel_hdr_v6->daddr);
  102. if (!(addr_type & IPV6_ADDR_MULTICAST)) {
  103. proto = bareudp->ethertype;
  104. } else if (bareudp->multi_proto_mode &&
  105. (addr_type & IPV6_ADDR_MULTICAST)) {
  106. proto = htons(ETH_P_MPLS_MC);
  107. } else {
  108. bareudp->dev->stats.rx_dropped++;
  109. goto drop;
  110. }
  111. }
  112. } else {
  113. proto = bareudp->ethertype;
  114. }
  115. if (iptunnel_pull_header(skb, BAREUDP_BASE_HLEN,
  116. proto,
  117. !net_eq(bareudp->net,
  118. dev_net(bareudp->dev)))) {
  119. bareudp->dev->stats.rx_dropped++;
  120. goto drop;
  121. }
  122. tun_dst = udp_tun_rx_dst(skb, family, TUNNEL_KEY, 0, 0);
  123. if (!tun_dst) {
  124. bareudp->dev->stats.rx_dropped++;
  125. goto drop;
  126. }
  127. skb_dst_set(skb, &tun_dst->dst);
  128. skb->dev = bareudp->dev;
  129. oiph = skb_network_header(skb);
  130. skb_reset_network_header(skb);
  131. skb_reset_mac_header(skb);
  132. if (!ipv6_mod_enabled() || family == AF_INET)
  133. err = IP_ECN_decapsulate(oiph, skb);
  134. else
  135. err = IP6_ECN_decapsulate(oiph, skb);
  136. if (unlikely(err)) {
  137. if (log_ecn_error) {
  138. if (!ipv6_mod_enabled() || family == AF_INET)
  139. net_info_ratelimited("non-ECT from %pI4 "
  140. "with TOS=%#x\n",
  141. &((struct iphdr *)oiph)->saddr,
  142. ((struct iphdr *)oiph)->tos);
  143. else
  144. net_info_ratelimited("non-ECT from %pI6\n",
  145. &((struct ipv6hdr *)oiph)->saddr);
  146. }
  147. if (err > 1) {
  148. ++bareudp->dev->stats.rx_frame_errors;
  149. ++bareudp->dev->stats.rx_errors;
  150. goto drop;
  151. }
  152. }
  153. len = skb->len;
  154. err = gro_cells_receive(&bareudp->gro_cells, skb);
  155. if (likely(err == NET_RX_SUCCESS))
  156. dev_sw_netstats_rx_add(bareudp->dev, len);
  157. return 0;
  158. drop:
  159. /* Consume bad packet */
  160. kfree_skb(skb);
  161. return 0;
  162. }
  163. static int bareudp_err_lookup(struct sock *sk, struct sk_buff *skb)
  164. {
  165. return 0;
  166. }
  167. static int bareudp_init(struct net_device *dev)
  168. {
  169. struct bareudp_dev *bareudp = netdev_priv(dev);
  170. int err;
  171. dev->tstats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
  172. if (!dev->tstats)
  173. return -ENOMEM;
  174. err = gro_cells_init(&bareudp->gro_cells, dev);
  175. if (err) {
  176. free_percpu(dev->tstats);
  177. return err;
  178. }
  179. return 0;
  180. }
  181. static void bareudp_uninit(struct net_device *dev)
  182. {
  183. struct bareudp_dev *bareudp = netdev_priv(dev);
  184. gro_cells_destroy(&bareudp->gro_cells);
  185. free_percpu(dev->tstats);
  186. }
  187. static struct socket *bareudp_create_sock(struct net *net, __be16 port)
  188. {
  189. struct udp_port_cfg udp_conf;
  190. struct socket *sock;
  191. int err;
  192. memset(&udp_conf, 0, sizeof(udp_conf));
  193. if (ipv6_mod_enabled())
  194. udp_conf.family = AF_INET6;
  195. else
  196. udp_conf.family = AF_INET;
  197. udp_conf.local_udp_port = port;
  198. /* Open UDP socket */
  199. err = udp_sock_create(net, &udp_conf, &sock);
  200. if (err < 0)
  201. return ERR_PTR(err);
  202. udp_allow_gso(sock->sk);
  203. return sock;
  204. }
  205. /* Create new listen socket if needed */
  206. static int bareudp_socket_create(struct bareudp_dev *bareudp, __be16 port)
  207. {
  208. struct udp_tunnel_sock_cfg tunnel_cfg;
  209. struct socket *sock;
  210. sock = bareudp_create_sock(bareudp->net, port);
  211. if (IS_ERR(sock))
  212. return PTR_ERR(sock);
  213. /* Mark socket as an encapsulation socket */
  214. memset(&tunnel_cfg, 0, sizeof(tunnel_cfg));
  215. tunnel_cfg.sk_user_data = bareudp;
  216. tunnel_cfg.encap_type = 1;
  217. tunnel_cfg.encap_rcv = bareudp_udp_encap_recv;
  218. tunnel_cfg.encap_err_lookup = bareudp_err_lookup;
  219. tunnel_cfg.encap_destroy = NULL;
  220. setup_udp_tunnel_sock(bareudp->net, sock, &tunnel_cfg);
  221. rcu_assign_pointer(bareudp->sock, sock);
  222. return 0;
  223. }
  224. static int bareudp_open(struct net_device *dev)
  225. {
  226. struct bareudp_dev *bareudp = netdev_priv(dev);
  227. int ret = 0;
  228. ret = bareudp_socket_create(bareudp, bareudp->port);
  229. return ret;
  230. }
  231. static void bareudp_sock_release(struct bareudp_dev *bareudp)
  232. {
  233. struct socket *sock;
  234. sock = bareudp->sock;
  235. rcu_assign_pointer(bareudp->sock, NULL);
  236. synchronize_net();
  237. udp_tunnel_sock_release(sock);
  238. }
  239. static int bareudp_stop(struct net_device *dev)
  240. {
  241. struct bareudp_dev *bareudp = netdev_priv(dev);
  242. bareudp_sock_release(bareudp);
  243. return 0;
  244. }
  245. static int bareudp_xmit_skb(struct sk_buff *skb, struct net_device *dev,
  246. struct bareudp_dev *bareudp,
  247. const struct ip_tunnel_info *info)
  248. {
  249. bool xnet = !net_eq(bareudp->net, dev_net(bareudp->dev));
  250. bool use_cache = ip_tunnel_dst_cache_usable(skb, info);
  251. struct socket *sock = rcu_dereference(bareudp->sock);
  252. bool udp_sum = !!(info->key.tun_flags & TUNNEL_CSUM);
  253. const struct ip_tunnel_key *key = &info->key;
  254. struct rtable *rt;
  255. __be16 sport, df;
  256. int min_headroom;
  257. __u8 tos, ttl;
  258. __be32 saddr;
  259. int err;
  260. if (!sock)
  261. return -ESHUTDOWN;
  262. rt = ip_route_output_tunnel(skb, dev, bareudp->net, &saddr, info,
  263. IPPROTO_UDP, use_cache);
  264. if (IS_ERR(rt))
  265. return PTR_ERR(rt);
  266. skb_tunnel_check_pmtu(skb, &rt->dst,
  267. BAREUDP_IPV4_HLEN + info->options_len, false);
  268. sport = udp_flow_src_port(bareudp->net, skb,
  269. bareudp->sport_min, USHRT_MAX,
  270. true);
  271. tos = ip_tunnel_ecn_encap(key->tos, ip_hdr(skb), skb);
  272. ttl = key->ttl;
  273. df = key->tun_flags & TUNNEL_DONT_FRAGMENT ? htons(IP_DF) : 0;
  274. skb_scrub_packet(skb, xnet);
  275. err = -ENOSPC;
  276. if (!skb_pull(skb, skb_network_offset(skb)))
  277. goto free_dst;
  278. min_headroom = LL_RESERVED_SPACE(rt->dst.dev) + rt->dst.header_len +
  279. BAREUDP_BASE_HLEN + info->options_len + sizeof(struct iphdr);
  280. err = skb_cow_head(skb, min_headroom);
  281. if (unlikely(err))
  282. goto free_dst;
  283. err = udp_tunnel_handle_offloads(skb, udp_sum);
  284. if (err)
  285. goto free_dst;
  286. skb_set_inner_protocol(skb, bareudp->ethertype);
  287. udp_tunnel_xmit_skb(rt, sock->sk, skb, saddr, info->key.u.ipv4.dst,
  288. tos, ttl, df, sport, bareudp->port,
  289. !net_eq(bareudp->net, dev_net(bareudp->dev)),
  290. !(info->key.tun_flags & TUNNEL_CSUM));
  291. return 0;
  292. free_dst:
  293. dst_release(&rt->dst);
  294. return err;
  295. }
  296. static int bareudp6_xmit_skb(struct sk_buff *skb, struct net_device *dev,
  297. struct bareudp_dev *bareudp,
  298. const struct ip_tunnel_info *info)
  299. {
  300. bool xnet = !net_eq(bareudp->net, dev_net(bareudp->dev));
  301. bool use_cache = ip_tunnel_dst_cache_usable(skb, info);
  302. struct socket *sock = rcu_dereference(bareudp->sock);
  303. bool udp_sum = !!(info->key.tun_flags & TUNNEL_CSUM);
  304. const struct ip_tunnel_key *key = &info->key;
  305. struct dst_entry *dst = NULL;
  306. struct in6_addr saddr, daddr;
  307. int min_headroom;
  308. __u8 prio, ttl;
  309. __be16 sport;
  310. int err;
  311. if (!sock)
  312. return -ESHUTDOWN;
  313. dst = ip6_dst_lookup_tunnel(skb, dev, bareudp->net, sock, &saddr, info,
  314. IPPROTO_UDP, use_cache);
  315. if (IS_ERR(dst))
  316. return PTR_ERR(dst);
  317. skb_tunnel_check_pmtu(skb, dst, BAREUDP_IPV6_HLEN + info->options_len,
  318. false);
  319. sport = udp_flow_src_port(bareudp->net, skb,
  320. bareudp->sport_min, USHRT_MAX,
  321. true);
  322. prio = ip_tunnel_ecn_encap(key->tos, ip_hdr(skb), skb);
  323. ttl = key->ttl;
  324. skb_scrub_packet(skb, xnet);
  325. err = -ENOSPC;
  326. if (!skb_pull(skb, skb_network_offset(skb)))
  327. goto free_dst;
  328. min_headroom = LL_RESERVED_SPACE(dst->dev) + dst->header_len +
  329. BAREUDP_BASE_HLEN + info->options_len + sizeof(struct ipv6hdr);
  330. err = skb_cow_head(skb, min_headroom);
  331. if (unlikely(err))
  332. goto free_dst;
  333. err = udp_tunnel_handle_offloads(skb, udp_sum);
  334. if (err)
  335. goto free_dst;
  336. daddr = info->key.u.ipv6.dst;
  337. udp_tunnel6_xmit_skb(dst, sock->sk, skb, dev,
  338. &saddr, &daddr, prio, ttl,
  339. info->key.label, sport, bareudp->port,
  340. !(info->key.tun_flags & TUNNEL_CSUM));
  341. return 0;
  342. free_dst:
  343. dst_release(dst);
  344. return err;
  345. }
  346. static bool bareudp_proto_valid(struct bareudp_dev *bareudp, __be16 proto)
  347. {
  348. if (bareudp->ethertype == proto)
  349. return true;
  350. if (!bareudp->multi_proto_mode)
  351. return false;
  352. if (bareudp->ethertype == htons(ETH_P_MPLS_UC) &&
  353. proto == htons(ETH_P_MPLS_MC))
  354. return true;
  355. if (bareudp->ethertype == htons(ETH_P_IP) &&
  356. proto == htons(ETH_P_IPV6))
  357. return true;
  358. return false;
  359. }
  360. static netdev_tx_t bareudp_xmit(struct sk_buff *skb, struct net_device *dev)
  361. {
  362. struct bareudp_dev *bareudp = netdev_priv(dev);
  363. struct ip_tunnel_info *info = NULL;
  364. int err;
  365. if (!bareudp_proto_valid(bareudp, skb->protocol)) {
  366. err = -EINVAL;
  367. goto tx_error;
  368. }
  369. info = skb_tunnel_info(skb);
  370. if (unlikely(!info || !(info->mode & IP_TUNNEL_INFO_TX))) {
  371. err = -EINVAL;
  372. goto tx_error;
  373. }
  374. rcu_read_lock();
  375. if (ipv6_mod_enabled() && info->mode & IP_TUNNEL_INFO_IPV6)
  376. err = bareudp6_xmit_skb(skb, dev, bareudp, info);
  377. else
  378. err = bareudp_xmit_skb(skb, dev, bareudp, info);
  379. rcu_read_unlock();
  380. if (likely(!err))
  381. return NETDEV_TX_OK;
  382. tx_error:
  383. dev_kfree_skb(skb);
  384. if (err == -ELOOP)
  385. dev->stats.collisions++;
  386. else if (err == -ENETUNREACH)
  387. dev->stats.tx_carrier_errors++;
  388. dev->stats.tx_errors++;
  389. return NETDEV_TX_OK;
  390. }
  391. static int bareudp_fill_metadata_dst(struct net_device *dev,
  392. struct sk_buff *skb)
  393. {
  394. struct ip_tunnel_info *info = skb_tunnel_info(skb);
  395. struct bareudp_dev *bareudp = netdev_priv(dev);
  396. bool use_cache;
  397. use_cache = ip_tunnel_dst_cache_usable(skb, info);
  398. if (!ipv6_mod_enabled() || ip_tunnel_info_af(info) == AF_INET) {
  399. struct rtable *rt;
  400. __be32 saddr;
  401. rt = ip_route_output_tunnel(skb, dev, bareudp->net, &saddr,
  402. info, IPPROTO_UDP, use_cache);
  403. if (IS_ERR(rt))
  404. return PTR_ERR(rt);
  405. ip_rt_put(rt);
  406. info->key.u.ipv4.src = saddr;
  407. } else if (ip_tunnel_info_af(info) == AF_INET6) {
  408. struct dst_entry *dst;
  409. struct in6_addr saddr;
  410. struct socket *sock = rcu_dereference(bareudp->sock);
  411. dst = ip6_dst_lookup_tunnel(skb, dev, bareudp->net, sock,
  412. &saddr, info, IPPROTO_UDP,
  413. use_cache);
  414. if (IS_ERR(dst))
  415. return PTR_ERR(dst);
  416. dst_release(dst);
  417. info->key.u.ipv6.src = saddr;
  418. } else {
  419. return -EINVAL;
  420. }
  421. info->key.tp_src = udp_flow_src_port(bareudp->net, skb,
  422. bareudp->sport_min,
  423. USHRT_MAX, true);
  424. info->key.tp_dst = bareudp->port;
  425. return 0;
  426. }
  427. static const struct net_device_ops bareudp_netdev_ops = {
  428. .ndo_init = bareudp_init,
  429. .ndo_uninit = bareudp_uninit,
  430. .ndo_open = bareudp_open,
  431. .ndo_stop = bareudp_stop,
  432. .ndo_start_xmit = bareudp_xmit,
  433. .ndo_get_stats64 = dev_get_tstats64,
  434. .ndo_fill_metadata_dst = bareudp_fill_metadata_dst,
  435. };
  436. static const struct nla_policy bareudp_policy[IFLA_BAREUDP_MAX + 1] = {
  437. [IFLA_BAREUDP_PORT] = { .type = NLA_U16 },
  438. [IFLA_BAREUDP_ETHERTYPE] = { .type = NLA_U16 },
  439. [IFLA_BAREUDP_SRCPORT_MIN] = { .type = NLA_U16 },
  440. [IFLA_BAREUDP_MULTIPROTO_MODE] = { .type = NLA_FLAG },
  441. };
  442. /* Info for udev, that this is a virtual tunnel endpoint */
  443. static const struct device_type bareudp_type = {
  444. .name = "bareudp",
  445. };
  446. /* Initialize the device structure. */
  447. static void bareudp_setup(struct net_device *dev)
  448. {
  449. dev->netdev_ops = &bareudp_netdev_ops;
  450. dev->needs_free_netdev = true;
  451. SET_NETDEV_DEVTYPE(dev, &bareudp_type);
  452. dev->features |= NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_FRAGLIST;
  453. dev->features |= NETIF_F_RXCSUM;
  454. dev->features |= NETIF_F_LLTX;
  455. dev->features |= NETIF_F_GSO_SOFTWARE;
  456. dev->hw_features |= NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_FRAGLIST;
  457. dev->hw_features |= NETIF_F_RXCSUM;
  458. dev->hw_features |= NETIF_F_GSO_SOFTWARE;
  459. dev->hard_header_len = 0;
  460. dev->addr_len = 0;
  461. dev->mtu = ETH_DATA_LEN;
  462. dev->min_mtu = IPV4_MIN_MTU;
  463. dev->max_mtu = IP_MAX_MTU - BAREUDP_BASE_HLEN;
  464. dev->type = ARPHRD_NONE;
  465. netif_keep_dst(dev);
  466. dev->priv_flags |= IFF_NO_QUEUE;
  467. dev->flags = IFF_POINTOPOINT | IFF_NOARP | IFF_MULTICAST;
  468. }
  469. static int bareudp_validate(struct nlattr *tb[], struct nlattr *data[],
  470. struct netlink_ext_ack *extack)
  471. {
  472. if (!data) {
  473. NL_SET_ERR_MSG(extack,
  474. "Not enough attributes provided to perform the operation");
  475. return -EINVAL;
  476. }
  477. return 0;
  478. }
  479. static int bareudp2info(struct nlattr *data[], struct bareudp_conf *conf,
  480. struct netlink_ext_ack *extack)
  481. {
  482. memset(conf, 0, sizeof(*conf));
  483. if (!data[IFLA_BAREUDP_PORT]) {
  484. NL_SET_ERR_MSG(extack, "port not specified");
  485. return -EINVAL;
  486. }
  487. if (!data[IFLA_BAREUDP_ETHERTYPE]) {
  488. NL_SET_ERR_MSG(extack, "ethertype not specified");
  489. return -EINVAL;
  490. }
  491. conf->port = nla_get_u16(data[IFLA_BAREUDP_PORT]);
  492. conf->ethertype = nla_get_u16(data[IFLA_BAREUDP_ETHERTYPE]);
  493. if (data[IFLA_BAREUDP_SRCPORT_MIN])
  494. conf->sport_min = nla_get_u16(data[IFLA_BAREUDP_SRCPORT_MIN]);
  495. if (data[IFLA_BAREUDP_MULTIPROTO_MODE])
  496. conf->multi_proto_mode = true;
  497. return 0;
  498. }
  499. static struct bareudp_dev *bareudp_find_dev(struct bareudp_net *bn,
  500. const struct bareudp_conf *conf)
  501. {
  502. struct bareudp_dev *bareudp, *t = NULL;
  503. list_for_each_entry(bareudp, &bn->bareudp_list, next) {
  504. if (conf->port == bareudp->port)
  505. t = bareudp;
  506. }
  507. return t;
  508. }
  509. static int bareudp_configure(struct net *net, struct net_device *dev,
  510. struct bareudp_conf *conf,
  511. struct netlink_ext_ack *extack)
  512. {
  513. struct bareudp_net *bn = net_generic(net, bareudp_net_id);
  514. struct bareudp_dev *t, *bareudp = netdev_priv(dev);
  515. int err;
  516. bareudp->net = net;
  517. bareudp->dev = dev;
  518. t = bareudp_find_dev(bn, conf);
  519. if (t) {
  520. NL_SET_ERR_MSG(extack, "Another bareudp device using the same port already exists");
  521. return -EBUSY;
  522. }
  523. if (conf->multi_proto_mode &&
  524. (conf->ethertype != htons(ETH_P_MPLS_UC) &&
  525. conf->ethertype != htons(ETH_P_IP))) {
  526. NL_SET_ERR_MSG(extack, "Cannot set multiproto mode for this ethertype (only IPv4 and unicast MPLS are supported)");
  527. return -EINVAL;
  528. }
  529. bareudp->port = conf->port;
  530. bareudp->ethertype = conf->ethertype;
  531. bareudp->sport_min = conf->sport_min;
  532. bareudp->multi_proto_mode = conf->multi_proto_mode;
  533. err = register_netdevice(dev);
  534. if (err)
  535. return err;
  536. list_add(&bareudp->next, &bn->bareudp_list);
  537. return 0;
  538. }
  539. static int bareudp_link_config(struct net_device *dev,
  540. struct nlattr *tb[])
  541. {
  542. int err;
  543. if (tb[IFLA_MTU]) {
  544. err = dev_set_mtu(dev, nla_get_u32(tb[IFLA_MTU]));
  545. if (err)
  546. return err;
  547. }
  548. return 0;
  549. }
  550. static void bareudp_dellink(struct net_device *dev, struct list_head *head)
  551. {
  552. struct bareudp_dev *bareudp = netdev_priv(dev);
  553. list_del(&bareudp->next);
  554. unregister_netdevice_queue(dev, head);
  555. }
  556. static int bareudp_newlink(struct net *net, struct net_device *dev,
  557. struct nlattr *tb[], struct nlattr *data[],
  558. struct netlink_ext_ack *extack)
  559. {
  560. struct bareudp_conf conf;
  561. int err;
  562. err = bareudp2info(data, &conf, extack);
  563. if (err)
  564. return err;
  565. err = bareudp_configure(net, dev, &conf, extack);
  566. if (err)
  567. return err;
  568. err = bareudp_link_config(dev, tb);
  569. if (err)
  570. goto err_unconfig;
  571. return 0;
  572. err_unconfig:
  573. bareudp_dellink(dev, NULL);
  574. return err;
  575. }
  576. static size_t bareudp_get_size(const struct net_device *dev)
  577. {
  578. return nla_total_size(sizeof(__be16)) + /* IFLA_BAREUDP_PORT */
  579. nla_total_size(sizeof(__be16)) + /* IFLA_BAREUDP_ETHERTYPE */
  580. nla_total_size(sizeof(__u16)) + /* IFLA_BAREUDP_SRCPORT_MIN */
  581. nla_total_size(0) + /* IFLA_BAREUDP_MULTIPROTO_MODE */
  582. 0;
  583. }
  584. static int bareudp_fill_info(struct sk_buff *skb, const struct net_device *dev)
  585. {
  586. struct bareudp_dev *bareudp = netdev_priv(dev);
  587. if (nla_put_be16(skb, IFLA_BAREUDP_PORT, bareudp->port))
  588. goto nla_put_failure;
  589. if (nla_put_be16(skb, IFLA_BAREUDP_ETHERTYPE, bareudp->ethertype))
  590. goto nla_put_failure;
  591. if (nla_put_u16(skb, IFLA_BAREUDP_SRCPORT_MIN, bareudp->sport_min))
  592. goto nla_put_failure;
  593. if (bareudp->multi_proto_mode &&
  594. nla_put_flag(skb, IFLA_BAREUDP_MULTIPROTO_MODE))
  595. goto nla_put_failure;
  596. return 0;
  597. nla_put_failure:
  598. return -EMSGSIZE;
  599. }
  600. static struct rtnl_link_ops bareudp_link_ops __read_mostly = {
  601. .kind = "bareudp",
  602. .maxtype = IFLA_BAREUDP_MAX,
  603. .policy = bareudp_policy,
  604. .priv_size = sizeof(struct bareudp_dev),
  605. .setup = bareudp_setup,
  606. .validate = bareudp_validate,
  607. .newlink = bareudp_newlink,
  608. .dellink = bareudp_dellink,
  609. .get_size = bareudp_get_size,
  610. .fill_info = bareudp_fill_info,
  611. };
  612. static __net_init int bareudp_init_net(struct net *net)
  613. {
  614. struct bareudp_net *bn = net_generic(net, bareudp_net_id);
  615. INIT_LIST_HEAD(&bn->bareudp_list);
  616. return 0;
  617. }
  618. static void bareudp_destroy_tunnels(struct net *net, struct list_head *head)
  619. {
  620. struct bareudp_net *bn = net_generic(net, bareudp_net_id);
  621. struct bareudp_dev *bareudp, *next;
  622. list_for_each_entry_safe(bareudp, next, &bn->bareudp_list, next)
  623. unregister_netdevice_queue(bareudp->dev, head);
  624. }
  625. static void __net_exit bareudp_exit_batch_net(struct list_head *net_list)
  626. {
  627. struct net *net;
  628. LIST_HEAD(list);
  629. rtnl_lock();
  630. list_for_each_entry(net, net_list, exit_list)
  631. bareudp_destroy_tunnels(net, &list);
  632. /* unregister the devices gathered above */
  633. unregister_netdevice_many(&list);
  634. rtnl_unlock();
  635. }
  636. static struct pernet_operations bareudp_net_ops = {
  637. .init = bareudp_init_net,
  638. .exit_batch = bareudp_exit_batch_net,
  639. .id = &bareudp_net_id,
  640. .size = sizeof(struct bareudp_net),
  641. };
  642. static int __init bareudp_init_module(void)
  643. {
  644. int rc;
  645. rc = register_pernet_subsys(&bareudp_net_ops);
  646. if (rc)
  647. goto out1;
  648. rc = rtnl_link_register(&bareudp_link_ops);
  649. if (rc)
  650. goto out2;
  651. return 0;
  652. out2:
  653. unregister_pernet_subsys(&bareudp_net_ops);
  654. out1:
  655. return rc;
  656. }
  657. late_initcall(bareudp_init_module);
  658. static void __exit bareudp_cleanup_module(void)
  659. {
  660. rtnl_link_unregister(&bareudp_link_ops);
  661. unregister_pernet_subsys(&bareudp_net_ops);
  662. }
  663. module_exit(bareudp_cleanup_module);
  664. MODULE_ALIAS_RTNL_LINK("bareudp");
  665. MODULE_LICENSE("GPL");
  666. MODULE_AUTHOR("Martin Varghese <[email protected]>");
  667. MODULE_DESCRIPTION("Interface driver for UDP encapsulated traffic");