device.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * Management Component Transport Protocol (MCTP) - device implementation.
  4. *
  5. * Copyright (c) 2021 Code Construct
  6. * Copyright (c) 2021 Google
  7. */
  8. #include <linux/if_arp.h>
  9. #include <linux/if_link.h>
  10. #include <linux/mctp.h>
  11. #include <linux/netdevice.h>
  12. #include <linux/rcupdate.h>
  13. #include <linux/rtnetlink.h>
  14. #include <net/addrconf.h>
  15. #include <net/netlink.h>
  16. #include <net/mctp.h>
  17. #include <net/mctpdevice.h>
  18. #include <net/sock.h>
  19. struct mctp_dump_cb {
  20. int h;
  21. int idx;
  22. size_t a_idx;
  23. };
  24. /* unlocked: caller must hold rcu_read_lock.
  25. * Returned mctp_dev has its refcount incremented, or NULL if unset.
  26. */
  27. struct mctp_dev *__mctp_dev_get(const struct net_device *dev)
  28. {
  29. struct mctp_dev *mdev = rcu_dereference(dev->mctp_ptr);
  30. /* RCU guarantees that any mdev is still live.
  31. * Zero refcount implies a pending free, return NULL.
  32. */
  33. if (mdev)
  34. if (!refcount_inc_not_zero(&mdev->refs))
  35. return NULL;
  36. return mdev;
  37. }
  38. /* Returned mctp_dev does not have refcount incremented. The returned pointer
  39. * remains live while rtnl_lock is held, as that prevents mctp_unregister()
  40. */
  41. struct mctp_dev *mctp_dev_get_rtnl(const struct net_device *dev)
  42. {
  43. return rtnl_dereference(dev->mctp_ptr);
  44. }
  45. static int mctp_addrinfo_size(void)
  46. {
  47. return NLMSG_ALIGN(sizeof(struct ifaddrmsg))
  48. + nla_total_size(1) // IFA_LOCAL
  49. + nla_total_size(1) // IFA_ADDRESS
  50. ;
  51. }
  52. /* flag should be NLM_F_MULTI for dump calls */
  53. static int mctp_fill_addrinfo(struct sk_buff *skb,
  54. struct mctp_dev *mdev, mctp_eid_t eid,
  55. int msg_type, u32 portid, u32 seq, int flag)
  56. {
  57. struct ifaddrmsg *hdr;
  58. struct nlmsghdr *nlh;
  59. nlh = nlmsg_put(skb, portid, seq,
  60. msg_type, sizeof(*hdr), flag);
  61. if (!nlh)
  62. return -EMSGSIZE;
  63. hdr = nlmsg_data(nlh);
  64. hdr->ifa_family = AF_MCTP;
  65. hdr->ifa_prefixlen = 0;
  66. hdr->ifa_flags = 0;
  67. hdr->ifa_scope = 0;
  68. hdr->ifa_index = mdev->dev->ifindex;
  69. if (nla_put_u8(skb, IFA_LOCAL, eid))
  70. goto cancel;
  71. if (nla_put_u8(skb, IFA_ADDRESS, eid))
  72. goto cancel;
  73. nlmsg_end(skb, nlh);
  74. return 0;
  75. cancel:
  76. nlmsg_cancel(skb, nlh);
  77. return -EMSGSIZE;
  78. }
  79. static int mctp_dump_dev_addrinfo(struct mctp_dev *mdev, struct sk_buff *skb,
  80. struct netlink_callback *cb)
  81. {
  82. struct mctp_dump_cb *mcb = (void *)cb->ctx;
  83. u32 portid, seq;
  84. int rc = 0;
  85. portid = NETLINK_CB(cb->skb).portid;
  86. seq = cb->nlh->nlmsg_seq;
  87. for (; mcb->a_idx < mdev->num_addrs; mcb->a_idx++) {
  88. rc = mctp_fill_addrinfo(skb, mdev, mdev->addrs[mcb->a_idx],
  89. RTM_NEWADDR, portid, seq, NLM_F_MULTI);
  90. if (rc < 0)
  91. break;
  92. }
  93. return rc;
  94. }
  95. static int mctp_dump_addrinfo(struct sk_buff *skb, struct netlink_callback *cb)
  96. {
  97. struct mctp_dump_cb *mcb = (void *)cb->ctx;
  98. struct net *net = sock_net(skb->sk);
  99. struct hlist_head *head;
  100. struct net_device *dev;
  101. struct ifaddrmsg *hdr;
  102. struct mctp_dev *mdev;
  103. int ifindex;
  104. int idx = 0, rc;
  105. hdr = nlmsg_data(cb->nlh);
  106. // filter by ifindex if requested
  107. ifindex = hdr->ifa_index;
  108. rcu_read_lock();
  109. for (; mcb->h < NETDEV_HASHENTRIES; mcb->h++, mcb->idx = 0) {
  110. idx = 0;
  111. head = &net->dev_index_head[mcb->h];
  112. hlist_for_each_entry_rcu(dev, head, index_hlist) {
  113. if (idx >= mcb->idx &&
  114. (ifindex == 0 || ifindex == dev->ifindex)) {
  115. mdev = __mctp_dev_get(dev);
  116. if (mdev) {
  117. rc = mctp_dump_dev_addrinfo(mdev,
  118. skb, cb);
  119. mctp_dev_put(mdev);
  120. // Error indicates full buffer, this
  121. // callback will get retried.
  122. if (rc < 0)
  123. goto out;
  124. }
  125. }
  126. idx++;
  127. // reset for next iteration
  128. mcb->a_idx = 0;
  129. }
  130. }
  131. out:
  132. rcu_read_unlock();
  133. mcb->idx = idx;
  134. return skb->len;
  135. }
  136. static void mctp_addr_notify(struct mctp_dev *mdev, mctp_eid_t eid, int msg_type,
  137. struct sk_buff *req_skb, struct nlmsghdr *req_nlh)
  138. {
  139. u32 portid = NETLINK_CB(req_skb).portid;
  140. struct net *net = dev_net(mdev->dev);
  141. struct sk_buff *skb;
  142. int rc = -ENOBUFS;
  143. skb = nlmsg_new(mctp_addrinfo_size(), GFP_KERNEL);
  144. if (!skb)
  145. goto out;
  146. rc = mctp_fill_addrinfo(skb, mdev, eid, msg_type,
  147. portid, req_nlh->nlmsg_seq, 0);
  148. if (rc < 0) {
  149. WARN_ON_ONCE(rc == -EMSGSIZE);
  150. goto out;
  151. }
  152. rtnl_notify(skb, net, portid, RTNLGRP_MCTP_IFADDR, req_nlh, GFP_KERNEL);
  153. return;
  154. out:
  155. kfree_skb(skb);
  156. rtnl_set_sk_err(net, RTNLGRP_MCTP_IFADDR, rc);
  157. }
  158. static const struct nla_policy ifa_mctp_policy[IFA_MAX + 1] = {
  159. [IFA_ADDRESS] = { .type = NLA_U8 },
  160. [IFA_LOCAL] = { .type = NLA_U8 },
  161. };
  162. static int mctp_rtm_newaddr(struct sk_buff *skb, struct nlmsghdr *nlh,
  163. struct netlink_ext_ack *extack)
  164. {
  165. struct net *net = sock_net(skb->sk);
  166. struct nlattr *tb[IFA_MAX + 1];
  167. struct net_device *dev;
  168. struct mctp_addr *addr;
  169. struct mctp_dev *mdev;
  170. struct ifaddrmsg *ifm;
  171. unsigned long flags;
  172. u8 *tmp_addrs;
  173. int rc;
  174. rc = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_mctp_policy,
  175. extack);
  176. if (rc < 0)
  177. return rc;
  178. ifm = nlmsg_data(nlh);
  179. if (tb[IFA_LOCAL])
  180. addr = nla_data(tb[IFA_LOCAL]);
  181. else if (tb[IFA_ADDRESS])
  182. addr = nla_data(tb[IFA_ADDRESS]);
  183. else
  184. return -EINVAL;
  185. /* find device */
  186. dev = __dev_get_by_index(net, ifm->ifa_index);
  187. if (!dev)
  188. return -ENODEV;
  189. mdev = mctp_dev_get_rtnl(dev);
  190. if (!mdev)
  191. return -ENODEV;
  192. if (!mctp_address_unicast(addr->s_addr))
  193. return -EINVAL;
  194. /* Prevent duplicates. Under RTNL so don't need to lock for reading */
  195. if (memchr(mdev->addrs, addr->s_addr, mdev->num_addrs))
  196. return -EEXIST;
  197. tmp_addrs = kmalloc(mdev->num_addrs + 1, GFP_KERNEL);
  198. if (!tmp_addrs)
  199. return -ENOMEM;
  200. memcpy(tmp_addrs, mdev->addrs, mdev->num_addrs);
  201. tmp_addrs[mdev->num_addrs] = addr->s_addr;
  202. /* Lock to write */
  203. spin_lock_irqsave(&mdev->addrs_lock, flags);
  204. mdev->num_addrs++;
  205. swap(mdev->addrs, tmp_addrs);
  206. spin_unlock_irqrestore(&mdev->addrs_lock, flags);
  207. kfree(tmp_addrs);
  208. mctp_addr_notify(mdev, addr->s_addr, RTM_NEWADDR, skb, nlh);
  209. mctp_route_add_local(mdev, addr->s_addr);
  210. return 0;
  211. }
  212. static int mctp_rtm_deladdr(struct sk_buff *skb, struct nlmsghdr *nlh,
  213. struct netlink_ext_ack *extack)
  214. {
  215. struct net *net = sock_net(skb->sk);
  216. struct nlattr *tb[IFA_MAX + 1];
  217. struct net_device *dev;
  218. struct mctp_addr *addr;
  219. struct mctp_dev *mdev;
  220. struct ifaddrmsg *ifm;
  221. unsigned long flags;
  222. u8 *pos;
  223. int rc;
  224. rc = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_mctp_policy,
  225. extack);
  226. if (rc < 0)
  227. return rc;
  228. ifm = nlmsg_data(nlh);
  229. if (tb[IFA_LOCAL])
  230. addr = nla_data(tb[IFA_LOCAL]);
  231. else if (tb[IFA_ADDRESS])
  232. addr = nla_data(tb[IFA_ADDRESS]);
  233. else
  234. return -EINVAL;
  235. /* find device */
  236. dev = __dev_get_by_index(net, ifm->ifa_index);
  237. if (!dev)
  238. return -ENODEV;
  239. mdev = mctp_dev_get_rtnl(dev);
  240. if (!mdev)
  241. return -ENODEV;
  242. pos = memchr(mdev->addrs, addr->s_addr, mdev->num_addrs);
  243. if (!pos)
  244. return -ENOENT;
  245. rc = mctp_route_remove_local(mdev, addr->s_addr);
  246. // we can ignore -ENOENT in the case a route was already removed
  247. if (rc < 0 && rc != -ENOENT)
  248. return rc;
  249. spin_lock_irqsave(&mdev->addrs_lock, flags);
  250. memmove(pos, pos + 1, mdev->num_addrs - 1 - (pos - mdev->addrs));
  251. mdev->num_addrs--;
  252. spin_unlock_irqrestore(&mdev->addrs_lock, flags);
  253. mctp_addr_notify(mdev, addr->s_addr, RTM_DELADDR, skb, nlh);
  254. return 0;
  255. }
  256. void mctp_dev_hold(struct mctp_dev *mdev)
  257. {
  258. refcount_inc(&mdev->refs);
  259. }
  260. void mctp_dev_put(struct mctp_dev *mdev)
  261. {
  262. if (mdev && refcount_dec_and_test(&mdev->refs)) {
  263. kfree(mdev->addrs);
  264. dev_put(mdev->dev);
  265. kfree_rcu(mdev, rcu);
  266. }
  267. }
  268. void mctp_dev_release_key(struct mctp_dev *dev, struct mctp_sk_key *key)
  269. __must_hold(&key->lock)
  270. {
  271. if (!dev)
  272. return;
  273. if (dev->ops && dev->ops->release_flow)
  274. dev->ops->release_flow(dev, key);
  275. key->dev = NULL;
  276. mctp_dev_put(dev);
  277. }
  278. void mctp_dev_set_key(struct mctp_dev *dev, struct mctp_sk_key *key)
  279. __must_hold(&key->lock)
  280. {
  281. mctp_dev_hold(dev);
  282. key->dev = dev;
  283. }
  284. static struct mctp_dev *mctp_add_dev(struct net_device *dev)
  285. {
  286. struct mctp_dev *mdev;
  287. ASSERT_RTNL();
  288. mdev = kzalloc(sizeof(*mdev), GFP_KERNEL);
  289. if (!mdev)
  290. return ERR_PTR(-ENOMEM);
  291. spin_lock_init(&mdev->addrs_lock);
  292. mdev->net = mctp_default_net(dev_net(dev));
  293. /* associate to net_device */
  294. refcount_set(&mdev->refs, 1);
  295. rcu_assign_pointer(dev->mctp_ptr, mdev);
  296. dev_hold(dev);
  297. mdev->dev = dev;
  298. return mdev;
  299. }
  300. static int mctp_fill_link_af(struct sk_buff *skb,
  301. const struct net_device *dev, u32 ext_filter_mask)
  302. {
  303. struct mctp_dev *mdev;
  304. mdev = mctp_dev_get_rtnl(dev);
  305. if (!mdev)
  306. return -ENODATA;
  307. if (nla_put_u32(skb, IFLA_MCTP_NET, mdev->net))
  308. return -EMSGSIZE;
  309. return 0;
  310. }
  311. static size_t mctp_get_link_af_size(const struct net_device *dev,
  312. u32 ext_filter_mask)
  313. {
  314. struct mctp_dev *mdev;
  315. unsigned int ret;
  316. /* caller holds RCU */
  317. mdev = __mctp_dev_get(dev);
  318. if (!mdev)
  319. return 0;
  320. ret = nla_total_size(4); /* IFLA_MCTP_NET */
  321. mctp_dev_put(mdev);
  322. return ret;
  323. }
  324. static const struct nla_policy ifla_af_mctp_policy[IFLA_MCTP_MAX + 1] = {
  325. [IFLA_MCTP_NET] = { .type = NLA_U32 },
  326. };
  327. static int mctp_set_link_af(struct net_device *dev, const struct nlattr *attr,
  328. struct netlink_ext_ack *extack)
  329. {
  330. struct nlattr *tb[IFLA_MCTP_MAX + 1];
  331. struct mctp_dev *mdev;
  332. int rc;
  333. rc = nla_parse_nested(tb, IFLA_MCTP_MAX, attr, ifla_af_mctp_policy,
  334. NULL);
  335. if (rc)
  336. return rc;
  337. mdev = mctp_dev_get_rtnl(dev);
  338. if (!mdev)
  339. return 0;
  340. if (tb[IFLA_MCTP_NET])
  341. WRITE_ONCE(mdev->net, nla_get_u32(tb[IFLA_MCTP_NET]));
  342. return 0;
  343. }
  344. /* Matches netdev types that should have MCTP handling */
  345. static bool mctp_known(struct net_device *dev)
  346. {
  347. /* only register specific types (inc. NONE for TUN devices) */
  348. return dev->type == ARPHRD_MCTP ||
  349. dev->type == ARPHRD_LOOPBACK ||
  350. dev->type == ARPHRD_NONE;
  351. }
  352. static void mctp_unregister(struct net_device *dev)
  353. {
  354. struct mctp_dev *mdev;
  355. mdev = mctp_dev_get_rtnl(dev);
  356. if (!mdev)
  357. return;
  358. RCU_INIT_POINTER(mdev->dev->mctp_ptr, NULL);
  359. mctp_route_remove_dev(mdev);
  360. mctp_neigh_remove_dev(mdev);
  361. mctp_dev_put(mdev);
  362. }
  363. static int mctp_register(struct net_device *dev)
  364. {
  365. struct mctp_dev *mdev;
  366. /* Already registered? */
  367. if (rtnl_dereference(dev->mctp_ptr))
  368. return 0;
  369. /* only register specific types */
  370. if (!mctp_known(dev))
  371. return 0;
  372. mdev = mctp_add_dev(dev);
  373. if (IS_ERR(mdev))
  374. return PTR_ERR(mdev);
  375. return 0;
  376. }
  377. static int mctp_dev_notify(struct notifier_block *this, unsigned long event,
  378. void *ptr)
  379. {
  380. struct net_device *dev = netdev_notifier_info_to_dev(ptr);
  381. int rc;
  382. switch (event) {
  383. case NETDEV_REGISTER:
  384. rc = mctp_register(dev);
  385. if (rc)
  386. return notifier_from_errno(rc);
  387. break;
  388. case NETDEV_UNREGISTER:
  389. mctp_unregister(dev);
  390. break;
  391. }
  392. return NOTIFY_OK;
  393. }
  394. static int mctp_register_netdevice(struct net_device *dev,
  395. const struct mctp_netdev_ops *ops)
  396. {
  397. struct mctp_dev *mdev;
  398. mdev = mctp_add_dev(dev);
  399. if (IS_ERR(mdev))
  400. return PTR_ERR(mdev);
  401. mdev->ops = ops;
  402. return register_netdevice(dev);
  403. }
  404. int mctp_register_netdev(struct net_device *dev,
  405. const struct mctp_netdev_ops *ops)
  406. {
  407. int rc;
  408. rtnl_lock();
  409. rc = mctp_register_netdevice(dev, ops);
  410. rtnl_unlock();
  411. return rc;
  412. }
  413. EXPORT_SYMBOL_GPL(mctp_register_netdev);
  414. void mctp_unregister_netdev(struct net_device *dev)
  415. {
  416. unregister_netdev(dev);
  417. }
  418. EXPORT_SYMBOL_GPL(mctp_unregister_netdev);
  419. static struct rtnl_af_ops mctp_af_ops = {
  420. .family = AF_MCTP,
  421. .fill_link_af = mctp_fill_link_af,
  422. .get_link_af_size = mctp_get_link_af_size,
  423. .set_link_af = mctp_set_link_af,
  424. };
  425. static struct notifier_block mctp_dev_nb = {
  426. .notifier_call = mctp_dev_notify,
  427. .priority = ADDRCONF_NOTIFY_PRIORITY,
  428. };
  429. void __init mctp_device_init(void)
  430. {
  431. register_netdevice_notifier(&mctp_dev_nb);
  432. rtnl_register_module(THIS_MODULE, PF_MCTP, RTM_GETADDR,
  433. NULL, mctp_dump_addrinfo, 0);
  434. rtnl_register_module(THIS_MODULE, PF_MCTP, RTM_NEWADDR,
  435. mctp_rtm_newaddr, NULL, 0);
  436. rtnl_register_module(THIS_MODULE, PF_MCTP, RTM_DELADDR,
  437. mctp_rtm_deladdr, NULL, 0);
  438. rtnl_af_register(&mctp_af_ops);
  439. }
  440. void __exit mctp_device_exit(void)
  441. {
  442. rtnl_af_unregister(&mctp_af_ops);
  443. rtnl_unregister(PF_MCTP, RTM_DELADDR);
  444. rtnl_unregister(PF_MCTP, RTM_NEWADDR);
  445. rtnl_unregister(PF_MCTP, RTM_GETADDR);
  446. unregister_netdevice_notifier(&mctp_dev_nb);
  447. }