main.c 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369
  1. // SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
  2. /* Copyright (c) 2015 - 2021 Intel Corporation */
  3. #include "main.h"
  4. #include "../../../net/ethernet/intel/ice/ice.h"
  5. MODULE_ALIAS("i40iw");
  6. MODULE_AUTHOR("Intel Corporation, <[email protected]>");
  7. MODULE_DESCRIPTION("Intel(R) Ethernet Protocol Driver for RDMA");
  8. MODULE_LICENSE("Dual BSD/GPL");
  9. static struct notifier_block irdma_inetaddr_notifier = {
  10. .notifier_call = irdma_inetaddr_event
  11. };
  12. static struct notifier_block irdma_inetaddr6_notifier = {
  13. .notifier_call = irdma_inet6addr_event
  14. };
  15. static struct notifier_block irdma_net_notifier = {
  16. .notifier_call = irdma_net_event
  17. };
  18. static struct notifier_block irdma_netdevice_notifier = {
  19. .notifier_call = irdma_netdevice_event
  20. };
  21. static void irdma_register_notifiers(void)
  22. {
  23. register_inetaddr_notifier(&irdma_inetaddr_notifier);
  24. register_inet6addr_notifier(&irdma_inetaddr6_notifier);
  25. register_netevent_notifier(&irdma_net_notifier);
  26. register_netdevice_notifier(&irdma_netdevice_notifier);
  27. }
  28. static void irdma_unregister_notifiers(void)
  29. {
  30. unregister_netevent_notifier(&irdma_net_notifier);
  31. unregister_inetaddr_notifier(&irdma_inetaddr_notifier);
  32. unregister_inet6addr_notifier(&irdma_inetaddr6_notifier);
  33. unregister_netdevice_notifier(&irdma_netdevice_notifier);
  34. }
  35. static void irdma_prep_tc_change(struct irdma_device *iwdev)
  36. {
  37. iwdev->vsi.tc_change_pending = true;
  38. irdma_sc_suspend_resume_qps(&iwdev->vsi, IRDMA_OP_SUSPEND);
  39. /* Wait for all qp's to suspend */
  40. wait_event_timeout(iwdev->suspend_wq,
  41. !atomic_read(&iwdev->vsi.qp_suspend_reqs),
  42. msecs_to_jiffies(IRDMA_EVENT_TIMEOUT_MS));
  43. irdma_ws_reset(&iwdev->vsi);
  44. }
  45. static void irdma_log_invalid_mtu(u16 mtu, struct irdma_sc_dev *dev)
  46. {
  47. if (mtu < IRDMA_MIN_MTU_IPV4)
  48. ibdev_warn(to_ibdev(dev), "MTU setting [%d] too low for RDMA traffic. Minimum MTU is 576 for IPv4\n", mtu);
  49. else if (mtu < IRDMA_MIN_MTU_IPV6)
  50. ibdev_warn(to_ibdev(dev), "MTU setting [%d] too low for RDMA traffic. Minimum MTU is 1280 for IPv6\\n", mtu);
  51. }
  52. static void irdma_fill_qos_info(struct irdma_l2params *l2params,
  53. struct iidc_qos_params *qos_info)
  54. {
  55. int i;
  56. l2params->num_tc = qos_info->num_tc;
  57. l2params->vsi_prio_type = qos_info->vport_priority_type;
  58. l2params->vsi_rel_bw = qos_info->vport_relative_bw;
  59. for (i = 0; i < l2params->num_tc; i++) {
  60. l2params->tc_info[i].egress_virt_up =
  61. qos_info->tc_info[i].egress_virt_up;
  62. l2params->tc_info[i].ingress_virt_up =
  63. qos_info->tc_info[i].ingress_virt_up;
  64. l2params->tc_info[i].prio_type = qos_info->tc_info[i].prio_type;
  65. l2params->tc_info[i].rel_bw = qos_info->tc_info[i].rel_bw;
  66. l2params->tc_info[i].tc_ctx = qos_info->tc_info[i].tc_ctx;
  67. }
  68. for (i = 0; i < IIDC_MAX_USER_PRIORITY; i++)
  69. l2params->up2tc[i] = qos_info->up2tc[i];
  70. if (qos_info->pfc_mode == IIDC_DSCP_PFC_MODE) {
  71. l2params->dscp_mode = true;
  72. memcpy(l2params->dscp_map, qos_info->dscp_map, sizeof(l2params->dscp_map));
  73. }
  74. }
  75. static void irdma_iidc_event_handler(struct ice_pf *pf, struct iidc_event *event)
  76. {
  77. struct irdma_device *iwdev = dev_get_drvdata(&pf->adev->dev);
  78. struct irdma_l2params l2params = {};
  79. if (*event->type & BIT(IIDC_EVENT_AFTER_MTU_CHANGE)) {
  80. ibdev_dbg(&iwdev->ibdev, "CLNT: new MTU = %d\n", iwdev->netdev->mtu);
  81. if (iwdev->vsi.mtu != iwdev->netdev->mtu) {
  82. l2params.mtu = iwdev->netdev->mtu;
  83. l2params.mtu_changed = true;
  84. irdma_log_invalid_mtu(l2params.mtu, &iwdev->rf->sc_dev);
  85. irdma_change_l2params(&iwdev->vsi, &l2params);
  86. }
  87. } else if (*event->type & BIT(IIDC_EVENT_BEFORE_TC_CHANGE)) {
  88. if (iwdev->vsi.tc_change_pending)
  89. return;
  90. irdma_prep_tc_change(iwdev);
  91. } else if (*event->type & BIT(IIDC_EVENT_AFTER_TC_CHANGE)) {
  92. struct iidc_qos_params qos_info = {};
  93. if (!iwdev->vsi.tc_change_pending)
  94. return;
  95. l2params.tc_changed = true;
  96. ibdev_dbg(&iwdev->ibdev, "CLNT: TC Change\n");
  97. ice_get_qos_params(pf, &qos_info);
  98. irdma_fill_qos_info(&l2params, &qos_info);
  99. if (iwdev->rf->protocol_used != IRDMA_IWARP_PROTOCOL_ONLY)
  100. iwdev->dcb_vlan_mode = qos_info.num_tc > 1 && !l2params.dscp_mode;
  101. irdma_change_l2params(&iwdev->vsi, &l2params);
  102. } else if (*event->type & BIT(IIDC_EVENT_CRIT_ERR)) {
  103. ibdev_warn(&iwdev->ibdev, "ICE OICR event notification: oicr = 0x%08x\n",
  104. event->reg);
  105. if (event->reg & IRDMAPFINT_OICR_PE_CRITERR_M) {
  106. u32 pe_criterr;
  107. pe_criterr = readl(iwdev->rf->sc_dev.hw_regs[IRDMA_GLPE_CRITERR]);
  108. #define IRDMA_Q1_RESOURCE_ERR 0x0001024d
  109. if (pe_criterr != IRDMA_Q1_RESOURCE_ERR) {
  110. ibdev_err(&iwdev->ibdev, "critical PE Error, GLPE_CRITERR=0x%08x\n",
  111. pe_criterr);
  112. iwdev->rf->reset = true;
  113. } else {
  114. ibdev_warn(&iwdev->ibdev, "Q1 Resource Check\n");
  115. }
  116. }
  117. if (event->reg & IRDMAPFINT_OICR_HMC_ERR_M) {
  118. ibdev_err(&iwdev->ibdev, "HMC Error\n");
  119. iwdev->rf->reset = true;
  120. }
  121. if (event->reg & IRDMAPFINT_OICR_PE_PUSH_M) {
  122. ibdev_err(&iwdev->ibdev, "PE Push Error\n");
  123. iwdev->rf->reset = true;
  124. }
  125. if (iwdev->rf->reset)
  126. iwdev->rf->gen_ops.request_reset(iwdev->rf);
  127. }
  128. }
  129. /**
  130. * irdma_request_reset - Request a reset
  131. * @rf: RDMA PCI function
  132. */
  133. static void irdma_request_reset(struct irdma_pci_f *rf)
  134. {
  135. struct ice_pf *pf = rf->cdev;
  136. ibdev_warn(&rf->iwdev->ibdev, "Requesting a reset\n");
  137. ice_rdma_request_reset(pf, IIDC_PFR);
  138. }
  139. /**
  140. * irdma_lan_register_qset - Register qset with LAN driver
  141. * @vsi: vsi structure
  142. * @tc_node: Traffic class node
  143. */
  144. static int irdma_lan_register_qset(struct irdma_sc_vsi *vsi,
  145. struct irdma_ws_node *tc_node)
  146. {
  147. struct irdma_device *iwdev = vsi->back_vsi;
  148. struct ice_pf *pf = iwdev->rf->cdev;
  149. struct iidc_rdma_qset_params qset = {};
  150. int ret;
  151. qset.qs_handle = tc_node->qs_handle;
  152. qset.tc = tc_node->traffic_class;
  153. qset.vport_id = vsi->vsi_idx;
  154. ret = ice_add_rdma_qset(pf, &qset);
  155. if (ret) {
  156. ibdev_dbg(&iwdev->ibdev, "WS: LAN alloc_res for rdma qset failed.\n");
  157. return ret;
  158. }
  159. tc_node->l2_sched_node_id = qset.teid;
  160. vsi->qos[tc_node->user_pri].l2_sched_node_id = qset.teid;
  161. return 0;
  162. }
  163. /**
  164. * irdma_lan_unregister_qset - Unregister qset with LAN driver
  165. * @vsi: vsi structure
  166. * @tc_node: Traffic class node
  167. */
  168. static void irdma_lan_unregister_qset(struct irdma_sc_vsi *vsi,
  169. struct irdma_ws_node *tc_node)
  170. {
  171. struct irdma_device *iwdev = vsi->back_vsi;
  172. struct ice_pf *pf = iwdev->rf->cdev;
  173. struct iidc_rdma_qset_params qset = {};
  174. qset.qs_handle = tc_node->qs_handle;
  175. qset.tc = tc_node->traffic_class;
  176. qset.vport_id = vsi->vsi_idx;
  177. qset.teid = tc_node->l2_sched_node_id;
  178. if (ice_del_rdma_qset(pf, &qset))
  179. ibdev_dbg(&iwdev->ibdev, "WS: LAN free_res for rdma qset failed.\n");
  180. }
  181. static void irdma_remove(struct auxiliary_device *aux_dev)
  182. {
  183. struct iidc_auxiliary_dev *iidc_adev = container_of(aux_dev,
  184. struct iidc_auxiliary_dev,
  185. adev);
  186. struct ice_pf *pf = iidc_adev->pf;
  187. struct irdma_device *iwdev = auxiliary_get_drvdata(aux_dev);
  188. irdma_ib_unregister_device(iwdev);
  189. ice_rdma_update_vsi_filter(pf, iwdev->vsi_num, false);
  190. pr_debug("INIT: Gen2 PF[%d] device remove success\n", PCI_FUNC(pf->pdev->devfn));
  191. }
  192. static void irdma_fill_device_info(struct irdma_device *iwdev, struct ice_pf *pf,
  193. struct ice_vsi *vsi)
  194. {
  195. struct irdma_pci_f *rf = iwdev->rf;
  196. rf->cdev = pf;
  197. rf->gen_ops.register_qset = irdma_lan_register_qset;
  198. rf->gen_ops.unregister_qset = irdma_lan_unregister_qset;
  199. rf->hw.hw_addr = pf->hw.hw_addr;
  200. rf->pcidev = pf->pdev;
  201. rf->msix_count = pf->num_rdma_msix;
  202. rf->pf_id = pf->hw.pf_id;
  203. rf->msix_entries = &pf->msix_entries[pf->rdma_base_vector];
  204. rf->default_vsi.vsi_idx = vsi->vsi_num;
  205. rf->protocol_used = pf->rdma_mode & IIDC_RDMA_PROTOCOL_ROCEV2 ?
  206. IRDMA_ROCE_PROTOCOL_ONLY : IRDMA_IWARP_PROTOCOL_ONLY;
  207. rf->rdma_ver = IRDMA_GEN_2;
  208. rf->rsrc_profile = IRDMA_HMC_PROFILE_DEFAULT;
  209. rf->rst_to = IRDMA_RST_TIMEOUT_HZ;
  210. rf->gen_ops.request_reset = irdma_request_reset;
  211. rf->limits_sel = 7;
  212. rf->iwdev = iwdev;
  213. mutex_init(&iwdev->ah_tbl_lock);
  214. iwdev->netdev = vsi->netdev;
  215. iwdev->vsi_num = vsi->vsi_num;
  216. iwdev->init_state = INITIAL_STATE;
  217. iwdev->roce_cwnd = IRDMA_ROCE_CWND_DEFAULT;
  218. iwdev->roce_ackcreds = IRDMA_ROCE_ACKCREDS_DEFAULT;
  219. iwdev->rcv_wnd = IRDMA_CM_DEFAULT_RCV_WND_SCALED;
  220. iwdev->rcv_wscale = IRDMA_CM_DEFAULT_RCV_WND_SCALE;
  221. if (rf->protocol_used == IRDMA_ROCE_PROTOCOL_ONLY)
  222. iwdev->roce_mode = true;
  223. }
  224. static int irdma_probe(struct auxiliary_device *aux_dev, const struct auxiliary_device_id *id)
  225. {
  226. struct iidc_auxiliary_dev *iidc_adev = container_of(aux_dev,
  227. struct iidc_auxiliary_dev,
  228. adev);
  229. struct ice_pf *pf = iidc_adev->pf;
  230. struct ice_vsi *vsi = ice_get_main_vsi(pf);
  231. struct iidc_qos_params qos_info = {};
  232. struct irdma_device *iwdev;
  233. struct irdma_pci_f *rf;
  234. struct irdma_l2params l2params = {};
  235. int err;
  236. if (!vsi)
  237. return -EIO;
  238. iwdev = ib_alloc_device(irdma_device, ibdev);
  239. if (!iwdev)
  240. return -ENOMEM;
  241. iwdev->rf = kzalloc(sizeof(*rf), GFP_KERNEL);
  242. if (!iwdev->rf) {
  243. ib_dealloc_device(&iwdev->ibdev);
  244. return -ENOMEM;
  245. }
  246. irdma_fill_device_info(iwdev, pf, vsi);
  247. rf = iwdev->rf;
  248. err = irdma_ctrl_init_hw(rf);
  249. if (err)
  250. goto err_ctrl_init;
  251. l2params.mtu = iwdev->netdev->mtu;
  252. ice_get_qos_params(pf, &qos_info);
  253. irdma_fill_qos_info(&l2params, &qos_info);
  254. if (iwdev->rf->protocol_used != IRDMA_IWARP_PROTOCOL_ONLY)
  255. iwdev->dcb_vlan_mode = l2params.num_tc > 1 && !l2params.dscp_mode;
  256. err = irdma_rt_init_hw(iwdev, &l2params);
  257. if (err)
  258. goto err_rt_init;
  259. err = irdma_ib_register_device(iwdev);
  260. if (err)
  261. goto err_ibreg;
  262. ice_rdma_update_vsi_filter(pf, iwdev->vsi_num, true);
  263. ibdev_dbg(&iwdev->ibdev, "INIT: Gen2 PF[%d] device probe success\n", PCI_FUNC(rf->pcidev->devfn));
  264. auxiliary_set_drvdata(aux_dev, iwdev);
  265. return 0;
  266. err_ibreg:
  267. irdma_rt_deinit_hw(iwdev);
  268. err_rt_init:
  269. irdma_ctrl_deinit_hw(rf);
  270. err_ctrl_init:
  271. kfree(iwdev->rf);
  272. ib_dealloc_device(&iwdev->ibdev);
  273. return err;
  274. }
  275. static const struct auxiliary_device_id irdma_auxiliary_id_table[] = {
  276. {.name = "ice.iwarp", },
  277. {.name = "ice.roce", },
  278. {},
  279. };
  280. MODULE_DEVICE_TABLE(auxiliary, irdma_auxiliary_id_table);
  281. static struct iidc_auxiliary_drv irdma_auxiliary_drv = {
  282. .adrv = {
  283. .id_table = irdma_auxiliary_id_table,
  284. .probe = irdma_probe,
  285. .remove = irdma_remove,
  286. },
  287. .event_handler = irdma_iidc_event_handler,
  288. };
  289. static int __init irdma_init_module(void)
  290. {
  291. int ret;
  292. ret = auxiliary_driver_register(&i40iw_auxiliary_drv);
  293. if (ret) {
  294. pr_err("Failed i40iw(gen_1) auxiliary_driver_register() ret=%d\n",
  295. ret);
  296. return ret;
  297. }
  298. ret = auxiliary_driver_register(&irdma_auxiliary_drv.adrv);
  299. if (ret) {
  300. auxiliary_driver_unregister(&i40iw_auxiliary_drv);
  301. pr_err("Failed irdma auxiliary_driver_register() ret=%d\n",
  302. ret);
  303. return ret;
  304. }
  305. irdma_register_notifiers();
  306. return 0;
  307. }
  308. static void __exit irdma_exit_module(void)
  309. {
  310. irdma_unregister_notifiers();
  311. auxiliary_driver_unregister(&irdma_auxiliary_drv.adrv);
  312. auxiliary_driver_unregister(&i40iw_auxiliary_drv);
  313. }
  314. module_init(irdma_init_module);
  315. module_exit(irdma_exit_module);