mhi_wwan_ctrl.c 7.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285
  1. // SPDX-License-Identifier: GPL-2.0-only
  2. /* Copyright (c) 2021, Linaro Ltd <[email protected]> */
  3. #include <linux/kernel.h>
  4. #include <linux/mhi.h>
  5. #include <linux/mod_devicetable.h>
  6. #include <linux/module.h>
  7. #include <linux/wwan.h>
  8. /* MHI wwan flags */
  9. enum mhi_wwan_flags {
  10. MHI_WWAN_DL_CAP,
  11. MHI_WWAN_UL_CAP,
  12. MHI_WWAN_RX_REFILL,
  13. };
  14. #define MHI_WWAN_MAX_MTU 0x8000
  15. struct mhi_wwan_dev {
  16. /* Lower level is a mhi dev, upper level is a wwan port */
  17. struct mhi_device *mhi_dev;
  18. struct wwan_port *wwan_port;
  19. /* State and capabilities */
  20. unsigned long flags;
  21. size_t mtu;
  22. /* Protect against concurrent TX and TX-completion (bh) */
  23. spinlock_t tx_lock;
  24. /* Protect RX budget and rx_refill scheduling */
  25. spinlock_t rx_lock;
  26. struct work_struct rx_refill;
  27. /* RX budget is initially set to the size of the MHI RX queue and is
  28. * used to limit the number of allocated and queued packets. It is
  29. * decremented on data queueing and incremented on data release.
  30. */
  31. unsigned int rx_budget;
  32. };
  33. /* Increment RX budget and schedule RX refill if necessary */
  34. static void mhi_wwan_rx_budget_inc(struct mhi_wwan_dev *mhiwwan)
  35. {
  36. spin_lock_bh(&mhiwwan->rx_lock);
  37. mhiwwan->rx_budget++;
  38. if (test_bit(MHI_WWAN_RX_REFILL, &mhiwwan->flags))
  39. schedule_work(&mhiwwan->rx_refill);
  40. spin_unlock_bh(&mhiwwan->rx_lock);
  41. }
  42. /* Decrement RX budget if non-zero and return true on success */
  43. static bool mhi_wwan_rx_budget_dec(struct mhi_wwan_dev *mhiwwan)
  44. {
  45. bool ret = false;
  46. spin_lock_bh(&mhiwwan->rx_lock);
  47. if (mhiwwan->rx_budget) {
  48. mhiwwan->rx_budget--;
  49. if (test_bit(MHI_WWAN_RX_REFILL, &mhiwwan->flags))
  50. ret = true;
  51. }
  52. spin_unlock_bh(&mhiwwan->rx_lock);
  53. return ret;
  54. }
  55. static void __mhi_skb_destructor(struct sk_buff *skb)
  56. {
  57. /* RX buffer has been consumed, increase the allowed budget */
  58. mhi_wwan_rx_budget_inc(skb_shinfo(skb)->destructor_arg);
  59. }
  60. static void mhi_wwan_ctrl_refill_work(struct work_struct *work)
  61. {
  62. struct mhi_wwan_dev *mhiwwan = container_of(work, struct mhi_wwan_dev, rx_refill);
  63. struct mhi_device *mhi_dev = mhiwwan->mhi_dev;
  64. while (mhi_wwan_rx_budget_dec(mhiwwan)) {
  65. struct sk_buff *skb;
  66. skb = alloc_skb(mhiwwan->mtu, GFP_KERNEL);
  67. if (!skb) {
  68. mhi_wwan_rx_budget_inc(mhiwwan);
  69. break;
  70. }
  71. /* To prevent unlimited buffer allocation if nothing consumes
  72. * the RX buffers (passed to WWAN core), track their lifespan
  73. * to not allocate more than allowed budget.
  74. */
  75. skb->destructor = __mhi_skb_destructor;
  76. skb_shinfo(skb)->destructor_arg = mhiwwan;
  77. if (mhi_queue_skb(mhi_dev, DMA_FROM_DEVICE, skb, mhiwwan->mtu, MHI_EOT)) {
  78. dev_err(&mhi_dev->dev, "Failed to queue buffer\n");
  79. kfree_skb(skb);
  80. break;
  81. }
  82. }
  83. }
  84. static int mhi_wwan_ctrl_start(struct wwan_port *port)
  85. {
  86. struct mhi_wwan_dev *mhiwwan = wwan_port_get_drvdata(port);
  87. int ret;
  88. /* Start mhi device's channel(s) */
  89. ret = mhi_prepare_for_transfer(mhiwwan->mhi_dev);
  90. if (ret)
  91. return ret;
  92. /* Don't allocate more buffers than MHI channel queue size */
  93. mhiwwan->rx_budget = mhi_get_free_desc_count(mhiwwan->mhi_dev, DMA_FROM_DEVICE);
  94. /* Add buffers to the MHI inbound queue */
  95. if (test_bit(MHI_WWAN_DL_CAP, &mhiwwan->flags)) {
  96. set_bit(MHI_WWAN_RX_REFILL, &mhiwwan->flags);
  97. mhi_wwan_ctrl_refill_work(&mhiwwan->rx_refill);
  98. }
  99. return 0;
  100. }
  101. static void mhi_wwan_ctrl_stop(struct wwan_port *port)
  102. {
  103. struct mhi_wwan_dev *mhiwwan = wwan_port_get_drvdata(port);
  104. spin_lock_bh(&mhiwwan->rx_lock);
  105. clear_bit(MHI_WWAN_RX_REFILL, &mhiwwan->flags);
  106. spin_unlock_bh(&mhiwwan->rx_lock);
  107. cancel_work_sync(&mhiwwan->rx_refill);
  108. mhi_unprepare_from_transfer(mhiwwan->mhi_dev);
  109. }
  110. static int mhi_wwan_ctrl_tx(struct wwan_port *port, struct sk_buff *skb)
  111. {
  112. struct mhi_wwan_dev *mhiwwan = wwan_port_get_drvdata(port);
  113. int ret;
  114. if (skb->len > mhiwwan->mtu)
  115. return -EMSGSIZE;
  116. if (!test_bit(MHI_WWAN_UL_CAP, &mhiwwan->flags))
  117. return -EOPNOTSUPP;
  118. /* Queue the packet for MHI transfer and check fullness of the queue */
  119. spin_lock_bh(&mhiwwan->tx_lock);
  120. ret = mhi_queue_skb(mhiwwan->mhi_dev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT);
  121. if (mhi_queue_is_full(mhiwwan->mhi_dev, DMA_TO_DEVICE))
  122. wwan_port_txoff(port);
  123. spin_unlock_bh(&mhiwwan->tx_lock);
  124. return ret;
  125. }
  126. static const struct wwan_port_ops wwan_pops = {
  127. .start = mhi_wwan_ctrl_start,
  128. .stop = mhi_wwan_ctrl_stop,
  129. .tx = mhi_wwan_ctrl_tx,
  130. };
  131. static void mhi_ul_xfer_cb(struct mhi_device *mhi_dev,
  132. struct mhi_result *mhi_result)
  133. {
  134. struct mhi_wwan_dev *mhiwwan = dev_get_drvdata(&mhi_dev->dev);
  135. struct wwan_port *port = mhiwwan->wwan_port;
  136. struct sk_buff *skb = mhi_result->buf_addr;
  137. dev_dbg(&mhi_dev->dev, "%s: status: %d xfer_len: %zu\n", __func__,
  138. mhi_result->transaction_status, mhi_result->bytes_xferd);
  139. /* MHI core has done with the buffer, release it */
  140. consume_skb(skb);
  141. /* There is likely new slot available in the MHI queue, re-allow TX */
  142. spin_lock_bh(&mhiwwan->tx_lock);
  143. if (!mhi_queue_is_full(mhiwwan->mhi_dev, DMA_TO_DEVICE))
  144. wwan_port_txon(port);
  145. spin_unlock_bh(&mhiwwan->tx_lock);
  146. }
  147. static void mhi_dl_xfer_cb(struct mhi_device *mhi_dev,
  148. struct mhi_result *mhi_result)
  149. {
  150. struct mhi_wwan_dev *mhiwwan = dev_get_drvdata(&mhi_dev->dev);
  151. struct wwan_port *port = mhiwwan->wwan_port;
  152. struct sk_buff *skb = mhi_result->buf_addr;
  153. dev_dbg(&mhi_dev->dev, "%s: status: %d receive_len: %zu\n", __func__,
  154. mhi_result->transaction_status, mhi_result->bytes_xferd);
  155. if (mhi_result->transaction_status &&
  156. mhi_result->transaction_status != -EOVERFLOW) {
  157. kfree_skb(skb);
  158. return;
  159. }
  160. /* MHI core does not update skb->len, do it before forward */
  161. skb_put(skb, mhi_result->bytes_xferd);
  162. wwan_port_rx(port, skb);
  163. /* Do not increment rx budget nor refill RX buffers now, wait for the
  164. * buffer to be consumed. Done from __mhi_skb_destructor().
  165. */
  166. }
  167. static int mhi_wwan_ctrl_probe(struct mhi_device *mhi_dev,
  168. const struct mhi_device_id *id)
  169. {
  170. struct mhi_controller *cntrl = mhi_dev->mhi_cntrl;
  171. struct mhi_wwan_dev *mhiwwan;
  172. struct wwan_port *port;
  173. mhiwwan = kzalloc(sizeof(*mhiwwan), GFP_KERNEL);
  174. if (!mhiwwan)
  175. return -ENOMEM;
  176. mhiwwan->mhi_dev = mhi_dev;
  177. mhiwwan->mtu = MHI_WWAN_MAX_MTU;
  178. INIT_WORK(&mhiwwan->rx_refill, mhi_wwan_ctrl_refill_work);
  179. spin_lock_init(&mhiwwan->tx_lock);
  180. spin_lock_init(&mhiwwan->rx_lock);
  181. if (mhi_dev->dl_chan)
  182. set_bit(MHI_WWAN_DL_CAP, &mhiwwan->flags);
  183. if (mhi_dev->ul_chan)
  184. set_bit(MHI_WWAN_UL_CAP, &mhiwwan->flags);
  185. dev_set_drvdata(&mhi_dev->dev, mhiwwan);
  186. /* Register as a wwan port, id->driver_data contains wwan port type */
  187. port = wwan_create_port(&cntrl->mhi_dev->dev, id->driver_data,
  188. &wwan_pops, mhiwwan);
  189. if (IS_ERR(port)) {
  190. kfree(mhiwwan);
  191. return PTR_ERR(port);
  192. }
  193. mhiwwan->wwan_port = port;
  194. return 0;
  195. };
  196. static void mhi_wwan_ctrl_remove(struct mhi_device *mhi_dev)
  197. {
  198. struct mhi_wwan_dev *mhiwwan = dev_get_drvdata(&mhi_dev->dev);
  199. wwan_remove_port(mhiwwan->wwan_port);
  200. kfree(mhiwwan);
  201. }
  202. static const struct mhi_device_id mhi_wwan_ctrl_match_table[] = {
  203. { .chan = "DUN", .driver_data = WWAN_PORT_AT },
  204. { .chan = "DUN2", .driver_data = WWAN_PORT_AT },
  205. { .chan = "MBIM", .driver_data = WWAN_PORT_MBIM },
  206. { .chan = "QMI", .driver_data = WWAN_PORT_QMI },
  207. { .chan = "DIAG", .driver_data = WWAN_PORT_QCDM },
  208. { .chan = "FIREHOSE", .driver_data = WWAN_PORT_FIREHOSE },
  209. {},
  210. };
  211. MODULE_DEVICE_TABLE(mhi, mhi_wwan_ctrl_match_table);
  212. static struct mhi_driver mhi_wwan_ctrl_driver = {
  213. .id_table = mhi_wwan_ctrl_match_table,
  214. .remove = mhi_wwan_ctrl_remove,
  215. .probe = mhi_wwan_ctrl_probe,
  216. .ul_xfer_cb = mhi_ul_xfer_cb,
  217. .dl_xfer_cb = mhi_dl_xfer_cb,
  218. .driver = {
  219. .name = "mhi_wwan_ctrl",
  220. },
  221. };
  222. module_mhi_driver(mhi_wwan_ctrl_driver);
  223. MODULE_LICENSE("GPL v2");
  224. MODULE_DESCRIPTION("MHI WWAN CTRL Driver");
  225. MODULE_AUTHOR("Loic Poulain <[email protected]>");