testmode.c 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676
  1. // SPDX-License-Identifier: ISC
  2. /* Copyright (C) 2020 Felix Fietkau <[email protected]> */
  3. #include <linux/random.h>
  4. #include "mt76.h"
  5. const struct nla_policy mt76_tm_policy[NUM_MT76_TM_ATTRS] = {
  6. [MT76_TM_ATTR_RESET] = { .type = NLA_FLAG },
  7. [MT76_TM_ATTR_STATE] = { .type = NLA_U8 },
  8. [MT76_TM_ATTR_TX_COUNT] = { .type = NLA_U32 },
  9. [MT76_TM_ATTR_TX_LENGTH] = { .type = NLA_U32 },
  10. [MT76_TM_ATTR_TX_RATE_MODE] = { .type = NLA_U8 },
  11. [MT76_TM_ATTR_TX_RATE_NSS] = { .type = NLA_U8 },
  12. [MT76_TM_ATTR_TX_RATE_IDX] = { .type = NLA_U8 },
  13. [MT76_TM_ATTR_TX_RATE_SGI] = { .type = NLA_U8 },
  14. [MT76_TM_ATTR_TX_RATE_LDPC] = { .type = NLA_U8 },
  15. [MT76_TM_ATTR_TX_RATE_STBC] = { .type = NLA_U8 },
  16. [MT76_TM_ATTR_TX_LTF] = { .type = NLA_U8 },
  17. [MT76_TM_ATTR_TX_ANTENNA] = { .type = NLA_U8 },
  18. [MT76_TM_ATTR_TX_SPE_IDX] = { .type = NLA_U8 },
  19. [MT76_TM_ATTR_TX_POWER_CONTROL] = { .type = NLA_U8 },
  20. [MT76_TM_ATTR_TX_POWER] = { .type = NLA_NESTED },
  21. [MT76_TM_ATTR_TX_DUTY_CYCLE] = { .type = NLA_U8 },
  22. [MT76_TM_ATTR_TX_IPG] = { .type = NLA_U32 },
  23. [MT76_TM_ATTR_TX_TIME] = { .type = NLA_U32 },
  24. [MT76_TM_ATTR_FREQ_OFFSET] = { .type = NLA_U32 },
  25. [MT76_TM_ATTR_DRV_DATA] = { .type = NLA_NESTED },
  26. };
  27. EXPORT_SYMBOL_GPL(mt76_tm_policy);
  28. void mt76_testmode_tx_pending(struct mt76_phy *phy)
  29. {
  30. struct mt76_testmode_data *td = &phy->test;
  31. struct mt76_dev *dev = phy->dev;
  32. struct mt76_wcid *wcid = &dev->global_wcid;
  33. struct sk_buff *skb = td->tx_skb;
  34. struct mt76_queue *q;
  35. u16 tx_queued_limit;
  36. int qid;
  37. if (!skb || !td->tx_pending)
  38. return;
  39. qid = skb_get_queue_mapping(skb);
  40. q = phy->q_tx[qid];
  41. tx_queued_limit = td->tx_queued_limit ? td->tx_queued_limit : 1000;
  42. spin_lock_bh(&q->lock);
  43. while (td->tx_pending > 0 &&
  44. td->tx_queued - td->tx_done < tx_queued_limit &&
  45. q->queued < q->ndesc / 2) {
  46. int ret;
  47. ret = dev->queue_ops->tx_queue_skb(dev, q, qid, skb_get(skb),
  48. wcid, NULL);
  49. if (ret < 0)
  50. break;
  51. td->tx_pending--;
  52. td->tx_queued++;
  53. }
  54. dev->queue_ops->kick(dev, q);
  55. spin_unlock_bh(&q->lock);
  56. }
  57. static u32
  58. mt76_testmode_max_mpdu_len(struct mt76_phy *phy, u8 tx_rate_mode)
  59. {
  60. switch (tx_rate_mode) {
  61. case MT76_TM_TX_MODE_HT:
  62. return IEEE80211_MAX_MPDU_LEN_HT_7935;
  63. case MT76_TM_TX_MODE_VHT:
  64. case MT76_TM_TX_MODE_HE_SU:
  65. case MT76_TM_TX_MODE_HE_EXT_SU:
  66. case MT76_TM_TX_MODE_HE_TB:
  67. case MT76_TM_TX_MODE_HE_MU:
  68. if (phy->sband_5g.sband.vht_cap.cap &
  69. IEEE80211_VHT_CAP_MAX_MPDU_LENGTH_7991)
  70. return IEEE80211_MAX_MPDU_LEN_VHT_7991;
  71. return IEEE80211_MAX_MPDU_LEN_VHT_11454;
  72. case MT76_TM_TX_MODE_CCK:
  73. case MT76_TM_TX_MODE_OFDM:
  74. default:
  75. return IEEE80211_MAX_FRAME_LEN;
  76. }
  77. }
  78. static void
  79. mt76_testmode_free_skb(struct mt76_phy *phy)
  80. {
  81. struct mt76_testmode_data *td = &phy->test;
  82. dev_kfree_skb(td->tx_skb);
  83. td->tx_skb = NULL;
  84. }
  85. int mt76_testmode_alloc_skb(struct mt76_phy *phy, u32 len)
  86. {
  87. #define MT_TXP_MAX_LEN 4095
  88. u16 fc = IEEE80211_FTYPE_DATA | IEEE80211_STYPE_DATA |
  89. IEEE80211_FCTL_FROMDS;
  90. struct mt76_testmode_data *td = &phy->test;
  91. struct sk_buff **frag_tail, *head;
  92. struct ieee80211_tx_info *info;
  93. struct ieee80211_hdr *hdr;
  94. u32 max_len, head_len;
  95. int nfrags, i;
  96. max_len = mt76_testmode_max_mpdu_len(phy, td->tx_rate_mode);
  97. if (len > max_len)
  98. len = max_len;
  99. else if (len < sizeof(struct ieee80211_hdr))
  100. len = sizeof(struct ieee80211_hdr);
  101. nfrags = len / MT_TXP_MAX_LEN;
  102. head_len = nfrags ? MT_TXP_MAX_LEN : len;
  103. if (len > IEEE80211_MAX_FRAME_LEN)
  104. fc |= IEEE80211_STYPE_QOS_DATA;
  105. head = alloc_skb(head_len, GFP_KERNEL);
  106. if (!head)
  107. return -ENOMEM;
  108. hdr = __skb_put_zero(head, sizeof(*hdr));
  109. hdr->frame_control = cpu_to_le16(fc);
  110. memcpy(hdr->addr1, td->addr[0], ETH_ALEN);
  111. memcpy(hdr->addr2, td->addr[1], ETH_ALEN);
  112. memcpy(hdr->addr3, td->addr[2], ETH_ALEN);
  113. skb_set_queue_mapping(head, IEEE80211_AC_BE);
  114. get_random_bytes(__skb_put(head, head_len - sizeof(*hdr)),
  115. head_len - sizeof(*hdr));
  116. info = IEEE80211_SKB_CB(head);
  117. info->flags = IEEE80211_TX_CTL_INJECTED |
  118. IEEE80211_TX_CTL_NO_ACK |
  119. IEEE80211_TX_CTL_NO_PS_BUFFER;
  120. info->hw_queue |= FIELD_PREP(MT_TX_HW_QUEUE_PHY, phy->band_idx);
  121. frag_tail = &skb_shinfo(head)->frag_list;
  122. for (i = 0; i < nfrags; i++) {
  123. struct sk_buff *frag;
  124. u16 frag_len;
  125. if (i == nfrags - 1)
  126. frag_len = len % MT_TXP_MAX_LEN;
  127. else
  128. frag_len = MT_TXP_MAX_LEN;
  129. frag = alloc_skb(frag_len, GFP_KERNEL);
  130. if (!frag) {
  131. mt76_testmode_free_skb(phy);
  132. dev_kfree_skb(head);
  133. return -ENOMEM;
  134. }
  135. get_random_bytes(__skb_put(frag, frag_len), frag_len);
  136. head->len += frag->len;
  137. head->data_len += frag->len;
  138. *frag_tail = frag;
  139. frag_tail = &(*frag_tail)->next;
  140. }
  141. mt76_testmode_free_skb(phy);
  142. td->tx_skb = head;
  143. return 0;
  144. }
  145. EXPORT_SYMBOL(mt76_testmode_alloc_skb);
  146. static int
  147. mt76_testmode_tx_init(struct mt76_phy *phy)
  148. {
  149. struct mt76_testmode_data *td = &phy->test;
  150. struct ieee80211_tx_info *info;
  151. struct ieee80211_tx_rate *rate;
  152. u8 max_nss = hweight8(phy->antenna_mask);
  153. int ret;
  154. ret = mt76_testmode_alloc_skb(phy, td->tx_mpdu_len);
  155. if (ret)
  156. return ret;
  157. if (td->tx_rate_mode > MT76_TM_TX_MODE_VHT)
  158. goto out;
  159. if (td->tx_antenna_mask)
  160. max_nss = min_t(u8, max_nss, hweight8(td->tx_antenna_mask));
  161. info = IEEE80211_SKB_CB(td->tx_skb);
  162. rate = &info->control.rates[0];
  163. rate->count = 1;
  164. rate->idx = td->tx_rate_idx;
  165. switch (td->tx_rate_mode) {
  166. case MT76_TM_TX_MODE_CCK:
  167. if (phy->chandef.chan->band != NL80211_BAND_2GHZ)
  168. return -EINVAL;
  169. if (rate->idx > 4)
  170. return -EINVAL;
  171. break;
  172. case MT76_TM_TX_MODE_OFDM:
  173. if (phy->chandef.chan->band != NL80211_BAND_2GHZ)
  174. break;
  175. if (rate->idx > 8)
  176. return -EINVAL;
  177. rate->idx += 4;
  178. break;
  179. case MT76_TM_TX_MODE_HT:
  180. if (rate->idx > 8 * max_nss &&
  181. !(rate->idx == 32 &&
  182. phy->chandef.width >= NL80211_CHAN_WIDTH_40))
  183. return -EINVAL;
  184. rate->flags |= IEEE80211_TX_RC_MCS;
  185. break;
  186. case MT76_TM_TX_MODE_VHT:
  187. if (rate->idx > 9)
  188. return -EINVAL;
  189. if (td->tx_rate_nss > max_nss)
  190. return -EINVAL;
  191. ieee80211_rate_set_vht(rate, td->tx_rate_idx, td->tx_rate_nss);
  192. rate->flags |= IEEE80211_TX_RC_VHT_MCS;
  193. break;
  194. default:
  195. break;
  196. }
  197. if (td->tx_rate_sgi)
  198. rate->flags |= IEEE80211_TX_RC_SHORT_GI;
  199. if (td->tx_rate_ldpc)
  200. info->flags |= IEEE80211_TX_CTL_LDPC;
  201. if (td->tx_rate_stbc)
  202. info->flags |= IEEE80211_TX_CTL_STBC;
  203. if (td->tx_rate_mode >= MT76_TM_TX_MODE_HT) {
  204. switch (phy->chandef.width) {
  205. case NL80211_CHAN_WIDTH_40:
  206. rate->flags |= IEEE80211_TX_RC_40_MHZ_WIDTH;
  207. break;
  208. case NL80211_CHAN_WIDTH_80:
  209. rate->flags |= IEEE80211_TX_RC_80_MHZ_WIDTH;
  210. break;
  211. case NL80211_CHAN_WIDTH_80P80:
  212. case NL80211_CHAN_WIDTH_160:
  213. rate->flags |= IEEE80211_TX_RC_160_MHZ_WIDTH;
  214. break;
  215. default:
  216. break;
  217. }
  218. }
  219. out:
  220. return 0;
  221. }
  222. static void
  223. mt76_testmode_tx_start(struct mt76_phy *phy)
  224. {
  225. struct mt76_testmode_data *td = &phy->test;
  226. struct mt76_dev *dev = phy->dev;
  227. td->tx_queued = 0;
  228. td->tx_done = 0;
  229. td->tx_pending = td->tx_count;
  230. mt76_worker_schedule(&dev->tx_worker);
  231. }
  232. static void
  233. mt76_testmode_tx_stop(struct mt76_phy *phy)
  234. {
  235. struct mt76_testmode_data *td = &phy->test;
  236. struct mt76_dev *dev = phy->dev;
  237. mt76_worker_disable(&dev->tx_worker);
  238. td->tx_pending = 0;
  239. mt76_worker_enable(&dev->tx_worker);
  240. wait_event_timeout(dev->tx_wait, td->tx_done == td->tx_queued,
  241. MT76_TM_TIMEOUT * HZ);
  242. mt76_testmode_free_skb(phy);
  243. }
  244. static inline void
  245. mt76_testmode_param_set(struct mt76_testmode_data *td, u16 idx)
  246. {
  247. td->param_set[idx / 32] |= BIT(idx % 32);
  248. }
  249. static inline bool
  250. mt76_testmode_param_present(struct mt76_testmode_data *td, u16 idx)
  251. {
  252. return td->param_set[idx / 32] & BIT(idx % 32);
  253. }
  254. static void
  255. mt76_testmode_init_defaults(struct mt76_phy *phy)
  256. {
  257. struct mt76_testmode_data *td = &phy->test;
  258. if (td->tx_mpdu_len > 0)
  259. return;
  260. td->tx_mpdu_len = 1024;
  261. td->tx_count = 1;
  262. td->tx_rate_mode = MT76_TM_TX_MODE_OFDM;
  263. td->tx_rate_nss = 1;
  264. memcpy(td->addr[0], phy->macaddr, ETH_ALEN);
  265. memcpy(td->addr[1], phy->macaddr, ETH_ALEN);
  266. memcpy(td->addr[2], phy->macaddr, ETH_ALEN);
  267. }
  268. static int
  269. __mt76_testmode_set_state(struct mt76_phy *phy, enum mt76_testmode_state state)
  270. {
  271. enum mt76_testmode_state prev_state = phy->test.state;
  272. struct mt76_dev *dev = phy->dev;
  273. int err;
  274. if (prev_state == MT76_TM_STATE_TX_FRAMES)
  275. mt76_testmode_tx_stop(phy);
  276. if (state == MT76_TM_STATE_TX_FRAMES) {
  277. err = mt76_testmode_tx_init(phy);
  278. if (err)
  279. return err;
  280. }
  281. err = dev->test_ops->set_state(phy, state);
  282. if (err) {
  283. if (state == MT76_TM_STATE_TX_FRAMES)
  284. mt76_testmode_tx_stop(phy);
  285. return err;
  286. }
  287. if (state == MT76_TM_STATE_TX_FRAMES)
  288. mt76_testmode_tx_start(phy);
  289. else if (state == MT76_TM_STATE_RX_FRAMES) {
  290. memset(&phy->test.rx_stats, 0, sizeof(phy->test.rx_stats));
  291. }
  292. phy->test.state = state;
  293. return 0;
  294. }
  295. int mt76_testmode_set_state(struct mt76_phy *phy, enum mt76_testmode_state state)
  296. {
  297. struct mt76_testmode_data *td = &phy->test;
  298. struct ieee80211_hw *hw = phy->hw;
  299. if (state == td->state && state == MT76_TM_STATE_OFF)
  300. return 0;
  301. if (state > MT76_TM_STATE_OFF &&
  302. (!test_bit(MT76_STATE_RUNNING, &phy->state) ||
  303. !(hw->conf.flags & IEEE80211_CONF_MONITOR)))
  304. return -ENOTCONN;
  305. if (state != MT76_TM_STATE_IDLE &&
  306. td->state != MT76_TM_STATE_IDLE) {
  307. int ret;
  308. ret = __mt76_testmode_set_state(phy, MT76_TM_STATE_IDLE);
  309. if (ret)
  310. return ret;
  311. }
  312. return __mt76_testmode_set_state(phy, state);
  313. }
  314. EXPORT_SYMBOL(mt76_testmode_set_state);
  315. static int
  316. mt76_tm_get_u8(struct nlattr *attr, u8 *dest, u8 min, u8 max)
  317. {
  318. u8 val;
  319. if (!attr)
  320. return 0;
  321. val = nla_get_u8(attr);
  322. if (val < min || val > max)
  323. return -EINVAL;
  324. *dest = val;
  325. return 0;
  326. }
  327. int mt76_testmode_cmd(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
  328. void *data, int len)
  329. {
  330. struct mt76_phy *phy = hw->priv;
  331. struct mt76_dev *dev = phy->dev;
  332. struct mt76_testmode_data *td = &phy->test;
  333. struct nlattr *tb[NUM_MT76_TM_ATTRS];
  334. u32 state;
  335. int err;
  336. int i;
  337. if (!dev->test_ops)
  338. return -EOPNOTSUPP;
  339. err = nla_parse_deprecated(tb, MT76_TM_ATTR_MAX, data, len,
  340. mt76_tm_policy, NULL);
  341. if (err)
  342. return err;
  343. err = -EINVAL;
  344. mutex_lock(&dev->mutex);
  345. if (tb[MT76_TM_ATTR_RESET]) {
  346. mt76_testmode_set_state(phy, MT76_TM_STATE_OFF);
  347. memset(td, 0, sizeof(*td));
  348. }
  349. mt76_testmode_init_defaults(phy);
  350. if (tb[MT76_TM_ATTR_TX_COUNT])
  351. td->tx_count = nla_get_u32(tb[MT76_TM_ATTR_TX_COUNT]);
  352. if (tb[MT76_TM_ATTR_TX_RATE_IDX])
  353. td->tx_rate_idx = nla_get_u8(tb[MT76_TM_ATTR_TX_RATE_IDX]);
  354. if (mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_RATE_MODE], &td->tx_rate_mode,
  355. 0, MT76_TM_TX_MODE_MAX) ||
  356. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_RATE_NSS], &td->tx_rate_nss,
  357. 1, hweight8(phy->antenna_mask)) ||
  358. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_RATE_SGI], &td->tx_rate_sgi, 0, 2) ||
  359. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_RATE_LDPC], &td->tx_rate_ldpc, 0, 1) ||
  360. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_RATE_STBC], &td->tx_rate_stbc, 0, 1) ||
  361. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_LTF], &td->tx_ltf, 0, 2) ||
  362. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_ANTENNA],
  363. &td->tx_antenna_mask, 0, 0xff) ||
  364. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_SPE_IDX], &td->tx_spe_idx, 0, 27) ||
  365. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_DUTY_CYCLE],
  366. &td->tx_duty_cycle, 0, 99) ||
  367. mt76_tm_get_u8(tb[MT76_TM_ATTR_TX_POWER_CONTROL],
  368. &td->tx_power_control, 0, 1))
  369. goto out;
  370. if (tb[MT76_TM_ATTR_TX_LENGTH]) {
  371. u32 val = nla_get_u32(tb[MT76_TM_ATTR_TX_LENGTH]);
  372. if (val > mt76_testmode_max_mpdu_len(phy, td->tx_rate_mode) ||
  373. val < sizeof(struct ieee80211_hdr))
  374. goto out;
  375. td->tx_mpdu_len = val;
  376. }
  377. if (tb[MT76_TM_ATTR_TX_IPG])
  378. td->tx_ipg = nla_get_u32(tb[MT76_TM_ATTR_TX_IPG]);
  379. if (tb[MT76_TM_ATTR_TX_TIME])
  380. td->tx_time = nla_get_u32(tb[MT76_TM_ATTR_TX_TIME]);
  381. if (tb[MT76_TM_ATTR_FREQ_OFFSET])
  382. td->freq_offset = nla_get_u32(tb[MT76_TM_ATTR_FREQ_OFFSET]);
  383. if (tb[MT76_TM_ATTR_STATE]) {
  384. state = nla_get_u32(tb[MT76_TM_ATTR_STATE]);
  385. if (state > MT76_TM_STATE_MAX)
  386. goto out;
  387. } else {
  388. state = td->state;
  389. }
  390. if (tb[MT76_TM_ATTR_TX_POWER]) {
  391. struct nlattr *cur;
  392. int idx = 0;
  393. int rem;
  394. nla_for_each_nested(cur, tb[MT76_TM_ATTR_TX_POWER], rem) {
  395. if (nla_len(cur) != 1 ||
  396. idx >= ARRAY_SIZE(td->tx_power))
  397. goto out;
  398. td->tx_power[idx++] = nla_get_u8(cur);
  399. }
  400. }
  401. if (tb[MT76_TM_ATTR_MAC_ADDRS]) {
  402. struct nlattr *cur;
  403. int idx = 0;
  404. int rem;
  405. nla_for_each_nested(cur, tb[MT76_TM_ATTR_MAC_ADDRS], rem) {
  406. if (nla_len(cur) != ETH_ALEN || idx >= 3)
  407. goto out;
  408. memcpy(td->addr[idx], nla_data(cur), ETH_ALEN);
  409. idx++;
  410. }
  411. }
  412. if (dev->test_ops->set_params) {
  413. err = dev->test_ops->set_params(phy, tb, state);
  414. if (err)
  415. goto out;
  416. }
  417. for (i = MT76_TM_ATTR_STATE; i < ARRAY_SIZE(tb); i++)
  418. if (tb[i])
  419. mt76_testmode_param_set(td, i);
  420. err = 0;
  421. if (tb[MT76_TM_ATTR_STATE])
  422. err = mt76_testmode_set_state(phy, state);
  423. out:
  424. mutex_unlock(&dev->mutex);
  425. return err;
  426. }
  427. EXPORT_SYMBOL(mt76_testmode_cmd);
  428. static int
  429. mt76_testmode_dump_stats(struct mt76_phy *phy, struct sk_buff *msg)
  430. {
  431. struct mt76_testmode_data *td = &phy->test;
  432. struct mt76_dev *dev = phy->dev;
  433. u64 rx_packets = 0;
  434. u64 rx_fcs_error = 0;
  435. int i;
  436. if (dev->test_ops->dump_stats) {
  437. int ret;
  438. ret = dev->test_ops->dump_stats(phy, msg);
  439. if (ret)
  440. return ret;
  441. }
  442. for (i = 0; i < ARRAY_SIZE(td->rx_stats.packets); i++) {
  443. rx_packets += td->rx_stats.packets[i];
  444. rx_fcs_error += td->rx_stats.fcs_error[i];
  445. }
  446. if (nla_put_u32(msg, MT76_TM_STATS_ATTR_TX_PENDING, td->tx_pending) ||
  447. nla_put_u32(msg, MT76_TM_STATS_ATTR_TX_QUEUED, td->tx_queued) ||
  448. nla_put_u32(msg, MT76_TM_STATS_ATTR_TX_DONE, td->tx_done) ||
  449. nla_put_u64_64bit(msg, MT76_TM_STATS_ATTR_RX_PACKETS, rx_packets,
  450. MT76_TM_STATS_ATTR_PAD) ||
  451. nla_put_u64_64bit(msg, MT76_TM_STATS_ATTR_RX_FCS_ERROR, rx_fcs_error,
  452. MT76_TM_STATS_ATTR_PAD))
  453. return -EMSGSIZE;
  454. return 0;
  455. }
  456. int mt76_testmode_dump(struct ieee80211_hw *hw, struct sk_buff *msg,
  457. struct netlink_callback *cb, void *data, int len)
  458. {
  459. struct mt76_phy *phy = hw->priv;
  460. struct mt76_dev *dev = phy->dev;
  461. struct mt76_testmode_data *td = &phy->test;
  462. struct nlattr *tb[NUM_MT76_TM_ATTRS] = {};
  463. int err = 0;
  464. void *a;
  465. int i;
  466. if (!dev->test_ops)
  467. return -EOPNOTSUPP;
  468. if (cb->args[2]++ > 0)
  469. return -ENOENT;
  470. if (data) {
  471. err = nla_parse_deprecated(tb, MT76_TM_ATTR_MAX, data, len,
  472. mt76_tm_policy, NULL);
  473. if (err)
  474. return err;
  475. }
  476. mutex_lock(&dev->mutex);
  477. if (tb[MT76_TM_ATTR_STATS]) {
  478. err = -EINVAL;
  479. a = nla_nest_start(msg, MT76_TM_ATTR_STATS);
  480. if (a) {
  481. err = mt76_testmode_dump_stats(phy, msg);
  482. nla_nest_end(msg, a);
  483. }
  484. goto out;
  485. }
  486. mt76_testmode_init_defaults(phy);
  487. err = -EMSGSIZE;
  488. if (nla_put_u32(msg, MT76_TM_ATTR_STATE, td->state))
  489. goto out;
  490. if (dev->test_mtd.name &&
  491. (nla_put_string(msg, MT76_TM_ATTR_MTD_PART, dev->test_mtd.name) ||
  492. nla_put_u32(msg, MT76_TM_ATTR_MTD_OFFSET, dev->test_mtd.offset)))
  493. goto out;
  494. if (nla_put_u32(msg, MT76_TM_ATTR_TX_COUNT, td->tx_count) ||
  495. nla_put_u32(msg, MT76_TM_ATTR_TX_LENGTH, td->tx_mpdu_len) ||
  496. nla_put_u8(msg, MT76_TM_ATTR_TX_RATE_MODE, td->tx_rate_mode) ||
  497. nla_put_u8(msg, MT76_TM_ATTR_TX_RATE_NSS, td->tx_rate_nss) ||
  498. nla_put_u8(msg, MT76_TM_ATTR_TX_RATE_IDX, td->tx_rate_idx) ||
  499. nla_put_u8(msg, MT76_TM_ATTR_TX_RATE_SGI, td->tx_rate_sgi) ||
  500. nla_put_u8(msg, MT76_TM_ATTR_TX_RATE_LDPC, td->tx_rate_ldpc) ||
  501. nla_put_u8(msg, MT76_TM_ATTR_TX_RATE_STBC, td->tx_rate_stbc) ||
  502. (mt76_testmode_param_present(td, MT76_TM_ATTR_TX_LTF) &&
  503. nla_put_u8(msg, MT76_TM_ATTR_TX_LTF, td->tx_ltf)) ||
  504. (mt76_testmode_param_present(td, MT76_TM_ATTR_TX_ANTENNA) &&
  505. nla_put_u8(msg, MT76_TM_ATTR_TX_ANTENNA, td->tx_antenna_mask)) ||
  506. (mt76_testmode_param_present(td, MT76_TM_ATTR_TX_SPE_IDX) &&
  507. nla_put_u8(msg, MT76_TM_ATTR_TX_SPE_IDX, td->tx_spe_idx)) ||
  508. (mt76_testmode_param_present(td, MT76_TM_ATTR_TX_DUTY_CYCLE) &&
  509. nla_put_u8(msg, MT76_TM_ATTR_TX_DUTY_CYCLE, td->tx_duty_cycle)) ||
  510. (mt76_testmode_param_present(td, MT76_TM_ATTR_TX_IPG) &&
  511. nla_put_u32(msg, MT76_TM_ATTR_TX_IPG, td->tx_ipg)) ||
  512. (mt76_testmode_param_present(td, MT76_TM_ATTR_TX_TIME) &&
  513. nla_put_u32(msg, MT76_TM_ATTR_TX_TIME, td->tx_time)) ||
  514. (mt76_testmode_param_present(td, MT76_TM_ATTR_TX_POWER_CONTROL) &&
  515. nla_put_u8(msg, MT76_TM_ATTR_TX_POWER_CONTROL, td->tx_power_control)) ||
  516. (mt76_testmode_param_present(td, MT76_TM_ATTR_FREQ_OFFSET) &&
  517. nla_put_u8(msg, MT76_TM_ATTR_FREQ_OFFSET, td->freq_offset)))
  518. goto out;
  519. if (mt76_testmode_param_present(td, MT76_TM_ATTR_TX_POWER)) {
  520. a = nla_nest_start(msg, MT76_TM_ATTR_TX_POWER);
  521. if (!a)
  522. goto out;
  523. for (i = 0; i < ARRAY_SIZE(td->tx_power); i++)
  524. if (nla_put_u8(msg, i, td->tx_power[i]))
  525. goto out;
  526. nla_nest_end(msg, a);
  527. }
  528. if (mt76_testmode_param_present(td, MT76_TM_ATTR_MAC_ADDRS)) {
  529. a = nla_nest_start(msg, MT76_TM_ATTR_MAC_ADDRS);
  530. if (!a)
  531. goto out;
  532. for (i = 0; i < 3; i++)
  533. if (nla_put(msg, i, ETH_ALEN, td->addr[i]))
  534. goto out;
  535. nla_nest_end(msg, a);
  536. }
  537. err = 0;
  538. out:
  539. mutex_unlock(&dev->mutex);
  540. return err;
  541. }
  542. EXPORT_SYMBOL(mt76_testmode_dump);