dm.c 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607
  1. // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
  2. /*
  3. * Copyright (c) 2021, Mellanox Technologies inc. All rights reserved.
  4. */
  5. #include <rdma/uverbs_std_types.h>
  6. #include "dm.h"
  7. #define UVERBS_MODULE_NAME mlx5_ib
  8. #include <rdma/uverbs_named_ioctl.h>
  9. static int mlx5_cmd_alloc_memic(struct mlx5_dm *dm, phys_addr_t *addr,
  10. u64 length, u32 alignment)
  11. {
  12. struct mlx5_core_dev *dev = dm->dev;
  13. u64 num_memic_hw_pages = MLX5_CAP_DEV_MEM(dev, memic_bar_size)
  14. >> PAGE_SHIFT;
  15. u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr);
  16. u32 max_alignment = MLX5_CAP_DEV_MEM(dev, log_max_memic_addr_alignment);
  17. u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE);
  18. u32 out[MLX5_ST_SZ_DW(alloc_memic_out)] = {};
  19. u32 in[MLX5_ST_SZ_DW(alloc_memic_in)] = {};
  20. u32 mlx5_alignment;
  21. u64 page_idx = 0;
  22. int ret = 0;
  23. if (!length || (length & MLX5_MEMIC_ALLOC_SIZE_MASK))
  24. return -EINVAL;
  25. /* mlx5 device sets alignment as 64*2^driver_value
  26. * so normalizing is needed.
  27. */
  28. mlx5_alignment = (alignment < MLX5_MEMIC_BASE_ALIGN) ? 0 :
  29. alignment - MLX5_MEMIC_BASE_ALIGN;
  30. if (mlx5_alignment > max_alignment)
  31. return -EINVAL;
  32. MLX5_SET(alloc_memic_in, in, opcode, MLX5_CMD_OP_ALLOC_MEMIC);
  33. MLX5_SET(alloc_memic_in, in, range_size, num_pages * PAGE_SIZE);
  34. MLX5_SET(alloc_memic_in, in, memic_size, length);
  35. MLX5_SET(alloc_memic_in, in, log_memic_addr_alignment,
  36. mlx5_alignment);
  37. while (page_idx < num_memic_hw_pages) {
  38. spin_lock(&dm->lock);
  39. page_idx = bitmap_find_next_zero_area(dm->memic_alloc_pages,
  40. num_memic_hw_pages,
  41. page_idx,
  42. num_pages, 0);
  43. if (page_idx < num_memic_hw_pages)
  44. bitmap_set(dm->memic_alloc_pages,
  45. page_idx, num_pages);
  46. spin_unlock(&dm->lock);
  47. if (page_idx >= num_memic_hw_pages)
  48. break;
  49. MLX5_SET64(alloc_memic_in, in, range_start_addr,
  50. hw_start_addr + (page_idx * PAGE_SIZE));
  51. ret = mlx5_cmd_exec_inout(dev, alloc_memic, in, out);
  52. if (ret) {
  53. spin_lock(&dm->lock);
  54. bitmap_clear(dm->memic_alloc_pages,
  55. page_idx, num_pages);
  56. spin_unlock(&dm->lock);
  57. if (ret == -EAGAIN) {
  58. page_idx++;
  59. continue;
  60. }
  61. return ret;
  62. }
  63. *addr = dev->bar_addr +
  64. MLX5_GET64(alloc_memic_out, out, memic_start_addr);
  65. return 0;
  66. }
  67. return -ENOMEM;
  68. }
  69. void mlx5_cmd_dealloc_memic(struct mlx5_dm *dm, phys_addr_t addr,
  70. u64 length)
  71. {
  72. struct mlx5_core_dev *dev = dm->dev;
  73. u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr);
  74. u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE);
  75. u32 in[MLX5_ST_SZ_DW(dealloc_memic_in)] = {};
  76. u64 start_page_idx;
  77. int err;
  78. addr -= dev->bar_addr;
  79. start_page_idx = (addr - hw_start_addr) >> PAGE_SHIFT;
  80. MLX5_SET(dealloc_memic_in, in, opcode, MLX5_CMD_OP_DEALLOC_MEMIC);
  81. MLX5_SET64(dealloc_memic_in, in, memic_start_addr, addr);
  82. MLX5_SET(dealloc_memic_in, in, memic_size, length);
  83. err = mlx5_cmd_exec_in(dev, dealloc_memic, in);
  84. if (err)
  85. return;
  86. spin_lock(&dm->lock);
  87. bitmap_clear(dm->memic_alloc_pages,
  88. start_page_idx, num_pages);
  89. spin_unlock(&dm->lock);
  90. }
  91. void mlx5_cmd_dealloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr,
  92. u8 operation)
  93. {
  94. u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {};
  95. struct mlx5_core_dev *dev = dm->dev;
  96. MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC);
  97. MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_DEALLOC);
  98. MLX5_SET(modify_memic_in, in, memic_operation_type, operation);
  99. MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr);
  100. mlx5_cmd_exec_in(dev, modify_memic, in);
  101. }
  102. static int mlx5_cmd_alloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr,
  103. u8 operation, phys_addr_t *op_addr)
  104. {
  105. u32 out[MLX5_ST_SZ_DW(modify_memic_out)] = {};
  106. u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {};
  107. struct mlx5_core_dev *dev = dm->dev;
  108. int err;
  109. MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC);
  110. MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_ALLOC);
  111. MLX5_SET(modify_memic_in, in, memic_operation_type, operation);
  112. MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr);
  113. err = mlx5_cmd_exec_inout(dev, modify_memic, in, out);
  114. if (err)
  115. return err;
  116. *op_addr = dev->bar_addr +
  117. MLX5_GET64(modify_memic_out, out, memic_operation_addr);
  118. return 0;
  119. }
  120. static int add_dm_mmap_entry(struct ib_ucontext *context,
  121. struct mlx5_user_mmap_entry *mentry, u8 mmap_flag,
  122. size_t size, u64 address)
  123. {
  124. mentry->mmap_flag = mmap_flag;
  125. mentry->address = address;
  126. return rdma_user_mmap_entry_insert_range(
  127. context, &mentry->rdma_entry, size,
  128. MLX5_IB_MMAP_DEVICE_MEM << 16,
  129. (MLX5_IB_MMAP_DEVICE_MEM << 16) + (1UL << 16) - 1);
  130. }
  131. static void mlx5_ib_dm_memic_free(struct kref *kref)
  132. {
  133. struct mlx5_ib_dm_memic *dm =
  134. container_of(kref, struct mlx5_ib_dm_memic, ref);
  135. struct mlx5_ib_dev *dev = to_mdev(dm->base.ibdm.device);
  136. mlx5_cmd_dealloc_memic(&dev->dm, dm->base.dev_addr, dm->base.size);
  137. kfree(dm);
  138. }
  139. static int copy_op_to_user(struct mlx5_ib_dm_op_entry *op_entry,
  140. struct uverbs_attr_bundle *attrs)
  141. {
  142. u64 start_offset;
  143. u16 page_idx;
  144. int err;
  145. page_idx = op_entry->mentry.rdma_entry.start_pgoff & 0xFFFF;
  146. start_offset = op_entry->op_addr & ~PAGE_MASK;
  147. err = uverbs_copy_to(attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX,
  148. &page_idx, sizeof(page_idx));
  149. if (err)
  150. return err;
  151. return uverbs_copy_to(attrs,
  152. MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET,
  153. &start_offset, sizeof(start_offset));
  154. }
  155. static int map_existing_op(struct mlx5_ib_dm_memic *dm, u8 op,
  156. struct uverbs_attr_bundle *attrs)
  157. {
  158. struct mlx5_ib_dm_op_entry *op_entry;
  159. op_entry = xa_load(&dm->ops, op);
  160. if (!op_entry)
  161. return -ENOENT;
  162. return copy_op_to_user(op_entry, attrs);
  163. }
  164. static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_MAP_OP_ADDR)(
  165. struct uverbs_attr_bundle *attrs)
  166. {
  167. struct ib_uobject *uobj = uverbs_attr_get_uobject(
  168. attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE);
  169. struct mlx5_ib_dev *dev = to_mdev(uobj->context->device);
  170. struct ib_dm *ibdm = uobj->object;
  171. struct mlx5_ib_dm_memic *dm = to_memic(ibdm);
  172. struct mlx5_ib_dm_op_entry *op_entry;
  173. int err;
  174. u8 op;
  175. err = uverbs_copy_from(&op, attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP);
  176. if (err)
  177. return err;
  178. if (op >= BITS_PER_TYPE(u32))
  179. return -EOPNOTSUPP;
  180. if (!(MLX5_CAP_DEV_MEM(dev->mdev, memic_operations) & BIT(op)))
  181. return -EOPNOTSUPP;
  182. mutex_lock(&dm->ops_xa_lock);
  183. err = map_existing_op(dm, op, attrs);
  184. if (!err || err != -ENOENT)
  185. goto err_unlock;
  186. op_entry = kzalloc(sizeof(*op_entry), GFP_KERNEL);
  187. if (!op_entry)
  188. goto err_unlock;
  189. err = mlx5_cmd_alloc_memic_op(&dev->dm, dm->base.dev_addr, op,
  190. &op_entry->op_addr);
  191. if (err) {
  192. kfree(op_entry);
  193. goto err_unlock;
  194. }
  195. op_entry->op = op;
  196. op_entry->dm = dm;
  197. err = add_dm_mmap_entry(uobj->context, &op_entry->mentry,
  198. MLX5_IB_MMAP_TYPE_MEMIC_OP, dm->base.size,
  199. op_entry->op_addr & PAGE_MASK);
  200. if (err) {
  201. mlx5_cmd_dealloc_memic_op(&dev->dm, dm->base.dev_addr, op);
  202. kfree(op_entry);
  203. goto err_unlock;
  204. }
  205. /* From this point, entry will be freed by mmap_free */
  206. kref_get(&dm->ref);
  207. err = copy_op_to_user(op_entry, attrs);
  208. if (err)
  209. goto err_remove;
  210. err = xa_insert(&dm->ops, op, op_entry, GFP_KERNEL);
  211. if (err)
  212. goto err_remove;
  213. mutex_unlock(&dm->ops_xa_lock);
  214. return 0;
  215. err_remove:
  216. rdma_user_mmap_entry_remove(&op_entry->mentry.rdma_entry);
  217. err_unlock:
  218. mutex_unlock(&dm->ops_xa_lock);
  219. return err;
  220. }
  221. static struct ib_dm *handle_alloc_dm_memic(struct ib_ucontext *ctx,
  222. struct ib_dm_alloc_attr *attr,
  223. struct uverbs_attr_bundle *attrs)
  224. {
  225. struct mlx5_dm *dm_db = &to_mdev(ctx->device)->dm;
  226. struct mlx5_ib_dm_memic *dm;
  227. u64 start_offset;
  228. u16 page_idx;
  229. int err;
  230. u64 address;
  231. if (!MLX5_CAP_DEV_MEM(dm_db->dev, memic))
  232. return ERR_PTR(-EOPNOTSUPP);
  233. dm = kzalloc(sizeof(*dm), GFP_KERNEL);
  234. if (!dm)
  235. return ERR_PTR(-ENOMEM);
  236. dm->base.type = MLX5_IB_UAPI_DM_TYPE_MEMIC;
  237. dm->base.size = roundup(attr->length, MLX5_MEMIC_BASE_SIZE);
  238. dm->base.ibdm.device = ctx->device;
  239. kref_init(&dm->ref);
  240. xa_init(&dm->ops);
  241. mutex_init(&dm->ops_xa_lock);
  242. dm->req_length = attr->length;
  243. err = mlx5_cmd_alloc_memic(dm_db, &dm->base.dev_addr,
  244. dm->base.size, attr->alignment);
  245. if (err) {
  246. kfree(dm);
  247. return ERR_PTR(err);
  248. }
  249. address = dm->base.dev_addr & PAGE_MASK;
  250. err = add_dm_mmap_entry(ctx, &dm->mentry, MLX5_IB_MMAP_TYPE_MEMIC,
  251. dm->base.size, address);
  252. if (err) {
  253. mlx5_cmd_dealloc_memic(dm_db, dm->base.dev_addr, dm->base.size);
  254. kfree(dm);
  255. return ERR_PTR(err);
  256. }
  257. page_idx = dm->mentry.rdma_entry.start_pgoff & 0xFFFF;
  258. err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX,
  259. &page_idx, sizeof(page_idx));
  260. if (err)
  261. goto err_copy;
  262. start_offset = dm->base.dev_addr & ~PAGE_MASK;
  263. err = uverbs_copy_to(attrs,
  264. MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
  265. &start_offset, sizeof(start_offset));
  266. if (err)
  267. goto err_copy;
  268. return &dm->base.ibdm;
  269. err_copy:
  270. rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry);
  271. return ERR_PTR(err);
  272. }
  273. static enum mlx5_sw_icm_type get_icm_type(int uapi_type)
  274. {
  275. switch (uapi_type) {
  276. case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
  277. return MLX5_SW_ICM_TYPE_HEADER_MODIFY;
  278. case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
  279. return MLX5_SW_ICM_TYPE_HEADER_MODIFY_PATTERN;
  280. case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
  281. default:
  282. return MLX5_SW_ICM_TYPE_STEERING;
  283. }
  284. }
  285. static struct ib_dm *handle_alloc_dm_sw_icm(struct ib_ucontext *ctx,
  286. struct ib_dm_alloc_attr *attr,
  287. struct uverbs_attr_bundle *attrs,
  288. int type)
  289. {
  290. struct mlx5_core_dev *dev = to_mdev(ctx->device)->mdev;
  291. enum mlx5_sw_icm_type icm_type;
  292. struct mlx5_ib_dm_icm *dm;
  293. u64 act_size;
  294. int err;
  295. if (!capable(CAP_SYS_RAWIO) || !capable(CAP_NET_RAW))
  296. return ERR_PTR(-EPERM);
  297. switch (type) {
  298. case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
  299. case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
  300. if (!(MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner) ||
  301. MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner) ||
  302. MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) ||
  303. MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2)))
  304. return ERR_PTR(-EOPNOTSUPP);
  305. break;
  306. case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
  307. if (!MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) ||
  308. !MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2))
  309. return ERR_PTR(-EOPNOTSUPP);
  310. break;
  311. default:
  312. return ERR_PTR(-EOPNOTSUPP);
  313. }
  314. dm = kzalloc(sizeof(*dm), GFP_KERNEL);
  315. if (!dm)
  316. return ERR_PTR(-ENOMEM);
  317. dm->base.type = type;
  318. dm->base.ibdm.device = ctx->device;
  319. /* Allocation size must a multiple of the basic block size
  320. * and a power of 2.
  321. */
  322. act_size = round_up(attr->length, MLX5_SW_ICM_BLOCK_SIZE(dev));
  323. act_size = roundup_pow_of_two(act_size);
  324. dm->base.size = act_size;
  325. icm_type = get_icm_type(type);
  326. err = mlx5_dm_sw_icm_alloc(dev, icm_type, act_size, attr->alignment,
  327. to_mucontext(ctx)->devx_uid,
  328. &dm->base.dev_addr, &dm->obj_id);
  329. if (err)
  330. goto free;
  331. err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
  332. &dm->base.dev_addr, sizeof(dm->base.dev_addr));
  333. if (err) {
  334. mlx5_dm_sw_icm_dealloc(dev, icm_type, dm->base.size,
  335. to_mucontext(ctx)->devx_uid,
  336. dm->base.dev_addr, dm->obj_id);
  337. goto free;
  338. }
  339. return &dm->base.ibdm;
  340. free:
  341. kfree(dm);
  342. return ERR_PTR(err);
  343. }
  344. struct ib_dm *mlx5_ib_alloc_dm(struct ib_device *ibdev,
  345. struct ib_ucontext *context,
  346. struct ib_dm_alloc_attr *attr,
  347. struct uverbs_attr_bundle *attrs)
  348. {
  349. enum mlx5_ib_uapi_dm_type type;
  350. int err;
  351. err = uverbs_get_const_default(&type, attrs,
  352. MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE,
  353. MLX5_IB_UAPI_DM_TYPE_MEMIC);
  354. if (err)
  355. return ERR_PTR(err);
  356. mlx5_ib_dbg(to_mdev(ibdev), "alloc_dm req: dm_type=%d user_length=0x%llx log_alignment=%d\n",
  357. type, attr->length, attr->alignment);
  358. switch (type) {
  359. case MLX5_IB_UAPI_DM_TYPE_MEMIC:
  360. return handle_alloc_dm_memic(context, attr, attrs);
  361. case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
  362. case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
  363. case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
  364. return handle_alloc_dm_sw_icm(context, attr, attrs, type);
  365. default:
  366. return ERR_PTR(-EOPNOTSUPP);
  367. }
  368. }
  369. static void dm_memic_remove_ops(struct mlx5_ib_dm_memic *dm)
  370. {
  371. struct mlx5_ib_dm_op_entry *entry;
  372. unsigned long idx;
  373. mutex_lock(&dm->ops_xa_lock);
  374. xa_for_each(&dm->ops, idx, entry) {
  375. xa_erase(&dm->ops, idx);
  376. rdma_user_mmap_entry_remove(&entry->mentry.rdma_entry);
  377. }
  378. mutex_unlock(&dm->ops_xa_lock);
  379. }
  380. static void mlx5_dm_memic_dealloc(struct mlx5_ib_dm_memic *dm)
  381. {
  382. dm_memic_remove_ops(dm);
  383. rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry);
  384. }
  385. static int mlx5_dm_icm_dealloc(struct mlx5_ib_ucontext *ctx,
  386. struct mlx5_ib_dm_icm *dm)
  387. {
  388. enum mlx5_sw_icm_type type = get_icm_type(dm->base.type);
  389. struct mlx5_core_dev *dev = to_mdev(dm->base.ibdm.device)->mdev;
  390. int err;
  391. err = mlx5_dm_sw_icm_dealloc(dev, type, dm->base.size, ctx->devx_uid,
  392. dm->base.dev_addr, dm->obj_id);
  393. if (!err)
  394. kfree(dm);
  395. return 0;
  396. }
  397. static int mlx5_ib_dealloc_dm(struct ib_dm *ibdm,
  398. struct uverbs_attr_bundle *attrs)
  399. {
  400. struct mlx5_ib_ucontext *ctx = rdma_udata_to_drv_context(
  401. &attrs->driver_udata, struct mlx5_ib_ucontext, ibucontext);
  402. struct mlx5_ib_dm *dm = to_mdm(ibdm);
  403. switch (dm->type) {
  404. case MLX5_IB_UAPI_DM_TYPE_MEMIC:
  405. mlx5_dm_memic_dealloc(to_memic(ibdm));
  406. return 0;
  407. case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
  408. case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
  409. case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
  410. return mlx5_dm_icm_dealloc(ctx, to_icm(ibdm));
  411. default:
  412. return -EOPNOTSUPP;
  413. }
  414. }
  415. static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_QUERY)(
  416. struct uverbs_attr_bundle *attrs)
  417. {
  418. struct ib_dm *ibdm =
  419. uverbs_attr_get_obj(attrs, MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE);
  420. struct mlx5_ib_dm *dm = to_mdm(ibdm);
  421. struct mlx5_ib_dm_memic *memic;
  422. u64 start_offset;
  423. u16 page_idx;
  424. int err;
  425. if (dm->type != MLX5_IB_UAPI_DM_TYPE_MEMIC)
  426. return -EOPNOTSUPP;
  427. memic = to_memic(ibdm);
  428. page_idx = memic->mentry.rdma_entry.start_pgoff & 0xFFFF;
  429. err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX,
  430. &page_idx, sizeof(page_idx));
  431. if (err)
  432. return err;
  433. start_offset = memic->base.dev_addr & ~PAGE_MASK;
  434. err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET,
  435. &start_offset, sizeof(start_offset));
  436. if (err)
  437. return err;
  438. return uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH,
  439. &memic->req_length,
  440. sizeof(memic->req_length));
  441. }
  442. void mlx5_ib_dm_mmap_free(struct mlx5_ib_dev *dev,
  443. struct mlx5_user_mmap_entry *mentry)
  444. {
  445. struct mlx5_ib_dm_op_entry *op_entry;
  446. struct mlx5_ib_dm_memic *mdm;
  447. switch (mentry->mmap_flag) {
  448. case MLX5_IB_MMAP_TYPE_MEMIC:
  449. mdm = container_of(mentry, struct mlx5_ib_dm_memic, mentry);
  450. kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
  451. break;
  452. case MLX5_IB_MMAP_TYPE_MEMIC_OP:
  453. op_entry = container_of(mentry, struct mlx5_ib_dm_op_entry,
  454. mentry);
  455. mdm = op_entry->dm;
  456. mlx5_cmd_dealloc_memic_op(&dev->dm, mdm->base.dev_addr,
  457. op_entry->op);
  458. kfree(op_entry);
  459. kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
  460. break;
  461. default:
  462. WARN_ON(true);
  463. }
  464. }
  465. DECLARE_UVERBS_NAMED_METHOD(
  466. MLX5_IB_METHOD_DM_QUERY,
  467. UVERBS_ATTR_IDR(MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE, UVERBS_OBJECT_DM,
  468. UVERBS_ACCESS_READ, UA_MANDATORY),
  469. UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET,
  470. UVERBS_ATTR_TYPE(u64), UA_MANDATORY),
  471. UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX,
  472. UVERBS_ATTR_TYPE(u16), UA_MANDATORY),
  473. UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH,
  474. UVERBS_ATTR_TYPE(u64), UA_MANDATORY));
  475. ADD_UVERBS_ATTRIBUTES_SIMPLE(
  476. mlx5_ib_dm, UVERBS_OBJECT_DM, UVERBS_METHOD_DM_ALLOC,
  477. UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
  478. UVERBS_ATTR_TYPE(u64), UA_MANDATORY),
  479. UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX,
  480. UVERBS_ATTR_TYPE(u16), UA_OPTIONAL),
  481. UVERBS_ATTR_CONST_IN(MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE,
  482. enum mlx5_ib_uapi_dm_type, UA_OPTIONAL));
  483. DECLARE_UVERBS_NAMED_METHOD(
  484. MLX5_IB_METHOD_DM_MAP_OP_ADDR,
  485. UVERBS_ATTR_IDR(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE,
  486. UVERBS_OBJECT_DM,
  487. UVERBS_ACCESS_READ,
  488. UA_MANDATORY),
  489. UVERBS_ATTR_PTR_IN(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP,
  490. UVERBS_ATTR_TYPE(u8),
  491. UA_MANDATORY),
  492. UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET,
  493. UVERBS_ATTR_TYPE(u64),
  494. UA_MANDATORY),
  495. UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX,
  496. UVERBS_ATTR_TYPE(u16),
  497. UA_OPTIONAL));
  498. DECLARE_UVERBS_GLOBAL_METHODS(UVERBS_OBJECT_DM,
  499. &UVERBS_METHOD(MLX5_IB_METHOD_DM_MAP_OP_ADDR),
  500. &UVERBS_METHOD(MLX5_IB_METHOD_DM_QUERY));
  501. const struct uapi_definition mlx5_ib_dm_defs[] = {
  502. UAPI_DEF_CHAIN_OBJ_TREE(UVERBS_OBJECT_DM, &mlx5_ib_dm),
  503. UAPI_DEF_CHAIN_OBJ_TREE_NAMED(UVERBS_OBJECT_DM),
  504. {},
  505. };
  506. const struct ib_device_ops mlx5_ib_dev_dm_ops = {
  507. .alloc_dm = mlx5_ib_alloc_dm,
  508. .dealloc_dm = mlx5_ib_dealloc_dm,
  509. .reg_dm_mr = mlx5_ib_reg_dm_mr,
  510. };