123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607 |
- // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
- /*
- * Copyright (c) 2021, Mellanox Technologies inc. All rights reserved.
- */
- #include <rdma/uverbs_std_types.h>
- #include "dm.h"
- #define UVERBS_MODULE_NAME mlx5_ib
- #include <rdma/uverbs_named_ioctl.h>
- static int mlx5_cmd_alloc_memic(struct mlx5_dm *dm, phys_addr_t *addr,
- u64 length, u32 alignment)
- {
- struct mlx5_core_dev *dev = dm->dev;
- u64 num_memic_hw_pages = MLX5_CAP_DEV_MEM(dev, memic_bar_size)
- >> PAGE_SHIFT;
- u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr);
- u32 max_alignment = MLX5_CAP_DEV_MEM(dev, log_max_memic_addr_alignment);
- u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE);
- u32 out[MLX5_ST_SZ_DW(alloc_memic_out)] = {};
- u32 in[MLX5_ST_SZ_DW(alloc_memic_in)] = {};
- u32 mlx5_alignment;
- u64 page_idx = 0;
- int ret = 0;
- if (!length || (length & MLX5_MEMIC_ALLOC_SIZE_MASK))
- return -EINVAL;
- /* mlx5 device sets alignment as 64*2^driver_value
- * so normalizing is needed.
- */
- mlx5_alignment = (alignment < MLX5_MEMIC_BASE_ALIGN) ? 0 :
- alignment - MLX5_MEMIC_BASE_ALIGN;
- if (mlx5_alignment > max_alignment)
- return -EINVAL;
- MLX5_SET(alloc_memic_in, in, opcode, MLX5_CMD_OP_ALLOC_MEMIC);
- MLX5_SET(alloc_memic_in, in, range_size, num_pages * PAGE_SIZE);
- MLX5_SET(alloc_memic_in, in, memic_size, length);
- MLX5_SET(alloc_memic_in, in, log_memic_addr_alignment,
- mlx5_alignment);
- while (page_idx < num_memic_hw_pages) {
- spin_lock(&dm->lock);
- page_idx = bitmap_find_next_zero_area(dm->memic_alloc_pages,
- num_memic_hw_pages,
- page_idx,
- num_pages, 0);
- if (page_idx < num_memic_hw_pages)
- bitmap_set(dm->memic_alloc_pages,
- page_idx, num_pages);
- spin_unlock(&dm->lock);
- if (page_idx >= num_memic_hw_pages)
- break;
- MLX5_SET64(alloc_memic_in, in, range_start_addr,
- hw_start_addr + (page_idx * PAGE_SIZE));
- ret = mlx5_cmd_exec_inout(dev, alloc_memic, in, out);
- if (ret) {
- spin_lock(&dm->lock);
- bitmap_clear(dm->memic_alloc_pages,
- page_idx, num_pages);
- spin_unlock(&dm->lock);
- if (ret == -EAGAIN) {
- page_idx++;
- continue;
- }
- return ret;
- }
- *addr = dev->bar_addr +
- MLX5_GET64(alloc_memic_out, out, memic_start_addr);
- return 0;
- }
- return -ENOMEM;
- }
- void mlx5_cmd_dealloc_memic(struct mlx5_dm *dm, phys_addr_t addr,
- u64 length)
- {
- struct mlx5_core_dev *dev = dm->dev;
- u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr);
- u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE);
- u32 in[MLX5_ST_SZ_DW(dealloc_memic_in)] = {};
- u64 start_page_idx;
- int err;
- addr -= dev->bar_addr;
- start_page_idx = (addr - hw_start_addr) >> PAGE_SHIFT;
- MLX5_SET(dealloc_memic_in, in, opcode, MLX5_CMD_OP_DEALLOC_MEMIC);
- MLX5_SET64(dealloc_memic_in, in, memic_start_addr, addr);
- MLX5_SET(dealloc_memic_in, in, memic_size, length);
- err = mlx5_cmd_exec_in(dev, dealloc_memic, in);
- if (err)
- return;
- spin_lock(&dm->lock);
- bitmap_clear(dm->memic_alloc_pages,
- start_page_idx, num_pages);
- spin_unlock(&dm->lock);
- }
- void mlx5_cmd_dealloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr,
- u8 operation)
- {
- u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {};
- struct mlx5_core_dev *dev = dm->dev;
- MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC);
- MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_DEALLOC);
- MLX5_SET(modify_memic_in, in, memic_operation_type, operation);
- MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr);
- mlx5_cmd_exec_in(dev, modify_memic, in);
- }
- static int mlx5_cmd_alloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr,
- u8 operation, phys_addr_t *op_addr)
- {
- u32 out[MLX5_ST_SZ_DW(modify_memic_out)] = {};
- u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {};
- struct mlx5_core_dev *dev = dm->dev;
- int err;
- MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC);
- MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_ALLOC);
- MLX5_SET(modify_memic_in, in, memic_operation_type, operation);
- MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr);
- err = mlx5_cmd_exec_inout(dev, modify_memic, in, out);
- if (err)
- return err;
- *op_addr = dev->bar_addr +
- MLX5_GET64(modify_memic_out, out, memic_operation_addr);
- return 0;
- }
- static int add_dm_mmap_entry(struct ib_ucontext *context,
- struct mlx5_user_mmap_entry *mentry, u8 mmap_flag,
- size_t size, u64 address)
- {
- mentry->mmap_flag = mmap_flag;
- mentry->address = address;
- return rdma_user_mmap_entry_insert_range(
- context, &mentry->rdma_entry, size,
- MLX5_IB_MMAP_DEVICE_MEM << 16,
- (MLX5_IB_MMAP_DEVICE_MEM << 16) + (1UL << 16) - 1);
- }
- static void mlx5_ib_dm_memic_free(struct kref *kref)
- {
- struct mlx5_ib_dm_memic *dm =
- container_of(kref, struct mlx5_ib_dm_memic, ref);
- struct mlx5_ib_dev *dev = to_mdev(dm->base.ibdm.device);
- mlx5_cmd_dealloc_memic(&dev->dm, dm->base.dev_addr, dm->base.size);
- kfree(dm);
- }
- static int copy_op_to_user(struct mlx5_ib_dm_op_entry *op_entry,
- struct uverbs_attr_bundle *attrs)
- {
- u64 start_offset;
- u16 page_idx;
- int err;
- page_idx = op_entry->mentry.rdma_entry.start_pgoff & 0xFFFF;
- start_offset = op_entry->op_addr & ~PAGE_MASK;
- err = uverbs_copy_to(attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX,
- &page_idx, sizeof(page_idx));
- if (err)
- return err;
- return uverbs_copy_to(attrs,
- MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET,
- &start_offset, sizeof(start_offset));
- }
- static int map_existing_op(struct mlx5_ib_dm_memic *dm, u8 op,
- struct uverbs_attr_bundle *attrs)
- {
- struct mlx5_ib_dm_op_entry *op_entry;
- op_entry = xa_load(&dm->ops, op);
- if (!op_entry)
- return -ENOENT;
- return copy_op_to_user(op_entry, attrs);
- }
- static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_MAP_OP_ADDR)(
- struct uverbs_attr_bundle *attrs)
- {
- struct ib_uobject *uobj = uverbs_attr_get_uobject(
- attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE);
- struct mlx5_ib_dev *dev = to_mdev(uobj->context->device);
- struct ib_dm *ibdm = uobj->object;
- struct mlx5_ib_dm_memic *dm = to_memic(ibdm);
- struct mlx5_ib_dm_op_entry *op_entry;
- int err;
- u8 op;
- err = uverbs_copy_from(&op, attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP);
- if (err)
- return err;
- if (op >= BITS_PER_TYPE(u32))
- return -EOPNOTSUPP;
- if (!(MLX5_CAP_DEV_MEM(dev->mdev, memic_operations) & BIT(op)))
- return -EOPNOTSUPP;
- mutex_lock(&dm->ops_xa_lock);
- err = map_existing_op(dm, op, attrs);
- if (!err || err != -ENOENT)
- goto err_unlock;
- op_entry = kzalloc(sizeof(*op_entry), GFP_KERNEL);
- if (!op_entry)
- goto err_unlock;
- err = mlx5_cmd_alloc_memic_op(&dev->dm, dm->base.dev_addr, op,
- &op_entry->op_addr);
- if (err) {
- kfree(op_entry);
- goto err_unlock;
- }
- op_entry->op = op;
- op_entry->dm = dm;
- err = add_dm_mmap_entry(uobj->context, &op_entry->mentry,
- MLX5_IB_MMAP_TYPE_MEMIC_OP, dm->base.size,
- op_entry->op_addr & PAGE_MASK);
- if (err) {
- mlx5_cmd_dealloc_memic_op(&dev->dm, dm->base.dev_addr, op);
- kfree(op_entry);
- goto err_unlock;
- }
- /* From this point, entry will be freed by mmap_free */
- kref_get(&dm->ref);
- err = copy_op_to_user(op_entry, attrs);
- if (err)
- goto err_remove;
- err = xa_insert(&dm->ops, op, op_entry, GFP_KERNEL);
- if (err)
- goto err_remove;
- mutex_unlock(&dm->ops_xa_lock);
- return 0;
- err_remove:
- rdma_user_mmap_entry_remove(&op_entry->mentry.rdma_entry);
- err_unlock:
- mutex_unlock(&dm->ops_xa_lock);
- return err;
- }
- static struct ib_dm *handle_alloc_dm_memic(struct ib_ucontext *ctx,
- struct ib_dm_alloc_attr *attr,
- struct uverbs_attr_bundle *attrs)
- {
- struct mlx5_dm *dm_db = &to_mdev(ctx->device)->dm;
- struct mlx5_ib_dm_memic *dm;
- u64 start_offset;
- u16 page_idx;
- int err;
- u64 address;
- if (!MLX5_CAP_DEV_MEM(dm_db->dev, memic))
- return ERR_PTR(-EOPNOTSUPP);
- dm = kzalloc(sizeof(*dm), GFP_KERNEL);
- if (!dm)
- return ERR_PTR(-ENOMEM);
- dm->base.type = MLX5_IB_UAPI_DM_TYPE_MEMIC;
- dm->base.size = roundup(attr->length, MLX5_MEMIC_BASE_SIZE);
- dm->base.ibdm.device = ctx->device;
- kref_init(&dm->ref);
- xa_init(&dm->ops);
- mutex_init(&dm->ops_xa_lock);
- dm->req_length = attr->length;
- err = mlx5_cmd_alloc_memic(dm_db, &dm->base.dev_addr,
- dm->base.size, attr->alignment);
- if (err) {
- kfree(dm);
- return ERR_PTR(err);
- }
- address = dm->base.dev_addr & PAGE_MASK;
- err = add_dm_mmap_entry(ctx, &dm->mentry, MLX5_IB_MMAP_TYPE_MEMIC,
- dm->base.size, address);
- if (err) {
- mlx5_cmd_dealloc_memic(dm_db, dm->base.dev_addr, dm->base.size);
- kfree(dm);
- return ERR_PTR(err);
- }
- page_idx = dm->mentry.rdma_entry.start_pgoff & 0xFFFF;
- err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX,
- &page_idx, sizeof(page_idx));
- if (err)
- goto err_copy;
- start_offset = dm->base.dev_addr & ~PAGE_MASK;
- err = uverbs_copy_to(attrs,
- MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
- &start_offset, sizeof(start_offset));
- if (err)
- goto err_copy;
- return &dm->base.ibdm;
- err_copy:
- rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry);
- return ERR_PTR(err);
- }
- static enum mlx5_sw_icm_type get_icm_type(int uapi_type)
- {
- switch (uapi_type) {
- case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
- return MLX5_SW_ICM_TYPE_HEADER_MODIFY;
- case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
- return MLX5_SW_ICM_TYPE_HEADER_MODIFY_PATTERN;
- case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
- default:
- return MLX5_SW_ICM_TYPE_STEERING;
- }
- }
- static struct ib_dm *handle_alloc_dm_sw_icm(struct ib_ucontext *ctx,
- struct ib_dm_alloc_attr *attr,
- struct uverbs_attr_bundle *attrs,
- int type)
- {
- struct mlx5_core_dev *dev = to_mdev(ctx->device)->mdev;
- enum mlx5_sw_icm_type icm_type;
- struct mlx5_ib_dm_icm *dm;
- u64 act_size;
- int err;
- if (!capable(CAP_SYS_RAWIO) || !capable(CAP_NET_RAW))
- return ERR_PTR(-EPERM);
- switch (type) {
- case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
- case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
- if (!(MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner) ||
- MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner) ||
- MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) ||
- MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2)))
- return ERR_PTR(-EOPNOTSUPP);
- break;
- case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
- if (!MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) ||
- !MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2))
- return ERR_PTR(-EOPNOTSUPP);
- break;
- default:
- return ERR_PTR(-EOPNOTSUPP);
- }
- dm = kzalloc(sizeof(*dm), GFP_KERNEL);
- if (!dm)
- return ERR_PTR(-ENOMEM);
- dm->base.type = type;
- dm->base.ibdm.device = ctx->device;
- /* Allocation size must a multiple of the basic block size
- * and a power of 2.
- */
- act_size = round_up(attr->length, MLX5_SW_ICM_BLOCK_SIZE(dev));
- act_size = roundup_pow_of_two(act_size);
- dm->base.size = act_size;
- icm_type = get_icm_type(type);
- err = mlx5_dm_sw_icm_alloc(dev, icm_type, act_size, attr->alignment,
- to_mucontext(ctx)->devx_uid,
- &dm->base.dev_addr, &dm->obj_id);
- if (err)
- goto free;
- err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
- &dm->base.dev_addr, sizeof(dm->base.dev_addr));
- if (err) {
- mlx5_dm_sw_icm_dealloc(dev, icm_type, dm->base.size,
- to_mucontext(ctx)->devx_uid,
- dm->base.dev_addr, dm->obj_id);
- goto free;
- }
- return &dm->base.ibdm;
- free:
- kfree(dm);
- return ERR_PTR(err);
- }
- struct ib_dm *mlx5_ib_alloc_dm(struct ib_device *ibdev,
- struct ib_ucontext *context,
- struct ib_dm_alloc_attr *attr,
- struct uverbs_attr_bundle *attrs)
- {
- enum mlx5_ib_uapi_dm_type type;
- int err;
- err = uverbs_get_const_default(&type, attrs,
- MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE,
- MLX5_IB_UAPI_DM_TYPE_MEMIC);
- if (err)
- return ERR_PTR(err);
- mlx5_ib_dbg(to_mdev(ibdev), "alloc_dm req: dm_type=%d user_length=0x%llx log_alignment=%d\n",
- type, attr->length, attr->alignment);
- switch (type) {
- case MLX5_IB_UAPI_DM_TYPE_MEMIC:
- return handle_alloc_dm_memic(context, attr, attrs);
- case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
- case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
- case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
- return handle_alloc_dm_sw_icm(context, attr, attrs, type);
- default:
- return ERR_PTR(-EOPNOTSUPP);
- }
- }
- static void dm_memic_remove_ops(struct mlx5_ib_dm_memic *dm)
- {
- struct mlx5_ib_dm_op_entry *entry;
- unsigned long idx;
- mutex_lock(&dm->ops_xa_lock);
- xa_for_each(&dm->ops, idx, entry) {
- xa_erase(&dm->ops, idx);
- rdma_user_mmap_entry_remove(&entry->mentry.rdma_entry);
- }
- mutex_unlock(&dm->ops_xa_lock);
- }
- static void mlx5_dm_memic_dealloc(struct mlx5_ib_dm_memic *dm)
- {
- dm_memic_remove_ops(dm);
- rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry);
- }
- static int mlx5_dm_icm_dealloc(struct mlx5_ib_ucontext *ctx,
- struct mlx5_ib_dm_icm *dm)
- {
- enum mlx5_sw_icm_type type = get_icm_type(dm->base.type);
- struct mlx5_core_dev *dev = to_mdev(dm->base.ibdm.device)->mdev;
- int err;
- err = mlx5_dm_sw_icm_dealloc(dev, type, dm->base.size, ctx->devx_uid,
- dm->base.dev_addr, dm->obj_id);
- if (!err)
- kfree(dm);
- return 0;
- }
- static int mlx5_ib_dealloc_dm(struct ib_dm *ibdm,
- struct uverbs_attr_bundle *attrs)
- {
- struct mlx5_ib_ucontext *ctx = rdma_udata_to_drv_context(
- &attrs->driver_udata, struct mlx5_ib_ucontext, ibucontext);
- struct mlx5_ib_dm *dm = to_mdm(ibdm);
- switch (dm->type) {
- case MLX5_IB_UAPI_DM_TYPE_MEMIC:
- mlx5_dm_memic_dealloc(to_memic(ibdm));
- return 0;
- case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM:
- case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM:
- case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM:
- return mlx5_dm_icm_dealloc(ctx, to_icm(ibdm));
- default:
- return -EOPNOTSUPP;
- }
- }
- static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_QUERY)(
- struct uverbs_attr_bundle *attrs)
- {
- struct ib_dm *ibdm =
- uverbs_attr_get_obj(attrs, MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE);
- struct mlx5_ib_dm *dm = to_mdm(ibdm);
- struct mlx5_ib_dm_memic *memic;
- u64 start_offset;
- u16 page_idx;
- int err;
- if (dm->type != MLX5_IB_UAPI_DM_TYPE_MEMIC)
- return -EOPNOTSUPP;
- memic = to_memic(ibdm);
- page_idx = memic->mentry.rdma_entry.start_pgoff & 0xFFFF;
- err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX,
- &page_idx, sizeof(page_idx));
- if (err)
- return err;
- start_offset = memic->base.dev_addr & ~PAGE_MASK;
- err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET,
- &start_offset, sizeof(start_offset));
- if (err)
- return err;
- return uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH,
- &memic->req_length,
- sizeof(memic->req_length));
- }
- void mlx5_ib_dm_mmap_free(struct mlx5_ib_dev *dev,
- struct mlx5_user_mmap_entry *mentry)
- {
- struct mlx5_ib_dm_op_entry *op_entry;
- struct mlx5_ib_dm_memic *mdm;
- switch (mentry->mmap_flag) {
- case MLX5_IB_MMAP_TYPE_MEMIC:
- mdm = container_of(mentry, struct mlx5_ib_dm_memic, mentry);
- kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
- break;
- case MLX5_IB_MMAP_TYPE_MEMIC_OP:
- op_entry = container_of(mentry, struct mlx5_ib_dm_op_entry,
- mentry);
- mdm = op_entry->dm;
- mlx5_cmd_dealloc_memic_op(&dev->dm, mdm->base.dev_addr,
- op_entry->op);
- kfree(op_entry);
- kref_put(&mdm->ref, mlx5_ib_dm_memic_free);
- break;
- default:
- WARN_ON(true);
- }
- }
- DECLARE_UVERBS_NAMED_METHOD(
- MLX5_IB_METHOD_DM_QUERY,
- UVERBS_ATTR_IDR(MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE, UVERBS_OBJECT_DM,
- UVERBS_ACCESS_READ, UA_MANDATORY),
- UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET,
- UVERBS_ATTR_TYPE(u64), UA_MANDATORY),
- UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX,
- UVERBS_ATTR_TYPE(u16), UA_MANDATORY),
- UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH,
- UVERBS_ATTR_TYPE(u64), UA_MANDATORY));
- ADD_UVERBS_ATTRIBUTES_SIMPLE(
- mlx5_ib_dm, UVERBS_OBJECT_DM, UVERBS_METHOD_DM_ALLOC,
- UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET,
- UVERBS_ATTR_TYPE(u64), UA_MANDATORY),
- UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX,
- UVERBS_ATTR_TYPE(u16), UA_OPTIONAL),
- UVERBS_ATTR_CONST_IN(MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE,
- enum mlx5_ib_uapi_dm_type, UA_OPTIONAL));
- DECLARE_UVERBS_NAMED_METHOD(
- MLX5_IB_METHOD_DM_MAP_OP_ADDR,
- UVERBS_ATTR_IDR(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE,
- UVERBS_OBJECT_DM,
- UVERBS_ACCESS_READ,
- UA_MANDATORY),
- UVERBS_ATTR_PTR_IN(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP,
- UVERBS_ATTR_TYPE(u8),
- UA_MANDATORY),
- UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET,
- UVERBS_ATTR_TYPE(u64),
- UA_MANDATORY),
- UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX,
- UVERBS_ATTR_TYPE(u16),
- UA_OPTIONAL));
- DECLARE_UVERBS_GLOBAL_METHODS(UVERBS_OBJECT_DM,
- &UVERBS_METHOD(MLX5_IB_METHOD_DM_MAP_OP_ADDR),
- &UVERBS_METHOD(MLX5_IB_METHOD_DM_QUERY));
- const struct uapi_definition mlx5_ib_dm_defs[] = {
- UAPI_DEF_CHAIN_OBJ_TREE(UVERBS_OBJECT_DM, &mlx5_ib_dm),
- UAPI_DEF_CHAIN_OBJ_TREE_NAMED(UVERBS_OBJECT_DM),
- {},
- };
- const struct ib_device_ops mlx5_ib_dev_dm_ops = {
- .alloc_dm = mlx5_ib_alloc_dm,
- .dealloc_dm = mlx5_ib_dealloc_dm,
- .reg_dm_mr = mlx5_ib_reg_dm_mr,
- };
|