Merge branch 'wr-cleanup' into k.o/for-4.4
This commit is contained in:
@@ -360,7 +360,7 @@ struct ipoib_dev_priv {
|
||||
unsigned tx_head;
|
||||
unsigned tx_tail;
|
||||
struct ib_sge tx_sge[MAX_SKB_FRAGS + 1];
|
||||
struct ib_send_wr tx_wr;
|
||||
struct ib_ud_wr tx_wr;
|
||||
unsigned tx_outstanding;
|
||||
struct ib_wc send_wc[MAX_SEND_CQE];
|
||||
|
||||
@@ -528,7 +528,7 @@ static inline void ipoib_build_sge(struct ipoib_dev_priv *priv,
|
||||
priv->tx_sge[i + off].addr = mapping[i + off];
|
||||
priv->tx_sge[i + off].length = skb_frag_size(&frags[i]);
|
||||
}
|
||||
priv->tx_wr.num_sge = nr_frags + off;
|
||||
priv->tx_wr.wr.num_sge = nr_frags + off;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_INFINIBAND_IPOIB_DEBUG
|
||||
|
@@ -700,9 +700,9 @@ static inline int post_send(struct ipoib_dev_priv *priv,
|
||||
|
||||
ipoib_build_sge(priv, tx_req);
|
||||
|
||||
priv->tx_wr.wr_id = wr_id | IPOIB_OP_CM;
|
||||
priv->tx_wr.wr.wr_id = wr_id | IPOIB_OP_CM;
|
||||
|
||||
return ib_post_send(tx->qp, &priv->tx_wr, &bad_wr);
|
||||
return ib_post_send(tx->qp, &priv->tx_wr.wr, &bad_wr);
|
||||
}
|
||||
|
||||
void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx)
|
||||
|
@@ -518,19 +518,19 @@ static inline int post_send(struct ipoib_dev_priv *priv,
|
||||
|
||||
ipoib_build_sge(priv, tx_req);
|
||||
|
||||
priv->tx_wr.wr_id = wr_id;
|
||||
priv->tx_wr.wr.ud.remote_qpn = qpn;
|
||||
priv->tx_wr.wr.ud.ah = address;
|
||||
priv->tx_wr.wr.wr_id = wr_id;
|
||||
priv->tx_wr.remote_qpn = qpn;
|
||||
priv->tx_wr.ah = address;
|
||||
|
||||
if (head) {
|
||||
priv->tx_wr.wr.ud.mss = skb_shinfo(skb)->gso_size;
|
||||
priv->tx_wr.wr.ud.header = head;
|
||||
priv->tx_wr.wr.ud.hlen = hlen;
|
||||
priv->tx_wr.opcode = IB_WR_LSO;
|
||||
priv->tx_wr.mss = skb_shinfo(skb)->gso_size;
|
||||
priv->tx_wr.header = head;
|
||||
priv->tx_wr.hlen = hlen;
|
||||
priv->tx_wr.wr.opcode = IB_WR_LSO;
|
||||
} else
|
||||
priv->tx_wr.opcode = IB_WR_SEND;
|
||||
priv->tx_wr.wr.opcode = IB_WR_SEND;
|
||||
|
||||
return ib_post_send(priv->qp, &priv->tx_wr, &bad_wr);
|
||||
return ib_post_send(priv->qp, &priv->tx_wr.wr, &bad_wr);
|
||||
}
|
||||
|
||||
void ipoib_send(struct net_device *dev, struct sk_buff *skb,
|
||||
@@ -583,9 +583,9 @@ void ipoib_send(struct net_device *dev, struct sk_buff *skb,
|
||||
}
|
||||
|
||||
if (skb->ip_summed == CHECKSUM_PARTIAL)
|
||||
priv->tx_wr.send_flags |= IB_SEND_IP_CSUM;
|
||||
priv->tx_wr.wr.send_flags |= IB_SEND_IP_CSUM;
|
||||
else
|
||||
priv->tx_wr.send_flags &= ~IB_SEND_IP_CSUM;
|
||||
priv->tx_wr.wr.send_flags &= ~IB_SEND_IP_CSUM;
|
||||
|
||||
if (++priv->tx_outstanding == ipoib_sendq_size) {
|
||||
ipoib_dbg(priv, "TX ring full, stopping kernel net queue\n");
|
||||
|
@@ -461,7 +461,7 @@ int ipoib_set_mode(struct net_device *dev, const char *buf)
|
||||
netdev_update_features(dev);
|
||||
dev_set_mtu(dev, ipoib_cm_max_mtu(dev));
|
||||
rtnl_unlock();
|
||||
priv->tx_wr.send_flags &= ~IB_SEND_IP_CSUM;
|
||||
priv->tx_wr.wr.send_flags &= ~IB_SEND_IP_CSUM;
|
||||
|
||||
ipoib_flush_paths(dev);
|
||||
rtnl_lock();
|
||||
|
@@ -245,7 +245,7 @@ static int ipoib_mcast_join_finish(struct ipoib_mcast *mcast,
|
||||
|
||||
priv->qkey = be32_to_cpu(priv->broadcast->mcmember.qkey);
|
||||
spin_unlock_irq(&priv->lock);
|
||||
priv->tx_wr.wr.ud.remote_qkey = priv->qkey;
|
||||
priv->tx_wr.remote_qkey = priv->qkey;
|
||||
set_qkey = 1;
|
||||
}
|
||||
|
||||
|
@@ -221,9 +221,9 @@ int ipoib_transport_dev_init(struct net_device *dev, struct ib_device *ca)
|
||||
for (i = 0; i < MAX_SKB_FRAGS + 1; ++i)
|
||||
priv->tx_sge[i].lkey = priv->pd->local_dma_lkey;
|
||||
|
||||
priv->tx_wr.opcode = IB_WR_SEND;
|
||||
priv->tx_wr.sg_list = priv->tx_sge;
|
||||
priv->tx_wr.send_flags = IB_SEND_SIGNALED;
|
||||
priv->tx_wr.wr.opcode = IB_WR_SEND;
|
||||
priv->tx_wr.wr.sg_list = priv->tx_sge;
|
||||
priv->tx_wr.wr.send_flags = IB_SEND_SIGNALED;
|
||||
|
||||
priv->rx_sge[0].lkey = priv->pd->local_dma_lkey;
|
||||
|
||||
|
@@ -295,7 +295,11 @@ struct iser_tx_desc {
|
||||
int num_sge;
|
||||
bool mapped;
|
||||
u8 wr_idx;
|
||||
struct ib_send_wr wrs[ISER_MAX_WRS];
|
||||
union iser_wr {
|
||||
struct ib_send_wr send;
|
||||
struct ib_fast_reg_wr fast_reg;
|
||||
struct ib_sig_handover_wr sig;
|
||||
} wrs[ISER_MAX_WRS];
|
||||
struct iser_mem_reg data_reg;
|
||||
struct iser_mem_reg prot_reg;
|
||||
struct ib_sig_attrs sig_attrs;
|
||||
@@ -707,11 +711,11 @@ iser_reg_desc_put_fmr(struct ib_conn *ib_conn,
|
||||
static inline struct ib_send_wr *
|
||||
iser_tx_next_wr(struct iser_tx_desc *tx_desc)
|
||||
{
|
||||
struct ib_send_wr *cur_wr = &tx_desc->wrs[tx_desc->wr_idx];
|
||||
struct ib_send_wr *cur_wr = &tx_desc->wrs[tx_desc->wr_idx].send;
|
||||
struct ib_send_wr *last_wr;
|
||||
|
||||
if (tx_desc->wr_idx) {
|
||||
last_wr = &tx_desc->wrs[tx_desc->wr_idx - 1];
|
||||
last_wr = &tx_desc->wrs[tx_desc->wr_idx - 1].send;
|
||||
last_wr->next = cur_wr;
|
||||
}
|
||||
tx_desc->wr_idx++;
|
||||
|
@@ -437,7 +437,7 @@ iser_reg_sig_mr(struct iscsi_iser_task *iser_task,
|
||||
{
|
||||
struct iser_tx_desc *tx_desc = &iser_task->desc;
|
||||
struct ib_sig_attrs *sig_attrs = &tx_desc->sig_attrs;
|
||||
struct ib_send_wr *wr;
|
||||
struct ib_sig_handover_wr *wr;
|
||||
int ret;
|
||||
|
||||
memset(sig_attrs, 0, sizeof(*sig_attrs));
|
||||
@@ -447,26 +447,24 @@ iser_reg_sig_mr(struct iscsi_iser_task *iser_task,
|
||||
|
||||
iser_set_prot_checks(iser_task->sc, &sig_attrs->check_mask);
|
||||
|
||||
if (!pi_ctx->sig_mr_valid) {
|
||||
wr = iser_tx_next_wr(tx_desc);
|
||||
iser_inv_rkey(wr, pi_ctx->sig_mr);
|
||||
}
|
||||
if (!pi_ctx->sig_mr_valid)
|
||||
iser_inv_rkey(iser_tx_next_wr(tx_desc), pi_ctx->sig_mr);
|
||||
|
||||
wr = iser_tx_next_wr(tx_desc);
|
||||
wr->opcode = IB_WR_REG_SIG_MR;
|
||||
wr->wr_id = ISER_FASTREG_LI_WRID;
|
||||
wr->sg_list = &data_reg->sge;
|
||||
wr->num_sge = 1;
|
||||
wr->send_flags = 0;
|
||||
wr->wr.sig_handover.sig_attrs = sig_attrs;
|
||||
wr->wr.sig_handover.sig_mr = pi_ctx->sig_mr;
|
||||
wr = sig_handover_wr(iser_tx_next_wr(tx_desc));
|
||||
wr->wr.opcode = IB_WR_REG_SIG_MR;
|
||||
wr->wr.wr_id = ISER_FASTREG_LI_WRID;
|
||||
wr->wr.sg_list = &data_reg->sge;
|
||||
wr->wr.num_sge = 1;
|
||||
wr->wr.send_flags = 0;
|
||||
wr->sig_attrs = sig_attrs;
|
||||
wr->sig_mr = pi_ctx->sig_mr;
|
||||
if (scsi_prot_sg_count(iser_task->sc))
|
||||
wr->wr.sig_handover.prot = &prot_reg->sge;
|
||||
wr->prot = &prot_reg->sge;
|
||||
else
|
||||
wr->wr.sig_handover.prot = NULL;
|
||||
wr->wr.sig_handover.access_flags = IB_ACCESS_LOCAL_WRITE |
|
||||
IB_ACCESS_REMOTE_READ |
|
||||
IB_ACCESS_REMOTE_WRITE;
|
||||
wr->prot = NULL;
|
||||
wr->access_flags = IB_ACCESS_LOCAL_WRITE |
|
||||
IB_ACCESS_REMOTE_READ |
|
||||
IB_ACCESS_REMOTE_WRITE;
|
||||
pi_ctx->sig_mr_valid = 0;
|
||||
|
||||
sig_reg->sge.lkey = pi_ctx->sig_mr->lkey;
|
||||
@@ -491,7 +489,7 @@ static int iser_fast_reg_mr(struct iscsi_iser_task *iser_task,
|
||||
struct ib_mr *mr = rsc->mr;
|
||||
struct ib_fast_reg_page_list *frpl = rsc->frpl;
|
||||
struct iser_tx_desc *tx_desc = &iser_task->desc;
|
||||
struct ib_send_wr *wr;
|
||||
struct ib_fast_reg_wr *wr;
|
||||
int offset, size, plen;
|
||||
|
||||
plen = iser_sg_to_page_vec(mem, device->ib_device, frpl->page_list,
|
||||
@@ -501,24 +499,22 @@ static int iser_fast_reg_mr(struct iscsi_iser_task *iser_task,
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (!rsc->mr_valid) {
|
||||
wr = iser_tx_next_wr(tx_desc);
|
||||
iser_inv_rkey(wr, mr);
|
||||
}
|
||||
if (!rsc->mr_valid)
|
||||
iser_inv_rkey(iser_tx_next_wr(tx_desc), mr);
|
||||
|
||||
wr = iser_tx_next_wr(tx_desc);
|
||||
wr->opcode = IB_WR_FAST_REG_MR;
|
||||
wr->wr_id = ISER_FASTREG_LI_WRID;
|
||||
wr->send_flags = 0;
|
||||
wr->wr.fast_reg.iova_start = frpl->page_list[0] + offset;
|
||||
wr->wr.fast_reg.page_list = frpl;
|
||||
wr->wr.fast_reg.page_list_len = plen;
|
||||
wr->wr.fast_reg.page_shift = SHIFT_4K;
|
||||
wr->wr.fast_reg.length = size;
|
||||
wr->wr.fast_reg.rkey = mr->rkey;
|
||||
wr->wr.fast_reg.access_flags = (IB_ACCESS_LOCAL_WRITE |
|
||||
IB_ACCESS_REMOTE_WRITE |
|
||||
IB_ACCESS_REMOTE_READ);
|
||||
wr = fast_reg_wr(iser_tx_next_wr(tx_desc));
|
||||
wr->wr.opcode = IB_WR_FAST_REG_MR;
|
||||
wr->wr.wr_id = ISER_FASTREG_LI_WRID;
|
||||
wr->wr.send_flags = 0;
|
||||
wr->iova_start = frpl->page_list[0] + offset;
|
||||
wr->page_list = frpl;
|
||||
wr->page_list_len = plen;
|
||||
wr->page_shift = SHIFT_4K;
|
||||
wr->length = size;
|
||||
wr->rkey = mr->rkey;
|
||||
wr->access_flags = (IB_ACCESS_LOCAL_WRITE |
|
||||
IB_ACCESS_REMOTE_WRITE |
|
||||
IB_ACCESS_REMOTE_READ);
|
||||
rsc->mr_valid = 0;
|
||||
|
||||
reg->sge.lkey = mr->lkey;
|
||||
|
@@ -1135,7 +1135,7 @@ int iser_post_send(struct ib_conn *ib_conn, struct iser_tx_desc *tx_desc,
|
||||
wr->opcode = IB_WR_SEND;
|
||||
wr->send_flags = signal ? IB_SEND_SIGNALED : 0;
|
||||
|
||||
ib_ret = ib_post_send(ib_conn->qp, &tx_desc->wrs[0], &bad_wr);
|
||||
ib_ret = ib_post_send(ib_conn->qp, &tx_desc->wrs[0].send, &bad_wr);
|
||||
if (ib_ret)
|
||||
iser_err("ib_post_send failed, ret:%d opcode:%d\n",
|
||||
ib_ret, bad_wr->opcode);
|
||||
|
@@ -1715,10 +1715,10 @@ isert_unmap_cmd(struct isert_cmd *isert_cmd, struct isert_conn *isert_conn)
|
||||
isert_unmap_data_buf(isert_conn, &wr->data);
|
||||
}
|
||||
|
||||
if (wr->send_wr) {
|
||||
if (wr->rdma_wr) {
|
||||
isert_dbg("Cmd %p free send_wr\n", isert_cmd);
|
||||
kfree(wr->send_wr);
|
||||
wr->send_wr = NULL;
|
||||
kfree(wr->rdma_wr);
|
||||
wr->rdma_wr = NULL;
|
||||
}
|
||||
|
||||
if (wr->ib_sge) {
|
||||
@@ -1753,7 +1753,7 @@ isert_unreg_rdma(struct isert_cmd *isert_cmd, struct isert_conn *isert_conn)
|
||||
}
|
||||
|
||||
wr->ib_sge = NULL;
|
||||
wr->send_wr = NULL;
|
||||
wr->rdma_wr = NULL;
|
||||
}
|
||||
|
||||
static void
|
||||
@@ -1922,7 +1922,7 @@ isert_completion_rdma_write(struct iser_tx_desc *tx_desc,
|
||||
}
|
||||
|
||||
device->unreg_rdma_mem(isert_cmd, isert_conn);
|
||||
wr->send_wr_num = 0;
|
||||
wr->rdma_wr_num = 0;
|
||||
if (ret)
|
||||
transport_send_check_condition_and_sense(se_cmd,
|
||||
se_cmd->pi_err, 0);
|
||||
@@ -1950,7 +1950,7 @@ isert_completion_rdma_read(struct iser_tx_desc *tx_desc,
|
||||
iscsit_stop_dataout_timer(cmd);
|
||||
device->unreg_rdma_mem(isert_cmd, isert_conn);
|
||||
cmd->write_data_done = wr->data.len;
|
||||
wr->send_wr_num = 0;
|
||||
wr->rdma_wr_num = 0;
|
||||
|
||||
isert_dbg("Cmd: %p RDMA_READ comp calling execute_cmd\n", isert_cmd);
|
||||
spin_lock_bh(&cmd->istate_lock);
|
||||
@@ -2402,7 +2402,7 @@ isert_put_text_rsp(struct iscsi_cmd *cmd, struct iscsi_conn *conn)
|
||||
|
||||
static int
|
||||
isert_build_rdma_wr(struct isert_conn *isert_conn, struct isert_cmd *isert_cmd,
|
||||
struct ib_sge *ib_sge, struct ib_send_wr *send_wr,
|
||||
struct ib_sge *ib_sge, struct ib_rdma_wr *rdma_wr,
|
||||
u32 data_left, u32 offset)
|
||||
{
|
||||
struct iscsi_cmd *cmd = isert_cmd->iscsi_cmd;
|
||||
@@ -2417,8 +2417,8 @@ isert_build_rdma_wr(struct isert_conn *isert_conn, struct isert_cmd *isert_cmd,
|
||||
sg_nents = min(cmd->se_cmd.t_data_nents - sg_off, isert_conn->max_sge);
|
||||
page_off = offset % PAGE_SIZE;
|
||||
|
||||
send_wr->sg_list = ib_sge;
|
||||
send_wr->wr_id = (uintptr_t)&isert_cmd->tx_desc;
|
||||
rdma_wr->wr.sg_list = ib_sge;
|
||||
rdma_wr->wr.wr_id = (uintptr_t)&isert_cmd->tx_desc;
|
||||
/*
|
||||
* Perform mapping of TCM scatterlist memory ib_sge dma_addr.
|
||||
*/
|
||||
@@ -2443,11 +2443,11 @@ isert_build_rdma_wr(struct isert_conn *isert_conn, struct isert_cmd *isert_cmd,
|
||||
isert_dbg("Incrementing ib_sge pointer to %p\n", ib_sge);
|
||||
}
|
||||
|
||||
send_wr->num_sge = ++i;
|
||||
rdma_wr->wr.num_sge = ++i;
|
||||
isert_dbg("Set outgoing sg_list: %p num_sg: %u from TCM SGLs\n",
|
||||
send_wr->sg_list, send_wr->num_sge);
|
||||
rdma_wr->wr.sg_list, rdma_wr->wr.num_sge);
|
||||
|
||||
return send_wr->num_sge;
|
||||
return rdma_wr->wr.num_sge;
|
||||
}
|
||||
|
||||
static int
|
||||
@@ -2458,7 +2458,7 @@ isert_map_rdma(struct iscsi_conn *conn, struct iscsi_cmd *cmd,
|
||||
struct isert_cmd *isert_cmd = iscsit_priv_cmd(cmd);
|
||||
struct isert_conn *isert_conn = conn->context;
|
||||
struct isert_data_buf *data = &wr->data;
|
||||
struct ib_send_wr *send_wr;
|
||||
struct ib_rdma_wr *rdma_wr;
|
||||
struct ib_sge *ib_sge;
|
||||
u32 offset, data_len, data_left, rdma_write_max, va_offset = 0;
|
||||
int ret = 0, i, ib_sge_cnt;
|
||||
@@ -2483,11 +2483,11 @@ isert_map_rdma(struct iscsi_conn *conn, struct iscsi_cmd *cmd,
|
||||
}
|
||||
wr->ib_sge = ib_sge;
|
||||
|
||||
wr->send_wr_num = DIV_ROUND_UP(data->nents, isert_conn->max_sge);
|
||||
wr->send_wr = kzalloc(sizeof(struct ib_send_wr) * wr->send_wr_num,
|
||||
wr->rdma_wr_num = DIV_ROUND_UP(data->nents, isert_conn->max_sge);
|
||||
wr->rdma_wr = kzalloc(sizeof(struct ib_rdma_wr) * wr->rdma_wr_num,
|
||||
GFP_KERNEL);
|
||||
if (!wr->send_wr) {
|
||||
isert_dbg("Unable to allocate wr->send_wr\n");
|
||||
if (!wr->rdma_wr) {
|
||||
isert_dbg("Unable to allocate wr->rdma_wr\n");
|
||||
ret = -ENOMEM;
|
||||
goto unmap_cmd;
|
||||
}
|
||||
@@ -2495,31 +2495,31 @@ isert_map_rdma(struct iscsi_conn *conn, struct iscsi_cmd *cmd,
|
||||
wr->isert_cmd = isert_cmd;
|
||||
rdma_write_max = isert_conn->max_sge * PAGE_SIZE;
|
||||
|
||||
for (i = 0; i < wr->send_wr_num; i++) {
|
||||
send_wr = &isert_cmd->rdma_wr.send_wr[i];
|
||||
for (i = 0; i < wr->rdma_wr_num; i++) {
|
||||
rdma_wr = &isert_cmd->rdma_wr.rdma_wr[i];
|
||||
data_len = min(data_left, rdma_write_max);
|
||||
|
||||
send_wr->send_flags = 0;
|
||||
rdma_wr->wr.send_flags = 0;
|
||||
if (wr->iser_ib_op == ISER_IB_RDMA_WRITE) {
|
||||
send_wr->opcode = IB_WR_RDMA_WRITE;
|
||||
send_wr->wr.rdma.remote_addr = isert_cmd->read_va + offset;
|
||||
send_wr->wr.rdma.rkey = isert_cmd->read_stag;
|
||||
if (i + 1 == wr->send_wr_num)
|
||||
send_wr->next = &isert_cmd->tx_desc.send_wr;
|
||||
rdma_wr->wr.opcode = IB_WR_RDMA_WRITE;
|
||||
rdma_wr->remote_addr = isert_cmd->read_va + offset;
|
||||
rdma_wr->rkey = isert_cmd->read_stag;
|
||||
if (i + 1 == wr->rdma_wr_num)
|
||||
rdma_wr->wr.next = &isert_cmd->tx_desc.send_wr;
|
||||
else
|
||||
send_wr->next = &wr->send_wr[i + 1];
|
||||
rdma_wr->wr.next = &wr->rdma_wr[i + 1].wr;
|
||||
} else {
|
||||
send_wr->opcode = IB_WR_RDMA_READ;
|
||||
send_wr->wr.rdma.remote_addr = isert_cmd->write_va + va_offset;
|
||||
send_wr->wr.rdma.rkey = isert_cmd->write_stag;
|
||||
if (i + 1 == wr->send_wr_num)
|
||||
send_wr->send_flags = IB_SEND_SIGNALED;
|
||||
rdma_wr->wr.opcode = IB_WR_RDMA_READ;
|
||||
rdma_wr->remote_addr = isert_cmd->write_va + va_offset;
|
||||
rdma_wr->rkey = isert_cmd->write_stag;
|
||||
if (i + 1 == wr->rdma_wr_num)
|
||||
rdma_wr->wr.send_flags = IB_SEND_SIGNALED;
|
||||
else
|
||||
send_wr->next = &wr->send_wr[i + 1];
|
||||
rdma_wr->wr.next = &wr->rdma_wr[i + 1].wr;
|
||||
}
|
||||
|
||||
ib_sge_cnt = isert_build_rdma_wr(isert_conn, isert_cmd, ib_sge,
|
||||
send_wr, data_len, offset);
|
||||
rdma_wr, data_len, offset);
|
||||
ib_sge += ib_sge_cnt;
|
||||
|
||||
offset += data_len;
|
||||
@@ -2599,8 +2599,8 @@ isert_fast_reg_mr(struct isert_conn *isert_conn,
|
||||
struct ib_device *ib_dev = device->ib_device;
|
||||
struct ib_mr *mr;
|
||||
struct ib_fast_reg_page_list *frpl;
|
||||
struct ib_send_wr fr_wr, inv_wr;
|
||||
struct ib_send_wr *bad_wr, *wr = NULL;
|
||||
struct ib_fast_reg_wr fr_wr;
|
||||
struct ib_send_wr inv_wr, *bad_wr, *wr = NULL;
|
||||
int ret, pagelist_len;
|
||||
u32 page_off;
|
||||
|
||||
@@ -2638,20 +2638,20 @@ isert_fast_reg_mr(struct isert_conn *isert_conn,
|
||||
|
||||
/* Prepare FASTREG WR */
|
||||
memset(&fr_wr, 0, sizeof(fr_wr));
|
||||
fr_wr.wr_id = ISER_FASTREG_LI_WRID;
|
||||
fr_wr.opcode = IB_WR_FAST_REG_MR;
|
||||
fr_wr.wr.fast_reg.iova_start = frpl->page_list[0] + page_off;
|
||||
fr_wr.wr.fast_reg.page_list = frpl;
|
||||
fr_wr.wr.fast_reg.page_list_len = pagelist_len;
|
||||
fr_wr.wr.fast_reg.page_shift = PAGE_SHIFT;
|
||||
fr_wr.wr.fast_reg.length = mem->len;
|
||||
fr_wr.wr.fast_reg.rkey = mr->rkey;
|
||||
fr_wr.wr.fast_reg.access_flags = IB_ACCESS_LOCAL_WRITE;
|
||||
fr_wr.wr.wr_id = ISER_FASTREG_LI_WRID;
|
||||
fr_wr.wr.opcode = IB_WR_FAST_REG_MR;
|
||||
fr_wr.iova_start = frpl->page_list[0] + page_off;
|
||||
fr_wr.page_list = frpl;
|
||||
fr_wr.page_list_len = pagelist_len;
|
||||
fr_wr.page_shift = PAGE_SHIFT;
|
||||
fr_wr.length = mem->len;
|
||||
fr_wr.rkey = mr->rkey;
|
||||
fr_wr.access_flags = IB_ACCESS_LOCAL_WRITE;
|
||||
|
||||
if (!wr)
|
||||
wr = &fr_wr;
|
||||
wr = &fr_wr.wr;
|
||||
else
|
||||
wr->next = &fr_wr;
|
||||
wr->next = &fr_wr.wr;
|
||||
|
||||
ret = ib_post_send(isert_conn->qp, wr, &bad_wr);
|
||||
if (ret) {
|
||||
@@ -2732,8 +2732,8 @@ isert_reg_sig_mr(struct isert_conn *isert_conn,
|
||||
struct isert_rdma_wr *rdma_wr,
|
||||
struct fast_reg_descriptor *fr_desc)
|
||||
{
|
||||
struct ib_send_wr sig_wr, inv_wr;
|
||||
struct ib_send_wr *bad_wr, *wr = NULL;
|
||||
struct ib_sig_handover_wr sig_wr;
|
||||
struct ib_send_wr inv_wr, *bad_wr, *wr = NULL;
|
||||
struct pi_context *pi_ctx = fr_desc->pi_ctx;
|
||||
struct ib_sig_attrs sig_attrs;
|
||||
int ret;
|
||||
@@ -2751,20 +2751,20 @@ isert_reg_sig_mr(struct isert_conn *isert_conn,
|
||||
}
|
||||
|
||||
memset(&sig_wr, 0, sizeof(sig_wr));
|
||||
sig_wr.opcode = IB_WR_REG_SIG_MR;
|
||||
sig_wr.wr_id = ISER_FASTREG_LI_WRID;
|
||||
sig_wr.sg_list = &rdma_wr->ib_sg[DATA];
|
||||
sig_wr.num_sge = 1;
|
||||
sig_wr.wr.sig_handover.access_flags = IB_ACCESS_LOCAL_WRITE;
|
||||
sig_wr.wr.sig_handover.sig_attrs = &sig_attrs;
|
||||
sig_wr.wr.sig_handover.sig_mr = pi_ctx->sig_mr;
|
||||
sig_wr.wr.opcode = IB_WR_REG_SIG_MR;
|
||||
sig_wr.wr.wr_id = ISER_FASTREG_LI_WRID;
|
||||
sig_wr.wr.sg_list = &rdma_wr->ib_sg[DATA];
|
||||
sig_wr.wr.num_sge = 1;
|
||||
sig_wr.access_flags = IB_ACCESS_LOCAL_WRITE;
|
||||
sig_wr.sig_attrs = &sig_attrs;
|
||||
sig_wr.sig_mr = pi_ctx->sig_mr;
|
||||
if (se_cmd->t_prot_sg)
|
||||
sig_wr.wr.sig_handover.prot = &rdma_wr->ib_sg[PROT];
|
||||
sig_wr.prot = &rdma_wr->ib_sg[PROT];
|
||||
|
||||
if (!wr)
|
||||
wr = &sig_wr;
|
||||
wr = &sig_wr.wr;
|
||||
else
|
||||
wr->next = &sig_wr;
|
||||
wr->next = &sig_wr.wr;
|
||||
|
||||
ret = ib_post_send(isert_conn->qp, wr, &bad_wr);
|
||||
if (ret) {
|
||||
@@ -2858,7 +2858,7 @@ isert_reg_rdma(struct iscsi_conn *conn, struct iscsi_cmd *cmd,
|
||||
struct isert_cmd *isert_cmd = iscsit_priv_cmd(cmd);
|
||||
struct isert_conn *isert_conn = conn->context;
|
||||
struct fast_reg_descriptor *fr_desc = NULL;
|
||||
struct ib_send_wr *send_wr;
|
||||
struct ib_rdma_wr *rdma_wr;
|
||||
struct ib_sge *ib_sg;
|
||||
u32 offset;
|
||||
int ret = 0;
|
||||
@@ -2899,26 +2899,26 @@ isert_reg_rdma(struct iscsi_conn *conn, struct iscsi_cmd *cmd,
|
||||
|
||||
memcpy(&wr->s_ib_sge, ib_sg, sizeof(*ib_sg));
|
||||
wr->ib_sge = &wr->s_ib_sge;
|
||||
wr->send_wr_num = 1;
|
||||
memset(&wr->s_send_wr, 0, sizeof(*send_wr));
|
||||
wr->send_wr = &wr->s_send_wr;
|
||||
wr->rdma_wr_num = 1;
|
||||
memset(&wr->s_rdma_wr, 0, sizeof(wr->s_rdma_wr));
|
||||
wr->rdma_wr = &wr->s_rdma_wr;
|
||||
wr->isert_cmd = isert_cmd;
|
||||
|
||||
send_wr = &isert_cmd->rdma_wr.s_send_wr;
|
||||
send_wr->sg_list = &wr->s_ib_sge;
|
||||
send_wr->num_sge = 1;
|
||||
send_wr->wr_id = (uintptr_t)&isert_cmd->tx_desc;
|
||||
rdma_wr = &isert_cmd->rdma_wr.s_rdma_wr;
|
||||
rdma_wr->wr.sg_list = &wr->s_ib_sge;
|
||||
rdma_wr->wr.num_sge = 1;
|
||||
rdma_wr->wr.wr_id = (uintptr_t)&isert_cmd->tx_desc;
|
||||
if (wr->iser_ib_op == ISER_IB_RDMA_WRITE) {
|
||||
send_wr->opcode = IB_WR_RDMA_WRITE;
|
||||
send_wr->wr.rdma.remote_addr = isert_cmd->read_va;
|
||||
send_wr->wr.rdma.rkey = isert_cmd->read_stag;
|
||||
send_wr->send_flags = !isert_prot_cmd(isert_conn, se_cmd) ?
|
||||
rdma_wr->wr.opcode = IB_WR_RDMA_WRITE;
|
||||
rdma_wr->remote_addr = isert_cmd->read_va;
|
||||
rdma_wr->rkey = isert_cmd->read_stag;
|
||||
rdma_wr->wr.send_flags = !isert_prot_cmd(isert_conn, se_cmd) ?
|
||||
0 : IB_SEND_SIGNALED;
|
||||
} else {
|
||||
send_wr->opcode = IB_WR_RDMA_READ;
|
||||
send_wr->wr.rdma.remote_addr = isert_cmd->write_va;
|
||||
send_wr->wr.rdma.rkey = isert_cmd->write_stag;
|
||||
send_wr->send_flags = IB_SEND_SIGNALED;
|
||||
rdma_wr->wr.opcode = IB_WR_RDMA_READ;
|
||||
rdma_wr->remote_addr = isert_cmd->write_va;
|
||||
rdma_wr->rkey = isert_cmd->write_stag;
|
||||
rdma_wr->wr.send_flags = IB_SEND_SIGNALED;
|
||||
}
|
||||
|
||||
return 0;
|
||||
@@ -2966,8 +2966,8 @@ isert_put_datain(struct iscsi_conn *conn, struct iscsi_cmd *cmd)
|
||||
isert_init_tx_hdrs(isert_conn, &isert_cmd->tx_desc);
|
||||
isert_init_send_wr(isert_conn, isert_cmd,
|
||||
&isert_cmd->tx_desc.send_wr);
|
||||
isert_cmd->rdma_wr.s_send_wr.next = &isert_cmd->tx_desc.send_wr;
|
||||
wr->send_wr_num += 1;
|
||||
isert_cmd->rdma_wr.s_rdma_wr.wr.next = &isert_cmd->tx_desc.send_wr;
|
||||
wr->rdma_wr_num += 1;
|
||||
|
||||
rc = isert_post_recv(isert_conn, isert_cmd->rx_desc);
|
||||
if (rc) {
|
||||
@@ -2976,7 +2976,7 @@ isert_put_datain(struct iscsi_conn *conn, struct iscsi_cmd *cmd)
|
||||
}
|
||||
}
|
||||
|
||||
rc = ib_post_send(isert_conn->qp, wr->send_wr, &wr_failed);
|
||||
rc = ib_post_send(isert_conn->qp, &wr->rdma_wr->wr, &wr_failed);
|
||||
if (rc)
|
||||
isert_warn("ib_post_send() failed for IB_WR_RDMA_WRITE\n");
|
||||
|
||||
@@ -3010,7 +3010,7 @@ isert_get_dataout(struct iscsi_conn *conn, struct iscsi_cmd *cmd, bool recovery)
|
||||
return rc;
|
||||
}
|
||||
|
||||
rc = ib_post_send(isert_conn->qp, wr->send_wr, &wr_failed);
|
||||
rc = ib_post_send(isert_conn->qp, &wr->rdma_wr->wr, &wr_failed);
|
||||
if (rc)
|
||||
isert_warn("ib_post_send() failed for IB_WR_RDMA_READ\n");
|
||||
|
||||
|
@@ -117,9 +117,9 @@ struct isert_rdma_wr {
|
||||
enum iser_ib_op_code iser_ib_op;
|
||||
struct ib_sge *ib_sge;
|
||||
struct ib_sge s_ib_sge;
|
||||
int send_wr_num;
|
||||
struct ib_send_wr *send_wr;
|
||||
struct ib_send_wr s_send_wr;
|
||||
int rdma_wr_num;
|
||||
struct ib_rdma_wr *rdma_wr;
|
||||
struct ib_rdma_wr s_rdma_wr;
|
||||
struct ib_sge ib_sg[3];
|
||||
struct isert_data_buf data;
|
||||
struct isert_data_buf prot;
|
||||
|
@@ -1306,7 +1306,7 @@ static int srp_map_finish_fr(struct srp_map_state *state,
|
||||
struct srp_target_port *target = ch->target;
|
||||
struct srp_device *dev = target->srp_host->srp_dev;
|
||||
struct ib_send_wr *bad_wr;
|
||||
struct ib_send_wr wr;
|
||||
struct ib_fast_reg_wr wr;
|
||||
struct srp_fr_desc *desc;
|
||||
u32 rkey;
|
||||
|
||||
@@ -1324,17 +1324,17 @@ static int srp_map_finish_fr(struct srp_map_state *state,
|
||||
sizeof(state->pages[0]) * state->npages);
|
||||
|
||||
memset(&wr, 0, sizeof(wr));
|
||||
wr.opcode = IB_WR_FAST_REG_MR;
|
||||
wr.wr_id = FAST_REG_WR_ID_MASK;
|
||||
wr.wr.fast_reg.iova_start = state->base_dma_addr;
|
||||
wr.wr.fast_reg.page_list = desc->frpl;
|
||||
wr.wr.fast_reg.page_list_len = state->npages;
|
||||
wr.wr.fast_reg.page_shift = ilog2(dev->mr_page_size);
|
||||
wr.wr.fast_reg.length = state->dma_len;
|
||||
wr.wr.fast_reg.access_flags = (IB_ACCESS_LOCAL_WRITE |
|
||||
IB_ACCESS_REMOTE_READ |
|
||||
IB_ACCESS_REMOTE_WRITE);
|
||||
wr.wr.fast_reg.rkey = desc->mr->lkey;
|
||||
wr.wr.opcode = IB_WR_FAST_REG_MR;
|
||||
wr.wr.wr_id = FAST_REG_WR_ID_MASK;
|
||||
wr.iova_start = state->base_dma_addr;
|
||||
wr.page_list = desc->frpl;
|
||||
wr.page_list_len = state->npages;
|
||||
wr.page_shift = ilog2(dev->mr_page_size);
|
||||
wr.length = state->dma_len;
|
||||
wr.access_flags = (IB_ACCESS_LOCAL_WRITE |
|
||||
IB_ACCESS_REMOTE_READ |
|
||||
IB_ACCESS_REMOTE_WRITE);
|
||||
wr.rkey = desc->mr->lkey;
|
||||
|
||||
*state->fr.next++ = desc;
|
||||
state->nmdesc++;
|
||||
@@ -1342,7 +1342,7 @@ static int srp_map_finish_fr(struct srp_map_state *state,
|
||||
srp_map_desc(state, state->base_dma_addr, state->dma_len,
|
||||
desc->mr->rkey);
|
||||
|
||||
return ib_post_send(ch->qp, &wr, &bad_wr);
|
||||
return ib_post_send(ch->qp, &wr.wr, &bad_wr);
|
||||
}
|
||||
|
||||
static int srp_finish_mapping(struct srp_map_state *state,
|
||||
|
@@ -2823,7 +2823,7 @@ static int srpt_cm_handler(struct ib_cm_id *cm_id, struct ib_cm_event *event)
|
||||
static int srpt_perform_rdmas(struct srpt_rdma_ch *ch,
|
||||
struct srpt_send_ioctx *ioctx)
|
||||
{
|
||||
struct ib_send_wr wr;
|
||||
struct ib_rdma_wr wr;
|
||||
struct ib_send_wr *bad_wr;
|
||||
struct rdma_iu *riu;
|
||||
int i;
|
||||
@@ -2851,29 +2851,29 @@ static int srpt_perform_rdmas(struct srpt_rdma_ch *ch,
|
||||
|
||||
for (i = 0; i < n_rdma; ++i, ++riu) {
|
||||
if (dir == DMA_FROM_DEVICE) {
|
||||
wr.opcode = IB_WR_RDMA_WRITE;
|
||||
wr.wr_id = encode_wr_id(i == n_rdma - 1 ?
|
||||
wr.wr.opcode = IB_WR_RDMA_WRITE;
|
||||
wr.wr.wr_id = encode_wr_id(i == n_rdma - 1 ?
|
||||
SRPT_RDMA_WRITE_LAST :
|
||||
SRPT_RDMA_MID,
|
||||
ioctx->ioctx.index);
|
||||
} else {
|
||||
wr.opcode = IB_WR_RDMA_READ;
|
||||
wr.wr_id = encode_wr_id(i == n_rdma - 1 ?
|
||||
wr.wr.opcode = IB_WR_RDMA_READ;
|
||||
wr.wr.wr_id = encode_wr_id(i == n_rdma - 1 ?
|
||||
SRPT_RDMA_READ_LAST :
|
||||
SRPT_RDMA_MID,
|
||||
ioctx->ioctx.index);
|
||||
}
|
||||
wr.next = NULL;
|
||||
wr.wr.rdma.remote_addr = riu->raddr;
|
||||
wr.wr.rdma.rkey = riu->rkey;
|
||||
wr.num_sge = riu->sge_cnt;
|
||||
wr.sg_list = riu->sge;
|
||||
wr.wr.next = NULL;
|
||||
wr.remote_addr = riu->raddr;
|
||||
wr.rkey = riu->rkey;
|
||||
wr.wr.num_sge = riu->sge_cnt;
|
||||
wr.wr.sg_list = riu->sge;
|
||||
|
||||
/* only get completion event for the last rdma write */
|
||||
if (i == (n_rdma - 1) && dir == DMA_TO_DEVICE)
|
||||
wr.send_flags = IB_SEND_SIGNALED;
|
||||
wr.wr.send_flags = IB_SEND_SIGNALED;
|
||||
|
||||
ret = ib_post_send(ch->qp, &wr, &bad_wr);
|
||||
ret = ib_post_send(ch->qp, &wr.wr, &bad_wr);
|
||||
if (ret)
|
||||
break;
|
||||
}
|
||||
@@ -2882,11 +2882,11 @@ static int srpt_perform_rdmas(struct srpt_rdma_ch *ch,
|
||||
pr_err("%s[%d]: ib_post_send() returned %d for %d/%d\n",
|
||||
__func__, __LINE__, ret, i, n_rdma);
|
||||
if (ret && i > 0) {
|
||||
wr.num_sge = 0;
|
||||
wr.wr_id = encode_wr_id(SRPT_RDMA_ABORT, ioctx->ioctx.index);
|
||||
wr.send_flags = IB_SEND_SIGNALED;
|
||||
wr.wr.num_sge = 0;
|
||||
wr.wr.wr_id = encode_wr_id(SRPT_RDMA_ABORT, ioctx->ioctx.index);
|
||||
wr.wr.send_flags = IB_SEND_SIGNALED;
|
||||
while (ch->state == CH_LIVE &&
|
||||
ib_post_send(ch->qp, &wr, &bad_wr) != 0) {
|
||||
ib_post_send(ch->qp, &wr.wr, &bad_wr) != 0) {
|
||||
pr_info("Trying to abort failed RDMA transfer [%d]\n",
|
||||
ioctx->ioctx.index);
|
||||
msleep(1000);
|
||||
|
Reference in New Issue
Block a user