Merge branch 'for-linus' of master.kernel.org:/pub/scm/linux/kernel/git/roland/infiniband
* 'for-linus' of master.kernel.org:/pub/scm/linux/kernel/git/roland/infiniband: IB/cm: Improve local id allocation IPoIB/cm: Fix SRQ WR leak IB/ipoib: Fix typos in error messages IB/mlx4: Check if SRQ is full when posting receive IB/mlx4: Pass send queue sizes from userspace to kernel IB/mlx4: Fix check of opcode in mlx4_ib_post_send() mlx4_core: Fix array overrun in dump_dev_cap_flags() IB/mlx4: Fix RESET to RESET and RESET to ERROR transitions IB/mthca: Fix RESET to ERROR transition IB/mlx4: Set GRH:HopLimit when sending globally routed MADs IB/mthca: Set GRH:HopLimit when building MLX headers IB/mlx4: Fix check of max_qp_dest_rdma in modify QP IB/mthca: Fix use-after-free on device restart IB/ehca: Return proper error code if register_mr fails IPoIB: Handle P_Key table reordering IB/core: Use start_port() and end_port() IB/core: Add helpers for uncached GID and P_Key searches IB/ipath: Fix potential deadlock with multicast spinlocks IB/core: Free umem when mm is already gone
This commit is contained in:
@@ -279,6 +279,7 @@ int mthca_read_ah(struct mthca_dev *dev, struct mthca_ah *ah,
|
||||
(be32_to_cpu(ah->av->sl_tclass_flowlabel) >> 20) & 0xff;
|
||||
header->grh.flow_label =
|
||||
ah->av->sl_tclass_flowlabel & cpu_to_be32(0xfffff);
|
||||
header->grh.hop_limit = ah->av->hop_limit;
|
||||
ib_get_cached_gid(&dev->ib_dev,
|
||||
be32_to_cpu(ah->av->port_pd) >> 24,
|
||||
ah->av->gid_index % dev->limits.gid_table_len,
|
||||
|
@@ -1250,12 +1250,14 @@ static void __mthca_remove_one(struct pci_dev *pdev)
|
||||
int __mthca_restart_one(struct pci_dev *pdev)
|
||||
{
|
||||
struct mthca_dev *mdev;
|
||||
int hca_type;
|
||||
|
||||
mdev = pci_get_drvdata(pdev);
|
||||
if (!mdev)
|
||||
return -ENODEV;
|
||||
hca_type = mdev->hca_type;
|
||||
__mthca_remove_one(pdev);
|
||||
return __mthca_init_one(pdev, mdev->hca_type);
|
||||
return __mthca_init_one(pdev, hca_type);
|
||||
}
|
||||
|
||||
static int __devinit mthca_init_one(struct pci_dev *pdev,
|
||||
|
@@ -296,7 +296,7 @@ static int to_mthca_st(int transport)
|
||||
}
|
||||
}
|
||||
|
||||
static void store_attrs(struct mthca_sqp *sqp, struct ib_qp_attr *attr,
|
||||
static void store_attrs(struct mthca_sqp *sqp, const struct ib_qp_attr *attr,
|
||||
int attr_mask)
|
||||
{
|
||||
if (attr_mask & IB_QP_PKEY_INDEX)
|
||||
@@ -328,7 +328,7 @@ static void init_port(struct mthca_dev *dev, int port)
|
||||
mthca_warn(dev, "INIT_IB returned status %02x.\n", status);
|
||||
}
|
||||
|
||||
static __be32 get_hw_access_flags(struct mthca_qp *qp, struct ib_qp_attr *attr,
|
||||
static __be32 get_hw_access_flags(struct mthca_qp *qp, const struct ib_qp_attr *attr,
|
||||
int attr_mask)
|
||||
{
|
||||
u8 dest_rd_atomic;
|
||||
@@ -511,7 +511,7 @@ out:
|
||||
return err;
|
||||
}
|
||||
|
||||
static int mthca_path_set(struct mthca_dev *dev, struct ib_ah_attr *ah,
|
||||
static int mthca_path_set(struct mthca_dev *dev, const struct ib_ah_attr *ah,
|
||||
struct mthca_qp_path *path, u8 port)
|
||||
{
|
||||
path->g_mylmc = ah->src_path_bits & 0x7f;
|
||||
@@ -539,12 +539,12 @@ static int mthca_path_set(struct mthca_dev *dev, struct ib_ah_attr *ah,
|
||||
return 0;
|
||||
}
|
||||
|
||||
int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
|
||||
struct ib_udata *udata)
|
||||
static int __mthca_modify_qp(struct ib_qp *ibqp,
|
||||
const struct ib_qp_attr *attr, int attr_mask,
|
||||
enum ib_qp_state cur_state, enum ib_qp_state new_state)
|
||||
{
|
||||
struct mthca_dev *dev = to_mdev(ibqp->device);
|
||||
struct mthca_qp *qp = to_mqp(ibqp);
|
||||
enum ib_qp_state cur_state, new_state;
|
||||
struct mthca_mailbox *mailbox;
|
||||
struct mthca_qp_param *qp_param;
|
||||
struct mthca_qp_context *qp_context;
|
||||
@@ -552,60 +552,6 @@ int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
|
||||
u8 status;
|
||||
int err = -EINVAL;
|
||||
|
||||
mutex_lock(&qp->mutex);
|
||||
|
||||
if (attr_mask & IB_QP_CUR_STATE) {
|
||||
cur_state = attr->cur_qp_state;
|
||||
} else {
|
||||
spin_lock_irq(&qp->sq.lock);
|
||||
spin_lock(&qp->rq.lock);
|
||||
cur_state = qp->state;
|
||||
spin_unlock(&qp->rq.lock);
|
||||
spin_unlock_irq(&qp->sq.lock);
|
||||
}
|
||||
|
||||
new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state;
|
||||
|
||||
if (!ib_modify_qp_is_ok(cur_state, new_state, ibqp->qp_type, attr_mask)) {
|
||||
mthca_dbg(dev, "Bad QP transition (transport %d) "
|
||||
"%d->%d with attr 0x%08x\n",
|
||||
qp->transport, cur_state, new_state,
|
||||
attr_mask);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (cur_state == new_state && cur_state == IB_QPS_RESET) {
|
||||
err = 0;
|
||||
goto out;
|
||||
}
|
||||
|
||||
if ((attr_mask & IB_QP_PKEY_INDEX) &&
|
||||
attr->pkey_index >= dev->limits.pkey_table_len) {
|
||||
mthca_dbg(dev, "P_Key index (%u) too large. max is %d\n",
|
||||
attr->pkey_index, dev->limits.pkey_table_len-1);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if ((attr_mask & IB_QP_PORT) &&
|
||||
(attr->port_num == 0 || attr->port_num > dev->limits.num_ports)) {
|
||||
mthca_dbg(dev, "Port number (%u) is invalid\n", attr->port_num);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC &&
|
||||
attr->max_rd_atomic > dev->limits.max_qp_init_rdma) {
|
||||
mthca_dbg(dev, "Max rdma_atomic as initiator %u too large (max is %d)\n",
|
||||
attr->max_rd_atomic, dev->limits.max_qp_init_rdma);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC &&
|
||||
attr->max_dest_rd_atomic > 1 << dev->qp_table.rdb_shift) {
|
||||
mthca_dbg(dev, "Max rdma_atomic as responder %u too large (max %d)\n",
|
||||
attr->max_dest_rd_atomic, 1 << dev->qp_table.rdb_shift);
|
||||
goto out;
|
||||
}
|
||||
|
||||
mailbox = mthca_alloc_mailbox(dev, GFP_KERNEL);
|
||||
if (IS_ERR(mailbox)) {
|
||||
err = PTR_ERR(mailbox);
|
||||
@@ -892,6 +838,98 @@ int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
|
||||
|
||||
out_mailbox:
|
||||
mthca_free_mailbox(dev, mailbox);
|
||||
out:
|
||||
return err;
|
||||
}
|
||||
|
||||
static const struct ib_qp_attr dummy_init_attr = { .port_num = 1 };
|
||||
static const int dummy_init_attr_mask[] = {
|
||||
[IB_QPT_UD] = (IB_QP_PKEY_INDEX |
|
||||
IB_QP_PORT |
|
||||
IB_QP_QKEY),
|
||||
[IB_QPT_UC] = (IB_QP_PKEY_INDEX |
|
||||
IB_QP_PORT |
|
||||
IB_QP_ACCESS_FLAGS),
|
||||
[IB_QPT_RC] = (IB_QP_PKEY_INDEX |
|
||||
IB_QP_PORT |
|
||||
IB_QP_ACCESS_FLAGS),
|
||||
[IB_QPT_SMI] = (IB_QP_PKEY_INDEX |
|
||||
IB_QP_QKEY),
|
||||
[IB_QPT_GSI] = (IB_QP_PKEY_INDEX |
|
||||
IB_QP_QKEY),
|
||||
};
|
||||
|
||||
int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
|
||||
struct ib_udata *udata)
|
||||
{
|
||||
struct mthca_dev *dev = to_mdev(ibqp->device);
|
||||
struct mthca_qp *qp = to_mqp(ibqp);
|
||||
enum ib_qp_state cur_state, new_state;
|
||||
int err = -EINVAL;
|
||||
|
||||
mutex_lock(&qp->mutex);
|
||||
if (attr_mask & IB_QP_CUR_STATE) {
|
||||
cur_state = attr->cur_qp_state;
|
||||
} else {
|
||||
spin_lock_irq(&qp->sq.lock);
|
||||
spin_lock(&qp->rq.lock);
|
||||
cur_state = qp->state;
|
||||
spin_unlock(&qp->rq.lock);
|
||||
spin_unlock_irq(&qp->sq.lock);
|
||||
}
|
||||
|
||||
new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state;
|
||||
|
||||
if (!ib_modify_qp_is_ok(cur_state, new_state, ibqp->qp_type, attr_mask)) {
|
||||
mthca_dbg(dev, "Bad QP transition (transport %d) "
|
||||
"%d->%d with attr 0x%08x\n",
|
||||
qp->transport, cur_state, new_state,
|
||||
attr_mask);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if ((attr_mask & IB_QP_PKEY_INDEX) &&
|
||||
attr->pkey_index >= dev->limits.pkey_table_len) {
|
||||
mthca_dbg(dev, "P_Key index (%u) too large. max is %d\n",
|
||||
attr->pkey_index, dev->limits.pkey_table_len-1);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if ((attr_mask & IB_QP_PORT) &&
|
||||
(attr->port_num == 0 || attr->port_num > dev->limits.num_ports)) {
|
||||
mthca_dbg(dev, "Port number (%u) is invalid\n", attr->port_num);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC &&
|
||||
attr->max_rd_atomic > dev->limits.max_qp_init_rdma) {
|
||||
mthca_dbg(dev, "Max rdma_atomic as initiator %u too large (max is %d)\n",
|
||||
attr->max_rd_atomic, dev->limits.max_qp_init_rdma);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC &&
|
||||
attr->max_dest_rd_atomic > 1 << dev->qp_table.rdb_shift) {
|
||||
mthca_dbg(dev, "Max rdma_atomic as responder %u too large (max %d)\n",
|
||||
attr->max_dest_rd_atomic, 1 << dev->qp_table.rdb_shift);
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (cur_state == new_state && cur_state == IB_QPS_RESET) {
|
||||
err = 0;
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (cur_state == IB_QPS_RESET && new_state == IB_QPS_ERR) {
|
||||
err = __mthca_modify_qp(ibqp, &dummy_init_attr,
|
||||
dummy_init_attr_mask[ibqp->qp_type],
|
||||
IB_QPS_RESET, IB_QPS_INIT);
|
||||
if (err)
|
||||
goto out;
|
||||
cur_state = IB_QPS_INIT;
|
||||
}
|
||||
|
||||
err = __mthca_modify_qp(ibqp, attr, attr_mask, cur_state, new_state);
|
||||
|
||||
out:
|
||||
mutex_unlock(&qp->mutex);
|
||||
|
Fai riferimento in un nuovo problema
Block a user