arch: Mass conversion of smp_mb__*()
Mostly scripted conversion of the smp_mb__* barriers. Signed-off-by: Peter Zijlstra <peterz@infradead.org> Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Link: http://lkml.kernel.org/n/tip-55dhyhocezdw1dg7u19hmh1u@git.kernel.org Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: linux-arch@vger.kernel.org Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:

committed by
Ingo Molnar

parent
1b15611e1c
commit
4e857c58ef
@@ -951,7 +951,7 @@ static int tcm_loop_port_link(
|
||||
struct tcm_loop_hba *tl_hba = tl_tpg->tl_hba;
|
||||
|
||||
atomic_inc(&tl_tpg->tl_tpg_port_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
/*
|
||||
* Add Linux/SCSI struct scsi_device by HCTL
|
||||
*/
|
||||
@@ -986,7 +986,7 @@ static void tcm_loop_port_unlink(
|
||||
scsi_device_put(sd);
|
||||
|
||||
atomic_dec(&tl_tpg->tl_tpg_port_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
pr_debug("TCM_Loop_ConfigFS: Port Unlink Successful\n");
|
||||
}
|
||||
|
@@ -393,7 +393,7 @@ target_emulate_set_target_port_groups(struct se_cmd *cmd)
|
||||
continue;
|
||||
|
||||
atomic_inc(&tg_pt_gp->tg_pt_gp_ref_cnt);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
spin_unlock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
|
||||
@@ -404,7 +404,7 @@ target_emulate_set_target_port_groups(struct se_cmd *cmd)
|
||||
|
||||
spin_lock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
atomic_dec(&tg_pt_gp->tg_pt_gp_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
break;
|
||||
}
|
||||
spin_unlock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
@@ -990,7 +990,7 @@ static void core_alua_do_transition_tg_pt_work(struct work_struct *work)
|
||||
* TARGET PORT GROUPS command
|
||||
*/
|
||||
atomic_inc(&mem->tg_pt_gp_mem_ref_cnt);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&tg_pt_gp->tg_pt_gp_lock);
|
||||
|
||||
spin_lock_bh(&port->sep_alua_lock);
|
||||
@@ -1020,7 +1020,7 @@ static void core_alua_do_transition_tg_pt_work(struct work_struct *work)
|
||||
|
||||
spin_lock(&tg_pt_gp->tg_pt_gp_lock);
|
||||
atomic_dec(&mem->tg_pt_gp_mem_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
spin_unlock(&tg_pt_gp->tg_pt_gp_lock);
|
||||
/*
|
||||
@@ -1054,7 +1054,7 @@ static void core_alua_do_transition_tg_pt_work(struct work_struct *work)
|
||||
core_alua_dump_state(tg_pt_gp->tg_pt_gp_alua_pending_state));
|
||||
spin_lock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
atomic_dec(&tg_pt_gp->tg_pt_gp_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
|
||||
if (tg_pt_gp->tg_pt_gp_transition_complete)
|
||||
@@ -1116,7 +1116,7 @@ static int core_alua_do_transition_tg_pt(
|
||||
*/
|
||||
spin_lock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
atomic_inc(&tg_pt_gp->tg_pt_gp_ref_cnt);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
|
||||
if (!explicit && tg_pt_gp->tg_pt_gp_implicit_trans_secs) {
|
||||
@@ -1159,7 +1159,7 @@ int core_alua_do_port_transition(
|
||||
spin_lock(&local_lu_gp_mem->lu_gp_mem_lock);
|
||||
lu_gp = local_lu_gp_mem->lu_gp;
|
||||
atomic_inc(&lu_gp->lu_gp_ref_cnt);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&local_lu_gp_mem->lu_gp_mem_lock);
|
||||
/*
|
||||
* For storage objects that are members of the 'default_lu_gp',
|
||||
@@ -1176,7 +1176,7 @@ int core_alua_do_port_transition(
|
||||
rc = core_alua_do_transition_tg_pt(l_tg_pt_gp,
|
||||
new_state, explicit);
|
||||
atomic_dec(&lu_gp->lu_gp_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
return rc;
|
||||
}
|
||||
/*
|
||||
@@ -1190,7 +1190,7 @@ int core_alua_do_port_transition(
|
||||
|
||||
dev = lu_gp_mem->lu_gp_mem_dev;
|
||||
atomic_inc(&lu_gp_mem->lu_gp_mem_ref_cnt);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&lu_gp->lu_gp_lock);
|
||||
|
||||
spin_lock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
@@ -1219,7 +1219,7 @@ int core_alua_do_port_transition(
|
||||
tg_pt_gp->tg_pt_gp_alua_nacl = NULL;
|
||||
}
|
||||
atomic_inc(&tg_pt_gp->tg_pt_gp_ref_cnt);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
/*
|
||||
* core_alua_do_transition_tg_pt() will always return
|
||||
@@ -1230,7 +1230,7 @@ int core_alua_do_port_transition(
|
||||
|
||||
spin_lock(&dev->t10_alua.tg_pt_gps_lock);
|
||||
atomic_dec(&tg_pt_gp->tg_pt_gp_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
if (rc)
|
||||
break;
|
||||
}
|
||||
@@ -1238,7 +1238,7 @@ int core_alua_do_port_transition(
|
||||
|
||||
spin_lock(&lu_gp->lu_gp_lock);
|
||||
atomic_dec(&lu_gp_mem->lu_gp_mem_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
spin_unlock(&lu_gp->lu_gp_lock);
|
||||
|
||||
@@ -1252,7 +1252,7 @@ int core_alua_do_port_transition(
|
||||
}
|
||||
|
||||
atomic_dec(&lu_gp->lu_gp_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
return rc;
|
||||
}
|
||||
|
||||
|
@@ -225,7 +225,7 @@ struct se_dev_entry *core_get_se_deve_from_rtpi(
|
||||
continue;
|
||||
|
||||
atomic_inc(&deve->pr_ref_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock_irq(&nacl->device_list_lock);
|
||||
|
||||
return deve;
|
||||
@@ -1392,7 +1392,7 @@ int core_dev_add_initiator_node_lun_acl(
|
||||
spin_lock(&lun->lun_acl_lock);
|
||||
list_add_tail(&lacl->lacl_list, &lun->lun_acl_list);
|
||||
atomic_inc(&lun->lun_acl_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&lun->lun_acl_lock);
|
||||
|
||||
pr_debug("%s_TPG[%hu]_LUN[%u->%u] - Added %s ACL for "
|
||||
@@ -1426,7 +1426,7 @@ int core_dev_del_initiator_node_lun_acl(
|
||||
spin_lock(&lun->lun_acl_lock);
|
||||
list_del(&lacl->lacl_list);
|
||||
atomic_dec(&lun->lun_acl_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&lun->lun_acl_lock);
|
||||
|
||||
core_disable_device_list_for_node(lun, NULL, lacl->mapped_lun,
|
||||
|
@@ -323,7 +323,7 @@ static void iblock_bio_done(struct bio *bio, int err)
|
||||
* Bump the ib_bio_err_cnt and release bio.
|
||||
*/
|
||||
atomic_inc(&ibr->ib_bio_err_cnt);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
|
||||
bio_put(bio);
|
||||
|
@@ -675,7 +675,7 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
|
||||
spin_lock(&dev->se_port_lock);
|
||||
list_for_each_entry_safe(port, port_tmp, &dev->dev_sep_list, sep_list) {
|
||||
atomic_inc(&port->sep_tg_pt_ref_cnt);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&dev->se_port_lock);
|
||||
|
||||
spin_lock_bh(&port->sep_alua_lock);
|
||||
@@ -710,7 +710,7 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
|
||||
continue;
|
||||
|
||||
atomic_inc(&deve_tmp->pr_ref_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock_bh(&port->sep_alua_lock);
|
||||
/*
|
||||
* Grab a configfs group dependency that is released
|
||||
@@ -723,9 +723,9 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
|
||||
pr_err("core_scsi3_lunacl_depend"
|
||||
"_item() failed\n");
|
||||
atomic_dec(&port->sep_tg_pt_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
atomic_dec(&deve_tmp->pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
goto out;
|
||||
}
|
||||
/*
|
||||
@@ -740,9 +740,9 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
|
||||
sa_res_key, all_tg_pt, aptpl);
|
||||
if (!pr_reg_atp) {
|
||||
atomic_dec(&port->sep_tg_pt_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
atomic_dec(&deve_tmp->pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
core_scsi3_lunacl_undepend_item(deve_tmp);
|
||||
goto out;
|
||||
}
|
||||
@@ -755,7 +755,7 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
|
||||
|
||||
spin_lock(&dev->se_port_lock);
|
||||
atomic_dec(&port->sep_tg_pt_ref_cnt);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
spin_unlock(&dev->se_port_lock);
|
||||
|
||||
@@ -1110,7 +1110,7 @@ static struct t10_pr_registration *__core_scsi3_locate_pr_reg(
|
||||
continue;
|
||||
}
|
||||
atomic_inc(&pr_reg->pr_res_holders);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&pr_tmpl->registration_lock);
|
||||
return pr_reg;
|
||||
}
|
||||
@@ -1125,7 +1125,7 @@ static struct t10_pr_registration *__core_scsi3_locate_pr_reg(
|
||||
continue;
|
||||
|
||||
atomic_inc(&pr_reg->pr_res_holders);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&pr_tmpl->registration_lock);
|
||||
return pr_reg;
|
||||
}
|
||||
@@ -1155,7 +1155,7 @@ static struct t10_pr_registration *core_scsi3_locate_pr_reg(
|
||||
static void core_scsi3_put_pr_reg(struct t10_pr_registration *pr_reg)
|
||||
{
|
||||
atomic_dec(&pr_reg->pr_res_holders);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
|
||||
static int core_scsi3_check_implicit_release(
|
||||
@@ -1349,7 +1349,7 @@ static void core_scsi3_tpg_undepend_item(struct se_portal_group *tpg)
|
||||
&tpg->tpg_group.cg_item);
|
||||
|
||||
atomic_dec(&tpg->tpg_pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
|
||||
static int core_scsi3_nodeacl_depend_item(struct se_node_acl *nacl)
|
||||
@@ -1369,7 +1369,7 @@ static void core_scsi3_nodeacl_undepend_item(struct se_node_acl *nacl)
|
||||
|
||||
if (nacl->dynamic_node_acl) {
|
||||
atomic_dec(&nacl->acl_pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -1377,7 +1377,7 @@ static void core_scsi3_nodeacl_undepend_item(struct se_node_acl *nacl)
|
||||
&nacl->acl_group.cg_item);
|
||||
|
||||
atomic_dec(&nacl->acl_pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
|
||||
static int core_scsi3_lunacl_depend_item(struct se_dev_entry *se_deve)
|
||||
@@ -1408,7 +1408,7 @@ static void core_scsi3_lunacl_undepend_item(struct se_dev_entry *se_deve)
|
||||
*/
|
||||
if (!lun_acl) {
|
||||
atomic_dec(&se_deve->pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
return;
|
||||
}
|
||||
nacl = lun_acl->se_lun_nacl;
|
||||
@@ -1418,7 +1418,7 @@ static void core_scsi3_lunacl_undepend_item(struct se_dev_entry *se_deve)
|
||||
&lun_acl->se_lun_group.cg_item);
|
||||
|
||||
atomic_dec(&se_deve->pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
|
||||
static sense_reason_t
|
||||
@@ -1552,14 +1552,14 @@ core_scsi3_decode_spec_i_port(
|
||||
continue;
|
||||
|
||||
atomic_inc(&tmp_tpg->tpg_pr_ref_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&dev->se_port_lock);
|
||||
|
||||
if (core_scsi3_tpg_depend_item(tmp_tpg)) {
|
||||
pr_err(" core_scsi3_tpg_depend_item()"
|
||||
" for tmp_tpg\n");
|
||||
atomic_dec(&tmp_tpg->tpg_pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
ret = TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
|
||||
goto out_unmap;
|
||||
}
|
||||
@@ -1573,7 +1573,7 @@ core_scsi3_decode_spec_i_port(
|
||||
tmp_tpg, i_str);
|
||||
if (dest_node_acl) {
|
||||
atomic_inc(&dest_node_acl->acl_pr_ref_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
spin_unlock_irq(&tmp_tpg->acl_node_lock);
|
||||
|
||||
@@ -1587,7 +1587,7 @@ core_scsi3_decode_spec_i_port(
|
||||
pr_err("configfs_depend_item() failed"
|
||||
" for dest_node_acl->acl_group\n");
|
||||
atomic_dec(&dest_node_acl->acl_pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
core_scsi3_tpg_undepend_item(tmp_tpg);
|
||||
ret = TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
|
||||
goto out_unmap;
|
||||
@@ -1647,7 +1647,7 @@ core_scsi3_decode_spec_i_port(
|
||||
pr_err("core_scsi3_lunacl_depend_item()"
|
||||
" failed\n");
|
||||
atomic_dec(&dest_se_deve->pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
core_scsi3_nodeacl_undepend_item(dest_node_acl);
|
||||
core_scsi3_tpg_undepend_item(dest_tpg);
|
||||
ret = TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
|
||||
@@ -3168,14 +3168,14 @@ core_scsi3_emulate_pro_register_and_move(struct se_cmd *cmd, u64 res_key,
|
||||
continue;
|
||||
|
||||
atomic_inc(&dest_se_tpg->tpg_pr_ref_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&dev->se_port_lock);
|
||||
|
||||
if (core_scsi3_tpg_depend_item(dest_se_tpg)) {
|
||||
pr_err("core_scsi3_tpg_depend_item() failed"
|
||||
" for dest_se_tpg\n");
|
||||
atomic_dec(&dest_se_tpg->tpg_pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
ret = TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
|
||||
goto out_put_pr_reg;
|
||||
}
|
||||
@@ -3273,7 +3273,7 @@ after_iport_check:
|
||||
initiator_str);
|
||||
if (dest_node_acl) {
|
||||
atomic_inc(&dest_node_acl->acl_pr_ref_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
spin_unlock_irq(&dest_se_tpg->acl_node_lock);
|
||||
|
||||
@@ -3289,7 +3289,7 @@ after_iport_check:
|
||||
pr_err("core_scsi3_nodeacl_depend_item() for"
|
||||
" dest_node_acl\n");
|
||||
atomic_dec(&dest_node_acl->acl_pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
dest_node_acl = NULL;
|
||||
ret = TCM_INVALID_PARAMETER_LIST;
|
||||
goto out;
|
||||
@@ -3314,7 +3314,7 @@ after_iport_check:
|
||||
if (core_scsi3_lunacl_depend_item(dest_se_deve)) {
|
||||
pr_err("core_scsi3_lunacl_depend_item() failed\n");
|
||||
atomic_dec(&dest_se_deve->pr_ref_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
dest_se_deve = NULL;
|
||||
ret = TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
|
||||
goto out;
|
||||
@@ -3880,7 +3880,7 @@ core_scsi3_pri_read_full_status(struct se_cmd *cmd)
|
||||
add_desc_len = 0;
|
||||
|
||||
atomic_inc(&pr_reg->pr_res_holders);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock(&pr_tmpl->registration_lock);
|
||||
/*
|
||||
* Determine expected length of $FABRIC_MOD specific
|
||||
@@ -3894,7 +3894,7 @@ core_scsi3_pri_read_full_status(struct se_cmd *cmd)
|
||||
" out of buffer: %d\n", cmd->data_length);
|
||||
spin_lock(&pr_tmpl->registration_lock);
|
||||
atomic_dec(&pr_reg->pr_res_holders);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
break;
|
||||
}
|
||||
/*
|
||||
@@ -3956,7 +3956,7 @@ core_scsi3_pri_read_full_status(struct se_cmd *cmd)
|
||||
|
||||
spin_lock(&pr_tmpl->registration_lock);
|
||||
atomic_dec(&pr_reg->pr_res_holders);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
/*
|
||||
* Set the ADDITIONAL DESCRIPTOR LENGTH
|
||||
*/
|
||||
|
@@ -736,7 +736,7 @@ void target_qf_do_work(struct work_struct *work)
|
||||
list_for_each_entry_safe(cmd, cmd_tmp, &qf_cmd_list, se_qf_node) {
|
||||
list_del(&cmd->se_qf_node);
|
||||
atomic_dec(&dev->dev_qf_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
pr_debug("Processing %s cmd: %p QUEUE_FULL in work queue"
|
||||
" context: %s\n", cmd->se_tfo->get_fabric_name(), cmd,
|
||||
@@ -1148,7 +1148,7 @@ transport_check_alloc_task_attr(struct se_cmd *cmd)
|
||||
* Dormant to Active status.
|
||||
*/
|
||||
cmd->se_ordered_id = atomic_inc_return(&dev->dev_ordered_id);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
pr_debug("Allocated se_ordered_id: %u for Task Attr: 0x%02x on %s\n",
|
||||
cmd->se_ordered_id, cmd->sam_task_attr,
|
||||
dev->transport->name);
|
||||
@@ -1705,7 +1705,7 @@ static bool target_handle_task_attr(struct se_cmd *cmd)
|
||||
return false;
|
||||
case MSG_ORDERED_TAG:
|
||||
atomic_inc(&dev->dev_ordered_sync);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
pr_debug("Added ORDERED for CDB: 0x%02x to ordered list, "
|
||||
" se_ordered_id: %u\n",
|
||||
@@ -1723,7 +1723,7 @@ static bool target_handle_task_attr(struct se_cmd *cmd)
|
||||
* For SIMPLE and UNTAGGED Task Attribute commands
|
||||
*/
|
||||
atomic_inc(&dev->simple_cmds);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
break;
|
||||
}
|
||||
|
||||
@@ -1828,7 +1828,7 @@ static void transport_complete_task_attr(struct se_cmd *cmd)
|
||||
|
||||
if (cmd->sam_task_attr == MSG_SIMPLE_TAG) {
|
||||
atomic_dec(&dev->simple_cmds);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
dev->dev_cur_ordered_id++;
|
||||
pr_debug("Incremented dev->dev_cur_ordered_id: %u for"
|
||||
" SIMPLE: %u\n", dev->dev_cur_ordered_id,
|
||||
@@ -1840,7 +1840,7 @@ static void transport_complete_task_attr(struct se_cmd *cmd)
|
||||
cmd->se_ordered_id);
|
||||
} else if (cmd->sam_task_attr == MSG_ORDERED_TAG) {
|
||||
atomic_dec(&dev->dev_ordered_sync);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
dev->dev_cur_ordered_id++;
|
||||
pr_debug("Incremented dev_cur_ordered_id: %u for ORDERED:"
|
||||
@@ -1899,7 +1899,7 @@ static void transport_handle_queue_full(
|
||||
spin_lock_irq(&dev->qf_cmd_lock);
|
||||
list_add_tail(&cmd->se_qf_node, &cmd->se_dev->qf_cmd_list);
|
||||
atomic_inc(&dev->dev_qf_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
spin_unlock_irq(&cmd->se_dev->qf_cmd_lock);
|
||||
|
||||
schedule_work(&cmd->se_dev->qf_work_queue);
|
||||
@@ -2875,7 +2875,7 @@ void transport_send_task_abort(struct se_cmd *cmd)
|
||||
if (cmd->se_tfo->write_pending_status(cmd) != 0) {
|
||||
cmd->transport_state |= CMD_T_ABORTED;
|
||||
cmd->se_cmd_flags |= SCF_SEND_DELAYED_TAS;
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
@@ -162,7 +162,7 @@ int core_scsi3_ua_allocate(
|
||||
spin_unlock_irq(&nacl->device_list_lock);
|
||||
|
||||
atomic_inc(&deve->ua_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
return 0;
|
||||
}
|
||||
list_add_tail(&ua->ua_nacl_list, &deve->ua_list);
|
||||
@@ -175,7 +175,7 @@ int core_scsi3_ua_allocate(
|
||||
asc, ascq);
|
||||
|
||||
atomic_inc(&deve->ua_count);
|
||||
smp_mb__after_atomic_inc();
|
||||
smp_mb__after_atomic();
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -190,7 +190,7 @@ void core_scsi3_ua_release_all(
|
||||
kmem_cache_free(se_ua_cache, ua);
|
||||
|
||||
atomic_dec(&deve->ua_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
spin_unlock(&deve->ua_lock);
|
||||
}
|
||||
@@ -251,7 +251,7 @@ void core_scsi3_ua_for_check_condition(
|
||||
kmem_cache_free(se_ua_cache, ua);
|
||||
|
||||
atomic_dec(&deve->ua_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
spin_unlock(&deve->ua_lock);
|
||||
spin_unlock_irq(&nacl->device_list_lock);
|
||||
@@ -310,7 +310,7 @@ int core_scsi3_ua_clear_for_request_sense(
|
||||
kmem_cache_free(se_ua_cache, ua);
|
||||
|
||||
atomic_dec(&deve->ua_count);
|
||||
smp_mb__after_atomic_dec();
|
||||
smp_mb__after_atomic();
|
||||
}
|
||||
spin_unlock(&deve->ua_lock);
|
||||
spin_unlock_irq(&nacl->device_list_lock);
|
||||
|
Reference in New Issue
Block a user