arch: Mass conversion of smp_mb__*()
Mostly scripted conversion of the smp_mb__* barriers. Signed-off-by: Peter Zijlstra <peterz@infradead.org> Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Link: http://lkml.kernel.org/n/tip-55dhyhocezdw1dg7u19hmh1u@git.kernel.org Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: linux-arch@vger.kernel.org Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:

committed by
Ingo Molnar

parent
1b15611e1c
commit
4e857c58ef
@@ -598,7 +598,7 @@ static void rds_ib_set_ack(struct rds_ib_connection *ic, u64 seq,
|
||||
{
|
||||
atomic64_set(&ic->i_ack_next, seq);
|
||||
if (ack_required) {
|
||||
smp_mb__before_clear_bit();
|
||||
smp_mb__before_atomic();
|
||||
set_bit(IB_ACK_REQUESTED, &ic->i_ack_flags);
|
||||
}
|
||||
}
|
||||
@@ -606,7 +606,7 @@ static void rds_ib_set_ack(struct rds_ib_connection *ic, u64 seq,
|
||||
static u64 rds_ib_get_ack(struct rds_ib_connection *ic)
|
||||
{
|
||||
clear_bit(IB_ACK_REQUESTED, &ic->i_ack_flags);
|
||||
smp_mb__after_clear_bit();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
return atomic64_read(&ic->i_ack_next);
|
||||
}
|
||||
|
@@ -429,7 +429,7 @@ static void rds_iw_set_ack(struct rds_iw_connection *ic, u64 seq,
|
||||
{
|
||||
atomic64_set(&ic->i_ack_next, seq);
|
||||
if (ack_required) {
|
||||
smp_mb__before_clear_bit();
|
||||
smp_mb__before_atomic();
|
||||
set_bit(IB_ACK_REQUESTED, &ic->i_ack_flags);
|
||||
}
|
||||
}
|
||||
@@ -437,7 +437,7 @@ static void rds_iw_set_ack(struct rds_iw_connection *ic, u64 seq,
|
||||
static u64 rds_iw_get_ack(struct rds_iw_connection *ic)
|
||||
{
|
||||
clear_bit(IB_ACK_REQUESTED, &ic->i_ack_flags);
|
||||
smp_mb__after_clear_bit();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
return atomic64_read(&ic->i_ack_next);
|
||||
}
|
||||
|
@@ -107,7 +107,7 @@ static int acquire_in_xmit(struct rds_connection *conn)
|
||||
static void release_in_xmit(struct rds_connection *conn)
|
||||
{
|
||||
clear_bit(RDS_IN_XMIT, &conn->c_flags);
|
||||
smp_mb__after_clear_bit();
|
||||
smp_mb__after_atomic();
|
||||
/*
|
||||
* We don't use wait_on_bit()/wake_up_bit() because our waking is in a
|
||||
* hot path and finding waiters is very rare. We don't want to walk
|
||||
@@ -661,7 +661,7 @@ void rds_send_drop_acked(struct rds_connection *conn, u64 ack,
|
||||
|
||||
/* order flag updates with spin locks */
|
||||
if (!list_empty(&list))
|
||||
smp_mb__after_clear_bit();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
spin_unlock_irqrestore(&conn->c_lock, flags);
|
||||
|
||||
@@ -691,7 +691,7 @@ void rds_send_drop_to(struct rds_sock *rs, struct sockaddr_in *dest)
|
||||
}
|
||||
|
||||
/* order flag updates with the rs lock */
|
||||
smp_mb__after_clear_bit();
|
||||
smp_mb__after_atomic();
|
||||
|
||||
spin_unlock_irqrestore(&rs->rs_lock, flags);
|
||||
|
||||
|
@@ -93,7 +93,7 @@ int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm,
|
||||
rm->m_ack_seq = tc->t_last_sent_nxt +
|
||||
sizeof(struct rds_header) +
|
||||
be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1;
|
||||
smp_mb__before_clear_bit();
|
||||
smp_mb__before_atomic();
|
||||
set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags);
|
||||
tc->t_last_expected_una = rm->m_ack_seq + 1;
|
||||
|
||||
|
Reference in New Issue
Block a user