net: rename NET_{ADD|INC}_STATS_BH()
Rename NET_INC_STATS_BH() to __NET_INC_STATS() and NET_ADD_STATS_BH() to __NET_ADD_STATS() Signed-off-by: Eric Dumazet <edumazet@google.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:

committed by
David S. Miller

parent
b15084ec7d
commit
02a1d6e7a6
@@ -30,7 +30,7 @@ static void tcp_write_err(struct sock *sk)
|
||||
sk->sk_error_report(sk);
|
||||
|
||||
tcp_done(sk);
|
||||
NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT);
|
||||
__NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT);
|
||||
}
|
||||
|
||||
/* Do not allow orphaned sockets to eat all our resources.
|
||||
@@ -68,7 +68,7 @@ static int tcp_out_of_resources(struct sock *sk, bool do_reset)
|
||||
if (do_reset)
|
||||
tcp_send_active_reset(sk, GFP_ATOMIC);
|
||||
tcp_done(sk);
|
||||
NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY);
|
||||
__NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY);
|
||||
return 1;
|
||||
}
|
||||
return 0;
|
||||
@@ -162,8 +162,8 @@ static int tcp_write_timeout(struct sock *sk)
|
||||
if (tp->syn_fastopen || tp->syn_data)
|
||||
tcp_fastopen_cache_set(sk, 0, NULL, true, 0);
|
||||
if (tp->syn_data && icsk->icsk_retransmits == 1)
|
||||
NET_INC_STATS_BH(sock_net(sk),
|
||||
LINUX_MIB_TCPFASTOPENACTIVEFAIL);
|
||||
__NET_INC_STATS(sock_net(sk),
|
||||
LINUX_MIB_TCPFASTOPENACTIVEFAIL);
|
||||
}
|
||||
retry_until = icsk->icsk_syn_retries ? : net->ipv4.sysctl_tcp_syn_retries;
|
||||
syn_set = true;
|
||||
@@ -178,8 +178,8 @@ static int tcp_write_timeout(struct sock *sk)
|
||||
tp->bytes_acked <= tp->rx_opt.mss_clamp) {
|
||||
tcp_fastopen_cache_set(sk, 0, NULL, true, 0);
|
||||
if (icsk->icsk_retransmits == net->ipv4.sysctl_tcp_retries1)
|
||||
NET_INC_STATS_BH(sock_net(sk),
|
||||
LINUX_MIB_TCPFASTOPENACTIVEFAIL);
|
||||
__NET_INC_STATS(sock_net(sk),
|
||||
LINUX_MIB_TCPFASTOPENACTIVEFAIL);
|
||||
}
|
||||
/* Black hole detection */
|
||||
tcp_mtu_probing(icsk, sk);
|
||||
@@ -228,7 +228,7 @@ void tcp_delack_timer_handler(struct sock *sk)
|
||||
if (!skb_queue_empty(&tp->ucopy.prequeue)) {
|
||||
struct sk_buff *skb;
|
||||
|
||||
NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPSCHEDULERFAILED);
|
||||
__NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSCHEDULERFAILED);
|
||||
|
||||
while ((skb = __skb_dequeue(&tp->ucopy.prequeue)) != NULL)
|
||||
sk_backlog_rcv(sk, skb);
|
||||
@@ -248,7 +248,7 @@ void tcp_delack_timer_handler(struct sock *sk)
|
||||
icsk->icsk_ack.ato = TCP_ATO_MIN;
|
||||
}
|
||||
tcp_send_ack(sk);
|
||||
NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_DELAYEDACKS);
|
||||
__NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS);
|
||||
}
|
||||
|
||||
out:
|
||||
@@ -265,7 +265,7 @@ static void tcp_delack_timer(unsigned long data)
|
||||
tcp_delack_timer_handler(sk);
|
||||
} else {
|
||||
inet_csk(sk)->icsk_ack.blocked = 1;
|
||||
NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED);
|
||||
__NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED);
|
||||
/* deleguate our work to tcp_release_cb() */
|
||||
if (!test_and_set_bit(TCP_DELACK_TIMER_DEFERRED, &tcp_sk(sk)->tsq_flags))
|
||||
sock_hold(sk);
|
||||
@@ -431,7 +431,7 @@ void tcp_retransmit_timer(struct sock *sk)
|
||||
} else {
|
||||
mib_idx = LINUX_MIB_TCPTIMEOUTS;
|
||||
}
|
||||
NET_INC_STATS_BH(sock_net(sk), mib_idx);
|
||||
__NET_INC_STATS(sock_net(sk), mib_idx);
|
||||
}
|
||||
|
||||
tcp_enter_loss(sk);
|
||||
@@ -549,7 +549,7 @@ void tcp_syn_ack_timeout(const struct request_sock *req)
|
||||
{
|
||||
struct net *net = read_pnet(&inet_rsk(req)->ireq_net);
|
||||
|
||||
NET_INC_STATS_BH(net, LINUX_MIB_TCPTIMEOUTS);
|
||||
__NET_INC_STATS(net, LINUX_MIB_TCPTIMEOUTS);
|
||||
}
|
||||
EXPORT_SYMBOL(tcp_syn_ack_timeout);
|
||||
|
||||
|
Reference in New Issue
Block a user