浏览代码

qcacmn: Update stats in dp_txrx_peer in per packet path

Update stats in dp_txrx_peer in per packet Tx completion and
Rx path.

Change-Id: I807cb5ca9fe2aeeabdd4cb95d6e30cb9781560f4
CRs-Fixed: 3092123
Harsh Kumar Bijlani 3 年之前
父节点
当前提交
93cd9e0af8
共有 10 个文件被更改,包括 473 次插入269 次删除
  1. 16 8
      dp/wifi3.0/be/dp_be_rx.c
  2. 62 23
      dp/wifi3.0/dp_internal.h
  3. 9 2
      dp/wifi3.0/dp_ipa.c
  4. 150 89
      dp/wifi3.0/dp_rx.c
  5. 3 2
      dp/wifi3.0/dp_rx.h
  6. 38 26
      dp/wifi3.0/dp_rx_err.c
  7. 11 6
      dp/wifi3.0/dp_stats.c
  8. 159 97
      dp/wifi3.0/dp_tx.c
  9. 21 12
      dp/wifi3.0/li/dp_li_rx.c
  10. 4 4
      dp/wifi3.0/li/dp_li_tx.c

+ 16 - 8
dp/wifi3.0/be/dp_be_rx.c

@@ -646,7 +646,9 @@ done:
 
 			if (qdf_nbuf_is_raw_frame(nbuf)) {
 				DP_STATS_INC(vdev->pdev, rx_raw_pkts, 1);
-				DP_STATS_INC_PKT(peer, rx.raw, 1, msdu_len);
+				DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+							      rx.raw, 1,
+							      msdu_len);
 			} else {
 				dp_rx_nbuf_free(nbuf);
 				DP_STATS_INC(soc, rx.err.scatter_msdu, 1);
@@ -671,7 +673,9 @@ done:
 		if (qdf_unlikely(vdev->multipass_en)) {
 			if (dp_rx_multipass_process(txrx_peer, nbuf,
 						    tid) == false) {
-				DP_STATS_INC(peer, rx.multipass_rx_pkt_drop, 1);
+				DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+							  rx.multipass_rx_pkt_drop,
+							  1);
 				dp_rx_nbuf_free(nbuf);
 				nbuf = next;
 				continue;
@@ -680,7 +684,8 @@ done:
 
 		if (!dp_wds_rx_policy_check(rx_tlv_hdr, vdev, txrx_peer)) {
 			dp_rx_err("%pK: Policy Check Drop pkt", soc);
-			DP_STATS_INC(peer, rx.policy_check_drop, 1);
+			DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+						  rx.policy_check_drop, 1);
 			tid_stats->fail_cnt[POLICY_CHECK_DROP]++;
 			/* Drop & free packet */
 			dp_rx_nbuf_free(nbuf);
@@ -695,7 +700,8 @@ done:
 								rx_tlv_hdr) ==
 				  false))) {
 			tid_stats->fail_cnt[NAWDS_MCAST_DROP]++;
-			DP_STATS_INC(peer, rx.nawds_mcast_drop, 1);
+			DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+						  rx.nawds_mcast_drop, 1);
 			dp_rx_nbuf_free(nbuf);
 			nbuf = next;
 			continue;
@@ -711,8 +717,9 @@ done:
 					qdf_nbuf_is_ipv4_wapi_pkt(nbuf);
 
 			if (!is_eapol) {
-				DP_STATS_INC(peer,
-					     rx.peer_unauth_rx_pkt_drop, 1);
+				DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+							  rx.peer_unauth_rx_pkt_drop,
+							  1);
 				dp_rx_nbuf_free(nbuf);
 				nbuf = next;
 				continue;
@@ -780,8 +787,9 @@ done:
 					  QDF_NBUF_CB_RX_PKT_LEN(nbuf),
 					  enh_flag);
 		if (qdf_unlikely(txrx_peer->in_twt))
-			DP_STATS_INC_PKT(peer, rx.to_stack_twt, 1,
-					 QDF_NBUF_CB_RX_PKT_LEN(nbuf));
+			DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+						      rx.to_stack_twt, 1,
+						      QDF_NBUF_CB_RX_PKT_LEN(nbuf));
 
 		tid_stats->delivered_to_stack++;
 		nbuf = next;

+ 62 - 23
dp/wifi3.0/dp_internal.h

@@ -754,7 +754,7 @@ void DP_PRINT_STATS(const char *fmt, ...);
 		_handle->stats._field += _delta; \
 }
 
-#define DP_STATS_FLAT_INC(_handle, _field, _delta) \
+#define DP_PEER_STATS_FLAT_INC(_handle, _field, _delta) \
 { \
 	if (likely(_handle)) \
 		_handle->_field += _delta; \
@@ -772,7 +772,7 @@ void DP_PRINT_STATS(const char *fmt, ...);
 		_handle->stats._field -= _delta; \
 }
 
-#define DP_STATS_FLAT_DEC(_handle, _field, _delta) \
+#define DP_PEER_STATS_FLAT_DEC(_handle, _field, _delta) \
 { \
 	if (likely(_handle)) \
 		_handle->_field -= _delta; \
@@ -790,10 +790,10 @@ void DP_PRINT_STATS(const char *fmt, ...);
 	DP_STATS_INC(_handle, _field.bytes, _bytes) \
 }
 
-#define DP_STATS_FLAT_INC_PKT(_handle, _field, _count, _bytes) \
+#define DP_PEER_STATS_FLAT_INC_PKT(_handle, _field, _count, _bytes) \
 { \
-	DP_STATS_FLAT_INC(_handle, _field.num, _count); \
-	DP_STATS_FLAT_INC(_handle, _field.bytes, _bytes) \
+	DP_PEER_STATS_FLAT_INC(_handle, _field.num, _count); \
+	DP_PEER_STATS_FLAT_INC(_handle, _field.bytes, _bytes) \
 }
 
 #define DP_STATS_INCC_PKT(_handle, _field, _count, _bytes, _cond) \
@@ -820,79 +820,118 @@ void DP_PRINT_STATS(const char *fmt, ...);
 
 #else
 #define DP_STATS_INC(_handle, _field, _delta)
-#define DP_STATS_FLAT_INC(_handle, _field, _delta)
+#define DP_PEER_STATS_FLAT_INC(_handle, _field, _delta)
 #define DP_STATS_INCC(_handle, _field, _delta, _cond)
 #define DP_STATS_DEC(_handle, _field, _delta)
-#define DP_STATS_FLAT_DEC(_handle, _field, _delta)
+#define DP_PEER_STATS_FLAT_DEC(_handle, _field, _delta)
 #define DP_STATS_UPD(_handle, _field, _delta)
 #define DP_STATS_INC_PKT(_handle, _field, _count, _bytes)
-#define DP_STATS_FLAT_INC_PKT(_handle, _field, _count, _bytes)
+#define DP_PEER_STATS_FLAT_INC_PKT(_handle, _field, _count, _bytes)
 #define DP_STATS_INCC_PKT(_handle, _field, _count, _bytes, _cond)
 #define DP_STATS_AGGR(_handle_a, _handle_b, _field)
 #define DP_STATS_AGGR_PKT(_handle_a, _handle_b, _field)
 #endif
 
+#define DP_PEER_PER_PKT_STATS_INC(_handle, _field, _delta) \
+{ \
+	DP_STATS_INC(_handle, per_pkt_stats._field, _delta); \
+}
+
+#define DP_PEER_PER_PKT_STATS_INCC(_handle, _field, _delta, _cond) \
+{ \
+	DP_STATS_INCC(_handle, per_pkt_stats._field, _delta, _cond); \
+}
+
+#define DP_PEER_PER_PKT_STATS_INC_PKT(_handle, _field, _count, _bytes) \
+{ \
+	DP_PEER_PER_PKT_STATS_INC(_handle, _field.num, _count); \
+	DP_PEER_PER_PKT_STATS_INC(_handle, _field.bytes, _bytes) \
+}
+
+#define DP_PEER_PER_PKT_STATS_INCC_PKT(_handle, _field, _count, _bytes, _cond) \
+{ \
+	DP_PEER_PER_PKT_STATS_INCC(_handle, _field.num, _count, _cond); \
+	DP_PEER_PER_PKT_STATS_INCC(_handle, _field.bytes, _bytes, _cond) \
+}
+
+#ifndef QCA_ENHANCED_STATS_SUPPORT
+#define DP_PEER_EXTD_STATS_INC(_handle, _field, _delta) \
+{ \
+	DP_STATS_INC(_handle, extd_stats._field, _delta); \
+}
+
+#define DP_PEER_EXTD_STATS_INCC(_handle, _field, _delta, _cond) \
+{ \
+	DP_STATS_INCC(_handle, extd_stats._field, _delta, _cond); \
+}
+
+#define DP_PEER_EXTD_STATS_UPD(_handle, _field, _delta) \
+{ \
+	DP_STATS_UPD(_handle, extd_stats._field, _delta); \
+}
+#endif
+
 #if defined(QCA_VDEV_STATS_HW_OFFLOAD_SUPPORT) && \
 	defined(QCA_ENHANCED_STATS_SUPPORT)
 #define DP_PEER_TO_STACK_INCC_PKT(_handle, _count, _bytes, _cond) \
 { \
 	if (!(_handle->hw_txrx_stats_en) || _cond) \
-		DP_STATS_FLAT_INC_PKT(_handle, to_stack, _count, _bytes); \
+		DP_PEER_STATS_FLAT_INC_PKT(_handle, to_stack, _count, _bytes); \
 }
 
 #define DP_PEER_TO_STACK_DECC(_handle, _count, _cond) \
 { \
 	if (!(_handle->hw_txrx_stats_en) || _cond) \
-		DP_STATS_FLAT_DEC(_handle, to_stack.num, _count); \
+		DP_PEER_STATS_FLAT_DEC(_handle, to_stack.num, _count); \
 }
 
 #define DP_PEER_MC_INCC_PKT(_handle, _count, _bytes, _cond) \
 { \
 	if (!(_handle->hw_txrx_stats_en) || _cond) \
-		DP_STATS_FLAT_INC_PKT(_handle, multicast, _count, _bytes); \
+		DP_PEER_PER_PKT_STATS_INC_PKT(_handle, rx.multicast, _count, _bytes); \
 }
 
 #define DP_PEER_BC_INCC_PKT(_handle, _count, _bytes, _cond) \
 { \
 	if (!(_handle->hw_txrx_stats_en) || _cond) \
-		DP_STATS_FLAT_INC_PKT(_handle, bcast, _count, _bytes); \
+		DP_PEER_PER_PKT_STATS_INC_PKT(_handle, rx.bcast, _count, _bytes); \
 }
 #elif defined(QCA_VDEV_STATS_HW_OFFLOAD_SUPPORT)
 #define DP_PEER_TO_STACK_INCC_PKT(_handle, _count, _bytes, _cond) \
 { \
 	if (!(_handle->hw_txrx_stats_en)) \
-		DP_STATS_FLAT_INC_PKT(_handle, to_stack, _count, _bytes); \
+		DP_PEER_STATS_FLAT_INC_PKT(_handle, to_stack, _count, _bytes); \
 }
 
 #define DP_PEER_TO_STACK_DECC(_handle, _count, _cond) \
 { \
 	if (!(_handle->hw_txrx_stats_en)) \
-		DP_STATS_FLAT_DEC(_handle, to_stack.num, _count); \
+		DP_PEER_STATS_FLAT_DEC(_handle, to_stack.num, _count); \
 }
 
 #define DP_PEER_MC_INCC_PKT(_handle, _count, _bytes, _cond) \
 { \
 	if (!(_handle->hw_txrx_stats_en)) \
-		DP_STATS_FLAT_INC_PKT(_handle, multicast, _count, _bytes); \
+		DP_PEER_PER_PKT_STATS_INC_PKT(_handle, rx.multicast, _count, _bytes); \
 }
 
 #define DP_PEER_BC_INCC_PKT(_handle, _count, _bytes, _cond) \
 { \
 	if (!(_handle->hw_txrx_stats_en)) \
-		DP_STATS_FLAT_INC_PKT(_handle, bcast, _count, _bytes); \
+		DP_PEER_PER_PKT_STATS_INC_PKT(_handle, rx.bcast, _count, _bytes); \
 }
 #else
 #define DP_PEER_TO_STACK_INCC_PKT(_handle, _count, _bytes, _cond) \
-	DP_STATS_FLAT_INC_PKT(_handle, to_stack, _count, _bytes);
+	DP_PEER_STATS_FLAT_INC_PKT(_handle, to_stack, _count, _bytes);
 
 #define DP_PEER_TO_STACK_DECC(_handle, _count, _cond) \
-	DP_STATS_FLAT_DEC(_handle, to_stack.num, _count);
+	DP_PEER_STATS_FLAT_DEC(_handle, to_stack.num, _count);
 
 #define DP_PEER_MC_INCC_PKT(_handle, _count, _bytes, _cond) \
-	DP_STATS_FLAT_INC_PKT(_handle, multicast, _count, _bytes);
+	DP_PEER_PER_PKT_STATS_INC_PKT(_handle, rx.multicast, _count, _bytes);
 
 #define DP_PEER_BC_INCC_PKT(_handle, _count, _bytes, _cond) \
-	DP_STATS_FLAT_INC_PKT(_handle, bcast, _count, _bytes);
+	DP_PEER_PER_PKT_STATS_INC_PKT(_handle, rx.bcast, _count, _bytes);
 #endif
 
 #ifdef ENABLE_DP_HIST_STATS
@@ -2072,7 +2111,7 @@ dp_hif_update_pipe_callback(struct dp_soc *dp_soc, void *cb_context,
  * dp_vdev_peer_stats_update_protocol_cnt() - update per-peer protocol counters
  * @vdev: VDEV DP object
  * @nbuf: data packet
- * @peer: Peer DP object
+ * @peer: DP TXRX Peer object
  * @is_egress: whether egress or ingress
  * @is_rx: whether rx or tx
  *
@@ -2081,7 +2120,7 @@ dp_hif_update_pipe_callback(struct dp_soc *dp_soc, void *cb_context,
  */
 void dp_vdev_peer_stats_update_protocol_cnt(struct dp_vdev *vdev,
 					    qdf_nbuf_t nbuf,
-					    struct dp_peer *peer,
+					    struct dp_txrx_peer *txrx_peer,
 					    bool is_egress,
 					    bool is_rx);
 
@@ -2107,7 +2146,7 @@ void dp_vdev_peer_stats_update_protocol_cnt_tx(struct dp_vdev *vdev_hdl,
 					       qdf_nbuf_t nbuf);
 
 #else
-#define dp_vdev_peer_stats_update_protocol_cnt(vdev, nbuf, peer, \
+#define dp_vdev_peer_stats_update_protocol_cnt(vdev, nbuf, txrx_peer, \
 					       is_egress, is_rx)
 
 static inline

+ 9 - 2
dp/wifi3.0/dp_ipa.c

@@ -2748,16 +2748,23 @@ static qdf_nbuf_t dp_ipa_intrabss_send(struct dp_pdev *pdev,
 	if (qdf_unlikely(!vdev_peer))
 		return nbuf;
 
+	if (qdf_unlikely(!vdev_peer->txrx_peer)) {
+		dp_peer_unref_delete(vdev_peer, DP_MOD_ID_IPA);
+		return nbuf;
+	}
+
 	qdf_mem_zero(nbuf->cb, sizeof(nbuf->cb));
 	len = qdf_nbuf_len(nbuf);
 
 	if (dp_tx_send((struct cdp_soc_t *)pdev->soc, vdev->vdev_id, nbuf)) {
-		DP_STATS_INC_PKT(vdev_peer, rx.intra_bss.fail, 1, len);
+		DP_PEER_PER_PKT_STATS_INC_PKT(vdev_peer->txrx_peer,
+					      rx.intra_bss.fail, 1, len);
 		dp_peer_unref_delete(vdev_peer, DP_MOD_ID_IPA);
 		return nbuf;
 	}
 
-	DP_STATS_INC_PKT(vdev_peer, rx.intra_bss.pkts, 1, len);
+	DP_PEER_PER_PKT_STATS_INC_PKT(vdev_peer->txrx_peer,
+				      rx.intra_bss.pkts, 1, len);
 	dp_peer_unref_delete(vdev_peer, DP_MOD_ID_IPA);
 	return NULL;
 }

+ 150 - 89
dp/wifi3.0/dp_rx.c

@@ -774,7 +774,8 @@ dp_rx_deliver_raw(struct dp_vdev *vdev, qdf_nbuf_t nbuf_list,
 		DP_RX_LIST_APPEND(deliver_list_head, deliver_list_tail, nbuf);
 
 		DP_STATS_INC(vdev->pdev, rx_raw_pkts, 1);
-		DP_STATS_INC_PKT(txrx_peer, rx.raw, 1, qdf_nbuf_len(nbuf));
+		DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, rx.raw, 1,
+					      qdf_nbuf_len(nbuf));
 		/*
 		 * reset the chfrag_start and chfrag_end bits in nbuf cb
 		 * as this is a non-amsdu pkt and RAW mode simulation expects
@@ -843,11 +844,13 @@ bool dp_rx_intrabss_mcbc_fwd(struct dp_soc *soc, struct dp_txrx_peer *ta_peer,
 				  CB_FTYPE_INTRABSS_FWD);
 	if (dp_tx_send((struct cdp_soc_t *)soc,
 		       ta_peer->vdev->vdev_id, nbuf_copy)) {
-		DP_STATS_INC_PKT(ta_peer, rx.intra_bss.fail, 1, len);
+		DP_PEER_PER_PKT_STATS_INC_PKT(ta_peer, rx.intra_bss.fail, 1,
+					      len);
 		tid_stats->fail_cnt[INTRABSS_DROP]++;
 		dp_rx_nbuf_free(nbuf_copy);
 	} else {
-		DP_STATS_INC_PKT(ta_peer, rx.intra_bss.pkts, 1, len);
+		DP_PEER_PER_PKT_STATS_INC_PKT(ta_peer, rx.intra_bss.pkts, 1,
+					      len);
 		tid_stats->intrabss_cnt++;
 	}
 	return false;
@@ -883,8 +886,9 @@ bool dp_rx_intrabss_ucast_fwd(struct dp_soc *soc, struct dp_txrx_peer *ta_peer,
 
 		nbuf = qdf_nbuf_unshare(nbuf);
 		if (!nbuf) {
-			DP_STATS_INC_PKT(ta_peer,
-					 rx.intra_bss.fail, 1, len);
+			DP_PEER_PER_PKT_STATS_INC_PKT(ta_peer,
+						      rx.intra_bss.fail,
+						      1, len);
 			/* return true even though the pkt is
 			 * not forwarded. Basically skb_unshare
 			 * failed and we want to continue with
@@ -897,11 +901,11 @@ bool dp_rx_intrabss_ucast_fwd(struct dp_soc *soc, struct dp_txrx_peer *ta_peer,
 
 	if (!dp_tx_send((struct cdp_soc_t *)soc,
 			tx_vdev_id, nbuf)) {
-		DP_STATS_INC_PKT(ta_peer, rx.intra_bss.pkts, 1,
-				 len);
+		DP_PEER_PER_PKT_STATS_INC_PKT(ta_peer, rx.intra_bss.pkts, 1,
+					      len);
 	} else {
-		DP_STATS_INC_PKT(ta_peer, rx.intra_bss.fail, 1,
-				 len);
+		DP_PEER_PER_PKT_STATS_INC_PKT(ta_peer, rx.intra_bss.fail, 1,
+					      len);
 		tid_stats->fail_cnt[INTRABSS_DROP]++;
 		return false;
 	}
@@ -940,6 +944,9 @@ void dp_rx_fill_mesh_stats(struct dp_vdev *vdev, qdf_nbuf_t nbuf,
 	uint32_t center_chan_freq;
 	struct dp_soc *soc = vdev->pdev->soc;
 	struct dp_peer *peer;
+	struct dp_peer *primary_link_peer;
+	struct dp_soc *link_peer_soc;
+	cdp_peer_stats_param_t buf = {0};
 
 	/* fill recv mesh stats */
 	rx_info = qdf_mem_malloc(sizeof(struct mesh_recv_hdr_s));
@@ -973,10 +980,22 @@ void dp_rx_fill_mesh_stats(struct dp_vdev *vdev, qdf_nbuf_t nbuf,
 						   rx_info->rs_keyix);
 		}
 
-		rx_info->rs_snr = peer->stats.rx.snr;
 		dp_peer_unref_delete(peer, DP_MOD_ID_MESH);
 	}
 
+	primary_link_peer = dp_get_primary_link_peer_by_id(soc,
+							   txrx_peer->peer_id,
+							   DP_MOD_ID_MESH);
+
+	if (qdf_likely(primary_link_peer)) {
+		link_peer_soc = primary_link_peer->vdev->pdev->soc;
+		dp_monitor_peer_get_stats_param(link_peer_soc,
+						primary_link_peer,
+						cdp_peer_rx_snr, &buf);
+		rx_info->rs_snr = buf.rx_snr;
+		dp_peer_unref_delete(primary_link_peer, DP_MOD_ID_MESH);
+	}
+
 	rx_info->rs_rssi = rx_info->rs_snr + DP_DEFAULT_NOISEFLOOR;
 
 	soc = vdev->pdev->soc;
@@ -1738,6 +1757,7 @@ dp_rx_enqueue_rx(struct dp_txrx_peer *txrx_peer, qdf_nbuf_t rx_buf_list)
 	struct dp_peer_cached_bufq *bufqi = &txrx_peer->bufq_info;
 	int num_buff_elem;
 	QDF_STATUS ret = QDF_STATUS_SUCCESS;
+	struct dp_soc *soc = txrx_peer->vdev->pdev->soc;
 	struct dp_peer *peer = dp_peer_get_ref_by_id(soc, txrx_peer->peer_id,
 						     DP_MOD_ID_RX);
 
@@ -1873,7 +1893,8 @@ static void dp_rx_check_delivery_to_stack(struct dp_soc *soc,
 		num_nbuf = dp_rx_drop_nbuf_list(vdev->pdev, nbuf_head);
 		DP_STATS_INC(soc, rx.err.rejected, num_nbuf);
 		if (txrx_peer)
-			DP_STATS_DEC(txrx_peer, to_stack.num, num_nbuf);
+			DP_PEER_STATS_FLAT_DEC(txrx_peer, to_stack.num,
+					       num_nbuf);
 	}
 }
 #endif /* ifdef DELIVERY_TO_STACK_STATUS_CHECK */
@@ -1966,27 +1987,128 @@ QDF_STATUS dp_rx_eapol_deliver_to_stack(struct dp_soc *soc,
 
 #ifndef QCA_HOST_MODE_WIFI_DISABLED
 #ifdef VDEV_PEER_PROTOCOL_COUNT
-#define dp_rx_msdu_stats_update_prot_cnts(vdev_hdl, nbuf, peer) \
+#define dp_rx_msdu_stats_update_prot_cnts(vdev_hdl, nbuf, txrx_peer) \
 { \
 	qdf_nbuf_t nbuf_local; \
-	struct dp_peer *peer_local; \
+	struct dp_txrx_peer *txrx_peer_local; \
 	struct dp_vdev *vdev_local = vdev_hdl; \
 	do { \
 		if (qdf_likely(!((vdev_local)->peer_protocol_count_track))) \
 			break; \
 		nbuf_local = nbuf; \
-		peer_local = peer; \
+		txrx_peer_local = txrx_peer; \
 		if (qdf_unlikely(qdf_nbuf_is_frag((nbuf_local)))) \
 			break; \
 		else if (qdf_unlikely(qdf_nbuf_is_raw_frame((nbuf_local)))) \
 			break; \
 		dp_vdev_peer_stats_update_protocol_cnt((vdev_local), \
 						       (nbuf_local), \
-						       (peer_local), 0, 1); \
+						       (txrx_peer_local), 0, 1); \
 	} while (0); \
 }
 #else
-#define dp_rx_msdu_stats_update_prot_cnts(vdev_hdl, nbuf, peer)
+#define dp_rx_msdu_stats_update_prot_cnts(vdev_hdl, nbuf, txrx_peer)
+#endif
+
+#ifndef QCA_ENHANCED_STATS_SUPPORT
+/**
+ * dp_rx_msdu_extd_stats_update(): Update Rx extended path stats for peer
+ *
+ * @soc: datapath soc handle
+ * @nbuf: received msdu buffer
+ * @rx_tlv_hdr: rx tlv header
+ * @txrx_peer: datapath txrx_peer handle
+ *
+ * Return: void
+ */
+static inline
+void dp_rx_msdu_extd_stats_update(struct dp_soc *soc, qdf_nbuf_t nbuf,
+				  uint8_t *rx_tlv_hdr,
+				  struct dp_txrx_peer *txrx_peer)
+{
+	bool is_ampdu;
+	uint32_t sgi, mcs, tid, nss, bw, reception_type, pkt_type;
+
+	/*
+	 * TODO - For KIWI this field is present in ring_desc
+	 * Try to use ring desc instead of tlv.
+	 */
+	is_ampdu = hal_rx_mpdu_info_ampdu_flag_get(soc->hal_soc, rx_tlv_hdr);
+	DP_PEER_EXTD_STATS_INCC(txrx_peer, rx.ampdu_cnt, 1, is_ampdu);
+	DP_PEER_EXTD_STATS_INCC(txrx_peer, rx.non_ampdu_cnt, 1, !(is_ampdu));
+
+	sgi = hal_rx_tlv_sgi_get(soc->hal_soc, rx_tlv_hdr);
+	mcs = hal_rx_tlv_rate_mcs_get(soc->hal_soc, rx_tlv_hdr);
+	tid = qdf_nbuf_get_tid_val(nbuf);
+	bw = hal_rx_tlv_bw_get(soc->hal_soc, rx_tlv_hdr);
+	reception_type = hal_rx_msdu_start_reception_type_get(soc->hal_soc,
+							      rx_tlv_hdr);
+	nss = hal_rx_msdu_start_nss_get(soc->hal_soc, rx_tlv_hdr);
+	pkt_type = hal_rx_tlv_get_pkt_type(soc->hal_soc, rx_tlv_hdr);
+
+	DP_PEER_EXTD_STATS_INCC(txrx_peer, rx.rx_mpdu_cnt[mcs], 1,
+		      ((mcs < MAX_MCS) && QDF_NBUF_CB_RX_CHFRAG_START(nbuf)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer, rx.rx_mpdu_cnt[MAX_MCS - 1], 1,
+		      ((mcs >= MAX_MCS) && QDF_NBUF_CB_RX_CHFRAG_START(nbuf)));
+	DP_PEER_EXTD_STATS_INC(txrx_peer, rx.bw[bw], 1);
+	/*
+	 * only if nss > 0 and pkt_type is 11N/AC/AX,
+	 * then increase index [nss - 1] in array counter.
+	 */
+	if (nss > 0 && (pkt_type == DOT11_N ||
+			pkt_type == DOT11_AC ||
+			pkt_type == DOT11_AX))
+		DP_PEER_EXTD_STATS_INC(txrx_peer, rx.nss[nss - 1], 1);
+
+	DP_PEER_EXTD_STATS_INC(txrx_peer, rx.sgi_count[sgi], 1);
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, rx.err.mic_err, 1,
+				   hal_rx_tlv_mic_err_get(soc->hal_soc,
+				   rx_tlv_hdr));
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, rx.err.decrypt_err, 1,
+				   hal_rx_tlv_decrypt_err_get(soc->hal_soc,
+				   rx_tlv_hdr));
+
+	DP_PEER_EXTD_STATS_INC(txrx_peer, rx.wme_ac_type[TID_TO_WME_AC(tid)], 1);
+	DP_PEER_EXTD_STATS_INC(txrx_peer, rx.reception_type[reception_type], 1);
+
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS_11A) && (pkt_type == DOT11_A)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs <= MAX_MCS_11A) && (pkt_type == DOT11_A)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS_11B) && (pkt_type == DOT11_B)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs <= MAX_MCS_11B) && (pkt_type == DOT11_B)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS_11A) && (pkt_type == DOT11_N)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs <= MAX_MCS_11A) && (pkt_type == DOT11_N)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS_11AC) && (pkt_type == DOT11_AC)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs <= MAX_MCS_11AC) && (pkt_type == DOT11_AC)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS) && (pkt_type == DOT11_AX)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				rx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs < MAX_MCS) && (pkt_type == DOT11_AX)));
+}
+#else
+static inline
+void dp_rx_msdu_extd_stats_update(struct dp_soc *soc, qdf_nbuf_t nbuf,
+				  uint8_t *rx_tlv_hdr,
+				  struct dp_txrx_peer *txrx_peer)
+{
+}
 #endif
 
 /**
@@ -2007,8 +2129,7 @@ void dp_rx_msdu_stats_update(struct dp_soc *soc, qdf_nbuf_t nbuf,
 			     uint8_t ring_id,
 			     struct cdp_tid_rx_stats *tid_stats)
 {
-	bool is_ampdu, is_not_amsdu;
-	uint32_t sgi, mcs, tid, nss, bw, reception_type, pkt_type;
+	bool is_not_amsdu;
 	struct dp_vdev *vdev = txrx_peer->vdev;
 	bool enh_flag;
 	qdf_ether_header_t *eh;
@@ -2017,89 +2138,29 @@ void dp_rx_msdu_stats_update(struct dp_soc *soc, qdf_nbuf_t nbuf,
 	dp_rx_msdu_stats_update_prot_cnts(vdev, nbuf, txrx_peer);
 	is_not_amsdu = qdf_nbuf_is_rx_chfrag_start(nbuf) &
 			qdf_nbuf_is_rx_chfrag_end(nbuf);
-	DP_STATS_INC_PKT(txrx_peer, rx.rcvd_reo[ring_id], 1, msdu_len);
-	DP_STATS_INCC(txrx_peer, rx.non_amsdu_cnt, 1, is_not_amsdu);
-	DP_STATS_INCC(txrx_peer, rx.amsdu_cnt, 1, !is_not_amsdu);
-	DP_STATS_INCC(peer, rx.rx_retries, 1, qdf_nbuf_is_rx_retry_flag(nbuf));
+	DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, rx.rcvd_reo[ring_id], 1,
+				      msdu_len);
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, rx.non_amsdu_cnt, 1,
+				   is_not_amsdu);
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, rx.amsdu_cnt, 1, !is_not_amsdu);
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, rx.rx_retries, 1,
+				   qdf_nbuf_is_rx_retry_flag(nbuf));
 	tid_stats->msdu_cnt++;
 	if (qdf_unlikely(qdf_nbuf_is_da_mcbc(nbuf) &&
 			 (vdev->rx_decap_type == htt_cmn_pkt_type_ethernet))) {
 		eh = (qdf_ether_header_t *)qdf_nbuf_data(nbuf);
 		enh_flag = vdev->pdev->enhanced_stats_en;
-		DP_PEER_MC_INCC_PKT(peer, 1, msdu_len, enh_flag);
+		DP_PEER_MC_INCC_PKT(txrx_peer, 1, msdu_len, enh_flag);
 		tid_stats->mcast_msdu_cnt++;
 		if (QDF_IS_ADDR_BROADCAST(eh->ether_dhost)) {
-			DP_PEER_BC_INCC_PKT(peer, 1, msdu_len, enh_flag);
+			DP_PEER_BC_INCC_PKT(txrx_peer, 1, msdu_len, enh_flag);
 			tid_stats->bcast_msdu_cnt++;
 		}
 	}
 
-	/*
-	 * currently we can return from here as we have similar stats
-	 * updated at per ppdu level instead of msdu level
-	 */
-	if (!soc->process_rx_status)
-		return;
+	txrx_peer->stats.per_pkt_stats.rx.last_rx_ts = qdf_system_ticks();
 
-	peer->stats.rx.last_rx_ts = qdf_system_ticks();
-
-	/*
-	 * TODO - For KIWI this field is present in ring_desc
-	 * Try to use ring desc instead of tlv.
-	 */
-	is_ampdu = hal_rx_mpdu_info_ampdu_flag_get(soc->hal_soc, rx_tlv_hdr);
-	DP_STATS_INCC(peer, rx.ampdu_cnt, 1, is_ampdu);
-	DP_STATS_INCC(peer, rx.non_ampdu_cnt, 1, !(is_ampdu));
-	sgi = hal_rx_tlv_sgi_get(soc->hal_soc, rx_tlv_hdr);
-	mcs = hal_rx_tlv_rate_mcs_get(soc->hal_soc, rx_tlv_hdr);
-	tid = qdf_nbuf_get_tid_val(nbuf);
-	bw = hal_rx_tlv_bw_get(soc->hal_soc, rx_tlv_hdr);
-	reception_type = hal_rx_msdu_start_reception_type_get(soc->hal_soc,
-							      rx_tlv_hdr);
-	nss = hal_rx_msdu_start_nss_get(soc->hal_soc, rx_tlv_hdr);
-	pkt_type = hal_rx_tlv_get_pkt_type(soc->hal_soc, rx_tlv_hdr);
-	DP_STATS_INCC(peer, rx.rx_mpdu_cnt[mcs], 1,
-		      ((mcs < MAX_MCS) && QDF_NBUF_CB_RX_CHFRAG_START(nbuf)));
-	DP_STATS_INCC(peer, rx.rx_mpdu_cnt[MAX_MCS - 1], 1,
-		      ((mcs >= MAX_MCS) && QDF_NBUF_CB_RX_CHFRAG_START(nbuf)));
-	DP_STATS_INC(peer, rx.bw[bw], 1);
-	/*
-	 * only if nss > 0 and pkt_type is 11N/AC/AX,
-	 * then increase index [nss - 1] in array counter.
-	 */
-	if (nss > 0 && (pkt_type == DOT11_N ||
-			pkt_type == DOT11_AC ||
-			pkt_type == DOT11_AX))
-		DP_STATS_INC(peer, rx.nss[nss - 1], 1);
-	DP_STATS_INC(peer, rx.sgi_count[sgi], 1);
-	DP_STATS_INCC(peer, rx.err.mic_err, 1,
-		      hal_rx_tlv_mic_err_get(soc->hal_soc, rx_tlv_hdr));
-	DP_STATS_INCC(peer, rx.err.decrypt_err, 1,
-		      hal_rx_tlv_decrypt_err_get(soc->hal_soc, rx_tlv_hdr));
-
-	DP_STATS_INC(peer, rx.wme_ac_type[TID_TO_WME_AC(tid)], 1);
-	DP_STATS_INC(peer, rx.reception_type[reception_type], 1);
-
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-		      ((mcs >= MAX_MCS_11A) && (pkt_type == DOT11_A)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[mcs], 1,
-		      ((mcs <= MAX_MCS_11A) && (pkt_type == DOT11_A)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-		      ((mcs >= MAX_MCS_11B) && (pkt_type == DOT11_B)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[mcs], 1,
-		      ((mcs <= MAX_MCS_11B) && (pkt_type == DOT11_B)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-		      ((mcs >= MAX_MCS_11A) && (pkt_type == DOT11_N)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[mcs], 1,
-		      ((mcs <= MAX_MCS_11A) && (pkt_type == DOT11_N)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-		      ((mcs >= MAX_MCS_11AC) && (pkt_type == DOT11_AC)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[mcs], 1,
-		      ((mcs <= MAX_MCS_11AC) && (pkt_type == DOT11_AC)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-		      ((mcs >= MAX_MCS) && (pkt_type == DOT11_AX)));
-	DP_STATS_INCC(peer, rx.pkt_type[pkt_type].mcs_count[mcs], 1,
-		      ((mcs < MAX_MCS) && (pkt_type == DOT11_AX)));
+	dp_rx_msdu_extd_stats_update(soc, nbuf, rx_tlv_hdr, txrx_peer);
 }
 
 #ifndef WDS_VENDOR_EXTENSION

+ 3 - 2
dp/wifi3.0/dp_rx.h

@@ -263,10 +263,11 @@ bool dp_rx_check_ndi_mdns_fwding(struct dp_txrx_peer *ta_txrx_peer,
 {
 	if (ta_txrx_peer->vdev->opmode == wlan_op_mode_ndi &&
 	    qdf_nbuf_is_ipv6_mdns_pkt(nbuf)) {
-		DP_STATS_INC(ta_txrx_peer, rx.intra_bss.mdns_no_fwd, 1);
+		DP_PEER_PER_PKT_STATS_INC(ta_txrx_peer,
+					  rx.intra_bss.mdns_no_fwd, 1);
 		return false;
 	}
-		return true;
+	return true;
 }
 #else
 static inline

+ 38 - 26
dp/wifi3.0/dp_rx_err.c

@@ -1404,7 +1404,8 @@ dp_rx_null_q_desc_handle(struct dp_soc *soc, qdf_nbuf_t nbuf,
 	if ((!soc->mec_fw_offload) &&
 	    dp_rx_mcast_echo_check(soc, txrx_peer, rx_tlv_hdr, nbuf)) {
 		/* this is a looped back MCBC pkt, drop it */
-		DP_STATS_INC_PKT(peer, rx.mec_drop, 1, qdf_nbuf_len(nbuf));
+		DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, rx.mec_drop, 1,
+					      qdf_nbuf_len(nbuf));
 		goto drop_nbuf;
 	}
 
@@ -1414,7 +1415,8 @@ dp_rx_null_q_desc_handle(struct dp_soc *soc, qdf_nbuf_t nbuf,
 	 * from any proxysta.
 	 */
 	if (check_qwrap_multicast_loopback(vdev, nbuf)) {
-		DP_STATS_INC_PKT(peer, rx.mec_drop, 1, qdf_nbuf_len(nbuf));
+		DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, rx.mec_drop, 1,
+					      qdf_nbuf_len(nbuf));
 		goto drop_nbuf;
 	}
 
@@ -1423,13 +1425,13 @@ dp_rx_null_q_desc_handle(struct dp_soc *soc, qdf_nbuf_t nbuf,
 			 hal_rx_msdu_end_da_is_mcbc_get(soc->hal_soc,
 							rx_tlv_hdr))) {
 		dp_err_rl("free buffer for multicast packet");
-		DP_STATS_INC(peer, rx.nawds_mcast_drop, 1);
+		DP_PEER_PER_PKT_STATS_INC(txrx_peer, rx.nawds_mcast_drop, 1);
 		goto drop_nbuf;
 	}
 
 	if (!dp_wds_rx_policy_check(rx_tlv_hdr, vdev, txrx_peer)) {
 		dp_err_rl("mcast Policy Check Drop pkt");
-		DP_STATS_INC(peer, rx.policy_check_drop, 1);
+		DP_PEER_PER_PKT_STATS_INC(txrx_peer, rx.policy_check_drop, 1);
 		goto drop_nbuf;
 	}
 	/* WDS Source Port Learning */
@@ -1440,6 +1442,7 @@ dp_rx_null_q_desc_handle(struct dp_soc *soc, qdf_nbuf_t nbuf,
 					msdu_metadata);
 
 	if (hal_rx_is_unicast(soc->hal_soc, rx_tlv_hdr)) {
+		struct dp_peer *peer;
 		tid = hal_rx_tid_get(soc->hal_soc, rx_tlv_hdr);
 		peer = dp_peer_get_ref_by_id(soc, txrx_peer->peer_id,
 					     DP_MOD_ID_RX_ERR);
@@ -1510,7 +1513,8 @@ dp_rx_null_q_desc_handle(struct dp_soc *soc, qdf_nbuf_t nbuf,
 					    enh_flag);
 
 			if (QDF_IS_ADDR_BROADCAST(eh->ether_dhost))
-				DP_PEER_BC_INCC_PKT(peer, 1, qdf_nbuf_len(nbuf),
+				DP_PEER_BC_INCC_PKT(txrx_peer, 1,
+						    qdf_nbuf_len(nbuf),
 						    enh_flag);
 		}
 
@@ -1669,10 +1673,11 @@ process_rx:
 		eh = (qdf_ether_header_t *)qdf_nbuf_data(nbuf);
 		is_broadcast = (QDF_IS_ADDR_BROADCAST
 				(eh->ether_dhost)) ? 1 : 0 ;
-		DP_STATS_INC_PKT(peer, rx.multicast, 1, qdf_nbuf_len(nbuf));
+		DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, rx.multicast, 1,
+					      qdf_nbuf_len(nbuf));
 		if (is_broadcast) {
-			DP_STATS_INC_PKT(peer, rx.bcast, 1,
-					qdf_nbuf_len(nbuf));
+			DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, rx.bcast, 1,
+						      qdf_nbuf_len(nbuf));
 		}
 	}
 
@@ -1684,7 +1689,7 @@ process_rx:
 					  EXCEPTION_DEST_RING_ID, true, true);
 		/* Update the flow tag in SKB based on FSE metadata */
 		dp_rx_update_flow_tag(soc, vdev, nbuf, rx_tlv_hdr, true);
-		DP_STATS_FLAT_INC(txrx_peer, to_stack.num, 1);
+		DP_PEER_STATS_FLAT_INC(txrx_peer, to_stack.num, 1);
 		qdf_nbuf_set_exc_frame(nbuf, 1);
 		dp_rx_deliver_to_stack(soc, vdev, txrx_peer, nbuf, NULL);
 	}
@@ -2760,10 +2765,10 @@ done:
 				/* TODO */
 				/* Add per error code accounting */
 				case HAL_REO_ERR_REGULAR_FRAME_2K_JUMP:
-					if (peer)
-						DP_STATS_INC(peer,
-							     rx.err.jump_2k_err,
-							     1);
+					if (txrx_peer)
+						DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+									  rx.err.jump_2k_err,
+									  1);
 
 					pool_id = wbm_err_info.pool_id;
 
@@ -2782,8 +2787,9 @@ done:
 					break;
 				case HAL_REO_ERR_REGULAR_FRAME_OOR:
 					if (txrx_peer)
-						DP_STATS_INC(peer,
-							     rx.err.oor_err, 1);
+						DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+									  rx.err.oor_err,
+									  1);
 					if (hal_rx_msdu_end_first_msdu_get(soc->hal_soc,
 									   rx_tlv_hdr)) {
 						tid =
@@ -2811,8 +2817,9 @@ done:
 				case HAL_REO_ERR_PN_CHECK_FAILED:
 				case HAL_REO_ERR_PN_ERROR_HANDLING_FLAG_SET:
 					if (txrx_peer)
-						DP_STATS_INC(txrx_peer,
-							     rx.err.pn_err, 1);
+						DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+									  rx.err.pn_err,
+									  1);
 					dp_rx_nbuf_free(nbuf);
 					break;
 
@@ -2851,10 +2858,10 @@ done:
 				case HAL_RXDMA_ERR_UNENCRYPTED:
 
 				case HAL_RXDMA_ERR_WIFI_PARSE:
-					if (peer)
-						DP_STATS_INC(peer,
-							     rx.err.rxdma_wifi_parse_err,
-							     1);
+					if (txrx_peer)
+						DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+									  rx.err.rxdma_wifi_parse_err,
+									  1);
 
 					pool_id = wbm_err_info.pool_id;
 					dp_rx_process_rxdma_err(soc, nbuf,
@@ -2870,14 +2877,17 @@ done:
 								rx_tlv_hdr,
 								txrx_peer);
 					if (txrx_peer)
-						DP_STATS_INC(peer, rx.err.mic_err, 1);
+						DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+									  rx.err.mic_err,
+									  1);
 					break;
 
 				case HAL_RXDMA_ERR_DECRYPT:
 
 					if (txrx_peer) {
-						DP_STATS_INC(peer, rx.err.
-							     decrypt_err, 1);
+						DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+									  rx.err.decrypt_err,
+									  1);
 						dp_rx_nbuf_free(nbuf);
 						break;
 					}
@@ -2896,8 +2906,10 @@ done:
 								pool_id);
 					break;
 				case HAL_RXDMA_MULTICAST_ECHO:
-					DP_STATS_INC_PKT(peer, rx.mec_drop, 1,
-							 qdf_nbuf_len(nbuf));
+					if (txrx_peer)
+						DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+									      rx.mec_drop, 1,
+									      qdf_nbuf_len(nbuf));
 					dp_rx_nbuf_free(nbuf);
 					break;
 				case HAL_RXDMA_UNAUTHORIZED_WDS:

+ 11 - 6
dp/wifi3.0/dp_stats.c

@@ -4458,17 +4458,18 @@ static QDF_STATUS dp_peer_stats_update_protocol_test_cnt(struct dp_vdev *vdev,
 
 void dp_vdev_peer_stats_update_protocol_cnt(struct dp_vdev *vdev,
 					    qdf_nbuf_t nbuf,
-					    struct dp_peer *peer,
+					    struct dp_txrx_peer *txrx_peer,
 					    bool is_egress,
 					    bool is_rx)
 {
-	struct cdp_peer_stats *peer_stats;
+	struct dp_peer_per_pkt_stats *per_pkt_stats;
 	struct protocol_trace_count *protocol_trace_cnt;
 	enum cdp_protocol_trace prot;
 	struct dp_soc *soc;
 	struct ether_header *eh;
 	char *mac;
 	bool new_peer_ref = false;
+	struct dp_peer *peer = NULL;
 
 	if (qdf_likely(!vdev->peer_protocol_count_track))
 		return;
@@ -4485,14 +4486,18 @@ void dp_vdev_peer_stats_update_protocol_cnt(struct dp_vdev *vdev,
 	else
 		mac = eh->ether_dhost;
 
-	if (!peer) {
+	if (!txrx_peer) {
 		peer = dp_peer_find_hash_find(soc, mac, 0, vdev->vdev_id,
 					      DP_MOD_ID_GENERIC_STATS);
 		new_peer_ref = true;
 		if (!peer)
 			return;
+
+		txrx_peer = peer->txrx_peer;
+		if (!txrx_peer)
+			goto dp_vdev_peer_stats_update_protocol_cnt_free_peer;
 	}
-	peer_stats = &peer->stats;
+	per_pkt_stats = &txrx_peer->stats.per_pkt_stats;
 
 	if (qdf_nbuf_is_icmp_pkt(nbuf) == true)
 		prot = CDP_TRACE_ICMP;
@@ -4504,9 +4509,9 @@ void dp_vdev_peer_stats_update_protocol_cnt(struct dp_vdev *vdev,
 		goto dp_vdev_peer_stats_update_protocol_cnt_free_peer;
 
 	if (is_rx)
-		protocol_trace_cnt = peer_stats->rx.protocol_trace_cnt;
+		protocol_trace_cnt = per_pkt_stats->rx.protocol_trace_cnt;
 	else
-		protocol_trace_cnt = peer_stats->tx.protocol_trace_cnt;
+		protocol_trace_cnt = per_pkt_stats->tx.protocol_trace_cnt;
 
 	if (is_egress)
 		protocol_trace_cnt[prot].egress_cnt++;

+ 159 - 97
dp/wifi3.0/dp_tx.c

@@ -3151,8 +3151,9 @@ void dp_tx_nawds_handler(struct dp_soc *soc, struct dp_vdev *vdev,
 			 */
 			if (sa_peer_id == peer->peer_id) {
 				dp_tx_debug("multicast packet");
-				DP_STATS_INC(peer, tx.nawds_mcast_drop,
-					     1);
+				DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+							  tx.nawds_mcast_drop,
+							  1);
 				continue;
 			}
 
@@ -3173,9 +3174,11 @@ void dp_tx_nawds_handler(struct dp_soc *soc, struct dp_vdev *vdev,
 				dp_tx_debug("pkt send failed");
 				qdf_nbuf_free(nbuf_clone);
 			} else {
-				if (peer_id != DP_INVALID_PEER)
-					DP_STATS_INC_PKT(peer, tx.nawds_mcast,
-							 1, qdf_nbuf_len(nbuf));
+				if (peer_id != DP_INVALID_PEER) {
+					DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+								      tx.nawds_mcast,
+								      1, qdf_nbuf_len(nbuf));
+				}
 			}
 		}
 	}
@@ -3788,7 +3791,8 @@ void dp_tx_compute_delay(struct dp_vdev *vdev, struct dp_tx_desc_s *tx_desc,
 
 #ifdef DISABLE_DP_STATS
 static
-inline void dp_update_no_ack_stats(qdf_nbuf_t nbuf, struct dp_txrx_peer *peer)
+inline void dp_update_no_ack_stats(qdf_nbuf_t nbuf,
+				   struct dp_txrx_peer *txrx_peer)
 {
 }
 #else
@@ -3799,7 +3803,81 @@ dp_update_no_ack_stats(qdf_nbuf_t nbuf, struct dp_txrx_peer *txrx_peer)
 
 	DPTRACE(qdf_dp_track_noack_check(nbuf, &subtype));
 	if (subtype != QDF_PROTO_INVALID)
-		DP_STATS_INC(peer, tx.no_ack_count[subtype], 1);
+		DP_PEER_PER_PKT_STATS_INC(txrx_peer, tx.no_ack_count[subtype],
+					  1);
+}
+#endif
+
+#ifndef QCA_ENHANCED_STATS_SUPPORT
+/**
+ * dp_tx_update_peer_extd_stats()- Update Tx extended path stats for peer
+ *
+ * @ts: Tx compltion status
+ * @txrx_peer: datapath txrx_peer handle
+ *
+ * Return: void
+ */
+static inline void
+dp_tx_update_peer_extd_stats(struct hal_tx_completion_status *ts,
+			     struct dp_txrx_peer *txrx_peer)
+{
+	uint8_t mcs, pkt_type;
+
+	mcs = ts->mcs;
+	pkt_type = ts->pkt_type;
+
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS_11A) && (pkt_type == DOT11_A)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs < (MAX_MCS_11A)) && (pkt_type == DOT11_A)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS_11B) && (pkt_type == DOT11_B)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs < MAX_MCS_11B) && (pkt_type == DOT11_B)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS_11A) && (pkt_type == DOT11_N)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs < MAX_MCS_11A) && (pkt_type == DOT11_N)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= MAX_MCS_11AC) && (pkt_type == DOT11_AC)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs < MAX_MCS_11AC) && (pkt_type == DOT11_AC)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
+				((mcs >= (MAX_MCS - 1)) && (pkt_type == DOT11_AX)));
+	DP_PEER_EXTD_STATS_INCC(txrx_peer,
+				tx.pkt_type[pkt_type].mcs_count[mcs], 1,
+				((mcs < (MAX_MCS - 1)) && (pkt_type == DOT11_AX)));
+
+	DP_PEER_EXTD_STATS_INC(txrx_peer, tx.sgi_count[ts->sgi], 1);
+	DP_PEER_EXTD_STATS_INC(txrx_peer, tx.bw[ts->bw], 1);
+	DP_PEER_EXTD_STATS_UPD(txrx_peer, tx.last_ack_rssi, ts->ack_frame_rssi);
+	DP_PEER_EXTD_STATS_INC(txrx_peer,
+			       tx.wme_ac_type[TID_TO_WME_AC(ts->tid)], 1);
+	DP_PEER_EXTD_STATS_INCC(txrx_peer, tx.stbc, 1, ts->stbc);
+	DP_PEER_EXTD_STATS_INCC(txrx_peer, tx.ldpc, 1, ts->ldpc);
+	DP_PEER_EXTD_STATS_INCC(txrx_peer, tx.retries, 1, ts->transmit_cnt > 1);
+	if (ts->first_msdu) {
+		DP_PEER_EXTD_STATS_INCC(txrx_peer, tx.retries_mpdu, 1,
+					ts->transmit_cnt > 1);
+		DP_PEER_EXTD_STATS_INCC(txrx_peer, tx.mpdu_success_with_retries,
+					qdf_do_div(ts->transmit_cnt, DP_RETRY_COUNT),
+					ts->transmit_cnt > DP_RETRY_COUNT);
+	}
+}
+#else
+static inline void
+dp_tx_update_peer_extd_stats(struct hal_tx_completion_status *ts,
+			     struct dp_txrx_peer *txrx_peer)
+{
 }
 #endif
 
@@ -3821,7 +3899,6 @@ dp_tx_update_peer_stats(struct dp_tx_desc_s *tx_desc,
 {
 	struct dp_pdev *pdev = txrx_peer->vdev->pdev;
 	struct dp_soc *soc = NULL;
-	uint8_t mcs, pkt_type;
 	uint8_t tid = ts->tid;
 	uint32_t length;
 	struct cdp_tid_tx_stats *tid_stats;
@@ -3835,38 +3912,35 @@ dp_tx_update_peer_stats(struct dp_tx_desc_s *tx_desc,
 	tid_stats = &pdev->stats.tid_stats.tid_tx_stats[ring_id][tid];
 	soc = pdev->soc;
 
-	mcs = ts->mcs;
-	pkt_type = ts->pkt_type;
-
 	if (ts->release_src != HAL_TX_COMP_RELEASE_SOURCE_TQM) {
 		dp_err("Release source is not from TQM");
 		return;
 	}
 
 	length = qdf_nbuf_len(tx_desc->nbuf);
-	DP_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, length);
+	DP_PEER_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, length);
 
 	if (qdf_unlikely(pdev->delay_stats_flag))
 		dp_tx_compute_delay(txrx_peer->vdev, tx_desc, tid, ring_id);
-	DP_STATS_INCC(peer, tx.dropped.age_out, 1,
-		     (ts->status == HAL_TX_TQM_RR_REM_CMD_AGED));
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.dropped.age_out, 1,
+			  (ts->status == HAL_TX_TQM_RR_REM_CMD_AGED));
 
-	DP_STATS_INCC_PKT(peer, tx.dropped.fw_rem, 1, length,
+	DP_PEER_PER_PKT_STATS_INCC_PKT(txrx_peer, tx.dropped.fw_rem, 1, length,
 			  (ts->status == HAL_TX_TQM_RR_REM_CMD_REM));
 
-	DP_STATS_INCC(peer, tx.dropped.fw_rem_notx, 1,
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.dropped.fw_rem_notx, 1,
 		     (ts->status == HAL_TX_TQM_RR_REM_CMD_NOTX));
 
-	DP_STATS_INCC(peer, tx.dropped.fw_rem_tx, 1,
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.dropped.fw_rem_tx, 1,
 		     (ts->status == HAL_TX_TQM_RR_REM_CMD_TX));
 
-	DP_STATS_INCC(peer, tx.dropped.fw_reason1, 1,
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.dropped.fw_reason1, 1,
 		     (ts->status == HAL_TX_TQM_RR_FW_REASON1));
 
-	DP_STATS_INCC(peer, tx.dropped.fw_reason2, 1,
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.dropped.fw_reason2, 1,
 		     (ts->status == HAL_TX_TQM_RR_FW_REASON2));
 
-	DP_STATS_INCC(peer, tx.dropped.fw_reason3, 1,
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.dropped.fw_reason3, 1,
 		     (ts->status == HAL_TX_TQM_RR_FW_REASON3));
 	/*
 	 * tx_failed is ideally supposed to be updated from HTT ppdu completion
@@ -3876,73 +3950,40 @@ dp_tx_update_peer_stats(struct dp_tx_desc_s *tx_desc,
 	 * then this has to be removed
 	 */
 	txrx_peer->tx_failed =
-				peer->stats.tx.dropped.fw_rem.num +
-				peer->stats.tx.dropped.fw_rem_notx +
-				peer->stats.tx.dropped.fw_rem_tx +
-				peer->stats.tx.dropped.age_out +
-				peer->stats.tx.dropped.fw_reason1 +
-				peer->stats.tx.dropped.fw_reason2 +
-				peer->stats.tx.dropped.fw_reason3;
+			txrx_peer->stats.per_pkt_stats.tx.dropped.fw_rem.num +
+			txrx_peer->stats.per_pkt_stats.tx.dropped.fw_rem_notx +
+			txrx_peer->stats.per_pkt_stats.tx.dropped.fw_rem_tx +
+			txrx_peer->stats.per_pkt_stats.tx.dropped.age_out +
+			txrx_peer->stats.per_pkt_stats.tx.dropped.fw_reason1 +
+			txrx_peer->stats.per_pkt_stats.tx.dropped.fw_reason2 +
+			txrx_peer->stats.per_pkt_stats.tx.dropped.fw_reason3;
 	if (ts->status < CDP_MAX_TX_TQM_STATUS) {
 		tid_stats->tqm_status_cnt[ts->status]++;
 	}
 
 	if (ts->status != HAL_TX_TQM_RR_FRAME_ACKED) {
-		DP_STATS_INCC(peer, tx.failed_retry_count, 1,
-			      ts->transmit_cnt > DP_RETRY_COUNT);
-		dp_update_no_ack_stats(tx_desc->nbuf, peer);
+		DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.failed_retry_count, 1,
+					   ts->transmit_cnt > DP_RETRY_COUNT);
+		dp_update_no_ack_stats(tx_desc->nbuf, txrx_peer);
 		return;
 	}
 
-	DP_STATS_INCC(peer, tx.retry_count, 1, ts->transmit_cnt > 1);
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.retry_count, 1,
+				   ts->transmit_cnt > 1);
 
-	DP_STATS_INCC(peer, tx.multiple_retry_count, 1, ts->transmit_cnt > 2);
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.multiple_retry_count, 1,
+				   ts->transmit_cnt > 2);
 
-	DP_STATS_INCC(peer, tx.ofdma, 1, ts->ofdma);
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.ofdma, 1, ts->ofdma);
 
-	DP_STATS_INCC(peer, tx.amsdu_cnt, 1, ts->msdu_part_of_amsdu);
-	DP_STATS_INCC(peer, tx.non_amsdu_cnt, 1, !ts->msdu_part_of_amsdu);
-	/*
-	 * Following Rate Statistics are updated from HTT PPDU events from FW.
-	 * Return from here if HTT PPDU events are enabled.
-	 */
-	if (!(soc->process_tx_status))
-		return;
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-			((mcs >= MAX_MCS_11A) && (pkt_type == DOT11_A)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[mcs], 1,
-			((mcs < (MAX_MCS_11A)) && (pkt_type == DOT11_A)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-			((mcs >= MAX_MCS_11B) && (pkt_type == DOT11_B)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[mcs], 1,
-			((mcs < MAX_MCS_11B) && (pkt_type == DOT11_B)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-			((mcs >= MAX_MCS_11A) && (pkt_type == DOT11_N)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[mcs], 1,
-			((mcs < MAX_MCS_11A) && (pkt_type == DOT11_N)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-			((mcs >= MAX_MCS_11AC) && (pkt_type == DOT11_AC)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[mcs], 1,
-			((mcs < MAX_MCS_11AC) && (pkt_type == DOT11_AC)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[MAX_MCS - 1], 1,
-			((mcs >= (MAX_MCS - 1)) && (pkt_type == DOT11_AX)));
-	DP_STATS_INCC(peer, tx.pkt_type[pkt_type].mcs_count[mcs], 1,
-			((mcs < (MAX_MCS - 1)) && (pkt_type == DOT11_AX)));
-
-	DP_STATS_INC(peer, tx.sgi_count[ts->sgi], 1);
-	DP_STATS_INC(peer, tx.bw[ts->bw], 1);
-	DP_STATS_UPD(peer, tx.last_ack_rssi, ts->ack_frame_rssi);
-	DP_STATS_INC(peer, tx.wme_ac_type[TID_TO_WME_AC(ts->tid)], 1);
-	DP_STATS_INCC(peer, tx.stbc, 1, ts->stbc);
-	DP_STATS_INCC(peer, tx.ldpc, 1, ts->ldpc);
-	DP_STATS_INCC(peer, tx.retries, 1, ts->transmit_cnt > 1);
-	if (ts->first_msdu) {
-		DP_STATS_INCC(peer, tx.retries_mpdu, 1, ts->transmit_cnt > 1);
-		DP_STATS_INCC(peer, tx.mpdu_success_with_retries,
-			      qdf_do_div(ts->transmit_cnt, DP_RETRY_COUNT),
-			      ts->transmit_cnt > DP_RETRY_COUNT);
-	}
-	peer->stats.tx.last_tx_ts = qdf_system_ticks();
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.amsdu_cnt, 1,
+				   ts->msdu_part_of_amsdu);
+	DP_PEER_PER_PKT_STATS_INCC(txrx_peer, tx.non_amsdu_cnt, 1,
+				   !ts->msdu_part_of_amsdu);
+
+	txrx_peer->stats.per_pkt_stats.tx.last_tx_ts = qdf_system_ticks();
+
+	dp_tx_update_peer_extd_stats(ts, txrx_peer);
 }
 
 #ifdef QCA_LL_TX_FLOW_CONTROL_V2
@@ -4055,6 +4096,8 @@ static inline void dp_tx_sojourn_stats_process(struct dp_pdev *pdev,
 {
 	uint64_t delta_ms;
 	struct cdp_tx_sojourn_stats *sojourn_stats;
+	struct dp_peer *primary_link_peer = NULL;
+	struct dp_soc *link_peer_soc = NULL;
 
 	if (qdf_unlikely(!pdev->enhanced_stats_en))
 		return;
@@ -4066,29 +4109,41 @@ static inline void dp_tx_sojourn_stats_process(struct dp_pdev *pdev,
 	if (qdf_unlikely(!pdev->sojourn_buf))
 		return;
 
+	primary_link_peer = dp_get_primary_link_peer_by_id(pdev->soc,
+							   txrx_peer->peer_id,
+							   DP_MOD_ID_TX_COMP);
+
+	if (qdf_unlikely(!primary_link_peer))
+		return;
+
 	sojourn_stats = (struct cdp_tx_sojourn_stats *)
 		qdf_nbuf_data(pdev->sojourn_buf);
 
-	sojourn_stats->cookie = (void *)peer->rdkstats_ctx;
+	link_peer_soc = primary_link_peer->vdev->pdev->soc;
+	sojourn_stats->cookie = (void *)
+			dp_monitor_peer_get_rdkstats_ctx(link_peer_soc,
+							 primary_link_peer);
 
 	delta_ms = qdf_ktime_to_ms(qdf_ktime_real_get()) -
 				txdesc_ts;
-	qdf_ewma_tx_lag_add(&peer->avg_sojourn_msdu[tid],
+	qdf_ewma_tx_lag_add(&txrx_peer->stats.per_pkt_stats.tx.avg_sojourn_msdu[tid],
 			    delta_ms);
 	sojourn_stats->sum_sojourn_msdu[tid] = delta_ms;
 	sojourn_stats->num_msdus[tid] = 1;
 	sojourn_stats->avg_sojourn_msdu[tid].internal =
-		peer->avg_sojourn_msdu[tid].internal;
+		txrx_peer->stats.per_pkt_stats.tx.avg_sojourn_msdu[tid].internal;
 	dp_wdi_event_handler(WDI_EVENT_TX_SOJOURN_STAT, pdev->soc,
 			     pdev->sojourn_buf, HTT_INVALID_PEER,
 			     WDI_NO_VAL, pdev->pdev_id);
 	sojourn_stats->sum_sojourn_msdu[tid] = 0;
 	sojourn_stats->num_msdus[tid] = 0;
 	sojourn_stats->avg_sojourn_msdu[tid].internal = 0;
+
+	dp_peer_unref_delete(primary_link_peer, DP_MOD_ID_TX_COMP);
 }
 #else
 static inline void dp_tx_sojourn_stats_process(struct dp_pdev *pdev,
-					       struct dp_peer *peer,
+					       struct dp_txrx_peer *txrx_peer,
 					       uint8_t tid,
 					       uint64_t txdesc_ts,
 					       uint32_t ppdu_id)
@@ -4445,22 +4500,26 @@ void dp_tx_comp_process_tx_status(struct dp_soc *soc,
 	if (qdf_unlikely(txrx_peer->bss_peer &&
 			 vdev->opmode == wlan_op_mode_ap)) {
 		if (ts->status != HAL_TX_TQM_RR_REM_CMD_REM) {
-			DP_STATS_INC_PKT(peer, tx.mcast, 1, length);
+			DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, tx.mcast, 1,
+						      length);
 
 			if (txrx_peer->vdev->tx_encap_type ==
 				htt_cmn_pkt_type_ethernet &&
 				QDF_IS_ADDR_BROADCAST(eh->ether_dhost)) {
-				DP_STATS_INC_PKT(peer, tx.bcast, 1, length);
+				DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+							      tx.bcast, 1,
+							      length);
 			}
 		}
 	} else {
-		DP_STATS_INC_PKT(peer, tx.ucast, 1, length);
+		DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, tx.ucast, 1, length);
 		if (ts->status == HAL_TX_TQM_RR_FRAME_ACKED) {
-			DP_STATS_INC_PKT(peer, tx.tx_success, 1, length);
+			DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer, tx.tx_success,
+						      1, length);
 			if (qdf_unlikely(txrx_peer->in_twt)) {
-				DP_STATS_INC_PKT(peer,
-						 tx.tx_success_twt,
-						 1, length);
+				DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+							      tx.tx_success_twt,
+							      1, length);
 			}
 		}
 	}
@@ -4483,7 +4542,7 @@ out:
 	defined(QCA_ENHANCED_STATS_SUPPORT)
 /*
  * dp_tx_update_peer_basic_stats(): Update peer basic stats
- * @peer: Datapath peer handle
+ * @txrx_peer: Datapath txrx_peer handle
  * @length: Length of the packet
  * @tx_status: Tx status from TQM/FW
  * @update: enhanced flag value present in dp_pdev
@@ -4495,9 +4554,10 @@ void dp_tx_update_peer_basic_stats(struct dp_txrx_peer *txrx_peer,
 				   bool update)
 {
 	if ((!txrx_peer->hw_txrx_stats_en) || update) {
-		DP_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, length);
-		DP_STATS_FLAT_INC(txrx_peer, tx_failed, 1,
-				  tx_status != HAL_TX_TQM_RR_FRAME_ACKED);
+		DP_PEER_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, length);
+
+		if (tx_status != HAL_TX_TQM_RR_FRAME_ACKED)
+			DP_PEER_STATS_FLAT_INC(txrx_peer, tx_failed, 1);
 	}
 }
 #elif defined(QCA_VDEV_STATS_HW_OFFLOAD_SUPPORT)
@@ -4506,9 +4566,10 @@ void dp_tx_update_peer_basic_stats(struct dp_txrx_peer *txrx_peer,
 				   bool update)
 {
 	if (!peer->hw_txrx_stats_en) {
-		DP_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, length);
-		DP_STATS_FLAT_INC(txrx_peer, tx_failed, 1,
-				  tx_status != HAL_TX_TQM_RR_FRAME_ACKED);
+		DP_PEER_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, length);
+
+		if (tx_status != HAL_TX_TQM_RR_FRAME_ACKED)
+			DP_PEER_STATS_FLAT_INC(txrx_peer, tx_failed, 1);
 	}
 }
 
@@ -4517,9 +4578,10 @@ void dp_tx_update_peer_basic_stats(struct dp_txrx_peer *txrx_peer,
 				   uint32_t length, uint8_t tx_status,
 				   bool update)
 {
-	DP_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, length);
-	DP_STATS_FLAT_INC(txrx_peer, tx_failed, 1,
-			  tx_status != HAL_TX_TQM_RR_FRAME_ACKED);
+	DP_PEER_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1, length);
+
+	if (tx_status != HAL_TX_TQM_RR_FRAME_ACKED)
+		DP_PEER_STATS_FLAT_INC(txrx_peer, tx_failed, 1);
 }
 #endif
 

+ 21 - 12
dp/wifi3.0/li/dp_li_rx.c

@@ -740,7 +740,9 @@ done:
 
 			if (qdf_nbuf_is_raw_frame(nbuf)) {
 				DP_STATS_INC(vdev->pdev, rx_raw_pkts, 1);
-				DP_STATS_INC_PKT(peer, rx.raw, 1, msdu_len);
+				DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+							      rx.raw, 1,
+							      msdu_len);
 			} else {
 				dp_rx_nbuf_free(nbuf);
 				DP_STATS_INC(soc, rx.err.scatter_msdu, 1);
@@ -765,7 +767,9 @@ done:
 		if (qdf_unlikely(vdev->multipass_en)) {
 			if (dp_rx_multipass_process(txrx_peer, nbuf,
 						    tid) == false) {
-				DP_STATS_INC(peer, rx.multipass_rx_pkt_drop, 1);
+				DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+							  rx.multipass_rx_pkt_drop,
+							  1);
 				dp_rx_nbuf_free(nbuf);
 				nbuf = next;
 				continue;
@@ -774,7 +778,8 @@ done:
 
 		if (!dp_wds_rx_policy_check(rx_tlv_hdr, vdev, txrx_peer)) {
 			dp_rx_err("%pK: Policy Check Drop pkt", soc);
-			DP_STATS_INC(peer, rx.policy_check_drop, 1);
+			DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+						  rx.policy_check_drop, 1);
 			tid_stats->fail_cnt[POLICY_CHECK_DROP]++;
 			/* Drop & free packet */
 			dp_rx_nbuf_free(nbuf);
@@ -789,7 +794,8 @@ done:
 								rx_tlv_hdr) ==
 				  false))) {
 			tid_stats->fail_cnt[NAWDS_MCAST_DROP]++;
-			DP_STATS_INC(peer, rx.nawds_mcast_drop, 1);
+			DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+						  rx.nawds_mcast_drop, 1);
 			dp_rx_nbuf_free(nbuf);
 			nbuf = next;
 			continue;
@@ -805,8 +811,9 @@ done:
 					qdf_nbuf_is_ipv4_wapi_pkt(nbuf);
 
 			if (!is_eapol) {
-				DP_STATS_INC(peer,
-					     rx.peer_unauth_rx_pkt_drop, 1);
+				DP_PEER_PER_PKT_STATS_INC(txrx_peer,
+							  rx.peer_unauth_rx_pkt_drop,
+							  1);
 				dp_rx_nbuf_free(nbuf);
 				nbuf = next;
 				continue;
@@ -867,8 +874,9 @@ done:
 								 rx_tlv_hdr,
 								 nbuf))) {
 				/* this is a looped back MCBC pkt,drop it */
-				DP_STATS_INC_PKT(peer, rx.mec_drop, 1,
-						 QDF_NBUF_CB_RX_PKT_LEN(nbuf));
+				DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+							      rx.mec_drop, 1,
+							      QDF_NBUF_CB_RX_PKT_LEN(nbuf));
 				dp_rx_nbuf_free(nbuf);
 				nbuf = next;
 				continue;
@@ -904,11 +912,12 @@ done:
 		DP_RX_LIST_APPEND(deliver_list_head,
 				  deliver_list_tail,
 				  nbuf);
-		DP_STATS_FLAT_INC_PKT(txrx_peer, to_stack, 1,
-				      QDF_NBUF_CB_RX_PKT_LEN(nbuf));
+		DP_PEER_STATS_FLAT_INC_PKT(txrx_peer, to_stack, 1,
+					   QDF_NBUF_CB_RX_PKT_LEN(nbuf));
 		if (qdf_unlikely(txrx_peer->in_twt))
-			DP_STATS_INC_PKT(peer, rx.to_stack_twt, 1,
-					 QDF_NBUF_CB_RX_PKT_LEN(nbuf));
+			DP_PEER_PER_PKT_STATS_INC_PKT(txrx_peer,
+						      rx.to_stack_twt, 1,
+						      QDF_NBUF_CB_RX_PKT_LEN(nbuf));
 
 		tid_stats->delivered_to_stack++;
 		nbuf = next;

+ 4 - 4
dp/wifi3.0/li/dp_li_tx.c

@@ -193,10 +193,10 @@ void dp_tx_process_htt_completion_li(struct dp_soc *soc,
 						       &txrx_ref_handle,
 						       DP_MOD_ID_HTT_COMP);
 		if (qdf_likely(txrx_peer)) {
-			DP_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1,
-					      qdf_nbuf_len(tx_desc->nbuf));
-			DP_STATS_FLAT_INC(txrx_peer, tx_failed, 1,
-					  tx_status != HTT_TX_FW2WBM_TX_STATUS_OK);
+			DP_PEER_STATS_FLAT_INC_PKT(txrx_peer, comp_pkt, 1,
+						   qdf_nbuf_len(tx_desc->nbuf));
+			if (tx_status != HTT_TX_FW2WBM_TX_STATUS_OK)
+				DP_PEER_STATS_FLAT_INC(txrx_peer, tx_failed, 1);
 		}
 
 		dp_tx_comp_process_tx_status(soc, tx_desc, &ts, txrx_peer,