Browse Source

qcacmn: Add support for Hash based steering in RX PATH

Add support for hash based steering in RX path, also
considered cases where a particular radio or both radios
are handled by NSS offload.
CRs-Fixed: 2092357

Change-Id: Ib0e88c28eecd7bfdb52c7337d4485ac41371be68
Tallapragada Kalyan 7 years ago
parent
commit
dbbb0c8085
5 changed files with 414 additions and 312 deletions
  1. 98 4
      dp/wifi3.0/dp_main.c
  2. 295 270
      dp/wifi3.0/dp_rx.c
  3. 5 0
      hal/wifi3.0/hal_api.h
  4. 15 37
      hal/wifi3.0/hal_rx.c
  5. 1 1
      wlan_cfg/wlan_cfg.c

+ 98 - 4
dp/wifi3.0/dp_main.c

@@ -1654,6 +1654,81 @@ static void dp_soc_reset_cpu_ring_map(struct dp_soc *soc)
 	}
 }
 
+#ifdef IPA_OFFLOAD
+/**
+ * dp_reo_remap_config() - configure reo remap register value based
+ *                         nss configuration.
+ *		based on offload_radio value below remap configuration
+ *		get applied.
+ *		0 - both Radios handled by host (remap rings 1, 2, 3 & 4)
+ *		1 - 1st Radio handled by NSS (remap rings 2, 3 & 4)
+ *		2 - 2nd Radio handled by NSS (remap rings 1, 2 & 4)
+ *		3 - both Radios handled by NSS (remap not required)
+ *		4 - IPA OFFLOAD enabled (remap rings 1,2 & 3)
+ *
+ * @remap1: output parameter indicates reo remap 1 register value
+ * @remap2: output parameter indicates reo remap 2 register value
+ * Return: bool type, true if remap is configured else false.
+ */
+static bool dp_reo_remap_config(struct dp_soc *soc,
+				uint32_t *remap1,
+				uint32_t *remap2)
+{
+
+	*remap1 = ((0x1 << 0) | (0x2 << 3) | (0x3 << 6) | (0x1 << 9) |
+		(0x2 << 12) | (0x3 << 15) | (0x1 << 18) | (0x2 << 21)) << 8;
+
+	*remap2 = ((0x3 << 0) | (0x1 << 3) | (0x2 << 6) | (0x3 << 9) |
+		(0x1 << 12) | (0x2 << 15) | (0x3 << 18) | (0x1 << 21)) << 8;
+
+	return true;
+}
+#else
+static bool dp_reo_remap_config(struct dp_soc *soc,
+				uint32_t *remap1,
+				uint32_t *remap2)
+{
+	uint8_t offload_radio = wlan_cfg_get_dp_soc_nss_cfg(soc->wlan_cfg_ctx);
+
+	switch (offload_radio) {
+	case 0:
+		*remap1 = ((0x1 << 0) | (0x2 << 3) | (0x3 << 6) |
+			(0x4 << 9) | (0x1 << 12) | (0x2 << 15) |
+			(0x3 << 18) | (0x4 << 21)) << 8;
+
+		*remap2 = ((0x1 << 0) | (0x2 << 3) | (0x3 << 6) |
+			(0x4 << 9) | (0x1 << 12) | (0x2 << 15) |
+			(0x3 << 18) | (0x4 << 21)) << 8;
+		break;
+
+	case 1:
+		*remap1 = ((0x2 << 0) | (0x3 << 3) | (0x4 << 6) |
+			(0x2 << 9) | (0x3 << 12) | (0x4 << 15) |
+			(0x2 << 18) | (0x3 << 21)) << 8;
+
+		*remap2 = ((0x4 << 0) | (0x2 << 3) | (0x3 << 6) |
+			(0x4 << 9) | (0x2 << 12) | (0x3 << 15) |
+			(0x4 << 18) | (0x2 << 21)) << 8;
+		break;
+
+	case 2:
+		*remap1 = ((0x1 << 0) | (0x3 << 3) | (0x4 << 6) |
+			(0x1 << 9) | (0x3 << 12) | (0x4 << 15) |
+			(0x1 << 18) | (0x3 << 21)) << 8;
+
+		*remap2 = ((0x4 << 0) | (0x1 << 3) | (0x3 << 6) |
+			(0x4 << 9) | (0x1 << 12) | (0x3 << 15) |
+			(0x4 << 18) | (0x1 << 21)) << 8;
+		break;
+
+	case 3:
+		/* return false if both radios are offloaded to NSS */
+		return false;
+	}
+	return true;
+}
+#endif
+
 /*
  * dp_soc_cmn_setup() - Common SoC level initializion
  * @soc:		Datapath SOC handle
@@ -1826,9 +1901,21 @@ static int dp_soc_cmn_setup(struct dp_soc *soc)
 	/* Setup HW REO */
 	qdf_mem_zero(&reo_params, sizeof(reo_params));
 
-	if (wlan_cfg_is_rx_hash_enabled(soc->wlan_cfg_ctx))
+	if (wlan_cfg_is_rx_hash_enabled(soc->wlan_cfg_ctx)) {
+
+		/*
+		 * Reo ring remap is not required if both radios
+		 * are offloaded to NSS
+		 */
+		if (!dp_reo_remap_config(soc,
+					&reo_params.remap1,
+					&reo_params.remap2))
+			goto out;
+
 		reo_params.rx_hash_enabled = true;
+	}
 
+out:
 	hal_reo_setup(soc->hal_soc, &reo_params);
 
 	qdf_atomic_set(&soc->cmn_init_done, 1);
@@ -2998,9 +3085,16 @@ static void dp_peer_setup_wifi3(struct cdp_vdev *vdev_hdl, void *peer_hdl)
 	peer->last_disassoc_rcvd = 0;
 	peer->last_deauth_rcvd = 0;
 
-	hash_based = wlan_cfg_is_rx_hash_enabled(soc->wlan_cfg_ctx);
-	QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_ERROR,
-		FL("hash based steering %d\n"), hash_based);
+	/*
+	 * hash based steering is disabled for Radios which are offloaded
+	 * to NSS
+	 */
+	if (!wlan_cfg_get_dp_pdev_nss_enabled(pdev->wlan_cfg_ctx))
+		hash_based = wlan_cfg_is_rx_hash_enabled(soc->wlan_cfg_ctx);
+
+	QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_INFO,
+		FL("hash based steering for pdev: %d is %d\n"),
+		pdev->pdev_id, hash_based);
 
 	if (!hash_based)
 		reo_dest = pdev->reo_dest;

+ 295 - 270
dp/wifi3.0/dp_rx.c

@@ -836,6 +836,33 @@ void dp_rx_sg_create(qdf_nbuf_t nbuf, uint8_t *rx_tlv_hdr,
 	}
 }
 
+static inline void dp_rx_deliver_to_stack(struct dp_vdev *vdev,
+						struct dp_peer *peer,
+						qdf_nbuf_t nbuf_list)
+{
+	/*
+	 * highly unlikely to have a vdev without a registerd rx
+	 * callback function. if so let us free the nbuf_list.
+	 */
+	if (qdf_unlikely(!vdev->osif_rx)) {
+		qdf_nbuf_t nbuf;
+		do {
+			nbuf = nbuf_list;
+			nbuf_list = nbuf_list->next;
+			qdf_nbuf_free(nbuf);
+		} while (nbuf_list);
+
+		return;
+	}
+
+	if (qdf_unlikely(vdev->rx_decap_type == htt_cmn_pkt_type_raw) ||
+			(vdev->rx_decap_type == htt_cmn_pkt_type_native_wifi))
+		dp_rx_deliver_raw(vdev, nbuf_list, peer);
+	else
+		vdev->osif_rx(vdev->osif_vdev, nbuf_list);
+
+}
+
 /**
  * dp_rx_process() - Brain of the Rx processing functionality
  *		     Called from the bottom half (tasklet/NET_RX_SOFTIRQ)
@@ -854,7 +881,7 @@ dp_rx_process(struct dp_intr *int_ctx, void *hal_ring, uint32_t quota)
 	void *hal_soc;
 	void *ring_desc;
 	struct dp_rx_desc *rx_desc = NULL;
-	qdf_nbuf_t nbuf;
+	qdf_nbuf_t nbuf, next;
 	union dp_rx_desc_list_elem_t *head[MAX_PDEV_CNT] = { NULL };
 	union dp_rx_desc_list_elem_t *tail[MAX_PDEV_CNT] = { NULL };
 	uint32_t rx_bufs_used = 0, rx_buf_cookie, l2_hdr_offset;
@@ -862,7 +889,6 @@ dp_rx_process(struct dp_intr *int_ctx, void *hal_ring, uint32_t quota)
 	uint16_t peer_id;
 	struct dp_peer *peer = NULL;
 	struct dp_vdev *vdev = NULL;
-	struct dp_vdev *vdev_list[WLAN_UMAC_PSOC_MAX_VDEVS] = { NULL };
 	uint32_t pkt_len;
 	struct hal_rx_mpdu_desc_info mpdu_desc_info;
 	struct hal_rx_msdu_desc_info msdu_desc_info;
@@ -871,11 +897,8 @@ dp_rx_process(struct dp_intr *int_ctx, void *hal_ring, uint32_t quota)
 	uint8_t *rx_tlv_hdr;
 	uint32_t rx_bufs_reaped[MAX_PDEV_CNT] = { 0 };
 	uint32_t sgi, mcs, tid, nss, bw, reception_type, pkt_type;
-	uint64_t vdev_map = 0;
 	uint8_t mac_id;
-	uint16_t i, vdev_cnt = 0;
 	uint32_t ampdu_flag, amsdu_flag;
-	struct ether_header *eh;
 	struct dp_pdev *pdev;
 	struct dp_srng *dp_rxdma_srng;
 	struct rx_desc_pool *rx_desc_pool;
@@ -888,6 +911,10 @@ dp_rx_process(struct dp_intr *int_ctx, void *hal_ring, uint32_t quota)
 	qdf_nbuf_t frag_list_head = NULL;
 	qdf_nbuf_t frag_list_tail = NULL;
 	uint16_t frag_list_len = 0;
+	qdf_nbuf_t nbuf_head = NULL;
+	qdf_nbuf_t nbuf_tail = NULL;
+	qdf_nbuf_t deliver_list_head = NULL;
+	qdf_nbuf_t deliver_list_tail = NULL;
 
 	DP_HIST_INIT();
 	/* Debug -- Remove later */
@@ -971,12 +998,6 @@ dp_rx_process(struct dp_intr *int_ctx, void *hal_ring, uint32_t quota)
 
 		}
 
-		if (!((vdev_map >> vdev->vdev_id) & 1)) {
-			vdev_map |= 1 << vdev->vdev_id;
-			vdev_list[vdev_cnt] = vdev;
-			vdev_cnt++;
-		}
-
 		/* Get MSDU DESC info */
 		hal_rx_msdu_desc_info_get(ring_desc, &msdu_desc_info);
 
@@ -1014,7 +1035,7 @@ dp_rx_process(struct dp_intr *int_ctx, void *hal_ring, uint32_t quota)
 				!(amsdu_flag));
 
 		DP_HIST_PACKET_COUNT_INC(vdev->pdev->pdev_id);
-		qdf_nbuf_queue_add(&vdev->rxq, rx_desc->nbuf);
+		DP_RX_LIST_APPEND(nbuf_head, nbuf_tail, rx_desc->nbuf);
 fail:
 		/*
 		 * if continuation bit is set then we have MSDU spread
@@ -1053,294 +1074,298 @@ done:
 					HAL_RX_BUF_RBM_SW3_BM);
 	}
 
-	for (i = 0; i < vdev_cnt; i++) {
-		qdf_nbuf_t deliver_list_head = NULL;
-		qdf_nbuf_t deliver_list_tail = NULL;
-
-		vdev = vdev_list[i];
-		while ((nbuf = qdf_nbuf_queue_remove(&vdev->rxq))) {
-			rx_tlv_hdr = qdf_nbuf_data(nbuf);
-			eh = (struct ether_header *)qdf_nbuf_data(nbuf);
+	vdev = NULL;
+	nbuf = nbuf_head;
+	while (nbuf) {
+		next = nbuf->next;
+		rx_tlv_hdr = qdf_nbuf_data(nbuf);
 
-			/*
-			 * Check if DMA completed -- msdu_done is the last bit
-			 * to be written
-			 */
-			if (!hal_rx_attn_msdu_done_get(rx_tlv_hdr)) {
+		peer_mdata = hal_rx_mpdu_peer_meta_data_get(rx_tlv_hdr);
+		peer_id = DP_PEER_METADATA_PEER_ID_GET(peer_mdata);
+		peer = dp_peer_find_by_id(soc, peer_id);
 
-				QDF_TRACE(QDF_MODULE_ID_DP,
-						QDF_TRACE_LEVEL_ERROR,
-						FL("MSDU DONE failure"));
-				DP_STATS_INC(vdev->pdev, dropped.msdu_not_done,
-						1);
-
-				hal_rx_dump_pkt_tlvs(rx_tlv_hdr,
-							QDF_TRACE_LEVEL_INFO);
-				qdf_assert(0);
-			}
+		if (deliver_list_head && (vdev != peer->vdev)) {
+			dp_rx_deliver_to_stack(vdev, peer, deliver_list_head);
+			deliver_list_head = NULL;
+			deliver_list_tail = NULL;
+		}
+		vdev = peer->vdev;
 
-			/*
-			 * The below condition happens when an MSDU is spread
-			 * across multiple buffers. This can happen in two cases
-			 * 1. The nbuf size is smaller then the received msdu.
-			 *    ex: we have set the nbuf size to 2048 during
-			 *        nbuf_alloc. but we received an msdu which is
-			 *        2304 bytes in size then this msdu is spread
-			 *        across 2 nbufs.
-			 *
-			 * 2. AMSDUs when RAW mode is enabled.
-			 *    ex: 1st MSDU is in 1st nbuf and 2nd MSDU is spread
-			 *        across 1st nbuf and 2nd nbuf and last MSDU is
-			 *        spread across 2nd nbuf and 3rd nbuf.
-			 *
-			 * for these scenarios let us create a skb frag_list and
-			 * append these buffers till the last MSDU of the AMSDU
-			 */
-			if (qdf_unlikely(vdev->rx_decap_type ==
-					htt_cmn_pkt_type_raw)) {
-
-				dp_rx_sg_create(nbuf, rx_tlv_hdr, &mpdu_len,
-						&is_first_frag, &frag_list_len,
-						&head_frag_nbuf,
-						&frag_list_head,
-						&frag_list_tail);
-
-				if (is_first_frag)
-					continue;
-				else {
-					frag_list_head = NULL;
-					frag_list_tail = NULL;
-					nbuf = head_frag_nbuf;
-					rx_tlv_hdr = qdf_nbuf_data(nbuf);
-				}
-			}
+		/*
+		 * Check if DMA completed -- msdu_done is the last bit
+		 * to be written
+		 */
+		if (!hal_rx_attn_msdu_done_get(rx_tlv_hdr)) {
 
-			peer_mdata = hal_rx_mpdu_peer_meta_data_get(rx_tlv_hdr);
-			peer_id = DP_PEER_METADATA_PEER_ID_GET(peer_mdata);
-			peer = dp_peer_find_by_id(soc, peer_id);
+			QDF_TRACE(QDF_MODULE_ID_DP,
+					QDF_TRACE_LEVEL_ERROR,
+					FL("MSDU DONE failure"));
+			DP_STATS_INC(vdev->pdev, dropped.msdu_not_done,
+					1);
+			hal_rx_dump_pkt_tlvs(rx_tlv_hdr,
+						QDF_TRACE_LEVEL_INFO);
+			qdf_assert(0);
+		}
 
-			/*
-			 * This is a redundant sanity check, Ideally peer
-			 * should never be NULL here. if for any reason it
-			 * is NULL we will assert.
-			 * Do nothing for LFR case.
-			 */
-			dp_rx_peer_validity_check(peer);
+		/*
+		 * The below condition happens when an MSDU is spread
+		 * across multiple buffers. This can happen in two cases
+		 * 1. The nbuf size is smaller then the received msdu.
+		 *    ex: we have set the nbuf size to 2048 during
+		 *        nbuf_alloc. but we received an msdu which is
+		 *        2304 bytes in size then this msdu is spread
+		 *        across 2 nbufs.
+		 *
+		 * 2. AMSDUs when RAW mode is enabled.
+		 *    ex: 1st MSDU is in 1st nbuf and 2nd MSDU is spread
+		 *        across 1st nbuf and 2nd nbuf and last MSDU is
+		 *        spread across 2nd nbuf and 3rd nbuf.
+		 *
+		 * for these scenarios let us create a skb frag_list and
+		 * append these buffers till the last MSDU of the AMSDU
+		 */
+		if (qdf_unlikely(vdev->rx_decap_type ==
+				htt_cmn_pkt_type_raw)) {
 
-			if (qdf_unlikely(peer->bss_peer)) {
-				QDF_TRACE(QDF_MODULE_ID_DP,
-					QDF_TRACE_LEVEL_INFO,
-					FL("received pkt with same src MAC"));
-				DP_STATS_INC(vdev->pdev, dropped.mec, 1);
+			dp_rx_sg_create(nbuf, rx_tlv_hdr, &mpdu_len,
+					&is_first_frag, &frag_list_len,
+					&head_frag_nbuf,
+					&frag_list_head,
+					&frag_list_tail);
 
-				/* Drop & free packet */
-				qdf_nbuf_free(nbuf);
-				/* Statistics */
+			if (is_first_frag) {
+				nbuf = next;
 				continue;
+			} else {
+				frag_list_head = NULL;
+				frag_list_tail = NULL;
+				nbuf = head_frag_nbuf;
+				rx_tlv_hdr = qdf_nbuf_data(nbuf);
 			}
+		}
 
-			pdev = vdev->pdev;
-			if (qdf_likely(
-				!hal_rx_attn_tcp_udp_cksum_fail_get(rx_tlv_hdr)
-				&&
-				!hal_rx_attn_ip_cksum_fail_get(rx_tlv_hdr))) {
-				qdf_nbuf_rx_cksum_t cksum = {0};
+		/*
+		 * This is a redundant sanity check, Ideally peer
+		 * should never be NULL here. if for any reason it
+		 * is NULL we will assert.
+		 * Do nothing for LFR case.
+		 */
+		dp_rx_peer_validity_check(peer);
 
-				cksum.l4_result =
-					QDF_NBUF_RX_CKSUM_TCP_UDP_UNNECESSARY;
+		if (qdf_unlikely(peer->bss_peer)) {
+			QDF_TRACE(QDF_MODULE_ID_DP,
+				QDF_TRACE_LEVEL_ERROR,
+				FL("received pkt with same src MAC"));
+			DP_STATS_INC(vdev->pdev, dropped.mec, 1);
 
-				qdf_nbuf_set_rx_cksum(nbuf, &cksum);
-			}
+			/* Drop & free packet */
+			qdf_nbuf_free(nbuf);
+			/* Statistics */
+			nbuf = next;
+			continue;
+		}
 
-			sgi = hal_rx_msdu_start_sgi_get(rx_tlv_hdr);
-			mcs = hal_rx_msdu_start_rate_mcs_get(rx_tlv_hdr);
-			tid = hal_rx_mpdu_start_tid_get(rx_tlv_hdr);
-
-			QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_INFO,
-				"%s: %d, SGI: %d, tid: %d",
-				__func__, __LINE__, sgi, tid);
-
-			bw = hal_rx_msdu_start_bw_get(rx_tlv_hdr);
-			reception_type = hal_rx_msdu_start_reception_type_get(
-					rx_tlv_hdr);
-			nss = hal_rx_msdu_start_nss_get(rx_tlv_hdr);
-			pkt_type = hal_rx_msdu_start_get_pkt_type(rx_tlv_hdr);
-
-			DP_STATS_INC(vdev->pdev, rx.bw[bw], 1);
-			DP_STATS_INC(vdev->pdev,
-					rx.reception_type[reception_type], 1);
-			DP_STATS_INCC(vdev->pdev, rx.nss[nss], 1,
-					((reception_type == REPT_MU_MIMO) ||
-					 (reception_type == REPT_MU_OFDMA_MIMO))
-					);
-			DP_STATS_INC(peer, rx.sgi_count[sgi], 1);
-			DP_STATS_INCC(peer, rx.err.mic_err, 1,
-					hal_rx_mpdu_end_mic_err_get(
-						rx_tlv_hdr));
-			DP_STATS_INCC(peer, rx.err.decrypt_err, 1,
-					hal_rx_mpdu_end_decrypt_err_get(
-						rx_tlv_hdr));
-
-			DP_STATS_INC(peer, rx.wme_ac_type[TID_TO_WME_AC(tid)],
-					1);
-			DP_STATS_INC(peer, rx.bw[bw], 1);
-			DP_STATS_INC(peer, rx.reception_type[reception_type],
-					1);
+		pdev = vdev->pdev;
+		if (qdf_likely(
+			!hal_rx_attn_tcp_udp_cksum_fail_get(rx_tlv_hdr)
+			&&
+			!hal_rx_attn_ip_cksum_fail_get(rx_tlv_hdr))) {
+			qdf_nbuf_rx_cksum_t cksum = {0};
 
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[MAX_MCS], 1,
-					((mcs >= MAX_MCS_11A) && (pkt_type
-						== DOT11_A)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[mcs], 1,
-					((mcs <= MAX_MCS_11A) && (pkt_type
-						== DOT11_A)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[MAX_MCS], 1,
-					((mcs >= MAX_MCS_11B)
-					 && (pkt_type == DOT11_B)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[mcs], 1,
-					((mcs <= MAX_MCS_11B)
-					 && (pkt_type == DOT11_B)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[MAX_MCS], 1,
-					((mcs >= MAX_MCS_11A)
-					 && (pkt_type == DOT11_N)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[mcs], 1,
-					((mcs <= MAX_MCS_11A)
-					 && (pkt_type == DOT11_N)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[MAX_MCS], 1,
-					((mcs >= MAX_MCS_11AC)
-					 && (pkt_type == DOT11_AC)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[mcs], 1,
-					((mcs <= MAX_MCS_11AC)
-					 && (pkt_type == DOT11_AC)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[MAX_MCS], 1,
-					((mcs >= (MAX_MCS-1))
-					 && (pkt_type == DOT11_AX)));
-			DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
-					mcs_count[mcs], 1,
-					((mcs <= (MAX_MCS-1))
-					 && (pkt_type == DOT11_AX)));
-
-			/*
-			 * HW structures call this L3 header padding --
-			 * even though this is actually the offset from
-			 * the buffer beginning where the L2 header
-			 * begins.
-			 */
-			QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_INFO,
-				FL("rxhash: flow id toeplitz: 0x%x\n"),
-				hal_rx_msdu_start_toeplitz_get(rx_tlv_hdr));
-
-			l2_hdr_offset =
-				hal_rx_msdu_end_l3_hdr_padding_get(rx_tlv_hdr);
-
-			msdu_len = hal_rx_msdu_start_msdu_len_get(rx_tlv_hdr);
-			pkt_len = msdu_len + l2_hdr_offset + RX_PKT_TLVS_LEN;
-
-			if (unlikely(qdf_nbuf_get_ext_list(nbuf)))
-				qdf_nbuf_pull_head(nbuf, RX_PKT_TLVS_LEN);
-			else {
-				qdf_nbuf_set_pktlen(nbuf, pkt_len);
-				qdf_nbuf_pull_head(nbuf,
-						RX_PKT_TLVS_LEN +
-						l2_hdr_offset);
-			}
+			cksum.l4_result =
+				QDF_NBUF_RX_CKSUM_TCP_UDP_UNNECESSARY;
 
-			if (qdf_unlikely(vdev->mesh_vdev)) {
-				if (dp_rx_filter_mesh_packets(vdev, nbuf,
-								rx_tlv_hdr)
-						== QDF_STATUS_SUCCESS) {
-					QDF_TRACE(QDF_MODULE_ID_DP,
-						QDF_TRACE_LEVEL_INFO_MED,
-						FL("mesh pkt filtered"));
-				DP_STATS_INC(vdev->pdev, dropped.mesh_filter,
-						1);
-
-					qdf_nbuf_free(nbuf);
-					continue;
-				}
-				dp_rx_fill_mesh_stats(vdev, nbuf,
-							rx_tlv_hdr, peer);
+			qdf_nbuf_set_rx_cksum(nbuf, &cksum);
+		}
+
+		sgi = hal_rx_msdu_start_sgi_get(rx_tlv_hdr);
+		mcs = hal_rx_msdu_start_rate_mcs_get(rx_tlv_hdr);
+		tid = hal_rx_mpdu_start_tid_get(rx_tlv_hdr);
+
+		QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_INFO,
+			"%s: %d, SGI: %d, tid: %d",
+			__func__, __LINE__, sgi, tid);
+
+		bw = hal_rx_msdu_start_bw_get(rx_tlv_hdr);
+		reception_type = hal_rx_msdu_start_reception_type_get(
+				rx_tlv_hdr);
+		nss = hal_rx_msdu_start_nss_get(rx_tlv_hdr);
+		pkt_type = hal_rx_msdu_start_get_pkt_type(rx_tlv_hdr);
+
+		DP_STATS_INC(vdev->pdev, rx.bw[bw], 1);
+		DP_STATS_INC(vdev->pdev,
+				rx.reception_type[reception_type], 1);
+		DP_STATS_INCC(vdev->pdev, rx.nss[nss], 1,
+				((reception_type == REPT_MU_MIMO) ||
+				 (reception_type == REPT_MU_OFDMA_MIMO))
+				);
+		DP_STATS_INC(peer, rx.sgi_count[sgi], 1);
+		DP_STATS_INCC(peer, rx.err.mic_err, 1,
+				hal_rx_mpdu_end_mic_err_get(
+					rx_tlv_hdr));
+		DP_STATS_INCC(peer, rx.err.decrypt_err, 1,
+				hal_rx_mpdu_end_decrypt_err_get(
+					rx_tlv_hdr));
+
+		DP_STATS_INC(peer, rx.wme_ac_type[TID_TO_WME_AC(tid)],
+				1);
+		DP_STATS_INC(peer, rx.bw[bw], 1);
+		DP_STATS_INC(peer, rx.reception_type[reception_type],
+				1);
+
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[MAX_MCS], 1,
+				((mcs >= MAX_MCS_11A) && (pkt_type
+					== DOT11_A)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[mcs], 1,
+				((mcs <= MAX_MCS_11A) && (pkt_type
+					== DOT11_A)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[MAX_MCS], 1,
+				((mcs >= MAX_MCS_11B)
+				 && (pkt_type == DOT11_B)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[mcs], 1,
+				((mcs <= MAX_MCS_11B)
+				 && (pkt_type == DOT11_B)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[MAX_MCS], 1,
+				((mcs >= MAX_MCS_11A)
+				 && (pkt_type == DOT11_N)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[mcs], 1,
+				((mcs <= MAX_MCS_11A)
+				 && (pkt_type == DOT11_N)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[MAX_MCS], 1,
+				((mcs >= MAX_MCS_11AC)
+				 && (pkt_type == DOT11_AC)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[mcs], 1,
+				((mcs <= MAX_MCS_11AC)
+				 && (pkt_type == DOT11_AC)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[MAX_MCS], 1,
+				((mcs >= (MAX_MCS-1))
+				 && (pkt_type == DOT11_AX)));
+		DP_STATS_INCC(peer, rx.pkt_type[pkt_type].
+				mcs_count[mcs], 1,
+				((mcs <= (MAX_MCS-1))
+				 && (pkt_type == DOT11_AX)));
+
+		/*
+		 * HW structures call this L3 header padding --
+		 * even though this is actually the offset from
+		 * the buffer beginning where the L2 header
+		 * begins.
+		 */
+		QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_INFO,
+			FL("rxhash: flow id toeplitz: 0x%x\n"),
+			hal_rx_msdu_start_toeplitz_get(rx_tlv_hdr));
+
+		l2_hdr_offset =
+			hal_rx_msdu_end_l3_hdr_padding_get(rx_tlv_hdr);
+
+		msdu_len = hal_rx_msdu_start_msdu_len_get(rx_tlv_hdr);
+		pkt_len = msdu_len + l2_hdr_offset + RX_PKT_TLVS_LEN;
+
+		if (unlikely(qdf_nbuf_get_ext_list(nbuf)))
+			qdf_nbuf_pull_head(nbuf, RX_PKT_TLVS_LEN);
+		else {
+			qdf_nbuf_set_pktlen(nbuf, pkt_len);
+			qdf_nbuf_pull_head(nbuf,
+					RX_PKT_TLVS_LEN +
+					l2_hdr_offset);
+		}
+
+		if (qdf_unlikely(vdev->mesh_vdev)) {
+			if (dp_rx_filter_mesh_packets(vdev, nbuf,
+							rx_tlv_hdr)
+					== QDF_STATUS_SUCCESS) {
+				QDF_TRACE(QDF_MODULE_ID_DP,
+					QDF_TRACE_LEVEL_INFO_MED,
+					FL("mesh pkt filtered"));
+			DP_STATS_INC(vdev->pdev, dropped.mesh_filter,
+					1);
+
+				qdf_nbuf_free(nbuf);
+				nbuf = next;
+				continue;
 			}
+			dp_rx_fill_mesh_stats(vdev, nbuf, rx_tlv_hdr, peer);
+		}
 
 #ifdef QCA_WIFI_NAPIER_EMULATION_DBG /* Debug code, remove later */
-			QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_ERROR,
-				"p_id %d msdu_len %d hdr_off %d",
-				peer_id, msdu_len, l2_hdr_offset);
+		QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_ERROR,
+			"p_id %d msdu_len %d hdr_off %d",
+			peer_id, msdu_len, l2_hdr_offset);
 
-			print_hex_dump(KERN_ERR,
-				       "\t Pkt Data:", DUMP_PREFIX_NONE, 32, 4,
-					qdf_nbuf_data(nbuf), 128, false);
+		print_hex_dump(KERN_ERR,
+			       "\t Pkt Data:", DUMP_PREFIX_NONE, 32, 4,
+				qdf_nbuf_data(nbuf), 128, false);
 #endif /* NAPIER_EMULATION */
 
-			if (qdf_likely(vdev->rx_decap_type ==
-						htt_cmn_pkt_type_ethernet) &&
-					(qdf_likely(!vdev->mesh_vdev))) {
-				/* WDS Source Port Learning */
-				dp_rx_wds_srcport_learn(soc,
-							rx_tlv_hdr,
+		if (qdf_likely(vdev->rx_decap_type ==
+					htt_cmn_pkt_type_ethernet) &&
+				(qdf_likely(!vdev->mesh_vdev))) {
+			/* WDS Source Port Learning */
+			dp_rx_wds_srcport_learn(soc,
+						rx_tlv_hdr,
+						peer,
+						nbuf);
+
+			/* Intrabss-fwd */
+			if ((vdev->opmode != wlan_op_mode_sta) &&
+					!vdev->nawds_enabled)
+				if (dp_rx_intrabss_fwd(soc,
 							peer,
-							nbuf);
-
-				/* Intrabss-fwd */
-				if ((vdev->opmode != wlan_op_mode_sta) &&
-						!vdev->nawds_enabled)
-					if (dp_rx_intrabss_fwd(soc,
-								peer,
-								rx_tlv_hdr,
-								nbuf))
-						continue; /* Get next desc */
-			}
+							rx_tlv_hdr,
+							nbuf)) {
+					nbuf = next;
+					continue; /* Get next desc */
+				}
+		}
 
-			rx_bufs_used++;
+		rx_bufs_used++;
 
-			dp_rx_lro(rx_tlv_hdr, peer, nbuf, int_ctx->lro_ctx);
+		dp_rx_lro(rx_tlv_hdr, peer, nbuf, int_ctx->lro_ctx);
 
-			DP_RX_LIST_APPEND(deliver_list_head,
-						deliver_list_tail,
-						nbuf);
+		DP_RX_LIST_APPEND(deliver_list_head,
+					deliver_list_tail,
+					nbuf);
 
-			DP_STATS_INCC_PKT(peer, rx.multicast, 1, pkt_len,
-					hal_rx_msdu_end_da_is_mcbc_get(
-						rx_tlv_hdr));
-
-			DP_STATS_INC_PKT(peer, rx.to_stack, 1,
-					pkt_len);
-
-			if ((pdev->enhanced_stats_en) && likely(peer) &&
-				hal_rx_attn_first_mpdu_get(rx_tlv_hdr)) {
-				if (soc->cdp_soc.ol_ops->update_dp_stats) {
-					soc->cdp_soc.ol_ops->update_dp_stats(
-							vdev->pdev->osif_pdev,
-							&peer->stats,
-							peer_id,
-							UPDATE_PEER_STATS);
-
-					dp_aggregate_vdev_stats(peer->vdev);
-
-					soc->cdp_soc.ol_ops->update_dp_stats(
-							vdev->pdev->osif_pdev,
-							&peer->vdev->stats,
-							peer->vdev->vdev_id,
-							UPDATE_VDEV_STATS);
-				}
+		DP_STATS_INCC_PKT(peer, rx.multicast, 1, pkt_len,
+				hal_rx_msdu_end_da_is_mcbc_get(
+					rx_tlv_hdr));
+
+		DP_STATS_INC_PKT(peer, rx.to_stack, 1,
+				pkt_len);
+
+		if ((pdev->enhanced_stats_en) && likely(peer) &&
+			hal_rx_attn_first_mpdu_get(rx_tlv_hdr)) {
+			if (soc->cdp_soc.ol_ops->update_dp_stats) {
+				soc->cdp_soc.ol_ops->update_dp_stats(
+						vdev->pdev->osif_pdev,
+						&peer->stats,
+						peer_id,
+						UPDATE_PEER_STATS);
+
+				dp_aggregate_vdev_stats(peer->vdev);
+
+				soc->cdp_soc.ol_ops->update_dp_stats(
+						vdev->pdev->osif_pdev,
+						&peer->vdev->stats,
+						peer->vdev->vdev_id,
+						UPDATE_VDEV_STATS);
 			}
 		}
-
-		if (qdf_unlikely(vdev->rx_decap_type == htt_cmn_pkt_type_raw) ||
-			(vdev->rx_decap_type == htt_cmn_pkt_type_native_wifi))
-			dp_rx_deliver_raw(vdev, deliver_list_head, peer);
-		else if (qdf_likely(vdev->osif_rx) && deliver_list_head)
-			vdev->osif_rx(vdev->osif_vdev, deliver_list_head);
+		nbuf = next;
 	}
 
+	if (deliver_list_head)
+		dp_rx_deliver_to_stack(vdev, peer, deliver_list_head);
+
 	return rx_bufs_used; /* Assume no scale factor for now */
 }
 

+ 5 - 0
hal/wifi3.0/hal_api.h

@@ -906,7 +906,12 @@ extern void hal_setup_link_idle_list(void *hal_soc,
 
 /* REO parameters to be passed to hal_reo_setup */
 struct hal_reo_params {
+	/* rx hash steering enabled or disabled */
 	bool rx_hash_enabled;
+	/* reo remap 1 register */
+	uint32_t remap1;
+	/* reo remap 2 register */
+	uint32_t remap2;
 };
 
 /**

+ 15 - 37
hal/wifi3.0/hal_rx.c

@@ -54,41 +54,6 @@ static inline void hal_uniform_desc_hdr_setup(uint32_t *desc, uint32_t owner,
 #endif
 #define HAL_NON_QOS_TID 16
 
-/**
- * When hash based routing is enabled, routing of the rx packet
- * is done based on the following value: 1 _ _ _ _ The last 4
- * bits are based on hash[3:0]. This means the possible values
- * are 0x10 to 0x1f. This value is used to look-up the
- * ring ID configured in Destination_Ring_Ctrl_IX_* register.
- * The Destination_Ring_Ctrl_IX_2 and Destination_Ring_Ctrl_IX_3
- * registers need to be configured to set-up the 16 entries to
- * map the hash values to a ring number. There are 3 bits per
- * hash entry – which are mapped as follows:
- * 0: TCL, 1:SW1, 2:SW2, * 3:SW3, 4:SW4, 5:Release, 6:FW(WIFI),
- * 7: NOT_USED.
-*/
-#ifdef IPA_OFFLOAD
-/**
- * When IPA is enabled, there will be 3 available rings.
- * Otherwise there will be 4.
- */
-#define REO_REMAP_REGISTER_2 ( \
-	((0x1 << 0) | (0x2 << 3) | (0x3 << 6) | (0x1 << 9) | \
-	(0x2 << 12) | (0x3 << 15) | (0x1 << 18) | (0x2 << 21)) << 8)
-
-#define REO_REMAP_REGISTER_3 ( \
-	((0x3 << 0) | (0x1 << 3) | (0x2 << 6) | (0x3 << 9) | \
-	(0x1 << 12) | (0x2 << 15) | (0x3 << 18) | (0x1 << 21)) << 8)
-#else
-#define REO_REMAP_REGISTER_2 ( \
-	((0x1 << 0) | (0x2 << 3) | (0x3 << 6) | (0x4 << 9) | \
-	(0x1 << 12) | (0x2 << 15) | (0x3 << 18) | (0x4 << 21)) << 8)
-
-#define REO_REMAP_REGISTER_3 ( \
-	((0x1 << 0) | (0x2 << 3) | (0x3 << 6) | (0x4 << 9) | \
-	(0x1 << 12) | (0x2 << 15) | (0x3 << 18) | (0x4 << 21)) << 8)
-#endif
-
 /**
  * hal_reo_qdesc_setup - Setup HW REO queue descriptor
  *
@@ -297,11 +262,24 @@ void hal_reo_setup(void *hal_soc,
 		SEQ_WCSS_UMAC_REO_REG_OFFSET),
 		(HAL_DEFAULT_REO_TIMEOUT_MS * 1000));
 
+	/*
+	 * When hash based routing is enabled, routing of the rx packet
+	 * is done based on the following value: 1 _ _ _ _ The last 4
+	 * bits are based on hash[3:0]. This means the possible values
+	 * are 0x10 to 0x1f. This value is used to look-up the
+	 * ring ID configured in Destination_Ring_Ctrl_IX_* register.
+	 * The Destination_Ring_Ctrl_IX_2 and Destination_Ring_Ctrl_IX_3
+	 * registers need to be configured to set-up the 16 entries to
+	 * map the hash values to a ring number. There are 3 bits per
+	 * hash entry – which are mapped as follows:
+	 * 0: TCL, 1:SW1, 2:SW2, * 3:SW3, 4:SW4, 5:Release, 6:FW(WIFI),
+	 * 7: NOT_USED.
+	*/
 	if (reo_params->rx_hash_enabled) {
 		HAL_REG_WRITE(soc,
 			HWIO_REO_R0_DESTINATION_RING_CTRL_IX_2_ADDR(
 			SEQ_WCSS_UMAC_REO_REG_OFFSET),
-			REO_REMAP_REGISTER_2);
+			reo_params->remap1);
 
 		QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_ERROR,
 			FL("HWIO_REO_R0_DESTINATION_RING_CTRL_IX_2_ADDR 0x%x\n"),
@@ -312,7 +290,7 @@ void hal_reo_setup(void *hal_soc,
 		HAL_REG_WRITE(soc,
 			HWIO_REO_R0_DESTINATION_RING_CTRL_IX_3_ADDR(
 			SEQ_WCSS_UMAC_REO_REG_OFFSET),
-			REO_REMAP_REGISTER_3);
+			reo_params->remap2);
 
 		QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_ERROR,
 			FL("HWIO_REO_R0_DESTINATION_RING_CTRL_IX_3_ADDR 0x%x\n"),

+ 1 - 1
wlan_cfg/wlan_cfg.c

@@ -60,7 +60,7 @@
 #endif
 
 #ifdef CONFIG_WIN
-#define WLAN_CFG_PER_PDEV_RX_RING 1
+#define WLAN_CFG_PER_PDEV_RX_RING 0
 #define NUM_RXDMA_RINGS_PER_PDEV 1
 #define WLAN_LRO_ENABLE 0