
Currently Refill buffer pool enqueue is using POOL_SIZE macro, but there is possibility of dynamically configuring max pool size less than that. So use max_bufq_len which will be configured dynamically during refill pool init instead of POOL_SIZE macro. Change-Id: I610b5dff71861254aa8b6d30e2c562bd7316e120 CRs-Fixed: 3575599
420 lines
11 KiB
C
420 lines
11 KiB
C
/*
|
|
* Copyright (c) 2020-2021 The Linux Foundation. All rights reserved.
|
|
* Copyright (c) 2022-2023 Qualcomm Innovation Center, Inc. All rights reserved.
|
|
*
|
|
* Permission to use, copy, modify, and/or distribute this software for
|
|
* any purpose with or without fee is hereby granted, provided that the
|
|
* above copyright notice and this permission notice appear in all
|
|
* copies.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL
|
|
* WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED
|
|
* WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE
|
|
* AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL
|
|
* DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR
|
|
* PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER
|
|
* TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR
|
|
* PERFORMANCE OF THIS SOFTWARE.
|
|
*/
|
|
|
|
#include "dp_rx_buffer_pool.h"
|
|
#include "dp_ipa.h"
|
|
|
|
#ifndef DP_RX_BUFFER_POOL_SIZE
|
|
#define DP_RX_BUFFER_POOL_SIZE 128
|
|
#endif
|
|
|
|
#ifndef DP_RX_BUFF_POOL_ALLOC_THRES
|
|
#define DP_RX_BUFF_POOL_ALLOC_THRES 1
|
|
#endif
|
|
|
|
#ifdef WLAN_FEATURE_RX_PREALLOC_BUFFER_POOL
|
|
bool dp_rx_buffer_pool_refill(struct dp_soc *soc, qdf_nbuf_t nbuf, u8 mac_id)
|
|
{
|
|
struct dp_pdev *pdev = dp_get_pdev_for_lmac_id(soc, mac_id);
|
|
struct rx_desc_pool *rx_desc_pool = &soc->rx_desc_buf[mac_id];
|
|
struct rx_buff_pool *bufpool = &soc->rx_buff_pool[mac_id];
|
|
qdf_nbuf_t next_nbuf, first_nbuf, refill_nbuf;
|
|
bool consumed = false;
|
|
|
|
if (!bufpool->is_initialized || !pdev)
|
|
return consumed;
|
|
|
|
/* process only buffers of RXDMA ring */
|
|
if (soc->wlan_cfg_ctx->rxdma1_enable)
|
|
return consumed;
|
|
|
|
first_nbuf = nbuf;
|
|
|
|
while (nbuf) {
|
|
next_nbuf = qdf_nbuf_next(nbuf);
|
|
|
|
if (qdf_likely(qdf_nbuf_queue_head_qlen(&bufpool->emerg_nbuf_q) >=
|
|
DP_RX_BUFFER_POOL_SIZE))
|
|
break;
|
|
|
|
refill_nbuf = qdf_nbuf_alloc(soc->osdev, rx_desc_pool->buf_size,
|
|
RX_BUFFER_RESERVATION,
|
|
rx_desc_pool->buf_alignment,
|
|
FALSE);
|
|
|
|
/* Failed to allocate new nbuf, reset and place it back
|
|
* in to the pool.
|
|
*/
|
|
if (!refill_nbuf) {
|
|
DP_STATS_INC(pdev,
|
|
rx_buffer_pool.num_bufs_consumed, 1);
|
|
consumed = true;
|
|
break;
|
|
}
|
|
|
|
/* Successful allocation!! */
|
|
DP_STATS_INC(pdev,
|
|
rx_buffer_pool.num_bufs_alloc_success, 1);
|
|
qdf_nbuf_queue_head_enqueue_tail(&bufpool->emerg_nbuf_q,
|
|
refill_nbuf);
|
|
nbuf = next_nbuf;
|
|
}
|
|
|
|
nbuf = first_nbuf;
|
|
if (consumed) {
|
|
/* Free the MSDU/scattered MSDU */
|
|
while (nbuf) {
|
|
next_nbuf = qdf_nbuf_next(nbuf);
|
|
dp_rx_buffer_pool_nbuf_free(soc, nbuf, mac_id);
|
|
nbuf = next_nbuf;
|
|
}
|
|
}
|
|
|
|
return consumed;
|
|
}
|
|
|
|
void dp_rx_buffer_pool_nbuf_free(struct dp_soc *soc, qdf_nbuf_t nbuf, u8 mac_id)
|
|
{
|
|
struct dp_pdev *dp_pdev = dp_get_pdev_for_lmac_id(soc, mac_id);
|
|
struct rx_desc_pool *rx_desc_pool;
|
|
struct rx_buff_pool *buff_pool;
|
|
|
|
if (!wlan_cfg_per_pdev_lmac_ring(soc->wlan_cfg_ctx))
|
|
mac_id = dp_pdev->lmac_id;
|
|
|
|
rx_desc_pool = &soc->rx_desc_buf[mac_id];
|
|
buff_pool = &soc->rx_buff_pool[mac_id];
|
|
|
|
if (qdf_likely(qdf_nbuf_queue_head_qlen(&buff_pool->emerg_nbuf_q) >=
|
|
DP_RX_BUFFER_POOL_SIZE) ||
|
|
!buff_pool->is_initialized)
|
|
return qdf_nbuf_free(nbuf);
|
|
|
|
qdf_nbuf_reset(nbuf, RX_BUFFER_RESERVATION,
|
|
rx_desc_pool->buf_alignment);
|
|
qdf_nbuf_queue_head_enqueue_tail(&buff_pool->emerg_nbuf_q, nbuf);
|
|
}
|
|
|
|
void dp_rx_refill_buff_pool_enqueue(struct dp_soc *soc)
|
|
{
|
|
struct rx_desc_pool *rx_desc_pool;
|
|
struct rx_refill_buff_pool *buff_pool;
|
|
qdf_device_t dev;
|
|
qdf_nbuf_t nbuf;
|
|
QDF_STATUS ret;
|
|
int count, i;
|
|
uint16_t num_refill;
|
|
uint16_t total_num_refill;
|
|
uint16_t total_count = 0;
|
|
uint16_t head, tail;
|
|
|
|
if (!soc)
|
|
return;
|
|
|
|
dev = soc->osdev;
|
|
buff_pool = &soc->rx_refill_buff_pool;
|
|
rx_desc_pool = &soc->rx_desc_buf[0];
|
|
if (!buff_pool->is_initialized)
|
|
return;
|
|
|
|
head = buff_pool->head;
|
|
tail = buff_pool->tail;
|
|
if (tail > head)
|
|
total_num_refill = (tail - head - 1);
|
|
else
|
|
total_num_refill = (buff_pool->max_bufq_len - head +
|
|
tail - 1);
|
|
|
|
while (total_num_refill) {
|
|
if (total_num_refill > DP_RX_REFILL_BUFF_POOL_BURST)
|
|
num_refill = DP_RX_REFILL_BUFF_POOL_BURST;
|
|
else
|
|
num_refill = total_num_refill;
|
|
|
|
count = 0;
|
|
for (i = 0; i < num_refill; i++) {
|
|
nbuf = qdf_nbuf_alloc(dev, rx_desc_pool->buf_size,
|
|
RX_BUFFER_RESERVATION,
|
|
rx_desc_pool->buf_alignment,
|
|
FALSE);
|
|
if (qdf_unlikely(!nbuf))
|
|
continue;
|
|
|
|
ret = qdf_nbuf_map_nbytes_single(dev, nbuf,
|
|
QDF_DMA_FROM_DEVICE,
|
|
rx_desc_pool->buf_size);
|
|
if (qdf_unlikely(QDF_IS_STATUS_ERROR(ret))) {
|
|
qdf_nbuf_free(nbuf);
|
|
continue;
|
|
}
|
|
|
|
dp_audio_smmu_map(dev,
|
|
qdf_mem_paddr_from_dmaaddr(dev,
|
|
QDF_NBUF_CB_PADDR(nbuf)),
|
|
QDF_NBUF_CB_PADDR(nbuf),
|
|
rx_desc_pool->buf_size);
|
|
|
|
buff_pool->buf_elem[head++] = nbuf;
|
|
head &= (buff_pool->max_bufq_len - 1);
|
|
count++;
|
|
}
|
|
|
|
if (count) {
|
|
buff_pool->head = head;
|
|
total_num_refill -= count;
|
|
total_count += count;
|
|
}
|
|
}
|
|
|
|
DP_STATS_INC(buff_pool->dp_pdev,
|
|
rx_refill_buff_pool.num_bufs_refilled,
|
|
total_count);
|
|
}
|
|
|
|
static inline qdf_nbuf_t dp_rx_refill_buff_pool_dequeue_nbuf(struct dp_soc *soc)
|
|
{
|
|
struct rx_refill_buff_pool *buff_pool = &soc->rx_refill_buff_pool;
|
|
qdf_nbuf_t nbuf = NULL;
|
|
uint16_t head, tail;
|
|
|
|
head = buff_pool->head;
|
|
tail = buff_pool->tail;
|
|
|
|
if (head == tail)
|
|
return NULL;
|
|
|
|
nbuf = buff_pool->buf_elem[tail++];
|
|
tail &= (buff_pool->max_bufq_len - 1);
|
|
buff_pool->tail = tail;
|
|
|
|
return nbuf;
|
|
}
|
|
|
|
qdf_nbuf_t
|
|
dp_rx_buffer_pool_nbuf_alloc(struct dp_soc *soc, uint32_t mac_id,
|
|
struct rx_desc_pool *rx_desc_pool,
|
|
uint32_t num_available_buffers)
|
|
{
|
|
struct dp_pdev *dp_pdev = dp_get_pdev_for_lmac_id(soc, mac_id);
|
|
struct rx_buff_pool *buff_pool;
|
|
struct dp_srng *dp_rxdma_srng;
|
|
qdf_nbuf_t nbuf;
|
|
|
|
nbuf = dp_rx_refill_buff_pool_dequeue_nbuf(soc);
|
|
if (qdf_likely(nbuf)) {
|
|
DP_STATS_INC(dp_pdev,
|
|
rx_refill_buff_pool.num_bufs_allocated, 1);
|
|
return nbuf;
|
|
}
|
|
|
|
if (!wlan_cfg_per_pdev_lmac_ring(soc->wlan_cfg_ctx))
|
|
mac_id = dp_pdev->lmac_id;
|
|
|
|
buff_pool = &soc->rx_buff_pool[mac_id];
|
|
dp_rxdma_srng = &soc->rx_refill_buf_ring[mac_id];
|
|
|
|
nbuf = qdf_nbuf_alloc(soc->osdev, rx_desc_pool->buf_size,
|
|
RX_BUFFER_RESERVATION,
|
|
rx_desc_pool->buf_alignment,
|
|
FALSE);
|
|
|
|
if (!buff_pool->is_initialized)
|
|
return nbuf;
|
|
|
|
if (qdf_likely(nbuf)) {
|
|
buff_pool->nbuf_fail_cnt = 0;
|
|
return nbuf;
|
|
}
|
|
|
|
buff_pool->nbuf_fail_cnt++;
|
|
|
|
/* Allocate buffer from the buffer pool */
|
|
if (buff_pool->nbuf_fail_cnt >= DP_RX_BUFFER_POOL_ALLOC_THRES ||
|
|
(num_available_buffers < dp_rxdma_srng->num_entries / 10)) {
|
|
nbuf = qdf_nbuf_queue_head_dequeue(&buff_pool->emerg_nbuf_q);
|
|
if (nbuf)
|
|
DP_STATS_INC(dp_pdev,
|
|
rx_buffer_pool.num_pool_bufs_replenish, 1);
|
|
}
|
|
|
|
return nbuf;
|
|
}
|
|
|
|
QDF_STATUS
|
|
dp_rx_buffer_pool_nbuf_map(struct dp_soc *soc,
|
|
struct rx_desc_pool *rx_desc_pool,
|
|
struct dp_rx_nbuf_frag_info *nbuf_frag_info_t)
|
|
{
|
|
QDF_STATUS ret = QDF_STATUS_SUCCESS;
|
|
|
|
if (!QDF_NBUF_CB_PADDR((nbuf_frag_info_t->virt_addr).nbuf)) {
|
|
ret = qdf_nbuf_map_nbytes_single(soc->osdev,
|
|
(nbuf_frag_info_t->virt_addr).nbuf,
|
|
QDF_DMA_FROM_DEVICE,
|
|
rx_desc_pool->buf_size);
|
|
if (QDF_IS_STATUS_SUCCESS(ret))
|
|
dp_audio_smmu_map(soc->osdev,
|
|
qdf_mem_paddr_from_dmaaddr(soc->osdev,
|
|
QDF_NBUF_CB_PADDR((nbuf_frag_info_t->virt_addr).nbuf)),
|
|
QDF_NBUF_CB_PADDR((nbuf_frag_info_t->virt_addr).nbuf),
|
|
rx_desc_pool->buf_size);
|
|
}
|
|
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void dp_rx_refill_buff_pool_init(struct dp_soc *soc, u8 mac_id)
|
|
{
|
|
struct rx_desc_pool *rx_desc_pool = &soc->rx_desc_buf[mac_id];
|
|
qdf_nbuf_t nbuf;
|
|
struct rx_refill_buff_pool *buff_pool = &soc->rx_refill_buff_pool;
|
|
QDF_STATUS ret;
|
|
uint16_t head = 0;
|
|
int i;
|
|
|
|
if (!wlan_cfg_is_rx_refill_buffer_pool_enabled(soc->wlan_cfg_ctx)) {
|
|
dp_err("RX refill buffer pool support is disabled");
|
|
buff_pool->is_initialized = false;
|
|
return;
|
|
}
|
|
|
|
buff_pool->max_bufq_len =
|
|
wlan_cfg_get_rx_refill_buf_pool_size(soc->wlan_cfg_ctx);
|
|
buff_pool->dp_pdev = dp_get_pdev_for_lmac_id(soc, 0);
|
|
buff_pool->tail = 0;
|
|
|
|
for (i = 0; i < (buff_pool->max_bufq_len - 1); i++) {
|
|
nbuf = qdf_nbuf_alloc(soc->osdev, rx_desc_pool->buf_size,
|
|
RX_BUFFER_RESERVATION,
|
|
rx_desc_pool->buf_alignment, FALSE);
|
|
if (!nbuf)
|
|
continue;
|
|
|
|
ret = qdf_nbuf_map_nbytes_single(soc->osdev, nbuf,
|
|
QDF_DMA_FROM_DEVICE,
|
|
rx_desc_pool->buf_size);
|
|
if (qdf_unlikely(QDF_IS_STATUS_ERROR(ret))) {
|
|
qdf_nbuf_free(nbuf);
|
|
continue;
|
|
}
|
|
|
|
dp_audio_smmu_map(soc->osdev,
|
|
qdf_mem_paddr_from_dmaaddr(soc->osdev,
|
|
QDF_NBUF_CB_PADDR(nbuf)),
|
|
QDF_NBUF_CB_PADDR(nbuf),
|
|
rx_desc_pool->buf_size);
|
|
|
|
buff_pool->buf_elem[head] = nbuf;
|
|
head++;
|
|
}
|
|
|
|
buff_pool->head = head;
|
|
|
|
dp_info("RX refill buffer pool required allocation: %u actual allocation: %u",
|
|
buff_pool->max_bufq_len,
|
|
buff_pool->head);
|
|
|
|
buff_pool->is_initialized = true;
|
|
}
|
|
|
|
void dp_rx_buffer_pool_init(struct dp_soc *soc, u8 mac_id)
|
|
{
|
|
struct rx_desc_pool *rx_desc_pool = &soc->rx_desc_buf[mac_id];
|
|
struct rx_buff_pool *buff_pool = &soc->rx_buff_pool[mac_id];
|
|
qdf_nbuf_t nbuf;
|
|
int i;
|
|
|
|
dp_rx_refill_buff_pool_init(soc, mac_id);
|
|
|
|
if (!wlan_cfg_is_rx_buffer_pool_enabled(soc->wlan_cfg_ctx)) {
|
|
dp_info("RX buffer pool support is disabled");
|
|
buff_pool->is_initialized = false;
|
|
return;
|
|
}
|
|
|
|
if (buff_pool->is_initialized)
|
|
return;
|
|
|
|
qdf_nbuf_queue_head_init(&buff_pool->emerg_nbuf_q);
|
|
|
|
for (i = 0; i < DP_RX_BUFFER_POOL_SIZE; i++) {
|
|
nbuf = qdf_nbuf_alloc(soc->osdev, rx_desc_pool->buf_size,
|
|
RX_BUFFER_RESERVATION,
|
|
rx_desc_pool->buf_alignment, FALSE);
|
|
if (!nbuf)
|
|
continue;
|
|
qdf_nbuf_queue_head_enqueue_tail(&buff_pool->emerg_nbuf_q,
|
|
nbuf);
|
|
}
|
|
|
|
dp_info("RX buffer pool required allocation: %u actual allocation: %u",
|
|
DP_RX_BUFFER_POOL_SIZE,
|
|
qdf_nbuf_queue_head_qlen(&buff_pool->emerg_nbuf_q));
|
|
|
|
buff_pool->is_initialized = true;
|
|
}
|
|
|
|
static void dp_rx_refill_buff_pool_deinit(struct dp_soc *soc, u8 mac_id)
|
|
{
|
|
struct rx_refill_buff_pool *buff_pool = &soc->rx_refill_buff_pool;
|
|
struct rx_desc_pool *rx_desc_pool = &soc->rx_desc_buf[mac_id];
|
|
qdf_nbuf_t nbuf;
|
|
uint32_t count = 0;
|
|
|
|
if (!buff_pool->is_initialized)
|
|
return;
|
|
|
|
while ((nbuf = dp_rx_refill_buff_pool_dequeue_nbuf(soc))) {
|
|
dp_audio_smmu_unmap(soc->osdev,
|
|
QDF_NBUF_CB_PADDR(nbuf),
|
|
rx_desc_pool->buf_size);
|
|
qdf_nbuf_unmap_nbytes_single(soc->osdev, nbuf,
|
|
QDF_DMA_BIDIRECTIONAL,
|
|
rx_desc_pool->buf_size);
|
|
qdf_nbuf_free(nbuf);
|
|
count++;
|
|
}
|
|
|
|
dp_info("Rx refill buffers freed during deinit %u head: %u, tail: %u",
|
|
count, buff_pool->head, buff_pool->tail);
|
|
|
|
buff_pool->is_initialized = false;
|
|
}
|
|
|
|
void dp_rx_buffer_pool_deinit(struct dp_soc *soc, u8 mac_id)
|
|
{
|
|
struct rx_buff_pool *buff_pool = &soc->rx_buff_pool[mac_id];
|
|
qdf_nbuf_t nbuf;
|
|
|
|
dp_rx_refill_buff_pool_deinit(soc, mac_id);
|
|
|
|
if (!buff_pool->is_initialized)
|
|
return;
|
|
|
|
dp_info("buffers in the RX buffer pool during deinit: %u",
|
|
qdf_nbuf_queue_head_qlen(&buff_pool->emerg_nbuf_q));
|
|
|
|
while ((nbuf = qdf_nbuf_queue_head_dequeue(&buff_pool->emerg_nbuf_q)))
|
|
qdf_nbuf_free(nbuf);
|
|
|
|
buff_pool->is_initialized = false;
|
|
}
|
|
#endif /* WLAN_FEATURE_RX_PREALLOC_BUFFER_POOL */
|