qcacmn: Add support for WLAN-IPA WDI2 with SMMU Stage 1

Update WLAN-IPA WDI-2 datapath for DMA buffer sharing with SMMU
Stage 1 translation support. When SMMU Stage 1 is enabled DMA
APIs return IO virtual address(IOVA) instead of physical address.
This IOVA need to mapped to physical address by IPA module
before accessing them.

Change-Id: I33082bc74760b0f12c348238c6f48f5ce5750172
CRS-Fixed: 2072953
This commit is contained in:
Sravan Kumar Kairam
2018-02-27 18:37:40 +05:30
committed by nshrivas
parent 59c25048cd
commit 58e0adfb53
14 changed files with 157 additions and 87 deletions

View File

@@ -282,27 +282,22 @@ struct txrx_pdev_cfg_param_t {
* ol_txrx_ipa_resources - Resources needed for IPA * ol_txrx_ipa_resources - Resources needed for IPA
*/ */
struct ol_txrx_ipa_resources { struct ol_txrx_ipa_resources {
qdf_dma_addr_t ce_sr_base_paddr; qdf_shared_mem_t *ce_sr;
uint32_t ce_sr_ring_size; uint32_t ce_sr_ring_size;
qdf_dma_addr_t ce_reg_paddr; qdf_dma_addr_t ce_reg_paddr;
qdf_dma_addr_t tx_comp_ring_base_paddr; qdf_shared_mem_t *tx_comp_ring;
uint32_t tx_comp_ring_size;
uint32_t tx_num_alloc_buffer; uint32_t tx_num_alloc_buffer;
qdf_dma_addr_t rx_rdy_ring_base_paddr; qdf_shared_mem_t *rx_rdy_ring;
uint32_t rx_rdy_ring_size; qdf_shared_mem_t *rx_proc_done_idx;
qdf_dma_addr_t rx_proc_done_idx_paddr;
void *rx_proc_done_idx_vaddr;
qdf_dma_addr_t rx2_rdy_ring_base_paddr; qdf_shared_mem_t *rx2_rdy_ring;
uint32_t rx2_rdy_ring_size; qdf_shared_mem_t *rx2_proc_done_idx;
qdf_dma_addr_t rx2_proc_done_idx_paddr;
void *rx2_proc_done_idx_vaddr;
/* IPA UC doorbell registers paddr */ /* IPA UC doorbell registers paddr */
qdf_dma_addr_t tx_comp_doorbell_paddr; qdf_dma_addr_t tx_comp_doorbell_dmaaddr;
qdf_dma_addr_t rx_ready_doorbell_paddr; qdf_dma_addr_t rx_ready_doorbell_dmaaddr;
uint32_t tx_pipe_handle; uint32_t tx_pipe_handle;
uint32_t rx_pipe_handle; uint32_t rx_pipe_handle;

View File

@@ -522,7 +522,7 @@ static inline void *hif_get_ce_handle(struct hif_opaque_softc *hif_ctx, int ret)
#define CONFIG_DISABLE_CDC_MAX_PERF_WAR 0 #define CONFIG_DISABLE_CDC_MAX_PERF_WAR 0
void hif_ipa_get_ce_resource(struct hif_opaque_softc *hif_ctx, void hif_ipa_get_ce_resource(struct hif_opaque_softc *hif_ctx,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr); qdf_dma_addr_t *ce_reg_paddr);

View File

@@ -1,5 +1,5 @@
/* /*
* Copyright (c) 2013-2017 The Linux Foundation. All rights reserved. * Copyright (c) 2013-2018 The Linux Foundation. All rights reserved.
* *
* Previously licensed under the ISC license by Qualcomm Atheros, Inc. * Previously licensed under the ISC license by Qualcomm Atheros, Inc.
* *
@@ -447,14 +447,14 @@ struct ce_sendlist {
#ifdef IPA_OFFLOAD #ifdef IPA_OFFLOAD
void ce_ipa_get_resource(struct CE_handle *ce, void ce_ipa_get_resource(struct CE_handle *ce,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr); qdf_dma_addr_t *ce_reg_paddr);
#else #else
/** /**
* ce_ipa_get_resource() - get uc resource on copyengine * ce_ipa_get_resource() - get uc resource on copyengine
* @ce: copyengine context * @ce: copyengine context
* @ce_sr_base_paddr: copyengine source ring base physical address * @ce_sr: copyengine source ring resource info
* @ce_sr_ring_size: copyengine source ring size * @ce_sr_ring_size: copyengine source ring size
* @ce_reg_paddr: copyengine register physical address * @ce_reg_paddr: copyengine register physical address
* *
@@ -467,7 +467,7 @@ void ce_ipa_get_resource(struct CE_handle *ce,
* Return: None * Return: None
*/ */
static inline void ce_ipa_get_resource(struct CE_handle *ce, static inline void ce_ipa_get_resource(struct CE_handle *ce,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr) qdf_dma_addr_t *ce_reg_paddr)
{ {

View File

@@ -698,6 +698,104 @@ static void ce_ring_test_initial_indexes(int ce_id, struct CE_ring_state *ring,
QDF_BUG(0); QDF_BUG(0);
} }
#ifdef IPA_OFFLOAD
/**
* ce_alloc_desc_ring() - Allocate copyengine descriptor ring
* @scn: softc instance
* @ce_id: ce in question
* @base_addr: pointer to copyengine ring base address
* @ce_ring: copyengine instance
* @nentries: number of entries should be allocated
* @desc_size: ce desc size
*
* Return: QDF_STATUS_SUCCESS - for success
*/
static QDF_STATUS ce_alloc_desc_ring(struct hif_softc *scn, unsigned int CE_id,
qdf_dma_addr_t *base_addr,
struct CE_ring_state *ce_ring,
unsigned int nentries, uint32_t desc_size)
{
if (CE_id == HIF_PCI_IPA_UC_ASSIGNED_CE) {
scn->ipa_ce_ring = qdf_mem_shared_mem_alloc(scn->qdf_dev,
nentries * desc_size + CE_DESC_RING_ALIGN);
if (!scn->ipa_ce_ring) {
HIF_ERROR("%s: Failed to allocate memory for IPA ce ring",
__func__);
return QDF_STATUS_E_NOMEM;
}
*base_addr = qdf_mem_get_dma_addr(scn->qdf_dev,
&scn->ipa_ce_ring->mem_info);
ce_ring->base_addr_owner_space_unaligned =
scn->ipa_ce_ring->vaddr;
} else {
ce_ring->base_addr_owner_space_unaligned =
qdf_mem_alloc_consistent(scn->qdf_dev,
scn->qdf_dev->dev,
(nentries * desc_size +
CE_DESC_RING_ALIGN),
base_addr);
if (!ce_ring->base_addr_owner_space_unaligned) {
HIF_ERROR("%s: Failed to allocate DMA memory for ce ring id : %u",
__func__, CE_id);
return QDF_STATUS_E_NOMEM;
}
}
return QDF_STATUS_SUCCESS;
}
/**
* ce_free_desc_ring() - Frees copyengine descriptor ring
* @scn: softc instance
* @ce_id: ce in question
* @ce_ring: copyengine instance
* @desc_size: ce desc size
*
* Return: None
*/
static void ce_free_desc_ring(struct hif_softc *scn, unsigned int CE_id,
struct CE_ring_state *ce_ring, uint32_t desc_size)
{
if (CE_id == HIF_PCI_IPA_UC_ASSIGNED_CE) {
qdf_mem_shared_mem_free(scn->qdf_dev,
scn->ipa_ce_ring);
ce_ring->base_addr_owner_space_unaligned = NULL;
} else {
qdf_mem_free_consistent(scn->qdf_dev, scn->qdf_dev->dev,
ce_ring->nentries * desc_size + CE_DESC_RING_ALIGN,
ce_ring->base_addr_owner_space_unaligned,
ce_ring->base_addr_CE_space, 0);
ce_ring->base_addr_owner_space_unaligned = NULL;
}
}
#else
static QDF_STATUS ce_alloc_desc_ring(struct hif_softc *scn, unsigned int CE_id,
qdf_dma_addr_t *base_addr,
struct CE_ring_state *ce_ring,
unsigned int nentries, uint32_t desc_size)
{
ce_ring->base_addr_owner_space_unaligned =
qdf_mem_alloc_consistent(scn->qdf_dev, scn->qdf_dev->dev,
(nentries * desc_size +
CE_DESC_RING_ALIGN), base_addr);
if (!ce_ring->base_addr_owner_space_unaligned) {
HIF_ERROR("%s: Failed to allocate DMA memory for ce ring id : %u",
__func__, CE_id);
return QDF_STATUS_E_NOMEM;
}
return QDF_STATUS_SUCCESS;
}
static void ce_free_desc_ring(struct hif_softc *scn, unsigned int CE_id,
struct CE_ring_state *ce_ring, uint32_t desc_size)
{
qdf_mem_free_consistent(scn->qdf_dev, scn->qdf_dev->dev,
ce_ring->nentries * desc_size + CE_DESC_RING_ALIGN,
ce_ring->base_addr_owner_space_unaligned,
ce_ring->base_addr_CE_space, 0);
ce_ring->base_addr_owner_space_unaligned = NULL;
}
#endif /* IPA_OFFLOAD */
/** /**
* ce_srng_based() - Does this target use srng * ce_srng_based() - Does this target use srng
* @ce_state : pointer to the state context of the CE * @ce_state : pointer to the state context of the CE
@@ -790,15 +888,10 @@ static struct CE_ring_state *ce_alloc_ring_state(struct CE_state *CE_state,
/* Legacy platforms that do not support cache /* Legacy platforms that do not support cache
* coherent DMA are unsupported * coherent DMA are unsupported
*/ */
ce_ring->base_addr_owner_space_unaligned = if (ce_alloc_desc_ring(scn, CE_state->id, &base_addr,
qdf_mem_alloc_consistent(scn->qdf_dev, ce_ring, nentries,
scn->qdf_dev->dev, desc_size) !=
(nentries * QDF_STATUS_SUCCESS) {
desc_size +
CE_DESC_RING_ALIGN),
&base_addr);
if (ce_ring->base_addr_owner_space_unaligned
== NULL) {
HIF_ERROR("%s: ring has no DMA mem", HIF_ERROR("%s: ring has no DMA mem",
__func__); __func__);
qdf_mem_free(ptr); qdf_mem_free(ptr);
@@ -1520,15 +1613,9 @@ void ce_fini(struct CE_handle *copyeng)
if (CE_state->src_ring->shadow_base_unaligned) if (CE_state->src_ring->shadow_base_unaligned)
qdf_mem_free(CE_state->src_ring->shadow_base_unaligned); qdf_mem_free(CE_state->src_ring->shadow_base_unaligned);
if (CE_state->src_ring->base_addr_owner_space_unaligned) if (CE_state->src_ring->base_addr_owner_space_unaligned)
qdf_mem_free_consistent(scn->qdf_dev, ce_free_desc_ring(scn, CE_state->id,
scn->qdf_dev->dev, CE_state->src_ring,
(CE_state->src_ring->nentries * desc_size);
desc_size +
CE_DESC_RING_ALIGN),
CE_state->src_ring->
base_addr_owner_space_unaligned,
CE_state->src_ring->
base_addr_CE_space, 0);
qdf_mem_free(CE_state->src_ring); qdf_mem_free(CE_state->src_ring);
} }
if (CE_state->dest_ring) { if (CE_state->dest_ring) {
@@ -1537,15 +1624,9 @@ void ce_fini(struct CE_handle *copyeng)
desc_size = ce_get_desc_size(scn, CE_RING_DEST); desc_size = ce_get_desc_size(scn, CE_RING_DEST);
if (CE_state->dest_ring->base_addr_owner_space_unaligned) if (CE_state->dest_ring->base_addr_owner_space_unaligned)
qdf_mem_free_consistent(scn->qdf_dev, ce_free_desc_ring(scn, CE_state->id,
scn->qdf_dev->dev, CE_state->dest_ring,
(CE_state->dest_ring->nentries * desc_size);
desc_size +
CE_DESC_RING_ALIGN),
CE_state->dest_ring->
base_addr_owner_space_unaligned,
CE_state->dest_ring->
base_addr_CE_space, 0);
qdf_mem_free(CE_state->dest_ring); qdf_mem_free(CE_state->dest_ring);
/* epping */ /* epping */
@@ -1563,15 +1644,9 @@ void ce_fini(struct CE_handle *copyeng)
desc_size = ce_get_desc_size(scn, CE_RING_STATUS); desc_size = ce_get_desc_size(scn, CE_RING_STATUS);
if (CE_state->status_ring->base_addr_owner_space_unaligned) if (CE_state->status_ring->base_addr_owner_space_unaligned)
qdf_mem_free_consistent(scn->qdf_dev, ce_free_desc_ring(scn, CE_state->id,
scn->qdf_dev->dev, CE_state->status_ring,
(CE_state->status_ring->nentries * desc_size);
desc_size +
CE_DESC_RING_ALIGN),
CE_state->status_ring->
base_addr_owner_space_unaligned,
CE_state->status_ring->
base_addr_CE_space, 0);
qdf_mem_free(CE_state->status_ring); qdf_mem_free(CE_state->status_ring);
} }
@@ -2840,7 +2915,7 @@ qdf_export_symbol(hif_ce_fastpath_cb_register);
* Return: None * Return: None
*/ */
void hif_ce_ipa_get_ce_resource(struct hif_softc *scn, void hif_ce_ipa_get_ce_resource(struct hif_softc *scn,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr) qdf_dma_addr_t *ce_reg_paddr)
{ {
@@ -2849,7 +2924,7 @@ void hif_ce_ipa_get_ce_resource(struct hif_softc *scn,
&(hif_state->pipe_info[HIF_PCI_IPA_UC_ASSIGNED_CE]); &(hif_state->pipe_info[HIF_PCI_IPA_UC_ASSIGNED_CE]);
struct CE_handle *ce_hdl = pipe_info->ce_hdl; struct CE_handle *ce_hdl = pipe_info->ce_hdl;
ce_ipa_get_resource(ce_hdl, ce_sr_base_paddr, ce_sr_ring_size, ce_ipa_get_resource(ce_hdl, ce_sr, ce_sr_ring_size,
ce_reg_paddr); ce_reg_paddr);
} }
#endif /* IPA_OFFLOAD */ #endif /* IPA_OFFLOAD */

View File

@@ -195,13 +195,13 @@ hif_ce_dump_target_memory(struct hif_softc *scn, void *ramdump_base,
#ifdef IPA_OFFLOAD #ifdef IPA_OFFLOAD
void hif_ce_ipa_get_ce_resource(struct hif_softc *scn, void hif_ce_ipa_get_ce_resource(struct hif_softc *scn,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr); qdf_dma_addr_t *ce_reg_paddr);
#else #else
static inline static inline
void hif_ce_ipa_get_ce_resource(struct hif_softc *scn, void hif_ce_ipa_get_ce_resource(struct hif_softc *scn,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr) qdf_dma_addr_t *ce_reg_paddr)
{ {

View File

@@ -2511,7 +2511,7 @@ qdf_export_symbol(ce_check_rx_pending);
/** /**
* ce_ipa_get_resource() - get uc resource on copyengine * ce_ipa_get_resource() - get uc resource on copyengine
* @ce: copyengine context * @ce: copyengine context
* @ce_sr_base_paddr: copyengine source ring base physical address * @ce_sr: copyengine source ring resource info
* @ce_sr_ring_size: copyengine source ring size * @ce_sr_ring_size: copyengine source ring size
* @ce_reg_paddr: copyengine register physical address * @ce_reg_paddr: copyengine register physical address
* *
@@ -2524,7 +2524,7 @@ qdf_export_symbol(ce_check_rx_pending);
* Return: None * Return: None
*/ */
void ce_ipa_get_resource(struct CE_handle *ce, void ce_ipa_get_resource(struct CE_handle *ce,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr) qdf_dma_addr_t *ce_reg_paddr)
{ {
@@ -2535,7 +2535,8 @@ void ce_ipa_get_resource(struct CE_handle *ce,
struct hif_softc *scn = CE_state->scn; struct hif_softc *scn = CE_state->scn;
if (CE_UNUSED == CE_state->state) { if (CE_UNUSED == CE_state->state) {
*ce_sr_base_paddr = 0; *qdf_mem_get_dma_addr_ptr(scn->qdf_dev,
&CE_state->scn->ipa_ce_ring->mem_info) = 0;
*ce_sr_ring_size = 0; *ce_sr_ring_size = 0;
return; return;
} }
@@ -2552,8 +2553,8 @@ void ce_ipa_get_resource(struct CE_handle *ce,
/* Get BAR address */ /* Get BAR address */
hif_read_phy_mem_base(CE_state->scn, &phy_mem_base); hif_read_phy_mem_base(CE_state->scn, &phy_mem_base);
*ce_sr_base_paddr = CE_state->src_ring->base_addr_CE_space; *ce_sr = CE_state->scn->ipa_ce_ring;
*ce_sr_ring_size = (uint32_t) (CE_state->src_ring->nentries * *ce_sr_ring_size = (uint32_t)(CE_state->src_ring->nentries *
sizeof(struct CE_src_desc)); sizeof(struct CE_src_desc));
*ce_reg_paddr = phy_mem_base + CE_BASE_ADDRESS(CE_state->id) + *ce_reg_paddr = phy_mem_base + CE_BASE_ADDRESS(CE_state->id) +
SR_WR_INDEX_ADDRESS; SR_WR_INDEX_ADDRESS;

View File

@@ -1,5 +1,5 @@
/* /*
* Copyright (c) 2016-2017 The Linux Foundation. All rights reserved. * Copyright (c) 2016-2018 The Linux Foundation. All rights reserved.
* *
* Previously licensed under the ISC license by Qualcomm Atheros, Inc. * Previously licensed under the ISC license by Qualcomm Atheros, Inc.
* *
@@ -299,14 +299,14 @@ void hif_dummy_dump_target_memory(struct hif_softc *hif_sc, void *ramdump_base,
/** /**
* hif_dummy_ipa_get_ce_resource - dummy call * hif_dummy_ipa_get_ce_resource - dummy call
* @scn: HIF context * @scn: HIF context
* @sr_base_paddr: source base address * @ce_sr: copyengine source ring resource info
* @sr_ring_size: source ring size * @sr_ring_size: source ring size
* @reg_paddr: bus physical address * @reg_paddr: bus physical address
* *
* Return: None * Return: None
*/ */
void hif_dummy_ipa_get_ce_resource(struct hif_softc *hif_sc, void hif_dummy_ipa_get_ce_resource(struct hif_softc *hif_sc,
qdf_dma_addr_t *sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *sr_ring_size, uint32_t *sr_ring_size,
qdf_dma_addr_t *reg_paddr) qdf_dma_addr_t *reg_paddr)
{ {

View File

@@ -1,5 +1,5 @@
/* /*
* Copyright (c) 2016-2017 The Linux Foundation. All rights reserved. * Copyright (c) 2016-2018 The Linux Foundation. All rights reserved.
* *
* Previously licensed under the ISC license by Qualcomm Atheros, Inc. * Previously licensed under the ISC license by Qualcomm Atheros, Inc.
* *
@@ -56,7 +56,7 @@ int hif_dummy_dump_registers(struct hif_softc *hif_sc);
void hif_dummy_dump_target_memory(struct hif_softc *hif_sc, void *ramdump_base, void hif_dummy_dump_target_memory(struct hif_softc *hif_sc, void *ramdump_base,
uint32_t address, uint32_t size); uint32_t address, uint32_t size);
void hif_dummy_ipa_get_ce_resource(struct hif_softc *hif_sc, void hif_dummy_ipa_get_ce_resource(struct hif_softc *hif_sc,
qdf_dma_addr_t *sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *sr_ring_size, uint32_t *sr_ring_size,
qdf_dma_addr_t *reg_paddr); qdf_dma_addr_t *reg_paddr);
void hif_dummy_mask_interrupt_call(struct hif_softc *hif_sc); void hif_dummy_mask_interrupt_call(struct hif_softc *hif_sc);

View File

@@ -351,14 +351,14 @@ void hif_dump_target_memory(struct hif_opaque_softc *hif_hdl,
} }
void hif_ipa_get_ce_resource(struct hif_opaque_softc *hif_hdl, void hif_ipa_get_ce_resource(struct hif_opaque_softc *hif_hdl,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr) qdf_dma_addr_t *ce_reg_paddr)
{ {
struct hif_softc *hif_sc = HIF_GET_SOFTC(hif_hdl); struct hif_softc *hif_sc = HIF_GET_SOFTC(hif_hdl);
hif_sc->bus_ops.hif_ipa_get_ce_resource(hif_sc, ce_sr_base_paddr, hif_sc->bus_ops.hif_ipa_get_ce_resource(hif_sc, ce_sr,
ce_sr_ring_size, ce_reg_paddr); ce_sr_ring_size, ce_reg_paddr);
} }
void hif_mask_interrupt_call(struct hif_opaque_softc *hif_hdl) void hif_mask_interrupt_call(struct hif_opaque_softc *hif_hdl)

View File

@@ -74,7 +74,7 @@ struct hif_bus_ops {
void *ramdump_base, void *ramdump_base,
uint32_t address, uint32_t size); uint32_t address, uint32_t size);
void (*hif_ipa_get_ce_resource)(struct hif_softc *hif_sc, void (*hif_ipa_get_ce_resource)(struct hif_softc *hif_sc,
qdf_dma_addr_t *sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *sr_ring_size, uint32_t *sr_ring_size,
qdf_dma_addr_t *reg_paddr); qdf_dma_addr_t *reg_paddr);
void (*hif_mask_interrupt_call)(struct hif_softc *hif_sc); void (*hif_mask_interrupt_call)(struct hif_softc *hif_sc);

View File

@@ -199,6 +199,9 @@ struct hif_softc {
#if HIF_CE_DEBUG_DATA_BUF #if HIF_CE_DEBUG_DATA_BUF
struct ce_desc_hist hif_ce_desc_hist; struct ce_desc_hist hif_ce_desc_hist;
#endif /* #if defined(HIF_CONFIG_SLUB_DEBUG_ON) || HIF_CE_DEBUG_DATA_BUF */ #endif /* #if defined(HIF_CONFIG_SLUB_DEBUG_ON) || HIF_CE_DEBUG_DATA_BUF */
#ifdef IPA_OFFLOAD
qdf_shared_mem_t *ipa_ce_ring;
#endif
}; };
static inline void *hif_get_hal_handle(void *hif_hdl) static inline void *hif_get_hal_handle(void *hif_hdl)

View File

@@ -237,12 +237,12 @@ static inline int hif_snoc_get_target_type(struct hif_softc *ol_sc,
} }
#ifdef IPA_OFFLOAD #ifdef IPA_OFFLOAD
static int hif_set_dma_coherent_mask(struct device *dev) static int hif_set_dma_coherent_mask(qdf_device_t osdev)
{ {
uint8_t addr_bits; uint8_t addr_bits;
if (false == hif_get_ipa_present()) if (false == hif_get_ipa_present())
return qdf_set_dma_coherent_mask(dev, return qdf_set_dma_coherent_mask(osdev->dev,
DMA_COHERENT_MASK_IPA_VER_3_AND_ABOVE); DMA_COHERENT_MASK_IPA_VER_3_AND_ABOVE);
if (hif_get_ipa_hw_type() < IPA_HW_v3_0) if (hif_get_ipa_hw_type() < IPA_HW_v3_0)
@@ -250,12 +250,12 @@ static int hif_set_dma_coherent_mask(struct device *dev)
else else
addr_bits = DMA_COHERENT_MASK_IPA_VER_3_AND_ABOVE; addr_bits = DMA_COHERENT_MASK_IPA_VER_3_AND_ABOVE;
return qdf_set_dma_coherent_mask(dev, addr_bits); return qdf_set_dma_coherent_mask(osdev->dev, addr_bits);
} }
#else #else
static int hif_set_dma_coherent_mask(struct device *dev) static int hif_set_dma_coherent_mask(qdf_device_t osdev)
{ {
return qdf_set_dma_coherent_mask(dev, 37); return qdf_set_dma_coherent_mask(osdev->dev, 37);
} }
#endif #endif
@@ -282,7 +282,7 @@ QDF_STATUS hif_snoc_enable_bus(struct hif_softc *ol_sc,
return QDF_STATUS_E_NOMEM; return QDF_STATUS_E_NOMEM;
} }
ret = hif_set_dma_coherent_mask(dev); ret = hif_set_dma_coherent_mask(ol_sc->qdf_dev);
if (ret) { if (ret) {
HIF_ERROR("%s: failed to set dma mask error = %d", HIF_ERROR("%s: failed to set dma mask error = %d",
__func__, ret); __func__, ret);

View File

@@ -945,17 +945,15 @@ void *htc_get_targetdef(HTC_HANDLE htc_handle)
* Return: None * Return: None
*/ */
void htc_ipa_get_ce_resource(HTC_HANDLE htc_handle, void htc_ipa_get_ce_resource(HTC_HANDLE htc_handle,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr) qdf_dma_addr_t *ce_reg_paddr)
{ {
HTC_TARGET *target = GET_HTC_TARGET_FROM_HANDLE(htc_handle); HTC_TARGET *target = GET_HTC_TARGET_FROM_HANDLE(htc_handle);
if (target->hif_dev != NULL) { if (target->hif_dev)
hif_ipa_get_ce_resource(target->hif_dev, hif_ipa_get_ce_resource(target->hif_dev,
ce_sr_base_paddr, ce_sr, ce_sr_ring_size, ce_reg_paddr);
ce_sr_ring_size, ce_reg_paddr);
}
} }
#endif /* IPA_OFFLOAD */ #endif /* IPA_OFFLOAD */

View File

@@ -747,14 +747,12 @@ void htc_vote_link_down(HTC_HANDLE HTCHandle);
void htc_vote_link_up(HTC_HANDLE HTCHandle); void htc_vote_link_up(HTC_HANDLE HTCHandle);
#ifdef IPA_OFFLOAD #ifdef IPA_OFFLOAD
void htc_ipa_get_ce_resource(HTC_HANDLE htc_handle, void htc_ipa_get_ce_resource(HTC_HANDLE htc_handle,
qdf_dma_addr_t *ce_sr_base_paddr, qdf_shared_mem_t **ce_sr,
uint32_t *ce_sr_ring_size, uint32_t *ce_sr_ring_size,
qdf_dma_addr_t *ce_reg_paddr); qdf_dma_addr_t *ce_reg_paddr);
#else #else
#define htc_ipa_get_ce_resource(htc_handle, \ #define htc_ipa_get_ce_resource(htc_handle, \
ce_sr_base_paddr, \ ce_sr, ce_sr_ring_size, ce_reg_paddr) /* NO-OP */
ce_sr_ring_size, \
ce_reg_paddr) /* NO-OP */
#endif /* IPA_OFFLOAD */ #endif /* IPA_OFFLOAD */
#if defined(DEBUG_HL_LOGGING) && defined(CONFIG_HL_SUPPORT) #if defined(DEBUG_HL_LOGGING) && defined(CONFIG_HL_SUPPORT)