qcacmn: qdf: Fix misspellings
Fix misspellings in qdf/... Change-Id: If61ee47fba94b2bb60f33ab74feda56dbe5cb7bd CRs-Fixed: 3277902
This commit is contained in:
@@ -113,7 +113,7 @@ __qdf_queue_work(__qdf_workqueue_t *wqueue, __qdf_work_t *work)
|
||||
/**
|
||||
* __qdf_sched_work - Schedule a deferred task on non-interrupt context
|
||||
* @work: pointer to work
|
||||
* Retrun: none
|
||||
* Return: none
|
||||
*/
|
||||
static inline QDF_STATUS __qdf_sched_work(__qdf_work_t *work)
|
||||
{
|
||||
|
@@ -26,7 +26,7 @@
|
||||
/**
|
||||
* __qdf_ipa_wdi_meter_evt_type_t - type of event client callback is
|
||||
* for AP+STA mode metering
|
||||
* @IPA_GET_WDI_SAP_STATS: get IPA_stats betwen SAP and STA -
|
||||
* @IPA_GET_WDI_SAP_STATS: get IPA_stats between SAP and STA -
|
||||
* use ipa_get_wdi_sap_stats structure
|
||||
* @IPA_SET_WIFI_QUOTA: set quota limit on STA -
|
||||
* use ipa_set_wifi_quota structure
|
||||
|
@@ -1,5 +1,6 @@
|
||||
/*
|
||||
* Copyright (c) 2014-2018, 2020 The Linux Foundation. All rights reserved.
|
||||
* Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
|
||||
*
|
||||
* Permission to use, copy, modify, and/or distribute this software for
|
||||
* any purpose with or without fee is hereby granted, provided that the
|
||||
@@ -101,7 +102,7 @@ typedef struct qdf_runtime_lock {
|
||||
|
||||
#define LINUX_LOCK_COOKIE 0x12345678
|
||||
|
||||
/* Function declarations and documenation */
|
||||
/* Function declarations and documentation */
|
||||
|
||||
#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 37)
|
||||
/**
|
||||
|
@@ -86,7 +86,7 @@ typedef struct mempool_elem {
|
||||
* @mem_size: Total size of the pool in bytes
|
||||
* @free_list: free pool list
|
||||
* @lock: spinlock object
|
||||
* @max_elem: Maximum number of elements in tha pool
|
||||
* @max_elem: Maximum number of elements in the pool
|
||||
* @free_cnt: Number of free elements available
|
||||
*/
|
||||
typedef struct __qdf_mempool_ctxt {
|
||||
@@ -150,7 +150,7 @@ enum dma_data_direction __qdf_dma_dir_to_os(qdf_dma_dir_t qdf_dir)
|
||||
* @buf: pointer to memory to be dma mapped
|
||||
* @dir: DMA map direction
|
||||
* @nbytes: number of bytes to be mapped.
|
||||
* @phy_addr: ponter to recive physical address.
|
||||
* @phy_addr: pointer to receive physical address.
|
||||
*
|
||||
* Return: success/failure
|
||||
*/
|
||||
@@ -441,7 +441,7 @@ __qdf_mem_paddr_from_dmaaddr(qdf_device_t osdev,
|
||||
|
||||
/**
|
||||
* __qdf_os_mem_dma_get_sgtable() - Returns DMA memory scatter gather table
|
||||
* @dev: device instace
|
||||
* @dev: device instance
|
||||
* @sgt: scatter gather table pointer
|
||||
* @cpu_addr: HLOS virtual address
|
||||
* @dma_addr: dma/iova
|
||||
|
@@ -126,7 +126,7 @@ typedef union {
|
||||
*
|
||||
* Notes:
|
||||
* 1. Hard limited to 48 bytes. Please count your bytes
|
||||
* 2. The size of this structure has to be easily calculatable and
|
||||
* 2. The size of this structure has to be easily calculable and
|
||||
* consistently so: do not use any conditional compile flags
|
||||
* 3. Split into a common part followed by a tx/rx overlay
|
||||
* 4. There is only one extra frag, which represents the HTC/HTT header
|
||||
@@ -155,7 +155,7 @@ typedef union {
|
||||
* @rx.dev.priv_cb_m.exc_frm: exception frame
|
||||
* @rx.dev.priv_cb_m.ipa_smmu_map: do IPA smmu map
|
||||
* @rx.dev.priv_cb_m.reo_dest_ind_or_sw_excpt: reo destination indication or
|
||||
sw execption bit from ring desc
|
||||
sw exception bit from ring desc
|
||||
* @rx.dev.priv_cb_m.lmac_id: lmac id for RX packet
|
||||
* @rx.dev.priv_cb_m.tcp_seq_num: TCP sequence number
|
||||
* @rx.dev.priv_cb_m.tcp_ack_num: TCP ACK number
|
||||
|
@@ -28,7 +28,7 @@
|
||||
#ifndef _I_QDF_NBUF_W_H
|
||||
#define _I_QDF_NBUF_W_H
|
||||
|
||||
/* ext_cb accesor macros and internal API's */
|
||||
/* ext_cb accessor macros and internal API's */
|
||||
|
||||
#define QDF_NBUF_CB_EXT_CB(skb) \
|
||||
(((struct qdf_nbuf_cb *)((skb)->cb))->u.tx.dev.priv_cb_w.ext_cb_ptr)
|
||||
@@ -41,7 +41,7 @@
|
||||
#define __qdf_nbuf_get_ext_cb(skb) \
|
||||
QDF_NBUF_CB_EXT_CB((skb))
|
||||
|
||||
/* fctx accesor macros and internal API's*/
|
||||
/* fctx accessor macros and internal API's*/
|
||||
|
||||
#define QDF_NBUF_CB_RX_FCTX(skb) \
|
||||
(((struct qdf_nbuf_cb *)((skb)->cb))->u.rx.dev.priv_cb_w.fctx)
|
||||
|
@@ -1,5 +1,6 @@
|
||||
/*
|
||||
* Copyright (c) 2014-2021 The Linux Foundation. All rights reserved.
|
||||
* Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
|
||||
*
|
||||
* Permission to use, copy, modify, and/or distribute this software for
|
||||
* any purpose with or without fee is hereby granted, provided that the
|
||||
@@ -321,7 +322,7 @@ static inline uint64_t __qdf_get_log_timestamp(void)
|
||||
/**
|
||||
* __qdf_get_log_timestamp - get time stamp for logging
|
||||
*
|
||||
* Return: system tick for non MSM platfroms
|
||||
* Return: system tick for non MSM platforms
|
||||
*/
|
||||
#if (LINUX_VERSION_CODE >= KERNEL_VERSION(3, 17, 0))
|
||||
static inline uint64_t __qdf_get_log_timestamp(void)
|
||||
|
@@ -55,7 +55,7 @@
|
||||
#endif
|
||||
|
||||
/* QDF_TRACE is the macro invoked to add trace messages to code. See the
|
||||
* documenation for qdf_trace_msg() for the parameters etc. for this function.
|
||||
* documentation for qdf_trace_msg() for the parameters etc. for this function.
|
||||
*
|
||||
* NOTE: Code QDF_TRACE() macros into the source code. Do not code directly
|
||||
* to the qdf_trace_msg() function.
|
||||
|
@@ -1,5 +1,6 @@
|
||||
/*
|
||||
* Copyright (c) 2014-2018, 2021 The Linux Foundation. All rights reserved.
|
||||
* Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
|
||||
*
|
||||
* Permission to use, copy, modify, and/or distribute this software for
|
||||
* any purpose with or without fee is hereby granted, provided that the
|
||||
@@ -28,7 +29,7 @@
|
||||
#include <qdf_list.h>
|
||||
#include <qdf_module.h>
|
||||
|
||||
/* Function declarations and documenation */
|
||||
/* Function declarations and documentation */
|
||||
|
||||
QDF_STATUS qdf_list_insert_before(qdf_list_t *list,
|
||||
qdf_list_node_t *new_node, qdf_list_node_t *node)
|
||||
|
@@ -769,7 +769,7 @@ void qdf_lock_stats_init(void)
|
||||
__qdf_put_lock_cookie(&lock_cookies[i]);
|
||||
|
||||
/* stats must be allocated for the spinlock before the cookie,
|
||||
* otherwise this qdf_lock_list_spinlock wouldnt get initialized
|
||||
* otherwise this qdf_lock_list_spinlock wouldn't get initialized
|
||||
* properly
|
||||
*/
|
||||
qdf_spinlock_create(&qdf_lock_list_spinlock);
|
||||
|
@@ -1,5 +1,6 @@
|
||||
/*
|
||||
* Copyright (c) 2015-2017, 2019 The Linux Foundation. All rights reserved.
|
||||
* Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
|
||||
*
|
||||
* Permission to use, copy, modify, and/or distribute this software for
|
||||
* any purpose with or without fee is hereby granted, provided that the
|
||||
@@ -142,7 +143,7 @@ qdf_lro_ctx_t qdf_lro_init(void)
|
||||
lro_ctx->lro_mgr = (struct net_lro_mgr *)lro_mem_ptr;
|
||||
lro_mem_ptr += lro_mgr_sz;
|
||||
|
||||
/* LRO decriptor array */
|
||||
/* LRO descriptor array */
|
||||
lro_ctx->lro_mgr->lro_arr = (struct net_lro_desc *)lro_mem_ptr;
|
||||
lro_mem_ptr += desc_arr_sz;
|
||||
|
||||
@@ -189,7 +190,7 @@ void qdf_lro_deinit(qdf_lro_ctx_t lro_ctx)
|
||||
* qdf_lro_tcp_flow_match() - function to check for a flow match
|
||||
* @iph: IP header
|
||||
* @tcph: TCP header
|
||||
* @lro_desc: LRO decriptor
|
||||
* @lro_desc: LRO descriptor
|
||||
*
|
||||
* Checks if the descriptor belongs to the same flow as the one
|
||||
* indicated by the TCP and IP header.
|
||||
@@ -443,7 +444,7 @@ void qdf_lro_flush(qdf_lro_ctx_t lro_ctx)
|
||||
* qdf_lro_get_desc() - LRO descriptor look-up function
|
||||
* @iph: IP header
|
||||
* @tcph: TCP header
|
||||
* @lro_arr: Array of LRO decriptors
|
||||
* @lro_arr: Array of LRO descriptors
|
||||
* @lro_mgr: LRO manager
|
||||
*
|
||||
* Looks-up the LRO descriptor for a given flow
|
||||
|
@@ -70,7 +70,7 @@ void qdf_register_mc_timer_callback(void (*callback) (qdf_mc_timer_t *))
|
||||
|
||||
qdf_export_symbol(qdf_register_mc_timer_callback);
|
||||
|
||||
/* Function declarations and documenation */
|
||||
/* Function declarations and documentation */
|
||||
|
||||
/**
|
||||
* qdf_try_allowing_sleep() - clean up timer states after it has been deactivated
|
||||
@@ -675,7 +675,7 @@ QDF_STATUS qdf_mc_timer_start(qdf_mc_timer_t *timer, uint32_t expiration_time)
|
||||
return QDF_STATUS_E_INVAL;
|
||||
}
|
||||
|
||||
/* make sure the remainer of the logic isn't interrupted */
|
||||
/* make sure the remainder of the logic isn't interrupted */
|
||||
qdf_spin_lock_irqsave(&timer->platform_info.spinlock);
|
||||
|
||||
/* ensure if the timer can be started */
|
||||
|
@@ -111,7 +111,7 @@ enum list_type {
|
||||
* @type: type of the list to be parsed
|
||||
* @threshold: configured by user by overwriting the respective debugfs
|
||||
* sys entry. This is to list the functions which requested
|
||||
* memory/dma allocations more than threshold nubmer of times.
|
||||
* memory/dma allocations more than threshold number of times.
|
||||
*/
|
||||
struct major_alloc_priv {
|
||||
enum list_type type;
|
||||
@@ -1491,7 +1491,7 @@ static void qdf_mem_debug_init(void)
|
||||
if (is_initial_mem_debug_disabled)
|
||||
return;
|
||||
|
||||
/* Initalizing the list with maximum size of 60000 */
|
||||
/* Initializing the list with maximum size of 60000 */
|
||||
for (i = 0; i < QDF_DEBUG_DOMAIN_COUNT; ++i)
|
||||
qdf_list_create(&qdf_mem_domains[i], 60000);
|
||||
qdf_spinlock_create(&qdf_mem_list_lock);
|
||||
|
@@ -2884,7 +2884,7 @@ static void qdf_nbuf_track_free(QDF_NBUF_TRACK *node)
|
||||
/* Try to shrink the freelist if free_list_count > than FREEQ_POOLSIZE
|
||||
* only shrink the freelist if it is bigger than twice the number of
|
||||
* nbufs in use. If the driver is stalling in a consistent bursty
|
||||
* fasion, this will keep 3/4 of thee allocations from the free list
|
||||
* fashion, this will keep 3/4 of thee allocations from the free list
|
||||
* while also allowing the system to recover memory as less frantic
|
||||
* traffic occurs.
|
||||
*/
|
||||
@@ -3950,7 +3950,7 @@ static inline void qdf_nbuf_tso_unmap_frag(
|
||||
* information
|
||||
* @osdev: qdf device handle
|
||||
* @skb: skb buffer
|
||||
* @tso_info: Parameters common to all segements
|
||||
* @tso_info: Parameters common to all segments
|
||||
*
|
||||
* Get the TSO information that is common across all the TCP
|
||||
* segments of the jumbo packet
|
||||
@@ -4000,7 +4000,7 @@ static uint8_t __qdf_nbuf_get_tso_cmn_seg_info(qdf_device_t osdev,
|
||||
return 1;
|
||||
|
||||
if (tso_info->ethproto == htons(ETH_P_IP)) {
|
||||
/* inlcude IPv4 header length for IPV4 (total length) */
|
||||
/* include IPv4 header length for IPV4 (total length) */
|
||||
tso_info->ip_tcp_hdr_len =
|
||||
tso_info->eit_hdr_len - tso_info->l2_len;
|
||||
} else if (tso_info->ethproto == htons(ETH_P_IPV6)) {
|
||||
@@ -4026,7 +4026,7 @@ static uint8_t __qdf_nbuf_get_tso_cmn_seg_info(qdf_device_t osdev,
|
||||
* __qdf_nbuf_fill_tso_cmn_seg_info() - Init function for each TSO nbuf segment
|
||||
*
|
||||
* @curr_seg: Segment whose contents are initialized
|
||||
* @tso_cmn_info: Parameters common to all segements
|
||||
* @tso_cmn_info: Parameters common to all segments
|
||||
*
|
||||
* Return: None
|
||||
*/
|
||||
@@ -4413,7 +4413,7 @@ uint32_t __qdf_nbuf_get_tso_num_seg(struct sk_buff *skb)
|
||||
* Remainder non-zero and nr_frags zero implies end of skb data.
|
||||
* In that case, one more tso seg is required to accommodate
|
||||
* remaining data, hence num_segs++. If nr_frags is non-zero,
|
||||
* then remaining data will be accomodated while doing the calculation
|
||||
* then remaining data will be accommodated while doing the calculation
|
||||
* for nr_frags data. Hence, frags_per_tso++.
|
||||
*/
|
||||
if (remainder) {
|
||||
@@ -4453,7 +4453,7 @@ uint32_t __qdf_nbuf_get_tso_num_seg(struct sk_buff *skb)
|
||||
* positive. If frags_per_tso reaches the (max-1),
|
||||
* [First frags always have EIT header, therefore max-1]
|
||||
* increment the num_segs as no more data can be
|
||||
* accomodated in the curr tso seg. Reset the remainder
|
||||
* accommodated in the curr tso seg. Reset the remainder
|
||||
* and frags per tso and keep looping.
|
||||
*/
|
||||
frags_per_tso++;
|
||||
@@ -4650,7 +4650,7 @@ __qdf_nbuf_dmamap_create(qdf_device_t osdev, __qdf_dma_map_t *dmap)
|
||||
{
|
||||
QDF_STATUS error = QDF_STATUS_SUCCESS;
|
||||
/*
|
||||
* driver can tell its SG capablity, it must be handled.
|
||||
* driver can tell its SG capability, it must be handled.
|
||||
* Bounce buffers if they are there
|
||||
*/
|
||||
(*dmap) = kzalloc(sizeof(struct __qdf_dma_map), GFP_KERNEL);
|
||||
@@ -5879,7 +5879,7 @@ qdf_export_symbol(qdf_nbuf_init_fast);
|
||||
|
||||
#ifdef QDF_NBUF_GLOBAL_COUNT
|
||||
/**
|
||||
* __qdf_nbuf_mod_init() - Intialization routine for qdf_nuf
|
||||
* __qdf_nbuf_mod_init() - Initialization routine for qdf_nuf
|
||||
*
|
||||
* Return void
|
||||
*/
|
||||
|
@@ -91,7 +91,7 @@ void __qdf_frag_mod_exit(void)
|
||||
* @p_frag: Pointer to frag
|
||||
* @alloc_func_name: Function where frag is allocated
|
||||
* @alloc_func_line: Allocation function line no.
|
||||
* @refcount: No. of refereces to the frag
|
||||
* @refcount: No. of references to the frag
|
||||
* @last_func_name: Function where frag recently accessed
|
||||
* @last_func_line_num: Line number of last function
|
||||
*
|
||||
|
@@ -1,5 +1,6 @@
|
||||
/*
|
||||
* Copyright (c) 2019-2020 The Linux Foundation. All rights reserved.
|
||||
* Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
|
||||
*
|
||||
* Permission to use, copy, modify, and/or distribute this software for
|
||||
* any purpose with or without fee is hereby granted, provided that the
|
||||
@@ -139,7 +140,7 @@ bool qdf_periodic_work_stop_sync(struct qdf_periodic_work *pwork)
|
||||
|
||||
/* To avoid using a lock, signal that the work shouldn't be restarted,
|
||||
* and cancel_sync in a loop. There is a very small race window, and
|
||||
* thus the work may ocassionally need to be cancelled more than once.
|
||||
* thus the work may occasionally need to be cancelled more than once.
|
||||
*/
|
||||
pwork->msec = 0;
|
||||
while (cancel_delayed_work_sync(&pwork->dwork))
|
||||
|
@@ -44,7 +44,7 @@
|
||||
#include <qdf_defer.h>
|
||||
#include <qdf_module.h>
|
||||
|
||||
/* Function declarations and documenation */
|
||||
/* Function declarations and documentation */
|
||||
|
||||
typedef int (*qdf_thread_os_func)(void *data);
|
||||
|
||||
|
@@ -396,7 +396,7 @@ void qdf_trace(uint8_t module, uint16_t code, uint16_t session, uint32_t data)
|
||||
return;
|
||||
|
||||
qdf_get_time_of_the_day_in_hr_min_sec_usec(time, sizeof(time));
|
||||
/* Aquire the lock so that only one thread at a time can fill the ring
|
||||
/* Acquire the lock so that only one thread at a time can fill the ring
|
||||
* buffer
|
||||
*/
|
||||
spin_lock_irqsave(<race_lock, flags);
|
||||
@@ -508,7 +508,7 @@ qdf_export_symbol(qdf_trace_register);
|
||||
* @session: Session id of log
|
||||
* @count: Number of lines to dump starting from tail to head
|
||||
*
|
||||
* This function will be called up on issueing ioctl call as mentioned following
|
||||
* This function will be called up on issuing ioctl call as mentioned following
|
||||
* [iwpriv wlan0 dumplog 0 0 <n> <bitmask_of_module>]
|
||||
*
|
||||
* <n> - number lines to dump starting from tail to head.
|
||||
@@ -537,7 +537,7 @@ void qdf_trace_dump_all(void *p_mac, uint8_t code, uint8_t session,
|
||||
g_qdf_trace_data.num, g_qdf_trace_data.head,
|
||||
g_qdf_trace_data.tail);
|
||||
|
||||
/* aquire the lock so that only one thread at a time can read
|
||||
/* acquire the lock so that only one thread at a time can read
|
||||
* the ring buffer
|
||||
*/
|
||||
spin_lock(<race_lock);
|
||||
@@ -3336,7 +3336,7 @@ void qdf_dp_trace_dump_all(uint32_t count, uint8_t pdev_id)
|
||||
g_qdf_dp_trace_data.num, g_qdf_dp_trace_data.head,
|
||||
g_qdf_dp_trace_data.tail);
|
||||
|
||||
/* aquire the lock so that only one thread at a time can read
|
||||
/* acquire the lock so that only one thread at a time can read
|
||||
* the ring buffer
|
||||
*/
|
||||
spin_lock_bh(&l_dp_trace_lock);
|
||||
@@ -3697,7 +3697,7 @@ qdf_export_symbol(qdf_rl_print_time_set);
|
||||
void qdf_rl_print_supressed_log(void)
|
||||
{
|
||||
if (qdf_rl_print_supressed) {
|
||||
pr_err("QDF Ratelimiting: %d prints supressed",
|
||||
pr_err("QDF Ratelimiting: %d prints suppressed",
|
||||
qdf_rl_print_supressed);
|
||||
qdf_rl_print_supressed = 0;
|
||||
}
|
||||
|
Reference in New Issue
Block a user