| /* |
| * Copyright (c) 2013-2017 The Linux Foundation. All rights reserved. |
| * |
| * Previously licensed under the ISC license by Qualcomm Atheros, Inc. |
| * |
| * |
| * Permission to use, copy, modify, and/or distribute this software for |
| * any purpose with or without fee is hereby granted, provided that the |
| * above copyright notice and this permission notice appear in all |
| * copies. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL |
| * WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED |
| * WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE |
| * AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL |
| * DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR |
| * PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER |
| * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR |
| * PERFORMANCE OF THIS SOFTWARE. |
| */ |
| |
| /* |
| * This file was originally distributed by Qualcomm Atheros, Inc. |
| * under proprietary terms before Copyright ownership was assigned |
| * to the Linux Foundation. |
| */ |
| /*======================================================================== |
| |
| \file wlan_hdd_ipa.c |
| |
| \brief WLAN HDD and ipa interface implementation |
| |
| ========================================================================*/ |
| |
| /*-------------------------------------------------------------------------- |
| Include Files |
| ------------------------------------------------------------------------*/ |
| #ifdef IPA_OFFLOAD |
| #include <wlan_hdd_includes.h> |
| #include <wlan_hdd_ipa.h> |
| |
| #include <linux/etherdevice.h> |
| #include <linux/atomic.h> |
| #include <linux/netdevice.h> |
| #include <linux/skbuff.h> |
| #include <linux/list.h> |
| #include <linux/debugfs.h> |
| #include <linux/inetdevice.h> |
| #include <linux/ip.h> |
| #include <wlan_hdd_softap_tx_rx.h> |
| |
| #include "vos_sched.h" |
| #include "tl_shim.h" |
| #include "wlan_qct_tl.h" |
| |
| #ifdef IPA_UC_OFFLOAD |
| #include "wma.h" |
| #include "wma_api.h" |
| #include "wal_rx_desc.h" |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| #define HDD_IPA_DESC_BUFFER_RATIO 4 |
| #define HDD_IPA_IPV4_NAME_EXT "_ipv4" |
| #define HDD_IPA_IPV6_NAME_EXT "_ipv6" |
| |
| #define HDD_IPA_RX_INACTIVITY_MSEC_DELAY 1000 |
| #ifdef IPA_UC_OFFLOAD |
| #define HDD_IPA_UC_WLAN_HDR_DES_MAC_OFFSET 12 |
| #define HDD_IPA_UC_WLAN_8023_HDR_SIZE 14 |
| /* WDI TX and RX PIPE */ |
| #define HDD_IPA_UC_NUM_WDI_PIPE 2 |
| #define HDD_IPA_UC_MAX_PENDING_EVENT 33 |
| |
| #define HDD_IPA_UC_DEBUG_DUMMY_MEM_SIZE 32000 |
| #define HDD_IPA_UC_RT_DEBUG_PERIOD 300 |
| #define HDD_IPA_UC_RT_DEBUG_BUF_COUNT 30 |
| #define HDD_IPA_UC_RT_DEBUG_FILL_INTERVAL 10000 |
| |
| #define MAX_PENDING_EVENT_COUNT 20 |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| #ifdef IPA_UC_OFFLOAD |
| typedef enum { |
| HDD_IPA_UC_OPCODE_TX_SUSPEND = 0, |
| HDD_IPA_UC_OPCODE_TX_RESUME = 1, |
| HDD_IPA_UC_OPCODE_RX_SUSPEND = 2, |
| HDD_IPA_UC_OPCODE_RX_RESUME = 3, |
| HDD_IPA_UC_OPCODE_STATS = 4, |
| HDD_IPA_UC_OPCODE_UC_READY = 5, |
| /* keep this last */ |
| HDD_IPA_UC_OPCODE_MAX |
| } hdd_ipa_uc_op_code; |
| |
| typedef enum { |
| HDD_IPA_UC_STAT_REASON_NONE, |
| HDD_IPA_UC_STAT_REASON_DEBUG, |
| HDD_IPA_UC_STAT_REASON_BW_CAL, |
| HDD_IPA_UC_STAT_REASON_DUMP_INFO |
| } hdd_ipa_uc_stat_reason; |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| struct llc_snap_hdr { |
| uint8_t dsap; |
| uint8_t ssap; |
| uint8_t resv[4]; |
| __be16 eth_type; |
| } __packed; |
| |
| struct hdd_ipa_tx_hdr { |
| struct ethhdr eth; |
| struct llc_snap_hdr llc_snap; |
| } __packed; |
| |
| /* For Tx pipes, use 802.3 Header format */ |
| static struct hdd_ipa_tx_hdr ipa_tx_hdr = { |
| { |
| {0xDE, 0xAD, 0xBE, 0xEF, 0xFF, 0xFF}, |
| {0xDE, 0xAD, 0xBE, 0xEF, 0xFF, 0xFF}, |
| 0x00 /* length can be zero */ |
| }, |
| { |
| /* LLC SNAP header 8 bytes */ |
| 0xaa, 0xaa, |
| {0x03, 0x00, 0x00, 0x00}, |
| 0x0008 /* type value(2 bytes) ,filled by wlan */ |
| /* 0x0800 - IPV4, 0x86dd - IPV6 */ |
| } |
| }; |
| |
| #ifdef IPA_UC_OFFLOAD |
| struct frag_header { |
| uint32 |
| length:16, /* length field is LSB of the FRAG DESC */ |
| reserved16:16; |
| uint32 reserved32; |
| } __packed; |
| |
| struct ipa_header { |
| uint32 |
| vdev_id:8, /* vdev_id field is LSB of IPA DESC */ |
| reserved:24; |
| } __packed; |
| |
| struct hdd_ipa_uc_tx_hdr { |
| struct frag_header frag_hd; |
| struct ipa_header ipa_hd; |
| struct ethhdr eth; |
| } __packed; |
| |
| /* For Tx pipes, use Ethernet-II Header format */ |
| struct hdd_ipa_uc_tx_hdr ipa_uc_tx_hdr = { |
| { |
| 0x00000000, |
| 0x00000000 |
| }, |
| { |
| 0x00000000 |
| }, |
| { |
| {0x00, 0x03, 0x7f, 0xaa, 0xbb, 0xcc}, |
| {0x00, 0x03, 0x7f, 0xdd, 0xee, 0xff}, |
| 0x0008 |
| } |
| }; |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| /* |
| +----------+----------+--------------+--------+ |
| | Reserved | QCMAP ID | interface id | STA ID | |
| +----------+----------+--------------+--------+ |
| */ |
| struct hdd_ipa_cld_hdr { |
| uint8_t reserved[2]; |
| uint8_t iface_id; |
| uint8_t sta_id; |
| } __packed; |
| |
| struct hdd_ipa_rx_hdr { |
| struct hdd_ipa_cld_hdr cld_hdr; |
| struct ethhdr eth; |
| } __packed; |
| |
| struct hdd_ipa_pm_tx_cb { |
| struct hdd_ipa_iface_context *iface_context; |
| struct ipa_rx_data *ipa_tx_desc; |
| }; |
| |
| #ifdef IPA_UC_OFFLOAD |
| struct hdd_ipa_uc_rx_hdr { |
| struct ethhdr eth; |
| } __packed; |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| #define HDD_IPA_WLAN_CLD_HDR_LEN sizeof(struct hdd_ipa_cld_hdr) |
| #ifdef IPA_UC_OFFLOAD |
| #define HDD_IPA_UC_WLAN_CLD_HDR_LEN 0 |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| #define HDD_IPA_WLAN_TX_HDR_LEN sizeof(ipa_tx_hdr) |
| #ifdef IPA_UC_OFFLOAD |
| #define HDD_IPA_UC_WLAN_TX_HDR_LEN sizeof(ipa_uc_tx_hdr) |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| #define HDD_IPA_WLAN_RX_HDR_LEN sizeof(struct hdd_ipa_rx_hdr) |
| #ifdef IPA_UC_OFFLOAD |
| #define HDD_IPA_UC_WLAN_RX_HDR_LEN sizeof(struct hdd_ipa_uc_rx_hdr) |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| #define HDD_IPA_WLAN_HDR_DES_MAC_OFFSET 0 |
| |
| #define HDD_IPA_GET_IFACE_ID(_data) \ |
| (((struct hdd_ipa_cld_hdr *) (_data))->iface_id) |
| |
| |
| #define HDD_IPA_LOG(LVL, fmt, args...) VOS_TRACE(VOS_MODULE_ID_HDD, LVL, \ |
| "%s:%d: "fmt, __func__, __LINE__, ## args) |
| #define HDD_IPA_DP_LOG(LVL, fmt, args...) VOS_TRACE(VOS_MODULE_ID_HDD_DATA, LVL, \ |
| "%s:%d: "fmt, __func__, __LINE__, ## args) |
| |
| |
| #define HDD_IPA_DBG_DUMP(_lvl, _prefix, _buf, _len) \ |
| do {\ |
| VOS_TRACE(VOS_MODULE_ID_HDD, _lvl, "%s:", _prefix); \ |
| VOS_TRACE_HEX_DUMP(VOS_MODULE_ID_HDD, _lvl, _buf, _len); \ |
| } while(0) |
| |
| #define DBG_DUMP_RX_LEN 32 |
| #define DBG_DUMP_TX_LEN 48 |
| |
| enum hdd_ipa_rm_state { |
| HDD_IPA_RM_RELEASED, |
| HDD_IPA_RM_GRANT_PENDING, |
| HDD_IPA_RM_GRANTED, |
| }; |
| |
| #define HDD_IPA_MAX_IFACE 3 |
| #define HDD_IPA_MAX_SYSBAM_PIPE 4 |
| #define HDD_IPA_RX_PIPE HDD_IPA_MAX_IFACE |
| |
| static struct hdd_ipa_adapter_2_client { |
| enum ipa_client_type cons_client; |
| enum ipa_client_type prod_client; |
| } hdd_ipa_adapter_2_client[HDD_IPA_MAX_IFACE] = { |
| #ifdef IPA_UC_OFFLOAD |
| {IPA_CLIENT_WLAN2_CONS, IPA_CLIENT_WLAN1_PROD}, |
| {IPA_CLIENT_WLAN3_CONS, IPA_CLIENT_WLAN1_PROD}, |
| {IPA_CLIENT_WLAN4_CONS, IPA_CLIENT_WLAN1_PROD}, |
| #else |
| {IPA_CLIENT_WLAN1_CONS, IPA_CLIENT_WLAN1_PROD}, |
| {IPA_CLIENT_WLAN2_CONS, IPA_CLIENT_WLAN1_PROD}, |
| {IPA_CLIENT_WLAN3_CONS, IPA_CLIENT_WLAN1_PROD}, |
| #endif |
| }; |
| |
| struct hdd_ipa_sys_pipe { |
| uint32_t conn_hdl; |
| uint8_t conn_hdl_valid; |
| struct ipa_sys_connect_params ipa_sys_params; |
| }; |
| |
| struct hdd_ipa_iface_stats { |
| uint64_t num_tx; |
| uint64_t num_tx_drop; |
| uint64_t num_tx_err; |
| uint64_t num_tx_cac_drop; |
| uint64_t num_rx_prefilter; |
| uint64_t num_rx_ipa_excep; |
| uint64_t num_rx_recv; |
| uint64_t num_rx_recv_mul; |
| uint64_t num_rx_send_desc_err; |
| uint64_t max_rx_mul; |
| }; |
| |
| struct hdd_ipa_priv; |
| |
| struct hdd_ipa_iface_context { |
| struct hdd_ipa_priv *hdd_ipa; |
| hdd_adapter_t *adapter; |
| void *tl_context; |
| |
| enum ipa_client_type cons_client; |
| enum ipa_client_type prod_client; |
| |
| uint8_t iface_id; /* This iface ID */ |
| uint8_t sta_id; /* This iface station ID */ |
| adf_os_spinlock_t interface_lock; |
| uint32_t ifa_address; |
| struct hdd_ipa_iface_stats stats; |
| }; |
| |
| |
| struct hdd_ipa_stats { |
| uint32_t event[IPA_WLAN_EVENT_MAX]; |
| uint64_t num_send_msg; |
| uint64_t num_free_msg; |
| |
| uint64_t num_rm_grant; |
| uint64_t num_rm_release; |
| uint64_t num_rm_grant_imm; |
| uint64_t num_cons_perf_req; |
| uint64_t num_prod_perf_req; |
| |
| uint64_t num_rx_drop; |
| uint64_t num_rx_ipa_tx_dp; |
| uint64_t num_rx_ipa_splice; |
| uint64_t num_rx_ipa_loop; |
| uint64_t num_rx_ipa_tx_dp_err; |
| uint64_t num_rx_ipa_write_done; |
| uint64_t num_max_ipa_tx_mul; |
| uint64_t num_rx_ipa_hw_maxed_out; |
| uint64_t max_pend_q_cnt; |
| |
| uint64_t num_tx_comp_cnt; |
| uint64_t num_tx_queued; |
| uint64_t num_tx_dequeued; |
| uint64_t num_max_pm_queue; |
| |
| uint64_t num_freeq_empty; |
| uint64_t num_pri_freeq_empty; |
| uint64_t num_rx_excep; |
| uint64_t num_tx_fwd_ok; |
| uint64_t num_tx_fwd_err; |
| }; |
| |
| #ifdef IPA_UC_OFFLOAD |
| struct ipa_uc_stas_map { |
| v_BOOL_t is_reserved; |
| uint8_t sta_id; |
| }; |
| |
| struct op_msg_type { |
| uint8_t msg_t; |
| uint8_t rsvd; |
| uint16_t op_code; |
| uint16_t len; |
| uint16_t rsvd_snd; |
| }; |
| |
| struct ipa_uc_fw_stats { |
| uint32_t tx_comp_ring_base; |
| uint32_t tx_comp_ring_size; |
| uint32_t tx_comp_ring_dbell_addr; |
| uint32_t tx_comp_ring_dbell_ind_val; |
| uint32_t tx_comp_ring_dbell_cached_val; |
| uint32_t tx_pkts_enqueued; |
| uint32_t tx_pkts_completed; |
| uint32_t tx_is_suspend; |
| uint32_t tx_reserved; |
| uint32_t rx_ind_ring_base; |
| uint32_t rx_ind_ring_size; |
| uint32_t rx_ind_ring_dbell_addr; |
| uint32_t rx_ind_ring_dbell_ind_val; |
| uint32_t rx_ind_ring_dbell_ind_cached_val; |
| uint32_t rx_ind_ring_rdidx_addr; |
| uint32_t rx_ind_ring_rd_idx_cached_val; |
| uint32_t rx_refill_idx; |
| uint32_t rx_num_pkts_indicated; |
| uint32_t rx_buf_refilled; |
| uint32_t rx_num_ind_drop_no_space; |
| uint32_t rx_num_ind_drop_no_buf; |
| uint32_t rx_is_suspend; |
| uint32_t rx_reserved; |
| }; |
| |
| struct ipa_uc_pending_event { |
| vos_list_node_t node; |
| hdd_adapter_t *adapter; |
| enum ipa_wlan_event type; |
| uint8_t sta_id; |
| uint8_t mac_addr[VOS_MAC_ADDR_SIZE]; |
| }; |
| |
| static const char *op_string[HDD_IPA_UC_OPCODE_MAX] = { |
| "TX_SUSPEND", |
| "TX_RESUME", |
| "RX_SUSPEND", |
| "RX_RESUME", |
| "STATS", |
| "OPCODE_MAX" |
| }; |
| |
| struct uc_rm_work_struct { |
| struct work_struct work; |
| enum ipa_rm_event event; |
| }; |
| |
| struct uc_op_work_struct { |
| struct work_struct work; |
| struct op_msg_type *msg; |
| }; |
| static uint8_t vdev_to_iface[CSR_ROAM_SESSION_MAX]; |
| |
| struct uc_rt_debug_info { |
| v_TIME_t time; |
| uint64_t ipa_excp_count; |
| uint64_t rx_drop_count; |
| uint64_t net_sent_count; |
| uint64_t rx_discard_count; |
| uint64_t tx_fwd_ok_count; |
| uint64_t tx_fwd_count; |
| uint64_t rx_destructor_call; |
| }; |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| struct hdd_ipa_priv { |
| struct hdd_ipa_sys_pipe sys_pipe[HDD_IPA_MAX_SYSBAM_PIPE]; |
| struct hdd_ipa_iface_context iface_context[HDD_IPA_MAX_IFACE]; |
| uint8_t num_iface; |
| enum hdd_ipa_rm_state rm_state; |
| /* |
| * IPA driver can send RM notifications with IRQ disabled so using adf |
| * APIs as it is taken care gracefully. Without this, kernel would throw |
| * an warning if spin_lock_bh is used while IRQ is disabled |
| */ |
| adf_os_spinlock_t rm_lock; |
| struct work_struct rm_work; |
| #ifdef IPA_UC_OFFLOAD |
| struct uc_rm_work_struct uc_rm_work; |
| struct uc_op_work_struct uc_op_work[HDD_IPA_UC_OPCODE_MAX]; |
| #endif |
| vos_wake_lock_t wake_lock; |
| struct delayed_work wake_lock_work; |
| bool wake_lock_released; |
| |
| enum ipa_client_type prod_client; |
| |
| atomic_t tx_ref_cnt; |
| adf_nbuf_queue_t pm_queue_head; |
| struct work_struct pm_work; |
| adf_os_spinlock_t pm_lock; |
| bool suspended; |
| |
| uint32_t pending_hw_desc_cnt; |
| uint32_t hw_desc_cnt; |
| spinlock_t q_lock; |
| uint32_t freeq_cnt; |
| struct list_head free_desc_head; |
| |
| uint32_t pend_q_cnt; |
| struct list_head pend_desc_head; |
| |
| hdd_context_t *hdd_ctx; |
| |
| struct dentry *debugfs_dir; |
| struct hdd_ipa_stats stats; |
| |
| struct notifier_block ipv4_notifier; |
| uint32_t curr_prod_bw; |
| uint32_t curr_cons_bw; |
| |
| #ifdef IPA_UC_OFFLOAD |
| uint8_t activated_fw_pipe; |
| uint8_t sap_num_connected_sta; |
| #ifdef IPA_UC_STA_OFFLOAD |
| uint8_t sta_connected; |
| #endif |
| uint32_t tx_pipe_handle; |
| uint32_t rx_pipe_handle; |
| v_BOOL_t resource_loading; |
| v_BOOL_t resource_unloading; |
| v_BOOL_t pending_cons_req; |
| struct ipa_uc_stas_map assoc_stas_map[WLAN_MAX_STA_COUNT]; |
| vos_list_t pending_event; |
| vos_lock_t event_lock; |
| uint32_t ipa_tx_packets_diff; |
| uint32_t ipa_rx_packets_diff; |
| uint32_t ipa_p_tx_packets; |
| uint32_t ipa_p_rx_packets; |
| uint64_t ipa_tx_forward; |
| uint64_t ipa_rx_discard; |
| uint64_t ipa_rx_net_send_count; |
| uint64_t ipa_rx_internel_drop_count; |
| uint64_t ipa_rx_destructor_count; |
| hdd_ipa_uc_stat_reason stat_req_reason; |
| struct ipa_wdi_in_params cons_pipe_in; |
| struct ipa_wdi_in_params prod_pipe_in; |
| v_BOOL_t uc_loaded; |
| v_BOOL_t wdi_enabled; |
| bool ipa_pipes_down; |
| vos_timer_t rt_debug_timer; |
| struct uc_rt_debug_info rt_bug_buffer[HDD_IPA_UC_RT_DEBUG_BUF_COUNT]; |
| unsigned int rt_buf_fill_index; |
| vos_timer_t rt_debug_fill_timer; |
| vos_lock_t rt_debug_lock; |
| vos_lock_t ipa_lock; |
| #endif /* IPA_UC_OFFLOAD */ |
| }; |
| |
| static struct hdd_ipa_priv *ghdd_ipa; |
| |
| #define HDD_IPA_ENABLE_MASK BIT(0) |
| #define HDD_IPA_PRE_FILTER_ENABLE_MASK BIT(1) |
| #define HDD_IPA_IPV6_ENABLE_MASK BIT(2) |
| #define HDD_IPA_RM_ENABLE_MASK BIT(3) |
| #define HDD_IPA_CLK_SCALING_ENABLE_MASK BIT(4) |
| #define HDD_IPA_REAL_TIME_DEBUGGING BIT(8) |
| |
| #define HDD_IPA_IS_CONFIG_ENABLED(_hdd_ctx, _mask)\ |
| (((_hdd_ctx)->cfg_ini->IpaConfig & (_mask)) == (_mask)) |
| |
| #ifdef IPA_UC_OFFLOAD |
| #define HDD_IPA_INCREASE_INTERNAL_DROP_COUNT(hdd_ipa) \ |
| hdd_ipa->ipa_rx_internel_drop_count++ |
| #define HDD_IPA_INCREASE_NET_SEND_COUNT(hdd_ipa) \ |
| hdd_ipa->ipa_rx_net_send_count++ |
| #else |
| #define HDD_IPA_INCREASE_INTERNAL_DROP_COUNT(hdd_ipa) {} |
| #define HDD_IPA_INCREASE_NET_SEND_COUNT(hdd_ipa) {} |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| /* Local Function Prototypes */ |
| static void hdd_ipa_i2w_cb(void *priv, enum ipa_dp_evt_type evt, |
| unsigned long data); |
| static void hdd_ipa_w2i_cb(void *priv, enum ipa_dp_evt_type evt, |
| unsigned long data); |
| static void hdd_ipa_msg_free_fn(void *buff, uint32_t len, uint32_t type); |
| |
| #ifdef IPA_UC_OFFLOAD |
| extern int process_wma_set_command(int sessid, int paramid, |
| int sval, int vpdev); |
| #endif /* IPA_UC_OFFLOAD */ |
| static void hdd_ipa_cleanup_iface(struct hdd_ipa_iface_context *iface_context); |
| |
| bool hdd_ipa_is_enabled(hdd_context_t *hdd_ctx) |
| { |
| return HDD_IPA_IS_CONFIG_ENABLED(hdd_ctx, HDD_IPA_ENABLE_MASK); |
| } |
| |
| static inline bool hdd_ipa_uc_is_enabled(struct hdd_ipa_priv *hdd_ipa) |
| { |
| #ifdef IPA_UC_OFFLOAD |
| return (hdd_ipa && hdd_ipa->hdd_ctx->cfg_ini->IpaUcOffloadEnabled); |
| #else |
| return false; |
| #endif /* IPA_UC_OFFLOAD */ |
| } |
| |
| static inline bool hdd_ipa_uc_sta_is_enabled(struct hdd_ipa_priv *hdd_ipa) |
| { |
| #ifdef IPA_UC_STA_OFFLOAD |
| return (hdd_ipa_uc_is_enabled(hdd_ipa) && |
| hdd_ipa->hdd_ctx->cfg_ini->ipa_uc_sta_offload); |
| #else |
| return false; |
| #endif /* IPA_UC_STA_OFFLOAD */ |
| } |
| |
| #ifdef IPA_UC_STA_OFFLOAD |
| static inline void hdd_ipa_uc_sta_reset_sta_connected( |
| struct hdd_ipa_priv *hdd_ipa) |
| { |
| vos_lock_acquire(&hdd_ipa->event_lock); |
| hdd_ipa->sta_connected = 0; |
| vos_lock_release(&hdd_ipa->event_lock); |
| } |
| #else |
| static inline void hdd_ipa_uc_sta_reset_sta_connected( |
| struct hdd_ipa_priv *hdd_ipa) |
| { |
| } |
| #endif |
| |
| static inline bool hdd_ipa_is_pre_filter_enabled(struct hdd_ipa_priv *hdd_ipa) |
| { |
| hdd_context_t *hdd_ctx = hdd_ipa->hdd_ctx; |
| return HDD_IPA_IS_CONFIG_ENABLED(hdd_ctx, HDD_IPA_PRE_FILTER_ENABLE_MASK); |
| } |
| |
| static inline bool hdd_ipa_is_ipv6_enabled(struct hdd_ipa_priv *hdd_ipa) |
| { |
| hdd_context_t *hdd_ctx = hdd_ipa->hdd_ctx; |
| return HDD_IPA_IS_CONFIG_ENABLED(hdd_ctx, HDD_IPA_IPV6_ENABLE_MASK); |
| } |
| |
| static inline bool hdd_ipa_is_rm_enabled(struct hdd_ipa_priv *hdd_ipa) |
| { |
| hdd_context_t *hdd_ctx = hdd_ipa->hdd_ctx; |
| return HDD_IPA_IS_CONFIG_ENABLED(hdd_ctx, HDD_IPA_RM_ENABLE_MASK); |
| } |
| |
| static inline bool hdd_ipa_is_clk_scaling_enabled(struct hdd_ipa_priv *hdd_ipa) |
| { |
| hdd_context_t *hdd_ctx = hdd_ipa->hdd_ctx; |
| return HDD_IPA_IS_CONFIG_ENABLED(hdd_ctx, |
| HDD_IPA_CLK_SCALING_ENABLE_MASK | |
| HDD_IPA_RM_ENABLE_MASK); |
| } |
| |
| static inline bool hdd_ipa_is_rt_debugging_enabled(hdd_context_t *hdd_ctx) |
| { |
| return HDD_IPA_IS_CONFIG_ENABLED(hdd_ctx, HDD_IPA_REAL_TIME_DEBUGGING); |
| } |
| |
| static struct ipa_tx_data_desc *hdd_ipa_alloc_data_desc( |
| struct hdd_ipa_priv *hdd_ipa, int priority) |
| { |
| struct ipa_tx_data_desc *desc = NULL; |
| |
| spin_lock_bh(&hdd_ipa->q_lock); |
| |
| /* Keep the descriptors for priority alloc which can be used for |
| * anchor nodes |
| */ |
| if (hdd_ipa->freeq_cnt < (HDD_IPA_DESC_BUFFER_RATIO * 2) && !priority) { |
| hdd_ipa->stats.num_freeq_empty++; |
| goto end; |
| } |
| |
| if (!list_empty(&hdd_ipa->free_desc_head)) { |
| desc = list_first_entry(&hdd_ipa->free_desc_head, |
| struct ipa_tx_data_desc, link); |
| list_del(&desc->link); |
| hdd_ipa->freeq_cnt--; |
| } else { |
| hdd_ipa->stats.num_pri_freeq_empty++; |
| } |
| |
| end: |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| |
| return desc; |
| } |
| |
| static void hdd_ipa_free_data_desc(struct hdd_ipa_priv *hdd_ipa, |
| struct ipa_tx_data_desc *desc) |
| { |
| desc->priv = NULL; |
| desc->pyld_buffer = NULL; |
| desc->pyld_len = 0; |
| spin_lock_bh(&hdd_ipa->q_lock); |
| list_add_tail(&desc->link, &hdd_ipa->free_desc_head); |
| hdd_ipa->freeq_cnt++; |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| } |
| |
| static struct iphdr * hdd_ipa_get_ip_pkt(void *data, uint16_t *eth_type) |
| { |
| struct ethhdr *eth = (struct ethhdr *)data; |
| struct llc_snap_hdr *ls_hdr; |
| struct iphdr *ip_hdr; |
| |
| ip_hdr = NULL; |
| *eth_type = be16_to_cpu(eth->h_proto); |
| if (*eth_type < 0x600) { |
| /* Non Ethernet II framing format */ |
| ls_hdr = (struct llc_snap_hdr *)((uint8_t *)data + |
| sizeof(struct ethhdr)); |
| |
| if (((ls_hdr->dsap == 0xAA) && (ls_hdr->ssap == 0xAA)) || |
| ((ls_hdr->dsap == 0xAB) && (ls_hdr->ssap == 0xAB))) |
| *eth_type = be16_to_cpu(ls_hdr->eth_type); |
| ip_hdr = (struct iphdr *)((uint8_t *)data + |
| sizeof(struct ethhdr) + sizeof(struct llc_snap_hdr)); |
| } else if (*eth_type == ETH_P_IP) { |
| ip_hdr = (struct iphdr *)((uint8_t *)data + |
| sizeof(struct ethhdr)); |
| } |
| |
| return ip_hdr; |
| } |
| |
| static bool hdd_ipa_can_send_to_ipa(hdd_adapter_t *adapter, struct hdd_ipa_priv *hdd_ipa, void *data) |
| { |
| uint16_t eth_type; |
| struct iphdr *ip_hdr = NULL; |
| |
| if (!hdd_ipa_is_pre_filter_enabled(hdd_ipa)) |
| return true; |
| ip_hdr = hdd_ipa_get_ip_pkt(data, ð_type); |
| |
| /* Check if the dest IP address is itself, then bypass IPA */ |
| if (eth_type == ETH_P_IP) { |
| if (ip_hdr->daddr != ((struct hdd_ipa_iface_context *)(adapter->ipa_context))->ifa_address) |
| return true; |
| else |
| return false; |
| } |
| |
| if (hdd_ipa_is_ipv6_enabled(hdd_ipa) && eth_type == ETH_P_IPV6) |
| return true; |
| |
| return false; |
| } |
| |
| #ifdef IPA_UC_OFFLOAD |
| /** |
| * hdd_ipa_uc_rt_debug_host_fill() - fill rt debug buffer |
| * @ctext: pointer to hdd context. |
| * |
| * If rt debug enabled, periodically called, and fill debug buffer |
| * |
| * Return: none |
| */ |
| static void hdd_ipa_uc_rt_debug_host_fill(void *ctext) |
| { |
| hdd_context_t *hdd_ctx = ctext; |
| struct hdd_ipa_priv *hdd_ipa; |
| struct uc_rt_debug_info *dump_info = NULL; |
| |
| if (wlan_hdd_validate_context(hdd_ctx)) |
| return; |
| |
| if (!hdd_ctx->hdd_ipa || |
| !hdd_ipa_uc_is_enabled((struct hdd_ipa_priv *)hdd_ctx->hdd_ipa)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: IPA UC is not enabled", __func__); |
| return; |
| } |
| |
| hdd_ipa = (struct hdd_ipa_priv *)hdd_ctx->hdd_ipa; |
| |
| vos_lock_acquire(&hdd_ipa->rt_debug_lock); |
| dump_info = &hdd_ipa->rt_bug_buffer[ |
| hdd_ipa->rt_buf_fill_index % HDD_IPA_UC_RT_DEBUG_BUF_COUNT]; |
| if (!dump_info) { |
| vos_lock_release(&hdd_ipa->rt_debug_lock); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: invalid dump pointer", __func__); |
| return; |
| } |
| |
| dump_info->time = vos_timer_get_system_time(); |
| dump_info->ipa_excp_count = hdd_ipa->stats.num_rx_excep; |
| dump_info->rx_drop_count = hdd_ipa->ipa_rx_internel_drop_count; |
| dump_info->net_sent_count = hdd_ipa->ipa_rx_net_send_count; |
| dump_info->tx_fwd_count = hdd_ipa->ipa_tx_forward; |
| dump_info->tx_fwd_ok_count = hdd_ipa->stats.num_tx_fwd_ok; |
| dump_info->rx_discard_count = hdd_ipa->ipa_rx_discard; |
| dump_info->rx_destructor_call = hdd_ipa->ipa_rx_destructor_count; |
| hdd_ipa->rt_buf_fill_index++; |
| vos_lock_release(&hdd_ipa->rt_debug_lock); |
| |
| vos_timer_start(&hdd_ipa->rt_debug_fill_timer, |
| HDD_IPA_UC_RT_DEBUG_FILL_INTERVAL); |
| } |
| |
| /** |
| * hdd_ipa_uc_rt_debug_host_dump() - dump rt debug buffer |
| * @pHddCtx: pointer to hdd context. |
| * |
| * If rt debug enabled, dump debug buffer contents based on requirement |
| * |
| * Return: none |
| */ |
| void hdd_ipa_uc_rt_debug_host_dump(hdd_context_t *hdd_ctx) |
| { |
| struct hdd_ipa_priv *hdd_ipa; |
| unsigned int dump_count; |
| unsigned int dump_index; |
| struct uc_rt_debug_info *dump_info = NULL; |
| |
| if (wlan_hdd_validate_context(hdd_ctx)) |
| return; |
| |
| hdd_ipa = hdd_ctx->hdd_ipa; |
| if (!hdd_ipa || |
| !hdd_ipa_uc_is_enabled((struct hdd_ipa_priv *)hdd_ctx->hdd_ipa)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: IPA UC is not enabled", __func__); |
| return; |
| } |
| |
| pr_err("========= WLAN-IPA DEBUG BUF DUMP ==========\n"); |
| pr_err(" TM : EXEP : DROP : NETS : FWOK : TXFD : DSTR : DSCD\n"); |
| |
| vos_lock_acquire(&hdd_ipa->rt_debug_lock); |
| for (dump_count = 0; |
| dump_count < HDD_IPA_UC_RT_DEBUG_BUF_COUNT; |
| dump_count++) { |
| dump_index = (hdd_ipa->rt_buf_fill_index + dump_count) % |
| HDD_IPA_UC_RT_DEBUG_BUF_COUNT; |
| dump_info = &hdd_ipa->rt_bug_buffer[dump_index]; |
| if ((dump_index > HDD_IPA_UC_RT_DEBUG_BUF_COUNT) || |
| (!dump_info)) { |
| vos_lock_release(&hdd_ipa->rt_debug_lock); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "INVALID"); |
| return; |
| } |
| pr_err("%12lu:%10llu:%10llu:%10llu:%10llu:%10llu:%10llu:%10llu\n", |
| dump_info->time, dump_info->ipa_excp_count, |
| dump_info->rx_drop_count, dump_info->net_sent_count, |
| dump_info->tx_fwd_ok_count, dump_info->tx_fwd_count, |
| dump_info->rx_destructor_call, |
| dump_info->rx_discard_count); |
| } |
| vos_lock_release(&hdd_ipa->rt_debug_lock); |
| pr_err("======= WLAN-IPA DEBUG BUF DUMP END ========\n"); |
| } |
| |
| /** |
| * hdd_ipa_uc_rt_debug_handler() - periodic memory health monitor handler |
| * @ctext: pointer to hdd context. |
| * |
| * periodically called by timer expire |
| * will try to alloc dummy memory and detect out of memory condition |
| * if out of memory detected, dump wlan-ipa stats |
| * |
| * Return: none |
| */ |
| static void hdd_ipa_uc_rt_debug_handler(void *ctext) |
| { |
| hdd_context_t *hdd_ctx= ctext; |
| struct hdd_ipa_priv *hdd_ipa = hdd_ctx->hdd_ipa; |
| void *dummy_ptr; |
| |
| if (wlan_hdd_validate_context(hdd_ctx)) |
| return; |
| |
| if (!hdd_ipa_is_rt_debugging_enabled(hdd_ctx)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: IPA RT debug is not enabled", __func__); |
| return; |
| } |
| |
| /* |
| * Allocate dummy buffer periodically and free immediately |
| * This will proactively detect OOM condition |
| * And if allocation fail, will dump WLAN IPA stats |
| */ |
| dummy_ptr = kmalloc(HDD_IPA_UC_DEBUG_DUMMY_MEM_SIZE, |
| GFP_KERNEL | GFP_ATOMIC); |
| if (!dummy_ptr) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_FATAL, |
| "%s: Dummy alloc fail", __func__); |
| hdd_ipa_uc_rt_debug_host_dump(hdd_ctx); |
| hdd_ipa_uc_stat_request( |
| hdd_get_adapter(hdd_ctx, WLAN_HDD_SOFTAP), 1); |
| } else { |
| kfree(dummy_ptr); |
| } |
| |
| vos_timer_start(&hdd_ipa->rt_debug_timer, |
| HDD_IPA_UC_RT_DEBUG_PERIOD); |
| } |
| |
| /** |
| * hdd_ipa_uc_rt_debug_destructor() - called by data packet free |
| * @skb: packet pinter |
| * |
| * when free data packet, will be invoked by wlan client and will increase |
| * free counter |
| * |
| * Return: none |
| */ |
| void hdd_ipa_uc_rt_debug_destructor(struct sk_buff *skb) |
| { |
| if (!ghdd_ipa) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: invalid hdd context", __func__); |
| return; |
| } |
| |
| ghdd_ipa->ipa_rx_destructor_count++; |
| } |
| |
| /** |
| * hdd_ipa_uc_rt_debug_deinit() - remove resources to handle rt debugging |
| * @pHddCtx: hdd main context |
| * |
| * free all rt debugging resources |
| * |
| * Return: none |
| */ |
| static void hdd_ipa_uc_rt_debug_deinit(hdd_context_t *pHddCtx) |
| { |
| struct hdd_ipa_priv *hdd_ipa = pHddCtx->hdd_ipa; |
| |
| if ( VOS_TIMER_STATE_STOPPED != |
| vos_timer_getCurrentState(&hdd_ipa->rt_debug_fill_timer)) { |
| vos_timer_stop(&hdd_ipa->rt_debug_fill_timer); |
| } |
| vos_timer_destroy(&hdd_ipa->rt_debug_fill_timer); |
| vos_lock_destroy(&hdd_ipa->rt_debug_lock); |
| |
| if (!hdd_ipa_is_rt_debugging_enabled(pHddCtx)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: IPA RT debug is not enabled", __func__); |
| return; |
| } |
| |
| if ( VOS_TIMER_STATE_STOPPED != |
| vos_timer_getCurrentState(&hdd_ipa->rt_debug_timer)) { |
| vos_timer_stop(&hdd_ipa->rt_debug_timer); |
| } |
| vos_timer_destroy(&hdd_ipa->rt_debug_timer); |
| } |
| |
| /** |
| * hdd_ipa_uc_rt_debug_init() - intialize resources to handle rt debugging |
| * @pHddCtx: hdd main context |
| * |
| * alloc and initialize all rt debugging resources |
| * |
| * Return: none |
| */ |
| static void hdd_ipa_uc_rt_debug_init(hdd_context_t *hdd_ctx) |
| { |
| struct hdd_ipa_priv *hdd_ipa = hdd_ctx->hdd_ipa; |
| |
| /* Histogram intialize by default */ |
| vos_lock_init(&hdd_ipa->rt_debug_lock); |
| vos_timer_init(&hdd_ipa->rt_debug_fill_timer, VOS_TIMER_TYPE_SW, |
| hdd_ipa_uc_rt_debug_host_fill, (void *)hdd_ctx); |
| |
| hdd_ipa->rt_buf_fill_index = 0; |
| vos_mem_zero(hdd_ipa->rt_bug_buffer, |
| sizeof(struct uc_rt_debug_info) * HDD_IPA_UC_RT_DEBUG_BUF_COUNT); |
| hdd_ipa->ipa_tx_forward = 0; |
| hdd_ipa->ipa_rx_discard = 0; |
| hdd_ipa->ipa_rx_net_send_count = 0; |
| hdd_ipa->ipa_rx_internel_drop_count = 0; |
| hdd_ipa->ipa_rx_destructor_count = 0; |
| |
| vos_timer_start(&hdd_ipa->rt_debug_fill_timer, |
| HDD_IPA_UC_RT_DEBUG_FILL_INTERVAL); |
| |
| /* Realtime debug enable only when feature enabled */ |
| if (!hdd_ipa_is_rt_debugging_enabled(hdd_ctx)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: IPA RT debug is not enabled", __func__); |
| return; |
| } |
| vos_timer_init(&hdd_ipa->rt_debug_timer, VOS_TIMER_TYPE_SW, |
| hdd_ipa_uc_rt_debug_handler, (void *)hdd_ctx); |
| |
| vos_timer_start(&hdd_ipa->rt_debug_timer, |
| HDD_IPA_UC_RT_DEBUG_PERIOD); |
| } |
| |
| /** |
| * hdd_ipa_dump_hdd_ipa() - dump entries in HDD IPA struct |
| * @hdd_ipa: HDD IPA struct |
| * |
| * Dump entries in struct hdd_ipa |
| * |
| * Return: none |
| */ |
| void hdd_ipa_dump_hdd_ipa(struct hdd_ipa_priv *hdd_ipa) |
| { |
| int i; |
| |
| /* HDD IPA */ |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== HDD IPA ====\n" |
| "num_iface: %d\n" |
| "rm_state: %d\n" |
| "rm_lock: %pK\n" |
| "rm_work: %pK\n" |
| "uc_rm_work: %pK\n" |
| "uc_op_work: %pK\n" |
| "wake_lock: %pK\n" |
| "wake_lock_work: %pK\n" |
| "wake_lock_released: %d\n" |
| "prod_client: %d\n" |
| "tx_ref_cnt: %d\n" |
| "pm_queue_head----\n" |
| "\thead: %pK\n" |
| "\ttail: %pK\n" |
| "\tqlen: %d\n" |
| "pm_work: %pK\n" |
| "pm_lock: %pK\n" |
| "suspended: %d\n", |
| hdd_ipa->num_iface, |
| hdd_ipa->rm_state, |
| &hdd_ipa->rm_lock, |
| &hdd_ipa->rm_work, |
| &hdd_ipa->uc_rm_work, |
| &hdd_ipa->uc_op_work, |
| &hdd_ipa->wake_lock, |
| &hdd_ipa->wake_lock_work, |
| hdd_ipa->wake_lock_released, |
| hdd_ipa->prod_client, |
| hdd_ipa->tx_ref_cnt.counter, |
| hdd_ipa->pm_queue_head.head, |
| hdd_ipa->pm_queue_head.tail, |
| hdd_ipa->pm_queue_head.qlen, |
| &hdd_ipa->pm_work, |
| &hdd_ipa->pm_lock, |
| hdd_ipa->suspended); |
| pr_err( |
| "pending_hw_desc_cnt: %d\n" |
| "hw_desc_cnt: %d\n" |
| "q_lock: %pK\n" |
| "freeq_cnt: %d\n" |
| "free_desc_head----\n" |
| "\tnext: %pK\n" |
| "\tprev: %pK\n" |
| "pend_q_cnt: %d\n" |
| "pend_desc_head----\n" |
| "\tnext: %pK\n" |
| "\tprev: %pK\n" |
| "hdd_ctx: %pK\n" |
| "debugfs_dir: %pK\n" |
| "stats: %pK\n" |
| "ipv4_notifier: %pK\n" |
| "curr_prod_bw: %d\n" |
| "curr_cons_bw: %d\n" |
| "activated_fw_pipe: %d\n" |
| "sap_num_connected_sta: %d\n" |
| #ifdef IPA_UC_STA_OFFLOAD |
| "sta_connected: %d\n" |
| #endif |
| , hdd_ipa->pending_hw_desc_cnt, |
| hdd_ipa->hw_desc_cnt, |
| &hdd_ipa->q_lock, |
| hdd_ipa->freeq_cnt, |
| hdd_ipa->free_desc_head.next, |
| hdd_ipa->free_desc_head.prev, |
| hdd_ipa->pend_q_cnt, |
| hdd_ipa->pend_desc_head.next, |
| hdd_ipa->pend_desc_head.prev, |
| hdd_ipa->hdd_ctx, |
| hdd_ipa->debugfs_dir, |
| &hdd_ipa->stats, |
| &hdd_ipa->ipv4_notifier, |
| hdd_ipa->curr_prod_bw, |
| hdd_ipa->curr_cons_bw, |
| hdd_ipa->activated_fw_pipe, |
| hdd_ipa->sap_num_connected_sta |
| #ifdef IPA_UC_STA_OFFLOAD |
| , (unsigned int)hdd_ipa->sta_connected |
| #endif |
| ); |
| pr_err( |
| "tx_pipe_handle: 0x%x\n" |
| "rx_pipe_handle: 0x%x\n" |
| "resource_loading: %d\n" |
| "resource_unloading: %d\n" |
| "pending_cons_req: %d\n" |
| "pending_event----\n" |
| "\tanchor.next: %pK\n" |
| "\tanchor.prev: %pK\n" |
| "\tcount: %d\n" |
| "\tlock: %pK\n" |
| "\tcookie: 0x%x\n" |
| "event_lock: %pK\n" |
| "ipa_tx_packets_diff: %d\n" |
| "ipa_rx_packets_diff: %d\n" |
| "ipa_p_tx_packets: %d\n" |
| "ipa_p_rx_packets: %d\n" |
| "stat_req_reason: %d\n", |
| hdd_ipa->tx_pipe_handle, |
| hdd_ipa->rx_pipe_handle, |
| hdd_ipa->resource_loading, |
| hdd_ipa->resource_unloading, |
| hdd_ipa->pending_cons_req, |
| hdd_ipa->pending_event.anchor.next, |
| hdd_ipa->pending_event.anchor.prev, |
| hdd_ipa->pending_event.count, |
| &hdd_ipa->pending_event.lock, |
| hdd_ipa->pending_event.cookie, |
| &hdd_ipa->event_lock, |
| hdd_ipa->ipa_tx_packets_diff, |
| hdd_ipa->ipa_rx_packets_diff, |
| hdd_ipa->ipa_p_tx_packets, |
| hdd_ipa->ipa_p_rx_packets, |
| hdd_ipa->stat_req_reason); |
| pr_err( |
| "cons_pipe_in----\n" |
| "\tsys: %pK\n" |
| "\tdl.comp_ring_base_pa: 0x%x\n" |
| "\tdl.comp_ring_size: %d\n" |
| "\tdl.ce_ring_base_pa: 0x%x\n" |
| "\tdl.ce_door_bell_pa: 0x%x\n" |
| "\tdl.ce_ring_size: %d\n" |
| "\tdl.num_tx_buffers: %d\n" |
| "prod_pipe_in----\n" |
| "\tsys: %pK\n" |
| "\tul.rdy_ring_base_pa: 0x%x\n" |
| "\tul.rdy_ring_size: %d\n" |
| "\tul.rdy_ring_rp_pa: 0x%x\n" |
| "uc_loaded: %d\n" |
| "wdi_enabled: %d\n", |
| &hdd_ipa->cons_pipe_in.sys, |
| (unsigned int)hdd_ipa->cons_pipe_in.u.dl.comp_ring_base_pa, |
| hdd_ipa->cons_pipe_in.u.dl.comp_ring_size, |
| (unsigned int)hdd_ipa->cons_pipe_in.u.dl.ce_ring_base_pa, |
| (unsigned int)hdd_ipa->cons_pipe_in.u.dl.ce_door_bell_pa, |
| hdd_ipa->cons_pipe_in.u.dl.ce_ring_size, |
| hdd_ipa->cons_pipe_in.u.dl.num_tx_buffers, |
| &hdd_ipa->prod_pipe_in.sys, |
| (unsigned int)hdd_ipa->prod_pipe_in.u.ul.rdy_ring_base_pa, |
| hdd_ipa->prod_pipe_in.u.ul.rdy_ring_size, |
| (unsigned int)hdd_ipa->prod_pipe_in.u.ul.rdy_ring_rp_pa, |
| hdd_ipa->uc_loaded, |
| hdd_ipa->wdi_enabled); |
| |
| pr_err("assoc_stas_map([id]is_reserved/sta_id): "); |
| for (i = 0; i < WLAN_MAX_STA_COUNT; i++) { |
| pr_err( |
| " [%d]%d/%d", |
| i, |
| hdd_ipa->assoc_stas_map[i].is_reserved, |
| hdd_ipa->assoc_stas_map[i].sta_id); |
| } |
| } |
| |
| /** |
| * hdd_ipa_dump_sys_pipe() - dump HDD IPA SYS Pipe struct |
| * @hdd_ipa: HDD IPA struct |
| * |
| * Dump entire struct hdd_ipa_sys_pipe |
| * |
| * Return: none |
| */ |
| void hdd_ipa_dump_sys_pipe(struct hdd_ipa_priv *hdd_ipa) |
| { |
| int i; |
| |
| /* IPA SYS Pipes */ |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "\n==== IPA SYS Pipes ====\n"); |
| |
| for (i = 0; i < HDD_IPA_MAX_SYSBAM_PIPE; i++) { |
| struct hdd_ipa_sys_pipe *sys_pipe; |
| struct ipa_sys_connect_params *ipa_sys_params; |
| |
| sys_pipe = &hdd_ipa->sys_pipe[i]; |
| ipa_sys_params = &sys_pipe->ipa_sys_params; |
| |
| pr_err( |
| "sys_pipe[%d]----\n" |
| "\tconn_hdl: 0x%x\n" |
| "\tconn_hdl_valid: %d\n" |
| "\tnat_en: %d\n" |
| "\thdr_len %d\n" |
| "\thdr_additional_const_len: %d\n" |
| "\thdr_ofst_pkt_size_valid: %d\n" |
| "\thdr_ofst_pkt_size: %d\n" |
| "\thdr_little_endian: %d\n" |
| "\tmode: %d\n" |
| "\tclient: %d\n" |
| "\tdesc_fifo_sz: %d\n" |
| "\tpriv: %pK\n" |
| "\tnotify: %pK\n" |
| "\tskip_ep_cfg: %d\n" |
| "\tkeep_ipa_awake: %d\n", |
| i, |
| sys_pipe->conn_hdl, |
| sys_pipe->conn_hdl_valid, |
| ipa_sys_params->ipa_ep_cfg.nat.nat_en, |
| ipa_sys_params->ipa_ep_cfg.hdr.hdr_len, |
| ipa_sys_params->ipa_ep_cfg.hdr.hdr_additional_const_len, |
| ipa_sys_params->ipa_ep_cfg.hdr.hdr_ofst_pkt_size_valid, |
| ipa_sys_params->ipa_ep_cfg.hdr.hdr_ofst_pkt_size, |
| ipa_sys_params->ipa_ep_cfg.hdr_ext.hdr_little_endian, |
| ipa_sys_params->ipa_ep_cfg.mode.mode, |
| ipa_sys_params->client, |
| ipa_sys_params->desc_fifo_sz, |
| ipa_sys_params->priv, |
| ipa_sys_params->notify, |
| ipa_sys_params->skip_ep_cfg, |
| ipa_sys_params->keep_ipa_awake); |
| } |
| } |
| |
| /** |
| * hdd_ipa_dump_iface_context() - dump HDD IPA Interface Context struct |
| * @hdd_ipa: HDD IPA struct |
| * |
| * Dump entire struct hdd_ipa_iface_context |
| * |
| * Return: none |
| */ |
| void hdd_ipa_dump_iface_context(struct hdd_ipa_priv *hdd_ipa) |
| { |
| int i; |
| |
| /* IPA Interface Contexts */ |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "\n==== IPA Interface Contexts ====\n"); |
| |
| for (i = 0; i < HDD_IPA_MAX_IFACE; i++) { |
| struct hdd_ipa_iface_context *iface_context; |
| |
| iface_context = &hdd_ipa->iface_context[i]; |
| |
| pr_err( |
| "iface_context[%d]----\n" |
| "\thdd_ipa: %pK\n" |
| "\tadapter: %pK\n" |
| "\ttl_context: %pK\n" |
| "\tcons_client: %d\n" |
| "\tprod_client: %d\n" |
| "\tiface_id: %d\n" |
| "\tsta_id: %d\n" |
| "\tinterface_lock: %pK\n" |
| "\tifa_address: 0x%x\n", |
| i, |
| iface_context->hdd_ipa, |
| iface_context->adapter, |
| iface_context->tl_context, |
| iface_context->cons_client, |
| iface_context->prod_client, |
| iface_context->iface_id, |
| iface_context->sta_id, |
| &iface_context->interface_lock, |
| iface_context->ifa_address); |
| } |
| } |
| |
| /** |
| * hdd_ipa_dump_info() - dump HDD IPA struct |
| * @pHddCtx: hdd main context |
| * |
| * Dump entire struct hdd_ipa |
| * |
| * Return: none |
| */ |
| void hdd_ipa_dump_info(hdd_context_t *hdd_ctx) |
| { |
| struct hdd_ipa_priv *hdd_ipa; |
| |
| if (wlan_hdd_validate_context(hdd_ctx)) |
| return; |
| |
| hdd_ipa = (struct hdd_ipa_priv *)hdd_ctx->hdd_ipa; |
| if (!hdd_ipa_is_enabled(hdd_ctx) || |
| !hdd_ipa_uc_is_enabled(hdd_ipa)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "IPA/IPA UC is not enabled, IpaConfig %u,IpaUcOffloadEnabled %u.", |
| hdd_ctx->cfg_ini->IpaConfig, |
| hdd_ctx->cfg_ini->IpaUcOffloadEnabled); |
| return; |
| } |
| |
| hdd_ipa_dump_hdd_ipa(hdd_ipa); |
| hdd_ipa_dump_sys_pipe(hdd_ipa); |
| hdd_ipa_dump_iface_context(hdd_ipa); |
| } |
| |
| void hdd_ipa_uc_stat_query(hdd_context_t *pHddCtx, |
| uint32_t *ipa_tx_diff, uint32_t *ipa_rx_diff) |
| { |
| struct hdd_ipa_priv *hdd_ipa; |
| |
| hdd_ipa = (struct hdd_ipa_priv *)pHddCtx->hdd_ipa; |
| *ipa_tx_diff = 0; |
| *ipa_rx_diff = 0; |
| |
| if (!hdd_ipa_is_enabled(pHddCtx) || |
| !(hdd_ipa_uc_is_enabled(hdd_ipa))) { |
| return; |
| } |
| |
| vos_lock_acquire(&hdd_ipa->ipa_lock); |
| if ((HDD_IPA_UC_NUM_WDI_PIPE == hdd_ipa->activated_fw_pipe) && |
| (VOS_FALSE == hdd_ipa->resource_loading)) { |
| *ipa_tx_diff = hdd_ipa->ipa_tx_packets_diff; |
| *ipa_rx_diff = hdd_ipa->ipa_rx_packets_diff; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_DEBUG, |
| "%s: STAT Query TX DIFF %d, RX DIFF %d", |
| __func__, *ipa_tx_diff, *ipa_rx_diff); |
| } |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| return; |
| } |
| |
| void hdd_ipa_uc_stat_request( hdd_adapter_t *adapter, uint8_t reason) |
| { |
| hdd_context_t *pHddCtx; |
| struct hdd_ipa_priv *hdd_ipa; |
| |
| if (!adapter) { |
| return; |
| } |
| |
| pHddCtx = (hdd_context_t *)adapter->pHddCtx; |
| hdd_ipa = (struct hdd_ipa_priv *)pHddCtx->hdd_ipa; |
| if (!hdd_ipa_is_enabled(pHddCtx) || |
| !(hdd_ipa_uc_is_enabled(hdd_ipa))) { |
| return; |
| } |
| |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_DEBUG, |
| "%s: STAT REQ Reason %d", |
| __func__, reason); |
| vos_lock_acquire(&hdd_ipa->ipa_lock); |
| if ((HDD_IPA_UC_NUM_WDI_PIPE == hdd_ipa->activated_fw_pipe) && |
| (VOS_FALSE == hdd_ipa->resource_loading)) { |
| hdd_ipa->stat_req_reason = (hdd_ipa_uc_stat_reason)reason; |
| process_wma_set_command( |
| (int)adapter->sessionId, |
| (int)WMA_VDEV_TXRX_GET_IPA_UC_FW_STATS_CMDID, |
| 0, VDEV_CMD); |
| } |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| } |
| |
| static v_BOOL_t hdd_ipa_uc_find_add_assoc_sta( |
| struct hdd_ipa_priv *hdd_ipa, |
| v_BOOL_t sta_add, |
| uint8_t sta_id) |
| { |
| /* Found associated sta */ |
| v_BOOL_t sta_found = VOS_FALSE; |
| uint8_t idx; |
| |
| for (idx = 0; idx < WLAN_MAX_STA_COUNT; idx++) { |
| if ((hdd_ipa->assoc_stas_map[idx].is_reserved) && |
| (hdd_ipa->assoc_stas_map[idx].sta_id == sta_id)) { |
| sta_found = VOS_TRUE; |
| break; |
| } |
| } |
| |
| /* Try to add sta which is already in |
| * If the sta is already in, just return sta_found */ |
| if (sta_add && sta_found) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: STA ID %d already exist, cannot add", |
| __func__, sta_id); |
| return sta_found; |
| } |
| |
| if (sta_add) { |
| /* Find first empty slot */ |
| for (idx = 0; idx < WLAN_MAX_STA_COUNT; idx++) { |
| if (!hdd_ipa->assoc_stas_map[idx].is_reserved) { |
| hdd_ipa->assoc_stas_map[idx].is_reserved = |
| VOS_TRUE; |
| hdd_ipa->assoc_stas_map[idx].sta_id = sta_id; |
| return sta_found; |
| } |
| } |
| } |
| |
| /* Delete STA from map, but could not find STA within the map |
| * Error case, add error log */ |
| if (!sta_add && !sta_found) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: STA ID %d does not exist, cannot delete", |
| __func__, sta_id); |
| return sta_found; |
| } |
| |
| if (!sta_add) { |
| for (idx = 0; idx < WLAN_MAX_STA_COUNT; idx++) { |
| if ((hdd_ipa->assoc_stas_map[idx].is_reserved) && |
| (hdd_ipa->assoc_stas_map[idx].sta_id == sta_id)) { |
| hdd_ipa->assoc_stas_map[idx].is_reserved = |
| VOS_FALSE; |
| hdd_ipa->assoc_stas_map[idx].sta_id = 0xFF; |
| return sta_found; |
| } |
| } |
| } |
| |
| return sta_found; |
| } |
| |
| static int hdd_ipa_uc_enable_pipes(struct hdd_ipa_priv *hdd_ipa) |
| { |
| int result; |
| |
| /* ACTIVATE TX PIPE */ |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: Enable TX PIPE(tx_pipe_handle=%d)", |
| __func__, hdd_ipa->tx_pipe_handle); |
| result = ipa_enable_wdi_pipe(hdd_ipa->tx_pipe_handle); |
| if (result) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Enable TX PIPE fail, code %d", |
| __func__, result); |
| return result; |
| } |
| result = ipa_resume_wdi_pipe(hdd_ipa->tx_pipe_handle); |
| if (result) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Resume TX PIPE fail, code %d", |
| __func__, result); |
| return result; |
| } |
| WLANTL_SetUcActive(hdd_ipa->hdd_ctx->pvosContext, |
| VOS_TRUE, VOS_TRUE); |
| |
| /* ACTIVATE RX PIPE */ |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: Enable RX PIPE(rx_pipe_handle=%d)" |
| , __func__, hdd_ipa->rx_pipe_handle); |
| result = ipa_enable_wdi_pipe(hdd_ipa->rx_pipe_handle); |
| if (result) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Enable RX PIPE fail, code %d", |
| __func__, result); |
| return result; |
| } |
| result = ipa_resume_wdi_pipe(hdd_ipa->rx_pipe_handle); |
| if (result) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Resume RX PIPE fail, code %d", |
| __func__, result); |
| return result; |
| } |
| WLANTL_SetUcActive(hdd_ipa->hdd_ctx->pvosContext, |
| VOS_TRUE, VOS_FALSE); |
| |
| hdd_ipa->ipa_pipes_down = false; |
| return 0; |
| } |
| |
| static int hdd_ipa_uc_disable_pipes(struct hdd_ipa_priv *hdd_ipa) |
| { |
| int result; |
| |
| hdd_ipa->ipa_pipes_down = true; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: Disable RX PIPE", __func__); |
| result = ipa_suspend_wdi_pipe(hdd_ipa->rx_pipe_handle); |
| if (result) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Suspend RX PIPE fail, code %d", |
| __func__, result); |
| return result; |
| } |
| result = ipa_disable_wdi_pipe(hdd_ipa->rx_pipe_handle); |
| if (result) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Disable RX PIPE fail, code %d", |
| __func__, result); |
| return result; |
| } |
| |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: Disable TX PIPE", __func__); |
| result = ipa_suspend_wdi_pipe(hdd_ipa->tx_pipe_handle); |
| if (result) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Suspend TX PIPE fail, code %d", |
| __func__, result); |
| return result; |
| } |
| result = ipa_disable_wdi_pipe(hdd_ipa->tx_pipe_handle); |
| if (result) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Disable TX PIPE fail, code %d", |
| __func__, result); |
| return result; |
| } |
| |
| return 0; |
| } |
| |
| static int hdd_ipa_uc_handle_first_con(struct hdd_ipa_priv *hdd_ipa) |
| { |
| hdd_ipa->activated_fw_pipe = 0; |
| hdd_ipa->resource_loading = VOS_TRUE; |
| |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, "+%s", __func__); |
| |
| /* If RM feature enabled |
| * Request PROD Resource first |
| * PROD resource may return sync or async manners */ |
| if (hdd_ipa_is_rm_enabled(hdd_ipa)) { |
| if (!ipa_rm_request_resource(IPA_RM_RESOURCE_WLAN_PROD)) { |
| /* RM PROD request sync return |
| * enable pipe immediately */ |
| if (hdd_ipa_uc_enable_pipes(hdd_ipa)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: IPA WDI Pipes activate fail", __func__); |
| hdd_ipa->resource_loading = VOS_FALSE; |
| return -EBUSY; |
| } |
| } |
| } else { |
| /* RM Disabled |
| * Just enabled all the PIPEs */ |
| if (hdd_ipa_uc_enable_pipes(hdd_ipa)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: IPA WDI Pipes activate fail", __func__); |
| hdd_ipa->resource_loading = VOS_FALSE; |
| return -EBUSY; |
| } |
| } |
| return 0; |
| } |
| |
| static int hdd_ipa_uc_handle_last_discon(struct hdd_ipa_priv *hdd_ipa) |
| { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, "+%s", __func__); |
| |
| hdd_ipa->resource_unloading = VOS_TRUE; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: Disable FW RX PIPE", __func__); |
| WLANTL_SetUcActive(hdd_ipa->hdd_ctx->pvosContext, |
| VOS_FALSE, VOS_FALSE); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: Disable FW TX PIPE", __func__); |
| WLANTL_SetUcActive(hdd_ipa->hdd_ctx->pvosContext, |
| VOS_FALSE, VOS_TRUE); |
| return 0; |
| } |
| |
| static void hdd_ipa_uc_rm_notify_handler(void *context, enum ipa_rm_event event) |
| { |
| struct hdd_ipa_priv *hdd_ipa = context; |
| VOS_STATUS status = VOS_STATUS_SUCCESS; |
| |
| /* |
| * When SSR is going on or driver is unloading, just return. |
| */ |
| status = wlan_hdd_validate_context(hdd_ipa->hdd_ctx); |
| if (0 != status) |
| return; |
| |
| if (!hdd_ipa_is_rm_enabled(hdd_ipa)) |
| return; |
| |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, "%s, event code %d", |
| __func__, event); |
| |
| switch (event) { |
| case IPA_RM_RESOURCE_GRANTED: |
| /* Differed RM Granted */ |
| vos_lock_acquire(&hdd_ipa->ipa_lock); |
| if ((VOS_FALSE == hdd_ipa->resource_unloading) && |
| (!hdd_ipa->activated_fw_pipe)) { |
| hdd_ipa_uc_enable_pipes(hdd_ipa); |
| } |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| break; |
| |
| case IPA_RM_RESOURCE_RELEASED: |
| /* Differed RM Released */ |
| hdd_ipa->resource_unloading = VOS_FALSE; |
| break; |
| |
| default: |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s, invalid event code %d", |
| __func__, event); |
| break; |
| } |
| } |
| |
| static void hdd_ipa_uc_rm_notify_defer(struct work_struct *work) |
| { |
| enum ipa_rm_event event; |
| struct uc_rm_work_struct *uc_rm_work = container_of(work, |
| struct uc_rm_work_struct, work); |
| struct hdd_ipa_priv *hdd_ipa = container_of(uc_rm_work, |
| struct hdd_ipa_priv, uc_rm_work); |
| |
| vos_ssr_protect(__func__); |
| event = uc_rm_work->event; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO_HIGH, |
| "%s, posted event %d", __func__, event); |
| |
| hdd_ipa_uc_rm_notify_handler(hdd_ipa, event); |
| vos_ssr_unprotect(__func__); |
| |
| return; |
| } |
| |
| static int hdd_ipa_uc_proc_pending_event(struct hdd_ipa_priv *hdd_ipa) |
| { |
| v_SIZE_t pending_event_count; |
| struct ipa_uc_pending_event *pending_event = NULL; |
| |
| vos_list_size(&hdd_ipa->pending_event, &pending_event_count); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s, Pending Event Count %d", __func__, pending_event_count); |
| if (!pending_event_count) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s, No Pending Event", __func__); |
| return 0; |
| } |
| |
| vos_list_remove_front(&hdd_ipa->pending_event, |
| (vos_list_node_t **)&pending_event); |
| while (pending_event != NULL) { |
| hdd_ipa_wlan_evt(pending_event->adapter, |
| pending_event->type, |
| pending_event->sta_id, |
| pending_event->mac_addr); |
| vos_mem_free(pending_event); |
| pending_event = NULL; |
| vos_list_remove_front(&hdd_ipa->pending_event, |
| (vos_list_node_t **)&pending_event); |
| } |
| return 0; |
| } |
| |
| static void hdd_ipa_uc_loaded_handler(struct hdd_ipa_priv *ipa_ctxt) |
| { |
| hdd_context_t *hdd_ctx = ipa_ctxt->hdd_ctx; |
| struct ipa_wdi_out_params pipe_out; |
| |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s : UC READY", __func__); |
| if (VOS_TRUE == ipa_ctxt->uc_loaded) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO_HIGH, |
| "%s : UC already loaded", __func__); |
| return; |
| } |
| |
| ipa_ctxt->uc_loaded = VOS_TRUE; |
| /* Connect pipe */ |
| ipa_connect_wdi_pipe(&ipa_ctxt->cons_pipe_in, &pipe_out); |
| ipa_ctxt->tx_pipe_handle = pipe_out.clnt_hdl; |
| hdd_ctx->tx_comp_doorbell_paddr = (v_U32_t)pipe_out.uc_door_bell_pa; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s : TX PIPE Handle %d, DBPA 0x%x", |
| __func__, ipa_ctxt->tx_pipe_handle, (v_U32_t)pipe_out.uc_door_bell_pa); |
| |
| ipa_connect_wdi_pipe(&ipa_ctxt->prod_pipe_in, &pipe_out); |
| ipa_ctxt->rx_pipe_handle = pipe_out.clnt_hdl; |
| hdd_ctx->rx_ready_doorbell_paddr = pipe_out.uc_door_bell_pa; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s : RX PIPE Handle %d, DBPA 0x%x", |
| __func__, ipa_ctxt->rx_pipe_handle, (v_U32_t)pipe_out.uc_door_bell_pa); |
| |
| /* If already any STA connected, enable IPA/FW PIPEs */ |
| if (ipa_ctxt->sap_num_connected_sta) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "Client already connected, enable IPA/FW PIPEs"); |
| hdd_ipa_uc_handle_first_con(ipa_ctxt); |
| } |
| } |
| void hdd_ipa_uc_loaded_uc_cb(void *priv_ctxt) |
| { |
| struct hdd_ipa_priv *hdd_ipa; |
| struct op_msg_type *msg; |
| struct uc_op_work_struct *uc_op_work; |
| |
| if (NULL == priv_ctxt) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, "Invalid IPA context"); |
| return; |
| } |
| |
| hdd_ipa = (struct hdd_ipa_priv *)priv_ctxt; |
| msg = (struct op_msg_type *)adf_os_mem_alloc(NULL, |
| sizeof(struct op_msg_type)); |
| if (!msg) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, "op_msg allocation fails"); |
| return; |
| } |
| |
| msg->op_code = HDD_IPA_UC_OPCODE_UC_READY; |
| |
| uc_op_work = &hdd_ipa->uc_op_work[msg->op_code]; |
| if (uc_op_work->msg) |
| /* When the same uC OPCODE is already pended, just return */ |
| return; |
| |
| uc_op_work->msg = msg; |
| schedule_work(&uc_op_work->work); |
| return; |
| } |
| |
| static void hdd_ipa_uc_op_cb(struct op_msg_type *op_msg, void *usr_ctxt) |
| { |
| struct op_msg_type *msg = op_msg; |
| struct ipa_uc_fw_stats *uc_fw_stat; |
| struct IpaHwStatsWDIInfoData_t ipa_stat; |
| struct hdd_ipa_priv *hdd_ipa; |
| hdd_context_t *hdd_ctx; |
| VOS_STATUS status = VOS_STATUS_SUCCESS; |
| struct ipa_msg_meta meta; |
| struct ipa_wlan_msg *ipa_msg; |
| int ret = 0; |
| |
| if (!op_msg || !usr_ctxt) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s, INVALID ARG", __func__); |
| return; |
| } |
| |
| if (HDD_IPA_UC_OPCODE_MAX <= msg->op_code) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s, INVALID OPCODE %d", __func__, msg->op_code); |
| adf_os_mem_free(op_msg); |
| return; |
| } |
| |
| hdd_ctx = (hdd_context_t *)usr_ctxt; |
| hdd_ipa = (struct hdd_ipa_priv *)hdd_ctx->hdd_ipa; |
| |
| /* |
| * When SSR is going on or driver is unloading, just return. |
| */ |
| status = wlan_hdd_validate_context(hdd_ctx); |
| if (0 != status) { |
| adf_os_mem_free(op_msg); |
| return; |
| } |
| |
| HDD_IPA_DP_LOG(VOS_TRACE_LEVEL_DEBUG, |
| "%s, OPCODE %s", __func__, op_string[msg->op_code]); |
| |
| if ((HDD_IPA_UC_OPCODE_TX_RESUME == msg->op_code) || |
| (HDD_IPA_UC_OPCODE_RX_RESUME == msg->op_code)) { |
| vos_lock_acquire(&hdd_ipa->ipa_lock); |
| hdd_ipa->activated_fw_pipe++; |
| if (HDD_IPA_UC_NUM_WDI_PIPE == hdd_ipa->activated_fw_pipe) { |
| hdd_ipa->resource_loading = VOS_FALSE; |
| if (VOS_FALSE == hdd_ipa->wdi_enabled) { |
| hdd_ipa->wdi_enabled = VOS_TRUE; |
| /* WDI enable message to IPA */ |
| meta.msg_len = sizeof(*ipa_msg); |
| ipa_msg = adf_os_mem_alloc(NULL, meta.msg_len); |
| if (ipa_msg == NULL) { |
| hddLog(VOS_TRACE_LEVEL_ERROR, |
| "msg allocation failed"); |
| adf_os_mem_free(op_msg); |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| return; |
| } |
| |
| meta.msg_type = WLAN_WDI_ENABLE; |
| hddLog(VOS_TRACE_LEVEL_INFO, |
| "ipa_send_msg(Evt:%d)", meta.msg_type); |
| ret = ipa_send_msg(&meta, ipa_msg, |
| hdd_ipa_msg_free_fn); |
| if (ret) { |
| hddLog(VOS_TRACE_LEVEL_ERROR, |
| "ipa_send_msg(Evt:%d)-fail=%d", |
| meta.msg_type, ret); |
| adf_os_mem_free(ipa_msg); |
| } |
| #ifdef IPA_UC_STA_OFFLOAD |
| else { |
| /* Send SCC/MCC Switching event to IPA */ |
| hdd_ipa_send_mcc_scc_msg(hdd_ctx, |
| hdd_ctx->mcc_mode); |
| } |
| #endif |
| } |
| |
| hdd_ipa_uc_proc_pending_event(hdd_ipa); |
| |
| if (hdd_ipa->pending_cons_req) |
| ipa_rm_notify_completion( |
| IPA_RM_RESOURCE_GRANTED, |
| IPA_RM_RESOURCE_WLAN_CONS); |
| hdd_ipa->pending_cons_req = VOS_FALSE; |
| } |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| } else if ((HDD_IPA_UC_OPCODE_TX_SUSPEND == msg->op_code) || |
| (HDD_IPA_UC_OPCODE_RX_SUSPEND == msg->op_code)) { |
| vos_lock_acquire(&hdd_ipa->ipa_lock); |
| hdd_ipa->activated_fw_pipe--; |
| if (!hdd_ipa->activated_fw_pipe) { |
| hdd_ipa_uc_disable_pipes(hdd_ipa); |
| if (hdd_ipa_is_rm_enabled(hdd_ipa)) |
| ipa_rm_release_resource( |
| IPA_RM_RESOURCE_WLAN_PROD); |
| /* Sync return success from IPA |
| * Enable/resume all the PIPEs */ |
| hdd_ipa->resource_unloading = VOS_FALSE; |
| hdd_ipa_uc_proc_pending_event(hdd_ipa); |
| hdd_ipa->pending_cons_req = VOS_FALSE; |
| } |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| } |
| |
| if ((HDD_IPA_UC_OPCODE_STATS == msg->op_code) && |
| (HDD_IPA_UC_STAT_REASON_DEBUG == hdd_ipa->stat_req_reason)) { |
| /* STATs from host */ |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== IPA_UC WLAN_HOST CE ====\n" |
| "CE RING BASE: 0x%x\n" |
| "CE RING SIZE: %d\n" |
| "CE REG ADDR : 0x%x", |
| hdd_ctx->ce_sr_base_paddr, |
| hdd_ctx->ce_sr_ring_size, |
| hdd_ctx->ce_reg_paddr); |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== IPA_UC WLAN_HOST TX ====\n" |
| "COMP RING BASE: 0x%x\n" |
| "COMP RING SIZE: %d\n" |
| "NUM ALLOC BUF: %d\n" |
| "COMP RING DBELL : 0x%x", |
| hdd_ctx->tx_comp_ring_base_paddr, |
| hdd_ctx->tx_comp_ring_size, |
| hdd_ctx->tx_num_alloc_buffer, |
| hdd_ctx->tx_comp_doorbell_paddr); |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== IPA_UC WLAN_HOST RX ====\n" |
| "IND RING BASE: 0x%x\n" |
| "IND RING SIZE: %d\n" |
| "IND RING DBELL : 0x%x\n" |
| "PROC DONE IND ADDR : 0x%x\n" |
| "NUM EXCP PKT : %llu\n" |
| "NUM TX FWD OK: %llu\n" |
| "NUM TX FWD ERR : %llu", |
| hdd_ctx->rx_rdy_ring_base_paddr, |
| hdd_ctx->rx_rdy_ring_size, |
| hdd_ctx->rx_ready_doorbell_paddr, |
| hdd_ctx->rx_proc_done_idx_paddr, |
| hdd_ipa->stats.num_rx_excep, |
| hdd_ipa->stats.num_tx_fwd_ok, |
| hdd_ipa->stats.num_tx_fwd_err); |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== IPA_UC WLAN_HOST CONTROL ====\n" |
| "SAP NUM STAs: %d\n" |
| #ifdef IPA_UC_STA_OFFLOAD |
| "STA CONNECTED: %d\n" |
| "CONCURRENT MODE: %s\n" |
| #endif |
| "TX PIPE HDL: %d\n" |
| "RX PIPE HDL : %d\n" |
| "RSC LOADING : %d\n" |
| "RSC UNLOADING : %d\n" |
| "PNDNG CNS RQT : %d", |
| hdd_ipa->sap_num_connected_sta, |
| #ifdef IPA_UC_STA_OFFLOAD |
| hdd_ipa->sta_connected, |
| (hdd_ctx->mcc_mode ? "MCC" : "SCC"), |
| #endif |
| hdd_ipa->tx_pipe_handle, |
| hdd_ipa->rx_pipe_handle, |
| (unsigned int)hdd_ipa->resource_loading, |
| (unsigned int)hdd_ipa->resource_unloading, |
| (unsigned int)hdd_ipa->pending_cons_req); |
| |
| /* STATs from FW */ |
| uc_fw_stat = (struct ipa_uc_fw_stats *) |
| ((v_U8_t *)op_msg + sizeof(struct op_msg_type)); |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== IPA_UC WLAN_FW TX ====\n" |
| "COMP RING BASE: 0x%x\n" |
| "COMP RING SIZE: %d\n" |
| "COMP RING DBELL : 0x%x\n" |
| "COMP RING DBELL IND VAL : %d\n" |
| "COMP RING DBELL CACHED VAL : %d\n" |
| "COMP RING DBELL CACHED VAL : %d\n" |
| "PKTS ENQ : %d\n" |
| "PKTS COMP : %d\n" |
| "IS SUSPEND : %d\n" |
| "RSVD : 0x%x", |
| uc_fw_stat->tx_comp_ring_base, |
| uc_fw_stat->tx_comp_ring_size, |
| uc_fw_stat->tx_comp_ring_dbell_addr, |
| uc_fw_stat->tx_comp_ring_dbell_ind_val, |
| uc_fw_stat->tx_comp_ring_dbell_cached_val, |
| uc_fw_stat->tx_comp_ring_dbell_cached_val, |
| uc_fw_stat->tx_pkts_enqueued, |
| uc_fw_stat->tx_pkts_completed, |
| uc_fw_stat->tx_is_suspend, |
| uc_fw_stat->tx_reserved); |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== IPA_UC WLAN_FW RX ====\n" |
| "IND RING BASE: 0x%x\n" |
| "IND RING SIZE: %d\n" |
| "IND RING DBELL : 0x%x\n" |
| "IND RING DBELL IND VAL : %d\n" |
| "IND RING DBELL CACHED VAL : %d\n" |
| "RDY IND ADDR : 0x%x\n" |
| "RDY IND CACHE VAL : %d\n" |
| "RFIL IND : %d\n" |
| "NUM PKT INDICAT : %d\n" |
| "BUF REFIL : %d\n" |
| "NUM DROP NO SPC : %d\n" |
| "NUM DROP NO BUF : %d\n" |
| "IS SUSPND : %d\n" |
| "RSVD : 0x%x\n", |
| uc_fw_stat->rx_ind_ring_base, |
| uc_fw_stat->rx_ind_ring_size, |
| uc_fw_stat->rx_ind_ring_dbell_addr, |
| uc_fw_stat->rx_ind_ring_dbell_ind_val, |
| uc_fw_stat->rx_ind_ring_dbell_ind_cached_val, |
| uc_fw_stat->rx_ind_ring_rdidx_addr, |
| uc_fw_stat->rx_ind_ring_rd_idx_cached_val, |
| uc_fw_stat->rx_refill_idx, |
| uc_fw_stat->rx_num_pkts_indicated, |
| uc_fw_stat->rx_buf_refilled, |
| uc_fw_stat->rx_num_ind_drop_no_space, |
| uc_fw_stat->rx_num_ind_drop_no_buf, |
| uc_fw_stat->rx_is_suspend, |
| uc_fw_stat->rx_reserved); |
| /* STATs from IPA */ |
| ipa_get_wdi_stats(&ipa_stat); |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== IPA_UC IPA TX ====\n" |
| "NUM PROCD : %d\n" |
| "CE DBELL : 0x%x\n" |
| "NUM DBELL FIRED : %d\n" |
| "COMP RNG FULL : %d\n" |
| "COMP RNG EMPT : %d\n" |
| "COMP RNG USE HGH : %d\n" |
| "COMP RNG USE LOW : %d\n" |
| "BAM FIFO FULL : %d\n" |
| "BAM FIFO EMPT : %d\n" |
| "BAM FIFO USE HGH : %d\n" |
| "BAM FIFO USE LOW : %d\n" |
| "NUM DBELL : %d\n" |
| "NUM UNEXP DBELL : %d\n" |
| "NUM BAM INT HDL : 0x%x\n" |
| "NUM BAM INT NON-RUN : 0x%x\n" |
| "NUM QMB INT HDL : 0x%x", |
| ipa_stat.tx_ch_stats.num_pkts_processed, |
| ipa_stat.tx_ch_stats.copy_engine_doorbell_value, |
| ipa_stat.tx_ch_stats.num_db_fired, |
| ipa_stat.tx_ch_stats.tx_comp_ring_stats.ringFull, |
| ipa_stat.tx_ch_stats.tx_comp_ring_stats.ringEmpty, |
| ipa_stat.tx_ch_stats.tx_comp_ring_stats.ringUsageHigh, |
| ipa_stat.tx_ch_stats.tx_comp_ring_stats.ringUsageLow, |
| ipa_stat.tx_ch_stats.bam_stats.bamFifoFull, |
| ipa_stat.tx_ch_stats.bam_stats.bamFifoEmpty, |
| ipa_stat.tx_ch_stats.bam_stats.bamFifoUsageHigh, |
| ipa_stat.tx_ch_stats.bam_stats.bamFifoUsageLow, |
| ipa_stat.tx_ch_stats.num_db, |
| ipa_stat.tx_ch_stats.num_unexpected_db, |
| ipa_stat.tx_ch_stats.num_bam_int_handled, |
| ipa_stat.tx_ch_stats.num_bam_int_in_non_runnning_state, |
| ipa_stat.tx_ch_stats.num_qmb_int_handled); |
| |
| VOS_TRACE(VOS_MODULE_ID_HDD, VOS_TRACE_LEVEL_ERROR, |
| "==== IPA_UC IPA RX ====\n" |
| "MAX OST PKT : %d\n" |
| "NUM PKT PRCSD : %d\n" |
| "RNG RP : 0x%x\n" |
| "COMP RNG FULL : %d\n" |
| "COMP RNG EMPT : %d\n" |
| "COMP RNG USE HGH : %d\n" |
| "COMP RNG USE LOW : %d\n" |
| "BAM FIFO FULL : %d\n" |
| "BAM FIFO EMPT : %d\n" |
| "BAM FIFO USE HGH : %d\n" |
| "BAM FIFO USE LOW : %d\n" |
| "NUM DB : %d\n" |
| "NUM UNEXP DB : %d\n" |
| "NUM BAM INT HNDL : 0x%x\n", |
| ipa_stat.rx_ch_stats.max_outstanding_pkts, |
| ipa_stat.rx_ch_stats.num_pkts_processed, |
| ipa_stat.rx_ch_stats.rx_ring_rp_value, |
| ipa_stat.rx_ch_stats.rx_ind_ring_stats.ringFull, |
| ipa_stat.rx_ch_stats.rx_ind_ring_stats.ringEmpty, |
| ipa_stat.rx_ch_stats.rx_ind_ring_stats.ringUsageHigh, |
| ipa_stat.rx_ch_stats.rx_ind_ring_stats.ringUsageLow, |
| ipa_stat.rx_ch_stats.bam_stats.bamFifoFull, |
| ipa_stat.rx_ch_stats.bam_stats.bamFifoEmpty, |
| ipa_stat.rx_ch_stats.bam_stats.bamFifoUsageHigh, |
| ipa_stat.rx_ch_stats.bam_stats.bamFifoUsageLow, |
| ipa_stat.rx_ch_stats.num_db, |
| ipa_stat.rx_ch_stats.num_unexpected_db, |
| ipa_stat.rx_ch_stats.num_bam_int_handled); |
| } else if ((HDD_IPA_UC_OPCODE_STATS == msg->op_code) && |
| (HDD_IPA_UC_STAT_REASON_BW_CAL == hdd_ipa->stat_req_reason)) { |
| /* STATs from FW */ |
| uc_fw_stat = (struct ipa_uc_fw_stats *) |
| ((v_U8_t *)op_msg + sizeof(struct op_msg_type)); |
| vos_lock_acquire(&hdd_ipa->ipa_lock); |
| hdd_ipa->ipa_tx_packets_diff = HDD_BW_GET_DIFF( |
| uc_fw_stat->tx_pkts_completed, |
| hdd_ipa->ipa_p_tx_packets); |
| hdd_ipa->ipa_rx_packets_diff = HDD_BW_GET_DIFF( |
| (uc_fw_stat->rx_num_ind_drop_no_space + |
| uc_fw_stat->rx_num_ind_drop_no_buf + |
| uc_fw_stat->rx_num_pkts_indicated), |
| hdd_ipa->ipa_p_rx_packets); |
| |
| hdd_ipa->ipa_p_tx_packets = uc_fw_stat->tx_pkts_completed; |
| hdd_ipa->ipa_p_rx_packets = |
| (uc_fw_stat->rx_num_ind_drop_no_space + |
| uc_fw_stat->rx_num_ind_drop_no_buf + |
| uc_fw_stat->rx_num_pkts_indicated); |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| } else if (HDD_IPA_UC_OPCODE_UC_READY == msg->op_code) { |
| vos_lock_acquire(&hdd_ipa->ipa_lock); |
| hdd_ipa_uc_loaded_handler(hdd_ipa); |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| } |
| |
| adf_os_mem_free(op_msg); |
| } |
| |
| static void hdd_ipa_uc_fw_op_event_handler(struct work_struct *work) |
| { |
| struct op_msg_type *msg; |
| struct uc_op_work_struct *uc_op_work = container_of(work, |
| struct uc_op_work_struct, work); |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| |
| vos_ssr_protect(__func__); |
| |
| msg = uc_op_work->msg; |
| uc_op_work->msg = NULL; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO_HIGH, |
| "%s, posted msg %d", __func__, msg->op_code); |
| |
| hdd_ipa_uc_op_cb(msg, hdd_ipa->hdd_ctx); |
| |
| vos_ssr_unprotect(__func__); |
| |
| return; |
| } |
| |
| static void hdd_ipa_uc_op_event_handler(v_U8_t *op_msg, void *hdd_ctx) |
| { |
| struct hdd_ipa_priv *hdd_ipa; |
| struct op_msg_type *msg; |
| struct uc_op_work_struct *uc_op_work; |
| |
| if (NULL == hdd_ctx) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, "Invalid HDD context"); |
| goto end; |
| } |
| |
| msg = (struct op_msg_type *)op_msg; |
| hdd_ipa = ((hdd_context_t *)hdd_ctx)->hdd_ipa; |
| |
| if (unlikely(!hdd_ipa)) |
| goto end; |
| |
| if (HDD_IPA_UC_OPCODE_MAX <= msg->op_code) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, "%s: Invalid OP Code (%d)", |
| __func__, msg->op_code); |
| goto end; |
| } |
| |
| uc_op_work = &hdd_ipa->uc_op_work[msg->op_code]; |
| if (uc_op_work->msg) |
| /* When the same uC OPCODE is already pended, just return */ |
| goto end; |
| |
| uc_op_work->msg = msg; |
| schedule_work(&uc_op_work->work); |
| return; |
| |
| end: |
| adf_os_mem_free(op_msg); |
| } |
| |
| static VOS_STATUS hdd_ipa_uc_ol_init(hdd_context_t *hdd_ctx) |
| { |
| struct ipa_wdi_in_params pipe_in; |
| struct ipa_wdi_out_params pipe_out; |
| struct hdd_ipa_priv *ipa_ctxt = (struct hdd_ipa_priv *)hdd_ctx->hdd_ipa; |
| uint8_t i; |
| struct ipa_wdi_db_params dbpa; |
| |
| vos_mem_zero(&ipa_ctxt->cons_pipe_in, sizeof(struct ipa_wdi_in_params)); |
| vos_mem_zero(&ipa_ctxt->prod_pipe_in, sizeof(struct ipa_wdi_in_params)); |
| |
| vos_mem_zero(&pipe_in, sizeof(struct ipa_wdi_in_params)); |
| vos_mem_zero(&pipe_out, sizeof(struct ipa_wdi_out_params)); |
| |
| vos_list_init(&ipa_ctxt->pending_event); |
| |
| if (!hdd_ctx->isLogpInProgress) { |
| vos_lock_init(&ipa_ctxt->event_lock); |
| vos_lock_init(&ipa_ctxt->ipa_lock); |
| } |
| |
| /* TX PIPE */ |
| pipe_in.sys.ipa_ep_cfg.nat.nat_en = IPA_BYPASS_NAT; |
| pipe_in.sys.ipa_ep_cfg.hdr.hdr_len = HDD_IPA_UC_WLAN_TX_HDR_LEN; |
| pipe_in.sys.ipa_ep_cfg.hdr.hdr_ofst_pkt_size_valid = 1; |
| pipe_in.sys.ipa_ep_cfg.hdr.hdr_ofst_pkt_size = 0; |
| pipe_in.sys.ipa_ep_cfg.hdr.hdr_additional_const_len = |
| HDD_IPA_UC_WLAN_8023_HDR_SIZE; |
| pipe_in.sys.ipa_ep_cfg.mode.mode = IPA_BASIC; |
| pipe_in.sys.client = IPA_CLIENT_WLAN1_CONS; |
| pipe_in.sys.desc_fifo_sz = hdd_ctx->cfg_ini->IpaDescSize; |
| pipe_in.sys.priv = hdd_ctx->hdd_ipa; |
| pipe_in.sys.ipa_ep_cfg.hdr_ext.hdr_little_endian = true; |
| pipe_in.sys.notify = hdd_ipa_i2w_cb; |
| if (!hdd_ipa_is_rm_enabled(ghdd_ipa)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: IPA RM DISABLED, IPA AWAKE", __func__); |
| pipe_in.sys.keep_ipa_awake = TRUE; |
| } |
| |
| pipe_in.u.dl.comp_ring_base_pa = hdd_ctx->tx_comp_ring_base_paddr; |
| pipe_in.u.dl.comp_ring_size = hdd_ctx->tx_comp_ring_size * 4; |
| pipe_in.u.dl.ce_ring_base_pa = hdd_ctx->ce_sr_base_paddr; |
| pipe_in.u.dl.ce_door_bell_pa = hdd_ctx->ce_reg_paddr; |
| pipe_in.u.dl.ce_ring_size = hdd_ctx->ce_sr_ring_size * 8; |
| pipe_in.u.dl.num_tx_buffers = hdd_ctx->tx_num_alloc_buffer; |
| |
| vos_mem_copy(&ipa_ctxt->cons_pipe_in, |
| &pipe_in, |
| sizeof(struct ipa_wdi_in_params)); |
| dbpa.client = IPA_CLIENT_WLAN1_CONS; |
| ipa_uc_wdi_get_dbpa(&dbpa); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s CONS DB get dbpa 0x%x", |
| __func__, (unsigned int)dbpa.uc_door_bell_pa); |
| hdd_ctx->tx_comp_doorbell_paddr = dbpa.uc_door_bell_pa; |
| if (VOS_TRUE == ipa_ctxt->uc_loaded) { |
| /* Connect WDI IPA PIPE */ |
| ipa_connect_wdi_pipe(&ipa_ctxt->cons_pipe_in, &pipe_out); |
| /* Micro Controller Doorbell register */ |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s CONS DB pipe out 0x%x TX PIPE Handle 0x%x", |
| __func__, (unsigned int)pipe_out.uc_door_bell_pa, |
| ipa_ctxt->tx_pipe_handle); |
| |
| hdd_ctx->tx_comp_doorbell_paddr = (v_U32_t)pipe_out.uc_door_bell_pa; |
| /* WLAN TX PIPE Handle */ |
| ipa_ctxt->tx_pipe_handle = pipe_out.clnt_hdl; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO_HIGH, |
| "TX : CRBPA 0x%x, CRS %d, CERBPA 0x%x, CEDPA 0x%x," |
| " CERZ %d, NB %d, CDBPAD 0x%x", |
| (unsigned int)pipe_in.u.dl.comp_ring_base_pa, |
| pipe_in.u.dl.comp_ring_size, |
| (unsigned int)pipe_in.u.dl.ce_ring_base_pa, |
| (unsigned int)pipe_in.u.dl.ce_door_bell_pa, |
| pipe_in.u.dl.ce_ring_size, |
| pipe_in.u.dl.num_tx_buffers, |
| (unsigned int)hdd_ctx->tx_comp_doorbell_paddr); |
| } |
| |
| /* RX PIPE */ |
| pipe_in.sys.ipa_ep_cfg.nat.nat_en = IPA_BYPASS_NAT; |
| pipe_in.sys.ipa_ep_cfg.hdr.hdr_len = HDD_IPA_UC_WLAN_RX_HDR_LEN; |
| pipe_in.sys.ipa_ep_cfg.hdr.hdr_ofst_metadata_valid = 0; |
| pipe_in.sys.ipa_ep_cfg.hdr.hdr_metadata_reg_valid = 1; |
| pipe_in.sys.ipa_ep_cfg.mode.mode = IPA_BASIC; |
| pipe_in.sys.client = IPA_CLIENT_WLAN1_PROD; |
| pipe_in.sys.desc_fifo_sz = hdd_ctx->cfg_ini->IpaDescSize + |
| sizeof(struct sps_iovec); |
| pipe_in.sys.notify = hdd_ipa_w2i_cb; |
| if (!hdd_ipa_is_rm_enabled(ghdd_ipa)) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: IPA RM DISABLED, IPA AWAKE", __func__); |
| pipe_in.sys.keep_ipa_awake = TRUE; |
| } |
| |
| pipe_in.u.ul.rdy_ring_base_pa = hdd_ctx->rx_rdy_ring_base_paddr; |
| pipe_in.u.ul.rdy_ring_size = hdd_ctx->rx_rdy_ring_size; |
| pipe_in.u.ul.rdy_ring_rp_pa = hdd_ctx->rx_proc_done_idx_paddr; |
| |
| vos_mem_copy(&ipa_ctxt->prod_pipe_in, |
| &pipe_in, |
| sizeof(struct ipa_wdi_in_params)); |
| dbpa.client = IPA_CLIENT_WLAN1_PROD; |
| ipa_uc_wdi_get_dbpa(&dbpa); |
| hdd_ctx->rx_ready_doorbell_paddr = dbpa.uc_door_bell_pa; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s PROD DB get dbpa 0x%x", |
| __func__, (unsigned int)dbpa.uc_door_bell_pa); |
| if (VOS_TRUE == ipa_ctxt->uc_loaded) { |
| ipa_connect_wdi_pipe(&ipa_ctxt->prod_pipe_in, &pipe_out); |
| hdd_ctx->rx_ready_doorbell_paddr = pipe_out.uc_door_bell_pa; |
| ipa_ctxt->rx_pipe_handle = pipe_out.clnt_hdl; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s PROD DB pipe out 0x%x", |
| __func__, (unsigned int)pipe_out.uc_door_bell_pa); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO_HIGH, |
| "RX : RRBPA 0x%x, RRS %d, PDIPA 0x%x, RDY_DB_PAD 0x%x", |
| (unsigned int)pipe_in.u.ul.rdy_ring_base_pa, |
| pipe_in.u.ul.rdy_ring_size, |
| (unsigned int)pipe_in.u.ul.rdy_ring_rp_pa, |
| (unsigned int)hdd_ctx->rx_ready_doorbell_paddr); |
| } |
| WLANTL_SetUcDoorbellPaddr((pVosContextType)(hdd_ctx->pvosContext), |
| (v_U32_t)hdd_ctx->tx_comp_doorbell_paddr, |
| (v_U32_t)hdd_ctx->rx_ready_doorbell_paddr); |
| |
| WLANTL_RegisterOPCbFnc((pVosContextType)(hdd_ctx->pvosContext), |
| hdd_ipa_uc_op_event_handler, (void *)hdd_ctx); |
| |
| for (i = 0; i < HDD_IPA_UC_OPCODE_MAX; i++) { |
| vos_init_work(&ipa_ctxt->uc_op_work[i].work, |
| hdd_ipa_uc_fw_op_event_handler); |
| ipa_ctxt->uc_op_work[i].msg = NULL; |
| } |
| |
| return VOS_STATUS_SUCCESS; |
| } |
| |
| /** |
| * hdd_ipa_uc_force_pipe_shutdown() - Force shutdown IPA pipe |
| * @hdd_ctx: hdd main context |
| * |
| * Force shutdown IPA pipe |
| * Independent of FW pipe status, IPA pipe shutdonw progress |
| * in case, any STA does not leave properly, IPA HW pipe should cleaned up |
| * independent from FW pipe status |
| * |
| * Return: NONE |
| */ |
| void hdd_ipa_uc_force_pipe_shutdown(hdd_context_t *hdd_ctx) |
| { |
| struct hdd_ipa_priv *hdd_ipa; |
| |
| if (!hdd_ipa_is_enabled(hdd_ctx) || !hdd_ctx->hdd_ipa) |
| return; |
| |
| hdd_ipa = (struct hdd_ipa_priv *)hdd_ctx->hdd_ipa; |
| if (false == hdd_ipa->ipa_pipes_down) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "IPA pipes are not down yet, force shutdown"); |
| hdd_ipa_uc_disable_pipes(hdd_ipa); |
| } else { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "IPA pipes are down, do nothing"); |
| } |
| |
| return; |
| } |
| |
| /** |
| * hdd_ipa_uc_send_evt() - send event to ipa |
| * @hdd_ctx: pointer to hdd context |
| * @type: event type |
| * @mac_addr: pointer to mac address |
| * |
| * Send event to IPA driver |
| * |
| * Return: 0 - Success |
| */ |
| static int hdd_ipa_uc_send_evt(hdd_adapter_t *adapter, |
| enum ipa_wlan_event type, uint8_t *mac_addr ) |
| { |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| struct ipa_msg_meta meta; |
| struct ipa_wlan_msg *msg; |
| int ret = 0; |
| |
| meta.msg_len = sizeof(struct ipa_wlan_msg); |
| msg = adf_os_mem_alloc(NULL, meta.msg_len); |
| if (msg == NULL) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "msg allocation failed"); |
| return -ENOMEM; |
| } |
| |
| meta.msg_type = type; |
| strlcpy(msg->name, adapter->dev->name, |
| IPA_RESOURCE_NAME_MAX); |
| memcpy(msg->mac_addr, mac_addr, ETH_ALEN); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, "%s: Evt: %d", |
| msg->name, meta.msg_type); |
| ret = ipa_send_msg(&meta, msg, hdd_ipa_msg_free_fn); |
| if (ret) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "%s: Evt: %d fail:%d", |
| msg->name, meta.msg_type, ret); |
| adf_os_mem_free(msg); |
| return ret; |
| } |
| |
| hdd_ipa->stats.num_send_msg++; |
| |
| return ret; |
| } |
| |
| /** |
| * hdd_ipa_uc_disconnect_client() - send client disconnect event |
| * @hdd_ctx: pointer to hdd adapter |
| * |
| * Send disconnect client event to IPA driver during SSR |
| * |
| * Return: 0 - Success |
| */ |
| static int hdd_ipa_uc_disconnect_client(hdd_adapter_t *adapter) |
| { |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| int ret = 0; |
| int i; |
| |
| for (i = 0; i < WLAN_MAX_STA_COUNT; i++) { |
| if (vos_is_macaddr_broadcast(&adapter->aStaInfo[i].macAddrSTA)) |
| continue; |
| if ((adapter->aStaInfo[i].isUsed) && |
| (!adapter->aStaInfo[i].isDeauthInProgress) && |
| hdd_ipa->sap_num_connected_sta) { |
| hdd_ipa_uc_send_evt(adapter, WLAN_CLIENT_DISCONNECT, |
| adapter->aStaInfo[i].macAddrSTA.bytes); |
| hdd_ipa->sap_num_connected_sta--; |
| } |
| } |
| |
| return ret; |
| } |
| |
| /** |
| * hdd_ipa_uc_disconnect_ap() - send ap disconnect event |
| * @hdd_ctx: pointer to hdd adapter |
| * |
| * Send disconnect ap event to IPA driver during SSR |
| * |
| * Return: 0 - Success |
| */ |
| |
| static int hdd_ipa_uc_disconnect_ap(hdd_adapter_t *adapter) |
| { |
| int ret = 0; |
| |
| if (adapter->ipa_context) |
| hdd_ipa_uc_send_evt(adapter, WLAN_AP_DISCONNECT, |
| adapter->dev->dev_addr); |
| |
| return ret; |
| } |
| |
| #ifdef IPA_UC_STA_OFFLOAD |
| /** |
| * hdd_ipa_uc_disconnect_sta() - send sta disconnect event |
| * @hdd_ctx: pointer to hdd adapter |
| * |
| * Send disconnect sta event to IPA driver during SSR |
| * |
| * Return: 0 - Success |
| */ |
| static int hdd_ipa_uc_disconnect_sta(hdd_adapter_t *adapter) |
| { |
| hdd_station_ctx_t *pHddStaCtx; |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| int ret = 0; |
| |
| if (hdd_ipa_uc_sta_is_enabled(hdd_ipa) && |
| hdd_ipa->sta_connected) { |
| pHddStaCtx = WLAN_HDD_GET_STATION_CTX_PTR(adapter); |
| hdd_ipa_uc_send_evt(adapter, WLAN_STA_DISCONNECT, |
| pHddStaCtx->conn_info.bssId); |
| } |
| |
| return ret; |
| } |
| #else |
| static int hdd_ipa_uc_disconnect_sta(hdd_adapter_t *adapter) |
| { |
| return 0; |
| } |
| |
| #endif |
| |
| /** |
| * hdd_ipa_uc_disconnect() - send disconnect ipa event |
| * @hdd_ctx: pointer to hdd context |
| * |
| * Send disconnect event to IPA driver during SSR |
| * |
| * Return: 0 - Success |
| */ |
| static int hdd_ipa_uc_disconnect(hdd_context_t *hdd_ctx) |
| { |
| hdd_adapter_list_node_t *adapter_node = NULL, *next = NULL; |
| VOS_STATUS status; |
| hdd_adapter_t *adapter; |
| int ret = 0; |
| |
| status = hdd_get_front_adapter (hdd_ctx, &adapter_node); |
| while (NULL != adapter_node && VOS_STATUS_SUCCESS == status) { |
| adapter = adapter_node->pAdapter; |
| if (adapter->device_mode == WLAN_HDD_SOFTAP) { |
| hdd_ipa_uc_disconnect_client(adapter); |
| hdd_ipa_uc_disconnect_ap(adapter); |
| } else if (adapter->device_mode == WLAN_HDD_INFRA_STATION) { |
| hdd_ipa_uc_disconnect_sta(adapter); |
| } |
| |
| status = hdd_get_next_adapter( |
| hdd_ctx, adapter_node, &next); |
| adapter_node = next; |
| } |
| |
| return ret; |
| } |
| |
| /** |
| * hdd_ipa_uc_ssr_deinit() - handle ipa deinit for SSR |
| * |
| * Deinit basic IPA UC host side to be in sync reloaded FW during |
| * SSR |
| * |
| * Return: 0 - Success |
| */ |
| int hdd_ipa_uc_ssr_deinit() |
| { |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| int idx; |
| struct hdd_ipa_iface_context *iface_context; |
| hdd_context_t *hdd_ctx; |
| |
| if (!hdd_ipa_uc_is_enabled(hdd_ipa)) |
| return 0; |
| |
| hdd_ctx = hdd_ipa->hdd_ctx; |
| /* send disconnect to ipa driver */ |
| hdd_ipa_uc_disconnect(hdd_ctx); |
| |
| /* Clean up HDD IPA interfaces */ |
| for (idx = 0; (hdd_ipa->num_iface > 0) && |
| (idx < HDD_IPA_MAX_IFACE); idx++) { |
| iface_context = &hdd_ipa->iface_context[idx]; |
| if (iface_context && iface_context->adapter) |
| hdd_ipa_cleanup_iface(iface_context); |
| } |
| |
| /* After SSR, wlan driver reloads FW again. But we need to protect |
| * IPA submodule during SSR transient state. So deinit basic IPA |
| * UC host side to be in sync with reloaded FW during SSR |
| */ |
| if (false == hdd_ipa->ipa_pipes_down) |
| hdd_ipa_uc_disable_pipes(hdd_ipa); |
| |
| vos_lock_acquire(&hdd_ipa->ipa_lock); |
| for (idx = 0; idx < WLAN_MAX_STA_COUNT; idx++) { |
| hdd_ipa->assoc_stas_map[idx].is_reserved = false; |
| hdd_ipa->assoc_stas_map[idx].sta_id = 0xFF; |
| } |
| vos_lock_release(&hdd_ipa->ipa_lock); |
| |
| /* |
| * Do WDI pipes disconnect here. During reinit new WDI pipes |
| * will be created. |
| */ |
| /* In MDM case pipes will be disconnected as part of ipa cleanup */ |
| if (hdd_ctx->cfg_ini->sap_internal_restart) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: Disconnect TX PIPE tx_pipe_handle=0x%x", |
| __func__, hdd_ipa->tx_pipe_handle); |
| ipa_disconnect_wdi_pipe(hdd_ipa->tx_pipe_handle); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "%s: Disconnect RX PIPE rx_pipe_handle=0x%x", |
| __func__, hdd_ipa->rx_pipe_handle); |
| ipa_disconnect_wdi_pipe(hdd_ipa->rx_pipe_handle); |
| } |
| |
| if (hdd_ipa_uc_sta_is_enabled(hdd_ipa)) { |
| hdd_ipa_uc_sta_reset_sta_connected(hdd_ipa); |
| } |
| |
| /* Full IPA driver cleanup not required since wlan driver is now |
| * unloaded and reloaded after SSR. |
| */ |
| return 0; |
| } |
| |
| /** |
| * hdd_ipa_uc_ssr_reinit() - handle ipa reinit after SSR |
| * |
| * Init basic IPA UC host side to be in sync with reloaded FW after |
| * SSR to resume IPA UC operations |
| * |
| * Return: 0 - Success |
| */ |
| int hdd_ipa_uc_ssr_reinit(hdd_context_t *hdd_ctx) |
| { |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| |
| if (!hdd_ipa_uc_is_enabled(hdd_ipa)) |
| return 0; |
| |
| if (hdd_ctx->cfg_ini->sap_internal_restart) |
| hdd_ipa_uc_ol_init(hdd_ctx); |
| |
| return 0; |
| } |
| #else |
| /** |
| * hdd_ipa_uc_rt_debug_destructor - called by data packet free |
| * @skb: packet pinter |
| * |
| * when free data packet, will be invoked by wlan client and will increase |
| * free counter |
| * |
| * Return: none |
| */ |
| void hdd_ipa_uc_rt_debug_destructor(struct sk_buff *skb) |
| { |
| return; |
| } |
| #endif /* IPA_UC_OFFLOAD */ |
| |
| static int hdd_ipa_rm_request(struct hdd_ipa_priv *hdd_ipa) |
| { |
| int ret = 0; |
| |
| if (!hdd_ipa_is_rm_enabled(hdd_ipa)) |
| return 0; |
| |
| adf_os_spin_lock_bh(&hdd_ipa->rm_lock); |
| |
| switch(hdd_ipa->rm_state) { |
| case HDD_IPA_RM_GRANTED: |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| return 0; |
| case HDD_IPA_RM_GRANT_PENDING: |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| return -EINPROGRESS; |
| case HDD_IPA_RM_RELEASED: |
| hdd_ipa->rm_state = HDD_IPA_RM_GRANT_PENDING; |
| break; |
| } |
| |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| |
| ret = ipa_rm_inactivity_timer_request_resource( |
| IPA_RM_RESOURCE_WLAN_PROD); |
| |
| adf_os_spin_lock_bh(&hdd_ipa->rm_lock); |
| if (ret == 0) { |
| hdd_ipa->rm_state = HDD_IPA_RM_GRANTED; |
| hdd_ipa->stats.num_rm_grant_imm++; |
| } |
| |
| cancel_delayed_work(&hdd_ipa->wake_lock_work); |
| if (hdd_ipa->wake_lock_released) { |
| vos_wake_lock_acquire(&hdd_ipa->wake_lock, |
| WIFI_POWER_EVENT_WAKELOCK_IPA); |
| hdd_ipa->wake_lock_released = false; |
| } |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| |
| return ret; |
| } |
| |
| static void hdd_ipa_wake_lock_timer_func(struct work_struct *work) |
| { |
| struct hdd_ipa_priv *hdd_ipa = container_of(to_delayed_work(work), |
| struct hdd_ipa_priv, wake_lock_work); |
| |
| adf_os_spin_lock_bh(&hdd_ipa->rm_lock); |
| |
| if (hdd_ipa->rm_state != HDD_IPA_RM_RELEASED) |
| goto end; |
| |
| hdd_ipa->wake_lock_released = true; |
| vos_wake_lock_release(&hdd_ipa->wake_lock, |
| WIFI_POWER_EVENT_WAKELOCK_IPA); |
| |
| end: |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| } |
| |
| static int hdd_ipa_rm_try_release(struct hdd_ipa_priv *hdd_ipa) |
| { |
| int ret = 0; |
| |
| if (!hdd_ipa_is_rm_enabled(hdd_ipa)) |
| return 0; |
| |
| if (atomic_read(&hdd_ipa->tx_ref_cnt)) |
| return -EAGAIN; |
| |
| #ifndef IPA_UC_STA_OFFLOAD |
| spin_lock_bh(&hdd_ipa->q_lock); |
| if (hdd_ipa->pending_hw_desc_cnt || hdd_ipa->pend_q_cnt) { |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| return -EAGAIN; |
| } |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| #endif |
| |
| adf_os_spin_lock_bh(&hdd_ipa->pm_lock); |
| |
| if (!adf_nbuf_is_queue_empty(&hdd_ipa->pm_queue_head)) { |
| adf_os_spin_unlock_bh(&hdd_ipa->pm_lock); |
| return -EAGAIN; |
| } |
| adf_os_spin_unlock_bh(&hdd_ipa->pm_lock); |
| |
| |
| adf_os_spin_lock_bh(&hdd_ipa->rm_lock); |
| switch(hdd_ipa->rm_state) { |
| case HDD_IPA_RM_GRANTED: |
| break; |
| case HDD_IPA_RM_GRANT_PENDING: |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| return -EINPROGRESS; |
| case HDD_IPA_RM_RELEASED: |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| return 0; |
| } |
| |
| /* IPA driver returns immediately so set the state here to avoid any |
| * race condition. |
| */ |
| hdd_ipa->rm_state = HDD_IPA_RM_RELEASED; |
| hdd_ipa->stats.num_rm_release++; |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| |
| ret = ipa_rm_inactivity_timer_release_resource( |
| IPA_RM_RESOURCE_WLAN_PROD); |
| |
| adf_os_spin_lock_bh(&hdd_ipa->rm_lock); |
| if (unlikely(ret != 0)) { |
| hdd_ipa->rm_state = HDD_IPA_RM_GRANTED; |
| WARN_ON(1); |
| } |
| |
| /* |
| * If wake_lock is released immediately, kernel would try to suspend |
| * immediately as well, Just avoid ping-pong between suspend-resume |
| * while there is healthy amount of data transfer going on by |
| * releasing the wake_lock after some delay. |
| */ |
| schedule_delayed_work(&hdd_ipa->wake_lock_work, |
| msecs_to_jiffies(HDD_IPA_RX_INACTIVITY_MSEC_DELAY)); |
| |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| |
| return ret; |
| } |
| |
| static void hdd_ipa_send_pkt_to_ipa(struct hdd_ipa_priv *hdd_ipa) |
| { |
| struct ipa_tx_data_desc *send_desc, *desc, *tmp; |
| uint32_t cur_send_cnt = 0, pend_q_cnt; |
| adf_nbuf_t buf; |
| struct ipa_tx_data_desc *send_desc_head = NULL; |
| |
| /* Unloading is in progress so do not proceed to send the packets to |
| * IPA |
| */ |
| if (hdd_ipa->hdd_ctx->isUnloadInProgress) |
| return; |
| |
| /* Make it priority queue request as send descriptor */ |
| send_desc_head = hdd_ipa_alloc_data_desc(hdd_ipa, 1); |
| |
| /* Try again later when descriptors are available */ |
| if (!send_desc_head) |
| return; |
| |
| INIT_LIST_HEAD(&send_desc_head->link); |
| |
| spin_lock_bh(&hdd_ipa->q_lock); |
| |
| if (hdd_ipa->pending_hw_desc_cnt >= hdd_ipa->hw_desc_cnt) { |
| hdd_ipa->stats.num_rx_ipa_hw_maxed_out++; |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| hdd_ipa_free_data_desc(hdd_ipa, send_desc_head); |
| return; |
| } |
| |
| pend_q_cnt = hdd_ipa->pend_q_cnt; |
| |
| if (pend_q_cnt == 0) { |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| hdd_ipa_free_data_desc(hdd_ipa, send_desc_head); |
| return; |
| } |
| |
| /* If hardware has more room than what is pending in the queue update |
| * the send_desc_head right away without going through the loop |
| */ |
| if ((hdd_ipa->pending_hw_desc_cnt + pend_q_cnt) < |
| hdd_ipa->hw_desc_cnt) { |
| list_splice_tail_init(&hdd_ipa->pend_desc_head, |
| &send_desc_head->link); |
| cur_send_cnt = pend_q_cnt; |
| hdd_ipa->pend_q_cnt = 0; |
| hdd_ipa->stats.num_rx_ipa_splice++; |
| } else { |
| while (((hdd_ipa->pending_hw_desc_cnt + cur_send_cnt) < |
| hdd_ipa->hw_desc_cnt) && pend_q_cnt > 0) |
| { |
| send_desc = list_first_entry(&hdd_ipa->pend_desc_head, |
| struct ipa_tx_data_desc, link); |
| list_del(&send_desc->link); |
| list_add_tail(&send_desc->link, &send_desc_head->link); |
| cur_send_cnt++; |
| pend_q_cnt--; |
| } |
| hdd_ipa->stats.num_rx_ipa_loop++; |
| |
| hdd_ipa->pend_q_cnt -= cur_send_cnt; |
| |
| VOS_ASSERT(hdd_ipa->pend_q_cnt == pend_q_cnt); |
| } |
| |
| hdd_ipa->pending_hw_desc_cnt += cur_send_cnt; |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| |
| if (ipa_tx_dp_mul(hdd_ipa->prod_client, send_desc_head) != 0) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "ipa_tx_dp_mul failed: %u, q_cnt: %u!", |
| hdd_ipa->pending_hw_desc_cnt, |
| hdd_ipa->pend_q_cnt); |
| goto ipa_tx_failed; |
| } |
| |
| hdd_ipa->stats.num_rx_ipa_tx_dp += cur_send_cnt; |
| if (cur_send_cnt > hdd_ipa->stats.num_max_ipa_tx_mul) |
| hdd_ipa->stats.num_max_ipa_tx_mul = cur_send_cnt; |
| |
| return; |
| |
| ipa_tx_failed: |
| |
| spin_lock_bh(&hdd_ipa->q_lock); |
| hdd_ipa->pending_hw_desc_cnt -= cur_send_cnt; |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| |
| list_for_each_entry_safe(desc, tmp, &send_desc_head->link, link) { |
| list_del(&desc->link); |
| buf = desc->priv; |
| adf_nbuf_free(buf); |
| hdd_ipa_free_data_desc(hdd_ipa, desc); |
| hdd_ipa->stats.num_rx_ipa_tx_dp_err++; |
| } |
| |
| /* Return anchor node */ |
| hdd_ipa_free_data_desc(hdd_ipa, send_desc_head); |
| } |
| |
| |
| static void hdd_ipa_rm_send_pkt_to_ipa(struct work_struct *work) |
| { |
| struct hdd_ipa_priv *hdd_ipa = container_of(work, |
| struct hdd_ipa_priv, rm_work); |
| |
| return hdd_ipa_send_pkt_to_ipa(hdd_ipa); |
| } |
| |
| static void hdd_ipa_rm_notify(void *user_data, enum ipa_rm_event event, |
| unsigned long data) |
| { |
| struct hdd_ipa_priv *hdd_ipa = user_data; |
| |
| if (unlikely(!hdd_ipa)) |
| return; |
| |
| if (!hdd_ipa_is_rm_enabled(hdd_ipa)) |
| return; |
| |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, "Evt: %d", event); |
| |
| switch(event) { |
| case IPA_RM_RESOURCE_GRANTED: |
| #ifdef IPA_UC_OFFLOAD |
| if (hdd_ipa_uc_is_enabled(hdd_ipa)) { |
| /* RM Notification comes with ISR context |
| * it should be serialized into work queue to avoid |
| * ISR sleep problem */ |
| hdd_ipa->uc_rm_work.event = event; |
| schedule_work(&hdd_ipa->uc_rm_work.work); |
| break; |
| } |
| #endif /* IPA_UC_OFFLOAD */ |
| adf_os_spin_lock_bh(&hdd_ipa->rm_lock); |
| hdd_ipa->rm_state = HDD_IPA_RM_GRANTED; |
| adf_os_spin_unlock_bh(&hdd_ipa->rm_lock); |
| hdd_ipa->stats.num_rm_grant++; |
| |
| schedule_work(&hdd_ipa->rm_work); |
| break; |
| case IPA_RM_RESOURCE_RELEASED: |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, "RM Release"); |
| #ifdef IPA_UC_OFFLOAD |
| hdd_ipa->resource_unloading = VOS_FALSE; |
| #endif /* IPA_UC_OFFLOAD */ |
| break; |
| default: |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, "Unknown RM Evt: %d", event); |
| break; |
| } |
| } |
| |
| static int hdd_ipa_rm_cons_release(void) |
| { |
| #ifdef IPA_UC_OFFLOAD |
| /* Do Nothing */ |
| #endif /* IPA_UC_OFFLOAD */ |
| return 0; |
| } |
| |
| static int hdd_ipa_rm_cons_request(void) |
| { |
| int ret = 0; |
| |
| #ifdef IPA_UC_OFFLOAD |
| if (ghdd_ipa->resource_loading) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_FATAL, |
| "%s: ipa resource loading in progress", |
| __func__); |
| ghdd_ipa->pending_cons_req = VOS_TRUE; |
| ret= -EINPROGRESS; |
| } else if (ghdd_ipa->resource_unloading) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_FATAL, |
| "%s: ipa resource unloading in progress", |
| __func__); |
| ghdd_ipa->pending_cons_req = VOS_TRUE; |
| ret = -EPERM; |
| } |
| #endif /* IPA_UC_OFFLOAD */ |
| return ret; |
| } |
| |
| int hdd_ipa_set_perf_level(hdd_context_t *hdd_ctx, uint64_t tx_packets, |
| uint64_t rx_packets) |
| { |
| uint32_t next_cons_bw, next_prod_bw; |
| struct hdd_ipa_priv *hdd_ipa = hdd_ctx->hdd_ipa; |
| struct ipa_rm_perf_profile profile; |
| int ret; |
| |
| if ((!hdd_ipa_is_enabled(hdd_ctx)) || |
| (!hdd_ipa_is_clk_scaling_enabled(hdd_ipa))) |
| return 0; |
| |
| memset(&profile, 0, sizeof(profile)); |
| |
| if (tx_packets > (hdd_ctx->cfg_ini->busBandwidthHighThreshold / 2)) |
| next_cons_bw = hdd_ctx->cfg_ini->IpaHighBandwidthMbps; |
| else if (tx_packets > |
| (hdd_ctx->cfg_ini->busBandwidthMediumThreshold / 2)) |
| next_cons_bw = hdd_ctx->cfg_ini->IpaMediumBandwidthMbps; |
| else |
| next_cons_bw = hdd_ctx->cfg_ini->IpaLowBandwidthMbps; |
| |
| if (rx_packets > (hdd_ctx->cfg_ini->busBandwidthHighThreshold / 2)) |
| next_prod_bw = hdd_ctx->cfg_ini->IpaHighBandwidthMbps; |
| else if (rx_packets > |
| (hdd_ctx->cfg_ini->busBandwidthMediumThreshold / 2)) |
| next_prod_bw = hdd_ctx->cfg_ini->IpaMediumBandwidthMbps; |
| else |
| next_prod_bw = hdd_ctx->cfg_ini->IpaLowBandwidthMbps; |
| |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_DEBUG, |
| "CONS perf curr: %d, next: %d", |
| hdd_ipa->curr_cons_bw, next_cons_bw); |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_DEBUG, |
| "PROD perf curr: %d, next: %d", |
| hdd_ipa->curr_prod_bw, next_prod_bw); |
| |
| if (hdd_ipa->curr_cons_bw != next_cons_bw) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_DEBUG, |
| "Requesting CONS perf curr: %d, next: %d", |
| hdd_ipa->curr_cons_bw, next_cons_bw); |
| profile.max_supported_bandwidth_mbps = next_cons_bw; |
| ret = ipa_rm_set_perf_profile(IPA_RM_RESOURCE_WLAN_CONS, |
| &profile); |
| if (ret) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "RM CONS set perf profile failed: %d", |
| ret); |
| |
| return ret; |
| } |
| hdd_ipa->curr_cons_bw = next_cons_bw; |
| hdd_ipa->stats.num_cons_perf_req++; |
| } |
| |
| if (hdd_ipa->curr_prod_bw != next_prod_bw) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_DEBUG, |
| "Requesting PROD perf curr: %d, next: %d", |
| hdd_ipa->curr_prod_bw, next_prod_bw); |
| profile.max_supported_bandwidth_mbps = next_prod_bw; |
| ret = ipa_rm_set_perf_profile(IPA_RM_RESOURCE_WLAN_PROD, |
| &profile); |
| if (ret) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "RM PROD set perf profile failed: %d", |
| ret); |
| return ret; |
| } |
| hdd_ipa->curr_prod_bw = next_prod_bw; |
| hdd_ipa->stats.num_prod_perf_req++; |
| } |
| |
| return 0; |
| } |
| |
| /** |
| * hdd_ipa_is_present() - get IPA hw status |
| * @hdd_ctx: pointer to hdd context |
| * |
| * ipa_uc_reg_rdyCB is not directly designed to check |
| * ipa hw status. This is an undocumented function which |
| * has confirmed with IPA team. |
| * |
| * Return: true - ipa hw present |
| * false - ipa hw not present |
| */ |
| bool hdd_ipa_is_present(hdd_context_t *hdd_ctx) |
| { |
| /* Check if ipa hw is enabled */ |
| if (ipa_uc_reg_rdyCB(NULL) != -EPERM) |
| return true; |
| else |
| return false; |
| } |
| |
| /** |
| * hdd_ipa_reset_ipaconfig() - reset IpaConfig |
| * @hdd_ctx: pointer to hdd context |
| * @ipaconfig: new value for IpaConfig |
| * |
| * Return: none |
| */ |
| void hdd_ipa_reset_ipaconfig(hdd_context_t *hdd_ctx, v_U32_t ipaconfig) |
| { |
| hdd_ctx->cfg_ini->IpaConfig = ipaconfig; |
| return; |
| } |
| |
| static int hdd_ipa_setup_rm(struct hdd_ipa_priv *hdd_ipa) |
| { |
| struct ipa_rm_create_params create_params = {0}; |
| int ret; |
| |
| if (!hdd_ipa_is_rm_enabled(hdd_ipa)) |
| return 0; |
| |
| vos_init_work(&hdd_ipa->rm_work, hdd_ipa_rm_send_pkt_to_ipa); |
| #ifdef IPA_UC_OFFLOAD |
| vos_init_work(&hdd_ipa->uc_rm_work.work, hdd_ipa_uc_rm_notify_defer); |
| #endif |
| memset(&create_params, 0, sizeof(create_params)); |
| create_params.name = IPA_RM_RESOURCE_WLAN_PROD; |
| create_params.reg_params.user_data = hdd_ipa; |
| create_params.reg_params.notify_cb = hdd_ipa_rm_notify; |
| create_params.floor_voltage = IPA_VOLTAGE_SVS; |
| |
| ret = ipa_rm_create_resource(&create_params); |
| if (ret) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "Create RM resource failed: %d", |
| ret); |
| goto setup_rm_fail; |
| } |
| |
| memset(&create_params, 0, sizeof(create_params)); |
| create_params.name = IPA_RM_RESOURCE_WLAN_CONS; |
| create_params.request_resource= hdd_ipa_rm_cons_request; |
| create_params.release_resource= hdd_ipa_rm_cons_release; |
| create_params.floor_voltage = IPA_VOLTAGE_SVS; |
| |
| ret = ipa_rm_create_resource(&create_params); |
| if (ret) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "Create RM CONS resource failed: %d", ret); |
| goto delete_prod; |
| } |
| |
| ipa_rm_add_dependency(IPA_RM_RESOURCE_WLAN_PROD, |
| IPA_RM_RESOURCE_APPS_CONS); |
| |
| ret = ipa_rm_inactivity_timer_init(IPA_RM_RESOURCE_WLAN_PROD, |
| HDD_IPA_RX_INACTIVITY_MSEC_DELAY); |
| if (ret) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, "Timer init failed: %d", |
| ret); |
| goto timer_init_failed; |
| } |
| |
| /* Set the lowest bandwidth to start with */ |
| ret = hdd_ipa_set_perf_level(hdd_ipa->hdd_ctx, 0, 0); |
| |
| if (ret) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "Set perf level failed: %d", ret); |
| goto set_perf_failed; |
| } |
| |
| vos_wake_lock_init(&hdd_ipa->wake_lock, "wlan_ipa"); |
| vos_init_delayed_work(&hdd_ipa->wake_lock_work, |
| hdd_ipa_wake_lock_timer_func); |
| adf_os_spinlock_init(&hdd_ipa->rm_lock); |
| hdd_ipa->rm_state = HDD_IPA_RM_RELEASED; |
| hdd_ipa->wake_lock_released = true; |
| atomic_set(&hdd_ipa->tx_ref_cnt, 0); |
| |
| return ret; |
| |
| set_perf_failed: |
| ipa_rm_inactivity_timer_destroy(IPA_RM_RESOURCE_WLAN_PROD); |
| |
| timer_init_failed: |
| ipa_rm_delete_resource(IPA_RM_RESOURCE_WLAN_CONS); |
| |
| delete_prod: |
| ipa_rm_delete_resource(IPA_RM_RESOURCE_WLAN_PROD); |
| |
| setup_rm_fail: |
| return ret; |
| } |
| |
| static void hdd_ipa_destory_rm_resource(struct hdd_ipa_priv *hdd_ipa) |
| { |
| int ret; |
| |
| if (!hdd_ipa_is_rm_enabled(hdd_ipa)) |
| return; |
| |
| cancel_delayed_work_sync(&hdd_ipa->wake_lock_work); |
| vos_wake_lock_destroy(&hdd_ipa->wake_lock); |
| |
| #ifdef WLAN_OPEN_SOURCE |
| cancel_work_sync(&hdd_ipa->rm_work); |
| #ifdef IPA_UC_OFFLOAD |
| cancel_work_sync(&hdd_ipa->uc_rm_work.work); |
| #endif |
| #endif |
| adf_os_spinlock_destroy(&hdd_ipa->rm_lock); |
| |
| ipa_rm_inactivity_timer_destroy(IPA_RM_RESOURCE_WLAN_PROD); |
| |
| ret = ipa_rm_delete_resource(IPA_RM_RESOURCE_WLAN_PROD); |
| if (ret) |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "RM PROD resource delete failed %d", ret); |
| |
| ret = ipa_rm_delete_resource(IPA_RM_RESOURCE_WLAN_CONS); |
| if (ret) |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_ERROR, |
| "RM CONS resource delete failed %d", ret); |
| } |
| |
| #define IPA_WLAN_RX_SOFTIRQ_THRESH 16 |
| |
| static void hdd_ipa_send_skb_to_network(adf_nbuf_t skb, hdd_adapter_t *adapter) |
| { |
| int result; |
| #ifndef QCA_CONFIG_SMP |
| struct iphdr* ip_h; |
| static atomic_t softirq_mitigation_cntr = |
| ATOMIC_INIT(IPA_WLAN_RX_SOFTIRQ_THRESH); |
| #endif |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| unsigned int cpu_index; |
| |
| if (!adapter || adapter->magic != WLAN_HDD_ADAPTER_MAGIC) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO_LOW, "Invalid adapter: 0x%pK", |
| adapter); |
| HDD_IPA_INCREASE_INTERNAL_DROP_COUNT(hdd_ipa); |
| adf_nbuf_free(skb); |
| return; |
| } |
| |
| if (hdd_ipa->hdd_ctx->isUnloadInProgress) { |
| HDD_IPA_INCREASE_INTERNAL_DROP_COUNT(hdd_ipa); |
| adf_nbuf_free(skb); |
| return; |
| } |
| |
| skb->destructor = hdd_ipa_uc_rt_debug_destructor; |
| skb->dev = adapter->dev; |
| skb->protocol = eth_type_trans(skb, skb->dev); |
| skb->ip_summed = CHECKSUM_NONE; |
| |
| cpu_index = wlan_hdd_get_cpu(); |
| |
| ++adapter->hdd_stats.hddTxRxStats.rxPackets[cpu_index]; |
| #ifdef QCA_CONFIG_SMP |
| result = netif_rx_ni(skb); |
| #else |
| ip_h = (struct iphdr*)((uint8_t*)skb->data); |
| if ((skb->protocol == htons(ETH_P_IP)) && |
| (ip_h->protocol == IPPROTO_ICMP)) { |
| result = netif_rx_ni(skb); |
| } else { |
| /* Call netif_rx_ni for every IPA_WLAN_RX_SOFTIRQ_THRESH packets |
| * to avoid excessive softirq's. |
| */ |
| if (atomic_dec_and_test(&softirq_mitigation_cntr)){ |
| result = netif_rx_ni(skb); |
| atomic_set(&softirq_mitigation_cntr, |
| IPA_WLAN_RX_SOFTIRQ_THRESH); |
| } else { |
| result = netif_rx(skb); |
| } |
| } |
| #endif |
| if (result == NET_RX_SUCCESS) |
| ++adapter->hdd_stats.hddTxRxStats.rxDelivered[cpu_index]; |
| else |
| ++adapter->hdd_stats.hddTxRxStats.rxRefused[cpu_index]; |
| |
| HDD_IPA_INCREASE_NET_SEND_COUNT(hdd_ipa); |
| adapter->dev->last_rx = jiffies; |
| } |
| |
| VOS_STATUS hdd_ipa_process_rxt(v_VOID_t *vosContext, adf_nbuf_t rx_buf_list, |
| v_U8_t sta_id) |
| { |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| hdd_adapter_t *adapter = NULL; |
| struct hdd_ipa_iface_context *iface_context = NULL; |
| adf_nbuf_t buf, next_buf; |
| uint8_t cur_cnt = 0; |
| struct hdd_ipa_cld_hdr *cld_hdr; |
| struct ipa_tx_data_desc *send_desc = NULL; |
| |
| if (!hdd_ipa_is_enabled(hdd_ipa->hdd_ctx)) |
| return VOS_STATUS_E_INVAL; |
| |
| adapter = hdd_ipa->hdd_ctx->sta_to_adapter[sta_id]; |
| if (!adapter || !adapter->ipa_context || |
| adapter->magic != WLAN_HDD_ADAPTER_MAGIC) { |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO_LOW, "Invalid sta_id: %d", |
| sta_id); |
| hdd_ipa->stats.num_rx_drop++; |
| if (adapter) |
| adapter->stats.rx_dropped++; |
| return VOS_STATUS_E_FAILURE; |
| } |
| |
| iface_context = (struct hdd_ipa_iface_context *) adapter->ipa_context; |
| |
| buf = rx_buf_list; |
| while (buf) { |
| HDD_IPA_DBG_DUMP(VOS_TRACE_LEVEL_DEBUG, "RX data", |
| buf->data, DBG_DUMP_RX_LEN); |
| |
| next_buf = adf_nbuf_queue_next(buf); |
| adf_nbuf_set_next(buf, NULL); |
| |
| adapter->stats.rx_packets++; |
| adapter->stats.rx_bytes += buf->len; |
| /* |
| * we want to send Rx packets to IPA only when it is |
| * IPV4 or IPV6 (if IPV6 is enabled). All other packets |
| * will be sent to network stack directly. |
| */ |
| if (!hdd_ipa_can_send_to_ipa(adapter, hdd_ipa, buf->data)) { |
| iface_context->stats.num_rx_prefilter++; |
| hdd_ipa_send_skb_to_network(buf, adapter); |
| buf = next_buf; |
| continue; |
| } |
| |
| cld_hdr = (struct hdd_ipa_cld_hdr *) skb_push(buf, |
| HDD_IPA_WLAN_CLD_HDR_LEN); |
| cld_hdr->sta_id = sta_id; |
| cld_hdr->iface_id = iface_context->iface_id; |
| |
| send_desc = hdd_ipa_alloc_data_desc(hdd_ipa, 0); |
| if (!send_desc) { |
| adf_nbuf_free(buf); /*No desc available; drop*/ |
| buf = next_buf; |
| iface_context->stats.num_rx_send_desc_err++; |
| continue; |
| } |
| |
| send_desc->priv = buf; |
| send_desc->pyld_buffer = buf->data; |
| send_desc->pyld_len = buf->len; |
| spin_lock_bh(&hdd_ipa->q_lock); |
| list_add_tail(&send_desc->link, &hdd_ipa->pend_desc_head); |
| hdd_ipa->pend_q_cnt++; |
| spin_unlock_bh(&hdd_ipa->q_lock); |
| cur_cnt++; |
| buf = next_buf; |
| } |
| |
| iface_context->stats.num_rx_recv += cur_cnt; |
| if (cur_cnt > 1) |
| iface_context->stats.num_rx_recv_mul++; |
| |
| if (cur_cnt > iface_context->stats.max_rx_mul) |
| iface_context->stats.max_rx_mul = cur_cnt; |
| |
| if (hdd_ipa->pend_q_cnt > hdd_ipa->stats.max_pend_q_cnt) |
| hdd_ipa->stats.max_pend_q_cnt = hdd_ipa->pend_q_cnt; |
| |
| if (cur_cnt && hdd_ipa_rm_request(hdd_ipa) == 0) { |
| hdd_ipa_send_pkt_to_ipa(hdd_ipa); |
| } |
| |
| return VOS_STATUS_SUCCESS; |
| } |
| |
| static void hdd_ipa_set_adapter_ip_filter(hdd_adapter_t *adapter) |
| { |
| struct in_ifaddr **ifap = NULL; |
| struct in_ifaddr *ifa = NULL; |
| struct in_device *in_dev; |
| struct net_device *dev; |
| struct hdd_ipa_iface_context *iface_context; |
| |
| iface_context = (struct hdd_ipa_iface_context *)adapter->ipa_context; |
| dev = adapter->dev; |
| if (!dev || !iface_context) { |
| return; |
| } |
| /* This optimization not needed for Station mode one of |
| * the reason being sta-usb tethered mode |
| */ |
| if (adapter->device_mode == WLAN_HDD_INFRA_STATION) { |
| iface_context->ifa_address = 0; |
| return; |
| } |
| |
| |
| /* Get IP address */ |
| if (dev->priv_flags & IFF_BRIDGE_PORT) { |
| #ifdef WLAN_OPEN_SOURCE |
| rcu_read_lock(); |
| #endif |
| dev = netdev_master_upper_dev_get_rcu(adapter->dev); |
| #ifdef WLAN_OPEN_SOURCE |
| rcu_read_unlock(); |
| #endif |
| if (!dev) |
| return; |
| } |
| if ((in_dev = __in_dev_get_rtnl(dev)) != NULL) { |
| for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL; |
| ifap = &ifa->ifa_next) { |
| if (dev->name && !strcmp(dev->name, ifa->ifa_label)) |
| break; /* found */ |
| } |
| } |
| if(ifa && ifa->ifa_address) { |
| iface_context->ifa_address = ifa->ifa_address; |
| |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO,"%s: %d.%d.%d.%d", dev->name, |
| iface_context->ifa_address & 0x000000ff, |
| iface_context->ifa_address >> 8 & 0x000000ff, |
| iface_context->ifa_address >> 16 & 0x000000ff, |
| iface_context->ifa_address >> 24 & 0x000000ff); |
| } |
| } |
| |
| static int hdd_ipa_ipv4_changed(struct notifier_block *nb, |
| unsigned long data, void *arg) |
| { |
| struct hdd_ipa_priv *hdd_ipa = ghdd_ipa; |
| hdd_adapter_list_node_t *padapter_node = NULL, *pnext = NULL; |
| hdd_adapter_t *padapter; |
| VOS_STATUS status; |
| HDD_IPA_LOG(VOS_TRACE_LEVEL_INFO, |
| "IPv4 Change detected. Updating wlan IPv4 local filters"); |
| |
| status = hdd_get_front_adapter(hdd_ipa->hdd_ctx, &padapter_node); |
| while (padapter_node && VOS_STATUS_SUCCESS == status) { |
| padapter = padapter_node->pAdapter; |
| if (padapter) |
| hdd_ipa_set_adapter_ip_filter(padapter); |
| |
| status = hdd_get_next_adapter(hdd_ipa->hdd_ctx, padapter_node, &pnext); |
| padapter_node = pnext; |
| } |
| return 0; |
| } |
| |
| #if defined(IPA_UC_OFFLOAD) && defined(INTRA_BSS_FWD_OFFLOAD) |
| /** |
| * hdd_ipa_intrabss_forward() - Forward intra bss packets. |
| * @hdd_ipa: pointer to HDD IPA struct |
| * @adapter: hdd adapter pointer |
| * @desc: Firmware descriptor |
| * @skb: Data buffer |
| * |
| * Return |
| * HDD_IPA_FORWARD_PKT_NONE |
| * HDD_IPA_FORWARD_PKT_DISCARD |
| * HDD_IPA_FORWARD_PKT_LOCAL_STACK |
| * |
| */ |
| |