| /****************************************************************************** |
| * |
| * This file is provided under a dual BSD/GPLv2 license. When using or |
| * redistributing this file, you may do so under either license. |
| * |
| * GPL LICENSE SUMMARY |
| * |
| * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. |
| * Copyright(c) 2013 - 2014 Intel Mobile Communications GmbH |
| * Copyright (C) 2015 Intel Deutschland GmbH |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of version 2 of the GNU General Public License as |
| * published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it will be useful, but |
| * WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110, |
| * USA |
| * |
| * The full GNU General Public License is included in this distribution |
| * in the file called COPYING. |
| * |
| * Contact Information: |
| * Intel Linux Wireless <linuxwifi@intel.com> |
| * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 |
| * |
| * BSD LICENSE |
| * |
| * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. |
| * Copyright(c) 2013 - 2014 Intel Mobile Communications GmbH |
| * All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * |
| * * Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * * Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in |
| * the documentation and/or other materials provided with the |
| * distribution. |
| * * Neither the name Intel Corporation nor the names of its |
| * contributors may be used to endorse or promote products derived |
| * from this software without specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| * |
| *****************************************************************************/ |
| #include <net/mac80211.h> |
| |
| #include "iwl-debug.h" |
| #include "iwl-io.h" |
| #include "iwl-prph.h" |
| #include "fw-dbg.h" |
| #include "mvm.h" |
| #include "fw-api-rs.h" |
| |
| /* |
| * Will return 0 even if the cmd failed when RFKILL is asserted unless |
| * CMD_WANT_SKB is set in cmd->flags. |
| */ |
| int iwl_mvm_send_cmd(struct iwl_mvm *mvm, struct iwl_host_cmd *cmd) |
| { |
| int ret; |
| |
| #if defined(CONFIG_IWLWIFI_DEBUGFS) && defined(CONFIG_PM_SLEEP) |
| if (WARN_ON(mvm->d3_test_active)) |
| return -EIO; |
| #endif |
| |
| /* |
| * Synchronous commands from this op-mode must hold |
| * the mutex, this ensures we don't try to send two |
| * (or more) synchronous commands at a time. |
| */ |
| if (!(cmd->flags & CMD_ASYNC)) { |
| lockdep_assert_held(&mvm->mutex); |
| if (!(cmd->flags & CMD_SEND_IN_IDLE)) |
| iwl_mvm_ref(mvm, IWL_MVM_REF_SENDING_CMD); |
| } |
| |
| ret = iwl_trans_send_cmd(mvm->trans, cmd); |
| |
| if (!(cmd->flags & (CMD_ASYNC | CMD_SEND_IN_IDLE))) |
| iwl_mvm_unref(mvm, IWL_MVM_REF_SENDING_CMD); |
| |
| /* |
| * If the caller wants the SKB, then don't hide any problems, the |
| * caller might access the response buffer which will be NULL if |
| * the command failed. |
| */ |
| if (cmd->flags & CMD_WANT_SKB) |
| return ret; |
| |
| /* Silently ignore failures if RFKILL is asserted */ |
| if (!ret || ret == -ERFKILL) |
| return 0; |
| return ret; |
| } |
| |
| int iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id, |
| u32 flags, u16 len, const void *data) |
| { |
| struct iwl_host_cmd cmd = { |
| .id = id, |
| .len = { len, }, |
| .data = { data, }, |
| .flags = flags, |
| }; |
| |
| return iwl_mvm_send_cmd(mvm, &cmd); |
| } |
| |
| /* |
| * We assume that the caller set the status to the success value |
| */ |
| int iwl_mvm_send_cmd_status(struct iwl_mvm *mvm, struct iwl_host_cmd *cmd, |
| u32 *status) |
| { |
| struct iwl_rx_packet *pkt; |
| struct iwl_cmd_response *resp; |
| int ret, resp_len; |
| |
| lockdep_assert_held(&mvm->mutex); |
| |
| #if defined(CONFIG_IWLWIFI_DEBUGFS) && defined(CONFIG_PM_SLEEP) |
| if (WARN_ON(mvm->d3_test_active)) |
| return -EIO; |
| #endif |
| |
| /* |
| * Only synchronous commands can wait for status, |
| * we use WANT_SKB so the caller can't. |
| */ |
| if (WARN_ONCE(cmd->flags & (CMD_ASYNC | CMD_WANT_SKB), |
| "cmd flags %x", cmd->flags)) |
| return -EINVAL; |
| |
| cmd->flags |= CMD_WANT_SKB; |
| |
| ret = iwl_trans_send_cmd(mvm->trans, cmd); |
| if (ret == -ERFKILL) { |
| /* |
| * The command failed because of RFKILL, don't update |
| * the status, leave it as success and return 0. |
| */ |
| return 0; |
| } else if (ret) { |
| return ret; |
| } |
| |
| pkt = cmd->resp_pkt; |
| /* Can happen if RFKILL is asserted */ |
| if (!pkt) { |
| ret = 0; |
| goto out_free_resp; |
| } |
| |
| resp_len = iwl_rx_packet_payload_len(pkt); |
| if (WARN_ON_ONCE(resp_len != sizeof(*resp))) { |
| ret = -EIO; |
| goto out_free_resp; |
| } |
| |
| resp = (void *)pkt->data; |
| *status = le32_to_cpu(resp->status); |
| out_free_resp: |
| iwl_free_resp(cmd); |
| return ret; |
| } |
| |
| /* |
| * We assume that the caller set the status to the sucess value |
| */ |
| int iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id, u16 len, |
| const void *data, u32 *status) |
| { |
| struct iwl_host_cmd cmd = { |
| .id = id, |
| .len = { len, }, |
| .data = { data, }, |
| }; |
| |
| return iwl_mvm_send_cmd_status(mvm, &cmd, status); |
| } |
| |
| #define IWL_DECLARE_RATE_INFO(r) \ |
| [IWL_RATE_##r##M_INDEX] = IWL_RATE_##r##M_PLCP |
| |
| /* |
| * Translate from fw_rate_index (IWL_RATE_XXM_INDEX) to PLCP |
| */ |
| static const u8 fw_rate_idx_to_plcp[IWL_RATE_COUNT] = { |
| IWL_DECLARE_RATE_INFO(1), |
| IWL_DECLARE_RATE_INFO(2), |
| IWL_DECLARE_RATE_INFO(5), |
| IWL_DECLARE_RATE_INFO(11), |
| IWL_DECLARE_RATE_INFO(6), |
| IWL_DECLARE_RATE_INFO(9), |
| IWL_DECLARE_RATE_INFO(12), |
| IWL_DECLARE_RATE_INFO(18), |
| IWL_DECLARE_RATE_INFO(24), |
| IWL_DECLARE_RATE_INFO(36), |
| IWL_DECLARE_RATE_INFO(48), |
| IWL_DECLARE_RATE_INFO(54), |
| }; |
| |
| int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags, |
| enum nl80211_band band) |
| { |
| int rate = rate_n_flags & RATE_LEGACY_RATE_MSK; |
| int idx; |
| int band_offset = 0; |
| |
| /* Legacy rate format, search for match in table */ |
| if (band == NL80211_BAND_5GHZ) |
| band_offset = IWL_FIRST_OFDM_RATE; |
| for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++) |
| if (fw_rate_idx_to_plcp[idx] == rate) |
| return idx - band_offset; |
| |
| return -1; |
| } |
| |
| u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx) |
| { |
| /* Get PLCP rate for tx_cmd->rate_n_flags */ |
| return fw_rate_idx_to_plcp[rate_idx]; |
| } |
| |
| void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb) |
| { |
| struct iwl_rx_packet *pkt = rxb_addr(rxb); |
| struct iwl_error_resp *err_resp = (void *)pkt->data; |
| |
| IWL_ERR(mvm, "FW Error notification: type 0x%08X cmd_id 0x%02X\n", |
| le32_to_cpu(err_resp->error_type), err_resp->cmd_id); |
| IWL_ERR(mvm, "FW Error notification: seq 0x%04X service 0x%08X\n", |
| le16_to_cpu(err_resp->bad_cmd_seq_num), |
| le32_to_cpu(err_resp->error_service)); |
| IWL_ERR(mvm, "FW Error notification: timestamp 0x%16llX\n", |
| le64_to_cpu(err_resp->timestamp)); |
| } |
| |
| /* |
| * Returns the first antenna as ANT_[ABC], as defined in iwl-config.h. |
| * The parameter should also be a combination of ANT_[ABC]. |
| */ |
| u8 first_antenna(u8 mask) |
| { |
| BUILD_BUG_ON(ANT_A != BIT(0)); /* using ffs is wrong if not */ |
| if (WARN_ON_ONCE(!mask)) /* ffs will return 0 if mask is zeroed */ |
| return BIT(0); |
| return BIT(ffs(mask) - 1); |
| } |
| |
| /* |
| * Toggles between TX antennas to send the probe request on. |
| * Receives the bitmask of valid TX antennas and the *index* used |
| * for the last TX, and returns the next valid *index* to use. |
| * In order to set it in the tx_cmd, must do BIT(idx). |
| */ |
| u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx) |
| { |
| u8 ind = last_idx; |
| int i; |
| |
| for (i = 0; i < RATE_MCS_ANT_NUM; i++) { |
| ind = (ind + 1) % RATE_MCS_ANT_NUM; |
| if (valid & BIT(ind)) |
| return ind; |
| } |
| |
| WARN_ONCE(1, "Failed to toggle between antennas 0x%x", valid); |
| return last_idx; |
| } |
| |
| static const struct { |
| const char *name; |
| u8 num; |
| } advanced_lookup[] = { |
| { "NMI_INTERRUPT_WDG", 0x34 }, |
| { "SYSASSERT", 0x35 }, |
| { "UCODE_VERSION_MISMATCH", 0x37 }, |
| { "BAD_COMMAND", 0x38 }, |
| { "NMI_INTERRUPT_DATA_ACTION_PT", 0x3C }, |
| { "FATAL_ERROR", 0x3D }, |
| { "NMI_TRM_HW_ERR", 0x46 }, |
| { "NMI_INTERRUPT_TRM", 0x4C }, |
| { "NMI_INTERRUPT_BREAK_POINT", 0x54 }, |
| { "NMI_INTERRUPT_WDG_RXF_FULL", 0x5C }, |
| { "NMI_INTERRUPT_WDG_NO_RBD_RXF_FULL", 0x64 }, |
| { "NMI_INTERRUPT_HOST", 0x66 }, |
| { "NMI_INTERRUPT_ACTION_PT", 0x7C }, |
| { "NMI_INTERRUPT_UNKNOWN", 0x84 }, |
| { "NMI_INTERRUPT_INST_ACTION_PT", 0x86 }, |
| { "ADVANCED_SYSASSERT", 0 }, |
| }; |
| |
| static const char *desc_lookup(u32 num) |
| { |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(advanced_lookup) - 1; i++) |
| if (advanced_lookup[i].num == num) |
| return advanced_lookup[i].name; |
| |
| /* No entry matches 'num', so it is the last: ADVANCED_SYSASSERT */ |
| return advanced_lookup[i].name; |
| } |
| |
| /* |
| * Note: This structure is read from the device with IO accesses, |
| * and the reading already does the endian conversion. As it is |
| * read with u32-sized accesses, any members with a different size |
| * need to be ordered correctly though! |
| */ |
| struct iwl_error_event_table_v1 { |
| u32 valid; /* (nonzero) valid, (0) log is empty */ |
| u32 error_id; /* type of error */ |
| u32 pc; /* program counter */ |
| u32 blink1; /* branch link */ |
| u32 blink2; /* branch link */ |
| u32 ilink1; /* interrupt link */ |
| u32 ilink2; /* interrupt link */ |
| u32 data1; /* error-specific data */ |
| u32 data2; /* error-specific data */ |
| u32 data3; /* error-specific data */ |
| u32 bcon_time; /* beacon timer */ |
| u32 tsf_low; /* network timestamp function timer */ |
| u32 tsf_hi; /* network timestamp function timer */ |
| u32 gp1; /* GP1 timer register */ |
| u32 gp2; /* GP2 timer register */ |
| u32 gp3; /* GP3 timer register */ |
| u32 ucode_ver; /* uCode version */ |
| u32 hw_ver; /* HW Silicon version */ |
| u32 brd_ver; /* HW board version */ |
| u32 log_pc; /* log program counter */ |
| u32 frame_ptr; /* frame pointer */ |
| u32 stack_ptr; /* stack pointer */ |
| u32 hcmd; /* last host command header */ |
| u32 isr0; /* isr status register LMPM_NIC_ISR0: |
| * rxtx_flag */ |
| u32 isr1; /* isr status register LMPM_NIC_ISR1: |
| * host_flag */ |
| u32 isr2; /* isr status register LMPM_NIC_ISR2: |
| * enc_flag */ |
| u32 isr3; /* isr status register LMPM_NIC_ISR3: |
| * time_flag */ |
| u32 isr4; /* isr status register LMPM_NIC_ISR4: |
| * wico interrupt */ |
| u32 isr_pref; /* isr status register LMPM_NIC_PREF_STAT */ |
| u32 wait_event; /* wait event() caller address */ |
| u32 l2p_control; /* L2pControlField */ |
| u32 l2p_duration; /* L2pDurationField */ |
| u32 l2p_mhvalid; /* L2pMhValidBits */ |
| u32 l2p_addr_match; /* L2pAddrMatchStat */ |
| u32 lmpm_pmg_sel; /* indicate which clocks are turned on |
| * (LMPM_PMG_SEL) */ |
| u32 u_timestamp; /* indicate when the date and time of the |
| * compilation */ |
| u32 flow_handler; /* FH read/write pointers, RX credit */ |
| } __packed /* LOG_ERROR_TABLE_API_S_VER_1 */; |
| |
| struct iwl_error_event_table { |
| u32 valid; /* (nonzero) valid, (0) log is empty */ |
| u32 error_id; /* type of error */ |
| u32 trm_hw_status0; /* TRM HW status */ |
| u32 trm_hw_status1; /* TRM HW status */ |
| u32 blink2; /* branch link */ |
| u32 ilink1; /* interrupt link */ |
| u32 ilink2; /* interrupt link */ |
| u32 data1; /* error-specific data */ |
| u32 data2; /* error-specific data */ |
| u32 data3; /* error-specific data */ |
| u32 bcon_time; /* beacon timer */ |
| u32 tsf_low; /* network timestamp function timer */ |
| u32 tsf_hi; /* network timestamp function timer */ |
| u32 gp1; /* GP1 timer register */ |
| u32 gp2; /* GP2 timer register */ |
| u32 fw_rev_type; /* firmware revision type */ |
| u32 major; /* uCode version major */ |
| u32 minor; /* uCode version minor */ |
| u32 hw_ver; /* HW Silicon version */ |
| u32 brd_ver; /* HW board version */ |
| u32 log_pc; /* log program counter */ |
| u32 frame_ptr; /* frame pointer */ |
| u32 stack_ptr; /* stack pointer */ |
| u32 hcmd; /* last host command header */ |
| u32 isr0; /* isr status register LMPM_NIC_ISR0: |
| * rxtx_flag */ |
| u32 isr1; /* isr status register LMPM_NIC_ISR1: |
| * host_flag */ |
| u32 isr2; /* isr status register LMPM_NIC_ISR2: |
| * enc_flag */ |
| u32 isr3; /* isr status register LMPM_NIC_ISR3: |
| * time_flag */ |
| u32 isr4; /* isr status register LMPM_NIC_ISR4: |
| * wico interrupt */ |
| u32 last_cmd_id; /* last HCMD id handled by the firmware */ |
| u32 wait_event; /* wait event() caller address */ |
| u32 l2p_control; /* L2pControlField */ |
| u32 l2p_duration; /* L2pDurationField */ |
| u32 l2p_mhvalid; /* L2pMhValidBits */ |
| u32 l2p_addr_match; /* L2pAddrMatchStat */ |
| u32 lmpm_pmg_sel; /* indicate which clocks are turned on |
| * (LMPM_PMG_SEL) */ |
| u32 u_timestamp; /* indicate when the date and time of the |
| * compilation */ |
| u32 flow_handler; /* FH read/write pointers, RX credit */ |
| } __packed /* LOG_ERROR_TABLE_API_S_VER_3 */; |
| |
| /* |
| * UMAC error struct - relevant starting from family 8000 chip. |
| * Note: This structure is read from the device with IO accesses, |
| * and the reading already does the endian conversion. As it is |
| * read with u32-sized accesses, any members with a different size |
| * need to be ordered correctly though! |
| */ |
| struct iwl_umac_error_event_table { |
| u32 valid; /* (nonzero) valid, (0) log is empty */ |
| u32 error_id; /* type of error */ |
| u32 blink1; /* branch link */ |
| u32 blink2; /* branch link */ |
| u32 ilink1; /* interrupt link */ |
| u32 ilink2; /* interrupt link */ |
| u32 data1; /* error-specific data */ |
| u32 data2; /* error-specific data */ |
| u32 data3; /* error-specific data */ |
| u32 umac_major; |
| u32 umac_minor; |
| u32 frame_pointer; /* core register 27*/ |
| u32 stack_pointer; /* core register 28 */ |
| u32 cmd_header; /* latest host cmd sent to UMAC */ |
| u32 nic_isr_pref; /* ISR status register */ |
| } __packed; |
| |
| #define ERROR_START_OFFSET (1 * sizeof(u32)) |
| #define ERROR_ELEM_SIZE (7 * sizeof(u32)) |
| |
| static void iwl_mvm_dump_umac_error_log(struct iwl_mvm *mvm) |
| { |
| struct iwl_trans *trans = mvm->trans; |
| struct iwl_umac_error_event_table table; |
| u32 base; |
| |
| base = mvm->umac_error_event_table; |
| |
| if (base < 0x800000) { |
| IWL_ERR(mvm, |
| "Not valid error log pointer 0x%08X for %s uCode\n", |
| base, |
| (mvm->cur_ucode == IWL_UCODE_INIT) |
| ? "Init" : "RT"); |
| return; |
| } |
| |
| iwl_trans_read_mem_bytes(trans, base, &table, sizeof(table)); |
| |
| if (ERROR_START_OFFSET <= table.valid * ERROR_ELEM_SIZE) { |
| IWL_ERR(trans, "Start IWL Error Log Dump:\n"); |
| IWL_ERR(trans, "Status: 0x%08lX, count: %d\n", |
| mvm->status, table.valid); |
| } |
| |
| IWL_ERR(mvm, "0x%08X | %s\n", table.error_id, |
| desc_lookup(table.error_id)); |
| IWL_ERR(mvm, "0x%08X | umac branchlink1\n", table.blink1); |
| IWL_ERR(mvm, "0x%08X | umac branchlink2\n", table.blink2); |
| IWL_ERR(mvm, "0x%08X | umac interruptlink1\n", table.ilink1); |
| IWL_ERR(mvm, "0x%08X | umac interruptlink2\n", table.ilink2); |
| IWL_ERR(mvm, "0x%08X | umac data1\n", table.data1); |
| IWL_ERR(mvm, "0x%08X | umac data2\n", table.data2); |
| IWL_ERR(mvm, "0x%08X | umac data3\n", table.data3); |
| IWL_ERR(mvm, "0x%08X | umac major\n", table.umac_major); |
| IWL_ERR(mvm, "0x%08X | umac minor\n", table.umac_minor); |
| IWL_ERR(mvm, "0x%08X | frame pointer\n", table.frame_pointer); |
| IWL_ERR(mvm, "0x%08X | stack pointer\n", table.stack_pointer); |
| IWL_ERR(mvm, "0x%08X | last host cmd\n", table.cmd_header); |
| IWL_ERR(mvm, "0x%08X | isr status reg\n", table.nic_isr_pref); |
| } |
| |
| void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm) |
| { |
| struct iwl_trans *trans = mvm->trans; |
| struct iwl_error_event_table table; |
| u32 base; |
| |
| base = mvm->error_event_table; |
| if (mvm->cur_ucode == IWL_UCODE_INIT) { |
| if (!base) |
| base = mvm->fw->init_errlog_ptr; |
| } else { |
| if (!base) |
| base = mvm->fw->inst_errlog_ptr; |
| } |
| |
| if (base < 0x400000) { |
| IWL_ERR(mvm, |
| "Not valid error log pointer 0x%08X for %s uCode\n", |
| base, |
| (mvm->cur_ucode == IWL_UCODE_INIT) |
| ? "Init" : "RT"); |
| return; |
| } |
| |
| iwl_trans_read_mem_bytes(trans, base, &table, sizeof(table)); |
| |
| if (ERROR_START_OFFSET <= table.valid * ERROR_ELEM_SIZE) { |
| IWL_ERR(trans, "Start IWL Error Log Dump:\n"); |
| IWL_ERR(trans, "Status: 0x%08lX, count: %d\n", |
| mvm->status, table.valid); |
| } |
| |
| /* Do not change this output - scripts rely on it */ |
| |
| IWL_ERR(mvm, "Loaded firmware version: %s\n", mvm->fw->fw_version); |
| |
| trace_iwlwifi_dev_ucode_error(trans->dev, table.error_id, table.tsf_low, |
| table.data1, table.data2, table.data3, |
| table.blink2, table.ilink1, |
| table.ilink2, table.bcon_time, table.gp1, |
| table.gp2, table.fw_rev_type, table.major, |
| table.minor, table.hw_ver, table.brd_ver); |
| IWL_ERR(mvm, "0x%08X | %-28s\n", table.error_id, |
| desc_lookup(table.error_id)); |
| IWL_ERR(mvm, "0x%08X | trm_hw_status0\n", table.trm_hw_status0); |
| IWL_ERR(mvm, "0x%08X | trm_hw_status1\n", table.trm_hw_status1); |
| IWL_ERR(mvm, "0x%08X | branchlink2\n", table.blink2); |
| IWL_ERR(mvm, "0x%08X | interruptlink1\n", table.ilink1); |
| IWL_ERR(mvm, "0x%08X | interruptlink2\n", table.ilink2); |
| IWL_ERR(mvm, "0x%08X | data1\n", table.data1); |
| IWL_ERR(mvm, "0x%08X | data2\n", table.data2); |
| IWL_ERR(mvm, "0x%08X | data3\n", table.data3); |
| IWL_ERR(mvm, "0x%08X | beacon time\n", table.bcon_time); |
| IWL_ERR(mvm, "0x%08X | tsf low\n", table.tsf_low); |
| IWL_ERR(mvm, "0x%08X | tsf hi\n", table.tsf_hi); |
| IWL_ERR(mvm, "0x%08X | time gp1\n", table.gp1); |
| IWL_ERR(mvm, "0x%08X | time gp2\n", table.gp2); |
| IWL_ERR(mvm, "0x%08X | uCode revision type\n", table.fw_rev_type); |
| IWL_ERR(mvm, "0x%08X | uCode version major\n", table.major); |
| IWL_ERR(mvm, "0x%08X | uCode version minor\n", table.minor); |
| IWL_ERR(mvm, "0x%08X | hw version\n", table.hw_ver); |
| IWL_ERR(mvm, "0x%08X | board version\n", table.brd_ver); |
| IWL_ERR(mvm, "0x%08X | hcmd\n", table.hcmd); |
| IWL_ERR(mvm, "0x%08X | isr0\n", table.isr0); |
| IWL_ERR(mvm, "0x%08X | isr1\n", table.isr1); |
| IWL_ERR(mvm, "0x%08X | isr2\n", table.isr2); |
| IWL_ERR(mvm, "0x%08X | isr3\n", table.isr3); |
| IWL_ERR(mvm, "0x%08X | isr4\n", table.isr4); |
| IWL_ERR(mvm, "0x%08X | last cmd Id\n", table.last_cmd_id); |
| IWL_ERR(mvm, "0x%08X | wait_event\n", table.wait_event); |
| IWL_ERR(mvm, "0x%08X | l2p_control\n", table.l2p_control); |
| IWL_ERR(mvm, "0x%08X | l2p_duration\n", table.l2p_duration); |
| IWL_ERR(mvm, "0x%08X | l2p_mhvalid\n", table.l2p_mhvalid); |
| IWL_ERR(mvm, "0x%08X | l2p_addr_match\n", table.l2p_addr_match); |
| IWL_ERR(mvm, "0x%08X | lmpm_pmg_sel\n", table.lmpm_pmg_sel); |
| IWL_ERR(mvm, "0x%08X | timestamp\n", table.u_timestamp); |
| IWL_ERR(mvm, "0x%08X | flow_handler\n", table.flow_handler); |
| |
| if (mvm->support_umac_log) |
| iwl_mvm_dump_umac_error_log(mvm); |
| } |
| |
| int iwl_mvm_find_free_queue(struct iwl_mvm *mvm, u8 sta_id, u8 minq, u8 maxq) |
| { |
| int i; |
| |
| lockdep_assert_held(&mvm->queue_info_lock); |
| |
| /* Start by looking for a free queue */ |
| for (i = minq; i <= maxq; i++) |
| if (mvm->queue_info[i].hw_queue_refcount == 0 && |
| mvm->queue_info[i].status == IWL_MVM_QUEUE_FREE) |
| return i; |
| |
| /* |
| * If no free queue found - settle for an inactive one to reconfigure |
| * Make sure that the inactive queue either already belongs to this STA, |
| * or that if it belongs to another one - it isn't the reserved queue |
| */ |
| for (i = minq; i <= maxq; i++) |
| if (mvm->queue_info[i].status == IWL_MVM_QUEUE_INACTIVE && |
| (sta_id == mvm->queue_info[i].ra_sta_id || |
| !mvm->queue_info[i].reserved)) |
| return i; |
| |
| return -ENOSPC; |
| } |
| |
| int iwl_mvm_reconfig_scd(struct iwl_mvm *mvm, int queue, int fifo, int sta_id, |
| int tid, int frame_limit, u16 ssn) |
| { |
| struct iwl_scd_txq_cfg_cmd cmd = { |
| .scd_queue = queue, |
| .action = SCD_CFG_ENABLE_QUEUE, |
| .window = frame_limit, |
| .sta_id = sta_id, |
| .ssn = cpu_to_le16(ssn), |
| .tx_fifo = fifo, |
| .aggregate = (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE || |
| queue == IWL_MVM_DQA_BSS_CLIENT_QUEUE), |
| .tid = tid, |
| }; |
| int ret; |
| |
| spin_lock_bh(&mvm->queue_info_lock); |
| if (WARN(mvm->queue_info[queue].hw_queue_refcount == 0, |
| "Trying to reconfig unallocated queue %d\n", queue)) { |
| spin_unlock_bh(&mvm->queue_info_lock); |
| return -ENXIO; |
| } |
| spin_unlock_bh(&mvm->queue_info_lock); |
| |
| IWL_DEBUG_TX_QUEUES(mvm, "Reconfig SCD for TXQ #%d\n", queue); |
| |
| ret = iwl_mvm_send_cmd_pdu(mvm, SCD_QUEUE_CFG, 0, sizeof(cmd), &cmd); |
| WARN_ONCE(ret, "Failed to re-configure queue %d on FIFO %d, ret=%d\n", |
| queue, fifo, ret); |
| |
| return ret; |
| } |
| |
| void iwl_mvm_enable_txq(struct iwl_mvm *mvm, int queue, int mac80211_queue, |
| u16 ssn, const struct iwl_trans_txq_scd_cfg *cfg, |
| unsigned int wdg_timeout) |
| { |
| bool enable_queue = true; |
| |
| spin_lock_bh(&mvm->queue_info_lock); |
| |
| /* Make sure this TID isn't already enabled */ |
| if (mvm->queue_info[queue].tid_bitmap & BIT(cfg->tid)) { |
| spin_unlock_bh(&mvm->queue_info_lock); |
| IWL_ERR(mvm, "Trying to enable TXQ with existing TID %d\n", |
| cfg->tid); |
| return; |
| } |
| |
| /* Update mappings and refcounts */ |
| if (mvm->queue_info[queue].hw_queue_refcount > 0) |
| enable_queue = false; |
| |
| mvm->queue_info[queue].hw_queue_to_mac80211 |= BIT(mac80211_queue); |
| mvm->queue_info[queue].hw_queue_refcount++; |
| mvm->queue_info[queue].tid_bitmap |= BIT(cfg->tid); |
| mvm->queue_info[queue].ra_sta_id = cfg->sta_id; |
| |
| if (enable_queue) { |
| if (cfg->tid != IWL_MAX_TID_COUNT) |
| mvm->queue_info[queue].mac80211_ac = |
| tid_to_mac80211_ac[cfg->tid]; |
| else |
| mvm->queue_info[queue].mac80211_ac = IEEE80211_AC_VO; |
| |
| mvm->queue_info[queue].txq_tid = cfg->tid; |
| } |
| |
| IWL_DEBUG_TX_QUEUES(mvm, |
| "Enabling TXQ #%d refcount=%d (mac80211 map:0x%x)\n", |
| queue, mvm->queue_info[queue].hw_queue_refcount, |
| mvm->queue_info[queue].hw_queue_to_mac80211); |
| |
| spin_unlock_bh(&mvm->queue_info_lock); |
| |
| /* Send the enabling command if we need to */ |
| if (enable_queue) { |
| struct iwl_scd_txq_cfg_cmd cmd = { |
| .scd_queue = queue, |
| .action = SCD_CFG_ENABLE_QUEUE, |
| .window = cfg->frame_limit, |
| .sta_id = cfg->sta_id, |
| .ssn = cpu_to_le16(ssn), |
| .tx_fifo = cfg->fifo, |
| .aggregate = cfg->aggregate, |
| .tid = cfg->tid, |
| }; |
| |
| /* Set sta_id in the command, if it exists */ |
| if (iwl_mvm_is_dqa_supported(mvm)) |
| cmd.sta_id = cfg->sta_id; |
| |
| iwl_trans_txq_enable_cfg(mvm->trans, queue, ssn, NULL, |
| wdg_timeout); |
| WARN(iwl_mvm_send_cmd_pdu(mvm, SCD_QUEUE_CFG, 0, sizeof(cmd), |
| &cmd), |
| "Failed to configure queue %d on FIFO %d\n", queue, |
| cfg->fifo); |
| } |
| } |
| |
| void iwl_mvm_disable_txq(struct iwl_mvm *mvm, int queue, int mac80211_queue, |
| u8 tid, u8 flags) |
| { |
| struct iwl_scd_txq_cfg_cmd cmd = { |
| .scd_queue = queue, |
| .action = SCD_CFG_DISABLE_QUEUE, |
| }; |
| bool remove_mac_queue = true; |
| int ret; |
| |
| spin_lock_bh(&mvm->queue_info_lock); |
| |
| if (WARN_ON(mvm->queue_info[queue].hw_queue_refcount == 0)) { |
| spin_unlock_bh(&mvm->queue_info_lock); |
| return; |
| } |
| |
| mvm->queue_info[queue].tid_bitmap &= ~BIT(tid); |
| |
| /* |
| * If there is another TID with the same AC - don't remove the MAC queue |
| * from the mapping |
| */ |
| if (tid < IWL_MAX_TID_COUNT) { |
| unsigned long tid_bitmap = |
| mvm->queue_info[queue].tid_bitmap; |
| int ac = tid_to_mac80211_ac[tid]; |
| int i; |
| |
| for_each_set_bit(i, &tid_bitmap, IWL_MAX_TID_COUNT) { |
| if (tid_to_mac80211_ac[i] == ac) |
| remove_mac_queue = false; |
| } |
| } |
| |
| if (remove_mac_queue) |
| mvm->queue_info[queue].hw_queue_to_mac80211 &= |
| ~BIT(mac80211_queue); |
| mvm->queue_info[queue].hw_queue_refcount--; |
| |
| cmd.action = mvm->queue_info[queue].hw_queue_refcount ? |
| SCD_CFG_ENABLE_QUEUE : SCD_CFG_DISABLE_QUEUE; |
| if (cmd.action == SCD_CFG_DISABLE_QUEUE) |
| mvm->queue_info[queue].status = IWL_MVM_QUEUE_FREE; |
| |
| IWL_DEBUG_TX_QUEUES(mvm, |
| "Disabling TXQ #%d refcount=%d (mac80211 map:0x%x)\n", |
| queue, |
| mvm->queue_info[queue].hw_queue_refcount, |
| mvm->queue_info[queue].hw_queue_to_mac80211); |
| |
| /* If the queue is still enabled - nothing left to do in this func */ |
| if (cmd.action == SCD_CFG_ENABLE_QUEUE) { |
| spin_unlock_bh(&mvm->queue_info_lock); |
| return; |
| } |
| |
| cmd.sta_id = mvm->queue_info[queue].ra_sta_id; |
| cmd.tid = mvm->queue_info[queue].txq_tid; |
| |
| /* Make sure queue info is correct even though we overwrite it */ |
| WARN(mvm->queue_info[queue].hw_queue_refcount || |
| mvm->queue_info[queue].tid_bitmap || |
| mvm->queue_info[queue].hw_queue_to_mac80211, |
| "TXQ #%d info out-of-sync - refcount=%d, mac map=0x%x, tid=0x%x\n", |
| queue, mvm->queue_info[queue].hw_queue_refcount, |
| mvm->queue_info[queue].hw_queue_to_mac80211, |
| mvm->queue_info[queue].tid_bitmap); |
| |
| /* If we are here - the queue is freed and we can zero out these vals */ |
| mvm->queue_info[queue].hw_queue_refcount = 0; |
| mvm->queue_info[queue].tid_bitmap = 0; |
| mvm->queue_info[queue].hw_queue_to_mac80211 = 0; |
| |
| /* Regardless if this is a reserved TXQ for a STA - mark it as false */ |
| mvm->queue_info[queue].reserved = false; |
| |
| spin_unlock_bh(&mvm->queue_info_lock); |
| |
| iwl_trans_txq_disable(mvm->trans, queue, false); |
| ret = iwl_mvm_send_cmd_pdu(mvm, SCD_QUEUE_CFG, flags, |
| sizeof(cmd), &cmd); |
| if (ret) |
| IWL_ERR(mvm, "Failed to disable queue %d (ret=%d)\n", |
| queue, ret); |
| } |
| |
| /** |
| * iwl_mvm_send_lq_cmd() - Send link quality command |
| * @init: This command is sent as part of station initialization right |
| * after station has been added. |
| * |
| * The link quality command is sent as the last step of station creation. |
| * This is the special case in which init is set and we call a callback in |
| * this case to clear the state indicating that station creation is in |
| * progress. |
| */ |
| int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq, bool init) |
| { |
| struct iwl_host_cmd cmd = { |
| .id = LQ_CMD, |
| .len = { sizeof(struct iwl_lq_cmd), }, |
| .flags = init ? 0 : CMD_ASYNC, |
| .data = { lq, }, |
| }; |
| |
| if (WARN_ON(lq->sta_id == IWL_MVM_STATION_COUNT)) |
| return -EINVAL; |
| |
| return iwl_mvm_send_cmd(mvm, &cmd); |
| } |
| |
| /** |
| * iwl_mvm_update_smps - Get a request to change the SMPS mode |
| * @req_type: The part of the driver who call for a change. |
| * @smps_requests: The request to change the SMPS mode. |
| * |
| * Get a requst to change the SMPS mode, |
| * and change it according to all other requests in the driver. |
| */ |
| void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif, |
| enum iwl_mvm_smps_type_request req_type, |
| enum ieee80211_smps_mode smps_request) |
| { |
| struct iwl_mvm_vif *mvmvif; |
| enum ieee80211_smps_mode smps_mode; |
| int i; |
| |
| lockdep_assert_held(&mvm->mutex); |
| |
| /* SMPS is irrelevant for NICs that don't have at least 2 RX antenna */ |
| if (num_of_ant(iwl_mvm_get_valid_rx_ant(mvm)) == 1) |
| return; |
| |
| if (vif->type == NL80211_IFTYPE_AP) |
| smps_mode = IEEE80211_SMPS_OFF; |
| else |
| smps_mode = IEEE80211_SMPS_AUTOMATIC; |
| |
| mvmvif = iwl_mvm_vif_from_mac80211(vif); |
| mvmvif->smps_requests[req_type] = smps_request; |
| for (i = 0; i < NUM_IWL_MVM_SMPS_REQ; i++) { |
| if (mvmvif->smps_requests[i] == IEEE80211_SMPS_STATIC) { |
| smps_mode = IEEE80211_SMPS_STATIC; |
| break; |
| } |
| if (mvmvif->smps_requests[i] == IEEE80211_SMPS_DYNAMIC) |
| smps_mode = IEEE80211_SMPS_DYNAMIC; |
| } |
| |
| ieee80211_request_smps(vif, smps_mode); |
| } |
| |
| int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear) |
| { |
| struct iwl_statistics_cmd scmd = { |
| .flags = clear ? cpu_to_le32(IWL_STATISTICS_FLG_CLEAR) : 0, |
| }; |
| struct iwl_host_cmd cmd = { |
| .id = STATISTICS_CMD, |
| .len[0] = sizeof(scmd), |
| .data[0] = &scmd, |
| .flags = CMD_WANT_SKB, |
| }; |
| int ret; |
| |
| ret = iwl_mvm_send_cmd(mvm, &cmd); |
| if (ret) |
| return ret; |
| |
| iwl_mvm_handle_rx_statistics(mvm, cmd.resp_pkt); |
| iwl_free_resp(&cmd); |
| |
| if (clear) |
| iwl_mvm_accu_radio_stats(mvm); |
| |
| return 0; |
| } |
| |
| void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm) |
| { |
| mvm->accu_radio_stats.rx_time += mvm->radio_stats.rx_time; |
| mvm->accu_radio_stats.tx_time += mvm->radio_stats.tx_time; |
| mvm->accu_radio_stats.on_time_rf += mvm->radio_stats.on_time_rf; |
| mvm->accu_radio_stats.on_time_scan += mvm->radio_stats.on_time_scan; |
| } |
| |
| static void iwl_mvm_diversity_iter(void *_data, u8 *mac, |
| struct ieee80211_vif *vif) |
| { |
| struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); |
| bool *result = _data; |
| int i; |
| |
| for (i = 0; i < NUM_IWL_MVM_SMPS_REQ; i++) { |
| if (mvmvif->smps_requests[i] == IEEE80211_SMPS_STATIC || |
| mvmvif->smps_requests[i] == IEEE80211_SMPS_DYNAMIC) |
| *result = false; |
| } |
| } |
| |
| bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm) |
| { |
| bool result = true; |
| |
| lockdep_assert_held(&mvm->mutex); |
| |
| if (num_of_ant(iwl_mvm_get_valid_rx_ant(mvm)) == 1) |
| return false; |
| |
| if (mvm->cfg->rx_with_siso_diversity) |
| return false; |
| |
| ieee80211_iterate_active_interfaces_atomic( |
| mvm->hw, IEEE80211_IFACE_ITER_NORMAL, |
| iwl_mvm_diversity_iter, &result); |
| |
| return result; |
| } |
| |
| int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif, |
| bool prev) |
| { |
| struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); |
| int res; |
| |
| lockdep_assert_held(&mvm->mutex); |
| |
| if (iwl_mvm_vif_low_latency(mvmvif) == prev) |
| return 0; |
| |
| res = iwl_mvm_update_quotas(mvm, false, NULL); |
| if (res) |
| return res; |
| |
| iwl_mvm_bt_coex_vif_change(mvm); |
| |
| return iwl_mvm_power_update_mac(mvm); |
| } |
| |
| static void iwl_mvm_ll_iter(void *_data, u8 *mac, struct ieee80211_vif *vif) |
| { |
| bool *result = _data; |
| |
| if (iwl_mvm_vif_low_latency(iwl_mvm_vif_from_mac80211(vif))) |
| *result = true; |
| } |
| |
| bool iwl_mvm_low_latency(struct iwl_mvm *mvm) |
| { |
| bool result = false; |
| |
| ieee80211_iterate_active_interfaces_atomic( |
| mvm->hw, IEEE80211_IFACE_ITER_NORMAL, |
| iwl_mvm_ll_iter, &result); |
| |
| return result; |
| } |
| |
| struct iwl_bss_iter_data { |
| struct ieee80211_vif *vif; |
| bool error; |
| }; |
| |
| static void iwl_mvm_bss_iface_iterator(void *_data, u8 *mac, |
| struct ieee80211_vif *vif) |
| { |
| struct iwl_bss_iter_data *data = _data; |
| |
| if (vif->type != NL80211_IFTYPE_STATION || vif->p2p) |
| return; |
| |
| if (data->vif) { |
| data->error = true; |
| return; |
| } |
| |
| data->vif = vif; |
| } |
| |
| struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm) |
| { |
| struct iwl_bss_iter_data bss_iter_data = {}; |
| |
| ieee80211_iterate_active_interfaces_atomic( |
| mvm->hw, IEEE80211_IFACE_ITER_NORMAL, |
| iwl_mvm_bss_iface_iterator, &bss_iter_data); |
| |
| if (bss_iter_data.error) { |
| IWL_ERR(mvm, "More than one managed interface active!\n"); |
| return ERR_PTR(-EINVAL); |
| } |
| |
| return bss_iter_data.vif; |
| } |
| |
| unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm, |
| struct ieee80211_vif *vif, |
| bool tdls, bool cmd_q) |
| { |
| struct iwl_fw_dbg_trigger_tlv *trigger; |
| struct iwl_fw_dbg_trigger_txq_timer *txq_timer; |
| unsigned int default_timeout = |
| cmd_q ? IWL_DEF_WD_TIMEOUT : mvm->cfg->base_params->wd_timeout; |
| |
| if (!iwl_fw_dbg_trigger_enabled(mvm->fw, FW_DBG_TRIGGER_TXQ_TIMERS)) |
| return iwlmvm_mod_params.tfd_q_hang_detect ? |
| default_timeout : IWL_WATCHDOG_DISABLED; |
| |
| trigger = iwl_fw_dbg_get_trigger(mvm->fw, FW_DBG_TRIGGER_TXQ_TIMERS); |
| txq_timer = (void *)trigger->data; |
| |
| if (tdls) |
| return le32_to_cpu(txq_timer->tdls); |
| |
| if (cmd_q) |
| return le32_to_cpu(txq_timer->command_queue); |
| |
| if (WARN_ON(!vif)) |
| return default_timeout; |
| |
| switch (ieee80211_vif_type_p2p(vif)) { |
| case NL80211_IFTYPE_ADHOC: |
| return le32_to_cpu(txq_timer->ibss); |
| case NL80211_IFTYPE_STATION: |
| return le32_to_cpu(txq_timer->bss); |
| case NL80211_IFTYPE_AP: |
| return le32_to_cpu(txq_timer->softap); |
| case NL80211_IFTYPE_P2P_CLIENT: |
| return le32_to_cpu(txq_timer->p2p_client); |
| case NL80211_IFTYPE_P2P_GO: |
| return le32_to_cpu(txq_timer->p2p_go); |
| case NL80211_IFTYPE_P2P_DEVICE: |
| return le32_to_cpu(txq_timer->p2p_device); |
| default: |
| WARN_ON(1); |
| return mvm->cfg->base_params->wd_timeout; |
| } |
| } |
| |
| void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif, |
| const char *errmsg) |
| { |
| struct iwl_fw_dbg_trigger_tlv *trig; |
| struct iwl_fw_dbg_trigger_mlme *trig_mlme; |
| |
| if (!iwl_fw_dbg_trigger_enabled(mvm->fw, FW_DBG_TRIGGER_MLME)) |
| goto out; |
| |
| trig = iwl_fw_dbg_get_trigger(mvm->fw, FW_DBG_TRIGGER_MLME); |
| trig_mlme = (void *)trig->data; |
| if (!iwl_fw_dbg_trigger_check_stop(mvm, vif, trig)) |
| goto out; |
| |
| if (trig_mlme->stop_connection_loss && |
| --trig_mlme->stop_connection_loss) |
| goto out; |
| |
| iwl_mvm_fw_dbg_collect_trig(mvm, trig, "%s", errmsg); |
| |
| out: |
| ieee80211_connection_loss(vif); |
| } |
| |
| /* |
| * Remove inactive TIDs of a given queue. |
| * If all queue TIDs are inactive - mark the queue as inactive |
| * If only some the queue TIDs are inactive - unmap them from the queue |
| */ |
| static void iwl_mvm_remove_inactive_tids(struct iwl_mvm *mvm, |
| struct iwl_mvm_sta *mvmsta, int queue, |
| unsigned long tid_bitmap) |
| { |
| int tid; |
| |
| lockdep_assert_held(&mvmsta->lock); |
| lockdep_assert_held(&mvm->queue_info_lock); |
| |
| /* Go over all non-active TIDs, incl. IWL_MAX_TID_COUNT (for mgmt) */ |
| for_each_set_bit(tid, &tid_bitmap, IWL_MAX_TID_COUNT + 1) { |
| /* If some TFDs are still queued - don't mark TID as inactive */ |
| if (iwl_mvm_tid_queued(&mvmsta->tid_data[tid])) |
| tid_bitmap &= ~BIT(tid); |
| } |
| |
| /* If all TIDs in the queue are inactive - mark queue as inactive. */ |
| if (tid_bitmap == mvm->queue_info[queue].tid_bitmap) { |
| mvm->queue_info[queue].status = IWL_MVM_QUEUE_INACTIVE; |
| |
| for_each_set_bit(tid, &tid_bitmap, IWL_MAX_TID_COUNT + 1) |
| mvmsta->tid_data[tid].is_tid_active = false; |
| |
| IWL_DEBUG_TX_QUEUES(mvm, "Queue %d marked as inactive\n", |
| queue); |
| return; |
| } |
| |
| /* |
| * If we are here, this is a shared queue and not all TIDs timed-out. |
| * Remove the ones that did. |
| */ |
| for_each_set_bit(tid, &tid_bitmap, IWL_MAX_TID_COUNT + 1) { |
| int mac_queue = mvmsta->vif->hw_queue[tid_to_mac80211_ac[tid]]; |
| |
| mvmsta->tid_data[tid].txq_id = IEEE80211_INVAL_HW_QUEUE; |
| mvm->queue_info[queue].hw_queue_to_mac80211 &= ~BIT(mac_queue); |
| mvm->queue_info[queue].hw_queue_refcount--; |
| mvm->queue_info[queue].tid_bitmap &= ~BIT(tid); |
| mvmsta->tid_data[tid].is_tid_active = false; |
| |
| IWL_DEBUG_TX_QUEUES(mvm, |
| "Removing inactive TID %d from shared Q:%d\n", |
| tid, queue); |
| } |
| |
| IWL_DEBUG_TX_QUEUES(mvm, |
| "TXQ #%d left with tid bitmap 0x%x\n", queue, |
| mvm->queue_info[queue].tid_bitmap); |
| |
| /* |
| * There may be different TIDs with the same mac queues, so make |
| * sure all TIDs have existing corresponding mac queues enabled |
| */ |
| tid_bitmap = mvm->queue_info[queue].tid_bitmap; |
| for_each_set_bit(tid, &tid_bitmap, IWL_MAX_TID_COUNT + 1) { |
| mvm->queue_info[queue].hw_queue_to_mac80211 |= |
| BIT(mvmsta->vif->hw_queue[tid_to_mac80211_ac[tid]]); |
| } |
| |
| /* If the queue is marked as shared - "unshare" it */ |
| if (mvm->queue_info[queue].hw_queue_refcount == 1 && |
| mvm->queue_info[queue].status == IWL_MVM_QUEUE_SHARED) { |
| mvm->queue_info[queue].status = IWL_MVM_QUEUE_RECONFIGURING; |
| IWL_DEBUG_TX_QUEUES(mvm, "Marking Q:%d for reconfig\n", |
| queue); |
| } |
| } |
| |
| void iwl_mvm_inactivity_check(struct iwl_mvm *mvm) |
| { |
| unsigned long timeout_queues_map = 0; |
| unsigned long now = jiffies; |
| int i; |
| |
| spin_lock_bh(&mvm->queue_info_lock); |
| for (i = 0; i < IWL_MAX_HW_QUEUES; i++) |
| if (mvm->queue_info[i].hw_queue_refcount > 0) |
| timeout_queues_map |= BIT(i); |
| spin_unlock_bh(&mvm->queue_info_lock); |
| |
| rcu_read_lock(); |
| |
| /* |
| * If a queue time outs - mark it as INACTIVE (don't remove right away |
| * if we don't have to.) This is an optimization in case traffic comes |
| * later, and we don't HAVE to use a currently-inactive queue |
| */ |
| for_each_set_bit(i, &timeout_queues_map, IWL_MAX_HW_QUEUES) { |
| struct ieee80211_sta *sta; |
| struct iwl_mvm_sta *mvmsta; |
| u8 sta_id; |
| int tid; |
| unsigned long inactive_tid_bitmap = 0; |
| unsigned long queue_tid_bitmap; |
| |
| spin_lock_bh(&mvm->queue_info_lock); |
| queue_tid_bitmap = mvm->queue_info[i].tid_bitmap; |
| |
| /* If TXQ isn't in active use anyway - nothing to do here... */ |
| if (mvm->queue_info[i].status != IWL_MVM_QUEUE_READY && |
| mvm->queue_info[i].status != IWL_MVM_QUEUE_SHARED) { |
| spin_unlock_bh(&mvm->queue_info_lock); |
| continue; |
| } |
| |
| /* Check to see if there are inactive TIDs on this queue */ |
| for_each_set_bit(tid, &queue_tid_bitmap, |
| IWL_MAX_TID_COUNT + 1) { |
| if (time_after(mvm->queue_info[i].last_frame_time[tid] + |
| IWL_MVM_DQA_QUEUE_TIMEOUT, now)) |
| continue; |
| |
| inactive_tid_bitmap |= BIT(tid); |
| } |
| spin_unlock_bh(&mvm->queue_info_lock); |
| |
| /* If all TIDs are active - finish check on this queue */ |
| if (!inactive_tid_bitmap) |
| continue; |
| |
| /* |
| * If we are here - the queue hadn't been served recently and is |
| * in use |
| */ |
| |
| sta_id = mvm->queue_info[i].ra_sta_id; |
| sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]); |
| |
| /* |
| * If the STA doesn't exist anymore, it isn't an error. It could |
| * be that it was removed since getting the queues, and in this |
| * case it should've inactivated its queues anyway. |
| */ |
| if (IS_ERR_OR_NULL(sta)) |
| continue; |
| |
| mvmsta = iwl_mvm_sta_from_mac80211(sta); |
| |
| spin_lock_bh(&mvmsta->lock); |
| spin_lock(&mvm->queue_info_lock); |
| iwl_mvm_remove_inactive_tids(mvm, mvmsta, i, |
| inactive_tid_bitmap); |
| spin_unlock(&mvm->queue_info_lock); |
| spin_unlock_bh(&mvmsta->lock); |
| } |
| |
| rcu_read_unlock(); |
| } |
| |
| void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, u32 *gp2, u64 *boottime) |
| { |
| bool ps_disabled; |
| |
| lockdep_assert_held(&mvm->mutex); |
| |
| /* Disable power save when reading GP2 */ |
| ps_disabled = mvm->ps_disabled; |
| if (!ps_disabled) { |
| mvm->ps_disabled = true; |
| iwl_mvm_power_update_device(mvm); |
| } |
| |
| *gp2 = iwl_read_prph(mvm->trans, DEVICE_SYSTEM_TIME_REG); |
| *boottime = ktime_get_boot_ns(); |
| |
| if (!ps_disabled) { |
| mvm->ps_disabled = ps_disabled; |
| iwl_mvm_power_update_device(mvm); |
| } |
| } |
| |
| int iwl_mvm_send_lqm_cmd(struct ieee80211_vif *vif, |
| enum iwl_lqm_cmd_operatrions operation, |
| u32 duration, u32 timeout) |
| { |
| struct iwl_mvm_vif *mvm_vif = iwl_mvm_vif_from_mac80211(vif); |
| struct iwl_link_qual_msrmnt_cmd cmd = { |
| .cmd_operation = cpu_to_le32(operation), |
| .mac_id = cpu_to_le32(mvm_vif->id), |
| .measurement_time = cpu_to_le32(duration), |
| .timeout = cpu_to_le32(timeout), |
| }; |
| u32 cmdid = |
| iwl_cmd_id(LINK_QUALITY_MEASUREMENT_CMD, MAC_CONF_GROUP, 0); |
| int ret; |
| |
| if (!fw_has_capa(&mvm_vif->mvm->fw->ucode_capa, |
| IWL_UCODE_TLV_CAPA_LQM_SUPPORT)) |
| return -EOPNOTSUPP; |
| |
| if (vif->type != NL80211_IFTYPE_STATION || vif->p2p) |
| return -EINVAL; |
| |
| switch (operation) { |
| case LQM_CMD_OPERATION_START_MEASUREMENT: |
| if (iwl_mvm_lqm_active(mvm_vif->mvm)) |
| return -EBUSY; |
| if (!vif->bss_conf.assoc) |
| return -EINVAL; |
| mvm_vif->lqm_active = true; |
| break; |
| case LQM_CMD_OPERATION_STOP_MEASUREMENT: |
| if (!iwl_mvm_lqm_active(mvm_vif->mvm)) |
| return -EINVAL; |
| break; |
| default: |
| return -EINVAL; |
| } |
| |
| ret = iwl_mvm_send_cmd_pdu(mvm_vif->mvm, cmdid, 0, sizeof(cmd), |
| &cmd); |
| |
| /* command failed - roll back lqm_active state */ |
| if (ret) { |
| mvm_vif->lqm_active = |
| operation == LQM_CMD_OPERATION_STOP_MEASUREMENT; |
| } |
| |
| return ret; |
| } |
| |
| static void iwl_mvm_lqm_active_iterator(void *_data, u8 *mac, |
| struct ieee80211_vif *vif) |
| { |
| struct iwl_mvm_vif *mvm_vif = iwl_mvm_vif_from_mac80211(vif); |
| bool *lqm_active = _data; |
| |
| *lqm_active = *lqm_active || mvm_vif->lqm_active; |
| } |
| |
| bool iwl_mvm_lqm_active(struct iwl_mvm *mvm) |
| { |
| bool ret = false; |
| |
| lockdep_assert_held(&mvm->mutex); |
| ieee80211_iterate_active_interfaces_atomic( |
| mvm->hw, IEEE80211_IFACE_ITER_NORMAL, |
| iwl_mvm_lqm_active_iterator, &ret); |
| |
| return ret; |
| } |