forked from Minki/linux
8d193ca26c
Our device needs two different firmwares: the INIT firmware and the operational (OPER) firmware. The first one is run when the driver loads and it returns calibrations results as well as the NVM. The second one implements the WiFi protocol. If the wlan interface is not brought up, the device is put to low power state: no firmware will be running. When the interface is brought up, we would run the OPER firmware only and reuse the results of the run of the INIT firmware when the driver was loaded. This is changing with this patch. We now run the INIT firmware every time mac80211 calls start(). The penalty for that is minimal since the INIT firwmare run fast. I now also avoid to power down the device between the INIT and OPER firmware on certains buses. The motivation for this change is that there are components on the device (MFUART) that are triggered by the INIT firmware and need the device to be powered up in order to keep running. Powering the device down between the INIT and OPER firmware would stop these components and prevent them from running again since they are triggered by the INIT firmware only. The new flow allows this and also allows to trigger these components again when the interface is brought up after it has been brought down. Signed-off-by: Eran Harary <eran.harary@intel.com> Signed-off-by: Emmanuel Grumbach <emmanuel.grumbach@intel.com>
865 lines
24 KiB
C
865 lines
24 KiB
C
/******************************************************************************
|
|
*
|
|
* This file is provided under a dual BSD/GPLv2 license. When using or
|
|
* redistributing this file, you may do so under either license.
|
|
*
|
|
* GPL LICENSE SUMMARY
|
|
*
|
|
* Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
|
|
* Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of version 2 of the GNU General Public License as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
|
|
* USA
|
|
*
|
|
* The full GNU General Public License is included in this distribution
|
|
* in the file called COPYING.
|
|
*
|
|
* Contact Information:
|
|
* Intel Linux Wireless <ilw@linux.intel.com>
|
|
* Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
|
|
*
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
|
|
* Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
*****************************************************************************/
|
|
#include <net/mac80211.h>
|
|
|
|
#include "iwl-trans.h"
|
|
#include "iwl-op-mode.h"
|
|
#include "iwl-fw.h"
|
|
#include "iwl-debug.h"
|
|
#include "iwl-csr.h" /* for iwl_mvm_rx_card_state_notif */
|
|
#include "iwl-io.h" /* for iwl_mvm_rx_card_state_notif */
|
|
#include "iwl-prph.h"
|
|
#include "iwl-eeprom-parse.h"
|
|
|
|
#include "mvm.h"
|
|
#include "iwl-phy-db.h"
|
|
|
|
#define MVM_UCODE_ALIVE_TIMEOUT HZ
|
|
#define MVM_UCODE_CALIB_TIMEOUT (2*HZ)
|
|
|
|
#define UCODE_VALID_OK cpu_to_le32(0x1)
|
|
|
|
struct iwl_mvm_alive_data {
|
|
bool valid;
|
|
u32 scd_base_addr;
|
|
};
|
|
|
|
static inline const struct fw_img *
|
|
iwl_get_ucode_image(struct iwl_mvm *mvm, enum iwl_ucode_type ucode_type)
|
|
{
|
|
if (ucode_type >= IWL_UCODE_TYPE_MAX)
|
|
return NULL;
|
|
|
|
return &mvm->fw->img[ucode_type];
|
|
}
|
|
|
|
static int iwl_send_tx_ant_cfg(struct iwl_mvm *mvm, u8 valid_tx_ant)
|
|
{
|
|
struct iwl_tx_ant_cfg_cmd tx_ant_cmd = {
|
|
.valid = cpu_to_le32(valid_tx_ant),
|
|
};
|
|
|
|
IWL_DEBUG_FW(mvm, "select valid tx ant: %u\n", valid_tx_ant);
|
|
return iwl_mvm_send_cmd_pdu(mvm, TX_ANT_CONFIGURATION_CMD, 0,
|
|
sizeof(tx_ant_cmd), &tx_ant_cmd);
|
|
}
|
|
|
|
static bool iwl_alive_fn(struct iwl_notif_wait_data *notif_wait,
|
|
struct iwl_rx_packet *pkt, void *data)
|
|
{
|
|
struct iwl_mvm *mvm =
|
|
container_of(notif_wait, struct iwl_mvm, notif_wait);
|
|
struct iwl_mvm_alive_data *alive_data = data;
|
|
struct mvm_alive_resp_ver1 *palive1;
|
|
struct mvm_alive_resp_ver2 *palive2;
|
|
struct mvm_alive_resp *palive;
|
|
|
|
if (iwl_rx_packet_payload_len(pkt) == sizeof(*palive1)) {
|
|
palive1 = (void *)pkt->data;
|
|
|
|
mvm->support_umac_log = false;
|
|
mvm->error_event_table =
|
|
le32_to_cpu(palive1->error_event_table_ptr);
|
|
mvm->log_event_table =
|
|
le32_to_cpu(palive1->log_event_table_ptr);
|
|
alive_data->scd_base_addr = le32_to_cpu(palive1->scd_base_ptr);
|
|
|
|
alive_data->valid = le16_to_cpu(palive1->status) ==
|
|
IWL_ALIVE_STATUS_OK;
|
|
IWL_DEBUG_FW(mvm,
|
|
"Alive VER1 ucode status 0x%04x revision 0x%01X 0x%01X flags 0x%01X\n",
|
|
le16_to_cpu(palive1->status), palive1->ver_type,
|
|
palive1->ver_subtype, palive1->flags);
|
|
} else if (iwl_rx_packet_payload_len(pkt) == sizeof(*palive2)) {
|
|
palive2 = (void *)pkt->data;
|
|
|
|
mvm->error_event_table =
|
|
le32_to_cpu(palive2->error_event_table_ptr);
|
|
mvm->log_event_table =
|
|
le32_to_cpu(palive2->log_event_table_ptr);
|
|
alive_data->scd_base_addr = le32_to_cpu(palive2->scd_base_ptr);
|
|
mvm->umac_error_event_table =
|
|
le32_to_cpu(palive2->error_info_addr);
|
|
mvm->sf_space.addr = le32_to_cpu(palive2->st_fwrd_addr);
|
|
mvm->sf_space.size = le32_to_cpu(palive2->st_fwrd_size);
|
|
|
|
alive_data->valid = le16_to_cpu(palive2->status) ==
|
|
IWL_ALIVE_STATUS_OK;
|
|
if (mvm->umac_error_event_table)
|
|
mvm->support_umac_log = true;
|
|
|
|
IWL_DEBUG_FW(mvm,
|
|
"Alive VER2 ucode status 0x%04x revision 0x%01X 0x%01X flags 0x%01X\n",
|
|
le16_to_cpu(palive2->status), palive2->ver_type,
|
|
palive2->ver_subtype, palive2->flags);
|
|
|
|
IWL_DEBUG_FW(mvm,
|
|
"UMAC version: Major - 0x%x, Minor - 0x%x\n",
|
|
palive2->umac_major, palive2->umac_minor);
|
|
} else if (iwl_rx_packet_payload_len(pkt) == sizeof(*palive)) {
|
|
palive = (void *)pkt->data;
|
|
|
|
mvm->error_event_table =
|
|
le32_to_cpu(palive->error_event_table_ptr);
|
|
mvm->log_event_table =
|
|
le32_to_cpu(palive->log_event_table_ptr);
|
|
alive_data->scd_base_addr = le32_to_cpu(palive->scd_base_ptr);
|
|
mvm->umac_error_event_table =
|
|
le32_to_cpu(palive->error_info_addr);
|
|
mvm->sf_space.addr = le32_to_cpu(palive->st_fwrd_addr);
|
|
mvm->sf_space.size = le32_to_cpu(palive->st_fwrd_size);
|
|
|
|
alive_data->valid = le16_to_cpu(palive->status) ==
|
|
IWL_ALIVE_STATUS_OK;
|
|
if (mvm->umac_error_event_table)
|
|
mvm->support_umac_log = true;
|
|
|
|
IWL_DEBUG_FW(mvm,
|
|
"Alive VER3 ucode status 0x%04x revision 0x%01X 0x%01X flags 0x%01X\n",
|
|
le16_to_cpu(palive->status), palive->ver_type,
|
|
palive->ver_subtype, palive->flags);
|
|
|
|
IWL_DEBUG_FW(mvm,
|
|
"UMAC version: Major - 0x%x, Minor - 0x%x\n",
|
|
le32_to_cpu(palive->umac_major),
|
|
le32_to_cpu(palive->umac_minor));
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool iwl_wait_phy_db_entry(struct iwl_notif_wait_data *notif_wait,
|
|
struct iwl_rx_packet *pkt, void *data)
|
|
{
|
|
struct iwl_phy_db *phy_db = data;
|
|
|
|
if (pkt->hdr.cmd != CALIB_RES_NOTIF_PHY_DB) {
|
|
WARN_ON(pkt->hdr.cmd != INIT_COMPLETE_NOTIF);
|
|
return true;
|
|
}
|
|
|
|
WARN_ON(iwl_phy_db_set_section(phy_db, pkt, GFP_ATOMIC));
|
|
|
|
return false;
|
|
}
|
|
|
|
static int iwl_mvm_load_ucode_wait_alive(struct iwl_mvm *mvm,
|
|
enum iwl_ucode_type ucode_type)
|
|
{
|
|
struct iwl_notification_wait alive_wait;
|
|
struct iwl_mvm_alive_data alive_data;
|
|
const struct fw_img *fw;
|
|
int ret, i;
|
|
enum iwl_ucode_type old_type = mvm->cur_ucode;
|
|
static const u8 alive_cmd[] = { MVM_ALIVE };
|
|
struct iwl_sf_region st_fwrd_space;
|
|
|
|
if (ucode_type == IWL_UCODE_REGULAR &&
|
|
iwl_fw_dbg_conf_usniffer(mvm->fw, FW_DBG_START_FROM_ALIVE))
|
|
fw = iwl_get_ucode_image(mvm, IWL_UCODE_REGULAR_USNIFFER);
|
|
else
|
|
fw = iwl_get_ucode_image(mvm, ucode_type);
|
|
if (WARN_ON(!fw))
|
|
return -EINVAL;
|
|
mvm->cur_ucode = ucode_type;
|
|
mvm->ucode_loaded = false;
|
|
|
|
iwl_init_notification_wait(&mvm->notif_wait, &alive_wait,
|
|
alive_cmd, ARRAY_SIZE(alive_cmd),
|
|
iwl_alive_fn, &alive_data);
|
|
|
|
ret = iwl_trans_start_fw(mvm->trans, fw, ucode_type == IWL_UCODE_INIT);
|
|
if (ret) {
|
|
mvm->cur_ucode = old_type;
|
|
iwl_remove_notification(&mvm->notif_wait, &alive_wait);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Some things may run in the background now, but we
|
|
* just wait for the ALIVE notification here.
|
|
*/
|
|
ret = iwl_wait_notification(&mvm->notif_wait, &alive_wait,
|
|
MVM_UCODE_ALIVE_TIMEOUT);
|
|
if (ret) {
|
|
mvm->cur_ucode = old_type;
|
|
return ret;
|
|
}
|
|
|
|
if (!alive_data.valid) {
|
|
IWL_ERR(mvm, "Loaded ucode is not valid!\n");
|
|
mvm->cur_ucode = old_type;
|
|
return -EIO;
|
|
}
|
|
|
|
/*
|
|
* update the sdio allocation according to the pointer we get in the
|
|
* alive notification.
|
|
*/
|
|
st_fwrd_space.addr = mvm->sf_space.addr;
|
|
st_fwrd_space.size = mvm->sf_space.size;
|
|
ret = iwl_trans_update_sf(mvm->trans, &st_fwrd_space);
|
|
if (ret) {
|
|
IWL_ERR(mvm, "Failed to update SF size. ret %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
iwl_trans_fw_alive(mvm->trans, alive_data.scd_base_addr);
|
|
|
|
/*
|
|
* Note: all the queues are enabled as part of the interface
|
|
* initialization, but in firmware restart scenarios they
|
|
* could be stopped, so wake them up. In firmware restart,
|
|
* mac80211 will have the queues stopped as well until the
|
|
* reconfiguration completes. During normal startup, they
|
|
* will be empty.
|
|
*/
|
|
|
|
for (i = 0; i < IWL_MAX_HW_QUEUES; i++) {
|
|
if (i < mvm->first_agg_queue && i != IWL_MVM_CMD_QUEUE)
|
|
mvm->queue_to_mac80211[i] = i;
|
|
else
|
|
mvm->queue_to_mac80211[i] = IWL_INVALID_MAC80211_QUEUE;
|
|
}
|
|
|
|
for (i = 0; i < IEEE80211_MAX_QUEUES; i++)
|
|
atomic_set(&mvm->mac80211_queue_stop_count[i], 0);
|
|
|
|
mvm->ucode_loaded = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int iwl_send_phy_cfg_cmd(struct iwl_mvm *mvm)
|
|
{
|
|
struct iwl_phy_cfg_cmd phy_cfg_cmd;
|
|
enum iwl_ucode_type ucode_type = mvm->cur_ucode;
|
|
|
|
/* Set parameters */
|
|
phy_cfg_cmd.phy_cfg = cpu_to_le32(iwl_mvm_get_phy_config(mvm));
|
|
phy_cfg_cmd.calib_control.event_trigger =
|
|
mvm->fw->default_calib[ucode_type].event_trigger;
|
|
phy_cfg_cmd.calib_control.flow_trigger =
|
|
mvm->fw->default_calib[ucode_type].flow_trigger;
|
|
|
|
IWL_DEBUG_INFO(mvm, "Sending Phy CFG command: 0x%x\n",
|
|
phy_cfg_cmd.phy_cfg);
|
|
|
|
return iwl_mvm_send_cmd_pdu(mvm, PHY_CONFIGURATION_CMD, 0,
|
|
sizeof(phy_cfg_cmd), &phy_cfg_cmd);
|
|
}
|
|
|
|
int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm)
|
|
{
|
|
struct iwl_notification_wait calib_wait;
|
|
static const u8 init_complete[] = {
|
|
INIT_COMPLETE_NOTIF,
|
|
CALIB_RES_NOTIF_PHY_DB
|
|
};
|
|
int ret;
|
|
|
|
lockdep_assert_held(&mvm->mutex);
|
|
|
|
if (WARN_ON_ONCE(mvm->calibrating))
|
|
return 0;
|
|
|
|
iwl_init_notification_wait(&mvm->notif_wait,
|
|
&calib_wait,
|
|
init_complete,
|
|
ARRAY_SIZE(init_complete),
|
|
iwl_wait_phy_db_entry,
|
|
mvm->phy_db);
|
|
|
|
/* Will also start the device */
|
|
ret = iwl_mvm_load_ucode_wait_alive(mvm, IWL_UCODE_INIT);
|
|
if (ret) {
|
|
IWL_ERR(mvm, "Failed to start INIT ucode: %d\n", ret);
|
|
goto error;
|
|
}
|
|
|
|
ret = iwl_send_bt_init_conf(mvm);
|
|
if (ret)
|
|
goto error;
|
|
|
|
/* Read the NVM only at driver load time, no need to do this twice */
|
|
if (read_nvm) {
|
|
/* Read nvm */
|
|
ret = iwl_nvm_init(mvm, true);
|
|
if (ret) {
|
|
IWL_ERR(mvm, "Failed to read NVM: %d\n", ret);
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
/* In case we read the NVM from external file, load it to the NIC */
|
|
if (mvm->nvm_file_name)
|
|
iwl_mvm_load_nvm_to_nic(mvm);
|
|
|
|
ret = iwl_nvm_check_version(mvm->nvm_data, mvm->trans);
|
|
WARN_ON(ret);
|
|
|
|
/*
|
|
* abort after reading the nvm in case RF Kill is on, we will complete
|
|
* the init seq later when RF kill will switch to off
|
|
*/
|
|
if (iwl_mvm_is_radio_killed(mvm)) {
|
|
IWL_DEBUG_RF_KILL(mvm,
|
|
"jump over all phy activities due to RF kill\n");
|
|
iwl_remove_notification(&mvm->notif_wait, &calib_wait);
|
|
ret = 1;
|
|
goto out;
|
|
}
|
|
|
|
mvm->calibrating = true;
|
|
|
|
/* Send TX valid antennas before triggering calibrations */
|
|
ret = iwl_send_tx_ant_cfg(mvm, iwl_mvm_get_valid_tx_ant(mvm));
|
|
if (ret)
|
|
goto error;
|
|
|
|
/*
|
|
* Send phy configurations command to init uCode
|
|
* to start the 16.0 uCode init image internal calibrations.
|
|
*/
|
|
ret = iwl_send_phy_cfg_cmd(mvm);
|
|
if (ret) {
|
|
IWL_ERR(mvm, "Failed to run INIT calibrations: %d\n",
|
|
ret);
|
|
goto error;
|
|
}
|
|
|
|
/*
|
|
* Some things may run in the background now, but we
|
|
* just wait for the calibration complete notification.
|
|
*/
|
|
ret = iwl_wait_notification(&mvm->notif_wait, &calib_wait,
|
|
MVM_UCODE_CALIB_TIMEOUT);
|
|
|
|
if (ret && iwl_mvm_is_radio_killed(mvm)) {
|
|
IWL_DEBUG_RF_KILL(mvm, "RFKILL while calibrating.\n");
|
|
ret = 1;
|
|
}
|
|
goto out;
|
|
|
|
error:
|
|
iwl_remove_notification(&mvm->notif_wait, &calib_wait);
|
|
out:
|
|
mvm->calibrating = false;
|
|
if (iwlmvm_mod_params.init_dbg && !mvm->nvm_data) {
|
|
/* we want to debug INIT and we have no NVM - fake */
|
|
mvm->nvm_data = kzalloc(sizeof(struct iwl_nvm_data) +
|
|
sizeof(struct ieee80211_channel) +
|
|
sizeof(struct ieee80211_rate),
|
|
GFP_KERNEL);
|
|
if (!mvm->nvm_data)
|
|
return -ENOMEM;
|
|
mvm->nvm_data->bands[0].channels = mvm->nvm_data->channels;
|
|
mvm->nvm_data->bands[0].n_channels = 1;
|
|
mvm->nvm_data->bands[0].n_bitrates = 1;
|
|
mvm->nvm_data->bands[0].bitrates =
|
|
(void *)mvm->nvm_data->channels + 1;
|
|
mvm->nvm_data->bands[0].bitrates->hw_value = 10;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void iwl_mvm_get_shared_mem_conf(struct iwl_mvm *mvm)
|
|
{
|
|
struct iwl_host_cmd cmd = {
|
|
.id = SHARED_MEM_CFG,
|
|
.flags = CMD_WANT_SKB,
|
|
.data = { NULL, },
|
|
.len = { 0, },
|
|
};
|
|
struct iwl_rx_packet *pkt;
|
|
struct iwl_shared_mem_cfg *mem_cfg;
|
|
u32 i;
|
|
|
|
lockdep_assert_held(&mvm->mutex);
|
|
|
|
if (WARN_ON(iwl_mvm_send_cmd(mvm, &cmd)))
|
|
return;
|
|
|
|
pkt = cmd.resp_pkt;
|
|
if (pkt->hdr.flags & IWL_CMD_FAILED_MSK) {
|
|
IWL_ERR(mvm, "Bad return from SHARED_MEM_CFG (0x%08X)\n",
|
|
pkt->hdr.flags);
|
|
goto exit;
|
|
}
|
|
|
|
mem_cfg = (void *)pkt->data;
|
|
|
|
mvm->shared_mem_cfg.shared_mem_addr =
|
|
le32_to_cpu(mem_cfg->shared_mem_addr);
|
|
mvm->shared_mem_cfg.shared_mem_size =
|
|
le32_to_cpu(mem_cfg->shared_mem_size);
|
|
mvm->shared_mem_cfg.sample_buff_addr =
|
|
le32_to_cpu(mem_cfg->sample_buff_addr);
|
|
mvm->shared_mem_cfg.sample_buff_size =
|
|
le32_to_cpu(mem_cfg->sample_buff_size);
|
|
mvm->shared_mem_cfg.txfifo_addr = le32_to_cpu(mem_cfg->txfifo_addr);
|
|
for (i = 0; i < ARRAY_SIZE(mvm->shared_mem_cfg.txfifo_size); i++)
|
|
mvm->shared_mem_cfg.txfifo_size[i] =
|
|
le32_to_cpu(mem_cfg->txfifo_size[i]);
|
|
for (i = 0; i < ARRAY_SIZE(mvm->shared_mem_cfg.rxfifo_size); i++)
|
|
mvm->shared_mem_cfg.rxfifo_size[i] =
|
|
le32_to_cpu(mem_cfg->rxfifo_size[i]);
|
|
mvm->shared_mem_cfg.page_buff_addr =
|
|
le32_to_cpu(mem_cfg->page_buff_addr);
|
|
mvm->shared_mem_cfg.page_buff_size =
|
|
le32_to_cpu(mem_cfg->page_buff_size);
|
|
IWL_DEBUG_INFO(mvm, "SHARED MEM CFG: got memory offsets/sizes\n");
|
|
|
|
exit:
|
|
iwl_free_resp(&cmd);
|
|
}
|
|
|
|
int iwl_mvm_fw_dbg_collect_desc(struct iwl_mvm *mvm,
|
|
struct iwl_mvm_dump_desc *desc,
|
|
unsigned int delay)
|
|
{
|
|
if (test_and_set_bit(IWL_MVM_STATUS_DUMPING_FW_LOG, &mvm->status))
|
|
return -EBUSY;
|
|
|
|
if (WARN_ON(mvm->fw_dump_desc))
|
|
iwl_mvm_free_fw_dump_desc(mvm);
|
|
|
|
IWL_WARN(mvm, "Collecting data: trigger %d fired.\n",
|
|
le32_to_cpu(desc->trig_desc.type));
|
|
|
|
mvm->fw_dump_desc = desc;
|
|
|
|
queue_delayed_work(system_wq, &mvm->fw_dump_wk, delay);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int iwl_mvm_fw_dbg_collect(struct iwl_mvm *mvm, enum iwl_fw_dbg_trigger trig,
|
|
const char *str, size_t len, unsigned int delay)
|
|
{
|
|
struct iwl_mvm_dump_desc *desc;
|
|
|
|
desc = kzalloc(sizeof(*desc) + len, GFP_ATOMIC);
|
|
if (!desc)
|
|
return -ENOMEM;
|
|
|
|
desc->len = len;
|
|
desc->trig_desc.type = cpu_to_le32(trig);
|
|
memcpy(desc->trig_desc.data, str, len);
|
|
|
|
return iwl_mvm_fw_dbg_collect_desc(mvm, desc, delay);
|
|
}
|
|
|
|
int iwl_mvm_fw_dbg_collect_trig(struct iwl_mvm *mvm,
|
|
struct iwl_fw_dbg_trigger_tlv *trigger,
|
|
const char *fmt, ...)
|
|
{
|
|
unsigned int delay = msecs_to_jiffies(le32_to_cpu(trigger->stop_delay));
|
|
u16 occurrences = le16_to_cpu(trigger->occurrences);
|
|
int ret, len = 0;
|
|
char buf[64];
|
|
|
|
if (!occurrences)
|
|
return 0;
|
|
|
|
if (fmt) {
|
|
va_list ap;
|
|
|
|
buf[sizeof(buf) - 1] = '\0';
|
|
|
|
va_start(ap, fmt);
|
|
vsnprintf(buf, sizeof(buf), fmt, ap);
|
|
va_end(ap);
|
|
|
|
/* check for truncation */
|
|
if (WARN_ON_ONCE(buf[sizeof(buf) - 1]))
|
|
buf[sizeof(buf) - 1] = '\0';
|
|
|
|
len = strlen(buf) + 1;
|
|
}
|
|
|
|
ret = iwl_mvm_fw_dbg_collect(mvm, le32_to_cpu(trigger->id), buf,
|
|
len, delay);
|
|
if (ret)
|
|
return ret;
|
|
|
|
trigger->occurrences = cpu_to_le16(occurrences - 1);
|
|
return 0;
|
|
}
|
|
|
|
static inline void iwl_mvm_restart_early_start(struct iwl_mvm *mvm)
|
|
{
|
|
if (mvm->cfg->device_family == IWL_DEVICE_FAMILY_7000)
|
|
iwl_clear_bits_prph(mvm->trans, MON_BUFF_SAMPLE_CTL, 0x100);
|
|
else
|
|
iwl_write_prph(mvm->trans, DBGC_IN_SAMPLE, 1);
|
|
}
|
|
|
|
int iwl_mvm_start_fw_dbg_conf(struct iwl_mvm *mvm, u8 conf_id)
|
|
{
|
|
u8 *ptr;
|
|
int ret;
|
|
int i;
|
|
|
|
if (WARN_ONCE(conf_id >= ARRAY_SIZE(mvm->fw->dbg_conf_tlv),
|
|
"Invalid configuration %d\n", conf_id))
|
|
return -EINVAL;
|
|
|
|
/* EARLY START - firmware's configuration is hard coded */
|
|
if ((!mvm->fw->dbg_conf_tlv[conf_id] ||
|
|
!mvm->fw->dbg_conf_tlv[conf_id]->num_of_hcmds) &&
|
|
conf_id == FW_DBG_START_FROM_ALIVE) {
|
|
iwl_mvm_restart_early_start(mvm);
|
|
return 0;
|
|
}
|
|
|
|
if (!mvm->fw->dbg_conf_tlv[conf_id])
|
|
return -EINVAL;
|
|
|
|
if (mvm->fw_dbg_conf != FW_DBG_INVALID)
|
|
IWL_WARN(mvm, "FW already configured (%d) - re-configuring\n",
|
|
mvm->fw_dbg_conf);
|
|
|
|
/* Send all HCMDs for configuring the FW debug */
|
|
ptr = (void *)&mvm->fw->dbg_conf_tlv[conf_id]->hcmd;
|
|
for (i = 0; i < mvm->fw->dbg_conf_tlv[conf_id]->num_of_hcmds; i++) {
|
|
struct iwl_fw_dbg_conf_hcmd *cmd = (void *)ptr;
|
|
|
|
ret = iwl_mvm_send_cmd_pdu(mvm, cmd->id, 0,
|
|
le16_to_cpu(cmd->len), cmd->data);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ptr += sizeof(*cmd);
|
|
ptr += le16_to_cpu(cmd->len);
|
|
}
|
|
|
|
mvm->fw_dbg_conf = conf_id;
|
|
return ret;
|
|
}
|
|
|
|
static int iwl_mvm_config_ltr_v1(struct iwl_mvm *mvm)
|
|
{
|
|
struct iwl_ltr_config_cmd_v1 cmd_v1 = {
|
|
.flags = cpu_to_le32(LTR_CFG_FLAG_FEATURE_ENABLE),
|
|
};
|
|
|
|
if (!mvm->trans->ltr_enabled)
|
|
return 0;
|
|
|
|
return iwl_mvm_send_cmd_pdu(mvm, LTR_CONFIG, 0,
|
|
sizeof(cmd_v1), &cmd_v1);
|
|
}
|
|
|
|
static int iwl_mvm_config_ltr(struct iwl_mvm *mvm)
|
|
{
|
|
struct iwl_ltr_config_cmd cmd = {
|
|
.flags = cpu_to_le32(LTR_CFG_FLAG_FEATURE_ENABLE),
|
|
};
|
|
|
|
if (!mvm->trans->ltr_enabled)
|
|
return 0;
|
|
|
|
if (!(mvm->fw->ucode_capa.api[0] & IWL_UCODE_TLV_API_HDC_PHASE_0))
|
|
return iwl_mvm_config_ltr_v1(mvm);
|
|
|
|
return iwl_mvm_send_cmd_pdu(mvm, LTR_CONFIG, 0,
|
|
sizeof(cmd), &cmd);
|
|
}
|
|
|
|
int iwl_mvm_up(struct iwl_mvm *mvm)
|
|
{
|
|
int ret, i;
|
|
struct ieee80211_channel *chan;
|
|
struct cfg80211_chan_def chandef;
|
|
|
|
lockdep_assert_held(&mvm->mutex);
|
|
|
|
ret = iwl_trans_start_hw(mvm->trans);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* If we haven't completed the run of the init ucode during
|
|
* module loading, load init ucode now
|
|
* (for example, if we were in RFKILL)
|
|
*/
|
|
ret = iwl_run_init_mvm_ucode(mvm, false);
|
|
if (ret && !iwlmvm_mod_params.init_dbg) {
|
|
IWL_ERR(mvm, "Failed to run INIT ucode: %d\n", ret);
|
|
/* this can't happen */
|
|
if (WARN_ON(ret > 0))
|
|
ret = -ERFKILL;
|
|
goto error;
|
|
}
|
|
if (!iwlmvm_mod_params.init_dbg) {
|
|
/*
|
|
* Stop and start the transport without entering low power
|
|
* mode. This will save the state of other components on the
|
|
* device that are triggered by the INIT firwmare (MFUART).
|
|
*/
|
|
_iwl_trans_stop_device(mvm->trans, false);
|
|
_iwl_trans_start_hw(mvm->trans, false);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (iwlmvm_mod_params.init_dbg)
|
|
return 0;
|
|
|
|
ret = iwl_mvm_load_ucode_wait_alive(mvm, IWL_UCODE_REGULAR);
|
|
if (ret) {
|
|
IWL_ERR(mvm, "Failed to start RT ucode: %d\n", ret);
|
|
goto error;
|
|
}
|
|
|
|
if (IWL_UCODE_API(mvm->fw->ucode_ver) >= 10)
|
|
iwl_mvm_get_shared_mem_conf(mvm);
|
|
|
|
ret = iwl_mvm_sf_update(mvm, NULL, false);
|
|
if (ret)
|
|
IWL_ERR(mvm, "Failed to initialize Smart Fifo\n");
|
|
|
|
mvm->fw_dbg_conf = FW_DBG_INVALID;
|
|
/* if we have a destination, assume EARLY START */
|
|
if (mvm->fw->dbg_dest_tlv)
|
|
mvm->fw_dbg_conf = FW_DBG_START_FROM_ALIVE;
|
|
iwl_mvm_start_fw_dbg_conf(mvm, FW_DBG_START_FROM_ALIVE);
|
|
|
|
ret = iwl_send_tx_ant_cfg(mvm, iwl_mvm_get_valid_tx_ant(mvm));
|
|
if (ret)
|
|
goto error;
|
|
|
|
ret = iwl_send_bt_init_conf(mvm);
|
|
if (ret)
|
|
goto error;
|
|
|
|
/* Send phy db control command and then phy db calibration*/
|
|
ret = iwl_send_phy_db_data(mvm->phy_db);
|
|
if (ret)
|
|
goto error;
|
|
|
|
ret = iwl_send_phy_cfg_cmd(mvm);
|
|
if (ret)
|
|
goto error;
|
|
|
|
/* init the fw <-> mac80211 STA mapping */
|
|
for (i = 0; i < IWL_MVM_STATION_COUNT; i++)
|
|
RCU_INIT_POINTER(mvm->fw_id_to_mac_id[i], NULL);
|
|
|
|
mvm->tdls_cs.peer.sta_id = IWL_MVM_STATION_COUNT;
|
|
|
|
/* reset quota debouncing buffer - 0xff will yield invalid data */
|
|
memset(&mvm->last_quota_cmd, 0xff, sizeof(mvm->last_quota_cmd));
|
|
|
|
/* Add auxiliary station for scanning */
|
|
ret = iwl_mvm_add_aux_sta(mvm);
|
|
if (ret)
|
|
goto error;
|
|
|
|
/* Add all the PHY contexts */
|
|
chan = &mvm->hw->wiphy->bands[IEEE80211_BAND_2GHZ]->channels[0];
|
|
cfg80211_chandef_create(&chandef, chan, NL80211_CHAN_NO_HT);
|
|
for (i = 0; i < NUM_PHY_CTX; i++) {
|
|
/*
|
|
* The channel used here isn't relevant as it's
|
|
* going to be overwritten in the other flows.
|
|
* For now use the first channel we have.
|
|
*/
|
|
ret = iwl_mvm_phy_ctxt_add(mvm, &mvm->phy_ctxts[i],
|
|
&chandef, 1, 1);
|
|
if (ret)
|
|
goto error;
|
|
}
|
|
|
|
/* Initialize tx backoffs to the minimal possible */
|
|
iwl_mvm_tt_tx_backoff(mvm, 0);
|
|
|
|
WARN_ON(iwl_mvm_config_ltr(mvm));
|
|
|
|
ret = iwl_mvm_power_update_device(mvm);
|
|
if (ret)
|
|
goto error;
|
|
|
|
/*
|
|
* RTNL is not taken during Ct-kill, but we don't need to scan/Tx
|
|
* anyway, so don't init MCC.
|
|
*/
|
|
if (!test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status)) {
|
|
ret = iwl_mvm_init_mcc(mvm);
|
|
if (ret)
|
|
goto error;
|
|
}
|
|
|
|
if (mvm->fw->ucode_capa.capa[0] & IWL_UCODE_TLV_CAPA_UMAC_SCAN) {
|
|
ret = iwl_mvm_config_scan(mvm);
|
|
if (ret)
|
|
goto error;
|
|
}
|
|
|
|
/* allow FW/transport low power modes if not during restart */
|
|
if (!test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status))
|
|
iwl_mvm_unref(mvm, IWL_MVM_REF_UCODE_DOWN);
|
|
|
|
IWL_DEBUG_INFO(mvm, "RT uCode started.\n");
|
|
return 0;
|
|
error:
|
|
iwl_trans_stop_device(mvm->trans);
|
|
return ret;
|
|
}
|
|
|
|
int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm)
|
|
{
|
|
int ret, i;
|
|
|
|
lockdep_assert_held(&mvm->mutex);
|
|
|
|
ret = iwl_trans_start_hw(mvm->trans);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = iwl_mvm_load_ucode_wait_alive(mvm, IWL_UCODE_WOWLAN);
|
|
if (ret) {
|
|
IWL_ERR(mvm, "Failed to start WoWLAN firmware: %d\n", ret);
|
|
goto error;
|
|
}
|
|
|
|
ret = iwl_send_tx_ant_cfg(mvm, iwl_mvm_get_valid_tx_ant(mvm));
|
|
if (ret)
|
|
goto error;
|
|
|
|
/* Send phy db control command and then phy db calibration*/
|
|
ret = iwl_send_phy_db_data(mvm->phy_db);
|
|
if (ret)
|
|
goto error;
|
|
|
|
ret = iwl_send_phy_cfg_cmd(mvm);
|
|
if (ret)
|
|
goto error;
|
|
|
|
/* init the fw <-> mac80211 STA mapping */
|
|
for (i = 0; i < IWL_MVM_STATION_COUNT; i++)
|
|
RCU_INIT_POINTER(mvm->fw_id_to_mac_id[i], NULL);
|
|
|
|
/* Add auxiliary station for scanning */
|
|
ret = iwl_mvm_add_aux_sta(mvm);
|
|
if (ret)
|
|
goto error;
|
|
|
|
return 0;
|
|
error:
|
|
iwl_trans_stop_device(mvm->trans);
|
|
return ret;
|
|
}
|
|
|
|
int iwl_mvm_rx_card_state_notif(struct iwl_mvm *mvm,
|
|
struct iwl_rx_cmd_buffer *rxb,
|
|
struct iwl_device_cmd *cmd)
|
|
{
|
|
struct iwl_rx_packet *pkt = rxb_addr(rxb);
|
|
struct iwl_card_state_notif *card_state_notif = (void *)pkt->data;
|
|
u32 flags = le32_to_cpu(card_state_notif->flags);
|
|
|
|
IWL_DEBUG_RF_KILL(mvm, "Card state received: HW:%s SW:%s CT:%s\n",
|
|
(flags & HW_CARD_DISABLED) ? "Kill" : "On",
|
|
(flags & SW_CARD_DISABLED) ? "Kill" : "On",
|
|
(flags & CT_KILL_CARD_DISABLED) ?
|
|
"Reached" : "Not reached");
|
|
|
|
return 0;
|
|
}
|
|
|
|
int iwl_mvm_rx_radio_ver(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
|
|
struct iwl_device_cmd *cmd)
|
|
{
|
|
struct iwl_rx_packet *pkt = rxb_addr(rxb);
|
|
struct iwl_radio_version_notif *radio_version = (void *)pkt->data;
|
|
|
|
/* TODO: what to do with that? */
|
|
IWL_DEBUG_INFO(mvm,
|
|
"Radio version: flavor: 0x%08x, step 0x%08x, dash 0x%08x\n",
|
|
le32_to_cpu(radio_version->radio_flavor),
|
|
le32_to_cpu(radio_version->radio_step),
|
|
le32_to_cpu(radio_version->radio_dash));
|
|
return 0;
|
|
}
|
|
|
|
int iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm,
|
|
struct iwl_rx_cmd_buffer *rxb,
|
|
struct iwl_device_cmd *cmd)
|
|
{
|
|
struct iwl_rx_packet *pkt = rxb_addr(rxb);
|
|
struct iwl_mfuart_load_notif *mfuart_notif = (void *)pkt->data;
|
|
|
|
IWL_DEBUG_INFO(mvm,
|
|
"MFUART: installed ver: 0x%08x, external ver: 0x%08x, status: 0x%08x, duration: 0x%08x\n",
|
|
le32_to_cpu(mfuart_notif->installed_ver),
|
|
le32_to_cpu(mfuart_notif->external_ver),
|
|
le32_to_cpu(mfuart_notif->status),
|
|
le32_to_cpu(mfuart_notif->duration));
|
|
return 0;
|
|
}
|