mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

The QCN9274 supports two memory profiles: a default profile and a low-memory profile. The driver signals the firmware to enable low-memory optimizations using the QMI initialization service. Add support to select the low-memory profile on system with less than 512 MB RAM. Tested-on: QCN9274 hw2.0 PCI WLAN.WBE.1.5-01651-QCAHKSWPL_SILICONZ-1 Tested-on: WCN7850 hw2.0 WLAN.HMT.1.1.c5-00284.1-QCAHMTSWPL_V1.0_V2.0_SILICONZ-3 Signed-off-by: Aaradhana Sahu <aaradhana.sahu@oss.qualcomm.com> Reviewed-by: Vasanthakumar Thiagarajan <vasanthakumar.thiagarajan@oss.qualcomm.com> Link: https://patch.msgid.link/20250708181102.4111054-5-aaradhana.sahu@oss.qualcomm.com Signed-off-by: Jeff Johnson <jeff.johnson@oss.qualcomm.com>
1156 lines
29 KiB
C
1156 lines
29 KiB
C
// SPDX-License-Identifier: BSD-3-Clause-Clear
|
|
/*
|
|
* Copyright (c) 2018-2019 The Linux Foundation. All rights reserved.
|
|
* Copyright (c) 2022-2025 Qualcomm Innovation Center, Inc. All rights reserved.
|
|
*/
|
|
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/firmware/qcom/qcom_scm.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_device.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/remoteproc.h>
|
|
#include <linux/soc/qcom/mdt_loader.h>
|
|
#include <linux/soc/qcom/smem_state.h>
|
|
#include "ahb.h"
|
|
#include "debug.h"
|
|
#include "hif.h"
|
|
|
|
static const struct of_device_id ath12k_ahb_of_match[] = {
|
|
{ .compatible = "qcom,ipq5332-wifi",
|
|
.data = (void *)ATH12K_HW_IPQ5332_HW10,
|
|
},
|
|
{ }
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(of, ath12k_ahb_of_match);
|
|
|
|
#define ATH12K_IRQ_CE0_OFFSET 4
|
|
#define ATH12K_MAX_UPDS 1
|
|
#define ATH12K_UPD_IRQ_WRD_LEN 18
|
|
static const char ath12k_userpd_irq[][9] = {"spawn",
|
|
"ready",
|
|
"stop-ack"};
|
|
|
|
static const char *irq_name[ATH12K_IRQ_NUM_MAX] = {
|
|
"misc-pulse1",
|
|
"misc-latch",
|
|
"sw-exception",
|
|
"watchdog",
|
|
"ce0",
|
|
"ce1",
|
|
"ce2",
|
|
"ce3",
|
|
"ce4",
|
|
"ce5",
|
|
"ce6",
|
|
"ce7",
|
|
"ce8",
|
|
"ce9",
|
|
"ce10",
|
|
"ce11",
|
|
"host2wbm-desc-feed",
|
|
"host2reo-re-injection",
|
|
"host2reo-command",
|
|
"host2rxdma-monitor-ring3",
|
|
"host2rxdma-monitor-ring2",
|
|
"host2rxdma-monitor-ring1",
|
|
"reo2ost-exception",
|
|
"wbm2host-rx-release",
|
|
"reo2host-status",
|
|
"reo2host-destination-ring4",
|
|
"reo2host-destination-ring3",
|
|
"reo2host-destination-ring2",
|
|
"reo2host-destination-ring1",
|
|
"rxdma2host-monitor-destination-mac3",
|
|
"rxdma2host-monitor-destination-mac2",
|
|
"rxdma2host-monitor-destination-mac1",
|
|
"ppdu-end-interrupts-mac3",
|
|
"ppdu-end-interrupts-mac2",
|
|
"ppdu-end-interrupts-mac1",
|
|
"rxdma2host-monitor-status-ring-mac3",
|
|
"rxdma2host-monitor-status-ring-mac2",
|
|
"rxdma2host-monitor-status-ring-mac1",
|
|
"host2rxdma-host-buf-ring-mac3",
|
|
"host2rxdma-host-buf-ring-mac2",
|
|
"host2rxdma-host-buf-ring-mac1",
|
|
"rxdma2host-destination-ring-mac3",
|
|
"rxdma2host-destination-ring-mac2",
|
|
"rxdma2host-destination-ring-mac1",
|
|
"host2tcl-input-ring4",
|
|
"host2tcl-input-ring3",
|
|
"host2tcl-input-ring2",
|
|
"host2tcl-input-ring1",
|
|
"wbm2host-tx-completions-ring4",
|
|
"wbm2host-tx-completions-ring3",
|
|
"wbm2host-tx-completions-ring2",
|
|
"wbm2host-tx-completions-ring1",
|
|
"tcl2host-status-ring",
|
|
};
|
|
|
|
enum ext_irq_num {
|
|
host2wbm_desc_feed = 16,
|
|
host2reo_re_injection,
|
|
host2reo_command,
|
|
host2rxdma_monitor_ring3,
|
|
host2rxdma_monitor_ring2,
|
|
host2rxdma_monitor_ring1,
|
|
reo2host_exception,
|
|
wbm2host_rx_release,
|
|
reo2host_status,
|
|
reo2host_destination_ring4,
|
|
reo2host_destination_ring3,
|
|
reo2host_destination_ring2,
|
|
reo2host_destination_ring1,
|
|
rxdma2host_monitor_destination_mac3,
|
|
rxdma2host_monitor_destination_mac2,
|
|
rxdma2host_monitor_destination_mac1,
|
|
ppdu_end_interrupts_mac3,
|
|
ppdu_end_interrupts_mac2,
|
|
ppdu_end_interrupts_mac1,
|
|
rxdma2host_monitor_status_ring_mac3,
|
|
rxdma2host_monitor_status_ring_mac2,
|
|
rxdma2host_monitor_status_ring_mac1,
|
|
host2rxdma_host_buf_ring_mac3,
|
|
host2rxdma_host_buf_ring_mac2,
|
|
host2rxdma_host_buf_ring_mac1,
|
|
rxdma2host_destination_ring_mac3,
|
|
rxdma2host_destination_ring_mac2,
|
|
rxdma2host_destination_ring_mac1,
|
|
host2tcl_input_ring4,
|
|
host2tcl_input_ring3,
|
|
host2tcl_input_ring2,
|
|
host2tcl_input_ring1,
|
|
wbm2host_tx_completions_ring4,
|
|
wbm2host_tx_completions_ring3,
|
|
wbm2host_tx_completions_ring2,
|
|
wbm2host_tx_completions_ring1,
|
|
tcl2host_status_ring,
|
|
};
|
|
|
|
static u32 ath12k_ahb_read32(struct ath12k_base *ab, u32 offset)
|
|
{
|
|
if (ab->ce_remap && offset < HAL_SEQ_WCSS_CMEM_OFFSET)
|
|
return ioread32(ab->mem_ce + offset);
|
|
return ioread32(ab->mem + offset);
|
|
}
|
|
|
|
static void ath12k_ahb_write32(struct ath12k_base *ab, u32 offset,
|
|
u32 value)
|
|
{
|
|
if (ab->ce_remap && offset < HAL_SEQ_WCSS_CMEM_OFFSET)
|
|
iowrite32(value, ab->mem_ce + offset);
|
|
else
|
|
iowrite32(value, ab->mem + offset);
|
|
}
|
|
|
|
static void ath12k_ahb_cancel_workqueue(struct ath12k_base *ab)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ab->hw_params->ce_count; i++) {
|
|
struct ath12k_ce_pipe *ce_pipe = &ab->ce.ce_pipe[i];
|
|
|
|
if (ath12k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR)
|
|
continue;
|
|
|
|
cancel_work_sync(&ce_pipe->intr_wq);
|
|
}
|
|
}
|
|
|
|
static void ath12k_ahb_ext_grp_disable(struct ath12k_ext_irq_grp *irq_grp)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < irq_grp->num_irq; i++)
|
|
disable_irq_nosync(irq_grp->ab->irq_num[irq_grp->irqs[i]]);
|
|
}
|
|
|
|
static void __ath12k_ahb_ext_irq_disable(struct ath12k_base *ab)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ATH12K_EXT_IRQ_GRP_NUM_MAX; i++) {
|
|
struct ath12k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i];
|
|
|
|
ath12k_ahb_ext_grp_disable(irq_grp);
|
|
if (irq_grp->napi_enabled) {
|
|
napi_synchronize(&irq_grp->napi);
|
|
napi_disable(&irq_grp->napi);
|
|
irq_grp->napi_enabled = false;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void ath12k_ahb_ext_grp_enable(struct ath12k_ext_irq_grp *irq_grp)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < irq_grp->num_irq; i++)
|
|
enable_irq(irq_grp->ab->irq_num[irq_grp->irqs[i]]);
|
|
}
|
|
|
|
static void ath12k_ahb_setbit32(struct ath12k_base *ab, u8 bit, u32 offset)
|
|
{
|
|
u32 val;
|
|
|
|
val = ath12k_ahb_read32(ab, offset);
|
|
ath12k_ahb_write32(ab, offset, val | BIT(bit));
|
|
}
|
|
|
|
static void ath12k_ahb_clearbit32(struct ath12k_base *ab, u8 bit, u32 offset)
|
|
{
|
|
u32 val;
|
|
|
|
val = ath12k_ahb_read32(ab, offset);
|
|
ath12k_ahb_write32(ab, offset, val & ~BIT(bit));
|
|
}
|
|
|
|
static void ath12k_ahb_ce_irq_enable(struct ath12k_base *ab, u16 ce_id)
|
|
{
|
|
const struct ce_attr *ce_attr;
|
|
const struct ce_ie_addr *ce_ie_addr = ab->hw_params->ce_ie_addr;
|
|
u32 ie1_reg_addr, ie2_reg_addr, ie3_reg_addr;
|
|
|
|
ie1_reg_addr = ce_ie_addr->ie1_reg_addr;
|
|
ie2_reg_addr = ce_ie_addr->ie2_reg_addr;
|
|
ie3_reg_addr = ce_ie_addr->ie3_reg_addr;
|
|
|
|
ce_attr = &ab->hw_params->host_ce_config[ce_id];
|
|
if (ce_attr->src_nentries)
|
|
ath12k_ahb_setbit32(ab, ce_id, ie1_reg_addr);
|
|
|
|
if (ce_attr->dest_nentries) {
|
|
ath12k_ahb_setbit32(ab, ce_id, ie2_reg_addr);
|
|
ath12k_ahb_setbit32(ab, ce_id + CE_HOST_IE_3_SHIFT,
|
|
ie3_reg_addr);
|
|
}
|
|
}
|
|
|
|
static void ath12k_ahb_ce_irq_disable(struct ath12k_base *ab, u16 ce_id)
|
|
{
|
|
const struct ce_attr *ce_attr;
|
|
const struct ce_ie_addr *ce_ie_addr = ab->hw_params->ce_ie_addr;
|
|
u32 ie1_reg_addr, ie2_reg_addr, ie3_reg_addr;
|
|
|
|
ie1_reg_addr = ce_ie_addr->ie1_reg_addr;
|
|
ie2_reg_addr = ce_ie_addr->ie2_reg_addr;
|
|
ie3_reg_addr = ce_ie_addr->ie3_reg_addr;
|
|
|
|
ce_attr = &ab->hw_params->host_ce_config[ce_id];
|
|
if (ce_attr->src_nentries)
|
|
ath12k_ahb_clearbit32(ab, ce_id, ie1_reg_addr);
|
|
|
|
if (ce_attr->dest_nentries) {
|
|
ath12k_ahb_clearbit32(ab, ce_id, ie2_reg_addr);
|
|
ath12k_ahb_clearbit32(ab, ce_id + CE_HOST_IE_3_SHIFT,
|
|
ie3_reg_addr);
|
|
}
|
|
}
|
|
|
|
static void ath12k_ahb_sync_ce_irqs(struct ath12k_base *ab)
|
|
{
|
|
int i;
|
|
int irq_idx;
|
|
|
|
for (i = 0; i < ab->hw_params->ce_count; i++) {
|
|
if (ath12k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR)
|
|
continue;
|
|
|
|
irq_idx = ATH12K_IRQ_CE0_OFFSET + i;
|
|
synchronize_irq(ab->irq_num[irq_idx]);
|
|
}
|
|
}
|
|
|
|
static void ath12k_ahb_sync_ext_irqs(struct ath12k_base *ab)
|
|
{
|
|
int i, j;
|
|
int irq_idx;
|
|
|
|
for (i = 0; i < ATH12K_EXT_IRQ_GRP_NUM_MAX; i++) {
|
|
struct ath12k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i];
|
|
|
|
for (j = 0; j < irq_grp->num_irq; j++) {
|
|
irq_idx = irq_grp->irqs[j];
|
|
synchronize_irq(ab->irq_num[irq_idx]);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void ath12k_ahb_ce_irqs_enable(struct ath12k_base *ab)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ab->hw_params->ce_count; i++) {
|
|
if (ath12k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR)
|
|
continue;
|
|
ath12k_ahb_ce_irq_enable(ab, i);
|
|
}
|
|
}
|
|
|
|
static void ath12k_ahb_ce_irqs_disable(struct ath12k_base *ab)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ab->hw_params->ce_count; i++) {
|
|
if (ath12k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR)
|
|
continue;
|
|
ath12k_ahb_ce_irq_disable(ab, i);
|
|
}
|
|
}
|
|
|
|
static int ath12k_ahb_start(struct ath12k_base *ab)
|
|
{
|
|
ath12k_ahb_ce_irqs_enable(ab);
|
|
ath12k_ce_rx_post_buf(ab);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath12k_ahb_ext_irq_enable(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ext_irq_grp *irq_grp;
|
|
int i;
|
|
|
|
for (i = 0; i < ATH12K_EXT_IRQ_GRP_NUM_MAX; i++) {
|
|
irq_grp = &ab->ext_irq_grp[i];
|
|
if (!irq_grp->napi_enabled) {
|
|
napi_enable(&irq_grp->napi);
|
|
irq_grp->napi_enabled = true;
|
|
}
|
|
ath12k_ahb_ext_grp_enable(irq_grp);
|
|
}
|
|
}
|
|
|
|
static void ath12k_ahb_ext_irq_disable(struct ath12k_base *ab)
|
|
{
|
|
__ath12k_ahb_ext_irq_disable(ab);
|
|
ath12k_ahb_sync_ext_irqs(ab);
|
|
}
|
|
|
|
static void ath12k_ahb_stop(struct ath12k_base *ab)
|
|
{
|
|
if (!test_bit(ATH12K_FLAG_CRASH_FLUSH, &ab->dev_flags))
|
|
ath12k_ahb_ce_irqs_disable(ab);
|
|
ath12k_ahb_sync_ce_irqs(ab);
|
|
ath12k_ahb_cancel_workqueue(ab);
|
|
timer_delete_sync(&ab->rx_replenish_retry);
|
|
ath12k_ce_cleanup_pipes(ab);
|
|
}
|
|
|
|
static int ath12k_ahb_power_up(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
char fw_name[ATH12K_USERPD_FW_NAME_LEN];
|
|
char fw2_name[ATH12K_USERPD_FW_NAME_LEN];
|
|
struct device *dev = ab->dev;
|
|
const struct firmware *fw, *fw2;
|
|
struct reserved_mem *rmem = NULL;
|
|
unsigned long time_left;
|
|
phys_addr_t mem_phys;
|
|
void *mem_region;
|
|
size_t mem_size;
|
|
u32 pasid;
|
|
int ret;
|
|
|
|
rmem = ath12k_core_get_reserved_mem(ab, 0);
|
|
if (!rmem)
|
|
return -ENODEV;
|
|
|
|
mem_phys = rmem->base;
|
|
mem_size = rmem->size;
|
|
mem_region = devm_memremap(dev, mem_phys, mem_size, MEMREMAP_WC);
|
|
if (IS_ERR(mem_region)) {
|
|
ath12k_err(ab, "unable to map memory region: %pa+%pa\n",
|
|
&rmem->base, &rmem->size);
|
|
return PTR_ERR(mem_region);
|
|
}
|
|
|
|
snprintf(fw_name, sizeof(fw_name), "%s/%s/%s%d%s", ATH12K_FW_DIR,
|
|
ab->hw_params->fw.dir, ATH12K_AHB_FW_PREFIX, ab_ahb->userpd_id,
|
|
ATH12K_AHB_FW_SUFFIX);
|
|
|
|
ret = request_firmware(&fw, fw_name, dev);
|
|
if (ret < 0) {
|
|
ath12k_err(ab, "request_firmware failed\n");
|
|
return ret;
|
|
}
|
|
|
|
ath12k_dbg(ab, ATH12K_DBG_AHB, "Booting fw image %s, size %zd\n", fw_name,
|
|
fw->size);
|
|
|
|
if (!fw->size) {
|
|
ath12k_err(ab, "Invalid firmware size\n");
|
|
ret = -EINVAL;
|
|
goto err_fw;
|
|
}
|
|
|
|
pasid = (u32_encode_bits(ab_ahb->userpd_id, ATH12K_USERPD_ID_MASK)) |
|
|
ATH12K_AHB_UPD_SWID;
|
|
|
|
/* Load FW image to a reserved memory location */
|
|
ret = qcom_mdt_load(dev, fw, fw_name, pasid, mem_region, mem_phys, mem_size,
|
|
&mem_phys);
|
|
if (ret) {
|
|
ath12k_err(ab, "Failed to load MDT segments: %d\n", ret);
|
|
goto err_fw;
|
|
}
|
|
|
|
snprintf(fw2_name, sizeof(fw2_name), "%s/%s/%s", ATH12K_FW_DIR,
|
|
ab->hw_params->fw.dir, ATH12K_AHB_FW2);
|
|
|
|
ret = request_firmware(&fw2, fw2_name, dev);
|
|
if (ret < 0) {
|
|
ath12k_err(ab, "request_firmware failed\n");
|
|
goto err_fw;
|
|
}
|
|
|
|
ath12k_dbg(ab, ATH12K_DBG_AHB, "Booting fw image %s, size %zd\n", fw2_name,
|
|
fw2->size);
|
|
|
|
if (!fw2->size) {
|
|
ath12k_err(ab, "Invalid firmware size\n");
|
|
ret = -EINVAL;
|
|
goto err_fw2;
|
|
}
|
|
|
|
ret = qcom_mdt_load_no_init(dev, fw2, fw2_name, pasid, mem_region, mem_phys,
|
|
mem_size, &mem_phys);
|
|
if (ret) {
|
|
ath12k_err(ab, "Failed to load MDT segments: %d\n", ret);
|
|
goto err_fw2;
|
|
}
|
|
|
|
/* Authenticate FW image using peripheral ID */
|
|
ret = qcom_scm_pas_auth_and_reset(pasid);
|
|
if (ret) {
|
|
ath12k_err(ab, "failed to boot the remote processor %d\n", ret);
|
|
goto err_fw2;
|
|
}
|
|
|
|
/* Instruct Q6 to spawn userPD thread */
|
|
ret = qcom_smem_state_update_bits(ab_ahb->spawn_state, BIT(ab_ahb->spawn_bit),
|
|
BIT(ab_ahb->spawn_bit));
|
|
if (ret) {
|
|
ath12k_err(ab, "Failed to update spawn state %d\n", ret);
|
|
goto err_fw2;
|
|
}
|
|
|
|
time_left = wait_for_completion_timeout(&ab_ahb->userpd_spawned,
|
|
ATH12K_USERPD_SPAWN_TIMEOUT);
|
|
if (!time_left) {
|
|
ath12k_err(ab, "UserPD spawn wait timed out\n");
|
|
ret = -ETIMEDOUT;
|
|
goto err_fw2;
|
|
}
|
|
|
|
time_left = wait_for_completion_timeout(&ab_ahb->userpd_ready,
|
|
ATH12K_USERPD_READY_TIMEOUT);
|
|
if (!time_left) {
|
|
ath12k_err(ab, "UserPD ready wait timed out\n");
|
|
ret = -ETIMEDOUT;
|
|
goto err_fw2;
|
|
}
|
|
|
|
qcom_smem_state_update_bits(ab_ahb->spawn_state, BIT(ab_ahb->spawn_bit), 0);
|
|
|
|
ath12k_dbg(ab, ATH12K_DBG_AHB, "UserPD%d is now UP\n", ab_ahb->userpd_id);
|
|
|
|
err_fw2:
|
|
release_firmware(fw2);
|
|
err_fw:
|
|
release_firmware(fw);
|
|
return ret;
|
|
}
|
|
|
|
static void ath12k_ahb_power_down(struct ath12k_base *ab, bool is_suspend)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
unsigned long time_left;
|
|
u32 pasid;
|
|
int ret;
|
|
|
|
qcom_smem_state_update_bits(ab_ahb->stop_state, BIT(ab_ahb->stop_bit),
|
|
BIT(ab_ahb->stop_bit));
|
|
|
|
time_left = wait_for_completion_timeout(&ab_ahb->userpd_stopped,
|
|
ATH12K_USERPD_STOP_TIMEOUT);
|
|
if (!time_left) {
|
|
ath12k_err(ab, "UserPD stop wait timed out\n");
|
|
return;
|
|
}
|
|
|
|
qcom_smem_state_update_bits(ab_ahb->stop_state, BIT(ab_ahb->stop_bit), 0);
|
|
|
|
pasid = (u32_encode_bits(ab_ahb->userpd_id, ATH12K_USERPD_ID_MASK)) |
|
|
ATH12K_AHB_UPD_SWID;
|
|
/* Release the firmware */
|
|
ret = qcom_scm_pas_shutdown(pasid);
|
|
if (ret)
|
|
ath12k_err(ab, "scm pas shutdown failed for userPD%d: %d\n",
|
|
ab_ahb->userpd_id, ret);
|
|
}
|
|
|
|
static void ath12k_ahb_init_qmi_ce_config(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_qmi_ce_cfg *cfg = &ab->qmi.ce_cfg;
|
|
|
|
cfg->tgt_ce_len = ab->hw_params->target_ce_count;
|
|
cfg->tgt_ce = ab->hw_params->target_ce_config;
|
|
cfg->svc_to_ce_map_len = ab->hw_params->svc_to_ce_map_len;
|
|
cfg->svc_to_ce_map = ab->hw_params->svc_to_ce_map;
|
|
ab->qmi.service_ins_id = ab->hw_params->qmi_service_ins_id;
|
|
}
|
|
|
|
static void ath12k_ahb_ce_workqueue(struct work_struct *work)
|
|
{
|
|
struct ath12k_ce_pipe *ce_pipe = from_work(ce_pipe, work, intr_wq);
|
|
|
|
ath12k_ce_per_engine_service(ce_pipe->ab, ce_pipe->pipe_num);
|
|
|
|
ath12k_ahb_ce_irq_enable(ce_pipe->ab, ce_pipe->pipe_num);
|
|
}
|
|
|
|
static irqreturn_t ath12k_ahb_ce_interrupt_handler(int irq, void *arg)
|
|
{
|
|
struct ath12k_ce_pipe *ce_pipe = arg;
|
|
|
|
/* last interrupt received for this CE */
|
|
ce_pipe->timestamp = jiffies;
|
|
|
|
ath12k_ahb_ce_irq_disable(ce_pipe->ab, ce_pipe->pipe_num);
|
|
|
|
queue_work(system_bh_wq, &ce_pipe->intr_wq);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int ath12k_ahb_ext_grp_napi_poll(struct napi_struct *napi, int budget)
|
|
{
|
|
struct ath12k_ext_irq_grp *irq_grp = container_of(napi,
|
|
struct ath12k_ext_irq_grp,
|
|
napi);
|
|
struct ath12k_base *ab = irq_grp->ab;
|
|
int work_done;
|
|
|
|
work_done = ath12k_dp_service_srng(ab, irq_grp, budget);
|
|
if (work_done < budget) {
|
|
napi_complete_done(napi, work_done);
|
|
ath12k_ahb_ext_grp_enable(irq_grp);
|
|
}
|
|
|
|
if (work_done > budget)
|
|
work_done = budget;
|
|
|
|
return work_done;
|
|
}
|
|
|
|
static irqreturn_t ath12k_ahb_ext_interrupt_handler(int irq, void *arg)
|
|
{
|
|
struct ath12k_ext_irq_grp *irq_grp = arg;
|
|
|
|
/* last interrupt received for this group */
|
|
irq_grp->timestamp = jiffies;
|
|
|
|
ath12k_ahb_ext_grp_disable(irq_grp);
|
|
|
|
napi_schedule(&irq_grp->napi);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int ath12k_ahb_config_ext_irq(struct ath12k_base *ab)
|
|
{
|
|
const struct ath12k_hw_ring_mask *ring_mask;
|
|
struct ath12k_ext_irq_grp *irq_grp;
|
|
const struct hal_ops *hal_ops;
|
|
int i, j, irq, irq_idx, ret;
|
|
u32 num_irq;
|
|
|
|
ring_mask = ab->hw_params->ring_mask;
|
|
hal_ops = ab->hw_params->hal_ops;
|
|
for (i = 0; i < ATH12K_EXT_IRQ_GRP_NUM_MAX; i++) {
|
|
irq_grp = &ab->ext_irq_grp[i];
|
|
num_irq = 0;
|
|
|
|
irq_grp->ab = ab;
|
|
irq_grp->grp_id = i;
|
|
|
|
irq_grp->napi_ndev = alloc_netdev_dummy(0);
|
|
if (!irq_grp->napi_ndev)
|
|
return -ENOMEM;
|
|
|
|
netif_napi_add(irq_grp->napi_ndev, &irq_grp->napi,
|
|
ath12k_ahb_ext_grp_napi_poll);
|
|
|
|
for (j = 0; j < ATH12K_EXT_IRQ_NUM_MAX; j++) {
|
|
/* For TX ring, ensure that the ring mask and the
|
|
* tcl_to_wbm_rbm_map point to the same ring number.
|
|
*/
|
|
if (ring_mask->tx[i] &
|
|
BIT(hal_ops->tcl_to_wbm_rbm_map[j].wbm_ring_num)) {
|
|
irq_grp->irqs[num_irq++] =
|
|
wbm2host_tx_completions_ring1 - j;
|
|
}
|
|
|
|
if (ring_mask->rx[i] & BIT(j)) {
|
|
irq_grp->irqs[num_irq++] =
|
|
reo2host_destination_ring1 - j;
|
|
}
|
|
|
|
if (ring_mask->rx_err[i] & BIT(j))
|
|
irq_grp->irqs[num_irq++] = reo2host_exception;
|
|
|
|
if (ring_mask->rx_wbm_rel[i] & BIT(j))
|
|
irq_grp->irqs[num_irq++] = wbm2host_rx_release;
|
|
|
|
if (ring_mask->reo_status[i] & BIT(j))
|
|
irq_grp->irqs[num_irq++] = reo2host_status;
|
|
|
|
if (ring_mask->rx_mon_dest[i] & BIT(j))
|
|
irq_grp->irqs[num_irq++] =
|
|
rxdma2host_monitor_destination_mac1;
|
|
}
|
|
|
|
irq_grp->num_irq = num_irq;
|
|
|
|
for (j = 0; j < irq_grp->num_irq; j++) {
|
|
irq_idx = irq_grp->irqs[j];
|
|
|
|
irq = platform_get_irq_byname(ab->pdev,
|
|
irq_name[irq_idx]);
|
|
ab->irq_num[irq_idx] = irq;
|
|
irq_set_status_flags(irq, IRQ_NOAUTOEN | IRQ_DISABLE_UNLAZY);
|
|
ret = devm_request_irq(ab->dev, irq,
|
|
ath12k_ahb_ext_interrupt_handler,
|
|
IRQF_TRIGGER_RISING,
|
|
irq_name[irq_idx], irq_grp);
|
|
if (ret)
|
|
ath12k_warn(ab, "failed request_irq for %d\n", irq);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath12k_ahb_config_irq(struct ath12k_base *ab)
|
|
{
|
|
int irq, irq_idx, i;
|
|
int ret;
|
|
|
|
/* Configure CE irqs */
|
|
for (i = 0; i < ab->hw_params->ce_count; i++) {
|
|
struct ath12k_ce_pipe *ce_pipe = &ab->ce.ce_pipe[i];
|
|
|
|
if (ath12k_ce_get_attr_flags(ab, i) & CE_ATTR_DIS_INTR)
|
|
continue;
|
|
|
|
irq_idx = ATH12K_IRQ_CE0_OFFSET + i;
|
|
|
|
INIT_WORK(&ce_pipe->intr_wq, ath12k_ahb_ce_workqueue);
|
|
irq = platform_get_irq_byname(ab->pdev, irq_name[irq_idx]);
|
|
ret = devm_request_irq(ab->dev, irq, ath12k_ahb_ce_interrupt_handler,
|
|
IRQF_TRIGGER_RISING, irq_name[irq_idx],
|
|
ce_pipe);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ab->irq_num[irq_idx] = irq;
|
|
}
|
|
|
|
/* Configure external interrupts */
|
|
ret = ath12k_ahb_config_ext_irq(ab);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ath12k_ahb_map_service_to_pipe(struct ath12k_base *ab, u16 service_id,
|
|
u8 *ul_pipe, u8 *dl_pipe)
|
|
{
|
|
const struct service_to_pipe *entry;
|
|
bool ul_set = false, dl_set = false;
|
|
u32 pipedir;
|
|
int i;
|
|
|
|
for (i = 0; i < ab->hw_params->svc_to_ce_map_len; i++) {
|
|
entry = &ab->hw_params->svc_to_ce_map[i];
|
|
|
|
if (__le32_to_cpu(entry->service_id) != service_id)
|
|
continue;
|
|
|
|
pipedir = __le32_to_cpu(entry->pipedir);
|
|
if (pipedir == PIPEDIR_IN || pipedir == PIPEDIR_INOUT) {
|
|
WARN_ON(dl_set);
|
|
*dl_pipe = __le32_to_cpu(entry->pipenum);
|
|
dl_set = true;
|
|
}
|
|
|
|
if (pipedir == PIPEDIR_OUT || pipedir == PIPEDIR_INOUT) {
|
|
WARN_ON(ul_set);
|
|
*ul_pipe = __le32_to_cpu(entry->pipenum);
|
|
ul_set = true;
|
|
}
|
|
}
|
|
|
|
if (WARN_ON(!ul_set || !dl_set))
|
|
return -ENOENT;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct ath12k_hif_ops ath12k_ahb_hif_ops_ipq5332 = {
|
|
.start = ath12k_ahb_start,
|
|
.stop = ath12k_ahb_stop,
|
|
.read32 = ath12k_ahb_read32,
|
|
.write32 = ath12k_ahb_write32,
|
|
.irq_enable = ath12k_ahb_ext_irq_enable,
|
|
.irq_disable = ath12k_ahb_ext_irq_disable,
|
|
.map_service_to_pipe = ath12k_ahb_map_service_to_pipe,
|
|
.power_up = ath12k_ahb_power_up,
|
|
.power_down = ath12k_ahb_power_down,
|
|
};
|
|
|
|
static irqreturn_t ath12k_userpd_irq_handler(int irq, void *data)
|
|
{
|
|
struct ath12k_base *ab = data;
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
|
|
if (irq == ab_ahb->userpd_irq_num[ATH12K_USERPD_SPAWN_IRQ]) {
|
|
complete(&ab_ahb->userpd_spawned);
|
|
} else if (irq == ab_ahb->userpd_irq_num[ATH12K_USERPD_READY_IRQ]) {
|
|
complete(&ab_ahb->userpd_ready);
|
|
} else if (irq == ab_ahb->userpd_irq_num[ATH12K_USERPD_STOP_ACK_IRQ]) {
|
|
complete(&ab_ahb->userpd_stopped);
|
|
} else {
|
|
ath12k_err(ab, "Invalid userpd interrupt\n");
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int ath12k_ahb_config_rproc_irq(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
int i, ret;
|
|
char *upd_irq_name;
|
|
|
|
for (i = 0; i < ATH12K_USERPD_MAX_IRQ; i++) {
|
|
ab_ahb->userpd_irq_num[i] = platform_get_irq_byname(ab->pdev,
|
|
ath12k_userpd_irq[i]);
|
|
if (ab_ahb->userpd_irq_num[i] < 0)
|
|
return ab_ahb->userpd_irq_num[i];
|
|
|
|
upd_irq_name = devm_kzalloc(&ab->pdev->dev, ATH12K_UPD_IRQ_WRD_LEN,
|
|
GFP_KERNEL);
|
|
if (!upd_irq_name)
|
|
return -ENOMEM;
|
|
|
|
scnprintf(upd_irq_name, ATH12K_UPD_IRQ_WRD_LEN, "UserPD%u-%s",
|
|
ab_ahb->userpd_id, ath12k_userpd_irq[i]);
|
|
ret = devm_request_threaded_irq(&ab->pdev->dev, ab_ahb->userpd_irq_num[i],
|
|
NULL, ath12k_userpd_irq_handler,
|
|
IRQF_TRIGGER_RISING | IRQF_ONESHOT,
|
|
upd_irq_name, ab);
|
|
if (ret)
|
|
return dev_err_probe(&ab->pdev->dev, ret,
|
|
"Request %s irq failed: %d\n",
|
|
ath12k_userpd_irq[i], ret);
|
|
}
|
|
|
|
ab_ahb->spawn_state = devm_qcom_smem_state_get(&ab->pdev->dev, "spawn",
|
|
&ab_ahb->spawn_bit);
|
|
if (IS_ERR(ab_ahb->spawn_state))
|
|
return dev_err_probe(&ab->pdev->dev, PTR_ERR(ab_ahb->spawn_state),
|
|
"Failed to acquire spawn state\n");
|
|
|
|
ab_ahb->stop_state = devm_qcom_smem_state_get(&ab->pdev->dev, "stop",
|
|
&ab_ahb->stop_bit);
|
|
if (IS_ERR(ab_ahb->stop_state))
|
|
return dev_err_probe(&ab->pdev->dev, PTR_ERR(ab_ahb->stop_state),
|
|
"Failed to acquire stop state\n");
|
|
|
|
init_completion(&ab_ahb->userpd_spawned);
|
|
init_completion(&ab_ahb->userpd_ready);
|
|
init_completion(&ab_ahb->userpd_stopped);
|
|
return 0;
|
|
}
|
|
|
|
static int ath12k_ahb_root_pd_state_notifier(struct notifier_block *nb,
|
|
const unsigned long event, void *data)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = container_of(nb, struct ath12k_ahb, root_pd_nb);
|
|
struct ath12k_base *ab = ab_ahb->ab;
|
|
|
|
if (event == ATH12K_RPROC_AFTER_POWERUP) {
|
|
ath12k_dbg(ab, ATH12K_DBG_AHB, "Root PD is UP\n");
|
|
complete(&ab_ahb->rootpd_ready);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath12k_ahb_register_rproc_notifier(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
|
|
ab_ahb->root_pd_nb.notifier_call = ath12k_ahb_root_pd_state_notifier;
|
|
init_completion(&ab_ahb->rootpd_ready);
|
|
|
|
ab_ahb->root_pd_notifier = qcom_register_ssr_notifier(ab_ahb->tgt_rproc->name,
|
|
&ab_ahb->root_pd_nb);
|
|
if (IS_ERR(ab_ahb->root_pd_notifier))
|
|
return PTR_ERR(ab_ahb->root_pd_notifier);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath12k_ahb_unregister_rproc_notifier(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
|
|
if (!ab_ahb->root_pd_notifier) {
|
|
ath12k_err(ab, "Rproc notifier not registered\n");
|
|
return;
|
|
}
|
|
|
|
qcom_unregister_ssr_notifier(ab_ahb->root_pd_notifier,
|
|
&ab_ahb->root_pd_nb);
|
|
ab_ahb->root_pd_notifier = NULL;
|
|
}
|
|
|
|
static int ath12k_ahb_get_rproc(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
struct device *dev = ab->dev;
|
|
struct device_node *np;
|
|
struct rproc *prproc;
|
|
|
|
np = of_parse_phandle(dev->of_node, "qcom,rproc", 0);
|
|
if (!np) {
|
|
ath12k_err(ab, "failed to get q6_rproc handle\n");
|
|
return -ENOENT;
|
|
}
|
|
|
|
prproc = rproc_get_by_phandle(np->phandle);
|
|
of_node_put(np);
|
|
if (!prproc)
|
|
return dev_err_probe(&ab->pdev->dev, -EPROBE_DEFER,
|
|
"failed to get rproc\n");
|
|
|
|
ab_ahb->tgt_rproc = prproc;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath12k_ahb_boot_root_pd(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
unsigned long time_left;
|
|
int ret;
|
|
|
|
ret = rproc_boot(ab_ahb->tgt_rproc);
|
|
if (ret < 0) {
|
|
ath12k_err(ab, "RootPD boot failed\n");
|
|
return ret;
|
|
}
|
|
|
|
time_left = wait_for_completion_timeout(&ab_ahb->rootpd_ready,
|
|
ATH12K_ROOTPD_READY_TIMEOUT);
|
|
if (!time_left) {
|
|
ath12k_err(ab, "RootPD ready wait timed out\n");
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath12k_ahb_configure_rproc(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
int ret;
|
|
|
|
ret = ath12k_ahb_get_rproc(ab);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
ret = ath12k_ahb_register_rproc_notifier(ab);
|
|
if (ret < 0) {
|
|
ret = dev_err_probe(&ab->pdev->dev, ret,
|
|
"failed to register rproc notifier\n");
|
|
goto err_put_rproc;
|
|
}
|
|
|
|
if (ab_ahb->tgt_rproc->state != RPROC_RUNNING) {
|
|
ret = ath12k_ahb_boot_root_pd(ab);
|
|
if (ret < 0) {
|
|
ath12k_err(ab, "failed to boot the remote processor Q6\n");
|
|
goto err_unreg_notifier;
|
|
}
|
|
}
|
|
|
|
return ath12k_ahb_config_rproc_irq(ab);
|
|
|
|
err_unreg_notifier:
|
|
ath12k_ahb_unregister_rproc_notifier(ab);
|
|
|
|
err_put_rproc:
|
|
rproc_put(ab_ahb->tgt_rproc);
|
|
return ret;
|
|
}
|
|
|
|
static void ath12k_ahb_deconfigure_rproc(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
|
|
ath12k_ahb_unregister_rproc_notifier(ab);
|
|
rproc_put(ab_ahb->tgt_rproc);
|
|
}
|
|
|
|
static int ath12k_ahb_resource_init(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
struct platform_device *pdev = ab->pdev;
|
|
struct resource *mem_res;
|
|
int ret;
|
|
|
|
ab->mem = devm_platform_get_and_ioremap_resource(pdev, 0, &mem_res);
|
|
if (IS_ERR(ab->mem)) {
|
|
ret = dev_err_probe(&pdev->dev, PTR_ERR(ab->mem), "ioremap error\n");
|
|
goto out;
|
|
}
|
|
|
|
ab->mem_len = resource_size(mem_res);
|
|
|
|
if (ab->hw_params->ce_remap) {
|
|
const struct ce_remap *ce_remap = ab->hw_params->ce_remap;
|
|
/* CE register space is moved out of WCSS and the space is not
|
|
* contiguous, hence remapping the CE registers to a new space
|
|
* for accessing them.
|
|
*/
|
|
ab->mem_ce = ioremap(ce_remap->base, ce_remap->size);
|
|
if (!ab->mem_ce) {
|
|
dev_err(&pdev->dev, "ce ioremap error\n");
|
|
ret = -ENOMEM;
|
|
goto err_mem_unmap;
|
|
}
|
|
ab->ce_remap = true;
|
|
ab->ce_remap_base_addr = HAL_IPQ5332_CE_WFSS_REG_BASE;
|
|
}
|
|
|
|
ab_ahb->xo_clk = devm_clk_get(ab->dev, "xo");
|
|
if (IS_ERR(ab_ahb->xo_clk)) {
|
|
ret = dev_err_probe(&pdev->dev, PTR_ERR(ab_ahb->xo_clk),
|
|
"failed to get xo clock\n");
|
|
goto err_mem_ce_unmap;
|
|
}
|
|
|
|
ret = clk_prepare_enable(ab_ahb->xo_clk);
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "failed to enable gcc_xo_clk: %d\n", ret);
|
|
goto err_clock_deinit;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_clock_deinit:
|
|
devm_clk_put(ab->dev, ab_ahb->xo_clk);
|
|
|
|
err_mem_ce_unmap:
|
|
ab_ahb->xo_clk = NULL;
|
|
if (ab->hw_params->ce_remap)
|
|
iounmap(ab->mem_ce);
|
|
|
|
err_mem_unmap:
|
|
ab->mem_ce = NULL;
|
|
devm_iounmap(ab->dev, ab->mem);
|
|
|
|
out:
|
|
ab->mem = NULL;
|
|
return ret;
|
|
}
|
|
|
|
static void ath12k_ahb_resource_deinit(struct ath12k_base *ab)
|
|
{
|
|
struct ath12k_ahb *ab_ahb = ath12k_ab_to_ahb(ab);
|
|
|
|
if (ab->mem)
|
|
devm_iounmap(ab->dev, ab->mem);
|
|
|
|
if (ab->mem_ce)
|
|
iounmap(ab->mem_ce);
|
|
|
|
ab->mem = NULL;
|
|
ab->mem_ce = NULL;
|
|
|
|
clk_disable_unprepare(ab_ahb->xo_clk);
|
|
devm_clk_put(ab->dev, ab_ahb->xo_clk);
|
|
ab_ahb->xo_clk = NULL;
|
|
}
|
|
|
|
static int ath12k_ahb_probe(struct platform_device *pdev)
|
|
{
|
|
struct ath12k_base *ab;
|
|
const struct ath12k_hif_ops *hif_ops;
|
|
struct ath12k_ahb *ab_ahb;
|
|
enum ath12k_hw_rev hw_rev;
|
|
u32 addr, userpd_id;
|
|
int ret;
|
|
|
|
ret = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "Failed to set 32-bit coherent dma\n");
|
|
return ret;
|
|
}
|
|
|
|
ab = ath12k_core_alloc(&pdev->dev, sizeof(struct ath12k_ahb),
|
|
ATH12K_BUS_AHB);
|
|
if (!ab)
|
|
return -ENOMEM;
|
|
|
|
hw_rev = (enum ath12k_hw_rev)(kernel_ulong_t)of_device_get_match_data(&pdev->dev);
|
|
switch (hw_rev) {
|
|
case ATH12K_HW_IPQ5332_HW10:
|
|
hif_ops = &ath12k_ahb_hif_ops_ipq5332;
|
|
userpd_id = ATH12K_IPQ5332_USERPD_ID;
|
|
break;
|
|
default:
|
|
ret = -EOPNOTSUPP;
|
|
goto err_core_free;
|
|
}
|
|
|
|
ab->hif.ops = hif_ops;
|
|
ab->pdev = pdev;
|
|
ab->hw_rev = hw_rev;
|
|
ab->target_mem_mode = ATH12K_QMI_MEMORY_MODE_DEFAULT;
|
|
platform_set_drvdata(pdev, ab);
|
|
ab_ahb = ath12k_ab_to_ahb(ab);
|
|
ab_ahb->ab = ab;
|
|
ab_ahb->userpd_id = userpd_id;
|
|
|
|
/* Set fixed_mem_region to true for platforms that support fixed memory
|
|
* reservation from DT. If memory is reserved from DT for FW, ath12k driver
|
|
* need not to allocate memory.
|
|
*/
|
|
if (!of_property_read_u32(ab->dev->of_node, "memory-region", &addr))
|
|
set_bit(ATH12K_FLAG_FIXED_MEM_REGION, &ab->dev_flags);
|
|
|
|
ret = ath12k_core_pre_init(ab);
|
|
if (ret)
|
|
goto err_core_free;
|
|
|
|
ret = ath12k_ahb_resource_init(ab);
|
|
if (ret)
|
|
goto err_core_free;
|
|
|
|
ret = ath12k_hal_srng_init(ab);
|
|
if (ret)
|
|
goto err_resource_deinit;
|
|
|
|
ret = ath12k_ce_alloc_pipes(ab);
|
|
if (ret) {
|
|
ath12k_err(ab, "failed to allocate ce pipes: %d\n", ret);
|
|
goto err_hal_srng_deinit;
|
|
}
|
|
|
|
ath12k_ahb_init_qmi_ce_config(ab);
|
|
|
|
ret = ath12k_ahb_configure_rproc(ab);
|
|
if (ret)
|
|
goto err_ce_free;
|
|
|
|
ret = ath12k_ahb_config_irq(ab);
|
|
if (ret) {
|
|
ath12k_err(ab, "failed to configure irq: %d\n", ret);
|
|
goto err_rproc_deconfigure;
|
|
}
|
|
|
|
ret = ath12k_core_init(ab);
|
|
if (ret) {
|
|
ath12k_err(ab, "failed to init core: %d\n", ret);
|
|
goto err_rproc_deconfigure;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_rproc_deconfigure:
|
|
ath12k_ahb_deconfigure_rproc(ab);
|
|
|
|
err_ce_free:
|
|
ath12k_ce_free_pipes(ab);
|
|
|
|
err_hal_srng_deinit:
|
|
ath12k_hal_srng_deinit(ab);
|
|
|
|
err_resource_deinit:
|
|
ath12k_ahb_resource_deinit(ab);
|
|
|
|
err_core_free:
|
|
ath12k_core_free(ab);
|
|
platform_set_drvdata(pdev, NULL);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ath12k_ahb_remove_prepare(struct ath12k_base *ab)
|
|
{
|
|
unsigned long left;
|
|
|
|
if (test_bit(ATH12K_FLAG_RECOVERY, &ab->dev_flags)) {
|
|
left = wait_for_completion_timeout(&ab->driver_recovery,
|
|
ATH12K_AHB_RECOVERY_TIMEOUT);
|
|
if (!left)
|
|
ath12k_warn(ab, "failed to receive recovery response completion\n");
|
|
}
|
|
|
|
set_bit(ATH12K_FLAG_UNREGISTERING, &ab->dev_flags);
|
|
cancel_work_sync(&ab->restart_work);
|
|
cancel_work_sync(&ab->qmi.event_work);
|
|
}
|
|
|
|
static void ath12k_ahb_free_resources(struct ath12k_base *ab)
|
|
{
|
|
struct platform_device *pdev = ab->pdev;
|
|
|
|
ath12k_hal_srng_deinit(ab);
|
|
ath12k_ce_free_pipes(ab);
|
|
ath12k_ahb_resource_deinit(ab);
|
|
ath12k_ahb_deconfigure_rproc(ab);
|
|
ath12k_core_free(ab);
|
|
platform_set_drvdata(pdev, NULL);
|
|
}
|
|
|
|
static void ath12k_ahb_remove(struct platform_device *pdev)
|
|
{
|
|
struct ath12k_base *ab = platform_get_drvdata(pdev);
|
|
|
|
if (test_bit(ATH12K_FLAG_QMI_FAIL, &ab->dev_flags)) {
|
|
ath12k_ahb_power_down(ab, false);
|
|
goto qmi_fail;
|
|
}
|
|
|
|
ath12k_ahb_remove_prepare(ab);
|
|
ath12k_core_hw_group_cleanup(ab->ag);
|
|
qmi_fail:
|
|
ath12k_core_deinit(ab);
|
|
ath12k_ahb_free_resources(ab);
|
|
}
|
|
|
|
static struct platform_driver ath12k_ahb_driver = {
|
|
.driver = {
|
|
.name = "ath12k_ahb",
|
|
.of_match_table = ath12k_ahb_of_match,
|
|
},
|
|
.probe = ath12k_ahb_probe,
|
|
.remove = ath12k_ahb_remove,
|
|
};
|
|
|
|
int ath12k_ahb_init(void)
|
|
{
|
|
return platform_driver_register(&ath12k_ahb_driver);
|
|
}
|
|
|
|
void ath12k_ahb_exit(void)
|
|
{
|
|
platform_driver_unregister(&ath12k_ahb_driver);
|
|
}
|