mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

On MIPS architecture with CPS-based SMP support, all CPU cores in the same cluster run at the same frequency since they share the same L2 cache, requiring a fixed CPU/L2 cache ratio. This allows to implement calibrate_delay_is_known(), which will return 0 (triggering calibration) only for the primary CPU of each cluster. For other CPUs, we can simply reuse the value from their cluster's primary CPU core. With the introduction of this patch, a configuration running 32 cores spread across two clusters sees a significant reduction in boot time by approximately 600 milliseconds. Reviewed-by: Jiaxun Yang <jiaxun.yang@flygoat.com> Signed-off-by: Gregory CLEMENT <gregory.clement@bootlin.com> Signed-off-by: Thomas Bogendoerfer <tsbogend@alpha.franken.de>
979 lines
25 KiB
C
979 lines
25 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Copyright (C) 2013 Imagination Technologies
|
|
* Author: Paul Burton <paul.burton@mips.com>
|
|
*/
|
|
|
|
#include <linux/cpu.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/io.h>
|
|
#include <linux/memblock.h>
|
|
#include <linux/sched/task_stack.h>
|
|
#include <linux/sched/hotplug.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/types.h>
|
|
#include <linux/irq.h>
|
|
|
|
#include <asm/bcache.h>
|
|
#include <asm/mips-cps.h>
|
|
#include <asm/mips_mt.h>
|
|
#include <asm/mipsregs.h>
|
|
#include <asm/pm-cps.h>
|
|
#include <asm/r4kcache.h>
|
|
#include <asm/regdef.h>
|
|
#include <asm/smp.h>
|
|
#include <asm/smp-cps.h>
|
|
#include <asm/time.h>
|
|
#include <asm/uasm.h>
|
|
|
|
#define BEV_VEC_SIZE 0x500
|
|
#define BEV_VEC_ALIGN 0x1000
|
|
|
|
enum label_id {
|
|
label_not_nmi = 1,
|
|
};
|
|
|
|
UASM_L_LA(_not_nmi)
|
|
|
|
static u64 core_entry_reg;
|
|
static phys_addr_t cps_vec_pa;
|
|
|
|
struct cluster_boot_config *mips_cps_cluster_bootcfg;
|
|
|
|
static void power_up_other_cluster(unsigned int cluster)
|
|
{
|
|
u32 stat, seq_state;
|
|
unsigned int timeout;
|
|
|
|
mips_cm_lock_other(cluster, CM_GCR_Cx_OTHER_CORE_CM, 0,
|
|
CM_GCR_Cx_OTHER_BLOCK_LOCAL);
|
|
stat = read_cpc_co_stat_conf();
|
|
mips_cm_unlock_other();
|
|
|
|
seq_state = stat & CPC_Cx_STAT_CONF_SEQSTATE;
|
|
seq_state >>= __ffs(CPC_Cx_STAT_CONF_SEQSTATE);
|
|
if (seq_state == CPC_Cx_STAT_CONF_SEQSTATE_U5)
|
|
return;
|
|
|
|
/* Set endianness & power up the CM */
|
|
mips_cm_lock_other(cluster, 0, 0, CM_GCR_Cx_OTHER_BLOCK_GLOBAL);
|
|
write_cpc_redir_sys_config(IS_ENABLED(CONFIG_CPU_BIG_ENDIAN));
|
|
write_cpc_redir_pwrup_ctl(1);
|
|
mips_cm_unlock_other();
|
|
|
|
/* Wait for the CM to start up */
|
|
timeout = 1000;
|
|
mips_cm_lock_other(cluster, CM_GCR_Cx_OTHER_CORE_CM, 0,
|
|
CM_GCR_Cx_OTHER_BLOCK_LOCAL);
|
|
while (1) {
|
|
stat = read_cpc_co_stat_conf();
|
|
seq_state = stat & CPC_Cx_STAT_CONF_SEQSTATE;
|
|
seq_state >>= __ffs(CPC_Cx_STAT_CONF_SEQSTATE);
|
|
if (seq_state == CPC_Cx_STAT_CONF_SEQSTATE_U5)
|
|
break;
|
|
|
|
if (timeout) {
|
|
mdelay(1);
|
|
timeout--;
|
|
} else {
|
|
pr_warn("Waiting for cluster %u CM to power up... STAT_CONF=0x%x\n",
|
|
cluster, stat);
|
|
mdelay(1000);
|
|
}
|
|
}
|
|
|
|
mips_cm_unlock_other();
|
|
}
|
|
|
|
static unsigned __init core_vpe_count(unsigned int cluster, unsigned core)
|
|
{
|
|
return min(smp_max_threads, mips_cps_numvps(cluster, core));
|
|
}
|
|
|
|
static void __init *mips_cps_build_core_entry(void *addr)
|
|
{
|
|
extern void (*nmi_handler)(void);
|
|
u32 *p = addr;
|
|
u32 val;
|
|
struct uasm_label labels[2];
|
|
struct uasm_reloc relocs[2];
|
|
struct uasm_label *l = labels;
|
|
struct uasm_reloc *r = relocs;
|
|
|
|
memset(labels, 0, sizeof(labels));
|
|
memset(relocs, 0, sizeof(relocs));
|
|
|
|
uasm_i_mfc0(&p, GPR_K0, C0_STATUS);
|
|
UASM_i_LA(&p, GPR_T9, ST0_NMI);
|
|
uasm_i_and(&p, GPR_K0, GPR_K0, GPR_T9);
|
|
|
|
uasm_il_bnez(&p, &r, GPR_K0, label_not_nmi);
|
|
uasm_i_nop(&p);
|
|
UASM_i_LA(&p, GPR_K0, (long)&nmi_handler);
|
|
|
|
uasm_l_not_nmi(&l, p);
|
|
|
|
val = CAUSEF_IV;
|
|
uasm_i_lui(&p, GPR_K0, val >> 16);
|
|
uasm_i_ori(&p, GPR_K0, GPR_K0, val & 0xffff);
|
|
uasm_i_mtc0(&p, GPR_K0, C0_CAUSE);
|
|
val = ST0_CU1 | ST0_CU0 | ST0_BEV | ST0_KX_IF_64;
|
|
uasm_i_lui(&p, GPR_K0, val >> 16);
|
|
uasm_i_ori(&p, GPR_K0, GPR_K0, val & 0xffff);
|
|
uasm_i_mtc0(&p, GPR_K0, C0_STATUS);
|
|
uasm_i_ehb(&p);
|
|
uasm_i_ori(&p, GPR_A0, 0, read_c0_config() & CONF_CM_CMASK);
|
|
UASM_i_LA(&p, GPR_A1, (long)mips_gcr_base);
|
|
#if defined(KBUILD_64BIT_SYM32) || defined(CONFIG_32BIT)
|
|
UASM_i_LA(&p, GPR_T9, CKSEG1ADDR(__pa_symbol(mips_cps_core_boot)));
|
|
#else
|
|
UASM_i_LA(&p, GPR_T9, TO_UNCAC(__pa_symbol(mips_cps_core_boot)));
|
|
#endif
|
|
uasm_i_jr(&p, GPR_T9);
|
|
uasm_i_nop(&p);
|
|
|
|
uasm_resolve_relocs(relocs, labels);
|
|
|
|
return p;
|
|
}
|
|
|
|
static bool __init check_64bit_reset(void)
|
|
{
|
|
bool cx_64bit_reset = false;
|
|
|
|
mips_cm_lock_other(0, 0, 0, CM_GCR_Cx_OTHER_BLOCK_LOCAL);
|
|
write_gcr_co_reset64_base(CM_GCR_Cx_RESET64_BASE_BEVEXCBASE);
|
|
if ((read_gcr_co_reset64_base() & CM_GCR_Cx_RESET64_BASE_BEVEXCBASE) ==
|
|
CM_GCR_Cx_RESET64_BASE_BEVEXCBASE)
|
|
cx_64bit_reset = true;
|
|
mips_cm_unlock_other();
|
|
|
|
return cx_64bit_reset;
|
|
}
|
|
|
|
static int __init allocate_cps_vecs(void)
|
|
{
|
|
/* Try to allocate in KSEG1 first */
|
|
cps_vec_pa = memblock_phys_alloc_range(BEV_VEC_SIZE, BEV_VEC_ALIGN,
|
|
0x0, CSEGX_SIZE - 1);
|
|
|
|
if (cps_vec_pa)
|
|
core_entry_reg = CKSEG1ADDR(cps_vec_pa) &
|
|
CM_GCR_Cx_RESET_BASE_BEVEXCBASE;
|
|
|
|
if (!cps_vec_pa && mips_cm_is64) {
|
|
phys_addr_t end;
|
|
|
|
if (check_64bit_reset()) {
|
|
pr_info("VP Local Reset Exception Base support 47 bits address\n");
|
|
end = MEMBLOCK_ALLOC_ANYWHERE;
|
|
} else {
|
|
end = SZ_4G - 1;
|
|
}
|
|
cps_vec_pa = memblock_phys_alloc_range(BEV_VEC_SIZE, BEV_VEC_ALIGN, 0, end);
|
|
if (cps_vec_pa) {
|
|
if (check_64bit_reset())
|
|
core_entry_reg = (cps_vec_pa & CM_GCR_Cx_RESET64_BASE_BEVEXCBASE) |
|
|
CM_GCR_Cx_RESET_BASE_MODE;
|
|
else
|
|
core_entry_reg = (cps_vec_pa & CM_GCR_Cx_RESET_BASE_BEVEXCBASE) |
|
|
CM_GCR_Cx_RESET_BASE_MODE;
|
|
}
|
|
}
|
|
|
|
if (!cps_vec_pa)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void __init setup_cps_vecs(void)
|
|
{
|
|
void *cps_vec;
|
|
|
|
cps_vec = (void *)CKSEG1ADDR_OR_64BIT(cps_vec_pa);
|
|
mips_cps_build_core_entry(cps_vec);
|
|
|
|
memcpy(cps_vec + 0x200, &excep_tlbfill, 0x80);
|
|
memcpy(cps_vec + 0x280, &excep_xtlbfill, 0x80);
|
|
memcpy(cps_vec + 0x300, &excep_cache, 0x80);
|
|
memcpy(cps_vec + 0x380, &excep_genex, 0x80);
|
|
memcpy(cps_vec + 0x400, &excep_intex, 0x80);
|
|
memcpy(cps_vec + 0x480, &excep_ejtag, 0x80);
|
|
|
|
/* Make sure no prefetched data in cache */
|
|
blast_inv_dcache_range(CKSEG0ADDR_OR_64BIT(cps_vec_pa), CKSEG0ADDR_OR_64BIT(cps_vec_pa) + BEV_VEC_SIZE);
|
|
bc_inv(CKSEG0ADDR_OR_64BIT(cps_vec_pa), BEV_VEC_SIZE);
|
|
__sync();
|
|
}
|
|
|
|
static void __init cps_smp_setup(void)
|
|
{
|
|
unsigned int nclusters, ncores, nvpes, core_vpes;
|
|
int cl, c, v;
|
|
|
|
/* Detect & record VPE topology */
|
|
nvpes = 0;
|
|
nclusters = mips_cps_numclusters();
|
|
pr_info("%s topology ", cpu_has_mips_r6 ? "VP" : "VPE");
|
|
for (cl = 0; cl < nclusters; cl++) {
|
|
if (cl > 0)
|
|
pr_cont(",");
|
|
pr_cont("{");
|
|
|
|
if (mips_cm_revision() >= CM_REV_CM3_5)
|
|
power_up_other_cluster(cl);
|
|
|
|
ncores = mips_cps_numcores(cl);
|
|
for (c = 0; c < ncores; c++) {
|
|
core_vpes = core_vpe_count(cl, c);
|
|
|
|
if (c > 0)
|
|
pr_cont(",");
|
|
pr_cont("%u", core_vpes);
|
|
|
|
/* Use the number of VPEs in cluster 0 core 0 for smp_num_siblings */
|
|
if (!cl && !c)
|
|
smp_num_siblings = core_vpes;
|
|
cpumask_set_cpu(nvpes, &__cpu_primary_thread_mask);
|
|
|
|
for (v = 0; v < min_t(int, core_vpes, NR_CPUS - nvpes); v++) {
|
|
cpu_set_cluster(&cpu_data[nvpes + v], cl);
|
|
cpu_set_core(&cpu_data[nvpes + v], c);
|
|
cpu_set_vpe_id(&cpu_data[nvpes + v], v);
|
|
}
|
|
|
|
nvpes += core_vpes;
|
|
}
|
|
|
|
pr_cont("}");
|
|
}
|
|
pr_cont(" total %u\n", nvpes);
|
|
|
|
/* Indicate present CPUs (CPU being synonymous with VPE) */
|
|
for (v = 0; v < min_t(unsigned, nvpes, NR_CPUS); v++) {
|
|
set_cpu_possible(v, true);
|
|
set_cpu_present(v, true);
|
|
__cpu_number_map[v] = v;
|
|
__cpu_logical_map[v] = v;
|
|
}
|
|
|
|
/* Set a coherent default CCA (CWB) */
|
|
change_c0_config(CONF_CM_CMASK, 0x5);
|
|
|
|
/* Initialise core 0 */
|
|
mips_cps_core_init();
|
|
|
|
/* Make core 0 coherent with everything */
|
|
write_gcr_cl_coherence(0xff);
|
|
|
|
if (allocate_cps_vecs())
|
|
pr_err("Failed to allocate CPS vectors\n");
|
|
|
|
if (core_entry_reg && mips_cm_revision() >= CM_REV_CM3)
|
|
write_gcr_bev_base(core_entry_reg);
|
|
|
|
#ifdef CONFIG_MIPS_MT_FPAFF
|
|
/* If we have an FPU, enroll ourselves in the FPU-full mask */
|
|
if (cpu_has_fpu)
|
|
cpumask_set_cpu(0, &mt_fpu_cpumask);
|
|
#endif /* CONFIG_MIPS_MT_FPAFF */
|
|
}
|
|
|
|
unsigned long calibrate_delay_is_known(void)
|
|
{
|
|
int first_cpu_cluster = 0;
|
|
|
|
/* The calibration has to be done on the primary CPU of the cluster */
|
|
if (mips_cps_first_online_in_cluster(&first_cpu_cluster))
|
|
return 0;
|
|
|
|
return cpu_data[first_cpu_cluster].udelay_val;
|
|
}
|
|
|
|
static void __init cps_prepare_cpus(unsigned int max_cpus)
|
|
{
|
|
unsigned int nclusters, ncores, core_vpes, nvpe = 0, c, cl, cca;
|
|
bool cca_unsuitable, cores_limited;
|
|
struct cluster_boot_config *cluster_bootcfg;
|
|
struct core_boot_config *core_bootcfg;
|
|
|
|
mips_mt_set_cpuoptions();
|
|
|
|
if (!core_entry_reg) {
|
|
pr_err("core_entry address unsuitable, disabling smp-cps\n");
|
|
goto err_out;
|
|
}
|
|
|
|
/* Detect whether the CCA is unsuited to multi-core SMP */
|
|
cca = read_c0_config() & CONF_CM_CMASK;
|
|
switch (cca) {
|
|
case 0x4: /* CWBE */
|
|
case 0x5: /* CWB */
|
|
/* The CCA is coherent, multi-core is fine */
|
|
cca_unsuitable = false;
|
|
break;
|
|
|
|
default:
|
|
/* CCA is not coherent, multi-core is not usable */
|
|
cca_unsuitable = true;
|
|
}
|
|
|
|
/* Warn the user if the CCA prevents multi-core */
|
|
cores_limited = false;
|
|
if (cca_unsuitable || cpu_has_dc_aliases) {
|
|
for_each_present_cpu(c) {
|
|
if (cpus_are_siblings(smp_processor_id(), c))
|
|
continue;
|
|
|
|
set_cpu_present(c, false);
|
|
cores_limited = true;
|
|
}
|
|
}
|
|
if (cores_limited)
|
|
pr_warn("Using only one core due to %s%s%s\n",
|
|
cca_unsuitable ? "unsuitable CCA" : "",
|
|
(cca_unsuitable && cpu_has_dc_aliases) ? " & " : "",
|
|
cpu_has_dc_aliases ? "dcache aliasing" : "");
|
|
|
|
setup_cps_vecs();
|
|
|
|
/* Allocate cluster boot configuration structs */
|
|
nclusters = mips_cps_numclusters();
|
|
mips_cps_cluster_bootcfg = kcalloc(nclusters,
|
|
sizeof(*mips_cps_cluster_bootcfg),
|
|
GFP_KERNEL);
|
|
if (!mips_cps_cluster_bootcfg)
|
|
goto err_out;
|
|
|
|
if (nclusters > 1)
|
|
mips_cm_update_property();
|
|
|
|
for (cl = 0; cl < nclusters; cl++) {
|
|
/* Allocate core boot configuration structs */
|
|
ncores = mips_cps_numcores(cl);
|
|
core_bootcfg = kcalloc(ncores, sizeof(*core_bootcfg),
|
|
GFP_KERNEL);
|
|
if (!core_bootcfg)
|
|
goto err_out;
|
|
mips_cps_cluster_bootcfg[cl].core_config = core_bootcfg;
|
|
|
|
mips_cps_cluster_bootcfg[cl].core_power =
|
|
kcalloc(BITS_TO_LONGS(ncores), sizeof(unsigned long),
|
|
GFP_KERNEL);
|
|
if (!mips_cps_cluster_bootcfg[cl].core_power)
|
|
goto err_out;
|
|
|
|
/* Allocate VPE boot configuration structs */
|
|
for (c = 0; c < ncores; c++) {
|
|
int v;
|
|
core_vpes = core_vpe_count(cl, c);
|
|
core_bootcfg[c].vpe_config = kcalloc(core_vpes,
|
|
sizeof(*core_bootcfg[c].vpe_config),
|
|
GFP_KERNEL);
|
|
for (v = 0; v < core_vpes; v++)
|
|
cpumask_set_cpu(nvpe++, &mips_cps_cluster_bootcfg[cl].cpumask);
|
|
if (!core_bootcfg[c].vpe_config)
|
|
goto err_out;
|
|
}
|
|
}
|
|
|
|
/* Mark this CPU as powered up & booted */
|
|
cl = cpu_cluster(¤t_cpu_data);
|
|
c = cpu_core(¤t_cpu_data);
|
|
cluster_bootcfg = &mips_cps_cluster_bootcfg[cl];
|
|
cpu_smt_set_num_threads(core_vpes, core_vpes);
|
|
core_bootcfg = &cluster_bootcfg->core_config[c];
|
|
bitmap_set(cluster_bootcfg->core_power, cpu_core(¤t_cpu_data), 1);
|
|
atomic_set(&core_bootcfg->vpe_mask, 1 << cpu_vpe_id(¤t_cpu_data));
|
|
|
|
return;
|
|
err_out:
|
|
/* Clean up allocations */
|
|
if (mips_cps_cluster_bootcfg) {
|
|
for (cl = 0; cl < nclusters; cl++) {
|
|
cluster_bootcfg = &mips_cps_cluster_bootcfg[cl];
|
|
ncores = mips_cps_numcores(cl);
|
|
for (c = 0; c < ncores; c++) {
|
|
core_bootcfg = &cluster_bootcfg->core_config[c];
|
|
kfree(core_bootcfg->vpe_config);
|
|
}
|
|
kfree(mips_cps_cluster_bootcfg[c].core_config);
|
|
}
|
|
kfree(mips_cps_cluster_bootcfg);
|
|
mips_cps_cluster_bootcfg = NULL;
|
|
}
|
|
|
|
/* Effectively disable SMP by declaring CPUs not present */
|
|
for_each_possible_cpu(c) {
|
|
if (c == 0)
|
|
continue;
|
|
set_cpu_present(c, false);
|
|
}
|
|
}
|
|
|
|
static void init_cluster_l2(void)
|
|
{
|
|
u32 l2_cfg, l2sm_cop, result;
|
|
|
|
while (!mips_cm_is_l2_hci_broken) {
|
|
l2_cfg = read_gcr_redir_l2_ram_config();
|
|
|
|
/* If HCI is not supported, use the state machine below */
|
|
if (!(l2_cfg & CM_GCR_L2_RAM_CONFIG_PRESENT))
|
|
break;
|
|
if (!(l2_cfg & CM_GCR_L2_RAM_CONFIG_HCI_SUPPORTED))
|
|
break;
|
|
|
|
/* If the HCI_DONE bit is set, we're finished */
|
|
if (l2_cfg & CM_GCR_L2_RAM_CONFIG_HCI_DONE)
|
|
return;
|
|
}
|
|
|
|
l2sm_cop = read_gcr_redir_l2sm_cop();
|
|
if (WARN(!(l2sm_cop & CM_GCR_L2SM_COP_PRESENT),
|
|
"L2 init not supported on this system yet"))
|
|
return;
|
|
|
|
/* Clear L2 tag registers */
|
|
write_gcr_redir_l2_tag_state(0);
|
|
write_gcr_redir_l2_ecc(0);
|
|
|
|
/* Ensure the L2 tag writes complete before the state machine starts */
|
|
mb();
|
|
|
|
/* Wait for the L2 state machine to be idle */
|
|
do {
|
|
l2sm_cop = read_gcr_redir_l2sm_cop();
|
|
} while (l2sm_cop & CM_GCR_L2SM_COP_RUNNING);
|
|
|
|
/* Start a store tag operation */
|
|
l2sm_cop = CM_GCR_L2SM_COP_TYPE_IDX_STORETAG;
|
|
l2sm_cop <<= __ffs(CM_GCR_L2SM_COP_TYPE);
|
|
l2sm_cop |= CM_GCR_L2SM_COP_CMD_START;
|
|
write_gcr_redir_l2sm_cop(l2sm_cop);
|
|
|
|
/* Ensure the state machine starts before we poll for completion */
|
|
mb();
|
|
|
|
/* Wait for the operation to be complete */
|
|
do {
|
|
l2sm_cop = read_gcr_redir_l2sm_cop();
|
|
result = l2sm_cop & CM_GCR_L2SM_COP_RESULT;
|
|
result >>= __ffs(CM_GCR_L2SM_COP_RESULT);
|
|
} while (!result);
|
|
|
|
WARN(result != CM_GCR_L2SM_COP_RESULT_DONE_OK,
|
|
"L2 state machine failed cache init with error %u\n", result);
|
|
}
|
|
|
|
static void boot_core(unsigned int cluster, unsigned int core,
|
|
unsigned int vpe_id)
|
|
{
|
|
struct cluster_boot_config *cluster_cfg;
|
|
u32 access, stat, seq_state;
|
|
unsigned int timeout, ncores;
|
|
|
|
cluster_cfg = &mips_cps_cluster_bootcfg[cluster];
|
|
ncores = mips_cps_numcores(cluster);
|
|
|
|
if ((cluster != cpu_cluster(¤t_cpu_data)) &&
|
|
bitmap_empty(cluster_cfg->core_power, ncores)) {
|
|
power_up_other_cluster(cluster);
|
|
|
|
mips_cm_lock_other(cluster, core, 0,
|
|
CM_GCR_Cx_OTHER_BLOCK_GLOBAL);
|
|
|
|
/* Ensure cluster GCRs are where we expect */
|
|
write_gcr_redir_base(read_gcr_base());
|
|
write_gcr_redir_cpc_base(read_gcr_cpc_base());
|
|
write_gcr_redir_gic_base(read_gcr_gic_base());
|
|
|
|
init_cluster_l2();
|
|
|
|
/* Mirror L2 configuration */
|
|
write_gcr_redir_l2_only_sync_base(read_gcr_l2_only_sync_base());
|
|
write_gcr_redir_l2_pft_control(read_gcr_l2_pft_control());
|
|
write_gcr_redir_l2_pft_control_b(read_gcr_l2_pft_control_b());
|
|
|
|
/* Mirror ECC/parity setup */
|
|
write_gcr_redir_err_control(read_gcr_err_control());
|
|
|
|
/* Set BEV base */
|
|
write_gcr_redir_bev_base(core_entry_reg);
|
|
|
|
mips_cm_unlock_other();
|
|
}
|
|
|
|
if (cluster != cpu_cluster(¤t_cpu_data)) {
|
|
mips_cm_lock_other(cluster, core, 0,
|
|
CM_GCR_Cx_OTHER_BLOCK_GLOBAL);
|
|
|
|
/* Ensure the core can access the GCRs */
|
|
access = read_gcr_redir_access();
|
|
access |= BIT(core);
|
|
write_gcr_redir_access(access);
|
|
|
|
mips_cm_unlock_other();
|
|
} else {
|
|
/* Ensure the core can access the GCRs */
|
|
access = read_gcr_access();
|
|
access |= BIT(core);
|
|
write_gcr_access(access);
|
|
}
|
|
|
|
/* Select the appropriate core */
|
|
mips_cm_lock_other(cluster, core, 0, CM_GCR_Cx_OTHER_BLOCK_LOCAL);
|
|
|
|
/* Set its reset vector */
|
|
if (mips_cm_is64)
|
|
write_gcr_co_reset64_base(core_entry_reg);
|
|
else
|
|
write_gcr_co_reset_base(core_entry_reg);
|
|
|
|
/* Ensure its coherency is disabled */
|
|
write_gcr_co_coherence(0);
|
|
|
|
/* Start it with the legacy memory map and exception base */
|
|
write_gcr_co_reset_ext_base(CM_GCR_Cx_RESET_EXT_BASE_UEB);
|
|
|
|
/* Ensure the core can access the GCRs */
|
|
if (mips_cm_revision() < CM_REV_CM3)
|
|
set_gcr_access(1 << core);
|
|
else
|
|
set_gcr_access_cm3(1 << core);
|
|
|
|
if (mips_cpc_present()) {
|
|
/* Reset the core */
|
|
mips_cpc_lock_other(core);
|
|
|
|
if (mips_cm_revision() >= CM_REV_CM3) {
|
|
/* Run only the requested VP following the reset */
|
|
write_cpc_co_vp_stop(0xf);
|
|
write_cpc_co_vp_run(1 << vpe_id);
|
|
|
|
/*
|
|
* Ensure that the VP_RUN register is written before the
|
|
* core leaves reset.
|
|
*/
|
|
wmb();
|
|
}
|
|
|
|
write_cpc_co_cmd(CPC_Cx_CMD_RESET);
|
|
|
|
timeout = 100;
|
|
while (true) {
|
|
stat = read_cpc_co_stat_conf();
|
|
seq_state = stat & CPC_Cx_STAT_CONF_SEQSTATE;
|
|
seq_state >>= __ffs(CPC_Cx_STAT_CONF_SEQSTATE);
|
|
|
|
/* U6 == coherent execution, ie. the core is up */
|
|
if (seq_state == CPC_Cx_STAT_CONF_SEQSTATE_U6)
|
|
break;
|
|
|
|
/* Delay a little while before we start warning */
|
|
if (timeout) {
|
|
timeout--;
|
|
mdelay(10);
|
|
continue;
|
|
}
|
|
|
|
pr_warn("Waiting for core %u to start... STAT_CONF=0x%x\n",
|
|
core, stat);
|
|
mdelay(1000);
|
|
}
|
|
|
|
mips_cpc_unlock_other();
|
|
} else {
|
|
/* Take the core out of reset */
|
|
write_gcr_co_reset_release(0);
|
|
}
|
|
|
|
mips_cm_unlock_other();
|
|
|
|
/* The core is now powered up */
|
|
bitmap_set(cluster_cfg->core_power, core, 1);
|
|
|
|
/*
|
|
* Restore CM_PWRUP=0 so that the CM can power down if all the cores in
|
|
* the cluster do (eg. if they're all removed via hotplug.
|
|
*/
|
|
if (mips_cm_revision() >= CM_REV_CM3_5) {
|
|
mips_cm_lock_other(cluster, 0, 0, CM_GCR_Cx_OTHER_BLOCK_GLOBAL);
|
|
write_cpc_redir_pwrup_ctl(0);
|
|
mips_cm_unlock_other();
|
|
}
|
|
}
|
|
|
|
static void remote_vpe_boot(void *dummy)
|
|
{
|
|
unsigned int cluster = cpu_cluster(¤t_cpu_data);
|
|
unsigned core = cpu_core(¤t_cpu_data);
|
|
struct cluster_boot_config *cluster_cfg =
|
|
&mips_cps_cluster_bootcfg[cluster];
|
|
struct core_boot_config *core_cfg = &cluster_cfg->core_config[core];
|
|
|
|
mips_cps_boot_vpes(core_cfg, cpu_vpe_id(¤t_cpu_data));
|
|
}
|
|
|
|
static int cps_boot_secondary(int cpu, struct task_struct *idle)
|
|
{
|
|
unsigned int cluster = cpu_cluster(&cpu_data[cpu]);
|
|
unsigned core = cpu_core(&cpu_data[cpu]);
|
|
unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]);
|
|
struct cluster_boot_config *cluster_cfg =
|
|
&mips_cps_cluster_bootcfg[cluster];
|
|
struct core_boot_config *core_cfg = &cluster_cfg->core_config[core];
|
|
struct vpe_boot_config *vpe_cfg = &core_cfg->vpe_config[vpe_id];
|
|
unsigned int remote;
|
|
int err;
|
|
|
|
vpe_cfg->pc = (unsigned long)&smp_bootstrap;
|
|
vpe_cfg->sp = __KSTK_TOS(idle);
|
|
vpe_cfg->gp = (unsigned long)task_thread_info(idle);
|
|
|
|
atomic_or(1 << cpu_vpe_id(&cpu_data[cpu]), &core_cfg->vpe_mask);
|
|
|
|
preempt_disable();
|
|
|
|
if (!test_bit(core, cluster_cfg->core_power)) {
|
|
/* Boot a VPE on a powered down core */
|
|
boot_core(cluster, core, vpe_id);
|
|
goto out;
|
|
}
|
|
|
|
if (cpu_has_vp) {
|
|
mips_cm_lock_other(cluster, core, vpe_id,
|
|
CM_GCR_Cx_OTHER_BLOCK_LOCAL);
|
|
if (mips_cm_is64)
|
|
write_gcr_co_reset64_base(core_entry_reg);
|
|
else
|
|
write_gcr_co_reset_base(core_entry_reg);
|
|
mips_cm_unlock_other();
|
|
}
|
|
|
|
if (!cpus_are_siblings(cpu, smp_processor_id())) {
|
|
/* Boot a VPE on another powered up core */
|
|
for (remote = 0; remote < NR_CPUS; remote++) {
|
|
if (!cpus_are_siblings(cpu, remote))
|
|
continue;
|
|
if (cpu_online(remote))
|
|
break;
|
|
}
|
|
if (remote >= NR_CPUS) {
|
|
pr_crit("No online CPU in core %u to start CPU%d\n",
|
|
core, cpu);
|
|
goto out;
|
|
}
|
|
|
|
err = smp_call_function_single(remote, remote_vpe_boot,
|
|
NULL, 1);
|
|
if (err)
|
|
panic("Failed to call remote CPU\n");
|
|
goto out;
|
|
}
|
|
|
|
BUG_ON(!cpu_has_mipsmt && !cpu_has_vp);
|
|
|
|
/* Boot a VPE on this core */
|
|
mips_cps_boot_vpes(core_cfg, vpe_id);
|
|
out:
|
|
preempt_enable();
|
|
return 0;
|
|
}
|
|
|
|
static void cps_init_secondary(void)
|
|
{
|
|
int core = cpu_core(¤t_cpu_data);
|
|
|
|
/* Disable MT - we only want to run 1 TC per VPE */
|
|
if (cpu_has_mipsmt)
|
|
dmt();
|
|
|
|
if (mips_cm_revision() >= CM_REV_CM3) {
|
|
unsigned int ident = read_gic_vl_ident();
|
|
|
|
/*
|
|
* Ensure that our calculation of the VP ID matches up with
|
|
* what the GIC reports, otherwise we'll have configured
|
|
* interrupts incorrectly.
|
|
*/
|
|
BUG_ON(ident != mips_cm_vp_id(smp_processor_id()));
|
|
}
|
|
|
|
if (core > 0 && !read_gcr_cl_coherence())
|
|
pr_warn("Core %u is not in coherent domain\n", core);
|
|
|
|
if (cpu_has_veic)
|
|
clear_c0_status(ST0_IM);
|
|
else
|
|
change_c0_status(ST0_IM, STATUSF_IP2 | STATUSF_IP3 |
|
|
STATUSF_IP4 | STATUSF_IP5 |
|
|
STATUSF_IP6 | STATUSF_IP7);
|
|
}
|
|
|
|
static void cps_smp_finish(void)
|
|
{
|
|
write_c0_compare(read_c0_count() + (8 * mips_hpt_frequency / HZ));
|
|
|
|
#ifdef CONFIG_MIPS_MT_FPAFF
|
|
/* If we have an FPU, enroll ourselves in the FPU-full mask */
|
|
if (cpu_has_fpu)
|
|
cpumask_set_cpu(smp_processor_id(), &mt_fpu_cpumask);
|
|
#endif /* CONFIG_MIPS_MT_FPAFF */
|
|
|
|
local_irq_enable();
|
|
}
|
|
|
|
#if defined(CONFIG_HOTPLUG_CPU) || defined(CONFIG_KEXEC_CORE)
|
|
|
|
enum cpu_death {
|
|
CPU_DEATH_HALT,
|
|
CPU_DEATH_POWER,
|
|
};
|
|
|
|
static void cps_shutdown_this_cpu(enum cpu_death death)
|
|
{
|
|
unsigned int cpu, core, vpe_id;
|
|
|
|
cpu = smp_processor_id();
|
|
core = cpu_core(&cpu_data[cpu]);
|
|
|
|
if (death == CPU_DEATH_HALT) {
|
|
vpe_id = cpu_vpe_id(&cpu_data[cpu]);
|
|
|
|
pr_debug("Halting core %d VP%d\n", core, vpe_id);
|
|
if (cpu_has_mipsmt) {
|
|
/* Halt this TC */
|
|
write_c0_tchalt(TCHALT_H);
|
|
instruction_hazard();
|
|
} else if (cpu_has_vp) {
|
|
write_cpc_cl_vp_stop(1 << vpe_id);
|
|
|
|
/* Ensure that the VP_STOP register is written */
|
|
wmb();
|
|
}
|
|
} else {
|
|
if (IS_ENABLED(CONFIG_HOTPLUG_CPU)) {
|
|
pr_debug("Gating power to core %d\n", core);
|
|
/* Power down the core */
|
|
cps_pm_enter_state(CPS_PM_POWER_GATED);
|
|
}
|
|
}
|
|
}
|
|
|
|
#ifdef CONFIG_KEXEC_CORE
|
|
|
|
static void cps_kexec_nonboot_cpu(void)
|
|
{
|
|
if (cpu_has_mipsmt || cpu_has_vp)
|
|
cps_shutdown_this_cpu(CPU_DEATH_HALT);
|
|
else
|
|
cps_shutdown_this_cpu(CPU_DEATH_POWER);
|
|
}
|
|
|
|
#endif /* CONFIG_KEXEC_CORE */
|
|
|
|
#endif /* CONFIG_HOTPLUG_CPU || CONFIG_KEXEC_CORE */
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
|
static int cps_cpu_disable(void)
|
|
{
|
|
unsigned cpu = smp_processor_id();
|
|
struct cluster_boot_config *cluster_cfg;
|
|
struct core_boot_config *core_cfg;
|
|
|
|
if (!cps_pm_support_state(CPS_PM_POWER_GATED))
|
|
return -EINVAL;
|
|
|
|
cluster_cfg = &mips_cps_cluster_bootcfg[cpu_cluster(¤t_cpu_data)];
|
|
core_cfg = &cluster_cfg->core_config[cpu_core(¤t_cpu_data)];
|
|
atomic_sub(1 << cpu_vpe_id(¤t_cpu_data), &core_cfg->vpe_mask);
|
|
smp_mb__after_atomic();
|
|
set_cpu_online(cpu, false);
|
|
calculate_cpu_foreign_map();
|
|
irq_migrate_all_off_this_cpu();
|
|
|
|
return 0;
|
|
}
|
|
|
|
static unsigned cpu_death_sibling;
|
|
static enum cpu_death cpu_death;
|
|
|
|
void play_dead(void)
|
|
{
|
|
unsigned int cpu;
|
|
|
|
local_irq_disable();
|
|
idle_task_exit();
|
|
cpu = smp_processor_id();
|
|
cpu_death = CPU_DEATH_POWER;
|
|
|
|
pr_debug("CPU%d going offline\n", cpu);
|
|
|
|
if (cpu_has_mipsmt || cpu_has_vp) {
|
|
/* Look for another online VPE within the core */
|
|
for_each_online_cpu(cpu_death_sibling) {
|
|
if (!cpus_are_siblings(cpu, cpu_death_sibling))
|
|
continue;
|
|
|
|
/*
|
|
* There is an online VPE within the core. Just halt
|
|
* this TC and leave the core alone.
|
|
*/
|
|
cpu_death = CPU_DEATH_HALT;
|
|
break;
|
|
}
|
|
}
|
|
|
|
cpuhp_ap_report_dead();
|
|
|
|
cps_shutdown_this_cpu(cpu_death);
|
|
|
|
/* This should never be reached */
|
|
panic("Failed to offline CPU %u", cpu);
|
|
}
|
|
|
|
static void wait_for_sibling_halt(void *ptr_cpu)
|
|
{
|
|
unsigned cpu = (unsigned long)ptr_cpu;
|
|
unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]);
|
|
unsigned halted;
|
|
unsigned long flags;
|
|
|
|
do {
|
|
local_irq_save(flags);
|
|
settc(vpe_id);
|
|
halted = read_tc_c0_tchalt();
|
|
local_irq_restore(flags);
|
|
} while (!(halted & TCHALT_H));
|
|
}
|
|
|
|
static void cps_cpu_die(unsigned int cpu) { }
|
|
|
|
static void cps_cleanup_dead_cpu(unsigned cpu)
|
|
{
|
|
unsigned int cluster = cpu_cluster(&cpu_data[cpu]);
|
|
unsigned core = cpu_core(&cpu_data[cpu]);
|
|
unsigned int vpe_id = cpu_vpe_id(&cpu_data[cpu]);
|
|
ktime_t fail_time;
|
|
unsigned stat;
|
|
int err;
|
|
struct cluster_boot_config *cluster_cfg;
|
|
|
|
cluster_cfg = &mips_cps_cluster_bootcfg[cluster];
|
|
|
|
/*
|
|
* Now wait for the CPU to actually offline. Without doing this that
|
|
* offlining may race with one or more of:
|
|
*
|
|
* - Onlining the CPU again.
|
|
* - Powering down the core if another VPE within it is offlined.
|
|
* - A sibling VPE entering a non-coherent state.
|
|
*
|
|
* In the non-MT halt case (ie. infinite loop) the CPU is doing nothing
|
|
* with which we could race, so do nothing.
|
|
*/
|
|
if (cpu_death == CPU_DEATH_POWER) {
|
|
/*
|
|
* Wait for the core to enter a powered down or clock gated
|
|
* state, the latter happening when a JTAG probe is connected
|
|
* in which case the CPC will refuse to power down the core.
|
|
*/
|
|
fail_time = ktime_add_ms(ktime_get(), 2000);
|
|
do {
|
|
mips_cm_lock_other(0, core, 0, CM_GCR_Cx_OTHER_BLOCK_LOCAL);
|
|
mips_cpc_lock_other(core);
|
|
stat = read_cpc_co_stat_conf();
|
|
stat &= CPC_Cx_STAT_CONF_SEQSTATE;
|
|
stat >>= __ffs(CPC_Cx_STAT_CONF_SEQSTATE);
|
|
mips_cpc_unlock_other();
|
|
mips_cm_unlock_other();
|
|
|
|
if (stat == CPC_Cx_STAT_CONF_SEQSTATE_D0 ||
|
|
stat == CPC_Cx_STAT_CONF_SEQSTATE_D2 ||
|
|
stat == CPC_Cx_STAT_CONF_SEQSTATE_U2)
|
|
break;
|
|
|
|
/*
|
|
* The core ought to have powered down, but didn't &
|
|
* now we don't really know what state it's in. It's
|
|
* likely that its _pwr_up pin has been wired to logic
|
|
* 1 & it powered back up as soon as we powered it
|
|
* down...
|
|
*
|
|
* The best we can do is warn the user & continue in
|
|
* the hope that the core is doing nothing harmful &
|
|
* might behave properly if we online it later.
|
|
*/
|
|
if (WARN(ktime_after(ktime_get(), fail_time),
|
|
"CPU%u hasn't powered down, seq. state %u\n",
|
|
cpu, stat))
|
|
break;
|
|
} while (1);
|
|
|
|
/* Indicate the core is powered off */
|
|
bitmap_clear(cluster_cfg->core_power, core, 1);
|
|
} else if (cpu_has_mipsmt) {
|
|
/*
|
|
* Have a CPU with access to the offlined CPUs registers wait
|
|
* for its TC to halt.
|
|
*/
|
|
err = smp_call_function_single(cpu_death_sibling,
|
|
wait_for_sibling_halt,
|
|
(void *)(unsigned long)cpu, 1);
|
|
if (err)
|
|
panic("Failed to call remote sibling CPU\n");
|
|
} else if (cpu_has_vp) {
|
|
do {
|
|
mips_cm_lock_other(0, core, vpe_id, CM_GCR_Cx_OTHER_BLOCK_LOCAL);
|
|
stat = read_cpc_co_vp_running();
|
|
mips_cm_unlock_other();
|
|
} while (stat & (1 << vpe_id));
|
|
}
|
|
}
|
|
|
|
#endif /* CONFIG_HOTPLUG_CPU */
|
|
|
|
static const struct plat_smp_ops cps_smp_ops = {
|
|
.smp_setup = cps_smp_setup,
|
|
.prepare_cpus = cps_prepare_cpus,
|
|
.boot_secondary = cps_boot_secondary,
|
|
.init_secondary = cps_init_secondary,
|
|
.smp_finish = cps_smp_finish,
|
|
.send_ipi_single = mips_smp_send_ipi_single,
|
|
.send_ipi_mask = mips_smp_send_ipi_mask,
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
.cpu_disable = cps_cpu_disable,
|
|
.cpu_die = cps_cpu_die,
|
|
.cleanup_dead_cpu = cps_cleanup_dead_cpu,
|
|
#endif
|
|
#ifdef CONFIG_KEXEC_CORE
|
|
.kexec_nonboot_cpu = cps_kexec_nonboot_cpu,
|
|
#endif
|
|
};
|
|
|
|
bool mips_cps_smp_in_use(void)
|
|
{
|
|
extern const struct plat_smp_ops *mp_ops;
|
|
return mp_ops == &cps_smp_ops;
|
|
}
|
|
|
|
int register_cps_smp_ops(void)
|
|
{
|
|
if (!mips_cm_present()) {
|
|
pr_warn("MIPS CPS SMP unable to proceed without a CM\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
/* check we have a GIC - we need one for IPIs */
|
|
if (!(read_gcr_gic_status() & CM_GCR_GIC_STATUS_EX)) {
|
|
pr_warn("MIPS CPS SMP unable to proceed without a GIC\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
register_smp_ops(&cps_smp_ops);
|
|
return 0;
|
|
}
|