mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-10-31 08:44:41 +00:00 
			
		
		
		
	 d8cc4fee3f
			
		
	
	
		d8cc4fee3f
		
	
	
	
	
		
			
			Cache attribute comes from GPA->HPA secondary mmu page table and is configured when kvm is enabled. It is the same for all VMs, so remove duplicated cache attribute setting on vCPU context switch. Signed-off-by: Bibo Mao <maobibo@loongson.cn> Signed-off-by: Huacai Chen <chenhuacai@loongson.cn>
		
			
				
	
	
		
			1770 lines
		
	
	
	
		
			44 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1770 lines
		
	
	
	
		
			44 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * Copyright (C) 2020-2023 Loongson Technology Corporation Limited
 | |
|  */
 | |
| 
 | |
| #include <linux/kvm_host.h>
 | |
| #include <linux/entry-kvm.h>
 | |
| #include <asm/fpu.h>
 | |
| #include <asm/lbt.h>
 | |
| #include <asm/loongarch.h>
 | |
| #include <asm/setup.h>
 | |
| #include <asm/time.h>
 | |
| 
 | |
| #define CREATE_TRACE_POINTS
 | |
| #include "trace.h"
 | |
| 
 | |
| const struct _kvm_stats_desc kvm_vcpu_stats_desc[] = {
 | |
| 	KVM_GENERIC_VCPU_STATS(),
 | |
| 	STATS_DESC_COUNTER(VCPU, int_exits),
 | |
| 	STATS_DESC_COUNTER(VCPU, idle_exits),
 | |
| 	STATS_DESC_COUNTER(VCPU, cpucfg_exits),
 | |
| 	STATS_DESC_COUNTER(VCPU, signal_exits),
 | |
| 	STATS_DESC_COUNTER(VCPU, hypercall_exits)
 | |
| };
 | |
| 
 | |
| const struct kvm_stats_header kvm_vcpu_stats_header = {
 | |
| 	.name_size = KVM_STATS_NAME_SIZE,
 | |
| 	.num_desc = ARRAY_SIZE(kvm_vcpu_stats_desc),
 | |
| 	.id_offset = sizeof(struct kvm_stats_header),
 | |
| 	.desc_offset = sizeof(struct kvm_stats_header) + KVM_STATS_NAME_SIZE,
 | |
| 	.data_offset = sizeof(struct kvm_stats_header) + KVM_STATS_NAME_SIZE +
 | |
| 		       sizeof(kvm_vcpu_stats_desc),
 | |
| };
 | |
| 
 | |
| static inline void kvm_save_host_pmu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	struct kvm_context *context;
 | |
| 
 | |
| 	context = this_cpu_ptr(vcpu->kvm->arch.vmcs);
 | |
| 	context->perf_cntr[0] = read_csr_perfcntr0();
 | |
| 	context->perf_cntr[1] = read_csr_perfcntr1();
 | |
| 	context->perf_cntr[2] = read_csr_perfcntr2();
 | |
| 	context->perf_cntr[3] = read_csr_perfcntr3();
 | |
| 	context->perf_ctrl[0] = write_csr_perfctrl0(0);
 | |
| 	context->perf_ctrl[1] = write_csr_perfctrl1(0);
 | |
| 	context->perf_ctrl[2] = write_csr_perfctrl2(0);
 | |
| 	context->perf_ctrl[3] = write_csr_perfctrl3(0);
 | |
| }
 | |
| 
 | |
| static inline void kvm_restore_host_pmu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	struct kvm_context *context;
 | |
| 
 | |
| 	context = this_cpu_ptr(vcpu->kvm->arch.vmcs);
 | |
| 	write_csr_perfcntr0(context->perf_cntr[0]);
 | |
| 	write_csr_perfcntr1(context->perf_cntr[1]);
 | |
| 	write_csr_perfcntr2(context->perf_cntr[2]);
 | |
| 	write_csr_perfcntr3(context->perf_cntr[3]);
 | |
| 	write_csr_perfctrl0(context->perf_ctrl[0]);
 | |
| 	write_csr_perfctrl1(context->perf_ctrl[1]);
 | |
| 	write_csr_perfctrl2(context->perf_ctrl[2]);
 | |
| 	write_csr_perfctrl3(context->perf_ctrl[3]);
 | |
| }
 | |
| 
 | |
| 
 | |
| static inline void kvm_save_guest_pmu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR0);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR1);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR2);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR3);
 | |
| 	kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL0);
 | |
| 	kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL1);
 | |
| 	kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL2);
 | |
| 	kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL3);
 | |
| }
 | |
| 
 | |
| static inline void kvm_restore_guest_pmu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR0);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR1);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR2);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR3);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL0);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL1);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL2);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL3);
 | |
| }
 | |
| 
 | |
| static int kvm_own_pmu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	unsigned long val;
 | |
| 
 | |
| 	if (!kvm_guest_has_pmu(&vcpu->arch))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	kvm_save_host_pmu(vcpu);
 | |
| 
 | |
| 	/* Set PM0-PM(num) to guest */
 | |
| 	val = read_csr_gcfg() & ~CSR_GCFG_GPERF;
 | |
| 	val |= (kvm_get_pmu_num(&vcpu->arch) + 1) << CSR_GCFG_GPERF_SHIFT;
 | |
| 	write_csr_gcfg(val);
 | |
| 
 | |
| 	kvm_restore_guest_pmu(vcpu);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void kvm_lose_pmu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	unsigned long val;
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	if (!(vcpu->arch.aux_inuse & KVM_LARCH_PMU))
 | |
| 		return;
 | |
| 
 | |
| 	kvm_save_guest_pmu(vcpu);
 | |
| 
 | |
| 	/* Disable pmu access from guest */
 | |
| 	write_csr_gcfg(read_csr_gcfg() & ~CSR_GCFG_GPERF);
 | |
| 
 | |
| 	/*
 | |
| 	 * Clear KVM_LARCH_PMU if the guest is not using PMU CSRs when
 | |
| 	 * exiting the guest, so that the next time trap into the guest.
 | |
| 	 * We don't need to deal with PMU CSRs contexts.
 | |
| 	 */
 | |
| 	val = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_PERFCTRL0);
 | |
| 	val |= kvm_read_sw_gcsr(csr, LOONGARCH_CSR_PERFCTRL1);
 | |
| 	val |= kvm_read_sw_gcsr(csr, LOONGARCH_CSR_PERFCTRL2);
 | |
| 	val |= kvm_read_sw_gcsr(csr, LOONGARCH_CSR_PERFCTRL3);
 | |
| 	if (!(val & KVM_PMU_EVENT_ENABLED))
 | |
| 		vcpu->arch.aux_inuse &= ~KVM_LARCH_PMU;
 | |
| 
 | |
| 	kvm_restore_host_pmu(vcpu);
 | |
| }
 | |
| 
 | |
| static void kvm_restore_pmu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	if ((vcpu->arch.aux_inuse & KVM_LARCH_PMU))
 | |
| 		kvm_make_request(KVM_REQ_PMU, vcpu);
 | |
| }
 | |
| 
 | |
| static void kvm_check_pmu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	if (kvm_check_request(KVM_REQ_PMU, vcpu)) {
 | |
| 		kvm_own_pmu(vcpu);
 | |
| 		vcpu->arch.aux_inuse |= KVM_LARCH_PMU;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void kvm_update_stolen_time(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	u32 version;
 | |
| 	u64 steal;
 | |
| 	gpa_t gpa;
 | |
| 	struct kvm_memslots *slots;
 | |
| 	struct kvm_steal_time __user *st;
 | |
| 	struct gfn_to_hva_cache *ghc;
 | |
| 
 | |
| 	ghc = &vcpu->arch.st.cache;
 | |
| 	gpa = vcpu->arch.st.guest_addr;
 | |
| 	if (!(gpa & KVM_STEAL_PHYS_VALID))
 | |
| 		return;
 | |
| 
 | |
| 	gpa &= KVM_STEAL_PHYS_MASK;
 | |
| 	slots = kvm_memslots(vcpu->kvm);
 | |
| 	if (slots->generation != ghc->generation || gpa != ghc->gpa) {
 | |
| 		if (kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, gpa, sizeof(*st))) {
 | |
| 			ghc->gpa = INVALID_GPA;
 | |
| 			return;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	st = (struct kvm_steal_time __user *)ghc->hva;
 | |
| 	unsafe_get_user(version, &st->version, out);
 | |
| 	if (version & 1)
 | |
| 		version += 1; /* first time write, random junk */
 | |
| 
 | |
| 	version += 1;
 | |
| 	unsafe_put_user(version, &st->version, out);
 | |
| 	smp_wmb();
 | |
| 
 | |
| 	unsafe_get_user(steal, &st->steal, out);
 | |
| 	steal += current->sched_info.run_delay - vcpu->arch.st.last_steal;
 | |
| 	vcpu->arch.st.last_steal = current->sched_info.run_delay;
 | |
| 	unsafe_put_user(steal, &st->steal, out);
 | |
| 
 | |
| 	smp_wmb();
 | |
| 	version += 1;
 | |
| 	unsafe_put_user(version, &st->version, out);
 | |
| out:
 | |
| 	mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * kvm_check_requests - check and handle pending vCPU requests
 | |
|  *
 | |
|  * Return: RESUME_GUEST if we should enter the guest
 | |
|  *         RESUME_HOST  if we should exit to userspace
 | |
|  */
 | |
| static int kvm_check_requests(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	if (!kvm_request_pending(vcpu))
 | |
| 		return RESUME_GUEST;
 | |
| 
 | |
| 	if (kvm_check_request(KVM_REQ_TLB_FLUSH, vcpu))
 | |
| 		vcpu->arch.vpid = 0;  /* Drop vpid for this vCPU */
 | |
| 
 | |
| 	if (kvm_dirty_ring_check_request(vcpu))
 | |
| 		return RESUME_HOST;
 | |
| 
 | |
| 	if (kvm_check_request(KVM_REQ_STEAL_UPDATE, vcpu))
 | |
| 		kvm_update_stolen_time(vcpu);
 | |
| 
 | |
| 	return RESUME_GUEST;
 | |
| }
 | |
| 
 | |
| static void kvm_late_check_requests(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	lockdep_assert_irqs_disabled();
 | |
| 	if (kvm_check_request(KVM_REQ_TLB_FLUSH_GPA, vcpu))
 | |
| 		if (vcpu->arch.flush_gpa != INVALID_GPA) {
 | |
| 			kvm_flush_tlb_gpa(vcpu, vcpu->arch.flush_gpa);
 | |
| 			vcpu->arch.flush_gpa = INVALID_GPA;
 | |
| 		}
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Check and handle pending signal and vCPU requests etc
 | |
|  * Run with irq enabled and preempt enabled
 | |
|  *
 | |
|  * Return: RESUME_GUEST if we should enter the guest
 | |
|  *         RESUME_HOST  if we should exit to userspace
 | |
|  *         < 0 if we should exit to userspace, where the return value
 | |
|  *         indicates an error
 | |
|  */
 | |
| static int kvm_enter_guest_check(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int idx, ret;
 | |
| 
 | |
| 	/*
 | |
| 	 * Check conditions before entering the guest
 | |
| 	 */
 | |
| 	ret = xfer_to_guest_mode_handle_work(vcpu);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	idx = srcu_read_lock(&vcpu->kvm->srcu);
 | |
| 	ret = kvm_check_requests(vcpu);
 | |
| 	srcu_read_unlock(&vcpu->kvm->srcu, idx);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Called with irq enabled
 | |
|  *
 | |
|  * Return: RESUME_GUEST if we should enter the guest, and irq disabled
 | |
|  *         Others if we should exit to userspace
 | |
|  */
 | |
| static int kvm_pre_enter_guest(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	do {
 | |
| 		ret = kvm_enter_guest_check(vcpu);
 | |
| 		if (ret != RESUME_GUEST)
 | |
| 			break;
 | |
| 
 | |
| 		/*
 | |
| 		 * Handle vcpu timer, interrupts, check requests and
 | |
| 		 * check vmid before vcpu enter guest
 | |
| 		 */
 | |
| 		local_irq_disable();
 | |
| 		kvm_deliver_intr(vcpu);
 | |
| 		kvm_deliver_exception(vcpu);
 | |
| 		/* Make sure the vcpu mode has been written */
 | |
| 		smp_store_mb(vcpu->mode, IN_GUEST_MODE);
 | |
| 		kvm_check_vpid(vcpu);
 | |
| 		kvm_check_pmu(vcpu);
 | |
| 
 | |
| 		/*
 | |
| 		 * Called after function kvm_check_vpid()
 | |
| 		 * Since it updates CSR.GSTAT used by kvm_flush_tlb_gpa(),
 | |
| 		 * and it may also clear KVM_REQ_TLB_FLUSH_GPA pending bit
 | |
| 		 */
 | |
| 		kvm_late_check_requests(vcpu);
 | |
| 		vcpu->arch.host_eentry = csr_read64(LOONGARCH_CSR_EENTRY);
 | |
| 		/* Clear KVM_LARCH_SWCSR_LATEST as CSR will change when enter guest */
 | |
| 		vcpu->arch.aux_inuse &= ~KVM_LARCH_SWCSR_LATEST;
 | |
| 
 | |
| 		if (kvm_request_pending(vcpu) || xfer_to_guest_mode_work_pending()) {
 | |
| 			/* make sure the vcpu mode has been written */
 | |
| 			smp_store_mb(vcpu->mode, OUTSIDE_GUEST_MODE);
 | |
| 			local_irq_enable();
 | |
| 			ret = -EAGAIN;
 | |
| 		}
 | |
| 	} while (ret != RESUME_GUEST);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Return 1 for resume guest and "<= 0" for resume host.
 | |
|  */
 | |
| static int kvm_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int ret = RESUME_GUEST;
 | |
| 	unsigned long estat = vcpu->arch.host_estat;
 | |
| 	u32 intr = estat & 0x1fff; /* Ignore NMI */
 | |
| 	u32 ecode = (estat & CSR_ESTAT_EXC) >> CSR_ESTAT_EXC_SHIFT;
 | |
| 
 | |
| 	vcpu->mode = OUTSIDE_GUEST_MODE;
 | |
| 
 | |
| 	/* Set a default exit reason */
 | |
| 	run->exit_reason = KVM_EXIT_UNKNOWN;
 | |
| 
 | |
| 	kvm_lose_pmu(vcpu);
 | |
| 
 | |
| 	guest_timing_exit_irqoff();
 | |
| 	guest_state_exit_irqoff();
 | |
| 	local_irq_enable();
 | |
| 
 | |
| 	trace_kvm_exit(vcpu, ecode);
 | |
| 	if (ecode) {
 | |
| 		ret = kvm_handle_fault(vcpu, ecode);
 | |
| 	} else {
 | |
| 		WARN(!intr, "vm exiting with suspicious irq\n");
 | |
| 		++vcpu->stat.int_exits;
 | |
| 	}
 | |
| 
 | |
| 	if (ret == RESUME_GUEST)
 | |
| 		ret = kvm_pre_enter_guest(vcpu);
 | |
| 
 | |
| 	if (ret != RESUME_GUEST) {
 | |
| 		local_irq_disable();
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	guest_timing_enter_irqoff();
 | |
| 	guest_state_enter_irqoff();
 | |
| 	trace_kvm_reenter(vcpu);
 | |
| 
 | |
| 	return RESUME_GUEST;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	return !!(vcpu->arch.irq_pending) &&
 | |
| 		vcpu->arch.mp_state.mp_state == KVM_MP_STATE_RUNNABLE;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	return kvm_vcpu_exiting_guest_mode(vcpu) == IN_GUEST_MODE;
 | |
| }
 | |
| 
 | |
| bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	return false;
 | |
| }
 | |
| 
 | |
| vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf)
 | |
| {
 | |
| 	return VM_FAULT_SIGBUS;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
 | |
| 				  struct kvm_translation *tr)
 | |
| {
 | |
| 	return -EINVAL;
 | |
| }
 | |
| 
 | |
| int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	/* Protect from TOD sync and vcpu_load/put() */
 | |
| 	preempt_disable();
 | |
| 	ret = kvm_pending_timer(vcpu) ||
 | |
| 		kvm_read_hw_gcsr(LOONGARCH_CSR_ESTAT) & (1 << INT_TI);
 | |
| 	preempt_enable();
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_dump_regs(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int i;
 | |
| 
 | |
| 	kvm_debug("vCPU Register Dump:\n");
 | |
| 	kvm_debug("\tPC = 0x%08lx\n", vcpu->arch.pc);
 | |
| 	kvm_debug("\tExceptions: %08lx\n", vcpu->arch.irq_pending);
 | |
| 
 | |
| 	for (i = 0; i < 32; i += 4) {
 | |
| 		kvm_debug("\tGPR%02d: %08lx %08lx %08lx %08lx\n", i,
 | |
| 		       vcpu->arch.gprs[i], vcpu->arch.gprs[i + 1],
 | |
| 		       vcpu->arch.gprs[i + 2], vcpu->arch.gprs[i + 3]);
 | |
| 	}
 | |
| 
 | |
| 	kvm_debug("\tCRMD: 0x%08lx, ESTAT: 0x%08lx\n",
 | |
| 		  kvm_read_hw_gcsr(LOONGARCH_CSR_CRMD),
 | |
| 		  kvm_read_hw_gcsr(LOONGARCH_CSR_ESTAT));
 | |
| 
 | |
| 	kvm_debug("\tERA: 0x%08lx\n", kvm_read_hw_gcsr(LOONGARCH_CSR_ERA));
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
 | |
| 				struct kvm_mp_state *mp_state)
 | |
| {
 | |
| 	*mp_state = vcpu->arch.mp_state;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
 | |
| 				struct kvm_mp_state *mp_state)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	switch (mp_state->mp_state) {
 | |
| 	case KVM_MP_STATE_RUNNABLE:
 | |
| 		vcpu->arch.mp_state = *mp_state;
 | |
| 		break;
 | |
| 	default:
 | |
| 		ret = -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
 | |
| 					struct kvm_guest_debug *dbg)
 | |
| {
 | |
| 	if (dbg->control & ~KVM_GUESTDBG_VALID_MASK)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (dbg->control & KVM_GUESTDBG_ENABLE)
 | |
| 		vcpu->guest_debug = dbg->control;
 | |
| 	else
 | |
| 		vcpu->guest_debug = 0;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static inline int kvm_set_cpuid(struct kvm_vcpu *vcpu, u64 val)
 | |
| {
 | |
| 	int cpuid;
 | |
| 	struct kvm_phyid_map *map;
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	if (val >= KVM_MAX_PHYID)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	map = vcpu->kvm->arch.phyid_map;
 | |
| 	cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_CPUID);
 | |
| 
 | |
| 	spin_lock(&vcpu->kvm->arch.phyid_map_lock);
 | |
| 	if ((cpuid < KVM_MAX_PHYID) && map->phys_map[cpuid].enabled) {
 | |
| 		/* Discard duplicated CPUID set operation */
 | |
| 		if (cpuid == val) {
 | |
| 			spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
 | |
| 			return 0;
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * CPUID is already set before
 | |
| 		 * Forbid changing to a different CPUID at runtime
 | |
| 		 */
 | |
| 		spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	if (map->phys_map[val].enabled) {
 | |
| 		/* Discard duplicated CPUID set operation */
 | |
| 		if (vcpu == map->phys_map[val].vcpu) {
 | |
| 			spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
 | |
| 			return 0;
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * New CPUID is already set with other vcpu
 | |
| 		 * Forbid sharing the same CPUID between different vcpus
 | |
| 		 */
 | |
| 		spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, val);
 | |
| 	map->phys_map[val].enabled	= true;
 | |
| 	map->phys_map[val].vcpu		= vcpu;
 | |
| 	spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static inline void kvm_drop_cpuid(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int cpuid;
 | |
| 	struct kvm_phyid_map *map;
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	map = vcpu->kvm->arch.phyid_map;
 | |
| 	cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_CPUID);
 | |
| 
 | |
| 	if (cpuid >= KVM_MAX_PHYID)
 | |
| 		return;
 | |
| 
 | |
| 	spin_lock(&vcpu->kvm->arch.phyid_map_lock);
 | |
| 	if (map->phys_map[cpuid].enabled) {
 | |
| 		map->phys_map[cpuid].vcpu = NULL;
 | |
| 		map->phys_map[cpuid].enabled = false;
 | |
| 		kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, KVM_MAX_PHYID);
 | |
| 	}
 | |
| 	spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
 | |
| }
 | |
| 
 | |
| struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid)
 | |
| {
 | |
| 	struct kvm_phyid_map *map;
 | |
| 
 | |
| 	if (cpuid >= KVM_MAX_PHYID)
 | |
| 		return NULL;
 | |
| 
 | |
| 	map = kvm->arch.phyid_map;
 | |
| 	if (!map->phys_map[cpuid].enabled)
 | |
| 		return NULL;
 | |
| 
 | |
| 	return map->phys_map[cpuid].vcpu;
 | |
| }
 | |
| 
 | |
| static int _kvm_getcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 *val)
 | |
| {
 | |
| 	unsigned long gintc;
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	if (get_gcsr_flag(id) & INVALID_GCSR)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (id == LOONGARCH_CSR_ESTAT) {
 | |
| 		preempt_disable();
 | |
| 		vcpu_load(vcpu);
 | |
| 		/*
 | |
| 		 * Sync pending interrupts into ESTAT so that interrupt
 | |
| 		 * remains during VM migration stage
 | |
| 		 */
 | |
| 		kvm_deliver_intr(vcpu);
 | |
| 		vcpu->arch.aux_inuse &= ~KVM_LARCH_SWCSR_LATEST;
 | |
| 		vcpu_put(vcpu);
 | |
| 		preempt_enable();
 | |
| 
 | |
| 		/* ESTAT IP0~IP7 get from GINTC */
 | |
| 		gintc = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_GINTC) & 0xff;
 | |
| 		*val = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_ESTAT) | (gintc << 2);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Get software CSR state since software state is consistent
 | |
| 	 * with hardware for synchronous ioctl
 | |
| 	 */
 | |
| 	*val = kvm_read_sw_gcsr(csr, id);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int _kvm_setcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 val)
 | |
| {
 | |
| 	int ret = 0, gintc;
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	if (get_gcsr_flag(id) & INVALID_GCSR)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (id == LOONGARCH_CSR_CPUID)
 | |
| 		return kvm_set_cpuid(vcpu, val);
 | |
| 
 | |
| 	if (id == LOONGARCH_CSR_ESTAT) {
 | |
| 		/* ESTAT IP0~IP7 inject through GINTC */
 | |
| 		gintc = (val >> 2) & 0xff;
 | |
| 		kvm_set_sw_gcsr(csr, LOONGARCH_CSR_GINTC, gintc);
 | |
| 
 | |
| 		gintc = val & ~(0xffUL << 2);
 | |
| 		kvm_set_sw_gcsr(csr, LOONGARCH_CSR_ESTAT, gintc);
 | |
| 
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	kvm_write_sw_gcsr(csr, id, val);
 | |
| 
 | |
| 	/*
 | |
| 	 * After modifying the PMU CSR register value of the vcpu.
 | |
| 	 * If the PMU CSRs are used, we need to set KVM_REQ_PMU.
 | |
| 	 */
 | |
| 	if (id >= LOONGARCH_CSR_PERFCTRL0 && id <= LOONGARCH_CSR_PERFCNTR3) {
 | |
| 		unsigned long val;
 | |
| 
 | |
| 		val = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_PERFCTRL0) |
 | |
| 		      kvm_read_sw_gcsr(csr, LOONGARCH_CSR_PERFCTRL1) |
 | |
| 		      kvm_read_sw_gcsr(csr, LOONGARCH_CSR_PERFCTRL2) |
 | |
| 		      kvm_read_sw_gcsr(csr, LOONGARCH_CSR_PERFCTRL3);
 | |
| 
 | |
| 		if (val & KVM_PMU_EVENT_ENABLED)
 | |
| 			kvm_make_request(KVM_REQ_PMU, vcpu);
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int _kvm_get_cpucfg_mask(int id, u64 *v)
 | |
| {
 | |
| 	if (id < 0 || id >= KVM_MAX_CPUCFG_REGS)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	switch (id) {
 | |
| 	case LOONGARCH_CPUCFG0:
 | |
| 		*v = GENMASK(31, 0);
 | |
| 		return 0;
 | |
| 	case LOONGARCH_CPUCFG1:
 | |
| 		/* CPUCFG1_MSGINT is not supported by KVM */
 | |
| 		*v = GENMASK(25, 0);
 | |
| 		return 0;
 | |
| 	case LOONGARCH_CPUCFG2:
 | |
| 		/* CPUCFG2 features unconditionally supported by KVM */
 | |
| 		*v = CPUCFG2_FP     | CPUCFG2_FPSP  | CPUCFG2_FPDP     |
 | |
| 		     CPUCFG2_FPVERS | CPUCFG2_LLFTP | CPUCFG2_LLFTPREV |
 | |
| 		     CPUCFG2_LSPW | CPUCFG2_LAM;
 | |
| 		/*
 | |
| 		 * For the ISA extensions listed below, if one is supported
 | |
| 		 * by the host, then it is also supported by KVM.
 | |
| 		 */
 | |
| 		if (cpu_has_lsx)
 | |
| 			*v |= CPUCFG2_LSX;
 | |
| 		if (cpu_has_lasx)
 | |
| 			*v |= CPUCFG2_LASX;
 | |
| 		if (cpu_has_lbt_x86)
 | |
| 			*v |= CPUCFG2_X86BT;
 | |
| 		if (cpu_has_lbt_arm)
 | |
| 			*v |= CPUCFG2_ARMBT;
 | |
| 		if (cpu_has_lbt_mips)
 | |
| 			*v |= CPUCFG2_MIPSBT;
 | |
| 
 | |
| 		return 0;
 | |
| 	case LOONGARCH_CPUCFG3:
 | |
| 		*v = GENMASK(16, 0);
 | |
| 		return 0;
 | |
| 	case LOONGARCH_CPUCFG4:
 | |
| 	case LOONGARCH_CPUCFG5:
 | |
| 		*v = GENMASK(31, 0);
 | |
| 		return 0;
 | |
| 	case LOONGARCH_CPUCFG6:
 | |
| 		if (cpu_has_pmp)
 | |
| 			*v = GENMASK(14, 0);
 | |
| 		else
 | |
| 			*v = 0;
 | |
| 		return 0;
 | |
| 	case LOONGARCH_CPUCFG16:
 | |
| 		*v = GENMASK(16, 0);
 | |
| 		return 0;
 | |
| 	case LOONGARCH_CPUCFG17 ... LOONGARCH_CPUCFG20:
 | |
| 		*v = GENMASK(30, 0);
 | |
| 		return 0;
 | |
| 	default:
 | |
| 		/*
 | |
| 		 * CPUCFG bits should be zero if reserved by HW or not
 | |
| 		 * supported by KVM.
 | |
| 		 */
 | |
| 		*v = 0;
 | |
| 		return 0;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int kvm_check_cpucfg(int id, u64 val)
 | |
| {
 | |
| 	int ret;
 | |
| 	u64 mask = 0;
 | |
| 
 | |
| 	ret = _kvm_get_cpucfg_mask(id, &mask);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	if (val & ~mask)
 | |
| 		/* Unsupported features and/or the higher 32 bits should not be set */
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	switch (id) {
 | |
| 	case LOONGARCH_CPUCFG2:
 | |
| 		if (!(val & CPUCFG2_LLFTP))
 | |
| 			/* Guests must have a constant timer */
 | |
| 			return -EINVAL;
 | |
| 		if ((val & CPUCFG2_FP) && (!(val & CPUCFG2_FPSP) || !(val & CPUCFG2_FPDP)))
 | |
| 			/* Single and double float point must both be set when FP is enabled */
 | |
| 			return -EINVAL;
 | |
| 		if ((val & CPUCFG2_LSX) && !(val & CPUCFG2_FP))
 | |
| 			/* LSX architecturally implies FP but val does not satisfy that */
 | |
| 			return -EINVAL;
 | |
| 		if ((val & CPUCFG2_LASX) && !(val & CPUCFG2_LSX))
 | |
| 			/* LASX architecturally implies LSX and FP but val does not satisfy that */
 | |
| 			return -EINVAL;
 | |
| 		return 0;
 | |
| 	case LOONGARCH_CPUCFG6:
 | |
| 		if (val & CPUCFG6_PMP) {
 | |
| 			u32 host = read_cpucfg(LOONGARCH_CPUCFG6);
 | |
| 			if ((val & CPUCFG6_PMBITS) != (host & CPUCFG6_PMBITS))
 | |
| 				return -EINVAL;
 | |
| 			if ((val & CPUCFG6_PMNUM) > (host & CPUCFG6_PMNUM))
 | |
| 				return -EINVAL;
 | |
| 			if ((val & CPUCFG6_UPM) && !(host & CPUCFG6_UPM))
 | |
| 				return -EINVAL;
 | |
| 		}
 | |
| 		return 0;
 | |
| 	default:
 | |
| 		/*
 | |
| 		 * Values for the other CPUCFG IDs are not being further validated
 | |
| 		 * besides the mask check above.
 | |
| 		 */
 | |
| 		return 0;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int kvm_get_one_reg(struct kvm_vcpu *vcpu,
 | |
| 		const struct kvm_one_reg *reg, u64 *v)
 | |
| {
 | |
| 	int id, ret = 0;
 | |
| 	u64 type = reg->id & KVM_REG_LOONGARCH_MASK;
 | |
| 
 | |
| 	switch (type) {
 | |
| 	case KVM_REG_LOONGARCH_CSR:
 | |
| 		id = KVM_GET_IOC_CSR_IDX(reg->id);
 | |
| 		ret = _kvm_getcsr(vcpu, id, v);
 | |
| 		break;
 | |
| 	case KVM_REG_LOONGARCH_CPUCFG:
 | |
| 		id = KVM_GET_IOC_CPUCFG_IDX(reg->id);
 | |
| 		if (id >= 0 && id < KVM_MAX_CPUCFG_REGS)
 | |
| 			*v = vcpu->arch.cpucfg[id];
 | |
| 		else
 | |
| 			ret = -EINVAL;
 | |
| 		break;
 | |
| 	case KVM_REG_LOONGARCH_LBT:
 | |
| 		if (!kvm_guest_has_lbt(&vcpu->arch))
 | |
| 			return -ENXIO;
 | |
| 
 | |
| 		switch (reg->id) {
 | |
| 		case KVM_REG_LOONGARCH_LBT_SCR0:
 | |
| 			*v = vcpu->arch.lbt.scr0;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_SCR1:
 | |
| 			*v = vcpu->arch.lbt.scr1;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_SCR2:
 | |
| 			*v = vcpu->arch.lbt.scr2;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_SCR3:
 | |
| 			*v = vcpu->arch.lbt.scr3;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_EFLAGS:
 | |
| 			*v = vcpu->arch.lbt.eflags;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_FTOP:
 | |
| 			*v = vcpu->arch.fpu.ftop;
 | |
| 			break;
 | |
| 		default:
 | |
| 			ret = -EINVAL;
 | |
| 			break;
 | |
| 		}
 | |
| 		break;
 | |
| 	case KVM_REG_LOONGARCH_KVM:
 | |
| 		switch (reg->id) {
 | |
| 		case KVM_REG_LOONGARCH_COUNTER:
 | |
| 			*v = drdtime() + vcpu->kvm->arch.time_offset;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_DEBUG_INST:
 | |
| 			*v = INSN_HVCL | KVM_HCALL_SWDBG;
 | |
| 			break;
 | |
| 		default:
 | |
| 			ret = -EINVAL;
 | |
| 			break;
 | |
| 		}
 | |
| 		break;
 | |
| 	default:
 | |
| 		ret = -EINVAL;
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int kvm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 	u64 v, size = reg->id & KVM_REG_SIZE_MASK;
 | |
| 
 | |
| 	switch (size) {
 | |
| 	case KVM_REG_SIZE_U64:
 | |
| 		ret = kvm_get_one_reg(vcpu, reg, &v);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		ret = put_user(v, (u64 __user *)(long)reg->addr);
 | |
| 		break;
 | |
| 	default:
 | |
| 		ret = -EINVAL;
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int kvm_set_one_reg(struct kvm_vcpu *vcpu,
 | |
| 			const struct kvm_one_reg *reg, u64 v)
 | |
| {
 | |
| 	int id, ret = 0;
 | |
| 	u64 type = reg->id & KVM_REG_LOONGARCH_MASK;
 | |
| 
 | |
| 	switch (type) {
 | |
| 	case KVM_REG_LOONGARCH_CSR:
 | |
| 		id = KVM_GET_IOC_CSR_IDX(reg->id);
 | |
| 		ret = _kvm_setcsr(vcpu, id, v);
 | |
| 		break;
 | |
| 	case KVM_REG_LOONGARCH_CPUCFG:
 | |
| 		id = KVM_GET_IOC_CPUCFG_IDX(reg->id);
 | |
| 		ret = kvm_check_cpucfg(id, v);
 | |
| 		if (ret)
 | |
| 			break;
 | |
| 		vcpu->arch.cpucfg[id] = (u32)v;
 | |
| 		if (id == LOONGARCH_CPUCFG6)
 | |
| 			vcpu->arch.max_pmu_csrid =
 | |
| 				LOONGARCH_CSR_PERFCTRL0 + 2 * kvm_get_pmu_num(&vcpu->arch) + 1;
 | |
| 		break;
 | |
| 	case KVM_REG_LOONGARCH_LBT:
 | |
| 		if (!kvm_guest_has_lbt(&vcpu->arch))
 | |
| 			return -ENXIO;
 | |
| 
 | |
| 		switch (reg->id) {
 | |
| 		case KVM_REG_LOONGARCH_LBT_SCR0:
 | |
| 			vcpu->arch.lbt.scr0 = v;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_SCR1:
 | |
| 			vcpu->arch.lbt.scr1 = v;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_SCR2:
 | |
| 			vcpu->arch.lbt.scr2 = v;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_SCR3:
 | |
| 			vcpu->arch.lbt.scr3 = v;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_EFLAGS:
 | |
| 			vcpu->arch.lbt.eflags = v;
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_LBT_FTOP:
 | |
| 			vcpu->arch.fpu.ftop = v;
 | |
| 			break;
 | |
| 		default:
 | |
| 			ret = -EINVAL;
 | |
| 			break;
 | |
| 		}
 | |
| 		break;
 | |
| 	case KVM_REG_LOONGARCH_KVM:
 | |
| 		switch (reg->id) {
 | |
| 		case KVM_REG_LOONGARCH_COUNTER:
 | |
| 			/*
 | |
| 			 * gftoffset is relative with board, not vcpu
 | |
| 			 * only set for the first time for smp system
 | |
| 			 */
 | |
| 			if (vcpu->vcpu_id == 0)
 | |
| 				vcpu->kvm->arch.time_offset = (signed long)(v - drdtime());
 | |
| 			break;
 | |
| 		case KVM_REG_LOONGARCH_VCPU_RESET:
 | |
| 			vcpu->arch.st.guest_addr = 0;
 | |
| 			memset(&vcpu->arch.irq_pending, 0, sizeof(vcpu->arch.irq_pending));
 | |
| 			memset(&vcpu->arch.irq_clear, 0, sizeof(vcpu->arch.irq_clear));
 | |
| 			break;
 | |
| 		default:
 | |
| 			ret = -EINVAL;
 | |
| 			break;
 | |
| 		}
 | |
| 		break;
 | |
| 	default:
 | |
| 		ret = -EINVAL;
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int kvm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 	u64 v, size = reg->id & KVM_REG_SIZE_MASK;
 | |
| 
 | |
| 	switch (size) {
 | |
| 	case KVM_REG_SIZE_U64:
 | |
| 		ret = get_user(v, (u64 __user *)(long)reg->addr);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		break;
 | |
| 	default:
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	return kvm_set_one_reg(vcpu, reg, v);
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
 | |
| {
 | |
| 	return -ENOIOCTLCMD;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
 | |
| {
 | |
| 	return -ENOIOCTLCMD;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
 | |
| {
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = 0; i < ARRAY_SIZE(vcpu->arch.gprs); i++)
 | |
| 		regs->gpr[i] = vcpu->arch.gprs[i];
 | |
| 
 | |
| 	regs->pc = vcpu->arch.pc;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
 | |
| {
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = 1; i < ARRAY_SIZE(vcpu->arch.gprs); i++)
 | |
| 		vcpu->arch.gprs[i] = regs->gpr[i];
 | |
| 
 | |
| 	vcpu->arch.gprs[0] = 0; /* zero is special, and cannot be set. */
 | |
| 	vcpu->arch.pc = regs->pc;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu,
 | |
| 				     struct kvm_enable_cap *cap)
 | |
| {
 | |
| 	/* FPU is enabled by default, will support LSX/LASX later. */
 | |
| 	return -EINVAL;
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_cpucfg_has_attr(struct kvm_vcpu *vcpu,
 | |
| 					 struct kvm_device_attr *attr)
 | |
| {
 | |
| 	switch (attr->attr) {
 | |
| 	case LOONGARCH_CPUCFG2:
 | |
| 	case LOONGARCH_CPUCFG6:
 | |
| 		return 0;
 | |
| 	case CPUCFG_KVM_FEATURE:
 | |
| 		return 0;
 | |
| 	default:
 | |
| 		return -ENXIO;
 | |
| 	}
 | |
| 
 | |
| 	return -ENXIO;
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_pvtime_has_attr(struct kvm_vcpu *vcpu,
 | |
| 					 struct kvm_device_attr *attr)
 | |
| {
 | |
| 	if (!kvm_guest_has_pv_feature(vcpu, KVM_FEATURE_STEAL_TIME)
 | |
| 			|| attr->attr != KVM_LOONGARCH_VCPU_PVTIME_GPA)
 | |
| 		return -ENXIO;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_vcpu_has_attr(struct kvm_vcpu *vcpu,
 | |
| 				       struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int ret = -ENXIO;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_LOONGARCH_VCPU_CPUCFG:
 | |
| 		ret = kvm_loongarch_cpucfg_has_attr(vcpu, attr);
 | |
| 		break;
 | |
| 	case KVM_LOONGARCH_VCPU_PVTIME_CTRL:
 | |
| 		ret = kvm_loongarch_pvtime_has_attr(vcpu, attr);
 | |
| 		break;
 | |
| 	default:
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_cpucfg_get_attr(struct kvm_vcpu *vcpu,
 | |
| 					 struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 	uint64_t val;
 | |
| 	uint64_t __user *uaddr = (uint64_t __user *)attr->addr;
 | |
| 
 | |
| 	switch (attr->attr) {
 | |
| 	case 0 ... (KVM_MAX_CPUCFG_REGS - 1):
 | |
| 		ret = _kvm_get_cpucfg_mask(attr->attr, &val);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		break;
 | |
| 	case CPUCFG_KVM_FEATURE:
 | |
| 		val = vcpu->kvm->arch.pv_features & LOONGARCH_PV_FEAT_MASK;
 | |
| 		break;
 | |
| 	default:
 | |
| 		return -ENXIO;
 | |
| 	}
 | |
| 
 | |
| 	put_user(val, uaddr);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_pvtime_get_attr(struct kvm_vcpu *vcpu,
 | |
| 					 struct kvm_device_attr *attr)
 | |
| {
 | |
| 	u64 gpa;
 | |
| 	u64 __user *user = (u64 __user *)attr->addr;
 | |
| 
 | |
| 	if (!kvm_guest_has_pv_feature(vcpu, KVM_FEATURE_STEAL_TIME)
 | |
| 			|| attr->attr != KVM_LOONGARCH_VCPU_PVTIME_GPA)
 | |
| 		return -ENXIO;
 | |
| 
 | |
| 	gpa = vcpu->arch.st.guest_addr;
 | |
| 	if (put_user(gpa, user))
 | |
| 		return -EFAULT;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_vcpu_get_attr(struct kvm_vcpu *vcpu,
 | |
| 				       struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int ret = -ENXIO;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_LOONGARCH_VCPU_CPUCFG:
 | |
| 		ret = kvm_loongarch_cpucfg_get_attr(vcpu, attr);
 | |
| 		break;
 | |
| 	case KVM_LOONGARCH_VCPU_PVTIME_CTRL:
 | |
| 		ret = kvm_loongarch_pvtime_get_attr(vcpu, attr);
 | |
| 		break;
 | |
| 	default:
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_cpucfg_set_attr(struct kvm_vcpu *vcpu,
 | |
| 					 struct kvm_device_attr *attr)
 | |
| {
 | |
| 	u64 val, valid;
 | |
| 	u64 __user *user = (u64 __user *)attr->addr;
 | |
| 	struct kvm *kvm = vcpu->kvm;
 | |
| 
 | |
| 	switch (attr->attr) {
 | |
| 	case CPUCFG_KVM_FEATURE:
 | |
| 		if (get_user(val, user))
 | |
| 			return -EFAULT;
 | |
| 
 | |
| 		valid = LOONGARCH_PV_FEAT_MASK;
 | |
| 		if (val & ~valid)
 | |
| 			return -EINVAL;
 | |
| 
 | |
| 		/* All vCPUs need set the same PV features */
 | |
| 		if ((kvm->arch.pv_features & LOONGARCH_PV_FEAT_UPDATED)
 | |
| 				&& ((kvm->arch.pv_features & valid) != val))
 | |
| 			return -EINVAL;
 | |
| 		kvm->arch.pv_features = val | LOONGARCH_PV_FEAT_UPDATED;
 | |
| 		return 0;
 | |
| 	default:
 | |
| 		return -ENXIO;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_pvtime_set_attr(struct kvm_vcpu *vcpu,
 | |
| 					 struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int idx, ret = 0;
 | |
| 	u64 gpa, __user *user = (u64 __user *)attr->addr;
 | |
| 	struct kvm *kvm = vcpu->kvm;
 | |
| 
 | |
| 	if (!kvm_guest_has_pv_feature(vcpu, KVM_FEATURE_STEAL_TIME)
 | |
| 			|| attr->attr != KVM_LOONGARCH_VCPU_PVTIME_GPA)
 | |
| 		return -ENXIO;
 | |
| 
 | |
| 	if (get_user(gpa, user))
 | |
| 		return -EFAULT;
 | |
| 
 | |
| 	if (gpa & ~(KVM_STEAL_PHYS_MASK | KVM_STEAL_PHYS_VALID))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (!(gpa & KVM_STEAL_PHYS_VALID)) {
 | |
| 		vcpu->arch.st.guest_addr = gpa;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	/* Check the address is in a valid memslot */
 | |
| 	idx = srcu_read_lock(&kvm->srcu);
 | |
| 	if (kvm_is_error_hva(gfn_to_hva(kvm, gpa >> PAGE_SHIFT)))
 | |
| 		ret = -EINVAL;
 | |
| 	srcu_read_unlock(&kvm->srcu, idx);
 | |
| 
 | |
| 	if (!ret) {
 | |
| 		vcpu->arch.st.guest_addr = gpa;
 | |
| 		vcpu->arch.st.last_steal = current->sched_info.run_delay;
 | |
| 		kvm_make_request(KVM_REQ_STEAL_UPDATE, vcpu);
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int kvm_loongarch_vcpu_set_attr(struct kvm_vcpu *vcpu,
 | |
| 				       struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int ret = -ENXIO;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_LOONGARCH_VCPU_CPUCFG:
 | |
| 		ret = kvm_loongarch_cpucfg_set_attr(vcpu, attr);
 | |
| 		break;
 | |
| 	case KVM_LOONGARCH_VCPU_PVTIME_CTRL:
 | |
| 		ret = kvm_loongarch_pvtime_set_attr(vcpu, attr);
 | |
| 		break;
 | |
| 	default:
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| long kvm_arch_vcpu_ioctl(struct file *filp,
 | |
| 			 unsigned int ioctl, unsigned long arg)
 | |
| {
 | |
| 	long r;
 | |
| 	struct kvm_device_attr attr;
 | |
| 	void __user *argp = (void __user *)arg;
 | |
| 	struct kvm_vcpu *vcpu = filp->private_data;
 | |
| 
 | |
| 	/*
 | |
| 	 * Only software CSR should be modified
 | |
| 	 *
 | |
| 	 * If any hardware CSR register is modified, vcpu_load/vcpu_put pair
 | |
| 	 * should be used. Since CSR registers owns by this vcpu, if switch
 | |
| 	 * to other vcpus, other vcpus need reload CSR registers.
 | |
| 	 *
 | |
| 	 * If software CSR is modified, bit KVM_LARCH_HWCSR_USABLE should
 | |
| 	 * be clear in vcpu->arch.aux_inuse, and vcpu_load will check
 | |
| 	 * aux_inuse flag and reload CSR registers form software.
 | |
| 	 */
 | |
| 
 | |
| 	switch (ioctl) {
 | |
| 	case KVM_SET_ONE_REG:
 | |
| 	case KVM_GET_ONE_REG: {
 | |
| 		struct kvm_one_reg reg;
 | |
| 
 | |
| 		r = -EFAULT;
 | |
| 		if (copy_from_user(®, argp, sizeof(reg)))
 | |
| 			break;
 | |
| 		if (ioctl == KVM_SET_ONE_REG) {
 | |
| 			r = kvm_set_reg(vcpu, ®);
 | |
| 			vcpu->arch.aux_inuse &= ~KVM_LARCH_HWCSR_USABLE;
 | |
| 		} else
 | |
| 			r = kvm_get_reg(vcpu, ®);
 | |
| 		break;
 | |
| 	}
 | |
| 	case KVM_ENABLE_CAP: {
 | |
| 		struct kvm_enable_cap cap;
 | |
| 
 | |
| 		r = -EFAULT;
 | |
| 		if (copy_from_user(&cap, argp, sizeof(cap)))
 | |
| 			break;
 | |
| 		r = kvm_vcpu_ioctl_enable_cap(vcpu, &cap);
 | |
| 		break;
 | |
| 	}
 | |
| 	case KVM_HAS_DEVICE_ATTR: {
 | |
| 		r = -EFAULT;
 | |
| 		if (copy_from_user(&attr, argp, sizeof(attr)))
 | |
| 			break;
 | |
| 		r = kvm_loongarch_vcpu_has_attr(vcpu, &attr);
 | |
| 		break;
 | |
| 	}
 | |
| 	case KVM_GET_DEVICE_ATTR: {
 | |
| 		r = -EFAULT;
 | |
| 		if (copy_from_user(&attr, argp, sizeof(attr)))
 | |
| 			break;
 | |
| 		r = kvm_loongarch_vcpu_get_attr(vcpu, &attr);
 | |
| 		break;
 | |
| 	}
 | |
| 	case KVM_SET_DEVICE_ATTR: {
 | |
| 		r = -EFAULT;
 | |
| 		if (copy_from_user(&attr, argp, sizeof(attr)))
 | |
| 			break;
 | |
| 		r = kvm_loongarch_vcpu_set_attr(vcpu, &attr);
 | |
| 		break;
 | |
| 	}
 | |
| 	default:
 | |
| 		r = -ENOIOCTLCMD;
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return r;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
 | |
| {
 | |
| 	int i = 0;
 | |
| 
 | |
| 	fpu->fcc = vcpu->arch.fpu.fcc;
 | |
| 	fpu->fcsr = vcpu->arch.fpu.fcsr;
 | |
| 	for (i = 0; i < NUM_FPU_REGS; i++)
 | |
| 		memcpy(&fpu->fpr[i], &vcpu->arch.fpu.fpr[i], FPU_REG_WIDTH / 64);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
 | |
| {
 | |
| 	int i = 0;
 | |
| 
 | |
| 	vcpu->arch.fpu.fcc = fpu->fcc;
 | |
| 	vcpu->arch.fpu.fcsr = fpu->fcsr;
 | |
| 	for (i = 0; i < NUM_FPU_REGS; i++)
 | |
| 		memcpy(&vcpu->arch.fpu.fpr[i], &fpu->fpr[i], FPU_REG_WIDTH / 64);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| #ifdef CONFIG_CPU_HAS_LBT
 | |
| int kvm_own_lbt(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	if (!kvm_guest_has_lbt(&vcpu->arch))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	preempt_disable();
 | |
| 	set_csr_euen(CSR_EUEN_LBTEN);
 | |
| 	_restore_lbt(&vcpu->arch.lbt);
 | |
| 	vcpu->arch.aux_inuse |= KVM_LARCH_LBT;
 | |
| 	preempt_enable();
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void kvm_lose_lbt(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	preempt_disable();
 | |
| 	if (vcpu->arch.aux_inuse & KVM_LARCH_LBT) {
 | |
| 		_save_lbt(&vcpu->arch.lbt);
 | |
| 		clear_csr_euen(CSR_EUEN_LBTEN);
 | |
| 		vcpu->arch.aux_inuse &= ~KVM_LARCH_LBT;
 | |
| 	}
 | |
| 	preempt_enable();
 | |
| }
 | |
| 
 | |
| static void kvm_check_fcsr(struct kvm_vcpu *vcpu, unsigned long fcsr)
 | |
| {
 | |
| 	/*
 | |
| 	 * If TM is enabled, top register save/restore will
 | |
| 	 * cause lbt exception, here enable lbt in advance
 | |
| 	 */
 | |
| 	if (fcsr & FPU_CSR_TM)
 | |
| 		kvm_own_lbt(vcpu);
 | |
| }
 | |
| 
 | |
| static void kvm_check_fcsr_alive(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	if (vcpu->arch.aux_inuse & KVM_LARCH_FPU) {
 | |
| 		if (vcpu->arch.aux_inuse & KVM_LARCH_LBT)
 | |
| 			return;
 | |
| 		kvm_check_fcsr(vcpu, read_fcsr(LOONGARCH_FCSR0));
 | |
| 	}
 | |
| }
 | |
| #else
 | |
| static inline void kvm_lose_lbt(struct kvm_vcpu *vcpu) { }
 | |
| static inline void kvm_check_fcsr(struct kvm_vcpu *vcpu, unsigned long fcsr) { }
 | |
| static inline void kvm_check_fcsr_alive(struct kvm_vcpu *vcpu) { }
 | |
| #endif
 | |
| 
 | |
| /* Enable FPU and restore context */
 | |
| void kvm_own_fpu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	preempt_disable();
 | |
| 
 | |
| 	/*
 | |
| 	 * Enable FPU for guest
 | |
| 	 * Set FR and FRE according to guest context
 | |
| 	 */
 | |
| 	kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr);
 | |
| 	set_csr_euen(CSR_EUEN_FPEN);
 | |
| 
 | |
| 	kvm_restore_fpu(&vcpu->arch.fpu);
 | |
| 	vcpu->arch.aux_inuse |= KVM_LARCH_FPU;
 | |
| 	trace_kvm_aux(vcpu, KVM_TRACE_AUX_RESTORE, KVM_TRACE_AUX_FPU);
 | |
| 
 | |
| 	preempt_enable();
 | |
| }
 | |
| 
 | |
| #ifdef CONFIG_CPU_HAS_LSX
 | |
| /* Enable LSX and restore context */
 | |
| int kvm_own_lsx(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	if (!kvm_guest_has_fpu(&vcpu->arch) || !kvm_guest_has_lsx(&vcpu->arch))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	preempt_disable();
 | |
| 
 | |
| 	/* Enable LSX for guest */
 | |
| 	kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr);
 | |
| 	set_csr_euen(CSR_EUEN_LSXEN | CSR_EUEN_FPEN);
 | |
| 	switch (vcpu->arch.aux_inuse & KVM_LARCH_FPU) {
 | |
| 	case KVM_LARCH_FPU:
 | |
| 		/*
 | |
| 		 * Guest FPU state already loaded,
 | |
| 		 * only restore upper LSX state
 | |
| 		 */
 | |
| 		_restore_lsx_upper(&vcpu->arch.fpu);
 | |
| 		break;
 | |
| 	default:
 | |
| 		/* Neither FP or LSX already active,
 | |
| 		 * restore full LSX state
 | |
| 		 */
 | |
| 		kvm_restore_lsx(&vcpu->arch.fpu);
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	trace_kvm_aux(vcpu, KVM_TRACE_AUX_RESTORE, KVM_TRACE_AUX_LSX);
 | |
| 	vcpu->arch.aux_inuse |= KVM_LARCH_LSX | KVM_LARCH_FPU;
 | |
| 	preempt_enable();
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| #endif
 | |
| 
 | |
| #ifdef CONFIG_CPU_HAS_LASX
 | |
| /* Enable LASX and restore context */
 | |
| int kvm_own_lasx(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	if (!kvm_guest_has_fpu(&vcpu->arch) || !kvm_guest_has_lsx(&vcpu->arch) || !kvm_guest_has_lasx(&vcpu->arch))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	preempt_disable();
 | |
| 
 | |
| 	kvm_check_fcsr(vcpu, vcpu->arch.fpu.fcsr);
 | |
| 	set_csr_euen(CSR_EUEN_FPEN | CSR_EUEN_LSXEN | CSR_EUEN_LASXEN);
 | |
| 	switch (vcpu->arch.aux_inuse & (KVM_LARCH_FPU | KVM_LARCH_LSX)) {
 | |
| 	case KVM_LARCH_LSX:
 | |
| 	case KVM_LARCH_LSX | KVM_LARCH_FPU:
 | |
| 		/* Guest LSX state already loaded, only restore upper LASX state */
 | |
| 		_restore_lasx_upper(&vcpu->arch.fpu);
 | |
| 		break;
 | |
| 	case KVM_LARCH_FPU:
 | |
| 		/* Guest FP state already loaded, only restore upper LSX & LASX state */
 | |
| 		_restore_lsx_upper(&vcpu->arch.fpu);
 | |
| 		_restore_lasx_upper(&vcpu->arch.fpu);
 | |
| 		break;
 | |
| 	default:
 | |
| 		/* Neither FP or LSX already active, restore full LASX state */
 | |
| 		kvm_restore_lasx(&vcpu->arch.fpu);
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	trace_kvm_aux(vcpu, KVM_TRACE_AUX_RESTORE, KVM_TRACE_AUX_LASX);
 | |
| 	vcpu->arch.aux_inuse |= KVM_LARCH_LASX | KVM_LARCH_LSX | KVM_LARCH_FPU;
 | |
| 	preempt_enable();
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| #endif
 | |
| 
 | |
| /* Save context and disable FPU */
 | |
| void kvm_lose_fpu(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	preempt_disable();
 | |
| 
 | |
| 	kvm_check_fcsr_alive(vcpu);
 | |
| 	if (vcpu->arch.aux_inuse & KVM_LARCH_LASX) {
 | |
| 		kvm_save_lasx(&vcpu->arch.fpu);
 | |
| 		vcpu->arch.aux_inuse &= ~(KVM_LARCH_LSX | KVM_LARCH_FPU | KVM_LARCH_LASX);
 | |
| 		trace_kvm_aux(vcpu, KVM_TRACE_AUX_SAVE, KVM_TRACE_AUX_LASX);
 | |
| 
 | |
| 		/* Disable LASX & LSX & FPU */
 | |
| 		clear_csr_euen(CSR_EUEN_FPEN | CSR_EUEN_LSXEN | CSR_EUEN_LASXEN);
 | |
| 	} else if (vcpu->arch.aux_inuse & KVM_LARCH_LSX) {
 | |
| 		kvm_save_lsx(&vcpu->arch.fpu);
 | |
| 		vcpu->arch.aux_inuse &= ~(KVM_LARCH_LSX | KVM_LARCH_FPU);
 | |
| 		trace_kvm_aux(vcpu, KVM_TRACE_AUX_SAVE, KVM_TRACE_AUX_LSX);
 | |
| 
 | |
| 		/* Disable LSX & FPU */
 | |
| 		clear_csr_euen(CSR_EUEN_FPEN | CSR_EUEN_LSXEN);
 | |
| 	} else if (vcpu->arch.aux_inuse & KVM_LARCH_FPU) {
 | |
| 		kvm_save_fpu(&vcpu->arch.fpu);
 | |
| 		vcpu->arch.aux_inuse &= ~KVM_LARCH_FPU;
 | |
| 		trace_kvm_aux(vcpu, KVM_TRACE_AUX_SAVE, KVM_TRACE_AUX_FPU);
 | |
| 
 | |
| 		/* Disable FPU */
 | |
| 		clear_csr_euen(CSR_EUEN_FPEN);
 | |
| 	}
 | |
| 	kvm_lose_lbt(vcpu);
 | |
| 
 | |
| 	preempt_enable();
 | |
| }
 | |
| 
 | |
| int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq)
 | |
| {
 | |
| 	int intr = (int)irq->irq;
 | |
| 
 | |
| 	if (intr > 0)
 | |
| 		kvm_queue_irq(vcpu, intr);
 | |
| 	else if (intr < 0)
 | |
| 		kvm_dequeue_irq(vcpu, -intr);
 | |
| 	else {
 | |
| 		kvm_err("%s: invalid interrupt ioctl %d\n", __func__, irq->irq);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	kvm_vcpu_kick(vcpu);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| long kvm_arch_vcpu_async_ioctl(struct file *filp,
 | |
| 			       unsigned int ioctl, unsigned long arg)
 | |
| {
 | |
| 	void __user *argp = (void __user *)arg;
 | |
| 	struct kvm_vcpu *vcpu = filp->private_data;
 | |
| 
 | |
| 	if (ioctl == KVM_INTERRUPT) {
 | |
| 		struct kvm_interrupt irq;
 | |
| 
 | |
| 		if (copy_from_user(&irq, argp, sizeof(irq)))
 | |
| 			return -EFAULT;
 | |
| 
 | |
| 		kvm_debug("[%d] %s: irq: %d\n", vcpu->vcpu_id, __func__, irq.irq);
 | |
| 
 | |
| 		return kvm_vcpu_ioctl_interrupt(vcpu, &irq);
 | |
| 	}
 | |
| 
 | |
| 	return -ENOIOCTLCMD;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_precreate(struct kvm *kvm, unsigned int id)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	unsigned long timer_hz;
 | |
| 	struct loongarch_csrs *csr;
 | |
| 
 | |
| 	vcpu->arch.vpid = 0;
 | |
| 	vcpu->arch.flush_gpa = INVALID_GPA;
 | |
| 
 | |
| 	hrtimer_init(&vcpu->arch.swtimer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS_PINNED_HARD);
 | |
| 	vcpu->arch.swtimer.function = kvm_swtimer_wakeup;
 | |
| 
 | |
| 	vcpu->arch.handle_exit = kvm_handle_exit;
 | |
| 	vcpu->arch.guest_eentry = (unsigned long)kvm_loongarch_ops->exc_entry;
 | |
| 	vcpu->arch.csr = kzalloc(sizeof(struct loongarch_csrs), GFP_KERNEL);
 | |
| 	if (!vcpu->arch.csr)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	/*
 | |
| 	 * All kvm exceptions share one exception entry, and host <-> guest
 | |
| 	 * switch also switch ECFG.VS field, keep host ECFG.VS info here.
 | |
| 	 */
 | |
| 	vcpu->arch.host_ecfg = (read_csr_ecfg() & CSR_ECFG_VS);
 | |
| 
 | |
| 	/* Init */
 | |
| 	vcpu->arch.last_sched_cpu = -1;
 | |
| 
 | |
| 	/* Init ipi_state lock */
 | |
| 	spin_lock_init(&vcpu->arch.ipi_state.lock);
 | |
| 
 | |
| 	/*
 | |
| 	 * Initialize guest register state to valid architectural reset state.
 | |
| 	 */
 | |
| 	timer_hz = calc_const_freq();
 | |
| 	kvm_init_timer(vcpu, timer_hz);
 | |
| 
 | |
| 	/* Set Initialize mode for guest */
 | |
| 	csr = vcpu->arch.csr;
 | |
| 	kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CRMD, CSR_CRMD_DA);
 | |
| 
 | |
| 	/* Set cpuid */
 | |
| 	kvm_write_sw_gcsr(csr, LOONGARCH_CSR_TMID, vcpu->vcpu_id);
 | |
| 	kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, KVM_MAX_PHYID);
 | |
| 
 | |
| 	/* Start with no pending virtual guest interrupts */
 | |
| 	csr->csrs[LOONGARCH_CSR_GINTC] = 0;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| }
 | |
| 
 | |
| void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int cpu;
 | |
| 	struct kvm_context *context;
 | |
| 
 | |
| 	hrtimer_cancel(&vcpu->arch.swtimer);
 | |
| 	kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
 | |
| 	kvm_drop_cpuid(vcpu);
 | |
| 	kfree(vcpu->arch.csr);
 | |
| 
 | |
| 	/*
 | |
| 	 * If the vCPU is freed and reused as another vCPU, we don't want the
 | |
| 	 * matching pointer wrongly hanging around in last_vcpu.
 | |
| 	 */
 | |
| 	for_each_possible_cpu(cpu) {
 | |
| 		context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
 | |
| 		if (context->last_vcpu == vcpu)
 | |
| 			context->last_vcpu = NULL;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int _kvm_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
 | |
| {
 | |
| 	bool migrated;
 | |
| 	struct kvm_context *context;
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	/*
 | |
| 	 * Have we migrated to a different CPU?
 | |
| 	 * If so, any old guest TLB state may be stale.
 | |
| 	 */
 | |
| 	migrated = (vcpu->arch.last_sched_cpu != cpu);
 | |
| 
 | |
| 	/*
 | |
| 	 * Was this the last vCPU to run on this CPU?
 | |
| 	 * If not, any old guest state from this vCPU will have been clobbered.
 | |
| 	 */
 | |
| 	context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
 | |
| 	if (migrated || (context->last_vcpu != vcpu))
 | |
| 		vcpu->arch.aux_inuse &= ~KVM_LARCH_HWCSR_USABLE;
 | |
| 	context->last_vcpu = vcpu;
 | |
| 
 | |
| 	/* Restore timer state regardless */
 | |
| 	kvm_restore_timer(vcpu);
 | |
| 	kvm_make_request(KVM_REQ_STEAL_UPDATE, vcpu);
 | |
| 
 | |
| 	/* Restore hardware PMU CSRs */
 | |
| 	kvm_restore_pmu(vcpu);
 | |
| 
 | |
| 	/* Don't bother restoring registers multiple times unless necessary */
 | |
| 	if (vcpu->arch.aux_inuse & KVM_LARCH_HWCSR_USABLE)
 | |
| 		return 0;
 | |
| 
 | |
| 	write_csr_gcntc((ulong)vcpu->kvm->arch.time_offset);
 | |
| 
 | |
| 	/* Restore guest CSR registers */
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_CRMD);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PRMD);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_EUEN);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_MISC);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_ECFG);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_ERA);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_BADV);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_BADI);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_EENTRY);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBIDX);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBEHI);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBELO0);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBELO1);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_ASID);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PGDL);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PGDH);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PWCTL0);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_PWCTL1);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_STLBPGSIZE);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_RVACFG);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_CPUID);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_KS0);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_KS1);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_KS2);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_KS3);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_KS4);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_KS5);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_KS6);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_KS7);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TMID);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_CNTC);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBRENTRY);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBRBADV);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBRERA);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBRSAVE);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBRELO0);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBRELO1);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBREHI);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TLBRPRMD);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_DMWIN0);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_DMWIN1);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_DMWIN2);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_DMWIN3);
 | |
| 	kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_LLBCTL);
 | |
| 
 | |
| 	/* Restore Root.GINTC from unused Guest.GINTC register */
 | |
| 	write_csr_gintc(csr->csrs[LOONGARCH_CSR_GINTC]);
 | |
| 
 | |
| 	/*
 | |
| 	 * We should clear linked load bit to break interrupted atomics. This
 | |
| 	 * prevents a SC on the next vCPU from succeeding by matching a LL on
 | |
| 	 * the previous vCPU.
 | |
| 	 */
 | |
| 	if (vcpu->kvm->created_vcpus > 1)
 | |
| 		set_gcsr_llbctl(CSR_LLBCTL_WCLLB);
 | |
| 
 | |
| 	vcpu->arch.aux_inuse |= KVM_LARCH_HWCSR_USABLE;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
 | |
| {
 | |
| 	unsigned long flags;
 | |
| 
 | |
| 	local_irq_save(flags);
 | |
| 	/* Restore guest state to registers */
 | |
| 	_kvm_vcpu_load(vcpu, cpu);
 | |
| 	local_irq_restore(flags);
 | |
| }
 | |
| 
 | |
| static int _kvm_vcpu_put(struct kvm_vcpu *vcpu, int cpu)
 | |
| {
 | |
| 	struct loongarch_csrs *csr = vcpu->arch.csr;
 | |
| 
 | |
| 	kvm_lose_fpu(vcpu);
 | |
| 
 | |
| 	/*
 | |
| 	 * Update CSR state from hardware if software CSR state is stale,
 | |
| 	 * most CSR registers are kept unchanged during process context
 | |
| 	 * switch except CSR registers like remaining timer tick value and
 | |
| 	 * injected interrupt state.
 | |
| 	 */
 | |
| 	if (vcpu->arch.aux_inuse & KVM_LARCH_SWCSR_LATEST)
 | |
| 		goto out;
 | |
| 
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_CRMD);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PRMD);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_EUEN);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_MISC);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_ECFG);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_ERA);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_BADV);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_BADI);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_EENTRY);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBIDX);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBEHI);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBELO0);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBELO1);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_ASID);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PGDL);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PGDH);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PWCTL0);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PWCTL1);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_STLBPGSIZE);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_RVACFG);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_CPUID);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PRCFG1);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PRCFG2);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PRCFG3);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_KS0);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_KS1);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_KS2);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_KS3);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_KS4);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_KS5);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_KS6);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_KS7);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TMID);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_CNTC);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_LLBCTL);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBRENTRY);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBRBADV);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBRERA);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBRSAVE);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBRELO0);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBRELO1);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBREHI);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_TLBRPRMD);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_DMWIN0);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_DMWIN1);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_DMWIN2);
 | |
| 	kvm_save_hw_gcsr(csr, LOONGARCH_CSR_DMWIN3);
 | |
| 
 | |
| 	vcpu->arch.aux_inuse |= KVM_LARCH_SWCSR_LATEST;
 | |
| 
 | |
| out:
 | |
| 	kvm_save_timer(vcpu);
 | |
| 	/* Save Root.GINTC into unused Guest.GINTC register */
 | |
| 	csr->csrs[LOONGARCH_CSR_GINTC] = read_csr_gintc();
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int cpu;
 | |
| 	unsigned long flags;
 | |
| 
 | |
| 	local_irq_save(flags);
 | |
| 	cpu = smp_processor_id();
 | |
| 	vcpu->arch.last_sched_cpu = cpu;
 | |
| 
 | |
| 	/* Save guest state in registers */
 | |
| 	_kvm_vcpu_put(vcpu, cpu);
 | |
| 	local_irq_restore(flags);
 | |
| }
 | |
| 
 | |
| int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	int r = -EINTR;
 | |
| 	struct kvm_run *run = vcpu->run;
 | |
| 
 | |
| 	if (vcpu->mmio_needed) {
 | |
| 		if (!vcpu->mmio_is_write)
 | |
| 			kvm_complete_mmio_read(vcpu, run);
 | |
| 		vcpu->mmio_needed = 0;
 | |
| 	}
 | |
| 
 | |
| 	switch (run->exit_reason) {
 | |
| 	case KVM_EXIT_HYPERCALL:
 | |
| 		kvm_complete_user_service(vcpu, run);
 | |
| 		break;
 | |
| 	case KVM_EXIT_LOONGARCH_IOCSR:
 | |
| 		if (!run->iocsr_io.is_write)
 | |
| 			kvm_complete_iocsr_read(vcpu, run);
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	if (!vcpu->wants_to_run)
 | |
| 		return r;
 | |
| 
 | |
| 	/* Clear exit_reason */
 | |
| 	run->exit_reason = KVM_EXIT_UNKNOWN;
 | |
| 	lose_fpu(1);
 | |
| 	vcpu_load(vcpu);
 | |
| 	kvm_sigset_activate(vcpu);
 | |
| 	r = kvm_pre_enter_guest(vcpu);
 | |
| 	if (r != RESUME_GUEST)
 | |
| 		goto out;
 | |
| 
 | |
| 	guest_timing_enter_irqoff();
 | |
| 	guest_state_enter_irqoff();
 | |
| 	trace_kvm_enter(vcpu);
 | |
| 	r = kvm_loongarch_ops->enter_guest(run, vcpu);
 | |
| 
 | |
| 	trace_kvm_out(vcpu);
 | |
| 	/*
 | |
| 	 * Guest exit is already recorded at kvm_handle_exit()
 | |
| 	 * return value must not be RESUME_GUEST
 | |
| 	 */
 | |
| 	local_irq_enable();
 | |
| out:
 | |
| 	kvm_sigset_deactivate(vcpu);
 | |
| 	vcpu_put(vcpu);
 | |
| 
 | |
| 	return r;
 | |
| }
 |