mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

Move TDX hardware setup to tdx.c, as the code is obviously TDX specific, co-locating the setup with tdx_bringup() makes it easier to see and document the success_disable_tdx "error" path, and configuring the TDX specific hooks in tdx.c reduces the number of globally visible TDX symbols. Reviewed-by: Kai Huang <kai.huang@intel.com> Reviewed-by: Xiaoyao Li <xiaoyao.li@intel.com> Link: https://lore.kernel.org/r/20250523001138.3182794-2-seanjc@google.com Signed-off-by: Sean Christopherson <seanjc@google.com>
205 lines
5.6 KiB
C
205 lines
5.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __KVM_X86_VMX_TDX_H
|
|
#define __KVM_X86_VMX_TDX_H
|
|
|
|
#include "tdx_arch.h"
|
|
#include "tdx_errno.h"
|
|
|
|
#ifdef CONFIG_KVM_INTEL_TDX
|
|
#include "common.h"
|
|
|
|
void tdx_hardware_setup(void);
|
|
int tdx_bringup(void);
|
|
void tdx_cleanup(void);
|
|
|
|
extern bool enable_tdx;
|
|
|
|
/* TDX module hardware states. These follow the TDX module OP_STATEs. */
|
|
enum kvm_tdx_state {
|
|
TD_STATE_UNINITIALIZED = 0,
|
|
TD_STATE_INITIALIZED,
|
|
TD_STATE_RUNNABLE,
|
|
};
|
|
|
|
struct kvm_tdx {
|
|
struct kvm kvm;
|
|
|
|
struct misc_cg *misc_cg;
|
|
int hkid;
|
|
enum kvm_tdx_state state;
|
|
|
|
u64 attributes;
|
|
u64 xfam;
|
|
|
|
u64 tsc_offset;
|
|
u64 tsc_multiplier;
|
|
|
|
struct tdx_td td;
|
|
|
|
/* For KVM_TDX_INIT_MEM_REGION. */
|
|
atomic64_t nr_premapped;
|
|
|
|
/*
|
|
* Prevent vCPUs from TD entry to ensure SEPT zap related SEAMCALLs do
|
|
* not contend with tdh_vp_enter() and TDCALLs.
|
|
* Set/unset is protected with kvm->mmu_lock.
|
|
*/
|
|
bool wait_for_sept_zap;
|
|
};
|
|
|
|
/* TDX module vCPU states */
|
|
enum vcpu_tdx_state {
|
|
VCPU_TD_STATE_UNINITIALIZED = 0,
|
|
VCPU_TD_STATE_INITIALIZED,
|
|
};
|
|
|
|
struct vcpu_tdx {
|
|
struct kvm_vcpu vcpu;
|
|
struct vcpu_vt vt;
|
|
u64 ext_exit_qualification;
|
|
gpa_t exit_gpa;
|
|
struct tdx_module_args vp_enter_args;
|
|
|
|
struct tdx_vp vp;
|
|
|
|
struct list_head cpu_list;
|
|
|
|
u64 vp_enter_ret;
|
|
|
|
enum vcpu_tdx_state state;
|
|
bool guest_entered;
|
|
|
|
u64 map_gpa_next;
|
|
u64 map_gpa_end;
|
|
};
|
|
|
|
void tdh_vp_rd_failed(struct vcpu_tdx *tdx, char *uclass, u32 field, u64 err);
|
|
void tdh_vp_wr_failed(struct vcpu_tdx *tdx, char *uclass, char *op, u32 field,
|
|
u64 val, u64 err);
|
|
|
|
static __always_inline u64 td_tdcs_exec_read64(struct kvm_tdx *kvm_tdx, u32 field)
|
|
{
|
|
u64 err, data;
|
|
|
|
err = tdh_mng_rd(&kvm_tdx->td, TDCS_EXEC(field), &data);
|
|
if (unlikely(err)) {
|
|
pr_err("TDH_MNG_RD[EXEC.0x%x] failed: 0x%llx\n", field, err);
|
|
return 0;
|
|
}
|
|
return data;
|
|
}
|
|
|
|
static __always_inline void tdvps_vmcs_check(u32 field, u8 bits)
|
|
{
|
|
#define VMCS_ENC_ACCESS_TYPE_MASK 0x1UL
|
|
#define VMCS_ENC_ACCESS_TYPE_FULL 0x0UL
|
|
#define VMCS_ENC_ACCESS_TYPE_HIGH 0x1UL
|
|
#define VMCS_ENC_ACCESS_TYPE(field) ((field) & VMCS_ENC_ACCESS_TYPE_MASK)
|
|
|
|
/* TDX is 64bit only. HIGH field isn't supported. */
|
|
BUILD_BUG_ON_MSG(__builtin_constant_p(field) &&
|
|
VMCS_ENC_ACCESS_TYPE(field) == VMCS_ENC_ACCESS_TYPE_HIGH,
|
|
"Read/Write to TD VMCS *_HIGH fields not supported");
|
|
|
|
BUILD_BUG_ON(bits != 16 && bits != 32 && bits != 64);
|
|
|
|
#define VMCS_ENC_WIDTH_MASK GENMASK(14, 13)
|
|
#define VMCS_ENC_WIDTH_16BIT (0UL << 13)
|
|
#define VMCS_ENC_WIDTH_64BIT (1UL << 13)
|
|
#define VMCS_ENC_WIDTH_32BIT (2UL << 13)
|
|
#define VMCS_ENC_WIDTH_NATURAL (3UL << 13)
|
|
#define VMCS_ENC_WIDTH(field) ((field) & VMCS_ENC_WIDTH_MASK)
|
|
|
|
/* TDX is 64bit only. i.e. natural width = 64bit. */
|
|
BUILD_BUG_ON_MSG(bits != 64 && __builtin_constant_p(field) &&
|
|
(VMCS_ENC_WIDTH(field) == VMCS_ENC_WIDTH_64BIT ||
|
|
VMCS_ENC_WIDTH(field) == VMCS_ENC_WIDTH_NATURAL),
|
|
"Invalid TD VMCS access for 64-bit field");
|
|
BUILD_BUG_ON_MSG(bits != 32 && __builtin_constant_p(field) &&
|
|
VMCS_ENC_WIDTH(field) == VMCS_ENC_WIDTH_32BIT,
|
|
"Invalid TD VMCS access for 32-bit field");
|
|
BUILD_BUG_ON_MSG(bits != 16 && __builtin_constant_p(field) &&
|
|
VMCS_ENC_WIDTH(field) == VMCS_ENC_WIDTH_16BIT,
|
|
"Invalid TD VMCS access for 16-bit field");
|
|
}
|
|
|
|
static __always_inline void tdvps_management_check(u64 field, u8 bits) {}
|
|
static __always_inline void tdvps_state_non_arch_check(u64 field, u8 bits) {}
|
|
|
|
#define TDX_BUILD_TDVPS_ACCESSORS(bits, uclass, lclass) \
|
|
static __always_inline u##bits td_##lclass##_read##bits(struct vcpu_tdx *tdx, \
|
|
u32 field) \
|
|
{ \
|
|
u64 err, data; \
|
|
\
|
|
tdvps_##lclass##_check(field, bits); \
|
|
err = tdh_vp_rd(&tdx->vp, TDVPS_##uclass(field), &data); \
|
|
if (unlikely(err)) { \
|
|
tdh_vp_rd_failed(tdx, #uclass, field, err); \
|
|
return 0; \
|
|
} \
|
|
return (u##bits)data; \
|
|
} \
|
|
static __always_inline void td_##lclass##_write##bits(struct vcpu_tdx *tdx, \
|
|
u32 field, u##bits val) \
|
|
{ \
|
|
u64 err; \
|
|
\
|
|
tdvps_##lclass##_check(field, bits); \
|
|
err = tdh_vp_wr(&tdx->vp, TDVPS_##uclass(field), val, \
|
|
GENMASK_ULL(bits - 1, 0)); \
|
|
if (unlikely(err)) \
|
|
tdh_vp_wr_failed(tdx, #uclass, " = ", field, (u64)val, err); \
|
|
} \
|
|
static __always_inline void td_##lclass##_setbit##bits(struct vcpu_tdx *tdx, \
|
|
u32 field, u64 bit) \
|
|
{ \
|
|
u64 err; \
|
|
\
|
|
tdvps_##lclass##_check(field, bits); \
|
|
err = tdh_vp_wr(&tdx->vp, TDVPS_##uclass(field), bit, bit); \
|
|
if (unlikely(err)) \
|
|
tdh_vp_wr_failed(tdx, #uclass, " |= ", field, bit, err); \
|
|
} \
|
|
static __always_inline void td_##lclass##_clearbit##bits(struct vcpu_tdx *tdx, \
|
|
u32 field, u64 bit) \
|
|
{ \
|
|
u64 err; \
|
|
\
|
|
tdvps_##lclass##_check(field, bits); \
|
|
err = tdh_vp_wr(&tdx->vp, TDVPS_##uclass(field), 0, bit); \
|
|
if (unlikely(err)) \
|
|
tdh_vp_wr_failed(tdx, #uclass, " &= ~", field, bit, err);\
|
|
}
|
|
|
|
|
|
bool tdx_interrupt_allowed(struct kvm_vcpu *vcpu);
|
|
int tdx_complete_emulated_msr(struct kvm_vcpu *vcpu, int err);
|
|
|
|
TDX_BUILD_TDVPS_ACCESSORS(16, VMCS, vmcs);
|
|
TDX_BUILD_TDVPS_ACCESSORS(32, VMCS, vmcs);
|
|
TDX_BUILD_TDVPS_ACCESSORS(64, VMCS, vmcs);
|
|
|
|
TDX_BUILD_TDVPS_ACCESSORS(8, MANAGEMENT, management);
|
|
TDX_BUILD_TDVPS_ACCESSORS(64, STATE_NON_ARCH, state_non_arch);
|
|
|
|
#else
|
|
static inline int tdx_bringup(void) { return 0; }
|
|
static inline void tdx_cleanup(void) {}
|
|
|
|
#define enable_tdx 0
|
|
|
|
struct kvm_tdx {
|
|
struct kvm kvm;
|
|
};
|
|
|
|
struct vcpu_tdx {
|
|
struct kvm_vcpu vcpu;
|
|
};
|
|
|
|
static inline bool tdx_interrupt_allowed(struct kvm_vcpu *vcpu) { return false; }
|
|
static inline int tdx_complete_emulated_msr(struct kvm_vcpu *vcpu, int err) { return 0; }
|
|
|
|
#endif
|
|
|
|
#endif
|