/* SPDX-License-Identifier: GPL-2.0 */ #ifndef __KVM_X86_VMX_TDX_H #define __KVM_X86_VMX_TDX_H #include "tdx_arch.h" #include "tdx_errno.h" #ifdef CONFIG_KVM_INTEL_TDX #include "common.h" int tdx_bringup(void); void tdx_cleanup(void); extern bool enable_tdx; /* TDX module hardware states. These follow the TDX module OP_STATEs. */ enum kvm_tdx_state { TD_STATE_UNINITIALIZED = 0, TD_STATE_INITIALIZED, TD_STATE_RUNNABLE, }; struct kvm_tdx { struct kvm kvm; struct misc_cg *misc_cg; int hkid; enum kvm_tdx_state state; u64 attributes; u64 xfam; u64 tsc_offset; u64 tsc_multiplier; struct tdx_td td; /* For KVM_TDX_INIT_MEM_REGION. */ atomic64_t nr_premapped; /* * Prevent vCPUs from TD entry to ensure SEPT zap related SEAMCALLs do * not contend with tdh_vp_enter() and TDCALLs. * Set/unset is protected with kvm->mmu_lock. */ bool wait_for_sept_zap; }; /* TDX module vCPU states */ enum vcpu_tdx_state { VCPU_TD_STATE_UNINITIALIZED = 0, VCPU_TD_STATE_INITIALIZED, }; struct vcpu_tdx { struct kvm_vcpu vcpu; struct vcpu_vt vt; u64 ext_exit_qualification; gpa_t exit_gpa; struct tdx_module_args vp_enter_args; struct tdx_vp vp; struct list_head cpu_list; u64 vp_enter_ret; enum vcpu_tdx_state state; bool guest_entered; u64 map_gpa_next; u64 map_gpa_end; }; void tdh_vp_rd_failed(struct vcpu_tdx *tdx, char *uclass, u32 field, u64 err); void tdh_vp_wr_failed(struct vcpu_tdx *tdx, char *uclass, char *op, u32 field, u64 val, u64 err); static __always_inline u64 td_tdcs_exec_read64(struct kvm_tdx *kvm_tdx, u32 field) { u64 err, data; err = tdh_mng_rd(&kvm_tdx->td, TDCS_EXEC(field), &data); if (unlikely(err)) { pr_err("TDH_MNG_RD[EXEC.0x%x] failed: 0x%llx\n", field, err); return 0; } return data; } static __always_inline void tdvps_vmcs_check(u32 field, u8 bits) { #define VMCS_ENC_ACCESS_TYPE_MASK 0x1UL #define VMCS_ENC_ACCESS_TYPE_FULL 0x0UL #define VMCS_ENC_ACCESS_TYPE_HIGH 0x1UL #define VMCS_ENC_ACCESS_TYPE(field) ((field) & VMCS_ENC_ACCESS_TYPE_MASK) /* TDX is 64bit only. HIGH field isn't supported. */ BUILD_BUG_ON_MSG(__builtin_constant_p(field) && VMCS_ENC_ACCESS_TYPE(field) == VMCS_ENC_ACCESS_TYPE_HIGH, "Read/Write to TD VMCS *_HIGH fields not supported"); BUILD_BUG_ON(bits != 16 && bits != 32 && bits != 64); #define VMCS_ENC_WIDTH_MASK GENMASK(14, 13) #define VMCS_ENC_WIDTH_16BIT (0UL << 13) #define VMCS_ENC_WIDTH_64BIT (1UL << 13) #define VMCS_ENC_WIDTH_32BIT (2UL << 13) #define VMCS_ENC_WIDTH_NATURAL (3UL << 13) #define VMCS_ENC_WIDTH(field) ((field) & VMCS_ENC_WIDTH_MASK) /* TDX is 64bit only. i.e. natural width = 64bit. */ BUILD_BUG_ON_MSG(bits != 64 && __builtin_constant_p(field) && (VMCS_ENC_WIDTH(field) == VMCS_ENC_WIDTH_64BIT || VMCS_ENC_WIDTH(field) == VMCS_ENC_WIDTH_NATURAL), "Invalid TD VMCS access for 64-bit field"); BUILD_BUG_ON_MSG(bits != 32 && __builtin_constant_p(field) && VMCS_ENC_WIDTH(field) == VMCS_ENC_WIDTH_32BIT, "Invalid TD VMCS access for 32-bit field"); BUILD_BUG_ON_MSG(bits != 16 && __builtin_constant_p(field) && VMCS_ENC_WIDTH(field) == VMCS_ENC_WIDTH_16BIT, "Invalid TD VMCS access for 16-bit field"); } static __always_inline void tdvps_management_check(u64 field, u8 bits) {} static __always_inline void tdvps_state_non_arch_check(u64 field, u8 bits) {} #define TDX_BUILD_TDVPS_ACCESSORS(bits, uclass, lclass) \ static __always_inline u##bits td_##lclass##_read##bits(struct vcpu_tdx *tdx, \ u32 field) \ { \ u64 err, data; \ \ tdvps_##lclass##_check(field, bits); \ err = tdh_vp_rd(&tdx->vp, TDVPS_##uclass(field), &data); \ if (unlikely(err)) { \ tdh_vp_rd_failed(tdx, #uclass, field, err); \ return 0; \ } \ return (u##bits)data; \ } \ static __always_inline void td_##lclass##_write##bits(struct vcpu_tdx *tdx, \ u32 field, u##bits val) \ { \ u64 err; \ \ tdvps_##lclass##_check(field, bits); \ err = tdh_vp_wr(&tdx->vp, TDVPS_##uclass(field), val, \ GENMASK_ULL(bits - 1, 0)); \ if (unlikely(err)) \ tdh_vp_wr_failed(tdx, #uclass, " = ", field, (u64)val, err); \ } \ static __always_inline void td_##lclass##_setbit##bits(struct vcpu_tdx *tdx, \ u32 field, u64 bit) \ { \ u64 err; \ \ tdvps_##lclass##_check(field, bits); \ err = tdh_vp_wr(&tdx->vp, TDVPS_##uclass(field), bit, bit); \ if (unlikely(err)) \ tdh_vp_wr_failed(tdx, #uclass, " |= ", field, bit, err); \ } \ static __always_inline void td_##lclass##_clearbit##bits(struct vcpu_tdx *tdx, \ u32 field, u64 bit) \ { \ u64 err; \ \ tdvps_##lclass##_check(field, bits); \ err = tdh_vp_wr(&tdx->vp, TDVPS_##uclass(field), 0, bit); \ if (unlikely(err)) \ tdh_vp_wr_failed(tdx, #uclass, " &= ~", field, bit, err);\ } bool tdx_interrupt_allowed(struct kvm_vcpu *vcpu); int tdx_complete_emulated_msr(struct kvm_vcpu *vcpu, int err); TDX_BUILD_TDVPS_ACCESSORS(16, VMCS, vmcs); TDX_BUILD_TDVPS_ACCESSORS(32, VMCS, vmcs); TDX_BUILD_TDVPS_ACCESSORS(64, VMCS, vmcs); TDX_BUILD_TDVPS_ACCESSORS(8, MANAGEMENT, management); TDX_BUILD_TDVPS_ACCESSORS(64, STATE_NON_ARCH, state_non_arch); #else static inline int tdx_bringup(void) { return 0; } static inline void tdx_cleanup(void) {} #define enable_tdx 0 struct kvm_tdx { struct kvm kvm; }; struct vcpu_tdx { struct kvm_vcpu vcpu; }; static inline bool tdx_interrupt_allowed(struct kvm_vcpu *vcpu) { return false; } static inline int tdx_complete_emulated_msr(struct kvm_vcpu *vcpu, int err) { return 0; } #endif #endif