4 #include <linux/arm-smccc.h>
5 #include <linux/kvm_host.h>
7 #include <asm/kvm_emulate.h>
12 #define KVM_ARM_SMCCC_STD_FEATURES \
13 GENMASK(KVM_REG_ARM_STD_BMAP_BIT_COUNT - 1, 0)
14 #define KVM_ARM_SMCCC_STD_HYP_FEATURES \
15 GENMASK(KVM_REG_ARM_STD_HYP_BMAP_BIT_COUNT - 1, 0)
16 #define KVM_ARM_SMCCC_VENDOR_HYP_FEATURES \
17 GENMASK(KVM_REG_ARM_VENDOR_HYP_BMAP_BIT_COUNT - 1, 0)
21 struct system_time_snapshot systime_snapshot;
29 ktime_get_snapshot(&systime_snapshot);
36 if (systime_snapshot.cs_id != CSID_ARM_ARCH_COUNTER)
46 case KVM_PTP_VIRT_COUNTER:
47 cycles = systime_snapshot.cycles - vcpu->kvm->arch.timer_data.voffset;
49 case KVM_PTP_PHYS_COUNTER:
50 cycles = systime_snapshot.cycles - vcpu->kvm->arch.timer_data.poffset;
62 val[0] = upper_32_bits(systime_snapshot.real);
63 val[1] = lower_32_bits(systime_snapshot.real);
64 val[2] = upper_32_bits(cycles);
65 val[3] = lower_32_bits(cycles);
76 case ARM_SMCCC_VERSION_FUNC_ID:
77 case ARM_SMCCC_ARCH_FEATURES_FUNC_ID:
81 if (ARM_SMCCC_OWNER_NUM(func_id) == ARM_SMCCC_OWNER_STANDARD &&
82 ARM_SMCCC_FUNC_NUM(func_id) <= 0x1f)
89 if (func_id >= KVM_PSCI_FN(0) && func_id <= KVM_PSCI_FN(3))
98 struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
101 case ARM_SMCCC_TRNG_VERSION:
102 case ARM_SMCCC_TRNG_FEATURES:
103 case ARM_SMCCC_TRNG_GET_UUID:
104 case ARM_SMCCC_TRNG_RND32:
105 case ARM_SMCCC_TRNG_RND64:
106 return test_bit(KVM_REG_ARM_STD_BIT_TRNG_V1_0,
107 &smccc_feat->std_bmap);
108 case ARM_SMCCC_HV_PV_TIME_FEATURES:
109 case ARM_SMCCC_HV_PV_TIME_ST:
110 return test_bit(KVM_REG_ARM_STD_HYP_BIT_PV_TIME,
111 &smccc_feat->std_hyp_bmap);
112 case ARM_SMCCC_VENDOR_HYP_KVM_FEATURES_FUNC_ID:
113 case ARM_SMCCC_VENDOR_HYP_CALL_UID_FUNC_ID:
114 return test_bit(KVM_REG_ARM_VENDOR_HYP_BIT_FUNC_FEAT,
115 &smccc_feat->vendor_hyp_bmap);
116 case ARM_SMCCC_VENDOR_HYP_KVM_PTP_FUNC_ID:
117 return test_bit(KVM_REG_ARM_VENDOR_HYP_BIT_PTP,
118 &smccc_feat->vendor_hyp_bmap);
124 #define SMC32_ARCH_RANGE_BEGIN ARM_SMCCC_VERSION_FUNC_ID
125 #define SMC32_ARCH_RANGE_END ARM_SMCCC_CALL_VAL(ARM_SMCCC_FAST_CALL, \
127 0, ARM_SMCCC_FUNC_MASK)
129 #define SMC64_ARCH_RANGE_BEGIN ARM_SMCCC_CALL_VAL(ARM_SMCCC_FAST_CALL, \
132 #define SMC64_ARCH_RANGE_END ARM_SMCCC_CALL_VAL(ARM_SMCCC_FAST_CALL, \
134 0, ARM_SMCCC_FUNC_MASK)
145 r = mtree_insert_range(&kvm->arch.smccc_filter,
147 xa_mk_value(KVM_SMCCC_FILTER_HANDLE),
152 r = mtree_insert_range(&kvm->arch.smccc_filter,
154 xa_mk_value(KVM_SMCCC_FILTER_HANDLE),
161 mtree_destroy(&kvm->arch.smccc_filter);
167 return !mtree_empty(&kvm->arch.smccc_filter);
172 const void *zero_page = page_to_virt(ZERO_PAGE(0));
173 struct kvm_smccc_filter filter;
177 if (copy_from_user(&filter, uaddr,
sizeof(filter)))
180 if (memcmp(filter.pad, zero_page,
sizeof(filter.pad)))
184 end = start + filter.nr_functions - 1;
186 if (end < start || filter.action >= NR_SMCCC_FILTER_ACTIONS)
189 mutex_lock(&kvm->arch.config_lock);
191 if (kvm_vm_has_ran_once(kvm)) {
202 r = mtree_insert_range(&kvm->arch.smccc_filter, start,
end,
203 xa_mk_value(filter.action), GFP_KERNEL_ACCOUNT);
205 mutex_unlock(&kvm->arch.config_lock);
211 unsigned long idx = func_id;
215 return KVM_SMCCC_FILTER_HANDLE;
223 val = mt_find(&kvm->arch.smccc_filter, &idx, idx);
224 return xa_to_value(val);
234 if (action != KVM_SMCCC_FILTER_HANDLE)
239 return KVM_SMCCC_FILTER_HANDLE;
241 return KVM_SMCCC_FILTER_DENY;
246 u8 ec = ESR_ELx_EC(kvm_vcpu_get_esr(vcpu));
247 struct kvm_run *run = vcpu->run;
250 if (ec == ESR_ELx_EC_SMC32 || ec == ESR_ELx_EC_SMC64)
251 flags |= KVM_HYPERCALL_EXIT_SMC;
253 if (!kvm_vcpu_trap_il_is32bit(vcpu))
254 flags |= KVM_HYPERCALL_EXIT_16BIT;
256 run->exit_reason = KVM_EXIT_HYPERCALL;
257 run->hypercall = (typeof(run->hypercall)) {
265 struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
267 u64 val[4] = {SMCCC_RET_NOT_SUPPORTED};
274 case KVM_SMCCC_FILTER_HANDLE:
276 case KVM_SMCCC_FILTER_DENY:
278 case KVM_SMCCC_FILTER_FWD_TO_USER:
282 WARN_RATELIMIT(1,
"Unhandled SMCCC filter action: %d\n", action);
287 case ARM_SMCCC_VERSION_FUNC_ID:
288 val[0] = ARM_SMCCC_VERSION_1_1;
290 case ARM_SMCCC_ARCH_FEATURES_FUNC_ID:
293 case ARM_SMCCC_ARCH_WORKAROUND_1:
294 switch (arm64_get_spectre_v2_state()) {
295 case SPECTRE_VULNERABLE:
297 case SPECTRE_MITIGATED:
298 val[0] = SMCCC_RET_SUCCESS;
300 case SPECTRE_UNAFFECTED:
301 val[0] = SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED;
305 case ARM_SMCCC_ARCH_WORKAROUND_2:
306 switch (arm64_get_spectre_v4_state()) {
307 case SPECTRE_VULNERABLE:
309 case SPECTRE_MITIGATED:
320 if (cpus_have_final_cap(ARM64_SSBS))
323 case SPECTRE_UNAFFECTED:
324 val[0] = SMCCC_RET_NOT_REQUIRED;
328 case ARM_SMCCC_ARCH_WORKAROUND_3:
329 switch (arm64_get_spectre_bhb_state()) {
330 case SPECTRE_VULNERABLE:
332 case SPECTRE_MITIGATED:
333 val[0] = SMCCC_RET_SUCCESS;
335 case SPECTRE_UNAFFECTED:
336 val[0] = SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED;
340 case ARM_SMCCC_HV_PV_TIME_FEATURES:
341 if (test_bit(KVM_REG_ARM_STD_HYP_BIT_PV_TIME,
342 &smccc_feat->std_hyp_bmap))
343 val[0] = SMCCC_RET_SUCCESS;
347 case ARM_SMCCC_HV_PV_TIME_FEATURES:
350 case ARM_SMCCC_HV_PV_TIME_ST:
352 if (gpa != INVALID_GPA)
355 case ARM_SMCCC_VENDOR_HYP_CALL_UID_FUNC_ID:
356 val[0] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_0;
357 val[1] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_1;
358 val[2] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_2;
359 val[3] = ARM_SMCCC_VENDOR_HYP_UID_KVM_REG_3;
361 case ARM_SMCCC_VENDOR_HYP_KVM_FEATURES_FUNC_ID:
362 val[0] = smccc_feat->vendor_hyp_bmap;
364 case ARM_SMCCC_VENDOR_HYP_KVM_PTP_FUNC_ID:
367 case ARM_SMCCC_TRNG_VERSION:
368 case ARM_SMCCC_TRNG_FEATURES:
369 case ARM_SMCCC_TRNG_GET_UUID:
370 case ARM_SMCCC_TRNG_RND32:
371 case ARM_SMCCC_TRNG_RND64:
383 KVM_REG_ARM_PSCI_VERSION,
384 KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1,
385 KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2,
386 KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3,
387 KVM_REG_ARM_STD_BMAP,
388 KVM_REG_ARM_STD_HYP_BMAP,
389 KVM_REG_ARM_VENDOR_HYP_BMAP,
394 struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
400 mt_init(&kvm->arch.smccc_filter);
405 mtree_destroy(&kvm->arch.smccc_filter);
425 #define KVM_REG_FEATURE_LEVEL_MASK GENMASK(3, 0)
434 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
435 switch (arm64_get_spectre_v2_state()) {
436 case SPECTRE_VULNERABLE:
437 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_AVAIL;
438 case SPECTRE_MITIGATED:
439 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_AVAIL;
440 case SPECTRE_UNAFFECTED:
441 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_REQUIRED;
443 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1_NOT_AVAIL;
444 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
445 switch (arm64_get_spectre_v4_state()) {
446 case SPECTRE_MITIGATED:
452 if (cpus_have_final_cap(ARM64_SSBS))
453 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
455 case SPECTRE_UNAFFECTED:
456 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED;
457 case SPECTRE_VULNERABLE:
458 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
461 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
462 switch (arm64_get_spectre_bhb_state()) {
463 case SPECTRE_VULNERABLE:
464 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_AVAIL;
465 case SPECTRE_MITIGATED:
466 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_AVAIL;
467 case SPECTRE_UNAFFECTED:
468 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_REQUIRED;
470 return KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3_NOT_AVAIL;
478 struct kvm_smccc_features *smccc_feat = &vcpu->kvm->arch.smccc_feat;
479 void __user *uaddr = (
void __user *)(
long)reg->addr;
483 case KVM_REG_ARM_PSCI_VERSION:
486 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
487 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
488 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
491 case KVM_REG_ARM_STD_BMAP:
492 val = READ_ONCE(smccc_feat->std_bmap);
494 case KVM_REG_ARM_STD_HYP_BMAP:
495 val = READ_ONCE(smccc_feat->std_hyp_bmap);
497 case KVM_REG_ARM_VENDOR_HYP_BMAP:
498 val = READ_ONCE(smccc_feat->vendor_hyp_bmap);
504 if (copy_to_user(uaddr, &val, KVM_REG_SIZE(reg->id)))
513 struct kvm *kvm = vcpu->kvm;
514 struct kvm_smccc_features *smccc_feat = &kvm->arch.smccc_feat;
515 unsigned long *fw_reg_bmap, fw_reg_features;
518 case KVM_REG_ARM_STD_BMAP:
519 fw_reg_bmap = &smccc_feat->std_bmap;
522 case KVM_REG_ARM_STD_HYP_BMAP:
523 fw_reg_bmap = &smccc_feat->std_hyp_bmap;
526 case KVM_REG_ARM_VENDOR_HYP_BMAP:
527 fw_reg_bmap = &smccc_feat->vendor_hyp_bmap;
535 if (val & ~fw_reg_features)
538 mutex_lock(&kvm->arch.config_lock);
540 if (kvm_vm_has_ran_once(kvm) && val != *fw_reg_bmap) {
545 WRITE_ONCE(*fw_reg_bmap, val);
547 mutex_unlock(&kvm->arch.config_lock);
553 void __user *uaddr = (
void __user *)(
long)reg->addr;
557 if (KVM_REG_SIZE(reg->id) !=
sizeof(val))
559 if (copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id)))
563 case KVM_REG_ARM_PSCI_VERSION:
567 wants_02 = vcpu_has_feature(vcpu, KVM_ARM_VCPU_PSCI_0_2);
573 vcpu->kvm->arch.psci_version = val;
580 vcpu->kvm->arch.psci_version = val;
586 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1:
587 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_3:
596 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2:
598 KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED))
602 if ((val & KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED) &&
611 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL:
612 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_UNKNOWN:
613 wa_level = KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_AVAIL;
615 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_AVAIL:
616 case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED:
617 wa_level = KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED;
631 case KVM_REG_ARM_STD_BMAP:
632 case KVM_REG_ARM_STD_HYP_BMAP:
633 case KVM_REG_ARM_VENDOR_HYP_BMAP:
644 switch (attr->attr) {
645 case KVM_ARM_VM_SMCCC_FILTER:
654 void __user *uaddr = (
void __user *)attr->addr;
656 switch (attr->attr) {
657 case KVM_ARM_VM_SMCCC_FILTER:
static u32 smccc_get_function(struct kvm_vcpu *vcpu)
static unsigned long smccc_get_arg1(struct kvm_vcpu *vcpu)
static void smccc_set_retval(struct kvm_vcpu *vcpu, unsigned long a0, unsigned long a1, unsigned long a2, unsigned long a3)
static int kvm_psci_version(struct kvm_vcpu *vcpu)
int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
#define SMC64_ARCH_RANGE_END
static const u64 kvm_arm_fw_reg_ids[]
#define SMC32_ARCH_RANGE_BEGIN
static u8 kvm_smccc_get_action(struct kvm_vcpu *vcpu, u32 func_id)
void kvm_arm_teardown_hypercalls(struct kvm *kvm)
static int kvm_arm_set_fw_reg_bmap(struct kvm_vcpu *vcpu, u64 reg_id, u64 val)
void kvm_arm_init_hypercalls(struct kvm *kvm)
#define KVM_ARM_SMCCC_STD_HYP_FEATURES
int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
static int get_kernel_wa_level(u64 regid)
int kvm_smccc_call_handler(struct kvm_vcpu *vcpu)
int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu)
static void kvm_ptp_get_time(struct kvm_vcpu *vcpu, u64 *val)
#define SMC64_ARCH_RANGE_BEGIN
static bool kvm_smccc_filter_configured(struct kvm *kvm)
int kvm_vm_smccc_set_attr(struct kvm *kvm, struct kvm_device_attr *attr)
#define KVM_REG_FEATURE_LEVEL_MASK
static bool kvm_smccc_test_fw_bmap(struct kvm_vcpu *vcpu, u32 func_id)
static int kvm_smccc_filter_insert_reserved(struct kvm *kvm)
int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
static bool kvm_smccc_default_allowed(u32 func_id)
#define KVM_ARM_SMCCC_VENDOR_HYP_FEATURES
#define KVM_ARM_SMCCC_STD_FEATURES
int kvm_vm_smccc_has_attr(struct kvm *kvm, struct kvm_device_attr *attr)
static void kvm_prepare_hypercall_exit(struct kvm_vcpu *vcpu, u32 func_id)
#define SMC32_ARCH_RANGE_END
static int kvm_smccc_set_filter(struct kvm *kvm, struct kvm_smccc_filter __user *uaddr)
static u8 kvm_smccc_filter_get_action(struct kvm *kvm, u32 func_id)
int kvm_psci_call(struct kvm_vcpu *vcpu)
long kvm_hypercall_pv_features(struct kvm_vcpu *vcpu)
gpa_t kvm_init_stolen_time(struct kvm_vcpu *vcpu)
int kvm_trng_call(struct kvm_vcpu *vcpu)