On Mon, Mar 24, 2025, Manali Shukla wrote:
- if (vmcb02->save.rip && (svm->nested.ctl.bus_lock_rip == vmcb02->save.rip)) {
vmcb02->control.bus_lock_counter = 1;
svm->bus_lock_rip = svm->nested.ctl.bus_lock_rip;
- } else {
vmcb02->control.bus_lock_counter = 0;
- }
- svm->nested.ctl.bus_lock_rip = INVALID_GPA;
- /* Done at vmrun: asid. */
/* Also overwritten later if necessary. */ @@ -1039,6 +1069,18 @@ int nested_svm_vmexit(struct vcpu_svm *svm) }
- /*
* If bus_lock_counter is nonzero and the guest has not moved past the
* guilty instruction, save bus_lock_rip in svm_nested_state. This will
* help determine at nested VMRUN whether to stash vmcb02's counter or
* reset it to '0'.
*/
- if (vmcb02->control.bus_lock_counter &&
svm->bus_lock_rip == vmcb02->save.rip)
svm->nested.ctl.bus_lock_rip = svm->bus_lock_rip;
- else
svm->nested.ctl.bus_lock_rip = INVALID_GPA;
- nested_svm_copy_common_state(svm->nested.vmcb02.ptr, svm->vmcb01.ptr);
svm_switch_vmcb(svm, &svm->vmcb01);
...
+static int bus_lock_exit(struct kvm_vcpu *vcpu) +{
- struct vcpu_svm *svm = to_svm(vcpu);
- vcpu->run->exit_reason = KVM_EXIT_X86_BUS_LOCK;
- vcpu->run->flags |= KVM_RUN_X86_BUS_LOCK;
- vcpu->arch.cui_linear_rip = kvm_get_linear_rip(vcpu);
- svm->bus_lock_rip = vcpu->arch.cui_linear_rip;
- vcpu->arch.complete_userspace_io = complete_userspace_buslock;
- return 0;
+}
@@ -327,6 +328,7 @@ struct vcpu_svm { /* Guest GIF value, used when vGIF is not enabled */ bool guest_gif;
- u64 bus_lock_rip;
I don't think this field is necessary. Rather than unconditionally invalidate on nested VMRUN and then conditionally restore on nested #VMEXIT, just leave svm->nested.ctl.bus_lock_rip set on VMRUN and conditionally invalidate on #VMEXIT. And then in bus_lock_exit(), update the field if the exit occurred while L2 is active.
Completely untested:
diff --git a/arch/x86/kvm/svm/nested.c b/arch/x86/kvm/svm/nested.c index a42ef7dd9143..98e065a93516 100644 --- a/arch/x86/kvm/svm/nested.c +++ b/arch/x86/kvm/svm/nested.c @@ -700,13 +700,10 @@ static void nested_vmcb02_prepare_control(struct vcpu_svm *svm, * L1 re-enters L2, the same instruction will trigger a VM-Exit and the * entire cycle start over. */ - if (vmcb02->save.rip && (svm->nested.ctl.bus_lock_rip == vmcb02->save.rip)) { + if (vmcb02->save.rip && (svm->nested.ctl.bus_lock_rip == vmcb02->save.rip)) vmcb02->control.bus_lock_counter = 1; - svm->bus_lock_rip = svm->nested.ctl.bus_lock_rip; - } else { + else vmcb02->control.bus_lock_counter = 0; - } - svm->nested.ctl.bus_lock_rip = INVALID_GPA;
/* Done at vmrun: asid. */
@@ -1070,15 +1067,10 @@ int nested_svm_vmexit(struct vcpu_svm *svm) }
/* - * If bus_lock_counter is nonzero and the guest has not moved past the - * guilty instruction, save bus_lock_rip in svm_nested_state. This will - * help determine at nested VMRUN whether to stash vmcb02's counter or - * reset it to '0'. + * Invalidate bus_lock_rip unless kVM is still waiting for the guest + * to make forward progress before re-enabling bus lock detection. */ - if (vmcb02->control.bus_lock_counter && - svm->bus_lock_rip == vmcb02->save.rip) - svm->nested.ctl.bus_lock_rip = svm->bus_lock_rip; - else + if (!vmcb02->control.bus_lock_counter) svm->nested.ctl.bus_lock_rip = INVALID_GPA;
nested_svm_copy_common_state(svm->nested.vmcb02.ptr, svm->vmcb01.ptr); diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c index ea12e93ae983..11ce031323fd 100644 --- a/arch/x86/kvm/svm/svm.c +++ b/arch/x86/kvm/svm/svm.c @@ -3333,9 +3333,10 @@ static int bus_lock_exit(struct kvm_vcpu *vcpu) vcpu->run->flags |= KVM_RUN_X86_BUS_LOCK;
vcpu->arch.cui_linear_rip = kvm_get_linear_rip(vcpu); - svm->bus_lock_rip = vcpu->arch.cui_linear_rip; vcpu->arch.complete_userspace_io = complete_userspace_buslock;
+ if (is_guest_mode(vcpu)) + svm->nested.ctl.bus_lock_rip = vcpu->arch.cui_linear_rip; return 0; }
diff --git a/arch/x86/kvm/svm/svm.h b/arch/x86/kvm/svm/svm.h index 7a4c5848c952..8667faccaedc 100644 --- a/arch/x86/kvm/svm/svm.h +++ b/arch/x86/kvm/svm/svm.h @@ -328,7 +328,6 @@ struct vcpu_svm {
/* Guest GIF value, used when vGIF is not enabled */ bool guest_gif; - u64 bus_lock_rip; };
struct svm_cpu_data {