Re: [PATCH v2 07/11] KVM: x86: add a delayed hardware NMI injection interface

From: Santosh Shukla
Date: Wed Feb 08 2023 - 04:33:30 EST


On 1/28/2023 6:39 AM, Sean Christopherson wrote:
> On Tue, Nov 29, 2022, Maxim Levitsky wrote:
>> This patch adds two new vendor callbacks:
>
> No "this patch" please, just say what it does.
>
Sure.

>> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
>> index 684a5519812fb2..46993ce61c92db 100644
>> --- a/arch/x86/include/asm/kvm_host.h
>> +++ b/arch/x86/include/asm/kvm_host.h
>> @@ -871,8 +871,13 @@ struct kvm_vcpu_arch {
>> u64 tsc_scaling_ratio; /* current scaling ratio */
>>
>> atomic_t nmi_queued; /* unprocessed asynchronous NMIs */
>> - unsigned nmi_pending; /* NMI queued after currently running handler */
>> +
>> + unsigned int nmi_pending; /*
>> + * NMI queued after currently running handler
>> + * (not including a hardware pending NMI (e.g vNMI))
>> + */
>
> Put the block comment above. I'd say collapse all of the comments about NMIs into
> a single big block comment.
>

ok.

>> bool nmi_injected; /* Trying to inject an NMI this entry */
>> +
>> bool smi_pending; /* SMI queued after currently running handler */
>> u8 handling_intr_from_guest;
>>
>> @@ -10015,13 +10022,34 @@ static void process_nmi(struct kvm_vcpu *vcpu)
>> * Otherwise, allow two (and we'll inject the first one immediately).
>> */
>> if (static_call(kvm_x86_get_nmi_mask)(vcpu) || vcpu->arch.nmi_injected)
>> - limit = 1;
>> + limit--;
>> +
>> + /* Also if there is already a NMI hardware queued to be injected,
>> + * decrease the limit again
>> + */
>
> /*
> * Block comment ...
> */
>

ok.

>> + if (static_call(kvm_x86_get_hw_nmi_pending)(vcpu))
>
> I'd prefer "is_hw_nmi_pending()" over "get", even if it means not pairing with
> "set". Though I think that's a good thing since they aren't perfect pairs.
>

Sure thing, will spin in v3.

>> + limit--;
>>
>> - vcpu->arch.nmi_pending += atomic_xchg(&vcpu->arch.nmi_queued, 0);
>> + if (limit <= 0)
>> + return;
>> +
>> + /* Attempt to use hardware NMI queueing */
>> + if (static_call(kvm_x86_set_hw_nmi_pending)(vcpu)) {
>> + limit--;
>> + nmi_to_queue--;
>> + }
>> +
>> + vcpu->arch.nmi_pending += nmi_to_queue;
>> vcpu->arch.nmi_pending = min(vcpu->arch.nmi_pending, limit);
>> kvm_make_request(KVM_REQ_EVENT, vcpu);
>> }
>>
>> +/* Return total number of NMIs pending injection to the VM */
>> +int kvm_get_total_nmi_pending(struct kvm_vcpu *vcpu)
>> +{
>> + return vcpu->arch.nmi_pending + static_call(kvm_x86_get_hw_nmi_pending)(vcpu);
>
> Nothing cares about the total count, this can just be;
>
>
> bool kvm_is_nmi_pending(struct kvm_vcpu *vcpu)
> {
> return vcpu->arch.nmi_pending ||
> static_call(kvm_x86_is_hw_nmi_pending)(vcpu);
> }
>

Yes, this simplifies things.

Thanks,
Santosh

>
>> +}
>> +
>> void kvm_make_scan_ioapic_request_mask(struct kvm *kvm,
>> unsigned long *vcpu_bitmap)
>> {
>> --
>> 2.26.3
>>