Re: [PATCH v8 4/6] LoongArch: KVM: Add vcpu search support from physical cpuid

From: Huacai Chen
Date: Wed May 08 2024 - 01:01:07 EST


On Tue, May 7, 2024 at 11:06 AM maobibo <maobibo@xxxxxxxxxxx> wrote:
>
>
>
> On 2024/5/7 上午10:05, Huacai Chen wrote:
> > On Tue, May 7, 2024 at 9:40 AM maobibo <maobibo@xxxxxxxxxxx> wrote:
> >>
> >>
> >>
> >> On 2024/5/6 下午10:17, Huacai Chen wrote:
> >>> On Mon, May 6, 2024 at 6:05 PM maobibo <maobibo@xxxxxxxxxxx> wrote:
> >>>>
> >>>>
> >>>>
> >>>> On 2024/5/6 下午5:40, Huacai Chen wrote:
> >>>>> On Mon, May 6, 2024 at 5:35 PM maobibo <maobibo@xxxxxxxxxxx> wrote:
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>> On 2024/5/6 下午4:59, Huacai Chen wrote:
> >>>>>>> On Mon, May 6, 2024 at 4:18 PM maobibo <maobibo@xxxxxxxxxxx> wrote:
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> On 2024/5/6 下午3:06, Huacai Chen wrote:
> >>>>>>>>> Hi, Bibo,
> >>>>>>>>>
> >>>>>>>>> On Mon, May 6, 2024 at 2:36 PM maobibo <maobibo@xxxxxxxxxxx> wrote:
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> On 2024/5/6 上午9:49, Huacai Chen wrote:
> >>>>>>>>>>> Hi, Bibo,
> >>>>>>>>>>>
> >>>>>>>>>>> On Sun, Apr 28, 2024 at 6:05 PM Bibo Mao <maobibo@xxxxxxxxxxx> wrote:
> >>>>>>>>>>>>
> >>>>>>>>>>>> Physical cpuid is used for interrupt routing for irqchips such as
> >>>>>>>>>>>> ipi/msi/extioi interrupt controller. And physical cpuid is stored
> >>>>>>>>>>>> at CSR register LOONGARCH_CSR_CPUID, it can not be changed once vcpu
> >>>>>>>>>>>> is created and physical cpuid of two vcpus cannot be the same.
> >>>>>>>>>>>>
> >>>>>>>>>>>> Different irqchips have different size declaration about physical cpuid,
> >>>>>>>>>>>> max cpuid value for CSR LOONGARCH_CSR_CPUID on 3A5000 is 512, max cpuid
> >>>>>>>>>>>> supported by IPI hardware is 1024, 256 for extioi irqchip, and 65536
> >>>>>>>>>>>> for MSI irqchip.
> >>>>>>>>>>>>
> >>>>>>>>>>>> The smallest value from all interrupt controllers is selected now,
> >>>>>>>>>>>> and the max cpuid size is defines as 256 by KVM which comes from
> >>>>>>>>>>>> extioi irqchip.
> >>>>>>>>>>>>
> >>>>>>>>>>>> Signed-off-by: Bibo Mao <maobibo@xxxxxxxxxxx>
> >>>>>>>>>>>> ---
> >>>>>>>>>>>> arch/loongarch/include/asm/kvm_host.h | 26 ++++++++
> >>>>>>>>>>>> arch/loongarch/include/asm/kvm_vcpu.h | 1 +
> >>>>>>>>>>>> arch/loongarch/kvm/vcpu.c | 93 ++++++++++++++++++++++++++-
> >>>>>>>>>>>> arch/loongarch/kvm/vm.c | 11 ++++
> >>>>>>>>>>>> 4 files changed, 130 insertions(+), 1 deletion(-)
> >>>>>>>>>>>>
> >>>>>>>>>>>> diff --git a/arch/loongarch/include/asm/kvm_host.h b/arch/loongarch/include/asm/kvm_host.h
> >>>>>>>>>>>> index 2d62f7b0d377..3ba16ef1fe69 100644
> >>>>>>>>>>>> --- a/arch/loongarch/include/asm/kvm_host.h
> >>>>>>>>>>>> +++ b/arch/loongarch/include/asm/kvm_host.h
> >>>>>>>>>>>> @@ -64,6 +64,30 @@ struct kvm_world_switch {
> >>>>>>>>>>>>
> >>>>>>>>>>>> #define MAX_PGTABLE_LEVELS 4
> >>>>>>>>>>>>
> >>>>>>>>>>>> +/*
> >>>>>>>>>>>> + * Physical cpu id is used for interrupt routing, there are different
> >>>>>>>>>>>> + * definitions about physical cpuid on different hardwares.
> >>>>>>>>>>>> + * For LOONGARCH_CSR_CPUID register, max cpuid size if 512
> >>>>>>>>>>>> + * For IPI HW, max dest CPUID size 1024
> >>>>>>>>>>>> + * For extioi interrupt controller, max dest CPUID size is 256
> >>>>>>>>>>>> + * For MSI interrupt controller, max supported CPUID size is 65536
> >>>>>>>>>>>> + *
> >>>>>>>>>>>> + * Currently max CPUID is defined as 256 for KVM hypervisor, in future
> >>>>>>>>>>>> + * it will be expanded to 4096, including 16 packages at most. And every
> >>>>>>>>>>>> + * package supports at most 256 vcpus
> >>>>>>>>>>>> + */
> >>>>>>>>>>>> +#define KVM_MAX_PHYID 256
> >>>>>>>>>>>> +
> >>>>>>>>>>>> +struct kvm_phyid_info {
> >>>>>>>>>>>> + struct kvm_vcpu *vcpu;
> >>>>>>>>>>>> + bool enabled;
> >>>>>>>>>>>> +};
> >>>>>>>>>>>> +
> >>>>>>>>>>>> +struct kvm_phyid_map {
> >>>>>>>>>>>> + int max_phyid;
> >>>>>>>>>>>> + struct kvm_phyid_info phys_map[KVM_MAX_PHYID];
> >>>>>>>>>>>> +};
> >>>>>>>>>>>> +
> >>>>>>>>>>>> struct kvm_arch {
> >>>>>>>>>>>> /* Guest physical mm */
> >>>>>>>>>>>> kvm_pte_t *pgd;
> >>>>>>>>>>>> @@ -71,6 +95,8 @@ struct kvm_arch {
> >>>>>>>>>>>> unsigned long invalid_ptes[MAX_PGTABLE_LEVELS];
> >>>>>>>>>>>> unsigned int pte_shifts[MAX_PGTABLE_LEVELS];
> >>>>>>>>>>>> unsigned int root_level;
> >>>>>>>>>>>> + spinlock_t phyid_map_lock;
> >>>>>>>>>>>> + struct kvm_phyid_map *phyid_map;
> >>>>>>>>>>>>
> >>>>>>>>>>>> s64 time_offset;
> >>>>>>>>>>>> struct kvm_context __percpu *vmcs;
> >>>>>>>>>>>> diff --git a/arch/loongarch/include/asm/kvm_vcpu.h b/arch/loongarch/include/asm/kvm_vcpu.h
> >>>>>>>>>>>> index 0cb4fdb8a9b5..9f53950959da 100644
> >>>>>>>>>>>> --- a/arch/loongarch/include/asm/kvm_vcpu.h
> >>>>>>>>>>>> +++ b/arch/loongarch/include/asm/kvm_vcpu.h
> >>>>>>>>>>>> @@ -81,6 +81,7 @@ void kvm_save_timer(struct kvm_vcpu *vcpu);
> >>>>>>>>>>>> void kvm_restore_timer(struct kvm_vcpu *vcpu);
> >>>>>>>>>>>>
> >>>>>>>>>>>> int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq);
> >>>>>>>>>>>> +struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid);
> >>>>>>>>>>>>
> >>>>>>>>>>>> /*
> >>>>>>>>>>>> * Loongarch KVM guest interrupt handling
> >>>>>>>>>>>> diff --git a/arch/loongarch/kvm/vcpu.c b/arch/loongarch/kvm/vcpu.c
> >>>>>>>>>>>> index 3a8779065f73..b633fd28b8db 100644
> >>>>>>>>>>>> --- a/arch/loongarch/kvm/vcpu.c
> >>>>>>>>>>>> +++ b/arch/loongarch/kvm/vcpu.c
> >>>>>>>>>>>> @@ -274,6 +274,95 @@ static int _kvm_getcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 *val)
> >>>>>>>>>>>> return 0;
> >>>>>>>>>>>> }
> >>>>>>>>>>>>
> >>>>>>>>>>>> +static inline int kvm_set_cpuid(struct kvm_vcpu *vcpu, u64 val)
> >>>>>>>>>>>> +{
> >>>>>>>>>>>> + int cpuid;
> >>>>>>>>>>>> + struct loongarch_csrs *csr = vcpu->arch.csr;
> >>>>>>>>>>>> + struct kvm_phyid_map *map;
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + if (val >= KVM_MAX_PHYID)
> >>>>>>>>>>>> + return -EINVAL;
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_ESTAT);
> >>>>>>>>>>>> + map = vcpu->kvm->arch.phyid_map;
> >>>>>>>>>>>> + spin_lock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>>>>>>>>> + if (map->phys_map[cpuid].enabled) {
> >>>>>>>>>>>> + /*
> >>>>>>>>>>>> + * Cpuid is already set before
> >>>>>>>>>>>> + * Forbid changing different cpuid at runtime
> >>>>>>>>>>>> + */
> >>>>>>>>>>>> + if (cpuid != val) {
> >>>>>>>>>>>> + /*
> >>>>>>>>>>>> + * Cpuid 0 is initial value for vcpu, maybe invalid
> >>>>>>>>>>>> + * unset value for vcpu
> >>>>>>>>>>>> + */
> >>>>>>>>>>>> + if (cpuid) {
> >>>>>>>>>>>> + spin_unlock(&vcpu->kvm->archphyid_map_lock);
> >>>>>>>>>>>> + return -EINVAL;
> >>>>>>>>>>>> + }
> >>>>>>>>>>>> + } else {
> >>>>>>>>>>>> + /* Discard duplicated cpuid set */
> >>>>>>>>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>>>>>>>>> + return 0;
> >>>>>>>>>>>> + }
> >>>>>>>>>>>> + }
> >>>>>>>>>>> I have changed the logic and comments when I apply, you can double
> >>>>>>>>>>> check whether it is correct.
> >>>>>>>>>> I checkout the latest version, the modification in function
> >>>>>>>>>> kvm_set_cpuid() is good for me.
> >>>>>>>>> Now the modified version is like this:
> >>>>>>>>>
> >>>>>>>>> + if (map->phys_map[cpuid].enabled) {
> >>>>>>>>> + /* Discard duplicated CPUID set operation */
> >>>>>>>>> + if (cpuid == val) {
> >>>>>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>>>>>> + return 0;
> >>>>>>>>> + }
> >>>>>>>>> +
> >>>>>>>>> + /*
> >>>>>>>>> + * CPUID is already set before
> >>>>>>>>> + * Forbid changing different CPUID at runtime
> >>>>>>>>> + * But CPUID 0 is the initial value for vcpu, so allow
> >>>>>>>>> + * changing from 0 to others
> >>>>>>>>> + */
> >>>>>>>>> + if (cpuid) {
> >>>>>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>>>>>> + return -EINVAL;
> >>>>>>>>> + }
> >>>>>>>>> + }
> >>>>>>>>> But I still doubt whether we should allow changing from 0 to others
> >>>>>>>>> while map->phys_map[cpuid].enabled is 1.
> >>>>>>>> It is necessary since the default sw cpuid is zero :-( And we can
> >>>>>>>> optimize it in later, such as set INVALID cpuid in function
> >>>>>>>> kvm_arch_vcpu_create() and logic will be simple in function kvm_set_cpuid().
> >>>>>>> In my opinion, if a vcpu with a uninitialized default physid=0, then
> >>>>>>> map->phys_map[cpuid].enabled should be 0, then code won't come here.
> >>>>>>> And if a vcpu with a real physid=0, then map->phys_map[cpuid].enabled
> >>>>>>> is 1, but we shouldn't allow it to change physid in this case.
> >>>>>> yes, that is actually a problem.
> >>>>>>
> >>>>>> vcpu0 firstly set physid=0, and vcpu0 set physid=1 again is not allowed.
> >>>>>> vcpu0 firstly set physid=0, and vcpu1 set physid=1 is allowed.
> >>>>>
> >>>>> So can we simply drop the if (cpuid) checking? That means:
> >>>>> + if (map->phys_map[cpuid].enabled) {
> >>>>> + /* Discard duplicated CPUID set operation */
> >>>>> + if (cpuid == val) {
> >>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>> + return 0;
> >>>>> + }
> >>>>> +
> >>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>> + return -EINVAL;
> >>>>> + }
> >>>> yes, the similar modification such as following, since the secondary
> >>>> scenario should be allowed.
> >>>> "vcpu0 firstly set physid=0, and vcpu1 set physid=1 is allowed though
> >>>> default sw cpuid is zero"
> >>>>
> >>>> --- a/arch/loongarch/kvm/vcpu.c
> >>>> +++ b/arch/loongarch/kvm/vcpu.c
> >>>> @@ -272,7 +272,7 @@ static inline int kvm_set_cpuid(struct kvm_vcpu
> >>>> *vcpu, u64 val)
> >>>> cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_CPUID);
> >>>>
> >>>> spin_lock(&vcpu->kvm->arch.phyid_map_lock);
> >>>> - if (map->phys_map[cpuid].enabled) {
> >>>> + if ((cpuid != KVM_MAX_PHYID) && map->phys_map[cpuid].enabled) {
> >>>> /* Discard duplicated CPUID set operation */
> >>>> if (cpuid == val) {
> >>>> spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>> @@ -282,13 +282,9 @@ static inline int kvm_set_cpuid(struct kvm_vcpu
> >>>> *vcpu, u64 val)
> >>>> /*
> >>>> * CPUID is already set before
> >>>> * Forbid changing different CPUID at runtime
> >>>> - * But CPUID 0 is the initial value for vcpu, so allow
> >>>> - * changing from 0 to others
> >>>> */
> >>>> - if (cpuid) {
> >>>> - spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>> - return -EINVAL;
> >>>> - }
> >>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>> + return -EINVAL;
> >>>> }
> >>>>
> >>>> if (map->phys_map[val].enabled) {
> >>>> @@ -1029,6 +1025,7 @@ int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
> >>>>
> >>>> /* Set cpuid */
> >>>> kvm_write_sw_gcsr(csr, LOONGARCH_CSR_TMID, vcpu->vcpu_id);
> >>>> + kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, KVM_MAX_PHYID);
> >>>>
> >>>> /* Start with no pending virtual guest interrupts */
> >>>> csr->csrs[LOONGARCH_CSR_GINTC] = 0;
> >>> Very nice, but I think kvm_drop_cpuid() should also set to KVM_MAX_PHYID.
> >>> Now I update my loongarch-kvm branch, you can test it again, and hope
> >>> it is in the perfect status.
> >> I sync and test the latest code from loongarch-kvm, pv ipi works well
> >> with 256 vcpus. And the code looks good to me, thanks for your review in
> >> short time.
> > OK, if SWDBG also works well, I will send PR to Paolo tomorrow.
> yes, sw debug works well with patch from qemu. And I will refresh patch
> to qemu after it is merged.
>
> https://lore.kernel.org/all/20240218070025.218680-1-maobibo@xxxxxxxxxxx/
>
> --- a/configs/targets/loongarch64-softmmu.mak
> +++ b/configs/targets/loongarch64-softmmu.mak
> @@ -1,5 +1,6 @@
> TARGET_ARCH=loongarch64
> TARGET_BASE_ARCH=loongarch
> TARGET_SUPPORTS_MTTCG=y
> +TARGET_KVM_HAVE_GUEST_DEBUG=y
> TARGET_XML_FILES= gdb-xml/loongarch-base32.xml
> gdb-xml/loongarch-base64.xml gdb-xml/loongarch-fpu.xml
> TARGET_NEED_FDT=y
Not enough, we need kvm_arch_update_guest_debug() and some other functions.

Huacai

>
> Regards
> Bibo Mao
> >
> > Huacai
> >
> >>
> >> Regards
> >> Bibo Mao
> >>>
> >>> Huacai
> >>>>
> >>>>
> >>>>>
> >>>>> Huacai
> >>>>>
> >>>>>>
> >>>>>>
> >>>>>>>
> >>>>>>> Huacai
> >>>>>>>
> >>>>>>>>
> >>>>>>>> Regards
> >>>>>>>> Bibo Mao
> >>>>>>>>
> >>>>>>>>>
> >>>>>>>>> Huacai
> >>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + if (map->phys_map[val].enabled) {
> >>>>>>>>>>>> + /*
> >>>>>>>>>>>> + * New cpuid is already set with other vcpu
> >>>>>>>>>>>> + * Forbid sharing the same cpuid between different vcpus
> >>>>>>>>>>>> + */
> >>>>>>>>>>>> + if (map->phys_map[val].vcpu != vcpu) {
> >>>>>>>>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>>>>>>>>> + return -EINVAL;
> >>>>>>>>>>>> + }
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + /* Discard duplicated cpuid set operation*/
> >>>>>>>>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>>>>>>>>> + return 0;
> >>>>>>>>>>>> + }
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, val);
> >>>>>>>>>>>> + map->phys_map[val].enabled = true;
> >>>>>>>>>>>> + map->phys_map[val].vcpu = vcpu;
> >>>>>>>>>>>> + if (map->max_phyid < val)
> >>>>>>>>>>>> + map->max_phyid = val;
> >>>>>>>>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
> >>>>>>>>>>>> + return 0;
> >>>>>>>>>>>> +}
> >>>>>>>>>>>> +
> >>>>>>>>>>>> +struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid)
> >>>>>>>>>>>> +{
> >>>>>>>>>>>> + struct kvm_phyid_map *map;
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + if (cpuid >= KVM_MAX_PHYID)
> >>>>>>>>>>>> + return NULL;
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + map = kvm->arch.phyid_map;
> >>>>>>>>>>>> + if (map->phys_map[cpuid].enabled)
> >>>>>>>>>>>> + return map->phys_map[cpuid].vcpu;
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + return NULL;
> >>>>>>>>>>>> +}
> >>>>>>>>>>>> +
> >>>>>>>>>>>> +static inline void kvm_drop_cpuid(struct kvm_vcpu *vcpu)
> >>>>>>>>>>>> +{
> >>>>>>>>>>>> + int cpuid;
> >>>>>>>>>>>> + struct loongarch_csrs *csr = vcpu->arch.csr;
> >>>>>>>>>>>> + struct kvm_phyid_map *map;
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + map = vcpu->kvm->arch.phyid_map;
> >>>>>>>>>>>> + cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_ESTAT);
> >>>>>>>>>>>> + if (cpuid >= KVM_MAX_PHYID)
> >>>>>>>>>>>> + return;
> >>>>>>>>>>>> +
> >>>>>>>>>>>> + if (map->phys_map[cpuid].enabled) {
> >>>>>>>>>>>> + map->phys_map[cpuid].vcpu = NULL;
> >>>>>>>>>>>> + map->phys_map[cpuid].enabled = false;
> >>>>>>>>>>>> + kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, 0);
> >>>>>>>>>>>> + }
> >>>>>>>>>>>> +}
> >>>>>>>>>>> While kvm_set_cpuid() is protected by a spinlock, do kvm_drop_cpuid()
> >>>>>>>>>>> and kvm_get_vcpu_by_cpuid() also need it?
> >>>>>>>>>>>
> >>>>>>>>>> It is good to me that spinlock is added in function kvm_drop_cpuid().
> >>>>>>>>>> And thinks for the efforts.
> >>>>>>>>>>
> >>>>>>>>>> Regards
> >>>>>>>>>> Bibo Mao
> >>>>>>>>>>>> +
> >>>>>>>>>>>> static int _kvm_setcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 val)
> >>>>>>>>>>>> {
> >>>>>>>>>>>> int ret = 0, gintc;
> >>>>>>>>>>>> @@ -291,7 +380,8 @@ static int _kvm_setcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 val)
> >>>>>>>>>>>> kvm_set_sw_gcsr(csr, LOONGARCH_CSR_ESTAT, gintc);
> >>>>>>>>>>>>
> >>>>>>>>>>>> return ret;
> >>>>>>>>>>>> - }
> >>>>>>>>>>>> + } else if (id == LOONGARCH_CSR_CPUID)
> >>>>>>>>>>>> + return kvm_set_cpuid(vcpu, val);
> >>>>>>>>>>>>
> >>>>>>>>>>>> kvm_write_sw_gcsr(csr, id, val);
> >>>>>>>>>>>>
> >>>>>>>>>>>> @@ -943,6 +1033,7 @@ void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
> >>>>>>>>>>>> hrtimer_cancel(&vcpu->arch.swtimer);
> >>>>>>>>>>>> kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
> >>>>>>>>>>>> kfree(vcpu->arch.csr);
> >>>>>>>>>>>> + kvm_drop_cpuid(vcpu);
> >>>>>>>>>>> I think this line should be before the above kfree(), otherwise you
> >>>>>>>>>>> get a "use after free".
> >>>>>>>>>>>
> >>>>>>>>>>> Huacai
> >>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> /*
> >>>>>>>>>>>> * If the vCPU is freed and reused as another vCPU, we don't want the
> >>>>>>>>>>>> diff --git a/arch/loongarch/kvm/vm.c b/arch/loongarch/kvm/vmc
> >>>>>>>>>>>> index 0a37f6fa8f2d..6006a28653ad 100644
> >>>>>>>>>>>> --- a/arch/loongarch/kvm/vm.c
> >>>>>>>>>>>> +++ b/arch/loongarch/kvm/vm.c
> >>>>>>>>>>>> @@ -30,6 +30,14 @@ int kvm_arch_init_vm(struct kvm *kvm, unsigned long type)
> >>>>>>>>>>>> if (!kvm->arch.pgd)
> >>>>>>>>>>>> return -ENOMEM;
> >>>>>>>>>>>>
> >>>>>>>>>>>> + kvm->arch.phyid_map = kvzalloc(sizeof(struct kvm_phyid_map),
> >>>>>>>>>>>> + GFP_KERNEL_ACCOUNT);
> >>>>>>>>>>>> + if (!kvm->arch.phyid_map) {
> >>>>>>>>>>>> + free_page((unsigned long)kvm->arch.pgd);
> >>>>>>>>>>>> + kvm->arch.pgd = NULL;
> >>>>>>>>>>>> + return -ENOMEM;
> >>>>>>>>>>>> + }
> >>>>>>>>>>>> +
> >>>>>>>>>>>> kvm_init_vmcs(kvm);
> >>>>>>>>>>>> kvm->arch.gpa_size = BIT(cpu_vabits - 1);
> >>>>>>>>>>>> kvm->arch.root_level = CONFIG_PGTABLE_LEVELS - 1;
> >>>>>>>>>>>> @@ -44,6 +52,7 @@ int kvm_arch_init_vm(struct kvm *kvm, unsigned long type)
> >>>>>>>>>>>> for (i = 0; i <= kvm->arch.root_level; i++)
> >>>>>>>>>>>> kvm->arch.pte_shifts[i] = PAGE_SHIFT + i * (PAGE_SHIFT - 3);
> >>>>>>>>>>>>
> >>>>>>>>>>>> + spin_lock_init(&kvm->arch.phyid_map_lock);
> >>>>>>>>>>>> return 0;
> >>>>>>>>>>>> }
> >>>>>>>>>>>>
> >>>>>>>>>>>> @@ -51,7 +60,9 @@ void kvm_arch_destroy_vm(struct kvm *kvm)
> >>>>>>>>>>>> {
> >>>>>>>>>>>> kvm_destroy_vcpus(kvm);
> >>>>>>>>>>>> free_page((unsigned long)kvm->arch.pgd);
> >>>>>>>>>>>> + kvfree(kvm->arch.phyid_map);
> >>>>>>>>>>>> kvm->arch.pgd = NULL;
> >>>>>>>>>>>> + kvm->arch.phyid_map = NULL;
> >>>>>>>>>>>> }
> >>>>>>>>>>>>
> >>>>>>>>>>>> int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext)
> >>>>>>>>>>>> --
> >>>>>>>>>>>> 2.39.3
> >>>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>
> >>>>>>
> >>>>
> >>
> >>
>
>