Re: [PATCH 3/6] kvm/x86/mmu: don't unnecessarily recalculate table_gfn in *fetch

From: Joerg Roedel
Date: Thu Mar 05 2009 - 09:36:49 EST


On Thu, Mar 05, 2009 at 01:12:30PM +0100, Joerg Roedel wrote:
> Signed-off-by: Joerg Roedel <joerg.roedel@xxxxxxx>
> ---
> arch/x86/kvm/paging_tmpl.h | 2 +-
> 1 files changed, 1 insertions(+), 1 deletions(-)
>
> diff --git a/arch/x86/kvm/paging_tmpl.h b/arch/x86/kvm/paging_tmpl.h
> index a0c11ea..79668ba 100644
> --- a/arch/x86/kvm/paging_tmpl.h
> +++ b/arch/x86/kvm/paging_tmpl.h
> @@ -315,7 +315,7 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
> direct = 1;
> if (!is_dirty_pte(gw->ptes[level - 1]))
> access &= ~ACC_WRITE_MASK;
> - table_gfn = gpte_to_gfn(gw->ptes[level - 1]);
> + table_gfn = gw->table_gfn[level - delta];
> } else {
> direct = 0;
> table_gfn = gw->table_gfn[level - 2];

Made a mistake here when rebasing these patches. Updated patch is below: