diff options
author | Xiao Guangrong <xiaoguangrong@cn.fujitsu.com> | 2011-05-15 17:20:27 +0200 |
---|---|---|
committer | Avi Kivity <avi@redhat.com> | 2011-07-12 10:45:02 +0200 |
commit | 332b207d65c1d7982489dbb83e5071c95e19eb75 (patch) | |
tree | 5a9917fc3b03b6f1c9d9e1436c89c3bc5f45b636 /arch/x86/include | |
parent | KVM: VMX: always_inline VMREADs (diff) | |
download | linux-332b207d65c1d7982489dbb83e5071c95e19eb75.tar.xz linux-332b207d65c1d7982489dbb83e5071c95e19eb75.zip |
KVM: MMU: optimize pte write path if don't have protected sp
Simply return from kvm_mmu_pte_write path if no shadow page is
write-protected, then we can avoid to walk all shadow pages and hold
mmu-lock
Signed-off-by: Xiao Guangrong <xiaoguangrong@cn.fujitsu.com>
Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>
Diffstat (limited to 'arch/x86/include')
-rw-r--r-- | arch/x86/include/asm/kvm_host.h | 1 |
1 files changed, 1 insertions, 0 deletions
diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h index db4b6543b830..387780eb97bb 100644 --- a/arch/x86/include/asm/kvm_host.h +++ b/arch/x86/include/asm/kvm_host.h @@ -441,6 +441,7 @@ struct kvm_arch { unsigned int n_used_mmu_pages; unsigned int n_requested_mmu_pages; unsigned int n_max_mmu_pages; + unsigned int indirect_shadow_pages; atomic_t invlpg_counter; struct hlist_head mmu_page_hash[KVM_NUM_MMU_PAGES]; /* |