Lines Matching refs:stat
855 ++kvm->stat.nx_lpage_splits; in track_possible_nx_huge_page()
890 --kvm->stat.nx_lpage_splits; in untrack_possible_nx_huge_page()
1651 if (rmap_count > kvm->stat.max_mmu_rmap_size) in __rmap_add()
1652 kvm->stat.max_mmu_rmap_size = rmap_count; in __rmap_add()
1883 --kvm->stat.mmu_unsync; in kvm_unlink_unsync_page()
2213 ++kvm->stat.mmu_cache_miss; in kvm_mmu_find_shadow_page()
2218 if (collisions > kvm->stat.max_mmu_page_hash_collisions) in kvm_mmu_find_shadow_page()
2219 kvm->stat.max_mmu_page_hash_collisions = collisions; in kvm_mmu_find_shadow_page()
2572 ++kvm->stat.mmu_shadow_zapped; in __kvm_mmu_prepare_zap_page()
2695 kvm->stat.mmu_recycled += total_zapped; in kvm_mmu_zap_oldest_mmu_pages()
2787 ++kvm->stat.mmu_unsync; in kvm_unsync_page()
2918 vcpu->stat.pf_mmio_spte_created++; in mmu_set_spte()
3533 vcpu->stat.pf_fast++; in fast_page_fault()
5668 ++vcpu->kvm->stat.mmu_pte_write; in kvm_mmu_track_write()
5674 ++vcpu->kvm->stat.mmu_flooded; in kvm_mmu_track_write()
5686 ++vcpu->kvm->stat.mmu_pde_zapped; in kvm_mmu_track_write()
5847 ++vcpu->stat.invlpg; in kvm_mmu_invlpg()
5869 ++vcpu->stat.invlpg; in kvm_mmu_invpcid_gva()
7012 unsigned long nx_lpage_splits = kvm->stat.nx_lpage_splits; in kvm_recover_nx_huge_pages()