Lines Matching defs:gfn

277 int kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages)  in kvm_arch_flush_remote_tlbs_range()
291 gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep)); in kvm_flush_remote_tlbs_sptep() local
296 static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, in mark_mmio_spte()
759 gfn_t gfn, unsigned int access) in kvm_mmu_page_set_translation()
780 gfn_t gfn = kvm_mmu_page_get_gfn(sp, index); in kvm_mmu_page_set_access() local
789 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot()
799 gfn_t gfn, int count) in update_gfn_disallow_lpage_count()
811 void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage()
816 void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage()
825 gfn_t gfn; in account_shadowed() local
873 gfn_t gfn; in unaccount_shadowed() local
902 gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
1080 static struct kvm_rmap_head *gfn_to_rmap(gfn_t gfn, int level, in gfn_to_rmap()
1094 gfn_t gfn; in rmap_remove() local
1409 struct kvm_memory_slot *slot, u64 gfn, in kvm_mmu_slot_gfn_write_protect()
1430 static bool kvm_vcpu_write_protect_gfn(struct kvm_vcpu *vcpu, u64 gfn) in kvm_vcpu_write_protect_gfn()
1445 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_zap_rmap()
1452 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_set_pte_rmap()
1497 gfn_t gfn; member
1607 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_age_rmap()
1621 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_test_age_rmap()
1638 u64 *spte, gfn_t gfn, unsigned int access) in __rmap_add()
1660 u64 *spte, gfn_t gfn, unsigned int access) in rmap_add()
1743 static unsigned kvm_page_table_hashfn(gfn_t gfn) in kvm_page_table_hashfn()
2147 gfn_t gfn, in kvm_mmu_find_shadow_page()
2232 gfn_t gfn, in kvm_mmu_alloc_shadow_page()
2269 gfn_t gfn, in __kvm_mmu_get_shadow_page()
2289 gfn_t gfn, in kvm_mmu_get_shadow_page()
2348 u64 *sptep, gfn_t gfn, in kvm_mmu_get_child_sp()
2751 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) in kvm_mmu_unprotect_page()
2800 gfn_t gfn, bool can_unsync, bool prefetch) in mmu_try_to_unsync_pages()
2901 u64 *sptep, unsigned int pte_access, gfn_t gfn, in mmu_set_spte()
2979 gfn_t gfn; in direct_pte_prefetch_many() local
3076 static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn, in host_pfn_mapping_level()
3141 const struct kvm_memory_slot *slot, gfn_t gfn, in kvm_mmu_max_mapping_level()
3261 static void kvm_send_hwpoison_signal(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_send_hwpoison_signal()
3642 static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant, in mmu_alloc_root()
4159 gfn_t gfn = get_mmio_spte_gfn(spte); in handle_mmio_page_fault() local
4222 gfn_t gfn) in kvm_arch_setup_async_pf()
4665 static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
5650 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mmu_track_write() local
6347 gfn_t gfn; in shadow_mmu_get_sp_for_split() local
6378 gfn_t gfn; in shadow_mmu_split_huge_page() local
6420 gfn_t gfn; in shadow_mmu_try_split_huge_page() local