Home
last modified time | relevance | path

Searched refs:base_gfn (Results 1 – 25 of 26) sorted by relevance

12

/openbmc/linux/arch/x86/kvm/
H A Dmmu.h269 static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) in gfn_to_index() argument
273 (base_gfn >> KVM_HPAGE_GFN_SHIFT(level)); in gfn_to_index()
280 return gfn_to_index(slot->base_gfn + npages - 1, in __kvm_mmu_slot_lpages()
281 slot->base_gfn, level) + 1; in __kvm_mmu_slot_lpages()
/openbmc/linux/arch/riscv/kvm/
H A Dmmu.c337 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in gstage_wp_memory_region()
338 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in gstage_wp_memory_region()
398 phys_addr_t base_gfn = slot->base_gfn + gfn_offset; in kvm_arch_mmu_enable_log_dirty_pt_masked() local
399 phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; in kvm_arch_mmu_enable_log_dirty_pt_masked()
400 phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; in kvm_arch_mmu_enable_log_dirty_pt_masked()
425 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot()
465 if ((new->base_gfn + new->npages) >= in kvm_arch_prepare_memory_region()
472 base_gpa = new->base_gfn << PAGE_SHIFT; in kvm_arch_prepare_memory_region()
/openbmc/linux/arch/x86/kvm/mmu/
H A Dpage_track.c67 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in update_gfn_write_track()
134 index = gfn_to_index(gfn, slot->base_gfn, PG_LEVEL_4K); in kvm_gfn_is_write_tracked()
244 n->track_remove_region(slot->base_gfn, slot->npages, n); in kvm_page_track_delete_slot()
H A Dtdp_mmu.c324 gfn_t base_gfn = sp->gfn; in handle_removed_pt() local
333 gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level); in handle_removed_pt()
1318 spte_set |= wrprot_gfn_range(kvm, root, slot->base_gfn, in kvm_tdp_mmu_wrprot_slot()
1319 slot->base_gfn + slot->npages, min_level); in kvm_tdp_mmu_wrprot_slot()
1576 spte_set |= clear_dirty_gfn_range(kvm, root, slot->base_gfn, in kvm_tdp_mmu_clear_dirty_slot()
1577 slot->base_gfn + slot->npages); in kvm_tdp_mmu_clear_dirty_slot()
1652 gfn_t start = slot->base_gfn; in zap_collapsible_spte_range()
H A Dpaging_tmpl.h634 gfn_t base_gfn = fault->gfn; in FNAME() local
636 WARN_ON_ONCE(gw->gfn != base_gfn); in FNAME()
731 base_gfn = gfn_round_for_level(fault->gfn, it.level); in FNAME()
737 sp = kvm_mmu_get_child_sp(vcpu, it.sptep, base_gfn, in FNAME()
752 base_gfn, fault->pfn, fault); in FNAME()
H A Dmmu.c794 idx = gfn_to_index(gfn, slot->base_gfn, level); in lpage_info_slot()
1085 idx = gfn_to_index(gfn, slot->base_gfn, level); in gfn_to_rmap()
1309 slot->base_gfn + gfn_offset, mask, true); in kvm_mmu_write_protect_pt_masked()
1342 slot->base_gfn + gfn_offset, mask, false); in kvm_mmu_clear_dirty_pt_masked()
1382 gfn_t end = slot->base_gfn + gfn_offset + __fls(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked()
3222 gfn_t base_gfn = fault->gfn; in direct_map() local
3235 base_gfn = gfn_round_for_level(fault->gfn, it.level); in direct_map()
3253 base_gfn, fault->pfn, fault); in direct_map()
5945 slot->base_gfn, slot->base_gfn + slot->npages - 1, in walk_slot_rmaps()
6221 start = max(gfn_start, memslot->base_gfn); in kvm_rmap_zap_gfn_range()
[all …]
/openbmc/linux/arch/arm64/kvm/
H A Dmmu.c339 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_flush_memslot()
945 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_unmap_memslot()
1134 start = memslot->base_gfn << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1135 end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1163 start = memslot->base_gfn << PAGE_SHIFT; in kvm_mmu_split_memory_region()
1164 end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in kvm_mmu_split_memory_region()
1186 phys_addr_t base_gfn = slot->base_gfn + gfn_offset; in kvm_arch_mmu_enable_log_dirty_pt_masked() local
1187 phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; in kvm_arch_mmu_enable_log_dirty_pt_masked()
1188 phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; in kvm_arch_mmu_enable_log_dirty_pt_masked()
1225 gpa_start = memslot->base_gfn << PAGE_SHIFT; in fault_supports_stage2_huge_mapping()
[all …]
/openbmc/linux/arch/powerpc/kvm/
H A Dtrace_hv.h305 __field(u64, base_gfn)
317 __entry->base_gfn = memslot ? memslot->base_gfn : -1UL;
325 __entry->base_gfn, __entry->slot_flags)
H A Dbook3s_hv_uvmem.c261 p->base_pfn = slot->base_gfn; in kvmppc_uvmem_slot_init()
279 if (p->base_pfn == slot->base_gfn) { in kvmppc_uvmem_slot_free()
394 unsigned long gfn = memslot->base_gfn; in kvmppc_memslot_page_merge()
450 memslot->base_gfn << PAGE_SHIFT, in __kvmppc_uvmem_memslot_create()
624 gfn = slot->base_gfn; in kvmppc_uvmem_drop_pages()
797 unsigned long gfn = memslot->base_gfn; in kvmppc_uv_migrate_mem_slot()
H A Dbook3s_64_mmu_hv.c592 if (gfn_base < memslot->base_gfn) in kvmppc_book3s_hv_page_fault()
706 rmap = &memslot->arch.rmap[gfn_base - memslot->base_gfn]; in kvmppc_book3s_hv_page_fault()
828 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_unmap_rmapp()
879 gfn = memslot->base_gfn; in kvmppc_core_flush_memslot_hv()
908 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_age_rmapp()
979 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_test_age_rmapp()
1124 if (gfn < memslot->base_gfn || in kvmppc_harvest_vpa_dirty()
1125 gfn >= memslot->base_gfn + memslot->npages) in kvmppc_harvest_vpa_dirty()
1130 __set_bit_le(gfn - memslot->base_gfn, map); in kvmppc_harvest_vpa_dirty()
1205 set_bit_le(gfn - memslot->base_gfn, memslot->dirty_bitmap); in kvmppc_unpin_guest_page()
[all …]
H A Dbook3s_64_mmu_radix.c1064 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_age_radix()
1096 unsigned long gfn = memslot->base_gfn + pagenum; in kvm_radix_test_clear_dirty()
1139 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_radix_test_clear_dirty()
1187 gpa = memslot->base_gfn << PAGE_SHIFT; in kvmppc_radix_flush_memslot()
H A Dbook3s_hv_rm_mmu.c104 gfn -= memslot->base_gfn; in kvmppc_update_dirty_map()
142 rmap = real_vmalloc_addr(&memslot->arch.rmap[gfn - memslot->base_gfn]); in revmap_for_hpte()
242 slot_fn = gfn - memslot->base_gfn; in kvmppc_do_h_enter()
H A De500_mmu_host.c381 slot_start = pfn - (gfn - slot->base_gfn); in kvmppc_e500_shadow_map()
H A Dbook3s_hv_nested.c1008 gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn; in kvmhv_remove_nest_rmap_range()
1643 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in __kvmhv_nested_page_fault()
H A Dbook3s_pr.c1888 ga = memslot->base_gfn << PAGE_SHIFT; in kvm_vm_ioctl_get_dirty_log_pr()
H A Dbook3s_hv.c928 if ((from + len) >= ((from_memslot->base_gfn + from_memslot->npages) in kvmppc_copy_guest()
940 if ((to + len) >= ((to_memslot->base_gfn + to_memslot->npages) in kvmppc_copy_guest()
/openbmc/linux/arch/mips/kvm/
H A Dmmu.c418 gfn_t base_gfn = slot->base_gfn + gfn_offset; in kvm_arch_mmu_enable_log_dirty_pt_masked() local
419 gfn_t start = base_gfn + __ffs(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked()
420 gfn_t end = base_gfn + __fls(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked()
H A Dmips.c200 kvm_mips_flush_gpa_pt(kvm, slot->base_gfn, in kvm_arch_flush_shadow_memslot()
201 slot->base_gfn + slot->npages - 1); in kvm_arch_flush_shadow_memslot()
235 needs_flush = kvm_mips_mkclean_gpa_pt(kvm, new->base_gfn, in kvm_arch_commit_memory_region()
236 new->base_gfn + new->npages - 1); in kvm_arch_commit_memory_region()
/openbmc/linux/include/linux/
H A Dkvm_host.h583 gfn_t base_gfn; member
1069 if (start < slot->base_gfn) { in kvm_memslot_iter_start()
1107 if (iter->slot->base_gfn + iter->slot->npages <= start) in kvm_memslot_iter_start()
1121 return iter->slot->base_gfn < end; in kvm_memslot_iter_is_valid()
1653 if (gfn >= slot->base_gfn && gfn < slot->base_gfn + slot->npages) in try_get_memslot()
1676 if (gfn >= slot->base_gfn) { in search_memslots()
1677 if (gfn < slot->base_gfn + slot->npages) in search_memslots()
1726 unsigned long offset = gfn - slot->base_gfn; in __gfn_to_hva_memslot()
1741 return slot->base_gfn + gfn_offset; in hva_to_gfn_memslot()
/openbmc/linux/virt/kvm/
H A Dkvm_main.c1456 if (slot->base_gfn < tmp->base_gfn) in kvm_insert_gfn_node()
1458 else if (slot->base_gfn > tmp->base_gfn) in kvm_insert_gfn_node()
1480 WARN_ON_ONCE(old->base_gfn != new->base_gfn); in kvm_replace_gfn_node()
1539 if (old && old->base_gfn == new->base_gfn) { in kvm_replace_memslot()
1735 dest->base_gfn = src->base_gfn; in kvm_copy_memslot()
1956 gfn_t base_gfn; in __kvm_set_memory_region() local
2004 base_gfn = (mem->guest_phys_addr >> PAGE_SHIFT); in __kvm_set_memory_region()
2022 if (base_gfn != old->base_gfn) in __kvm_set_memory_region()
2031 kvm_check_memslot_overlap(slots, id, base_gfn, base_gfn + npages)) in __kvm_set_memory_region()
2041 new->base_gfn = base_gfn; in __kvm_set_memory_region()
[all …]
/openbmc/linux/arch/s390/kvm/
H A Dpv.c261 while (slot && slot->base_gfn < pages_2g) { in kvm_s390_destroy_lower_2g()
262 len = min_t(u64, slot->npages, pages_2g - slot->base_gfn) * PAGE_SIZE; in kvm_s390_destroy_lower_2g()
265 slot = gfn_to_memslot(kvm, slot->base_gfn + slot->npages); in kvm_s390_destroy_lower_2g()
H A Dkvm-s390.c679 cur_gfn = memslot->base_gfn; in kvm_arch_sync_dirty_log()
680 last_gfn = memslot->base_gfn + memslot->npages; in kvm_arch_sync_dirty_log()
2225 unsigned long ofs = cur_gfn - ms->base_gfn; in kvm_s390_next_dirty_cmma()
2228 if (ms->base_gfn + ms->npages <= cur_gfn) { in kvm_s390_next_dirty_cmma()
2238 if (cur_gfn < ms->base_gfn) in kvm_s390_next_dirty_cmma()
2246 return ms->base_gfn + ofs; in kvm_s390_next_dirty_cmma()
2291 if (cur_gfn - ms->base_gfn >= ms->npages) { in kvm_s390_get_cmma()
5787 if ((new->base_gfn * PAGE_SIZE) + size > kvm->arch.mem_limit) in kvm_arch_prepare_memory_region()
5819 rc = gmap_unmap_segment(kvm->arch.gmap, old->base_gfn * PAGE_SIZE, in kvm_arch_commit_memory_region()
5823 rc = gmap_unmap_segment(kvm->arch.gmap, old->base_gfn * PAGE_SIZE, in kvm_arch_commit_memory_region()
[all …]
H A Dkvm-s390.h232 return ms->base_gfn + ms->npages; in kvm_s390_get_gfn_end()
H A Dpriv.c1208 if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms))) in __do_essa()
/openbmc/linux/arch/powerpc/include/asm/
H A Dkvm_book3s_64.h495 return !(memslot->base_gfn & mask) && !(memslot->npages & mask); in slot_is_aligned()

12