Lines Matching defs:kvm

157 __weak void kvm_arch_guest_memory_reclaimed(struct kvm *kvm)  in kvm_arch_guest_memory_reclaimed()
290 bool kvm_make_vcpus_request_mask(struct kvm *kvm, unsigned int req, in kvm_make_vcpus_request_mask()
316 bool kvm_make_all_cpus_request_except(struct kvm *kvm, unsigned int req, in kvm_make_all_cpus_request_except()
342 bool kvm_make_all_cpus_request(struct kvm *kvm, unsigned int req) in kvm_make_all_cpus_request()
348 void kvm_flush_remote_tlbs(struct kvm *kvm) in kvm_flush_remote_tlbs()
369 void kvm_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages) in kvm_flush_remote_tlbs_range()
382 void kvm_flush_remote_tlbs_memslot(struct kvm *kvm, in kvm_flush_remote_tlbs_memslot()
396 static void kvm_flush_shadow_all(struct kvm *kvm) in kvm_flush_shadow_all()
484 static void kvm_vcpu_init(struct kvm_vcpu *vcpu, struct kvm *kvm, unsigned id) in kvm_vcpu_init()
524 void kvm_destroy_vcpus(struct kvm *kvm) in kvm_destroy_vcpus()
583 static __always_inline int __kvm_handle_hva_range(struct kvm *kvm, in __kvm_handle_hva_range()
665 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_handle_hva_range() local
685 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_handle_hva_range_no_flush() local
699 static bool kvm_change_spte_gfn(struct kvm *kvm, struct kvm_gfn_range *range) in kvm_change_spte_gfn()
721 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_change_pte() local
741 void kvm_mmu_invalidate_begin(struct kvm *kvm, unsigned long start, in kvm_mmu_invalidate_begin()
773 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_invalidate_range_start() local
816 void kvm_mmu_invalidate_end(struct kvm *kvm, unsigned long start, in kvm_mmu_invalidate_end()
837 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_invalidate_range_end() local
913 struct kvm *kvm = mmu_notifier_to_kvm(mn); in kvm_mmu_notifier_release() local
931 static int kvm_init_mmu_notifier(struct kvm *kvm) in kvm_init_mmu_notifier()
939 static int kvm_init_mmu_notifier(struct kvm *kvm) in kvm_init_mmu_notifier()
951 struct kvm *kvm = container_of(bl, struct kvm, pm_notifier); in kvm_pm_notifier_call() local
956 static void kvm_init_pm_notifier(struct kvm *kvm) in kvm_init_pm_notifier()
964 static void kvm_destroy_pm_notifier(struct kvm *kvm) in kvm_destroy_pm_notifier()
969 static void kvm_init_pm_notifier(struct kvm *kvm) in kvm_init_pm_notifier()
973 static void kvm_destroy_pm_notifier(struct kvm *kvm) in kvm_destroy_pm_notifier()
988 static void kvm_free_memslot(struct kvm *kvm, struct kvm_memory_slot *slot) in kvm_free_memslot()
997 static void kvm_free_memslots(struct kvm *kvm, struct kvm_memslots *slots) in kvm_free_memslots()
1029 static void kvm_destroy_vm_debugfs(struct kvm *kvm) in kvm_destroy_vm_debugfs()
1047 static int kvm_create_vm_debugfs(struct kvm *kvm, const char *fdname) in kvm_create_vm_debugfs()
1126 int __weak kvm_arch_post_init_vm(struct kvm *kvm) in kvm_arch_post_init_vm()
1135 void __weak kvm_arch_pre_destroy_vm(struct kvm *kvm) in kvm_arch_pre_destroy_vm()
1145 int __weak kvm_arch_create_vm_debugfs(struct kvm *kvm) in kvm_arch_create_vm_debugfs()
1152 struct kvm *kvm = kvm_arch_alloc_vm(); in kvm_create_vm() local
1286 static void kvm_destroy_devices(struct kvm *kvm) in kvm_destroy_devices()
1301 static void kvm_destroy_vm(struct kvm *kvm) in kvm_destroy_vm()
1354 void kvm_get_kvm(struct kvm *kvm) in kvm_get_kvm()
1364 bool kvm_get_kvm_safe(struct kvm *kvm) in kvm_get_kvm_safe()
1370 void kvm_put_kvm(struct kvm *kvm) in kvm_put_kvm()
1384 void kvm_put_kvm_no_destroy(struct kvm *kvm) in kvm_put_kvm_no_destroy()
1392 struct kvm *kvm = filp->private_data; in kvm_vm_release() local
1415 static struct kvm_memslots *kvm_get_inactive_memslots(struct kvm *kvm, int as_id) in kvm_get_inactive_memslots()
1495 static void kvm_replace_memslot(struct kvm *kvm, in kvm_replace_memslot()
1562 static void kvm_swap_active_memslots(struct kvm *kvm, int as_id) in kvm_swap_active_memslots()
1620 static int kvm_prepare_memory_region(struct kvm *kvm, in kvm_prepare_memory_region()
1658 static void kvm_commit_memory_region(struct kvm *kvm, in kvm_commit_memory_region()
1720 static void kvm_activate_memslot(struct kvm *kvm, in kvm_activate_memslot()
1745 static void kvm_invalidate_memslot(struct kvm *kvm, in kvm_invalidate_memslot()
1787 static void kvm_create_memslot(struct kvm *kvm, in kvm_create_memslot()
1795 static void kvm_delete_memslot(struct kvm *kvm, in kvm_delete_memslot()
1807 static void kvm_move_memslot(struct kvm *kvm, in kvm_move_memslot()
1820 static void kvm_update_flags_memslot(struct kvm *kvm, in kvm_update_flags_memslot()
1833 static int kvm_set_memslot(struct kvm *kvm, in kvm_set_memslot()
1949 int __kvm_set_memory_region(struct kvm *kvm, in __kvm_set_memory_region()
2053 int kvm_set_memory_region(struct kvm *kvm, in kvm_set_memory_region()
2065 static int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, in kvm_vm_ioctl_set_memory_region()
2082 int kvm_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log, in kvm_get_dirty_log()
2145 static int kvm_get_dirty_log_protect(struct kvm *kvm, struct kvm_dirty_log *log) in kvm_get_dirty_log_protect()
2236 static int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, in kvm_vm_ioctl_get_dirty_log()
2255 static int kvm_clear_dirty_log_protect(struct kvm *kvm, in kvm_clear_dirty_log_protect()
2331 static int kvm_vm_ioctl_clear_dirty_log(struct kvm *kvm, in kvm_vm_ioctl_clear_dirty_log()
2345 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) in gfn_to_memslot()
2384 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) in kvm_is_visible_gfn()
2457 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn) in gfn_to_hva()
2488 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable) in gfn_to_hva_prot()
2778 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, in gfn_to_pfn_prot()
2806 kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn) in gfn_to_pfn()
2841 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn) in gfn_to_page()
3043 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset, in kvm_read_guest_page()
3061 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len) in kvm_read_guest()
3129 static int __kvm_write_guest_page(struct kvm *kvm, in __kvm_write_guest_page()
3146 int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, in kvm_write_guest_page()
3164 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data, in kvm_write_guest()
3247 int kvm_gfn_to_hva_cache_init(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_gfn_to_hva_cache_init()
3255 int kvm_write_guest_offset_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_write_guest_offset_cached()
3286 int kvm_write_guest_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_write_guest_cached()
3293 int kvm_read_guest_offset_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_read_guest_offset_cached()
3323 int kvm_read_guest_cached(struct kvm *kvm, struct gfn_to_hva_cache *ghc, in kvm_read_guest_cached()
3330 int kvm_clear_guest(struct kvm *kvm, gpa_t gpa, unsigned long len) in kvm_clear_guest()
3350 void mark_page_dirty_in_slot(struct kvm *kvm, in mark_page_dirty_in_slot()
3375 void mark_page_dirty(struct kvm *kvm, gfn_t gfn) in mark_page_dirty()
3535 struct kvm *kvm = vcpu->kvm; in kvm_vcpu_max_halt_poll_ns() local
3773 struct kvm *kvm = me->kvm; in kvm_vcpu_on_spin() local
3828 static bool kvm_page_in_dirty_ring(struct kvm *kvm, unsigned long pgoff) in kvm_page_in_dirty_ring()
3944 static int kvm_vm_ioctl_create_vcpu(struct kvm *kvm, u32 id) in kvm_vm_ioctl_create_vcpu()
4422 struct kvm *kvm = dev->kvm; in kvm_device_release() local
4475 static int kvm_ioctl_create_device(struct kvm *kvm, in kvm_ioctl_create_device()
4533 static int kvm_vm_ioctl_check_extension_generic(struct kvm *kvm, long arg) in kvm_vm_ioctl_check_extension_generic()
4595 static int kvm_vm_ioctl_enable_dirty_log_ring(struct kvm *kvm, u32 size) in kvm_vm_ioctl_enable_dirty_log_ring()
4633 static int kvm_vm_ioctl_reset_dirty_pages(struct kvm *kvm) in kvm_vm_ioctl_reset_dirty_pages()
4655 int __attribute__((weak)) kvm_vm_ioctl_enable_cap(struct kvm *kvm, in kvm_vm_ioctl_enable_cap()
4661 bool kvm_are_all_memslots_empty(struct kvm *kvm) in kvm_are_all_memslots_empty()
4676 static int kvm_vm_ioctl_enable_cap_generic(struct kvm *kvm, in kvm_vm_ioctl_enable_cap_generic()
4747 struct kvm *kvm = file->private_data; in kvm_vm_stats_read() local
4756 struct kvm *kvm = file->private_data; in kvm_vm_stats_release() local
4768 static int kvm_vm_ioctl_get_stats_fd(struct kvm *kvm) in kvm_vm_ioctl_get_stats_fd()
4795 struct kvm *kvm = filp->private_data; in kvm_vm_ioctl() local
5013 struct kvm *kvm = filp->private_data; in kvm_vm_compat_ioctl() local
5081 struct kvm *kvm; in kvm_dev_ioctl_create_vm() local
5551 int kvm_io_bus_register_dev(struct kvm *kvm, enum kvm_bus bus_idx, gpa_t addr, in kvm_io_bus_register_dev()
5593 int kvm_io_bus_unregister_dev(struct kvm *kvm, enum kvm_bus bus_idx, in kvm_io_bus_unregister_dev()
5641 struct kvm_io_device *kvm_io_bus_get_dev(struct kvm *kvm, enum kvm_bus bus_idx, in kvm_io_bus_get_dev()
5701 static int kvm_get_stat_per_vm(struct kvm *kvm, size_t offset, u64 *val) in kvm_get_stat_per_vm()
5708 static int kvm_clear_stat_per_vm(struct kvm *kvm, size_t offset) in kvm_clear_stat_per_vm()
5715 static int kvm_get_stat_per_vcpu(struct kvm *kvm, size_t offset, u64 *val) in kvm_get_stat_per_vcpu()
5728 static int kvm_clear_stat_per_vcpu(struct kvm *kvm, size_t offset) in kvm_clear_stat_per_vcpu()
5799 struct kvm *kvm; in vm_stat_get() local
5815 struct kvm *kvm; in vm_stat_clear() local
5835 struct kvm *kvm; in vcpu_stat_get() local
5851 struct kvm *kvm; in vcpu_stat_clear() local
5869 static void kvm_uevent_notify_change(unsigned int type, struct kvm *kvm) in kvm_uevent_notify_change()
6168 struct kvm *kvm; member
6184 struct kvm *kvm = init_context->kvm; in kvm_vm_worker_thread() local
6240 int kvm_vm_create_worker_thread(struct kvm *kvm, kvm_vm_thread_fn_t thread_fn, in kvm_vm_create_worker_thread()