/openbmc/linux/arch/x86/hyperv/ |
H A D | hv_spinlock.c | 80 pv_ops.lock.queued_spin_lock_slowpath = __pv_queued_spin_lock_slowpath; in hv_init_spinlocks() 81 pv_ops.lock.queued_spin_unlock = PV_CALLEE_SAVE(__pv_queued_spin_unlock); in hv_init_spinlocks() 82 pv_ops.lock.wait = hv_qlock_wait; in hv_init_spinlocks() 83 pv_ops.lock.kick = hv_qlock_kick; in hv_init_spinlocks() 84 pv_ops.lock.vcpu_is_preempted = PV_CALLEE_SAVE(hv_vcpu_is_preempted); in hv_init_spinlocks()
|
H A D | mmu.c | 242 pv_ops.mmu.flush_tlb_multi = hyperv_flush_tlb_multi; in hyperv_setup_mmu_ops() 243 pv_ops.mmu.tlb_remove_table = tlb_remove_table; in hyperv_setup_mmu_ops()
|
/openbmc/linux/Documentation/translations/zh_CN/virt/ |
H A D | paravirt_ops.rst | 22 (hypervisor,下文简称超级管理器),需要不同的二进制内核,这个限制已经被pv_ops移 23 除了。Linux pv_ops是一个虚拟化API,它能够支持不同的管理程序。它允许每个管理程序 27 pv_ops提供了一组函数指针,代表了与低级关键指令和各领域高级功能相对应的操作。 30 pv_ops操作被分为三类:
|
/openbmc/linux/arch/x86/xen/ |
H A D | spinlock.c | 139 pv_ops.lock.queued_spin_lock_slowpath = __pv_queued_spin_lock_slowpath; in xen_init_spinlocks() 140 pv_ops.lock.queued_spin_unlock = in xen_init_spinlocks() 142 pv_ops.lock.wait = xen_qlock_wait; in xen_init_spinlocks() 143 pv_ops.lock.kick = xen_qlock_kick; in xen_init_spinlocks() 144 pv_ops.lock.vcpu_is_preempted = PV_CALLEE_SAVE(xen_vcpu_stolen); in xen_init_spinlocks()
|
H A D | irq.c | 43 static const typeof(pv_ops) xen_irq_ops __initconst = { 57 pv_ops.irq = xen_irq_ops.irq; in xen_init_irq_ops()
|
H A D | enlighten_pv.c | 1129 pv_ops.irq.save_fl = __PV_IS_CALLEE_SAVE(xen_save_fl_direct); in xen_setup_vcpu_info_placement() 1130 pv_ops.irq.irq_disable = __PV_IS_CALLEE_SAVE(xen_irq_disable_direct); in xen_setup_vcpu_info_placement() 1131 pv_ops.irq.irq_enable = __PV_IS_CALLEE_SAVE(xen_irq_enable_direct); in xen_setup_vcpu_info_placement() 1132 pv_ops.mmu.read_cr2 = __PV_IS_CALLEE_SAVE(xen_read_cr2_direct); in xen_setup_vcpu_info_placement() 1140 static const typeof(pv_ops) xen_cpu_ops __initconst = { 1291 pv_ops.cpu.write_gdt_entry = xen_write_gdt_entry_boot; in xen_setup_gdt() 1292 pv_ops.cpu.load_gdt = xen_load_gdt_boot; in xen_setup_gdt() 1296 pv_ops.cpu.write_gdt_entry = xen_write_gdt_entry; in xen_setup_gdt() 1297 pv_ops.cpu.load_gdt = xen_load_gdt; in xen_setup_gdt() 1340 pv_ops.cpu = xen_cpu_ops.cpu; in xen_start_kernel()
|
H A D | mmu_pv.c | 1267 pv_ops.mmu.set_pte = __xen_set_pte; in xen_pagetable_init() 2154 pv_ops.mmu.set_pte = xen_set_pte; in xen_post_allocator_init() 2155 pv_ops.mmu.set_pmd = xen_set_pmd; in xen_post_allocator_init() 2156 pv_ops.mmu.set_pud = xen_set_pud; in xen_post_allocator_init() 2157 pv_ops.mmu.set_p4d = xen_set_p4d; in xen_post_allocator_init() 2161 pv_ops.mmu.alloc_pte = xen_alloc_pte; in xen_post_allocator_init() 2162 pv_ops.mmu.alloc_pmd = xen_alloc_pmd; in xen_post_allocator_init() 2163 pv_ops.mmu.release_pte = xen_release_pte; in xen_post_allocator_init() 2164 pv_ops.mmu.release_pmd = xen_release_pmd; in xen_post_allocator_init() 2165 pv_ops.mmu.alloc_pud = xen_alloc_pud; in xen_post_allocator_init() [all …]
|
H A D | mmu_hvm.c | 65 pv_ops.mmu.exit_mmap = xen_hvm_exit_mmap; in xen_hvm_init_mmu_ops()
|
/openbmc/linux/Documentation/virt/ |
H A D | paravirt_ops.rst | 9 different hypervisors; this restriction was removed with pv_ops. 10 Linux pv_ops is a virtualization API which enables support for different 15 pv_ops provides a set of function pointers which represent operations 17 functionalities in various areas. pv_ops allows for optimizations at run 21 pv_ops operations are classified into three categories:
|
/openbmc/linux/tools/objtool/ |
H A D | objtool.c | 120 if (!f->pv_ops) { in objtool_pv_add() 137 list_add(&func->pv_target, &f->pv_ops[idx].targets); in objtool_pv_add() 138 f->pv_ops[idx].clean = false; in objtool_pv_add()
|
H A D | check.c | 539 const char *pv_ops; in init_pv_ops() local 546 file->pv_ops = NULL; in init_pv_ops() 553 file->pv_ops = calloc(sizeof(struct pv_state), nr); in init_pv_ops() 554 if (!file->pv_ops) in init_pv_ops() 558 INIT_LIST_HEAD(&file->pv_ops[idx].targets); in init_pv_ops() 560 for (idx = 0; (pv_ops = pv_ops_tables[idx]); idx++) in init_pv_ops() 561 add_pv_ops(file, pv_ops); in init_pv_ops() 3381 if (file->pv_ops[idx].clean) in pv_call_dest() 3384 file->pv_ops[idx].clean = true; in pv_call_dest() 3386 list_for_each_entry(target, &file->pv_ops[idx].targets, pv_target) { in pv_call_dest() [all …]
|
/openbmc/linux/arch/x86/kernel/ |
H A D | paravirt-spinlocks.c | 20 return pv_ops.lock.queued_spin_unlock.func == in pv_is_native_spin_unlock() 32 return pv_ops.lock.vcpu_is_preempted.func == in pv_is_native_vcpu_is_preempted()
|
H A D | kvm.c | 313 pv_ops.cpu.io_delay = kvm_io_delay; in paravirt_ops_setup() 823 pv_ops.lock.vcpu_is_preempted = in kvm_guest_init() 837 pv_ops.mmu.flush_tlb_multi = kvm_flush_tlb_multi; in kvm_guest_init() 838 pv_ops.mmu.tlb_remove_table = tlb_remove_table; in kvm_guest_init() 938 pv_ops.mmu.notify_page_enc_status_changed = in kvm_init_platform() 1101 pv_ops.lock.queued_spin_lock_slowpath = __pv_queued_spin_lock_slowpath; in kvm_spinlock_init() 1102 pv_ops.lock.queued_spin_unlock = in kvm_spinlock_init() 1104 pv_ops.lock.wait = kvm_wait; in kvm_spinlock_init() 1105 pv_ops.lock.kick = kvm_kick_cpu; in kvm_spinlock_init()
|
H A D | paravirt.c | 94 void *opfunc = *((void **)&pv_ops + type); in paravirt_patch() 182 struct paravirt_patch_template pv_ops = { variable 317 EXPORT_SYMBOL(pv_ops);
|
/openbmc/linux/arch/x86/include/asm/ |
H A D | paravirt_types.h | 251 extern struct paravirt_patch_template pv_ops; 258 [paravirt_opptr] "m" (pv_ops.op) 409 #define PVOP_TEST_NULL(op) BUG_ON(pv_ops.op == NULL) 411 #define PVOP_TEST_NULL(op) ((void)pv_ops.op)
|
H A D | gsseg.h | 48 pv_ops.cpu.load_gs_index = native_lkgs; in lkgs_init()
|
H A D | paravirt.h | 577 pv_ops.mmu.set_fixmap(idx, phys, flags); in __set_fixmap() 770 call PARA_INDIRECT(pv_ops+PV_IRQ_save_fl);)
|
/openbmc/linux/tools/objtool/include/objtool/ |
H A D | objtool.h | 39 struct pv_state *pv_ops; member
|
/openbmc/linux/arch/x86/kernel/cpu/ |
H A D | vmware.c | 332 pv_ops.cpu.io_delay = paravirt_nop; in vmware_paravirt_ops_setup()
|