/openbmc/linux/arch/arm64/kernel/ |
H A D | vdso.c | 155 struct vm_area_struct *vma, struct vm_fault *vmf) in vvar_fault() argument 160 switch (vmf->pgoff) { in vvar_fault() 185 return vmf_insert_pfn(vma, vmf->address, pfn); in vvar_fault()
|
/openbmc/linux/drivers/media/v4l2-core/ |
H A D | videobuf-dma-sg.c | 430 static vm_fault_t videobuf_vm_fault(struct vm_fault *vmf) in videobuf_vm_fault() argument 432 struct vm_area_struct *vma = vmf->vma; in videobuf_vm_fault() 436 vmf->address, vma->vm_start, vma->vm_end); in videobuf_vm_fault() 441 clear_user_highpage(page, vmf->address); in videobuf_vm_fault() 442 vmf->page = page; in videobuf_vm_fault()
|
/openbmc/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem.c | 163 static vm_fault_t etnaviv_gem_fault(struct vm_fault *vmf) in etnaviv_gem_fault() argument 165 struct vm_area_struct *vma = vmf->vma; in etnaviv_gem_fault() 191 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in etnaviv_gem_fault() 195 VERB("Inserting %p pfn %lx, pa %lx", (void *)vmf->address, in etnaviv_gem_fault() 198 return vmf_insert_pfn(vma, vmf->address, pfn); in etnaviv_gem_fault()
|
/openbmc/linux/arch/powerpc/kvm/ |
H A D | book3s_xive_native.c | 228 static vm_fault_t xive_native_esb_fault(struct vm_fault *vmf) in xive_native_esb_fault() argument 230 struct vm_area_struct *vma = vmf->vma; in xive_native_esb_fault() 246 page_offset = vmf->pgoff - vma->vm_pgoff; in xive_native_esb_fault() 280 vmf_insert_pfn(vma, vmf->address, page >> PAGE_SHIFT); in xive_native_esb_fault() 288 static vm_fault_t xive_native_tima_fault(struct vm_fault *vmf) in xive_native_tima_fault() argument 290 struct vm_area_struct *vma = vmf->vma; in xive_native_tima_fault() 292 switch (vmf->pgoff - vma->vm_pgoff) { in xive_native_tima_fault() 297 vmf_insert_pfn(vma, vmf->address, xive_tima_os >> PAGE_SHIFT); in xive_native_tima_fault()
|
H A D | book3s_64_vio.c | 239 static vm_fault_t kvm_spapr_tce_fault(struct vm_fault *vmf) in kvm_spapr_tce_fault() argument 241 struct kvmppc_spapr_tce_table *stt = vmf->vma->vm_file->private_data; in kvm_spapr_tce_fault() 244 if (vmf->pgoff >= kvmppc_tce_pages(stt->size)) in kvm_spapr_tce_fault() 247 page = kvm_spapr_get_tce_page(stt, vmf->pgoff); in kvm_spapr_tce_fault() 252 vmf->page = page; in kvm_spapr_tce_fault()
|
H A D | book3s_hv_uvmem.c | 999 static vm_fault_t kvmppc_uvmem_migrate_to_ram(struct vm_fault *vmf) in kvmppc_uvmem_migrate_to_ram() argument 1001 struct kvmppc_uvmem_page_pvt *pvt = vmf->page->zone_device_data; in kvmppc_uvmem_migrate_to_ram() 1003 if (kvmppc_svm_page_out(vmf->vma, vmf->address, in kvmppc_uvmem_migrate_to_ram() 1004 vmf->address + PAGE_SIZE, PAGE_SHIFT, in kvmppc_uvmem_migrate_to_ram() 1005 pvt->kvm, pvt->gpa, vmf->page)) in kvmppc_uvmem_migrate_to_ram()
|
/openbmc/linux/arch/hexagon/include/asm/ |
H A D | cacheflush.h | 61 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() argument
|
/openbmc/linux/arch/microblaze/include/asm/ |
H A D | tlbflush.h | 36 #define update_mmu_cache_range(vmf, vma, addr, ptep, nr) do { } while (0) argument
|
/openbmc/linux/fs/erofs/ |
H A D | data.c | 416 static vm_fault_t erofs_dax_huge_fault(struct vm_fault *vmf, in erofs_dax_huge_fault() argument 419 return dax_iomap_fault(vmf, order, NULL, NULL, &erofs_iomap_ops); in erofs_dax_huge_fault() 422 static vm_fault_t erofs_dax_fault(struct vm_fault *vmf) in erofs_dax_fault() argument 424 return erofs_dax_huge_fault(vmf, 0); in erofs_dax_fault()
|
/openbmc/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_ttm.c | 1041 static vm_fault_t vm_fault_ttm(struct vm_fault *vmf) in vm_fault_ttm() argument 1043 struct vm_area_struct *area = vmf->vma; in vm_fault_ttm() 1056 ret = ttm_bo_vm_reserve(bo, vmf); in vm_fault_ttm() 1116 ret = ttm_bo_vm_fault_reserved(vmf, vmf->vma->vm_page_prot, in vm_fault_ttm() 1120 ret = ttm_bo_vm_dummy_page(vmf, vmf->vma->vm_page_prot); in vm_fault_ttm() 1123 if (ret == VM_FAULT_RETRY && !(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) in vm_fault_ttm()
|
/openbmc/linux/include/linux/ |
H A D | userfaultfd_k.h | 39 extern vm_fault_t handle_userfault(struct vm_fault *vmf, unsigned long reason); 204 static inline vm_fault_t handle_userfault(struct vm_fault *vmf, in handle_userfault() argument
|
/openbmc/linux/drivers/gpu/drm/radeon/ |
H A D | radeon_gem.c | 48 static vm_fault_t radeon_gem_fault(struct vm_fault *vmf) in radeon_gem_fault() argument 50 struct ttm_buffer_object *bo = vmf->vma->vm_private_data; in radeon_gem_fault() 56 ret = ttm_bo_vm_reserve(bo, vmf); in radeon_gem_fault() 64 ret = ttm_bo_vm_fault_reserved(vmf, vmf->vma->vm_page_prot, in radeon_gem_fault() 66 if (ret == VM_FAULT_RETRY && !(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) in radeon_gem_fault()
|
/openbmc/linux/fs/kernfs/ |
H A D | file.c | 369 static vm_fault_t kernfs_vma_fault(struct vm_fault *vmf) in kernfs_vma_fault() argument 371 struct file *file = vmf->vma->vm_file; in kernfs_vma_fault() 383 ret = of->vm_ops->fault(vmf); in kernfs_vma_fault() 389 static vm_fault_t kernfs_vma_page_mkwrite(struct vm_fault *vmf) in kernfs_vma_page_mkwrite() argument 391 struct file *file = vmf->vma->vm_file; in kernfs_vma_page_mkwrite() 403 ret = of->vm_ops->page_mkwrite(vmf); in kernfs_vma_page_mkwrite()
|
/openbmc/linux/arch/csky/abiv1/ |
H A D | cacheflush.c | 44 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, in update_mmu_cache_range() argument
|
/openbmc/linux/arch/powerpc/include/asm/nohash/ |
H A D | pgtable.h | 279 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, 282 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() argument
|
/openbmc/linux/fs/orangefs/ |
H A D | file.c | 376 static vm_fault_t orangefs_fault(struct vm_fault *vmf) in orangefs_fault() argument 378 struct file *file = vmf->vma->vm_file; in orangefs_fault() 389 return filemap_fault(vmf); in orangefs_fault()
|
/openbmc/linux/drivers/gpu/drm/ |
H A D | drm_gem_shmem_helper.c | 509 static vm_fault_t drm_gem_shmem_fault(struct vm_fault *vmf) in drm_gem_shmem_fault() argument 511 struct vm_area_struct *vma = vmf->vma; in drm_gem_shmem_fault() 520 page_offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in drm_gem_shmem_fault() 531 ret = vmf_insert_pfn(vma, vmf->address, page_to_pfn(page)); in drm_gem_shmem_fault()
|
/openbmc/linux/sound/usb/usx2y/ |
H A D | usx2yhwdeppcm.c | 673 static vm_fault_t snd_usx2y_hwdep_pcm_vm_fault(struct vm_fault *vmf) in snd_usx2y_hwdep_pcm_vm_fault() argument 678 offset = vmf->pgoff << PAGE_SHIFT; in snd_usx2y_hwdep_pcm_vm_fault() 679 vaddr = (char *)((struct usx2ydev *)vmf->vma->vm_private_data)->hwdep_pcm_shm + offset; in snd_usx2y_hwdep_pcm_vm_fault() 680 vmf->page = virt_to_page(vaddr); in snd_usx2y_hwdep_pcm_vm_fault() 681 get_page(vmf->page); in snd_usx2y_hwdep_pcm_vm_fault()
|
/openbmc/linux/mm/ |
H A D | internal.h | 106 vm_fault_t do_swap_page(struct vm_fault *vmf); 693 static inline struct file *maybe_unlock_mmap_for_io(struct vm_fault *vmf, in maybe_unlock_mmap_for_io() argument 696 int flags = vmf->flags; in maybe_unlock_mmap_for_io() 708 fpin = get_file(vmf->vma->vm_file); in maybe_unlock_mmap_for_io() 709 release_fault_lock(vmf); in maybe_unlock_mmap_for_io()
|
/openbmc/linux/drivers/hwtracing/intel_th/ |
H A D | msu.c | 1608 static vm_fault_t msc_mmap_fault(struct vm_fault *vmf) in msc_mmap_fault() argument 1610 struct msc_iter *iter = vmf->vma->vm_file->private_data; in msc_mmap_fault() 1613 vmf->page = msc_buffer_get_page(msc, vmf->pgoff); in msc_mmap_fault() 1614 if (!vmf->page) in msc_mmap_fault() 1617 get_page(vmf->page); in msc_mmap_fault() 1618 vmf->page->mapping = vmf->vma->vm_file->f_mapping; in msc_mmap_fault() 1619 vmf->page->index = vmf->pgoff; in msc_mmap_fault()
|
/openbmc/linux/arch/powerpc/platforms/book3s/ |
H A D | vas-api.c | 395 static vm_fault_t vas_mmap_fault(struct vm_fault *vmf) in vas_mmap_fault() argument 397 struct vm_area_struct *vma = vmf->vma; in vas_mmap_fault() 423 if (txwin->task_ref.vma != vmf->vma) { in vas_mmap_fault()
|
/openbmc/linux/drivers/vdpa/vdpa_user/ |
H A D | iova_domain.c | 485 static vm_fault_t vduse_domain_mmap_fault(struct vm_fault *vmf) in vduse_domain_mmap_fault() argument 487 struct vduse_iova_domain *domain = vmf->vma->vm_private_data; in vduse_domain_mmap_fault() 488 unsigned long iova = vmf->pgoff << PAGE_SHIFT; in vduse_domain_mmap_fault() 502 vmf->page = page; in vduse_domain_mmap_fault()
|
/openbmc/linux/fs/afs/ |
H A D | file.c | 33 static vm_fault_t afs_vm_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff); 576 static vm_fault_t afs_vm_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff) in afs_vm_map_pages() argument 578 struct afs_vnode *vnode = AFS_FS_I(file_inode(vmf->vma->vm_file)); in afs_vm_map_pages() 581 return filemap_map_pages(vmf, start_pgoff, end_pgoff); in afs_vm_map_pages()
|
/openbmc/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_migrate.c | 904 static vm_fault_t svm_migrate_to_ram(struct vm_fault *vmf) in svm_migrate_to_ram() argument 906 unsigned long addr = vmf->address; in svm_migrate_to_ram() 915 svm_bo = vmf->page->zone_device_data; in svm_migrate_to_ram() 926 if (mm != vmf->vma->vm_mm) in svm_migrate_to_ram() 972 r = svm_migrate_vram_to_ram(prange, vmf->vma->vm_mm, in svm_migrate_to_ram() 974 vmf->page); in svm_migrate_to_ram()
|
/openbmc/linux/arch/csky/abiv2/ |
H A D | cacheflush.c | 10 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, in update_mmu_cache_range() argument
|