/openbmc/linux/mm/ |
H A D | hmm.c | 138 for (i = 0; i < npages; ++i) { in hmm_range_need_fault() 153 unsigned long i, npages; in hmm_vma_walk_hole() local 157 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_walk_hole() 160 hmm_range_need_fault(hmm_vma_walk, hmm_pfns, npages, 0); in hmm_vma_walk_hole() 193 unsigned long pfn, npages, i; in hmm_vma_handle_pmd() local 197 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_handle_pmd() 329 unsigned long npages = (end - start) >> PAGE_SHIFT; in hmm_vma_walk_pmd() local 433 unsigned long i, npages, pfn; in hmm_vma_walk_pud() local 444 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_walk_pud() 449 npages, cpu_flags); in hmm_vma_walk_pud() [all …]
|
/openbmc/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem_prime.c | 20 int npages = obj->size >> PAGE_SHIFT; in etnaviv_gem_prime_get_sg_table() local 25 return drm_prime_pages_to_sg(obj->dev, etnaviv_obj->pages, npages); in etnaviv_gem_prime_get_sg_table() 117 int ret, npages; in etnaviv_gem_prime_import_sg_table() local 126 npages = size / PAGE_SIZE; in etnaviv_gem_prime_import_sg_table() 129 etnaviv_obj->pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in etnaviv_gem_prime_import_sg_table() 135 ret = drm_prime_sg_to_page_array(sgt, etnaviv_obj->pages, npages); in etnaviv_gem_prime_import_sg_table()
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx4/ |
H A D | icm.c | 63 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages() 72 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent() 159 while (npages > 0) { in mlx4_alloc_icm() 179 while (1 << cur_order > npages) in mlx4_alloc_icm() 188 &chunk->buf[chunk->npages], in mlx4_alloc_icm() 202 ++chunk->npages; in mlx4_alloc_icm() 208 chunk->sg, chunk->npages, in mlx4_alloc_icm() 215 if (chunk->npages == MLX4_ICM_CHUNK_LEN) in mlx4_alloc_icm() 218 npages -= 1 << cur_order; in mlx4_alloc_icm() 223 chunk->npages, DMA_BIDIRECTIONAL); in mlx4_alloc_icm() [all …]
|
H A D | mr.c | 199 if (!npages) { in mlx4_mtt_init() 538 access, npages, page_shift, mr); in mlx4_mr_alloc() 590 u64 iova, u64 size, int npages, in mlx4_mr_rereg_mem_write() argument 707 npages * sizeof(u64), DMA_TO_DEVICE); in mlx4_write_mtt_chunk() 709 for (i = 0; i < npages; ++i) in mlx4_write_mtt_chunk() 733 while (npages > 0) { in __mlx4_write_mtt() 737 npages -= chunk; in __mlx4_write_mtt() 764 while (npages > 0) { in mlx4_write_mtt() 766 npages); in mlx4_write_mtt() 778 npages -= chunk; in mlx4_write_mtt() [all …]
|
/openbmc/linux/arch/powerpc/sysdev/ |
H A D | dart_iommu.c | 173 long npages, unsigned long uaddr, in dart_build() argument 181 DBG("dart: build at: %lx, %lx, addr: %x\n", index, npages, uaddr); in dart_build() 188 l = npages; in dart_build() 196 dart_cache_sync(orig_dp, npages); in dart_build() 200 while (npages--) in dart_build() 209 static void dart_free(struct iommu_table *tbl, long index, long npages) in dart_free() argument 212 long orig_npages = npages; in dart_free() 219 DBG("dart: free at: %lx, %lx\n", index, npages); in dart_free() 223 while (npages--) in dart_free()
|
/openbmc/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_mr.c | 122 int ret, npages; in pvrdma_reg_user_mr() local 136 npages = ib_umem_num_dma_blocks(umem, PAGE_SIZE); in pvrdma_reg_user_mr() 137 if (npages < 0 || npages > PVRDMA_PAGE_DIR_MAX_PAGES) { in pvrdma_reg_user_mr() 139 npages); in pvrdma_reg_user_mr() 154 ret = pvrdma_page_dir_init(dev, &mr->pdir, npages, false); in pvrdma_reg_user_mr() 171 cmd->nchunks = npages; in pvrdma_reg_user_mr() 306 if (mr->npages == mr->max_pages) in pvrdma_set_page() 309 mr->pages[mr->npages++] = addr; in pvrdma_set_page() 320 mr->npages = 0; in pvrdma_map_mr_sg()
|
/openbmc/linux/arch/x86/mm/ |
H A D | mem_encrypt_amd.c | 62 unsigned long npages = PAGE_ALIGN(sz) >> PAGE_SHIFT; in snp_memcpy() local 69 early_snp_set_memory_shared((unsigned long)__va(paddr), paddr, npages); in snp_memcpy() 74 early_snp_set_memory_private((unsigned long)__va(paddr), paddr, npages); in snp_memcpy() 322 static bool amd_enc_status_change_prepare(unsigned long vaddr, int npages, bool enc) in amd_enc_status_change_prepare() argument 329 snp_set_memory_shared(vaddr, npages); in amd_enc_status_change_prepare() 335 static bool amd_enc_status_change_finish(unsigned long vaddr, int npages, bool enc) in amd_enc_status_change_finish() argument 342 snp_set_memory_private(vaddr, npages); in amd_enc_status_change_finish() 345 enc_dec_hypercall(vaddr, npages << PAGE_SHIFT, enc); in amd_enc_status_change_finish() 553 unsigned long vaddr, vaddr_end, npages; in mem_encrypt_free_decrypted_mem() local 558 npages = (vaddr_end - vaddr) >> PAGE_SHIFT; in mem_encrypt_free_decrypted_mem() [all …]
|
/openbmc/linux/drivers/iommu/iommufd/ |
H A D | pages.c | 188 unpin_user_pages(page_list, npages); in iopt_pages_err_unpin() 189 iopt_pages_sub_npinned(pages, npages); in iopt_pages_err_unpin() 616 size_t npages) in batch_from_pages() argument 618 struct page **end = pages + npages; in batch_from_pages() 637 while (npages) { in batch_unpin() 647 npages -= to_unpin; in batch_unpin() 742 unsigned long npages; in pfn_reader_user_pin() local 814 new_pages = cur_pages + npages; in incr_user_locked_vm() 869 decr_user_locked_vm(pages, npages); in do_update_pinned() 906 unsigned long npages; in pfn_reader_user_update_pinned() local [all …]
|
/openbmc/linux/arch/alpha/kernel/ |
H A D | pci_iommu.c | 238 long npages, dma_ofs, i; in pci_map_single_1() local 281 npages = iommu_num_pages(paddr, size, PAGE_SIZE); in pci_map_single_1() 294 for (i = 0; i < npages; ++i, paddr += PAGE_SIZE) in pci_map_single_1() 359 long dma_ofs, npages; in alpha_pci_unmap_page() local 396 iommu_arena_free(arena, dma_ofs, npages); in alpha_pci_unmap_page() 543 long npages, dma_ofs, i; in sg_fill() local 576 npages = iommu_num_pages(paddr, size, PAGE_SIZE); in sg_fill() 615 for (i = 0; i < npages; ++i, paddr += PAGE_SIZE) in sg_fill() 621 last_sg->length, npages); in sg_fill() 745 long npages, ofs; in alpha_pci_unmap_sg() local [all …]
|
/openbmc/linux/arch/powerpc/platforms/pasemi/ |
H A D | iommu.c | 78 long npages, unsigned long uaddr, in iobmap_build() argument 86 pr_debug("iobmap: build at: %lx, %lx, addr: %lx\n", index, npages, uaddr); in iobmap_build() 92 while (npages--) { in iobmap_build() 107 long npages) in iobmap_free() argument 112 pr_debug("iobmap: free at: %lx, %lx\n", index, npages); in iobmap_free() 118 while (npages--) { in iobmap_free()
|
/openbmc/linux/arch/powerpc/include/asm/ |
H A D | ultravisor.h | 34 static inline int uv_share_page(u64 pfn, u64 npages) in uv_share_page() argument 36 return ucall_norets(UV_SHARE_PAGE, pfn, npages); in uv_share_page() 39 static inline int uv_unshare_page(u64 pfn, u64 npages) in uv_unshare_page() argument 41 return ucall_norets(UV_UNSHARE_PAGE, pfn, npages); in uv_unshare_page()
|
H A D | kvm_book3s_64.h | 495 return !(memslot->base_gfn & mask) && !(memslot->npages & mask); in slot_is_aligned() 562 unsigned long npages) in set_dirty_bits() argument 565 if (npages >= 8) in set_dirty_bits() 566 memset((char *)map + i / 8, 0xff, npages / 8); in set_dirty_bits() 568 for (; npages; ++i, --npages) in set_dirty_bits() 573 unsigned long npages) in set_dirty_bits_atomic() argument 575 if (npages >= 8) in set_dirty_bits_atomic() 576 memset((char *)map + i / 8, 0xff, npages / 8); in set_dirty_bits_atomic() 578 for (; npages; ++i, --npages) in set_dirty_bits_atomic()
|
/openbmc/linux/drivers/infiniband/hw/usnic/ |
H A D | usnic_uiom.c | 95 unsigned long npages; in usnic_uiom_get_pages() local 121 npages = PAGE_ALIGN(size + (addr & ~PAGE_MASK)) >> PAGE_SHIFT; in usnic_uiom_get_pages() 139 while (npages) { in usnic_uiom_get_pages() 141 min_t(unsigned long, npages, in usnic_uiom_get_pages() 148 npages -= ret; in usnic_uiom_get_pages() 181 atomic64_sub(npages, ¤t->mm->pinned_vm); in usnic_uiom_get_pages() 213 int npages; in __usnic_uiom_reg_release() local 221 vpn_last = vpn_start + npages - 1; in __usnic_uiom_reg_release() 335 unsigned long npages; in usnic_uiom_reg_get() local 350 npages = PAGE_ALIGN(size + offset) >> PAGE_SHIFT; in usnic_uiom_reg_get() [all …]
|
/openbmc/linux/fs/nfs/ |
H A D | nfs3acl.c | 201 unsigned int npages = 1 + ((args.len - 1) >> PAGE_SHIFT); in __nfs3_proc_setacls() local 205 args.pages[args.npages] = alloc_page(GFP_KERNEL); in __nfs3_proc_setacls() 206 if (args.pages[args.npages] == NULL) in __nfs3_proc_setacls() 208 args.npages++; in __nfs3_proc_setacls() 209 } while (args.npages < npages); in __nfs3_proc_setacls() 240 while (args.npages != 0) { in __nfs3_proc_setacls() 241 args.npages--; in __nfs3_proc_setacls() 242 __free_page(args.pages[args.npages]); in __nfs3_proc_setacls()
|
/openbmc/linux/arch/sparc/kernel/ |
H A D | iommu-common.c | 99 unsigned long npages, in iommu_tbl_range_alloc() argument 112 bool largealloc = (large_pool && npages > iommu_large_alloc); in iommu_tbl_range_alloc() 120 if (unlikely(npages == 0)) { in iommu_tbl_range_alloc() 181 n = iommu_area_alloc(iommu->map, limit, start, npages, shift, in iommu_tbl_range_alloc() 211 end = n + npages; in iommu_tbl_range_alloc() 247 unsigned long npages, unsigned long entry) in iommu_tbl_range_free() argument 258 bitmap_clear(iommu->map, entry, npages); in iommu_tbl_range_free()
|
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_gem_object.c | 32 unsigned int npages; /* restricted by sg_alloc_table */ in huge_get_pages() local 37 if (overflows_type(obj->base.size / PAGE_SIZE, npages)) in huge_get_pages() 40 npages = obj->base.size / PAGE_SIZE; in huge_get_pages() 45 if (sg_alloc_table(pages, npages, GFP)) { in huge_get_pages() 63 if (nreal < npages) { in huge_get_pages()
|
/openbmc/linux/drivers/infiniband/hw/mlx4/ |
H A D | mr.c | 98 int *start_index, int *npages) in mlx4_ib_umem_write_mtt_block() argument 128 (*npages)++; in mlx4_ib_umem_write_mtt_block() 133 if (*npages == PAGE_SIZE / sizeof(u64)) { in mlx4_ib_umem_write_mtt_block() 135 *npages, pages); in mlx4_ib_umem_write_mtt_block() 139 (*start_index) += *npages; in mlx4_ib_umem_write_mtt_block() 140 *npages = 0; in mlx4_ib_umem_write_mtt_block() 192 int npages = 0; in mlx4_ib_umem_write_mtt() local 218 &npages); in mlx4_ib_umem_write_mtt() 235 &start_index, &npages); in mlx4_ib_umem_write_mtt() 240 if (npages) in mlx4_ib_umem_write_mtt() [all …]
|
/openbmc/linux/fs/ramfs/ |
H A D | file-nommu.c | 64 unsigned long npages, xpages, loop; in ramfs_nommu_expand_for_mapping() local 90 npages = (newsize + PAGE_SIZE - 1) >> PAGE_SHIFT; in ramfs_nommu_expand_for_mapping() 95 for (loop = npages; loop < xpages; loop++) in ramfs_nommu_expand_for_mapping() 99 newsize = PAGE_SIZE * npages; in ramfs_nommu_expand_for_mapping() 104 for (loop = 0; loop < npages; loop++) { in ramfs_nommu_expand_for_mapping() 123 while (loop < npages) in ramfs_nommu_expand_for_mapping()
|
/openbmc/linux/arch/powerpc/kvm/ |
H A D | book3s_64_vio.c | 211 for (i = 0; i < npages; i++) in release_spapr_tce_table() 305 unsigned long npages; in kvm_vm_ioctl_create_spapr_tce() local 312 npages = kvmppc_tce_pages(args->size); in kvm_vm_ioctl_create_spapr_tce() 641 unsigned long tce_list, unsigned long npages) in kvmppc_h_put_tce_indirect() argument 659 if (npages > 512) in kvmppc_h_put_tce_indirect() 665 ret = kvmppc_ioba_validate(stt, ioba, npages); in kvmppc_h_put_tce_indirect() 676 for (i = 0; i < npages; ++i) { in kvmppc_h_put_tce_indirect() 688 for (i = 0; i < npages; ++i) { in kvmppc_h_put_tce_indirect() 734 unsigned long tce_value, unsigned long npages) in kvmppc_h_stuff_tce() argument 744 ret = kvmppc_ioba_validate(stt, ioba, npages); in kvmppc_h_stuff_tce() [all …]
|
/openbmc/linux/fs/sysv/ |
H A D | dir.c | 76 unsigned long npages = dir_pages(inode); in sysv_readdir() local 87 for ( ; n < npages; n++, offset = 0) { in sysv_readdir() 145 unsigned long npages = dir_pages(dir); in sysv_find_entry() local 152 if (start >= npages) in sysv_find_entry() 172 if (++n >= npages) in sysv_find_entry() 191 unsigned long npages = dir_pages(dir); in sysv_add_link() local 198 for (n = 0; n <= npages; n++) { in sysv_add_link() 296 unsigned long i, npages = dir_pages(inode); in sysv_empty_dir() local 299 for (i = 0; i < npages; i++) { in sysv_empty_dir()
|
/openbmc/linux/mm/kmsan/ |
H A D | kmsan_test.c | 308 const int npages = 2; in test_init_kmsan_vmap_vunmap() local 314 pages = kmalloc_array(npages, sizeof(*pages), GFP_KERNEL); in test_init_kmsan_vmap_vunmap() 315 for (int i = 0; i < npages; i++) in test_init_kmsan_vmap_vunmap() 317 vbuf = vmap(pages, npages, VM_MAP, PAGE_KERNEL); in test_init_kmsan_vmap_vunmap() 318 memset(vbuf, 0xfe, npages * PAGE_SIZE); in test_init_kmsan_vmap_vunmap() 319 for (int i = 0; i < npages; i++) in test_init_kmsan_vmap_vunmap() 324 for (int i = 0; i < npages; i++) { in test_init_kmsan_vmap_vunmap() 339 int npages = 8; in test_init_vmalloc() local 343 buf = vmalloc(PAGE_SIZE * npages); in test_init_vmalloc() 345 memset(buf, 0xfe, PAGE_SIZE * npages); in test_init_vmalloc() [all …]
|
/openbmc/linux/arch/sparc/mm/ |
H A D | io-unit.c | 99 int i, j, k, npages; in iounit_get_area() local 103 npages = ((vaddr & ~PAGE_MASK) + size + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_get_area() 106 switch (npages) { in iounit_get_area() 112 IOD(("iounit_get_area(%08lx,%d[%d])=", vaddr, size, npages)); in iounit_get_area() 119 if (scan + npages > limit) { in iounit_get_area() 130 for (k = 1, scan++; k < npages; k++) in iounit_get_area() 134 scan -= npages; in iounit_get_area() 137 for (k = 0; k < npages; k++, iopte = __iopte(iopte_val(iopte) + 0x100), scan++) { in iounit_get_area()
|
/openbmc/linux/drivers/firmware/efi/ |
H A D | efi-init.c | 152 u64 paddr, npages, size; in reserve_regions() local 167 npages = md->num_pages; in reserve_regions() 173 paddr, paddr + (npages << EFI_PAGE_SHIFT) - 1, in reserve_regions() 177 memrange_efi_to_native(&paddr, &npages); in reserve_regions() 178 size = npages << PAGE_SHIFT; in reserve_regions()
|
/openbmc/linux/drivers/vfio/pci/pds/ |
H A D | lm.c | 18 unsigned long long npages; in pds_vfio_get_lm_file() local 44 npages = DIV_ROUND_UP_ULL(size, PAGE_SIZE); in pds_vfio_get_lm_file() 45 pages = kmalloc_array(npages, sizeof(*pages), GFP_KERNEL); in pds_vfio_get_lm_file() 54 for (unsigned long long i = 0; i < npages; i++) { in pds_vfio_get_lm_file() 66 if (sg_alloc_table_from_pages(&lm_file->sg_table, pages, npages, 0, in pds_vfio_get_lm_file() 72 lm_file->npages = npages; in pds_vfio_get_lm_file() 74 lm_file->alloc_size = npages * PAGE_SIZE; in pds_vfio_get_lm_file() 269 lm_file->size, lm_file->alloc_size, lm_file->npages); in pds_vfio_get_save_file()
|
/openbmc/linux/drivers/gpu/drm/gma500/ |
H A D | gem.c | 35 unsigned int npages; in psb_gem_pin() local 51 npages = obj->size / PAGE_SIZE; in psb_gem_pin() 53 set_pages_array_wc(pages, npages); in psb_gem_pin() 57 (gpu_base + pobj->offset), npages, 0, 0, in psb_gem_pin() 79 unsigned long npages; in psb_gem_unpin() local 93 npages = obj->size / PAGE_SIZE; in psb_gem_unpin() 96 (gpu_base + pobj->offset), npages, 0, 0); in psb_gem_unpin() 100 set_pages_array_wb(pobj->pages, npages); in psb_gem_unpin()
|