/openbmc/linux/drivers/media/pci/ivtv/ |
H A D | ivtv-udma.c | 104 dma->SG_length, dma->page_count); in ivtv_udma_setup() 110 if (user_dma.page_count <= 0) { in ivtv_udma_setup() 120 if (user_dma.page_count != err) { in ivtv_udma_setup() 122 err, user_dma.page_count); in ivtv_udma_setup() 130 dma->page_count = user_dma.page_count; in ivtv_udma_setup() 137 dma->page_count = 0; in ivtv_udma_setup() 143 dma->page_count, DMA_TO_DEVICE); in ivtv_udma_setup() 147 dma->page_count = 0; in ivtv_udma_setup() 158 return dma->page_count; in ivtv_udma_setup() 168 if (dma->page_count == 0) in ivtv_udma_unmap() [all …]
|
H A D | ivtv-yuv.c | 54 if (dma->SG_length || dma->page_count) { in ivtv_yuv_prep_user_dma() 57 dma->SG_length, dma->page_count); in ivtv_yuv_prep_user_dma() 66 y_dma.page_count, &dma->map[0], 0); in ivtv_yuv_prep_user_dma() 68 if (y_pages == y_dma.page_count) { in ivtv_yuv_prep_user_dma() 73 if (y_pages != y_dma.page_count || uv_pages != uv_dma.page_count) { in ivtv_yuv_prep_user_dma() 76 if (y_pages == y_dma.page_count) { in ivtv_yuv_prep_user_dma() 79 uv_pages, uv_dma.page_count); in ivtv_yuv_prep_user_dma() 90 y_pages, y_dma.page_count); in ivtv_yuv_prep_user_dma() 106 dma->page_count = y_pages + uv_pages; in ivtv_yuv_prep_user_dma() 112 dma->page_count = 0; in ivtv_yuv_prep_user_dma() [all …]
|
/openbmc/linux/tools/testing/selftests/mm/ |
H A D | ksm_tests.c | 367 if (assert_ksm_pages_count(page_count)) { in check_ksm_merge() 369 munmap(map_ptr, page_size * page_count); in check_ksm_merge() 377 munmap(map_ptr, page_size * page_count); in check_ksm_merge() 385 int page_count = 2; in check_ksm_unmerge() local 417 munmap(map_ptr, page_size * page_count); in check_ksm_unmerge() 456 munmap(map_ptr, page_size * page_count); in check_ksm_zero_page_merge() 461 munmap(map_ptr, page_size * page_count); in check_ksm_zero_page_merge() 491 int page_count = 2; in check_ksm_numa_merge() local 713 size_t page_count = 4000; in ksm_cow_time() local 797 page_count = atol(optarg); in main() [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_bo85b5.c | 47 u32 page_count = PFN_UP(new_reg->size); in nva3_bo_move_copy() local 50 page_count = PFN_UP(new_reg->size); in nva3_bo_move_copy() 51 while (page_count) { in nva3_bo_move_copy() 52 int line_count = (page_count > 8191) ? 8191 : page_count; in nva3_bo_move_copy() 68 page_count -= line_count; in nva3_bo_move_copy()
|
H A D | nouveau_bo90b5.c | 40 u32 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_copy() local 43 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_copy() 44 while (page_count) { in nvc0_bo_move_copy() 45 int line_count = (page_count > 8191) ? 8191 : page_count; in nvc0_bo_move_copy() 61 page_count -= line_count; in nvc0_bo_move_copy()
|
H A D | nouveau_bo9039.c | 45 u32 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_m2mf() local 48 page_count = PFN_UP(new_reg->size); in nvc0_bo_move_m2mf() 49 while (page_count) { in nvc0_bo_move_m2mf() 50 int line_count = (page_count > 2047) ? 2047 : page_count; in nvc0_bo_move_m2mf() 78 page_count -= line_count; in nvc0_bo_move_m2mf()
|
H A D | nouveau_bo0039.c | 55 u32 page_count = PFN_UP(new_reg->size); in nv04_bo_move_m2mf() local 65 page_count = PFN_UP(new_reg->size); in nv04_bo_move_m2mf() 66 while (page_count) { in nv04_bo_move_m2mf() 67 int line_count = (page_count > 2047) ? 2047 : page_count; in nv04_bo_move_m2mf() 88 page_count -= line_count; in nv04_bo_move_m2mf()
|
/openbmc/linux/drivers/firewire/ |
H A D | core-iso.c | 33 buffer->page_count = 0; in fw_iso_buffer_alloc() 40 for (i = 0; i < page_count; i++) { in fw_iso_buffer_alloc() 45 buffer->page_count = i; in fw_iso_buffer_alloc() 46 if (i < page_count) { in fw_iso_buffer_alloc() 62 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_map_dma() 71 if (i < buffer->page_count) in fw_iso_buffer_map_dma() 78 int page_count, enum dma_data_direction direction) in fw_iso_buffer_init() argument 82 ret = fw_iso_buffer_alloc(buffer, page_count); in fw_iso_buffer_init() 105 for (i = 0; i < buffer->page_count; i++) in fw_iso_buffer_destroy() 110 buffer->page_count = 0; in fw_iso_buffer_destroy() [all …]
|
/openbmc/linux/drivers/char/agp/ |
H A D | generic.c | 181 if (curr->page_count != 0) { in agp_free_memory() 186 for (i = 0; i < curr->page_count; i++) { in agp_free_memory() 219 size_t page_count, u32 type) in agp_allocate_memory() argument 231 (cur_memory + page_count < page_count)) in agp_allocate_memory() 264 for (i = 0; i < page_count; i++) { in agp_allocate_memory() 272 new->page_count++; in agp_allocate_memory() 1038 if (mem->page_count == 0) in agp_generic_insert_memory() 1117 if (mem->page_count == 0) in agp_generic_remove_memory() 1170 for (i = 0; i < page_count; i++) in agp_generic_alloc_user() 1172 new->page_count = 0; in agp_generic_alloc_user() [all …]
|
H A D | i460-agp.c | 311 if ((io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count) > num_entries) { in i460_insert_memory_small_io_page() 317 while (j < (io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count)) { in i460_insert_memory_small_io_page() 327 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { in i460_insert_memory_small_io_page() 346 for (i = pg_start; i < (pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count); i++) in i460_remove_memory_small_io_page() 415 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; in i460_insert_memory_large_io_page() 417 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; in i460_insert_memory_large_io_page() 473 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; in i460_remove_memory_large_io_page() 475 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; in i460_remove_memory_large_io_page()
|
H A D | nvidia-agp.c | 210 if (mem->page_count == 0) in nvidia_insert_memory() 213 if ((pg_start + mem->page_count) > in nvidia_insert_memory() 217 for (j = pg_start; j < (pg_start + mem->page_count); j++) { in nvidia_insert_memory() 226 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in nvidia_insert_memory() 250 if (mem->page_count == 0) in nvidia_remove_memory() 253 for (i = pg_start; i < (mem->page_count + pg_start); i++) in nvidia_remove_memory()
|
H A D | intel-gtt.c | 128 DBG("try unmapping %lu pages\n", (unsigned long)mem->page_count); in intel_gtt_unmap_memory() 218 if ((pg_start + mem->page_count) in i810_insert_dcache_entries() 225 for (i = pg_start; i < (pg_start + mem->page_count); i++) { in i810_insert_dcache_entries() 270 new->page_count = pg_count; in alloc_agpphysmem_i8xx() 281 if (curr->page_count == 4) in intel_i810_free_by_type() 912 if (mem->page_count == 0) in intel_fake_agp_insert_entries() 915 if (pg_start + mem->page_count > intel_private.gtt_total_entries) in intel_fake_agp_insert_entries() 930 ret = intel_gtt_map_memory(mem->pages, mem->page_count, &st); in intel_fake_agp_insert_entries() 965 if (mem->page_count == 0) in intel_fake_agp_remove_entries() 968 intel_gmch_gtt_clear_range(pg_start, mem->page_count); in intel_fake_agp_remove_entries() [all …]
|
H A D | ati-agp.c | 273 if (mem->page_count == 0) in ati_insert_memory() 276 if ((pg_start + mem->page_count) > num_entries) in ati_insert_memory() 280 while (j < (pg_start + mem->page_count)) { in ati_insert_memory() 294 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in ati_insert_memory() 319 if (mem->page_count == 0) in ati_remove_memory() 322 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in ati_remove_memory()
|
/openbmc/linux/drivers/vfio/pci/pds/ |
H A D | dirty.c | 66 le32_to_cpu(region_info[i].page_count), in pds_vfio_print_guest_region_info() 128 u32 page_count) in __pds_vfio_dirty_alloc_sgl() argument 137 max_sge = DIV_ROUND_UP(page_count, PAGE_SIZE * 8); in __pds_vfio_dirty_alloc_sgl() 158 u32 page_count) in pds_vfio_dirty_alloc_sgl() argument 164 page_count); in pds_vfio_dirty_alloc_sgl() 169 page_count); in pds_vfio_dirty_alloc_sgl() 191 u32 page_count; in pds_vfio_dirty_enable() local 251 region_info->page_count = cpu_to_le32(page_count); in pds_vfio_dirty_enable() 270 page_count = le32_to_cpu(region_info->page_count); in pds_vfio_dirty_enable() 274 regions_dma, region_start, page_count, in pds_vfio_dirty_enable() [all …]
|
/openbmc/u-boot/drivers/mtd/nand/raw/ |
H A D | mxc_nand_spl.c | 135 unsigned int page_count; in nfc_nand_page_address() local 143 page_count = CONFIG_SYS_NAND_SIZE / CONFIG_SYS_NAND_PAGE_SIZE; in nfc_nand_page_address() 145 if (page_address <= page_count) { in nfc_nand_page_address() 146 page_count--; /* transform 0x01000000 to 0x00ffffff */ in nfc_nand_page_address() 150 page_count = page_count >> 8; in nfc_nand_page_address() 151 } while (page_count); in nfc_nand_page_address()
|
/openbmc/linux/sound/soc/sof/amd/ |
H A D | acp-loader.c | 55 u32 dma_size, page_count; in acp_dsp_block_write() local 64 page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT; in acp_dsp_block_write() 65 dma_size = page_count * ACP_PAGE_SIZE; in acp_dsp_block_write() 157 u32 page_count, dma_size; in acp_dsp_pre_fw_run() local 167 page_count = PAGE_ALIGN(size_fw) >> PAGE_SHIFT; in acp_dsp_pre_fw_run() 168 adata->fw_bin_page_count = page_count; in acp_dsp_pre_fw_run() 170 configure_pte_for_fw_loading(FW_BIN, page_count, adata); in acp_dsp_pre_fw_run() 179 src_addr = ACP_SYSTEM_MEMORY_WINDOW + page_count * ACP_PAGE_SIZE; in acp_dsp_pre_fw_run()
|
/openbmc/linux/fs/squashfs/ |
H A D | block.c | 81 u64 read_start, u64 read_end, int page_count) in squashfs_bio_read_cached() argument 111 else if (idx == page_count - 1 && index + length != read_end) in squashfs_bio_read_cached() 200 const int page_count = DIV_ROUND_UP(total_len + offset, PAGE_SIZE); in squashfs_bio_read() local 204 bio = bio_kmalloc(page_count, GFP_NOIO); in squashfs_bio_read() 207 bio_init(bio, sb->s_bdev, bio->bi_inline_vecs, page_count, REQ_OP_READ); in squashfs_bio_read() 210 for (i = 0; i < page_count; ++i) { in squashfs_bio_read() 237 page_count); in squashfs_bio_read()
|
/openbmc/linux/arch/x86/hyperv/ |
H A D | hv_proc.c | 29 int i, j, page_count; in hv_call_deposit_pages() local 87 for (i = 0, page_count = 0; i < num_allocations; ++i) { in hv_call_deposit_pages() 89 for (j = 0; j < counts[i]; ++j, ++page_count) in hv_call_deposit_pages() 90 input_page->gpa_page_list[page_count] = base_pfn + j; in hv_call_deposit_pages() 93 page_count, 0, input_page, NULL); in hv_call_deposit_pages()
|
/openbmc/linux/drivers/target/ |
H A D | target_core_rd.c | 68 u32 i, j, page_count = 0, sg_per_table; in rd_release_sgl_table() local 78 page_count++; in rd_release_sgl_table() 85 return page_count; in rd_release_sgl_table() 90 u32 page_count; in rd_release_device_space() local 95 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_table_array, in rd_release_device_space() 100 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, in rd_release_device_space() 101 rd_dev->sg_table_count, (unsigned long)page_count * PAGE_SIZE); in rd_release_device_space() 216 u32 page_count; in rd_release_prot_space() local 221 page_count = rd_release_sgl_table(rd_dev, rd_dev->sg_prot_array, in rd_release_prot_space() 226 rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count, in rd_release_prot_space() [all …]
|
/openbmc/linux/fs/orangefs/ |
H A D | orangefs-bufmap.c | 152 int page_count; member 171 unpin_user_pages(bufmap->page_array, bufmap->page_count); in orangefs_bufmap_unmap() 239 bufmap->page_count = bufmap->total_size / PAGE_SIZE; in orangefs_bufmap_alloc() 243 kcalloc(bufmap->page_count, sizeof(struct page *), GFP_KERNEL); in orangefs_bufmap_alloc() 268 bufmap->page_count, FOLL_WRITE, bufmap->page_array); in orangefs_bufmap_map() 273 if (ret != bufmap->page_count) { in orangefs_bufmap_map() 275 bufmap->page_count, ret); in orangefs_bufmap_map() 290 for (i = 0; i < bufmap->page_count; i++) in orangefs_bufmap_map()
|
/openbmc/qemu/migration/ |
H A D | multifd.c | 142 p->iov = g_new0(struct iovec, p->page_count + 1); in nocomp_send_setup() 144 p->iov = g_new0(struct iovec, p->page_count); in nocomp_send_setup() 240 p->iov = g_new0(struct iovec, p->page_count); in nocomp_recv_setup() 475 if (packet->pages_alloc > p->page_count) { in multifd_recv_unfill_packet() 478 packet->pages_alloc, p->page_count) ; in multifd_recv_unfill_packet() 1184 p->pages = multifd_pages_init(page_count); in multifd_send_setup() 1195 p->page_count = page_count; in multifd_send_setup() 1610 + sizeof(uint64_t) * page_count; in multifd_recv_setup() 1614 p->normal = g_new0(ram_addr_t, page_count); in multifd_recv_setup() 1615 p->zero = g_new0(ram_addr_t, page_count); in multifd_recv_setup() [all …]
|
/openbmc/linux/tools/testing/radix-tree/ |
H A D | regression2.c | 61 unsigned long page_count = 0; variable 71 p->index = page_count++; in page_alloc()
|
/openbmc/linux/drivers/virt/vboxguest/ |
H A D | vboxguest_utils.c | 199 u32 page_count; in hgcm_call_add_pagelist_size() local 201 page_count = hgcm_call_buf_size_in_pages(buf, len); in hgcm_call_add_pagelist_size() 202 *extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]); in hgcm_call_add_pagelist_size() 337 u32 i, page_count; in hgcm_call_init_linaddr() local 348 page_count = hgcm_call_buf_size_in_pages(buf, len); in hgcm_call_init_linaddr() 356 dst_pg_lst->page_count = page_count; in hgcm_call_init_linaddr() 358 for (i = 0; i < page_count; i++) { in hgcm_call_init_linaddr() 368 *off_extra += offsetof(struct vmmdev_hgcm_pagelist, pages[page_count]); in hgcm_call_init_linaddr()
|
/openbmc/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shmem.c | 68 unsigned int page_count; /* restricted by sg_alloc_table */ in shmem_sg_alloc_table() local 75 if (overflows_type(size / PAGE_SIZE, page_count)) in shmem_sg_alloc_table() 78 page_count = size / PAGE_SIZE; in shmem_sg_alloc_table() 86 if (sg_alloc_table(st, page_count, GFP_KERNEL | __GFP_NOWARN)) in shmem_sg_alloc_table() 101 for (i = 0; i < page_count; i++) { in shmem_sg_alloc_table() 121 i915_gem_shrink(NULL, i915, 2 * page_count, NULL, *s++); in shmem_sg_alloc_table() 155 folio_nr_pages(folio), page_count - i); in shmem_sg_alloc_table()
|
/openbmc/linux/drivers/gpu/drm/ttm/ |
H A D | ttm_agp_backend.c | 65 mem->page_count = 0; in ttm_agp_bind() 72 mem->pages[mem->page_count++] = page; in ttm_agp_bind()
|