/openbmc/linux/mm/ |
H A D | mmu_gather.c | 27 if (batch->next) { in tlb_next_batch() 36 if (!batch) in tlb_next_batch() 41 batch->nr = 0; in tlb_next_batch() 89 for (batch = &tlb->local; batch && batch->nr; batch = batch->next) { in tlb_batch_pages_flush() 112 for (batch = tlb->local.next; batch; batch = next) { in tlb_batch_list_free() 134 batch->encoded_pages[batch->nr++] = page; in __tlb_remove_page_size() 135 if (batch->nr == batch->max) { in __tlb_remove_page_size() 249 struct mmu_table_batch **batch = &tlb->batch; in tlb_table_flush() local 251 if (*batch) { in tlb_table_flush() 260 struct mmu_table_batch **batch = &tlb->batch; in tlb_remove_table() local [all …]
|
/openbmc/linux/arch/powerpc/mm/book3s64/ |
H A D | hash_tlb.c | 51 i = batch->index; in hpte_need_flush() 103 if (!batch->active) { in hpte_need_flush() 119 if (i != 0 && (mm != batch->mm || batch->psize != psize || in hpte_need_flush() 125 batch->mm = mm; in hpte_need_flush() 129 batch->pte[i] = rpte; in hpte_need_flush() 130 batch->vpn[i] = vpn; in hpte_need_flush() 131 batch->index = ++i; in hpte_need_flush() 148 i = batch->index; in __flush_tlb_pending() 151 flush_hash_page(batch->vpn[0], batch->pte[0], in __flush_tlb_pending() 152 batch->psize, batch->ssize, local); in __flush_tlb_pending() [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/selftests/ |
H A D | igt_spinner.c | 97 if (!spin->batch) { in igt_spinner_pin() 105 spin->batch = vaddr; in igt_spinner_pin() 131 u32 *batch; in igt_spinner_create_request() local 139 if (!spin->batch) { in igt_spinner_create_request() 160 batch = spin->batch; in igt_spinner_create_request() 168 *batch++ = 0; in igt_spinner_create_request() 172 *batch++ = 0; in igt_spinner_create_request() 178 *batch++ = rq->fence.seqno; in igt_spinner_create_request() 180 *batch++ = arbitration_command; in igt_spinner_create_request() 226 if (!spin->batch) in igt_spinner_end() [all …]
|
/openbmc/linux/drivers/iommu/iommufd/ |
H A D | pages.c | 294 batch->npfns[batch->end - 1] < keep_pfns); in batch_clear_carry() 297 batch->pfns[0] = batch->pfns[batch->end - 1] + in batch_clear_carry() 298 (batch->npfns[batch->end - 1] - keep_pfns); in batch_clear_carry() 308 WARN_ON(batch->total_pfns != batch->npfns[0]); in batch_skip_carry() 327 batch->npfns = (u32 *)(batch->pfns + batch->array_size); in __batch_init() 355 pfn == batch->pfns[batch->end - 1] + batch->npfns[batch->end - 1] && in batch_add_pfn() 357 batch->npfns[batch->end - 1]++; in batch_add_pfn() 361 if (batch->end == batch->array_size) in batch_add_pfn() 364 batch->pfns[batch->end] = pfn; in batch_add_pfn() 365 batch->npfns[batch->end] = 1; in batch_add_pfn() [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gt/ |
H A D | gen7_renderclear.c | 257 batch_advance(batch, cs); in gen7_emit_state_base_address() 286 batch_advance(batch, cs); in gen7_emit_vfe_state() 294 u32 *cs = batch_alloc_items(batch, 8, 4); in gen7_emit_interface_descriptor_load() 305 batch_advance(batch, cs); in gen7_emit_interface_descriptor_load() 317 cs = batch_alloc_items(batch, 8, pkt); in gen7_emit_media_object() 337 batch_advance(batch, cs); in gen7_emit_media_object() 352 batch_advance(batch, cs); in gen7_emit_pipeline_flush() 373 batch_advance(batch, cs); in gen7_emit_pipeline_invalidate() 435 u32 *batch; in gen7_setup_clear_gpr_bb() local 444 if (IS_ERR(batch)) in gen7_setup_clear_gpr_bb() [all …]
|
H A D | intel_lrc.c | 1553 *batch++ = 0; in gen8_emit_flush_coherentl3_wa() 1559 batch = gen8_emit_pipe_control(batch, in gen8_emit_flush_coherentl3_wa() 1568 *batch++ = 0; in gen8_emit_flush_coherentl3_wa() 1570 return batch; in gen8_emit_flush_coherentl3_wa() 1595 batch = gen8_emit_flush_coherentl3_wa(engine, batch); in gen8_init_indirectctx_bb() 1599 batch = gen8_emit_pipe_control(batch, in gen8_init_indirectctx_bb() 1618 return batch; in gen8_init_indirectctx_bb() 1637 return batch; in emit_lri() 1668 batch = gen8_emit_flush_coherentl3_wa(engine, batch); in gen9_init_indirectctx_bb() 1671 batch = gen8_emit_pipe_control(batch, in gen9_init_indirectctx_bb() [all …]
|
H A D | gen8_engine_cs.h | 53 __gen8_emit_pipe_control(u32 *batch, u32 bit_group_0, in __gen8_emit_pipe_control() argument 56 memset(batch, 0, 6 * sizeof(u32)); in __gen8_emit_pipe_control() 58 batch[0] = GFX_OP_PIPE_CONTROL(6) | bit_group_0; in __gen8_emit_pipe_control() 59 batch[1] = bit_group_1; in __gen8_emit_pipe_control() 60 batch[2] = offset; in __gen8_emit_pipe_control() 62 return batch + 6; in __gen8_emit_pipe_control() 65 static inline u32 *gen8_emit_pipe_control(u32 *batch, in gen8_emit_pipe_control() argument 68 return __gen8_emit_pipe_control(batch, 0, bit_group_1, offset); in gen8_emit_pipe_control() 71 static inline u32 *gen12_emit_pipe_control(u32 *batch, u32 bit_group_0, in gen12_emit_pipe_control() argument 74 return __gen8_emit_pipe_control(batch, bit_group_0, in gen12_emit_pipe_control()
|
H A D | selftest_hangcheck.c | 38 u32 *batch; member 81 h->batch = vaddr; in hang_init() 113 u32 *batch; in hang_create_request() local 133 h->batch = vaddr; in hang_create_request() 171 batch = h->batch; in hang_create_request() 180 batch += 1024 / sizeof(*batch); in hang_create_request() 188 *batch++ = 0; in hang_create_request() 194 batch += 1024 / sizeof(*batch); in hang_create_request() 201 *batch++ = 0; in hang_create_request() 207 batch += 1024 / sizeof(*batch); in hang_create_request() [all …]
|
H A D | selftest_lrc.c | 971 if (IS_ERR(batch)) in store_context() 972 return batch; in store_context() 976 i915_vma_put(batch); in store_context() 983 i915_vma_put(batch); in store_context() 1047 return batch; in store_context() 1142 if (IS_ERR(batch)) in load_context() 1143 return batch; in load_context() 1147 i915_vma_put(batch); in load_context() 1205 return batch; in load_context() 1216 if (IS_ERR(batch)) in poison_registers() [all …]
|
H A D | selftest_tlb.c | 42 struct drm_i915_gem_object *batch; in pte_tlbinv() local 50 batch = i915_gem_object_create_internal(ce->vm->i915, 4096); in pte_tlbinv() 51 if (IS_ERR(batch)) in pte_tlbinv() 52 return PTR_ERR(batch); in pte_tlbinv() 54 vma = i915_vma_instance(batch, ce->vm, NULL); in pte_tlbinv() 100 cs = i915_gem_object_pin_map_unlocked(batch, I915_MAP_WC); in pte_tlbinv() 122 i915_gem_object_flush_map(batch); in pte_tlbinv() 177 cs = page_mask_bits(batch->mm.mapping); in pte_tlbinv() 188 i915_gem_object_put(batch); in pte_tlbinv()
|
H A D | selftest_workarounds.c | 504 struct i915_vma *batch; in check_dirty_whitelist() local 513 batch = create_batch(ce->vm); in check_dirty_whitelist() 514 if (IS_ERR(batch)) { in check_dirty_whitelist() 515 err = PTR_ERR(batch); in check_dirty_whitelist() 750 i915_vma_unpin_and_release(&batch, 0); in check_dirty_whitelist() 890 struct i915_vma *batch; in scrub_whitelisted_registers() local 894 batch = create_batch(ce->vm); in scrub_whitelisted_registers() 895 if (IS_ERR(batch)) in scrub_whitelisted_registers() 896 return PTR_ERR(batch); in scrub_whitelisted_registers() 919 i915_gem_object_flush_map(batch->obj); in scrub_whitelisted_registers() [all …]
|
H A D | selftest_engine_cs.c | 144 struct i915_vma *batch; in perf_mi_bb_start() local 153 batch = create_empty_batch(ce); in perf_mi_bb_start() 154 if (IS_ERR(batch)) { in perf_mi_bb_start() 155 err = PTR_ERR(batch); in perf_mi_bb_start() 160 err = i915_vma_sync(batch); in perf_mi_bb_start() 163 i915_vma_put(batch); in perf_mi_bb_start() 181 i915_vma_offset(batch), 8, in perf_mi_bb_start() 202 i915_vma_put(batch); in perf_mi_bb_start()
|
/openbmc/linux/arch/powerpc/include/asm/book3s/64/ |
H A D | tlbflush-hash.h | 25 extern void __flush_tlb_pending(struct ppc64_tlb_batch *batch); 31 struct ppc64_tlb_batch *batch; in arch_enter_lazy_mmu_mode() local 40 batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_enter_lazy_mmu_mode() 41 batch->active = 1; in arch_enter_lazy_mmu_mode() 46 struct ppc64_tlb_batch *batch; in arch_leave_lazy_mmu_mode() local 50 batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_leave_lazy_mmu_mode() 52 if (batch->index) in arch_leave_lazy_mmu_mode() 53 __flush_tlb_pending(batch); in arch_leave_lazy_mmu_mode() 54 batch->active = 0; in arch_leave_lazy_mmu_mode()
|
/openbmc/qemu/tests/avocado/acpi-bits/bits-config/ |
H A D | bits-cfg.txt | 4 # To run BITS in batch mode, set batch to a list of one or more of the 12 # Leave batch set to an empty string to disable batch mode. 13 # batch = 15 # Uncomment the following to run all available batch operations 18 batch = test acpi smbios
|
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/ |
H A D | igt_gem_utils.c | 116 struct i915_vma *batch; in igt_gpu_fill_dw() local 123 batch = igt_emit_store_dw(vma, offset, count, val); in igt_gpu_fill_dw() 124 if (IS_ERR(batch)) in igt_gpu_fill_dw() 125 return PTR_ERR(batch); in igt_gpu_fill_dw() 133 err = igt_vma_move_to_active_unlocked(batch, rq, 0); in igt_gpu_fill_dw() 146 i915_vma_offset(batch), in igt_gpu_fill_dw() 147 i915_vma_size(batch), in igt_gpu_fill_dw() 155 i915_vma_unpin_and_release(&batch, 0); in igt_gpu_fill_dw()
|
H A D | i915_gem_client_blt.c | 104 struct i915_vma *batch; member 143 struct drm_i915_gem_object *batch) in prepare_blit() argument 254 i915_gem_object_flush_map(batch); in prepare_blit() 255 i915_gem_object_unpin_map(batch); in prepare_blit() 268 i915_vma_put(t->batch); in tiled_blits_destroy_buffers() 308 if (IS_ERR(t->batch)) in tiled_blits_create_buffers() 309 return PTR_ERR(t->batch); in tiled_blits_create_buffers() 313 i915_vma_put(t->batch); in tiled_blits_create_buffers() 521 i915_vma_offset(t->batch), in tiled_blit() 522 i915_vma_size(t->batch), in tiled_blit() [all …]
|
/openbmc/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_mob.c | 238 struct vmw_otable_batch *batch) in vmw_otable_batch_setup() argument 257 &batch->otable_bo); in vmw_otable_batch_setup() 263 if (!batch->otables[i].enabled) in vmw_otable_batch_setup() 267 &batch->otable_bo->tbo, in vmw_otable_batch_setup() 279 if (batch->otables[i].enabled) in vmw_otable_batch_setup() 281 &batch->otables[i]); in vmw_otable_batch_setup() 285 ttm_bo_put(&batch->otable_bo->tbo); in vmw_otable_batch_setup() 286 batch->otable_bo = NULL; in vmw_otable_batch_setup() 340 if (batch->otables[i].enabled) in vmw_otable_batch_takedown() 342 &batch->otables[i]); in vmw_otable_batch_takedown() [all …]
|
/openbmc/linux/drivers/xen/ |
H A D | gntdev.c | 811 batch->pages[batch->nr_pages++] = page; in gntdev_get_page() 821 unpin_user_pages_dirty_lock(batch->pages, batch->nr_pages, batch->writeable); in gntdev_put_pages() 822 batch->nr_pages = 0; in gntdev_put_pages() 830 gnttab_batch_copy(batch->ops, batch->nr_ops); in gntdev_copy() 854 batch->nr_ops = 0; in gntdev_copy() 900 op = &batch->ops[batch->nr_ops]; in gntdev_grant_copy_seg() 946 batch->status[batch->nr_ops] = status; in gntdev_grant_copy_seg() 947 batch->nr_ops++; in gntdev_grant_copy_seg() 963 batch.nr_ops = 0; in gntdev_ioctl_grant_copy() 964 batch.nr_pages = 0; in gntdev_ioctl_grant_copy() [all …]
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | test_bpf_ma.c | 42 static __always_inline void batch_alloc_free(struct bpf_map *map, unsigned int batch, in batch_alloc_free() argument 49 for (i = 0; i < batch; i++) { in batch_alloc_free() 68 for (i = 0; i < batch; i++) { in batch_alloc_free() 84 #define CALL_BATCH_ALLOC_FREE(size, batch, idx) \ argument 85 batch_alloc_free((struct bpf_map *)(&array_##size), batch, idx)
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | rx.c | 26 int batch, i; in mlx5e_xsk_alloc_rx_mpwqe() local 35 batch = xsk_buff_alloc_batch(rq->xsk_pool, xsk_buffs, in mlx5e_xsk_alloc_rx_mpwqe() 44 for (; batch < rq->mpwqe.pages_per_wqe; batch++) { in mlx5e_xsk_alloc_rx_mpwqe() 45 xsk_buffs[batch] = xsk_buff_alloc(rq->xsk_pool); in mlx5e_xsk_alloc_rx_mpwqe() 46 if (unlikely(!xsk_buffs[batch])) in mlx5e_xsk_alloc_rx_mpwqe() 55 for (i = 0; i < batch; i++) { in mlx5e_xsk_alloc_rx_mpwqe() 65 for (i = 0; i < batch; i++) { in mlx5e_xsk_alloc_rx_mpwqe() 78 for (i = 0; i < batch; i++) { in mlx5e_xsk_alloc_rx_mpwqe() 105 for (i = 0; i < batch; i++) { in mlx5e_xsk_alloc_rx_mpwqe() 152 while (--batch >= 0) in mlx5e_xsk_alloc_rx_mpwqe() [all …]
|
/openbmc/linux/tools/testing/selftests/bpf/map_tests/ |
H A D | htab_map_batch_ops.c | 79 __u32 batch, count, total, total_success; in __test_map_lookup_and_delete_batch() local 109 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 119 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 127 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 153 total ? &batch : NULL, in __test_map_lookup_and_delete_batch() 154 &batch, keys + total, in __test_map_lookup_and_delete_batch() 216 total ? &batch : NULL, in __test_map_lookup_and_delete_batch() 217 &batch, keys + total, in __test_map_lookup_and_delete_batch()
|
/openbmc/linux/tools/virtio/ |
H A D | virtio_test.c | 170 bool delayed, int batch, int reset_n, int bufs) in run_test() argument 178 const bool random_batch = batch == RANDOM_BATCH; in run_test() 193 batch = (random() % vq->vring.num) + 1; in run_test() 196 (started - completed) < batch) { in run_test() 349 long batch = 1, reset = 0; in main() local 376 batch = RANDOM_BATCH; in main() 378 batch = strtol(optarg, NULL, 10); in main() 379 assert(batch > 0); in main() 380 assert(batch < (long)INT_MAX + 1); in main() 401 run_test(&dev, &dev.vqs[0], delayed, batch, reset, 0x100000); in main()
|
/openbmc/linux/drivers/net/ethernet/netronome/nfp/flower/ |
H A D | lag_conf.c | 234 unsigned int member_cnt, enum nfp_fl_lag_batch *batch) in nfp_fl_lag_config_group() argument 254 if (*batch == NFP_FL_LAG_BATCH_FIRST) { in nfp_fl_lag_config_group() 257 *batch = NFP_FL_LAG_BATCH_MEMBER; in nfp_fl_lag_config_group() 263 *batch = NFP_FL_LAG_BATCH_FINISHED; in nfp_fl_lag_config_group() 269 if (*batch == NFP_FL_LAG_BATCH_FINISHED) { in nfp_fl_lag_config_group() 296 enum nfp_fl_lag_batch batch = NFP_FL_LAG_BATCH_FIRST; in nfp_fl_lag_do_work() local 318 &batch); in nfp_fl_lag_do_work() 392 active_count, &batch); in nfp_fl_lag_do_work() 406 if (batch == NFP_FL_LAG_BATCH_MEMBER) { in nfp_fl_lag_do_work() 407 batch = NFP_FL_LAG_BATCH_FINISHED; in nfp_fl_lag_do_work() [all …]
|
/openbmc/linux/net/core/ |
H A D | netclassid_cgroup.c | 64 unsigned int batch; member 76 if (--ctx->batch == 0) { in update_classid_sock() 77 ctx->batch = UPDATE_CLASSID_BATCH; in update_classid_sock() 87 .batch = UPDATE_CLASSID_BATCH in update_classid_task()
|
/openbmc/linux/tools/virtio/ringtest/ |
H A D | main.c | 22 int batch = 1; variable 116 int tokick = batch; in run_guest() 129 tokick = batch; in run_guest() 348 batch = c; in main() 372 if (batch > max_outstanding) in main() 373 batch = max_outstanding; in main()
|