/openbmc/linux/drivers/gpu/drm/msm/ |
H A D | msm_gem_submit.c | 41 if (!submit) in submit_create() 47 kfree(submit); in submit_create() 54 kfree(submit); in submit_create() 62 submit->cmd = (void *)&submit->bos[nr_bos]; in submit_create() 73 return submit; in submit_create() 84 idr_remove(&submit->queue->fence_idr, submit->fence_id); in __msm_gem_submit_destroy() 109 kfree(submit); in __msm_gem_submit_destroy() 876 ret = submit_bo(submit, submit->cmd[i].idx, in msm_ioctl_gem_submit() 882 ((submit->cmd[i].size + submit->cmd[i].offset) > in msm_ioctl_gem_submit() 905 submit->cmd[i].nr_relocs, submit->cmd[i].relocs); in msm_ioctl_gem_submit() [all …]
|
H A D | msm_gpu.c | 279 if (submit) { in msm_gpu_crashstate_capture() 320 return submit; in find_submit() 369 if (submit) { in recover_worker() 371 submit->queue->faults++; in recover_worker() 372 if (submit->aspace) in recover_worker() 437 gpu->funcs->submit(gpu, submit); in recover_worker() 459 if (submit && submit->fault_dumped) in fault_worker() 462 if (submit) { in fault_worker() 716 if (submit && dma_fence_is_signaled(submit->hw_fence)) { in retire_submits() 758 submit->seqno = submit->hw_fence->seqno; in msm_gpu_submit() [all …]
|
H A D | msm_rd.c | 318 offset = iova - submit->bos[idx].iova; in snapshot_buf() 320 iova = submit->bos[idx].iova; in snapshot_buf() 377 pid_nr(submit->pid), submit->seqno); in msm_rd_dump_submit() 380 pid_nr(submit->pid), submit->seqno); in msm_rd_dump_submit() 386 for (i = 0; i < submit->nr_bos; i++) in msm_rd_dump_submit() 387 snapshot_buf(rd, submit, i, 0, 0, should_dump(submit, i)); in msm_rd_dump_submit() 389 for (i = 0; i < submit->nr_cmds; i++) { in msm_rd_dump_submit() 393 if (!should_dump(submit, i)) { in msm_rd_dump_submit() 394 snapshot_buf(rd, submit, submit->cmd[i].idx, in msm_rd_dump_submit() 400 uint64_t iova = submit->cmd[i].iova; in msm_rd_dump_submit() [all …]
|
H A D | msm_ringbuffer.c | 16 struct msm_gem_submit *submit = to_msm_submit(job); in msm_job_run() local 17 struct msm_fence_context *fctx = submit->ring->fctx; in msm_job_run() 18 struct msm_gpu *gpu = submit->gpu; in msm_job_run() 22 msm_fence_init(submit->hw_fence, fctx); in msm_job_run() 26 for (i = 0; i < submit->nr_bos; i++) { in msm_job_run() 27 struct drm_gem_object *obj = submit->bos[i].obj; in msm_job_run() 30 submit->bos[i].flags &= ~BO_PINNED; in msm_job_run() 38 msm_gpu_submit(gpu, submit); in msm_job_run() 42 return dma_fence_get(submit->hw_fence); in msm_job_run() 47 struct msm_gem_submit *submit = to_msm_submit(job); in msm_job_free() local [all …]
|
H A D | msm_gpu_trace.h | 34 TP_PROTO(struct msm_gem_submit *submit, u64 ticks), 35 TP_ARGS(submit, ticks), 44 __entry->pid = pid_nr(submit->pid); 45 __entry->id = submit->ident; 46 __entry->ringid = submit->ring->id; 47 __entry->seqno = submit->seqno; 59 TP_ARGS(submit, elapsed, clock, start, end), 71 __entry->pid = pid_nr(submit->pid); 72 __entry->id = submit->ident; 73 __entry->ringid = submit->ring->id; [all …]
|
H A D | msm_gem.h | 310 static inline void msm_gem_submit_get(struct msm_gem_submit *submit) in msm_gem_submit_get() argument 312 kref_get(&submit->ref); in msm_gem_submit_get() 315 static inline void msm_gem_submit_put(struct msm_gem_submit *submit) in msm_gem_submit_put() argument 317 kref_put(&submit->ref, __msm_gem_submit_destroy); in msm_gem_submit_put() 320 void msm_submit_retire(struct msm_gem_submit *submit); 326 should_dump(struct msm_gem_submit *submit, int idx) in should_dump() argument 329 return rd_full || (submit->bos[idx].flags & MSM_SUBMIT_BO_DUMP); in should_dump()
|
/openbmc/linux/drivers/gpu/drm/virtio/ |
H A D | virtgpu_submit.c | 51 u32 context = submit->fence_ctx + submit->ring_idx; in virtio_gpu_do_fence_wait() 332 virtio_gpu_free_syncobjs(submit->in_syncobjs, submit->num_in_syncobjs); in virtio_gpu_cleanup_submit() 333 virtio_gpu_free_post_deps(submit->post_deps, submit->num_out_syncobjs); in virtio_gpu_cleanup_submit() 338 if (submit->buflist) in virtio_gpu_cleanup_submit() 353 virtio_gpu_cmd_submit(submit->vgdev, submit->buf, submit->exbuf->size, in virtio_gpu_submit() 354 submit->vfpriv->ctx_id, submit->buflist, in virtio_gpu_submit() 361 submit->buf = NULL; in virtio_gpu_complete_submit() 379 memset(submit, 0, sizeof(*submit)); in virtio_gpu_init_submit() 410 submit->file = file; in virtio_gpu_init_submit() 460 submit->exbuf->fence_fd = submit->out_fence_fd; in virtio_gpu_install_out_fence_fd() [all …]
|
/openbmc/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem_submit.c | 35 size_t sz = size_vstruct(nr_bos, sizeof(submit->bos[0]), sizeof(*submit)); in submit_create() 38 if (!submit) in submit_create() 43 if (!submit->pmrs) { in submit_create() 44 kfree(submit); in submit_create() 49 submit->gpu = gpu; in submit_create() 52 return submit; in submit_create() 105 submit->nr_bos = i; in submit_lookup_objects() 397 xa_erase(&submit->gpu->user_fences, submit->out_fence_id); in submit_cleanup() 404 kfree(submit); in submit_cleanup() 519 if (!submit) { in etnaviv_ioctl_gem_submit() [all …]
|
H A D | etnaviv_sched.c | 26 fence = etnaviv_gpu_submit(submit); in etnaviv_sched_run_job() 37 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_sched_timedout_job() 48 if (dma_fence_is_signaled(submit->out_fence)) in etnaviv_sched_timedout_job() 70 etnaviv_core_dump(submit); in etnaviv_sched_timedout_job() 71 etnaviv_gpu_recover_hang(submit); in etnaviv_sched_timedout_job() 90 etnaviv_submit_put(submit); in etnaviv_sched_free_job() 101 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_sched_push_job() 111 drm_sched_job_arm(&submit->sched_job); in etnaviv_sched_push_job() 113 submit->out_fence = dma_fence_get(&submit->sched_job.s_fence->finished); in etnaviv_sched_push_job() 115 submit->out_fence, xa_limit_32b, in etnaviv_sched_push_job() [all …]
|
H A D | etnaviv_dump.c | 120 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_core_dump() 133 mutex_lock(&submit->mmu_context->lock); in etnaviv_core_dump() 145 for (i = 0; i < submit->nr_bos; i++) { in etnaviv_core_dump() 146 obj = submit->bos[i].obj; in etnaviv_core_dump() 165 mutex_unlock(&submit->mmu_context->lock); in etnaviv_core_dump() 181 &submit->mmu_context->cmdbuf_mapping)); in etnaviv_core_dump() 184 submit->cmdbuf.vaddr, submit->cmdbuf.size, in etnaviv_core_dump() 188 mutex_unlock(&submit->mmu_context->lock); in etnaviv_core_dump() 201 for (i = 0; i < submit->nr_bos; i++) { in etnaviv_core_dump() 206 obj = submit->bos[i].obj; in etnaviv_core_dump() [all …]
|
/openbmc/linux/crypto/async_tx/ |
H A D | async_xor.c | 40 submit->flags = flags_orig; in do_async_xor() 48 submit->cb_fn = NULL; in do_async_xor() 49 submit->cb_param = NULL; in do_async_xor() 54 if (submit->cb_fn) in do_async_xor() 84 submit->depend_tx = tx; in do_async_xor() 110 if (submit->scribble) in do_sync_xor_offs() 111 srcs = submit->scribble; in do_sync_xor_offs() 244 src_cnt, len, submit); in async_xor_offs() 277 src_cnt, len, submit); in async_xor() 337 if (submit->cb_fn) in async_xor_val_offs() [all …]
|
H A D | async_raid6_recov.c | 39 if (submit->flags & ASYNC_TX_FENCE) in async_sum_product() 58 async_tx_submit(chan, tx, submit); in async_sum_product() 89 struct async_submit_ctl *submit) in async_mult() argument 155 struct async_submit_ctl *submit) in __2data_recov_4() argument 165 void *cb_param = submit->cb_param; in __2data_recov_4() 205 struct async_submit_ctl *submit) in __2data_recov_5() argument 423 async_tx_sync_epilog(submit); in async_raid6_2data_recov() 445 blocks, offs, submit); in async_raid6_2data_recov() 453 blocks, offs, submit); in async_raid6_2data_recov() 456 blocks, offs, submit); in async_raid6_2data_recov() [all …]
|
H A D | async_pq.c | 52 submit->flags = flags_orig; in do_async_gen_syndrome() 61 submit->cb_fn = NULL; in do_async_gen_syndrome() 62 submit->cb_param = NULL; in do_async_gen_syndrome() 64 submit->cb_fn = cb_fn_orig; in do_async_gen_syndrome() 91 submit->depend_tx = tx; in do_async_gen_syndrome() 114 if (submit->scribble) in do_sync_gen_syndrome() 115 srcs = submit->scribble; in do_sync_gen_syndrome() 139 async_tx_sync_epilog(submit); in do_sync_gen_syndrome() 417 len, submit); in async_syndrome_val() 430 submit->cb_fn = cb_fn_orig; in async_syndrome_val() [all …]
|
H A D | raid6test.c | 60 struct async_submit_ctl submit; in raid6_dual_recov() local 71 init_async_submit(&submit, 0, NULL, NULL, NULL, addr_conv); in raid6_dual_recov() 73 disks, bytes, &submit); in raid6_dual_recov() 91 init_async_submit(&submit, ASYNC_TX_XOR_ZERO_DST, NULL, in raid6_dual_recov() 93 tx = async_xor(dest, blocks, 0, count, bytes, &submit); in raid6_dual_recov() 95 init_async_submit(&submit, 0, tx, NULL, NULL, addr_conv); in raid6_dual_recov() 97 disks, bytes, &submit); in raid6_dual_recov() 104 faila, ptrs, offs, &submit); in raid6_dual_recov() 109 faila, failb, ptrs, offs, &submit); in raid6_dual_recov() 115 disks, bytes, &result, spare, 0, &submit); in raid6_dual_recov() [all …]
|
H A D | async_tx.c | 43 __async_tx_find_channel(struct async_submit_ctl *submit, in __async_tx_find_channel() argument 46 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in __async_tx_find_channel() 144 struct async_submit_ctl *submit) in async_tx_submit() argument 146 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in async_tx_submit() 148 tx->callback = submit->cb_fn; in async_tx_submit() 149 tx->callback_param = submit->cb_param; in async_tx_submit() 204 if (submit->flags & ASYNC_TX_ACK) in async_tx_submit() 221 async_trigger_callback(struct async_submit_ctl *submit) in async_trigger_callback() argument 245 async_tx_submit(chan, tx, submit); in async_trigger_callback() 250 async_tx_quiesce(&submit->depend_tx); in async_trigger_callback() [all …]
|
H A D | async_memcpy.c | 34 struct async_submit_ctl *submit) in async_memcpy() argument 36 struct dma_chan *chan = async_tx_find_channel(submit, DMA_MEMCPY, in async_memcpy() 48 if (submit->cb_fn) in async_memcpy() 50 if (submit->flags & ASYNC_TX_FENCE) in async_memcpy() 70 async_tx_submit(chan, tx, submit); in async_memcpy() 76 async_tx_quiesce(&submit->depend_tx); in async_memcpy() 86 async_tx_sync_epilog(submit); in async_memcpy()
|
/openbmc/linux/include/linux/ |
H A D | async_tx.h | 118 async_tx_find_channel(struct async_submit_ctl *submit, in async_tx_find_channel() argument 133 async_tx_sync_epilog(struct async_submit_ctl *submit) in async_tx_sync_epilog() argument 135 if (submit->cb_fn) in async_tx_sync_epilog() 136 submit->cb_fn(submit->cb_param); in async_tx_sync_epilog() 159 struct async_submit_ctl *submit); 173 struct async_submit_ctl *submit); 179 struct async_submit_ctl *submit); 184 struct async_submit_ctl *submit); 190 size_t len, struct async_submit_ctl *submit); 200 struct async_submit_ctl *submit); [all …]
|
/openbmc/linux/drivers/dma/ti/ |
H A D | cppi41.c | 119 u16 submit; member 157 [ 0] = { .submit = 32, .complete = 93}, 158 [ 1] = { .submit = 34, .complete = 94}, 159 [ 2] = { .submit = 36, .complete = 95}, 160 [ 3] = { .submit = 38, .complete = 96}, 161 [ 4] = { .submit = 40, .complete = 97}, 162 [ 5] = { .submit = 42, .complete = 98}, 235 [0] = { .submit = 1, .complete = 26}, 236 [1] = { .submit = 3, .complete = 26}, 237 [2] = { .submit = 5, .complete = 26}, [all …]
|
/openbmc/linux/drivers/gpu/drm/lima/ |
H A D | lima_gem.c | 285 if (!submit->in_sync[i]) in lima_gem_add_deps() 289 submit->in_sync[i], 0); in lima_gem_add_deps() 305 struct lima_bo **bos = submit->lbos; in lima_gem_submit() 307 if (submit->out_sync) { in lima_gem_submit() 338 submit->nr_bos, &ctx); in lima_gem_submit() 343 submit->task, submit->ctx->context + submit->pipe, in lima_gem_submit() 344 bos, submit->nr_bos, vm); in lima_gem_submit() 354 submit->task, bos[i], in lima_gem_submit() 370 submit->nr_bos, &ctx); in lima_gem_submit() 385 lima_sched_task_fini(submit->task); in lima_gem_submit() [all …]
|
H A D | lima_drv.c | 111 struct lima_submit submit = {0}; in lima_ioctl_gem_submit() local 125 bos = kvcalloc(args->nr_bos, sizeof(*submit.bos) + sizeof(*submit.lbos), GFP_KERNEL); in lima_ioctl_gem_submit() 157 submit.pipe = args->pipe; in lima_ioctl_gem_submit() 158 submit.bos = bos; in lima_ioctl_gem_submit() 159 submit.lbos = (void *)bos + size; in lima_ioctl_gem_submit() 160 submit.nr_bos = args->nr_bos; in lima_ioctl_gem_submit() 161 submit.task = task; in lima_ioctl_gem_submit() 162 submit.ctx = ctx; in lima_ioctl_gem_submit() 163 submit.flags = args->flags; in lima_ioctl_gem_submit() 164 submit.in_sync[0] = args->in_sync[0]; in lima_ioctl_gem_submit() [all …]
|
/openbmc/linux/fs/iomap/ |
H A D | direct-io.c | 46 } submit; member 175 struct task_struct *waiter = dio->submit.waiter; in iomap_dio_bio_end_io() 177 WRITE_ONCE(dio->submit.waiter, NULL); in iomap_dio_bio_end_io() 330 orig_count = iov_iter_count(dio->submit.iter); in iomap_dio_bio_iter() 331 iov_iter_truncate(dio->submit.iter, length); in iomap_dio_bio_iter() 333 if (!iov_iter_count(dio->submit.iter)) in iomap_dio_bio_iter() 374 iov_iter_revert(dio->submit.iter, copied); in iomap_dio_bio_iter() 458 struct iov_iter *iter = dio->submit.iter; in iomap_dio_inline_iter() 576 dio->submit.iter = iter; in __iomap_dio_rw() 577 dio->submit.waiter = current; in __iomap_dio_rw() [all …]
|
/openbmc/linux/drivers/gpu/drm/msm/adreno/ |
H A D | a2xx_gpu.c | 13 static void a2xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) in a2xx_submit() argument 15 struct msm_ringbuffer *ring = submit->ring; in a2xx_submit() 18 for (i = 0; i < submit->nr_cmds; i++) { in a2xx_submit() 19 switch (submit->cmd[i].type) { in a2xx_submit() 25 if (gpu->cur_ctx_seqno == submit->queue->ctx->seqno) in a2xx_submit() 30 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in a2xx_submit() 31 OUT_RING(ring, submit->cmd[i].size); in a2xx_submit() 38 OUT_RING(ring, submit->seqno); in a2xx_submit() 47 OUT_RING(ring, submit->seqno); in a2xx_submit() 498 .submit = a2xx_submit,
|
/openbmc/u-boot/doc/driver-model/ |
H A D | MIGRATION.txt | 14 The subsystem itself has been converted and maintainers should submit patches 25 and maintainers should submit patches switching over to using CONFIG_DM_USB and 35 and maintainers should submit patches switching over to using CONFIG_AHCI and 55 Board Maintainers should submit the patches for enabling DM_SPI and DM_SPI_FLASH 96 submit patches switching over to using CONFIG_DM_PCI and other base driver 105 should submit patches switching over to using CONFIG_DM_VIDEO and other base
|
/openbmc/linux/drivers/gpu/drm/tegra/ |
H A D | firewall.c | 9 struct tegra_drm_submit_data *submit; member 31 for (i = 0; i < fw->submit->num_used_mappings; i++) { in fw_check_addr_valid() 32 struct tegra_drm_mapping *m = fw->submit->used_mappings[i].mapping; in fw_check_addr_valid() 145 u32 words, struct tegra_drm_submit_data *submit, in tegra_drm_fw_validate() argument 149 .submit = submit, in tegra_drm_fw_validate()
|
/openbmc/linux/drivers/gpu/drm/i915/selftests/ |
H A D | i915_active.c | 81 struct i915_sw_fence *submit; in __live_active_setup() local 90 submit = heap_fence_create(GFP_KERNEL); in __live_active_setup() 91 if (!submit) { in __live_active_setup() 109 err = i915_sw_fence_await_sw_fence_gfp(&rq->submit, in __live_active_setup() 110 submit, in __live_active_setup() 135 i915_sw_fence_commit(submit); in __live_active_setup() 136 heap_fence_put(submit); in __live_active_setup()
|