Lines Matching refs:gpu

28 		dev_dbg(submit->gpu->dev, "skipping bad job\n");  in etnaviv_sched_run_job()
37 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_sched_timedout_job() local
42 drm_sched_stop(&gpu->sched, sched_job); in etnaviv_sched_timedout_job()
56 dma_addr = gpu_read(gpu, VIVS_FE_DMA_ADDRESS); in etnaviv_sched_timedout_job()
57 change = dma_addr - gpu->hangcheck_dma_addr; in etnaviv_sched_timedout_job()
58 if (gpu->state == ETNA_GPU_STATE_RUNNING && in etnaviv_sched_timedout_job()
59 (gpu->completed_fence != gpu->hangcheck_fence || in etnaviv_sched_timedout_job()
61 gpu->hangcheck_dma_addr = dma_addr; in etnaviv_sched_timedout_job()
62 gpu->hangcheck_fence = gpu->completed_fence; in etnaviv_sched_timedout_job()
73 drm_sched_resubmit_jobs(&gpu->sched); in etnaviv_sched_timedout_job()
75 drm_sched_start(&gpu->sched, true); in etnaviv_sched_timedout_job()
80 drm_sched_start(&gpu->sched, true); in etnaviv_sched_timedout_job()
101 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_sched_push_job() local
109 mutex_lock(&gpu->sched_lock); in etnaviv_sched_push_job()
114 ret = xa_alloc_cyclic(&gpu->user_fences, &submit->out_fence_id, in etnaviv_sched_push_job()
116 &gpu->next_user_fence, GFP_KERNEL); in etnaviv_sched_push_job()
128 mutex_unlock(&gpu->sched_lock); in etnaviv_sched_push_job()
133 int etnaviv_sched_init(struct etnaviv_gpu *gpu) in etnaviv_sched_init() argument
137 ret = drm_sched_init(&gpu->sched, &etnaviv_sched_ops, in etnaviv_sched_init()
140 dev_name(gpu->dev), gpu->dev); in etnaviv_sched_init()
147 void etnaviv_sched_fini(struct etnaviv_gpu *gpu) in etnaviv_sched_fini() argument
149 drm_sched_fini(&gpu->sched); in etnaviv_sched_fini()