Lines Matching refs:drv
63 (u64)atomic64_read(&fence->drv->last_fence_id)); in virtio_gpu_timeline_value_str()
79 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_alloc() local
86 fence->drv = drv; in virtio_gpu_fence_alloc()
88 fence->emit_fence_info = !(base_fence_ctx == drv->context); in virtio_gpu_fence_alloc()
95 dma_fence_init(&fence->f, &virtio_gpu_fence_ops, &drv->lock, in virtio_gpu_fence_alloc()
105 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_emit() local
108 spin_lock_irqsave(&drv->lock, irq_flags); in virtio_gpu_fence_emit()
109 fence->fence_id = fence->f.seqno = ++drv->current_fence_id; in virtio_gpu_fence_emit()
111 list_add_tail(&fence->node, &drv->fences); in virtio_gpu_fence_emit()
112 spin_unlock_irqrestore(&drv->lock, irq_flags); in virtio_gpu_fence_emit()
130 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_event_process() local
134 spin_lock_irqsave(&drv->lock, irq_flags); in virtio_gpu_fence_event_process()
136 list_for_each_entry_safe(curr, tmp, &drv->fences, node) { in virtio_gpu_fence_event_process()
146 list_for_each_entry_safe(curr, tmp, &drv->fences, node) { in virtio_gpu_fence_event_process()
174 spin_unlock_irqrestore(&drv->lock, irq_flags); in virtio_gpu_fence_event_process()