1caab277bSThomas Gleixner /* SPDX-License-Identifier: GPL-2.0-only */ 2c8afe684SRob Clark /* 3c8afe684SRob Clark * Copyright (C) 2013 Red Hat 4c8afe684SRob Clark * Author: Rob Clark <robdclark@gmail.com> 5c8afe684SRob Clark */ 6c8afe684SRob Clark 7c8afe684SRob Clark #ifndef __MSM_GEM_H__ 8c8afe684SRob Clark #define __MSM_GEM_H__ 9c8afe684SRob Clark 10ee546cd3SJordan Crouse #include <linux/kref.h> 1152791eeeSChristian König #include <linux/dma-resv.h> 12*1d8a5ca4SRob Clark #include "drm/gpu_scheduler.h" 13c8afe684SRob Clark #include "msm_drv.h" 14c8afe684SRob Clark 1590643a24SRob Clark /* Make all GEM related WARN_ON()s ratelimited.. when things go wrong they 1690643a24SRob Clark * tend to go wrong 1000s of times in a short timespan. 1790643a24SRob Clark */ 1890643a24SRob Clark #define GEM_WARN_ON(x) WARN_RATELIMIT(x, "%s", __stringify(x)) 1990643a24SRob Clark 20072f1f91SRob Clark /* Additional internal-use only BO flags: */ 21072f1f91SRob Clark #define MSM_BO_STOLEN 0x10000000 /* try to use stolen/splash memory */ 220b462d7aSJonathan Marek #define MSM_BO_MAP_PRIV 0x20000000 /* use IOMMU_PRIV when mapping */ 23072f1f91SRob Clark 24667ce33eSRob Clark struct msm_gem_address_space { 25667ce33eSRob Clark const char *name; 26667ce33eSRob Clark /* NOTE: mm managed at the page level, size is in # of pages 27667ce33eSRob Clark * and position mm_node->start is in # of pages: 28667ce33eSRob Clark */ 29667ce33eSRob Clark struct drm_mm mm; 300e08270aSSushmita Susheelendra spinlock_t lock; /* Protects drm_mm node allocation/removal */ 31667ce33eSRob Clark struct msm_mmu *mmu; 32ee546cd3SJordan Crouse struct kref kref; 3325faf2f2SRob Clark 3425faf2f2SRob Clark /* For address spaces associated with a specific process, this 3525faf2f2SRob Clark * will be non-NULL: 3625faf2f2SRob Clark */ 3725faf2f2SRob Clark struct pid *pid; 38667ce33eSRob Clark }; 39667ce33eSRob Clark 40667ce33eSRob Clark struct msm_gem_vma { 41667ce33eSRob Clark struct drm_mm_node node; 42667ce33eSRob Clark uint64_t iova; 434b85f7f5SRob Clark struct msm_gem_address_space *aspace; 444b85f7f5SRob Clark struct list_head list; /* node in msm_gem_object::vmas */ 45c0ee9794SJordan Crouse bool mapped; 467ad0e8cfSJordan Crouse int inuse; 47667ce33eSRob Clark }; 48667ce33eSRob Clark 49c8afe684SRob Clark struct msm_gem_object { 50c8afe684SRob Clark struct drm_gem_object base; 51c8afe684SRob Clark 52c8afe684SRob Clark uint32_t flags; 53c8afe684SRob Clark 544cd33c48SRob Clark /** 554cd33c48SRob Clark * Advice: are the backing pages purgeable? 564cd33c48SRob Clark */ 574cd33c48SRob Clark uint8_t madv; 584cd33c48SRob Clark 59e1e9db2cSRob Clark /** 60cc8a4d5aSRob Clark * Is object on inactive_dontneed list (ie. counted in priv->shrinkable_count)? 61cc8a4d5aSRob Clark */ 62cc8a4d5aSRob Clark bool dontneed : 1; 63cc8a4d5aSRob Clark 64cc8a4d5aSRob Clark /** 6564fcbde7SRob Clark * Is object evictable (ie. counted in priv->evictable_count)? 6664fcbde7SRob Clark */ 6764fcbde7SRob Clark bool evictable : 1; 6864fcbde7SRob Clark 6964fcbde7SRob Clark /** 70e1e9db2cSRob Clark * count of active vmap'ing 71e1e9db2cSRob Clark */ 72e1e9db2cSRob Clark uint8_t vmap_count; 73e1e9db2cSRob Clark 74cc8a4d5aSRob Clark /** 756ed0897cSRob Clark * Node in list of all objects (mainly for debugfs, protected by 766ed0897cSRob Clark * priv->obj_lock 776ed0897cSRob Clark */ 786ed0897cSRob Clark struct list_head node; 796ed0897cSRob Clark 806ed0897cSRob Clark /** 81cc8a4d5aSRob Clark * An object is either: 826ed0897cSRob Clark * inactive - on priv->inactive_dontneed or priv->inactive_willneed 8364fcbde7SRob Clark * (depending on purgeability status) 847198e6b0SRob Clark * active - on one one of the gpu's active_list.. well, at 857198e6b0SRob Clark * least for now we don't have (I don't think) hw sync between 867198e6b0SRob Clark * 2d and 3d one devices which have both, meaning we need to 877198e6b0SRob Clark * block on submit if a bo is already on other ring 887198e6b0SRob Clark */ 89c8afe684SRob Clark struct list_head mm_list; 907198e6b0SRob Clark 917198e6b0SRob Clark /* Transiently in the process of submit ioctl, objects associated 927198e6b0SRob Clark * with the submit are on submit->bo_list.. this only lasts for 937198e6b0SRob Clark * the duration of the ioctl, so one bo can never be on multiple 947198e6b0SRob Clark * submit lists. 957198e6b0SRob Clark */ 967198e6b0SRob Clark struct list_head submit_entry; 977198e6b0SRob Clark 98c8afe684SRob Clark struct page **pages; 99c8afe684SRob Clark struct sg_table *sgt; 100c8afe684SRob Clark void *vaddr; 101c8afe684SRob Clark 1024b85f7f5SRob Clark struct list_head vmas; /* list of msm_gem_vma */ 1037198e6b0SRob Clark 104871d812aSRob Clark /* For physically contiguous buffers. Used when we don't have 105072f1f91SRob Clark * an IOMMU. Also used for stolen/splashscreen buffer. 106871d812aSRob Clark */ 107871d812aSRob Clark struct drm_mm_node *vram_node; 1080815d774SJordan Crouse 1090815d774SJordan Crouse char name[32]; /* Identifier to print for the debugfs files */ 1109d8baa2bSAkhil P Oommen 111ab5c54cbSRob Clark int active_count; 11264fcbde7SRob Clark int pin_count; 113c8afe684SRob Clark }; 114c8afe684SRob Clark #define to_msm_bo(x) container_of(x, struct msm_gem_object, base) 115c8afe684SRob Clark 1168f642378SRob Clark int msm_gem_mmap_obj(struct drm_gem_object *obj, 1178f642378SRob Clark struct vm_area_struct *vma); 1188f642378SRob Clark int msm_gem_mmap(struct file *filp, struct vm_area_struct *vma); 1198f642378SRob Clark uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj); 1208f642378SRob Clark int msm_gem_get_iova(struct drm_gem_object *obj, 1218f642378SRob Clark struct msm_gem_address_space *aspace, uint64_t *iova); 1228f642378SRob Clark int msm_gem_get_and_pin_iova_range(struct drm_gem_object *obj, 1238f642378SRob Clark struct msm_gem_address_space *aspace, uint64_t *iova, 1248f642378SRob Clark u64 range_start, u64 range_end); 125e4b87d22SRob Clark int msm_gem_get_and_pin_iova_locked(struct drm_gem_object *obj, 126e4b87d22SRob Clark struct msm_gem_address_space *aspace, uint64_t *iova); 1278f642378SRob Clark int msm_gem_get_and_pin_iova(struct drm_gem_object *obj, 1288f642378SRob Clark struct msm_gem_address_space *aspace, uint64_t *iova); 1298f642378SRob Clark uint64_t msm_gem_iova(struct drm_gem_object *obj, 1308f642378SRob Clark struct msm_gem_address_space *aspace); 131e4b87d22SRob Clark void msm_gem_unpin_iova_locked(struct drm_gem_object *obj, 132e4b87d22SRob Clark struct msm_gem_address_space *aspace); 1338f642378SRob Clark void msm_gem_unpin_iova(struct drm_gem_object *obj, 1348f642378SRob Clark struct msm_gem_address_space *aspace); 1358f642378SRob Clark struct page **msm_gem_get_pages(struct drm_gem_object *obj); 1368f642378SRob Clark void msm_gem_put_pages(struct drm_gem_object *obj); 1378f642378SRob Clark int msm_gem_dumb_create(struct drm_file *file, struct drm_device *dev, 1388f642378SRob Clark struct drm_mode_create_dumb *args); 1398f642378SRob Clark int msm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, 1408f642378SRob Clark uint32_t handle, uint64_t *offset); 141e4b87d22SRob Clark void *msm_gem_get_vaddr_locked(struct drm_gem_object *obj); 1428f642378SRob Clark void *msm_gem_get_vaddr(struct drm_gem_object *obj); 1438f642378SRob Clark void *msm_gem_get_vaddr_active(struct drm_gem_object *obj); 144e4b87d22SRob Clark void msm_gem_put_vaddr_locked(struct drm_gem_object *obj); 1458f642378SRob Clark void msm_gem_put_vaddr(struct drm_gem_object *obj); 1468f642378SRob Clark int msm_gem_madvise(struct drm_gem_object *obj, unsigned madv); 1478f642378SRob Clark void msm_gem_active_get(struct drm_gem_object *obj, struct msm_gpu *gpu); 1488f642378SRob Clark void msm_gem_active_put(struct drm_gem_object *obj); 1498f642378SRob Clark int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, ktime_t *timeout); 1508f642378SRob Clark int msm_gem_cpu_fini(struct drm_gem_object *obj); 1518f642378SRob Clark void msm_gem_free_object(struct drm_gem_object *obj); 1528f642378SRob Clark int msm_gem_new_handle(struct drm_device *dev, struct drm_file *file, 1538f642378SRob Clark uint32_t size, uint32_t flags, uint32_t *handle, char *name); 1548f642378SRob Clark struct drm_gem_object *msm_gem_new(struct drm_device *dev, 1558f642378SRob Clark uint32_t size, uint32_t flags); 1568f642378SRob Clark void *msm_gem_kernel_new(struct drm_device *dev, uint32_t size, 1578f642378SRob Clark uint32_t flags, struct msm_gem_address_space *aspace, 1588f642378SRob Clark struct drm_gem_object **bo, uint64_t *iova); 1598f642378SRob Clark void msm_gem_kernel_put(struct drm_gem_object *bo, 160030af2b0SRob Clark struct msm_gem_address_space *aspace); 1618f642378SRob Clark struct drm_gem_object *msm_gem_import(struct drm_device *dev, 1628f642378SRob Clark struct dma_buf *dmabuf, struct sg_table *sgt); 1638f642378SRob Clark __printf(2, 3) 1648f642378SRob Clark void msm_gem_object_set_name(struct drm_gem_object *bo, const char *fmt, ...); 165528107c8SRob Clark 166f48f3563SRob Clark #ifdef CONFIG_DEBUG_FS 167528107c8SRob Clark struct msm_gem_stats { 168528107c8SRob Clark struct { 169528107c8SRob Clark unsigned count; 170528107c8SRob Clark size_t size; 171f48f3563SRob Clark } all, active, resident, purgeable, purged; 172528107c8SRob Clark }; 173528107c8SRob Clark 174528107c8SRob Clark void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m, 175528107c8SRob Clark struct msm_gem_stats *stats); 1768f642378SRob Clark void msm_gem_describe_objects(struct list_head *list, struct seq_file *m); 1778f642378SRob Clark #endif 1788f642378SRob Clark 179a6ae74c9SRob Clark static inline void 180a6ae74c9SRob Clark msm_gem_lock(struct drm_gem_object *obj) 181a6ae74c9SRob Clark { 1826c0e3ea2SRob Clark dma_resv_lock(obj->resv, NULL); 183a6ae74c9SRob Clark } 184a6ae74c9SRob Clark 185599089c6SRob Clark static inline bool __must_check 186599089c6SRob Clark msm_gem_trylock(struct drm_gem_object *obj) 187599089c6SRob Clark { 1886c0e3ea2SRob Clark return dma_resv_trylock(obj->resv); 189599089c6SRob Clark } 190599089c6SRob Clark 191a6ae74c9SRob Clark static inline int 192a6ae74c9SRob Clark msm_gem_lock_interruptible(struct drm_gem_object *obj) 193a6ae74c9SRob Clark { 1946c0e3ea2SRob Clark return dma_resv_lock_interruptible(obj->resv, NULL); 195a6ae74c9SRob Clark } 196a6ae74c9SRob Clark 197a6ae74c9SRob Clark static inline void 198a6ae74c9SRob Clark msm_gem_unlock(struct drm_gem_object *obj) 199a6ae74c9SRob Clark { 2006c0e3ea2SRob Clark dma_resv_unlock(obj->resv); 201a6ae74c9SRob Clark } 202a6ae74c9SRob Clark 203a6ae74c9SRob Clark static inline bool 204a6ae74c9SRob Clark msm_gem_is_locked(struct drm_gem_object *obj) 205a6ae74c9SRob Clark { 2066c0e3ea2SRob Clark return dma_resv_is_locked(obj->resv); 207a6ae74c9SRob Clark } 208a6ae74c9SRob Clark 2097198e6b0SRob Clark static inline bool is_active(struct msm_gem_object *msm_obj) 2107198e6b0SRob Clark { 21190643a24SRob Clark GEM_WARN_ON(!msm_gem_is_locked(&msm_obj->base)); 212ab5c54cbSRob Clark return msm_obj->active_count; 2137198e6b0SRob Clark } 2147198e6b0SRob Clark 2150054eeb7SRob Clark /* imported/exported objects are not purgeable: */ 2160054eeb7SRob Clark static inline bool is_unpurgeable(struct msm_gem_object *msm_obj) 217cc8a4d5aSRob Clark { 21810f76165SRob Clark return msm_obj->base.import_attach || msm_obj->pin_count; 219cc8a4d5aSRob Clark } 220cc8a4d5aSRob Clark 22168209390SRob Clark static inline bool is_purgeable(struct msm_gem_object *msm_obj) 22268209390SRob Clark { 22368209390SRob Clark return (msm_obj->madv == MSM_MADV_DONTNEED) && msm_obj->sgt && 2240054eeb7SRob Clark !is_unpurgeable(msm_obj); 22568209390SRob Clark } 22668209390SRob Clark 227e1e9db2cSRob Clark static inline bool is_vunmapable(struct msm_gem_object *msm_obj) 228e1e9db2cSRob Clark { 22990643a24SRob Clark GEM_WARN_ON(!msm_gem_is_locked(&msm_obj->base)); 230e1e9db2cSRob Clark return (msm_obj->vmap_count == 0) && msm_obj->vaddr; 231e1e9db2cSRob Clark } 232e1e9db2cSRob Clark 2330054eeb7SRob Clark static inline void mark_purgeable(struct msm_gem_object *msm_obj) 234cc8a4d5aSRob Clark { 235cc8a4d5aSRob Clark struct msm_drm_private *priv = msm_obj->base.dev->dev_private; 236cc8a4d5aSRob Clark 23790643a24SRob Clark GEM_WARN_ON(!mutex_is_locked(&priv->mm_lock)); 238cc8a4d5aSRob Clark 2390054eeb7SRob Clark if (is_unpurgeable(msm_obj)) 240cc8a4d5aSRob Clark return; 241cc8a4d5aSRob Clark 24290643a24SRob Clark if (GEM_WARN_ON(msm_obj->dontneed)) 243cc8a4d5aSRob Clark return; 244cc8a4d5aSRob Clark 245cc8a4d5aSRob Clark priv->shrinkable_count += msm_obj->base.size >> PAGE_SHIFT; 246cc8a4d5aSRob Clark msm_obj->dontneed = true; 247cc8a4d5aSRob Clark } 248cc8a4d5aSRob Clark 2490054eeb7SRob Clark static inline void mark_unpurgeable(struct msm_gem_object *msm_obj) 250cc8a4d5aSRob Clark { 251cc8a4d5aSRob Clark struct msm_drm_private *priv = msm_obj->base.dev->dev_private; 252cc8a4d5aSRob Clark 25390643a24SRob Clark GEM_WARN_ON(!mutex_is_locked(&priv->mm_lock)); 254cc8a4d5aSRob Clark 2550054eeb7SRob Clark if (is_unpurgeable(msm_obj)) 256cc8a4d5aSRob Clark return; 257cc8a4d5aSRob Clark 25890643a24SRob Clark if (GEM_WARN_ON(!msm_obj->dontneed)) 259cc8a4d5aSRob Clark return; 260cc8a4d5aSRob Clark 261cc8a4d5aSRob Clark priv->shrinkable_count -= msm_obj->base.size >> PAGE_SHIFT; 26290643a24SRob Clark GEM_WARN_ON(priv->shrinkable_count < 0); 263cc8a4d5aSRob Clark msm_obj->dontneed = false; 264cc8a4d5aSRob Clark } 265cc8a4d5aSRob Clark 26664fcbde7SRob Clark static inline bool is_unevictable(struct msm_gem_object *msm_obj) 26764fcbde7SRob Clark { 26810f76165SRob Clark return is_unpurgeable(msm_obj) || msm_obj->vaddr; 26964fcbde7SRob Clark } 27064fcbde7SRob Clark 27164fcbde7SRob Clark static inline void mark_evictable(struct msm_gem_object *msm_obj) 27264fcbde7SRob Clark { 27364fcbde7SRob Clark struct msm_drm_private *priv = msm_obj->base.dev->dev_private; 27464fcbde7SRob Clark 27564fcbde7SRob Clark WARN_ON(!mutex_is_locked(&priv->mm_lock)); 27664fcbde7SRob Clark 27764fcbde7SRob Clark if (is_unevictable(msm_obj)) 27864fcbde7SRob Clark return; 27964fcbde7SRob Clark 28064fcbde7SRob Clark if (WARN_ON(msm_obj->evictable)) 28164fcbde7SRob Clark return; 28264fcbde7SRob Clark 28364fcbde7SRob Clark priv->evictable_count += msm_obj->base.size >> PAGE_SHIFT; 28464fcbde7SRob Clark msm_obj->evictable = true; 28564fcbde7SRob Clark } 28664fcbde7SRob Clark 28764fcbde7SRob Clark static inline void mark_unevictable(struct msm_gem_object *msm_obj) 28864fcbde7SRob Clark { 28964fcbde7SRob Clark struct msm_drm_private *priv = msm_obj->base.dev->dev_private; 29064fcbde7SRob Clark 29164fcbde7SRob Clark WARN_ON(!mutex_is_locked(&priv->mm_lock)); 29264fcbde7SRob Clark 29364fcbde7SRob Clark if (is_unevictable(msm_obj)) 29464fcbde7SRob Clark return; 29564fcbde7SRob Clark 29664fcbde7SRob Clark if (WARN_ON(!msm_obj->evictable)) 29764fcbde7SRob Clark return; 29864fcbde7SRob Clark 29964fcbde7SRob Clark priv->evictable_count -= msm_obj->base.size >> PAGE_SHIFT; 30064fcbde7SRob Clark WARN_ON(priv->evictable_count < 0); 30164fcbde7SRob Clark msm_obj->evictable = false; 30264fcbde7SRob Clark } 30364fcbde7SRob Clark 304599089c6SRob Clark void msm_gem_purge(struct drm_gem_object *obj); 30564fcbde7SRob Clark void msm_gem_evict(struct drm_gem_object *obj); 306599089c6SRob Clark void msm_gem_vunmap(struct drm_gem_object *obj); 3070e08270aSSushmita Susheelendra 3087198e6b0SRob Clark /* Created per submit-ioctl, to track bo's and cmdstream bufs, etc, 3097198e6b0SRob Clark * associated with the cmdstream submission for synchronization (and 310375f9a63SRob Clark * make it easier to unwind when things go wrong, etc). 3117198e6b0SRob Clark */ 3127198e6b0SRob Clark struct msm_gem_submit { 313*1d8a5ca4SRob Clark struct drm_sched_job base; 314964d2f97SRob Clark struct kref ref; 3157198e6b0SRob Clark struct drm_device *dev; 3167198e6b0SRob Clark struct msm_gpu *gpu; 317295b22aeSJordan Crouse struct msm_gem_address_space *aspace; 318f97decacSJordan Crouse struct list_head node; /* node in ring submit list */ 3197198e6b0SRob Clark struct list_head bo_list; 3207198e6b0SRob Clark struct ww_acquire_ctx ticket; 321f97decacSJordan Crouse uint32_t seqno; /* Sequence number of the submit on the ring */ 322*1d8a5ca4SRob Clark 323*1d8a5ca4SRob Clark /* Array of struct dma_fence * to block on before submitting this job. 324*1d8a5ca4SRob Clark */ 325*1d8a5ca4SRob Clark struct xarray deps; 326*1d8a5ca4SRob Clark unsigned long last_dep; 327*1d8a5ca4SRob Clark 328*1d8a5ca4SRob Clark /* Hw fence, which is created when the scheduler executes the job, and 329*1d8a5ca4SRob Clark * is signaled when the hw finishes (via seqno write from cmdstream) 330*1d8a5ca4SRob Clark */ 331*1d8a5ca4SRob Clark struct dma_fence *hw_fence; 332*1d8a5ca4SRob Clark 333*1d8a5ca4SRob Clark /* Userspace visible fence, which is signaled by the scheduler after 334*1d8a5ca4SRob Clark * the hw_fence is signaled. 335*1d8a5ca4SRob Clark */ 336*1d8a5ca4SRob Clark struct dma_fence *user_fence; 337*1d8a5ca4SRob Clark 338a61acbbeSRob Clark int fence_id; /* key into queue->fence_idr */ 339f7de1545SJordan Crouse struct msm_gpu_submitqueue *queue; 3404816b626SRob Clark struct pid *pid; /* submitting process */ 341e25e92e0SRob Clark bool fault_dumped; /* Limit devcoredump dumping to one per submit */ 342340faef2SRob Clark bool valid; /* true if no cmdstream patching needed */ 3436a8bd08dSRob Clark bool in_rb; /* "sudo" mode, copy cmds into RB */ 344f97decacSJordan Crouse struct msm_ringbuffer *ring; 345cf655d61SJordan Crouse struct msm_file_private *ctx; 3467198e6b0SRob Clark unsigned int nr_cmds; 3477198e6b0SRob Clark unsigned int nr_bos; 3484241db42SJordan Crouse u32 ident; /* A "identifier" for the submit for logging */ 3497198e6b0SRob Clark struct { 3507198e6b0SRob Clark uint32_t type; 3517198e6b0SRob Clark uint32_t size; /* in dwords */ 35278babc16SRob Clark uint64_t iova; 35320224d71SRob Clark uint32_t offset;/* in dwords */ 354a7d3c950SRob Clark uint32_t idx; /* cmdstream buffer idx in bos[] */ 35520224d71SRob Clark uint32_t nr_relocs; 35620224d71SRob Clark struct drm_msm_gem_submit_reloc *relocs; 3576b597ce2SRob Clark } *cmd; /* array of size nr_cmds */ 3587198e6b0SRob Clark struct { 3597198e6b0SRob Clark uint32_t flags; 360b673499aSKristian H. Kristensen union { 3617198e6b0SRob Clark struct msm_gem_object *obj; 362b673499aSKristian H. Kristensen uint32_t handle; 363b673499aSKristian H. Kristensen }; 36478babc16SRob Clark uint64_t iova; 3654c145df1SGustavo A. R. Silva } bos[]; 3667198e6b0SRob Clark }; 3677198e6b0SRob Clark 368*1d8a5ca4SRob Clark static inline struct msm_gem_submit *to_msm_submit(struct drm_sched_job *job) 369*1d8a5ca4SRob Clark { 370*1d8a5ca4SRob Clark return container_of(job, struct msm_gem_submit, base); 371*1d8a5ca4SRob Clark } 372*1d8a5ca4SRob Clark 373964d2f97SRob Clark void __msm_gem_submit_destroy(struct kref *kref); 374964d2f97SRob Clark 375964d2f97SRob Clark static inline void msm_gem_submit_get(struct msm_gem_submit *submit) 376964d2f97SRob Clark { 377964d2f97SRob Clark kref_get(&submit->ref); 378964d2f97SRob Clark } 379964d2f97SRob Clark 380964d2f97SRob Clark static inline void msm_gem_submit_put(struct msm_gem_submit *submit) 381964d2f97SRob Clark { 382964d2f97SRob Clark kref_put(&submit->ref, __msm_gem_submit_destroy); 383964d2f97SRob Clark } 384964d2f97SRob Clark 385be40596bSRob Clark void msm_submit_retire(struct msm_gem_submit *submit); 386be40596bSRob Clark 387e515af8dSRob Clark /* helper to determine of a buffer in submit should be dumped, used for both 388e515af8dSRob Clark * devcoredump and debugfs cmdstream dumping: 389e515af8dSRob Clark */ 390e515af8dSRob Clark static inline bool 391e515af8dSRob Clark should_dump(struct msm_gem_submit *submit, int idx) 392e515af8dSRob Clark { 393e515af8dSRob Clark extern bool rd_full; 394e515af8dSRob Clark return rd_full || (submit->bos[idx].flags & MSM_SUBMIT_BO_DUMP); 395e515af8dSRob Clark } 396e515af8dSRob Clark 397c8afe684SRob Clark #endif /* __MSM_GEM_H__ */ 398