1caab277bSThomas Gleixner /* SPDX-License-Identifier: GPL-2.0-only */ 2c8afe684SRob Clark /* 3c8afe684SRob Clark * Copyright (C) 2013 Red Hat 4c8afe684SRob Clark * Author: Rob Clark <robdclark@gmail.com> 5c8afe684SRob Clark */ 6c8afe684SRob Clark 7c8afe684SRob Clark #ifndef __MSM_GEM_H__ 8c8afe684SRob Clark #define __MSM_GEM_H__ 9c8afe684SRob Clark 10ee546cd3SJordan Crouse #include <linux/kref.h> 1152791eeeSChristian König #include <linux/dma-resv.h> 12c8afe684SRob Clark #include "msm_drv.h" 13c8afe684SRob Clark 14072f1f91SRob Clark /* Additional internal-use only BO flags: */ 15072f1f91SRob Clark #define MSM_BO_STOLEN 0x10000000 /* try to use stolen/splash memory */ 160b462d7aSJonathan Marek #define MSM_BO_MAP_PRIV 0x20000000 /* use IOMMU_PRIV when mapping */ 17072f1f91SRob Clark 18667ce33eSRob Clark struct msm_gem_address_space { 19667ce33eSRob Clark const char *name; 20667ce33eSRob Clark /* NOTE: mm managed at the page level, size is in # of pages 21667ce33eSRob Clark * and position mm_node->start is in # of pages: 22667ce33eSRob Clark */ 23667ce33eSRob Clark struct drm_mm mm; 240e08270aSSushmita Susheelendra spinlock_t lock; /* Protects drm_mm node allocation/removal */ 25667ce33eSRob Clark struct msm_mmu *mmu; 26ee546cd3SJordan Crouse struct kref kref; 2725faf2f2SRob Clark 2825faf2f2SRob Clark /* For address spaces associated with a specific process, this 2925faf2f2SRob Clark * will be non-NULL: 3025faf2f2SRob Clark */ 3125faf2f2SRob Clark struct pid *pid; 32667ce33eSRob Clark }; 33667ce33eSRob Clark 34667ce33eSRob Clark struct msm_gem_vma { 35667ce33eSRob Clark struct drm_mm_node node; 36667ce33eSRob Clark uint64_t iova; 374b85f7f5SRob Clark struct msm_gem_address_space *aspace; 384b85f7f5SRob Clark struct list_head list; /* node in msm_gem_object::vmas */ 39c0ee9794SJordan Crouse bool mapped; 407ad0e8cfSJordan Crouse int inuse; 41667ce33eSRob Clark }; 42667ce33eSRob Clark 43c8afe684SRob Clark struct msm_gem_object { 44c8afe684SRob Clark struct drm_gem_object base; 45c8afe684SRob Clark 46c8afe684SRob Clark uint32_t flags; 47c8afe684SRob Clark 484cd33c48SRob Clark /** 494cd33c48SRob Clark * Advice: are the backing pages purgeable? 504cd33c48SRob Clark */ 514cd33c48SRob Clark uint8_t madv; 524cd33c48SRob Clark 53e1e9db2cSRob Clark /** 54e1e9db2cSRob Clark * count of active vmap'ing 55e1e9db2cSRob Clark */ 56e1e9db2cSRob Clark uint8_t vmap_count; 57e1e9db2cSRob Clark 587198e6b0SRob Clark /* And object is either: 597198e6b0SRob Clark * inactive - on priv->inactive_list 607198e6b0SRob Clark * active - on one one of the gpu's active_list.. well, at 617198e6b0SRob Clark * least for now we don't have (I don't think) hw sync between 627198e6b0SRob Clark * 2d and 3d one devices which have both, meaning we need to 637198e6b0SRob Clark * block on submit if a bo is already on other ring 647198e6b0SRob Clark * 657198e6b0SRob Clark */ 66c8afe684SRob Clark struct list_head mm_list; 677198e6b0SRob Clark struct msm_gpu *gpu; /* non-null if active */ 687198e6b0SRob Clark 697198e6b0SRob Clark /* Transiently in the process of submit ioctl, objects associated 707198e6b0SRob Clark * with the submit are on submit->bo_list.. this only lasts for 717198e6b0SRob Clark * the duration of the ioctl, so one bo can never be on multiple 727198e6b0SRob Clark * submit lists. 737198e6b0SRob Clark */ 747198e6b0SRob Clark struct list_head submit_entry; 757198e6b0SRob Clark 76c8afe684SRob Clark struct page **pages; 77c8afe684SRob Clark struct sg_table *sgt; 78c8afe684SRob Clark void *vaddr; 79c8afe684SRob Clark 804b85f7f5SRob Clark struct list_head vmas; /* list of msm_gem_vma */ 817198e6b0SRob Clark 8248e7f183SKristian H. Kristensen struct llist_node freed; 8348e7f183SKristian H. Kristensen 84871d812aSRob Clark /* For physically contiguous buffers. Used when we don't have 85072f1f91SRob Clark * an IOMMU. Also used for stolen/splashscreen buffer. 86871d812aSRob Clark */ 87871d812aSRob Clark struct drm_mm_node *vram_node; 880e08270aSSushmita Susheelendra struct mutex lock; /* Protects resources associated with bo */ 890815d774SJordan Crouse 900815d774SJordan Crouse char name[32]; /* Identifier to print for the debugfs files */ 919d8baa2bSAkhil P Oommen 929d8baa2bSAkhil P Oommen atomic_t active_count; 93c8afe684SRob Clark }; 94c8afe684SRob Clark #define to_msm_bo(x) container_of(x, struct msm_gem_object, base) 95c8afe684SRob Clark 96*8f642378SRob Clark int msm_gem_mmap_obj(struct drm_gem_object *obj, 97*8f642378SRob Clark struct vm_area_struct *vma); 98*8f642378SRob Clark int msm_gem_mmap(struct file *filp, struct vm_area_struct *vma); 99*8f642378SRob Clark vm_fault_t msm_gem_fault(struct vm_fault *vmf); 100*8f642378SRob Clark uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj); 101*8f642378SRob Clark int msm_gem_get_iova(struct drm_gem_object *obj, 102*8f642378SRob Clark struct msm_gem_address_space *aspace, uint64_t *iova); 103*8f642378SRob Clark int msm_gem_get_and_pin_iova_range(struct drm_gem_object *obj, 104*8f642378SRob Clark struct msm_gem_address_space *aspace, uint64_t *iova, 105*8f642378SRob Clark u64 range_start, u64 range_end); 106*8f642378SRob Clark int msm_gem_get_and_pin_iova(struct drm_gem_object *obj, 107*8f642378SRob Clark struct msm_gem_address_space *aspace, uint64_t *iova); 108*8f642378SRob Clark uint64_t msm_gem_iova(struct drm_gem_object *obj, 109*8f642378SRob Clark struct msm_gem_address_space *aspace); 110*8f642378SRob Clark void msm_gem_unpin_iova(struct drm_gem_object *obj, 111*8f642378SRob Clark struct msm_gem_address_space *aspace); 112*8f642378SRob Clark struct page **msm_gem_get_pages(struct drm_gem_object *obj); 113*8f642378SRob Clark void msm_gem_put_pages(struct drm_gem_object *obj); 114*8f642378SRob Clark int msm_gem_dumb_create(struct drm_file *file, struct drm_device *dev, 115*8f642378SRob Clark struct drm_mode_create_dumb *args); 116*8f642378SRob Clark int msm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, 117*8f642378SRob Clark uint32_t handle, uint64_t *offset); 118*8f642378SRob Clark void *msm_gem_get_vaddr(struct drm_gem_object *obj); 119*8f642378SRob Clark void *msm_gem_get_vaddr_active(struct drm_gem_object *obj); 120*8f642378SRob Clark void msm_gem_put_vaddr(struct drm_gem_object *obj); 121*8f642378SRob Clark int msm_gem_madvise(struct drm_gem_object *obj, unsigned madv); 122*8f642378SRob Clark int msm_gem_sync_object(struct drm_gem_object *obj, 123*8f642378SRob Clark struct msm_fence_context *fctx, bool exclusive); 124*8f642378SRob Clark void msm_gem_active_get(struct drm_gem_object *obj, struct msm_gpu *gpu); 125*8f642378SRob Clark void msm_gem_active_put(struct drm_gem_object *obj); 126*8f642378SRob Clark int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, ktime_t *timeout); 127*8f642378SRob Clark int msm_gem_cpu_fini(struct drm_gem_object *obj); 128*8f642378SRob Clark void msm_gem_free_object(struct drm_gem_object *obj); 129*8f642378SRob Clark int msm_gem_new_handle(struct drm_device *dev, struct drm_file *file, 130*8f642378SRob Clark uint32_t size, uint32_t flags, uint32_t *handle, char *name); 131*8f642378SRob Clark struct drm_gem_object *msm_gem_new(struct drm_device *dev, 132*8f642378SRob Clark uint32_t size, uint32_t flags); 133*8f642378SRob Clark struct drm_gem_object *msm_gem_new_locked(struct drm_device *dev, 134*8f642378SRob Clark uint32_t size, uint32_t flags); 135*8f642378SRob Clark void *msm_gem_kernel_new(struct drm_device *dev, uint32_t size, 136*8f642378SRob Clark uint32_t flags, struct msm_gem_address_space *aspace, 137*8f642378SRob Clark struct drm_gem_object **bo, uint64_t *iova); 138*8f642378SRob Clark void *msm_gem_kernel_new_locked(struct drm_device *dev, uint32_t size, 139*8f642378SRob Clark uint32_t flags, struct msm_gem_address_space *aspace, 140*8f642378SRob Clark struct drm_gem_object **bo, uint64_t *iova); 141*8f642378SRob Clark void msm_gem_kernel_put(struct drm_gem_object *bo, 142*8f642378SRob Clark struct msm_gem_address_space *aspace, bool locked); 143*8f642378SRob Clark struct drm_gem_object *msm_gem_import(struct drm_device *dev, 144*8f642378SRob Clark struct dma_buf *dmabuf, struct sg_table *sgt); 145*8f642378SRob Clark __printf(2, 3) 146*8f642378SRob Clark void msm_gem_object_set_name(struct drm_gem_object *bo, const char *fmt, ...); 147*8f642378SRob Clark #ifdef CONFIG_DEBUG_FS 148*8f642378SRob Clark void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m); 149*8f642378SRob Clark void msm_gem_describe_objects(struct list_head *list, struct seq_file *m); 150*8f642378SRob Clark #endif 151*8f642378SRob Clark 152a6ae74c9SRob Clark static inline void 153a6ae74c9SRob Clark msm_gem_lock(struct drm_gem_object *obj) 154a6ae74c9SRob Clark { 155a6ae74c9SRob Clark struct msm_gem_object *msm_obj = to_msm_bo(obj); 156a6ae74c9SRob Clark mutex_lock(&msm_obj->lock); 157a6ae74c9SRob Clark } 158a6ae74c9SRob Clark 159a6ae74c9SRob Clark static inline int 160a6ae74c9SRob Clark msm_gem_lock_interruptible(struct drm_gem_object *obj) 161a6ae74c9SRob Clark { 162a6ae74c9SRob Clark struct msm_gem_object *msm_obj = to_msm_bo(obj); 163a6ae74c9SRob Clark return mutex_lock_interruptible(&msm_obj->lock); 164a6ae74c9SRob Clark } 165a6ae74c9SRob Clark 166a6ae74c9SRob Clark static inline void 167a6ae74c9SRob Clark msm_gem_unlock(struct drm_gem_object *obj) 168a6ae74c9SRob Clark { 169a6ae74c9SRob Clark struct msm_gem_object *msm_obj = to_msm_bo(obj); 170a6ae74c9SRob Clark mutex_unlock(&msm_obj->lock); 171a6ae74c9SRob Clark } 172a6ae74c9SRob Clark 173a6ae74c9SRob Clark static inline bool 174a6ae74c9SRob Clark msm_gem_is_locked(struct drm_gem_object *obj) 175a6ae74c9SRob Clark { 176a6ae74c9SRob Clark struct msm_gem_object *msm_obj = to_msm_bo(obj); 177a6ae74c9SRob Clark return mutex_is_locked(&msm_obj->lock); 178a6ae74c9SRob Clark } 179a6ae74c9SRob Clark 1807198e6b0SRob Clark static inline bool is_active(struct msm_gem_object *msm_obj) 1817198e6b0SRob Clark { 1829d8baa2bSAkhil P Oommen return atomic_read(&msm_obj->active_count); 1837198e6b0SRob Clark } 1847198e6b0SRob Clark 18568209390SRob Clark static inline bool is_purgeable(struct msm_gem_object *msm_obj) 18668209390SRob Clark { 1870e08270aSSushmita Susheelendra WARN_ON(!mutex_is_locked(&msm_obj->base.dev->struct_mutex)); 18868209390SRob Clark return (msm_obj->madv == MSM_MADV_DONTNEED) && msm_obj->sgt && 18968209390SRob Clark !msm_obj->base.dma_buf && !msm_obj->base.import_attach; 19068209390SRob Clark } 19168209390SRob Clark 192e1e9db2cSRob Clark static inline bool is_vunmapable(struct msm_gem_object *msm_obj) 193e1e9db2cSRob Clark { 194e1e9db2cSRob Clark return (msm_obj->vmap_count == 0) && msm_obj->vaddr; 195e1e9db2cSRob Clark } 196e1e9db2cSRob Clark 1970e08270aSSushmita Susheelendra /* The shrinker can be triggered while we hold objA->lock, and need 1980e08270aSSushmita Susheelendra * to grab objB->lock to purge it. Lockdep just sees these as a single 1990e08270aSSushmita Susheelendra * class of lock, so we use subclasses to teach it the difference. 2000e08270aSSushmita Susheelendra * 2010e08270aSSushmita Susheelendra * OBJ_LOCK_NORMAL is implicit (ie. normal mutex_lock() call), and 2020e08270aSSushmita Susheelendra * OBJ_LOCK_SHRINKER is used by shrinker. 2030e08270aSSushmita Susheelendra * 2040e08270aSSushmita Susheelendra * It is *essential* that we never go down paths that could trigger the 2050e08270aSSushmita Susheelendra * shrinker for a purgable object. This is ensured by checking that 2060e08270aSSushmita Susheelendra * msm_obj->madv == MSM_MADV_WILLNEED. 2070e08270aSSushmita Susheelendra */ 2080e08270aSSushmita Susheelendra enum msm_gem_lock { 2090e08270aSSushmita Susheelendra OBJ_LOCK_NORMAL, 2100e08270aSSushmita Susheelendra OBJ_LOCK_SHRINKER, 2110e08270aSSushmita Susheelendra }; 2120e08270aSSushmita Susheelendra 2130e08270aSSushmita Susheelendra void msm_gem_purge(struct drm_gem_object *obj, enum msm_gem_lock subclass); 2140e08270aSSushmita Susheelendra void msm_gem_vunmap(struct drm_gem_object *obj, enum msm_gem_lock subclass); 21548e7f183SKristian H. Kristensen void msm_gem_free_work(struct work_struct *work); 2160e08270aSSushmita Susheelendra 2177198e6b0SRob Clark /* Created per submit-ioctl, to track bo's and cmdstream bufs, etc, 2187198e6b0SRob Clark * associated with the cmdstream submission for synchronization (and 2197198e6b0SRob Clark * make it easier to unwind when things go wrong, etc). This only 2207198e6b0SRob Clark * lasts for the duration of the submit-ioctl. 2217198e6b0SRob Clark */ 2227198e6b0SRob Clark struct msm_gem_submit { 2237198e6b0SRob Clark struct drm_device *dev; 2247198e6b0SRob Clark struct msm_gpu *gpu; 225295b22aeSJordan Crouse struct msm_gem_address_space *aspace; 226f97decacSJordan Crouse struct list_head node; /* node in ring submit list */ 2277198e6b0SRob Clark struct list_head bo_list; 2287198e6b0SRob Clark struct ww_acquire_ctx ticket; 229f97decacSJordan Crouse uint32_t seqno; /* Sequence number of the submit on the ring */ 230f54d1867SChris Wilson struct dma_fence *fence; 231f7de1545SJordan Crouse struct msm_gpu_submitqueue *queue; 2324816b626SRob Clark struct pid *pid; /* submitting process */ 233340faef2SRob Clark bool valid; /* true if no cmdstream patching needed */ 2346a8bd08dSRob Clark bool in_rb; /* "sudo" mode, copy cmds into RB */ 235f97decacSJordan Crouse struct msm_ringbuffer *ring; 236cf655d61SJordan Crouse struct msm_file_private *ctx; 2377198e6b0SRob Clark unsigned int nr_cmds; 2387198e6b0SRob Clark unsigned int nr_bos; 2394241db42SJordan Crouse u32 ident; /* A "identifier" for the submit for logging */ 2407198e6b0SRob Clark struct { 2417198e6b0SRob Clark uint32_t type; 2427198e6b0SRob Clark uint32_t size; /* in dwords */ 24378babc16SRob Clark uint64_t iova; 244a7d3c950SRob Clark uint32_t idx; /* cmdstream buffer idx in bos[] */ 2456b597ce2SRob Clark } *cmd; /* array of size nr_cmds */ 2467198e6b0SRob Clark struct { 2477198e6b0SRob Clark uint32_t flags; 248b673499aSKristian H. Kristensen union { 2497198e6b0SRob Clark struct msm_gem_object *obj; 250b673499aSKristian H. Kristensen uint32_t handle; 251b673499aSKristian H. Kristensen }; 25278babc16SRob Clark uint64_t iova; 2534c145df1SGustavo A. R. Silva } bos[]; 2547198e6b0SRob Clark }; 2557198e6b0SRob Clark 256e515af8dSRob Clark /* helper to determine of a buffer in submit should be dumped, used for both 257e515af8dSRob Clark * devcoredump and debugfs cmdstream dumping: 258e515af8dSRob Clark */ 259e515af8dSRob Clark static inline bool 260e515af8dSRob Clark should_dump(struct msm_gem_submit *submit, int idx) 261e515af8dSRob Clark { 262e515af8dSRob Clark extern bool rd_full; 263e515af8dSRob Clark return rd_full || (submit->bos[idx].flags & MSM_SUBMIT_BO_DUMP); 264e515af8dSRob Clark } 265e515af8dSRob Clark 266c8afe684SRob Clark #endif /* __MSM_GEM_H__ */ 267