1caab277bSThomas Gleixner /* SPDX-License-Identifier: GPL-2.0-only */ 2c8afe684SRob Clark /* 3c8afe684SRob Clark * Copyright (C) 2013 Red Hat 4c8afe684SRob Clark * Author: Rob Clark <robdclark@gmail.com> 5c8afe684SRob Clark */ 6c8afe684SRob Clark 7c8afe684SRob Clark #ifndef __MSM_GEM_H__ 8c8afe684SRob Clark #define __MSM_GEM_H__ 9c8afe684SRob Clark 10ee546cd3SJordan Crouse #include <linux/kref.h> 1152791eeeSChristian König #include <linux/dma-resv.h> 12c8afe684SRob Clark #include "msm_drv.h" 13c8afe684SRob Clark 14072f1f91SRob Clark /* Additional internal-use only BO flags: */ 15072f1f91SRob Clark #define MSM_BO_STOLEN 0x10000000 /* try to use stolen/splash memory */ 160b462d7aSJonathan Marek #define MSM_BO_MAP_PRIV 0x20000000 /* use IOMMU_PRIV when mapping */ 17072f1f91SRob Clark 18667ce33eSRob Clark struct msm_gem_address_space { 19667ce33eSRob Clark const char *name; 20667ce33eSRob Clark /* NOTE: mm managed at the page level, size is in # of pages 21667ce33eSRob Clark * and position mm_node->start is in # of pages: 22667ce33eSRob Clark */ 23667ce33eSRob Clark struct drm_mm mm; 240e08270aSSushmita Susheelendra spinlock_t lock; /* Protects drm_mm node allocation/removal */ 25667ce33eSRob Clark struct msm_mmu *mmu; 26ee546cd3SJordan Crouse struct kref kref; 2725faf2f2SRob Clark 2825faf2f2SRob Clark /* For address spaces associated with a specific process, this 2925faf2f2SRob Clark * will be non-NULL: 3025faf2f2SRob Clark */ 3125faf2f2SRob Clark struct pid *pid; 32667ce33eSRob Clark }; 33667ce33eSRob Clark 34667ce33eSRob Clark struct msm_gem_vma { 35667ce33eSRob Clark struct drm_mm_node node; 36667ce33eSRob Clark uint64_t iova; 374b85f7f5SRob Clark struct msm_gem_address_space *aspace; 384b85f7f5SRob Clark struct list_head list; /* node in msm_gem_object::vmas */ 39c0ee9794SJordan Crouse bool mapped; 407ad0e8cfSJordan Crouse int inuse; 41667ce33eSRob Clark }; 42667ce33eSRob Clark 43c8afe684SRob Clark struct msm_gem_object { 44c8afe684SRob Clark struct drm_gem_object base; 45c8afe684SRob Clark 46c8afe684SRob Clark uint32_t flags; 47c8afe684SRob Clark 484cd33c48SRob Clark /** 494cd33c48SRob Clark * Advice: are the backing pages purgeable? 504cd33c48SRob Clark */ 514cd33c48SRob Clark uint8_t madv; 524cd33c48SRob Clark 53e1e9db2cSRob Clark /** 54e1e9db2cSRob Clark * count of active vmap'ing 55e1e9db2cSRob Clark */ 56e1e9db2cSRob Clark uint8_t vmap_count; 57e1e9db2cSRob Clark 587198e6b0SRob Clark /* And object is either: 597198e6b0SRob Clark * inactive - on priv->inactive_list 607198e6b0SRob Clark * active - on one one of the gpu's active_list.. well, at 617198e6b0SRob Clark * least for now we don't have (I don't think) hw sync between 627198e6b0SRob Clark * 2d and 3d one devices which have both, meaning we need to 637198e6b0SRob Clark * block on submit if a bo is already on other ring 647198e6b0SRob Clark * 657198e6b0SRob Clark */ 66c8afe684SRob Clark struct list_head mm_list; 677198e6b0SRob Clark struct msm_gpu *gpu; /* non-null if active */ 687198e6b0SRob Clark 697198e6b0SRob Clark /* Transiently in the process of submit ioctl, objects associated 707198e6b0SRob Clark * with the submit are on submit->bo_list.. this only lasts for 717198e6b0SRob Clark * the duration of the ioctl, so one bo can never be on multiple 727198e6b0SRob Clark * submit lists. 737198e6b0SRob Clark */ 747198e6b0SRob Clark struct list_head submit_entry; 757198e6b0SRob Clark 76c8afe684SRob Clark struct page **pages; 77c8afe684SRob Clark struct sg_table *sgt; 78c8afe684SRob Clark void *vaddr; 79c8afe684SRob Clark 804b85f7f5SRob Clark struct list_head vmas; /* list of msm_gem_vma */ 817198e6b0SRob Clark 8248e7f183SKristian H. Kristensen struct llist_node freed; 8348e7f183SKristian H. Kristensen 84871d812aSRob Clark /* For physically contiguous buffers. Used when we don't have 85072f1f91SRob Clark * an IOMMU. Also used for stolen/splashscreen buffer. 86871d812aSRob Clark */ 87871d812aSRob Clark struct drm_mm_node *vram_node; 880e08270aSSushmita Susheelendra struct mutex lock; /* Protects resources associated with bo */ 890815d774SJordan Crouse 900815d774SJordan Crouse char name[32]; /* Identifier to print for the debugfs files */ 919d8baa2bSAkhil P Oommen 929d8baa2bSAkhil P Oommen atomic_t active_count; 93c8afe684SRob Clark }; 94c8afe684SRob Clark #define to_msm_bo(x) container_of(x, struct msm_gem_object, base) 95c8afe684SRob Clark 967198e6b0SRob Clark static inline bool is_active(struct msm_gem_object *msm_obj) 977198e6b0SRob Clark { 989d8baa2bSAkhil P Oommen return atomic_read(&msm_obj->active_count); 997198e6b0SRob Clark } 1007198e6b0SRob Clark 10168209390SRob Clark static inline bool is_purgeable(struct msm_gem_object *msm_obj) 10268209390SRob Clark { 1030e08270aSSushmita Susheelendra WARN_ON(!mutex_is_locked(&msm_obj->base.dev->struct_mutex)); 10468209390SRob Clark return (msm_obj->madv == MSM_MADV_DONTNEED) && msm_obj->sgt && 10568209390SRob Clark !msm_obj->base.dma_buf && !msm_obj->base.import_attach; 10668209390SRob Clark } 10768209390SRob Clark 108e1e9db2cSRob Clark static inline bool is_vunmapable(struct msm_gem_object *msm_obj) 109e1e9db2cSRob Clark { 110e1e9db2cSRob Clark return (msm_obj->vmap_count == 0) && msm_obj->vaddr; 111e1e9db2cSRob Clark } 112e1e9db2cSRob Clark 1130e08270aSSushmita Susheelendra /* The shrinker can be triggered while we hold objA->lock, and need 1140e08270aSSushmita Susheelendra * to grab objB->lock to purge it. Lockdep just sees these as a single 1150e08270aSSushmita Susheelendra * class of lock, so we use subclasses to teach it the difference. 1160e08270aSSushmita Susheelendra * 1170e08270aSSushmita Susheelendra * OBJ_LOCK_NORMAL is implicit (ie. normal mutex_lock() call), and 1180e08270aSSushmita Susheelendra * OBJ_LOCK_SHRINKER is used by shrinker. 1190e08270aSSushmita Susheelendra * 1200e08270aSSushmita Susheelendra * It is *essential* that we never go down paths that could trigger the 1210e08270aSSushmita Susheelendra * shrinker for a purgable object. This is ensured by checking that 1220e08270aSSushmita Susheelendra * msm_obj->madv == MSM_MADV_WILLNEED. 1230e08270aSSushmita Susheelendra */ 1240e08270aSSushmita Susheelendra enum msm_gem_lock { 1250e08270aSSushmita Susheelendra OBJ_LOCK_NORMAL, 1260e08270aSSushmita Susheelendra OBJ_LOCK_SHRINKER, 1270e08270aSSushmita Susheelendra }; 1280e08270aSSushmita Susheelendra 1290e08270aSSushmita Susheelendra void msm_gem_purge(struct drm_gem_object *obj, enum msm_gem_lock subclass); 1300e08270aSSushmita Susheelendra void msm_gem_vunmap(struct drm_gem_object *obj, enum msm_gem_lock subclass); 13148e7f183SKristian H. Kristensen void msm_gem_free_work(struct work_struct *work); 1320e08270aSSushmita Susheelendra 1337198e6b0SRob Clark /* Created per submit-ioctl, to track bo's and cmdstream bufs, etc, 1347198e6b0SRob Clark * associated with the cmdstream submission for synchronization (and 1357198e6b0SRob Clark * make it easier to unwind when things go wrong, etc). This only 1367198e6b0SRob Clark * lasts for the duration of the submit-ioctl. 1377198e6b0SRob Clark */ 1387198e6b0SRob Clark struct msm_gem_submit { 1397198e6b0SRob Clark struct drm_device *dev; 1407198e6b0SRob Clark struct msm_gpu *gpu; 141295b22aeSJordan Crouse struct msm_gem_address_space *aspace; 142f97decacSJordan Crouse struct list_head node; /* node in ring submit list */ 1437198e6b0SRob Clark struct list_head bo_list; 1447198e6b0SRob Clark struct ww_acquire_ctx ticket; 145f97decacSJordan Crouse uint32_t seqno; /* Sequence number of the submit on the ring */ 146f54d1867SChris Wilson struct dma_fence *fence; 147f7de1545SJordan Crouse struct msm_gpu_submitqueue *queue; 1484816b626SRob Clark struct pid *pid; /* submitting process */ 149340faef2SRob Clark bool valid; /* true if no cmdstream patching needed */ 1506a8bd08dSRob Clark bool in_rb; /* "sudo" mode, copy cmds into RB */ 151f97decacSJordan Crouse struct msm_ringbuffer *ring; 152cf655d61SJordan Crouse struct msm_file_private *ctx; 1537198e6b0SRob Clark unsigned int nr_cmds; 1547198e6b0SRob Clark unsigned int nr_bos; 1554241db42SJordan Crouse u32 ident; /* A "identifier" for the submit for logging */ 1567198e6b0SRob Clark struct { 1577198e6b0SRob Clark uint32_t type; 1587198e6b0SRob Clark uint32_t size; /* in dwords */ 15978babc16SRob Clark uint64_t iova; 160a7d3c950SRob Clark uint32_t idx; /* cmdstream buffer idx in bos[] */ 1616b597ce2SRob Clark } *cmd; /* array of size nr_cmds */ 1627198e6b0SRob Clark struct { 1637198e6b0SRob Clark uint32_t flags; 164b673499aSKristian H. Kristensen union { 1657198e6b0SRob Clark struct msm_gem_object *obj; 166b673499aSKristian H. Kristensen uint32_t handle; 167b673499aSKristian H. Kristensen }; 16878babc16SRob Clark uint64_t iova; 1694c145df1SGustavo A. R. Silva } bos[]; 1707198e6b0SRob Clark }; 1717198e6b0SRob Clark 172e515af8dSRob Clark /* helper to determine of a buffer in submit should be dumped, used for both 173e515af8dSRob Clark * devcoredump and debugfs cmdstream dumping: 174e515af8dSRob Clark */ 175e515af8dSRob Clark static inline bool 176e515af8dSRob Clark should_dump(struct msm_gem_submit *submit, int idx) 177e515af8dSRob Clark { 178e515af8dSRob Clark extern bool rd_full; 179e515af8dSRob Clark return rd_full || (submit->bos[idx].flags & MSM_SUBMIT_BO_DUMP); 180e515af8dSRob Clark } 181e515af8dSRob Clark 182c8afe684SRob Clark #endif /* __MSM_GEM_H__ */ 183