1caab277bSThomas Gleixner /* SPDX-License-Identifier: GPL-2.0-only */ 2c8afe684SRob Clark /* 3c8afe684SRob Clark * Copyright (C) 2013 Red Hat 4c8afe684SRob Clark * Author: Rob Clark <robdclark@gmail.com> 5c8afe684SRob Clark */ 6c8afe684SRob Clark 7c8afe684SRob Clark #ifndef __MSM_GEM_H__ 8c8afe684SRob Clark #define __MSM_GEM_H__ 9c8afe684SRob Clark 10ee546cd3SJordan Crouse #include <linux/kref.h> 1152791eeeSChristian König #include <linux/dma-resv.h> 12c8afe684SRob Clark #include "msm_drv.h" 13c8afe684SRob Clark 14072f1f91SRob Clark /* Additional internal-use only BO flags: */ 15072f1f91SRob Clark #define MSM_BO_STOLEN 0x10000000 /* try to use stolen/splash memory */ 160b462d7aSJonathan Marek #define MSM_BO_MAP_PRIV 0x20000000 /* use IOMMU_PRIV when mapping */ 17072f1f91SRob Clark 18667ce33eSRob Clark struct msm_gem_address_space { 19667ce33eSRob Clark const char *name; 20667ce33eSRob Clark /* NOTE: mm managed at the page level, size is in # of pages 21667ce33eSRob Clark * and position mm_node->start is in # of pages: 22667ce33eSRob Clark */ 23667ce33eSRob Clark struct drm_mm mm; 240e08270aSSushmita Susheelendra spinlock_t lock; /* Protects drm_mm node allocation/removal */ 25667ce33eSRob Clark struct msm_mmu *mmu; 26ee546cd3SJordan Crouse struct kref kref; 27667ce33eSRob Clark }; 28667ce33eSRob Clark 29667ce33eSRob Clark struct msm_gem_vma { 30667ce33eSRob Clark struct drm_mm_node node; 31667ce33eSRob Clark uint64_t iova; 324b85f7f5SRob Clark struct msm_gem_address_space *aspace; 334b85f7f5SRob Clark struct list_head list; /* node in msm_gem_object::vmas */ 34c0ee9794SJordan Crouse bool mapped; 357ad0e8cfSJordan Crouse int inuse; 36667ce33eSRob Clark }; 37667ce33eSRob Clark 38c8afe684SRob Clark struct msm_gem_object { 39c8afe684SRob Clark struct drm_gem_object base; 40c8afe684SRob Clark 41c8afe684SRob Clark uint32_t flags; 42c8afe684SRob Clark 434cd33c48SRob Clark /** 444cd33c48SRob Clark * Advice: are the backing pages purgeable? 454cd33c48SRob Clark */ 464cd33c48SRob Clark uint8_t madv; 474cd33c48SRob Clark 48e1e9db2cSRob Clark /** 49e1e9db2cSRob Clark * count of active vmap'ing 50e1e9db2cSRob Clark */ 51e1e9db2cSRob Clark uint8_t vmap_count; 52e1e9db2cSRob Clark 537198e6b0SRob Clark /* And object is either: 547198e6b0SRob Clark * inactive - on priv->inactive_list 557198e6b0SRob Clark * active - on one one of the gpu's active_list.. well, at 567198e6b0SRob Clark * least for now we don't have (I don't think) hw sync between 577198e6b0SRob Clark * 2d and 3d one devices which have both, meaning we need to 587198e6b0SRob Clark * block on submit if a bo is already on other ring 597198e6b0SRob Clark * 607198e6b0SRob Clark */ 61c8afe684SRob Clark struct list_head mm_list; 627198e6b0SRob Clark struct msm_gpu *gpu; /* non-null if active */ 637198e6b0SRob Clark 647198e6b0SRob Clark /* Transiently in the process of submit ioctl, objects associated 657198e6b0SRob Clark * with the submit are on submit->bo_list.. this only lasts for 667198e6b0SRob Clark * the duration of the ioctl, so one bo can never be on multiple 677198e6b0SRob Clark * submit lists. 687198e6b0SRob Clark */ 697198e6b0SRob Clark struct list_head submit_entry; 707198e6b0SRob Clark 71c8afe684SRob Clark struct page **pages; 72c8afe684SRob Clark struct sg_table *sgt; 73c8afe684SRob Clark void *vaddr; 74c8afe684SRob Clark 754b85f7f5SRob Clark struct list_head vmas; /* list of msm_gem_vma */ 767198e6b0SRob Clark 7748e7f183SKristian H. Kristensen struct llist_node freed; 7848e7f183SKristian H. Kristensen 79871d812aSRob Clark /* For physically contiguous buffers. Used when we don't have 80072f1f91SRob Clark * an IOMMU. Also used for stolen/splashscreen buffer. 81871d812aSRob Clark */ 82871d812aSRob Clark struct drm_mm_node *vram_node; 830e08270aSSushmita Susheelendra struct mutex lock; /* Protects resources associated with bo */ 840815d774SJordan Crouse 850815d774SJordan Crouse char name[32]; /* Identifier to print for the debugfs files */ 86c8afe684SRob Clark }; 87c8afe684SRob Clark #define to_msm_bo(x) container_of(x, struct msm_gem_object, base) 88c8afe684SRob Clark 897198e6b0SRob Clark static inline bool is_active(struct msm_gem_object *msm_obj) 907198e6b0SRob Clark { 917198e6b0SRob Clark return msm_obj->gpu != NULL; 927198e6b0SRob Clark } 937198e6b0SRob Clark 9468209390SRob Clark static inline bool is_purgeable(struct msm_gem_object *msm_obj) 9568209390SRob Clark { 960e08270aSSushmita Susheelendra WARN_ON(!mutex_is_locked(&msm_obj->base.dev->struct_mutex)); 9768209390SRob Clark return (msm_obj->madv == MSM_MADV_DONTNEED) && msm_obj->sgt && 9868209390SRob Clark !msm_obj->base.dma_buf && !msm_obj->base.import_attach; 9968209390SRob Clark } 10068209390SRob Clark 101e1e9db2cSRob Clark static inline bool is_vunmapable(struct msm_gem_object *msm_obj) 102e1e9db2cSRob Clark { 103e1e9db2cSRob Clark return (msm_obj->vmap_count == 0) && msm_obj->vaddr; 104e1e9db2cSRob Clark } 105e1e9db2cSRob Clark 1060e08270aSSushmita Susheelendra /* The shrinker can be triggered while we hold objA->lock, and need 1070e08270aSSushmita Susheelendra * to grab objB->lock to purge it. Lockdep just sees these as a single 1080e08270aSSushmita Susheelendra * class of lock, so we use subclasses to teach it the difference. 1090e08270aSSushmita Susheelendra * 1100e08270aSSushmita Susheelendra * OBJ_LOCK_NORMAL is implicit (ie. normal mutex_lock() call), and 1110e08270aSSushmita Susheelendra * OBJ_LOCK_SHRINKER is used by shrinker. 1120e08270aSSushmita Susheelendra * 1130e08270aSSushmita Susheelendra * It is *essential* that we never go down paths that could trigger the 1140e08270aSSushmita Susheelendra * shrinker for a purgable object. This is ensured by checking that 1150e08270aSSushmita Susheelendra * msm_obj->madv == MSM_MADV_WILLNEED. 1160e08270aSSushmita Susheelendra */ 1170e08270aSSushmita Susheelendra enum msm_gem_lock { 1180e08270aSSushmita Susheelendra OBJ_LOCK_NORMAL, 1190e08270aSSushmita Susheelendra OBJ_LOCK_SHRINKER, 1200e08270aSSushmita Susheelendra }; 1210e08270aSSushmita Susheelendra 1220e08270aSSushmita Susheelendra void msm_gem_purge(struct drm_gem_object *obj, enum msm_gem_lock subclass); 1230e08270aSSushmita Susheelendra void msm_gem_vunmap(struct drm_gem_object *obj, enum msm_gem_lock subclass); 12448e7f183SKristian H. Kristensen void msm_gem_free_work(struct work_struct *work); 1250e08270aSSushmita Susheelendra 1267198e6b0SRob Clark /* Created per submit-ioctl, to track bo's and cmdstream bufs, etc, 1277198e6b0SRob Clark * associated with the cmdstream submission for synchronization (and 1287198e6b0SRob Clark * make it easier to unwind when things go wrong, etc). This only 1297198e6b0SRob Clark * lasts for the duration of the submit-ioctl. 1307198e6b0SRob Clark */ 1317198e6b0SRob Clark struct msm_gem_submit { 1327198e6b0SRob Clark struct drm_device *dev; 1337198e6b0SRob Clark struct msm_gpu *gpu; 134295b22aeSJordan Crouse struct msm_gem_address_space *aspace; 135f97decacSJordan Crouse struct list_head node; /* node in ring submit list */ 1367198e6b0SRob Clark struct list_head bo_list; 1377198e6b0SRob Clark struct ww_acquire_ctx ticket; 138f97decacSJordan Crouse uint32_t seqno; /* Sequence number of the submit on the ring */ 139f54d1867SChris Wilson struct dma_fence *fence; 140f7de1545SJordan Crouse struct msm_gpu_submitqueue *queue; 1414816b626SRob Clark struct pid *pid; /* submitting process */ 142340faef2SRob Clark bool valid; /* true if no cmdstream patching needed */ 1436a8bd08dSRob Clark bool in_rb; /* "sudo" mode, copy cmds into RB */ 144f97decacSJordan Crouse struct msm_ringbuffer *ring; 1457198e6b0SRob Clark unsigned int nr_cmds; 1467198e6b0SRob Clark unsigned int nr_bos; 1474241db42SJordan Crouse u32 ident; /* A "identifier" for the submit for logging */ 1487198e6b0SRob Clark struct { 1497198e6b0SRob Clark uint32_t type; 1507198e6b0SRob Clark uint32_t size; /* in dwords */ 15178babc16SRob Clark uint64_t iova; 152a7d3c950SRob Clark uint32_t idx; /* cmdstream buffer idx in bos[] */ 1536b597ce2SRob Clark } *cmd; /* array of size nr_cmds */ 1547198e6b0SRob Clark struct { 1557198e6b0SRob Clark uint32_t flags; 156b673499aSKristian H. Kristensen union { 1577198e6b0SRob Clark struct msm_gem_object *obj; 158b673499aSKristian H. Kristensen uint32_t handle; 159b673499aSKristian H. Kristensen }; 16078babc16SRob Clark uint64_t iova; 1614c145df1SGustavo A. R. Silva } bos[]; 1627198e6b0SRob Clark }; 1637198e6b0SRob Clark 164e515af8dSRob Clark /* helper to determine of a buffer in submit should be dumped, used for both 165e515af8dSRob Clark * devcoredump and debugfs cmdstream dumping: 166e515af8dSRob Clark */ 167e515af8dSRob Clark static inline bool 168e515af8dSRob Clark should_dump(struct msm_gem_submit *submit, int idx) 169e515af8dSRob Clark { 170e515af8dSRob Clark extern bool rd_full; 171e515af8dSRob Clark return rd_full || (submit->bos[idx].flags & MSM_SUBMIT_BO_DUMP); 172e515af8dSRob Clark } 173e515af8dSRob Clark 174c8afe684SRob Clark #endif /* __MSM_GEM_H__ */ 175