1 /* 2 * SPDX-License-Identifier: MIT 3 * 4 * Copyright © 2016 Intel Corporation 5 */ 6 7 #ifndef __I915_GEM_CONTEXT_H__ 8 #define __I915_GEM_CONTEXT_H__ 9 10 #include "i915_gem_context_types.h" 11 12 #include "gt/intel_context.h" 13 14 #include "i915_drv.h" 15 #include "i915_gem.h" 16 #include "i915_gem_gtt.h" 17 #include "i915_scheduler.h" 18 #include "intel_device_info.h" 19 20 struct drm_device; 21 struct drm_file; 22 23 static inline bool i915_gem_context_is_closed(const struct i915_gem_context *ctx) 24 { 25 return test_bit(CONTEXT_CLOSED, &ctx->flags); 26 } 27 28 static inline void i915_gem_context_set_closed(struct i915_gem_context *ctx) 29 { 30 GEM_BUG_ON(i915_gem_context_is_closed(ctx)); 31 set_bit(CONTEXT_CLOSED, &ctx->flags); 32 } 33 34 static inline bool i915_gem_context_no_error_capture(const struct i915_gem_context *ctx) 35 { 36 return test_bit(UCONTEXT_NO_ERROR_CAPTURE, &ctx->user_flags); 37 } 38 39 static inline void i915_gem_context_set_no_error_capture(struct i915_gem_context *ctx) 40 { 41 set_bit(UCONTEXT_NO_ERROR_CAPTURE, &ctx->user_flags); 42 } 43 44 static inline void i915_gem_context_clear_no_error_capture(struct i915_gem_context *ctx) 45 { 46 clear_bit(UCONTEXT_NO_ERROR_CAPTURE, &ctx->user_flags); 47 } 48 49 static inline bool i915_gem_context_is_bannable(const struct i915_gem_context *ctx) 50 { 51 return test_bit(UCONTEXT_BANNABLE, &ctx->user_flags); 52 } 53 54 static inline void i915_gem_context_set_bannable(struct i915_gem_context *ctx) 55 { 56 set_bit(UCONTEXT_BANNABLE, &ctx->user_flags); 57 } 58 59 static inline void i915_gem_context_clear_bannable(struct i915_gem_context *ctx) 60 { 61 clear_bit(UCONTEXT_BANNABLE, &ctx->user_flags); 62 } 63 64 static inline bool i915_gem_context_is_recoverable(const struct i915_gem_context *ctx) 65 { 66 return test_bit(UCONTEXT_RECOVERABLE, &ctx->user_flags); 67 } 68 69 static inline void i915_gem_context_set_recoverable(struct i915_gem_context *ctx) 70 { 71 set_bit(UCONTEXT_RECOVERABLE, &ctx->user_flags); 72 } 73 74 static inline void i915_gem_context_clear_recoverable(struct i915_gem_context *ctx) 75 { 76 clear_bit(UCONTEXT_RECOVERABLE, &ctx->user_flags); 77 } 78 79 static inline bool i915_gem_context_is_banned(const struct i915_gem_context *ctx) 80 { 81 return test_bit(CONTEXT_BANNED, &ctx->flags); 82 } 83 84 static inline void i915_gem_context_set_banned(struct i915_gem_context *ctx) 85 { 86 set_bit(CONTEXT_BANNED, &ctx->flags); 87 } 88 89 static inline bool i915_gem_context_force_single_submission(const struct i915_gem_context *ctx) 90 { 91 return test_bit(CONTEXT_FORCE_SINGLE_SUBMISSION, &ctx->flags); 92 } 93 94 static inline void i915_gem_context_set_force_single_submission(struct i915_gem_context *ctx) 95 { 96 __set_bit(CONTEXT_FORCE_SINGLE_SUBMISSION, &ctx->flags); 97 } 98 99 static inline bool 100 i915_gem_context_user_engines(const struct i915_gem_context *ctx) 101 { 102 return test_bit(CONTEXT_USER_ENGINES, &ctx->flags); 103 } 104 105 static inline void 106 i915_gem_context_set_user_engines(struct i915_gem_context *ctx) 107 { 108 set_bit(CONTEXT_USER_ENGINES, &ctx->flags); 109 } 110 111 static inline void 112 i915_gem_context_clear_user_engines(struct i915_gem_context *ctx) 113 { 114 clear_bit(CONTEXT_USER_ENGINES, &ctx->flags); 115 } 116 117 static inline bool i915_gem_context_is_kernel(struct i915_gem_context *ctx) 118 { 119 return !ctx->file_priv; 120 } 121 122 /* i915_gem_context.c */ 123 int __must_check i915_gem_init_contexts(struct drm_i915_private *i915); 124 void i915_gem_driver_release__contexts(struct drm_i915_private *i915); 125 126 int i915_gem_context_open(struct drm_i915_private *i915, 127 struct drm_file *file); 128 void i915_gem_context_close(struct drm_file *file); 129 130 void i915_gem_context_release(struct kref *ctx_ref); 131 132 int i915_gem_vm_create_ioctl(struct drm_device *dev, void *data, 133 struct drm_file *file); 134 int i915_gem_vm_destroy_ioctl(struct drm_device *dev, void *data, 135 struct drm_file *file); 136 137 int i915_gem_context_create_ioctl(struct drm_device *dev, void *data, 138 struct drm_file *file); 139 int i915_gem_context_destroy_ioctl(struct drm_device *dev, void *data, 140 struct drm_file *file); 141 int i915_gem_context_getparam_ioctl(struct drm_device *dev, void *data, 142 struct drm_file *file_priv); 143 int i915_gem_context_setparam_ioctl(struct drm_device *dev, void *data, 144 struct drm_file *file_priv); 145 int i915_gem_context_reset_stats_ioctl(struct drm_device *dev, void *data, 146 struct drm_file *file); 147 148 struct i915_gem_context * 149 i915_gem_context_create_kernel(struct drm_i915_private *i915, int prio); 150 151 static inline struct i915_gem_context * 152 i915_gem_context_get(struct i915_gem_context *ctx) 153 { 154 kref_get(&ctx->ref); 155 return ctx; 156 } 157 158 static inline void i915_gem_context_put(struct i915_gem_context *ctx) 159 { 160 kref_put(&ctx->ref, i915_gem_context_release); 161 } 162 163 static inline struct i915_address_space * 164 i915_gem_context_vm(struct i915_gem_context *ctx) 165 { 166 return rcu_dereference_protected(ctx->vm, lockdep_is_held(&ctx->mutex)); 167 } 168 169 static inline struct i915_address_space * 170 i915_gem_context_get_vm_rcu(struct i915_gem_context *ctx) 171 { 172 struct i915_address_space *vm; 173 174 rcu_read_lock(); 175 vm = rcu_dereference(ctx->vm); 176 if (!vm) 177 vm = &ctx->i915->ggtt.vm; 178 vm = i915_vm_get(vm); 179 rcu_read_unlock(); 180 181 return vm; 182 } 183 184 static inline struct i915_gem_engines * 185 i915_gem_context_engines(struct i915_gem_context *ctx) 186 { 187 return rcu_dereference_protected(ctx->engines, 188 lockdep_is_held(&ctx->engines_mutex)); 189 } 190 191 static inline struct i915_gem_engines * 192 i915_gem_context_lock_engines(struct i915_gem_context *ctx) 193 __acquires(&ctx->engines_mutex) 194 { 195 mutex_lock(&ctx->engines_mutex); 196 return i915_gem_context_engines(ctx); 197 } 198 199 static inline void 200 i915_gem_context_unlock_engines(struct i915_gem_context *ctx) 201 __releases(&ctx->engines_mutex) 202 { 203 mutex_unlock(&ctx->engines_mutex); 204 } 205 206 static inline struct intel_context * 207 i915_gem_context_get_engine(struct i915_gem_context *ctx, unsigned int idx) 208 { 209 struct intel_context *ce = ERR_PTR(-EINVAL); 210 211 rcu_read_lock(); { 212 struct i915_gem_engines *e = rcu_dereference(ctx->engines); 213 if (likely(idx < e->num_engines && e->engines[idx])) 214 ce = intel_context_get(e->engines[idx]); 215 } rcu_read_unlock(); 216 217 return ce; 218 } 219 220 static inline void 221 i915_gem_engines_iter_init(struct i915_gem_engines_iter *it, 222 struct i915_gem_engines *engines) 223 { 224 GEM_BUG_ON(!engines); 225 it->engines = engines; 226 it->idx = 0; 227 } 228 229 struct intel_context * 230 i915_gem_engines_iter_next(struct i915_gem_engines_iter *it); 231 232 #define for_each_gem_engine(ce, engines, it) \ 233 for (i915_gem_engines_iter_init(&(it), (engines)); \ 234 ((ce) = i915_gem_engines_iter_next(&(it)));) 235 236 struct i915_lut_handle *i915_lut_handle_alloc(void); 237 void i915_lut_handle_free(struct i915_lut_handle *lut); 238 239 #endif /* !__I915_GEM_CONTEXT_H__ */ 240