Home
last modified time | relevance | path

Searched refs:fctx (Results 1 – 25 of 33) sorted by relevance

12

/openbmc/linux/drivers/gpu/drm/msm/
H A Dmsm_fence.c35 if (msm_fence_completed(fctx, fctx->next_deadline_fence)) in deadline_work()
49 fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); in msm_fence_context_alloc()
50 if (!fctx) in msm_fence_context_alloc()
54 strscpy(fctx->name, name, sizeof(fctx->name)); in msm_fence_context_alloc()
65 fctx->completed_fence = fctx->last_fence; in msm_fence_context_alloc()
66 *fctx->fenceptr = fctx->last_fence; in msm_fence_context_alloc()
75 return fctx; in msm_fence_context_alloc()
80 kfree(fctx); in msm_fence_context_free()
136 struct msm_fence_context *fctx = f->fctx; in msm_fence_set_deadline() local
192 f->fctx = fctx; in msm_fence_init()
[all …]
H A Dmsm_fence.h79 void msm_fence_context_free(struct msm_fence_context *fctx);
81 bool msm_fence_completed(struct msm_fence_context *fctx, uint32_t fence);
82 void msm_update_fence(struct msm_fence_context *fctx, uint32_t fence);
85 void msm_fence_init(struct dma_fence *fence, struct msm_fence_context *fctx);
H A Dmsm_ringbuffer.c17 struct msm_fence_context *fctx = submit->ring->fctx; in msm_job_run() local
22 msm_fence_init(submit->hw_fence, fctx); in msm_job_run()
113 ring->fctx = msm_fence_context_alloc(gpu->dev, &ring->memptrs->fence, name); in msm_ringbuffer_new()
129 msm_fence_context_free(ring->fctx); in msm_ringbuffer_destroy()
H A Dmsm_ringbuffer.h70 struct msm_fence_context *fctx; member
H A Dmsm_gpu.c418 msm_update_fence(ring->fctx, fence); in recover_worker()
520 } else if (fence_before(fence, ring->fctx->last_fence) && in hangcheck_handler()
530 gpu->name, ring->fctx->last_fence); in hangcheck_handler()
536 if (fence_after(ring->fctx->last_fence, ring->hangcheck_fence)) in hangcheck_handler()
740 msm_update_fence(gpu->rb[i]->fctx, gpu->rb[i]->memptrs->fence); in msm_gpu_retire()
/openbmc/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_fence.c65 if (!--fctx->notify_ref) in nouveau_fence_signal()
99 fctx->killed = 1; in nouveau_fence_context_kill()
109 fctx->dead = 1; in nouveau_fence_context_del()
135 u32 seq = fctx->read(chan); in nouveau_fence_update()
176 struct nouveau_fence_chan *fctx = container_of(event, typeof(*fctx), event); in nouveau_fence_wait_uevent_handler() local
231 &fctx->lock, fctx->context, ++fctx->sequence); in nouveau_fence_emit()
234 &fctx->lock, fctx->context, ++fctx->sequence); in nouveau_fence_emit()
235 kref_get(&fctx->fence_ref); in nouveau_fence_emit()
237 ret = fctx->emit(fence); in nouveau_fence_emit()
463 return !fctx->dead ? fctx->name : "dead channel"; in nouveau_fence_get_timeline_name()
[all …]
H A Dnv10_fence.c61 struct nv10_fence_chan *fctx = chan->fence; in nv10_fence_context_del() local
62 nouveau_fence_context_del(&fctx->base); in nv10_fence_context_del()
63 nvif_object_dtor(&fctx->sema); in nv10_fence_context_del()
65 nouveau_fence_context_free(&fctx->base); in nv10_fence_context_del()
71 struct nv10_fence_chan *fctx; in nv10_fence_context_new() local
73 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv10_fence_context_new()
74 if (!fctx) in nv10_fence_context_new()
77 nouveau_fence_context_new(chan, &fctx->base); in nv10_fence_context_new()
78 fctx->base.emit = nv10_fence_emit; in nv10_fence_context_new()
79 fctx->base.read = nv10_fence_read; in nv10_fence_context_new()
[all …]
H A Dnv84_fence.c120 nouveau_vma_del(&fctx->vma); in nv84_fence_context_del()
122 nouveau_fence_context_del(&fctx->base); in nv84_fence_context_del()
124 nouveau_fence_context_free(&fctx->base); in nv84_fence_context_del()
131 struct nv84_fence_chan *fctx; in nv84_fence_context_new() local
134 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv84_fence_context_new()
135 if (!fctx) in nv84_fence_context_new()
139 fctx->base.emit = nv84_fence_emit; in nv84_fence_context_new()
140 fctx->base.sync = nv84_fence_sync; in nv84_fence_context_new()
141 fctx->base.read = nv84_fence_read; in nv84_fence_context_new()
142 fctx->base.emit32 = nv84_fence_emit32; in nv84_fence_context_new()
[all …]
H A Dnv04_fence.c70 struct nv04_fence_chan *fctx = chan->fence; in nv04_fence_context_del() local
71 nouveau_fence_context_del(&fctx->base); in nv04_fence_context_del()
73 nouveau_fence_context_free(&fctx->base); in nv04_fence_context_del()
79 struct nv04_fence_chan *fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv04_fence_context_new() local
80 if (fctx) { in nv04_fence_context_new()
81 nouveau_fence_context_new(chan, &fctx->base); in nv04_fence_context_new()
82 fctx->base.emit = nv04_fence_emit; in nv04_fence_context_new()
83 fctx->base.sync = nv04_fence_sync; in nv04_fence_context_new()
84 fctx->base.read = nv04_fence_read; in nv04_fence_context_new()
85 chan->fence = fctx; in nv04_fence_context_new()
H A Dnv17_fence.c41 struct nv10_fence_chan *fctx = chan->fence; in nv17_fence_sync() local
57 PUSH_MTHD(ppush, NV176E, SET_CONTEXT_DMA_SEMAPHORE, fctx->sema.handle, in nv17_fence_sync()
65 PUSH_MTHD(npush, NV176E, SET_CONTEXT_DMA_SEMAPHORE, fctx->sema.handle, in nv17_fence_sync()
81 struct nv10_fence_chan *fctx; in nv17_fence_context_new() local
86 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv17_fence_context_new()
87 if (!fctx) in nv17_fence_context_new()
90 nouveau_fence_context_new(chan, &fctx->base); in nv17_fence_context_new()
91 fctx->base.emit = nv10_fence_emit; in nv17_fence_context_new()
92 fctx->base.read = nv10_fence_read; in nv17_fence_context_new()
93 fctx->base.sync = nv17_fence_sync; in nv17_fence_context_new()
[all …]
H A Dnv50_fence.c39 struct nv10_fence_chan *fctx; in nv50_fence_context_new() local
45 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv50_fence_context_new()
46 if (!fctx) in nv50_fence_context_new()
49 nouveau_fence_context_new(chan, &fctx->base); in nv50_fence_context_new()
50 fctx->base.emit = nv10_fence_emit; in nv50_fence_context_new()
51 fctx->base.read = nv10_fence_read; in nv50_fence_context_new()
52 fctx->base.sync = nv17_fence_sync; in nv50_fence_context_new()
62 &fctx->sema); in nv50_fence_context_new()
H A Dnvc0_fence.c82 struct nv84_fence_chan *fctx = chan->fence; in nvc0_fence_context_new() local
83 fctx->base.emit32 = nvc0_fence_emit32; in nvc0_fence_context_new()
84 fctx->base.sync32 = nvc0_fence_sync32; in nvc0_fence_context_new()
/openbmc/linux/drivers/crypto/cavium/nitrox/
H A Dnitrox_aead.c38 struct flexi_crypto_context *fctx; in nitrox_aes_gcm_setkey() local
46 fctx = nctx->u.fctx; in nitrox_aes_gcm_setkey()
47 flags.fu = be64_to_cpu(fctx->flags.f); in nitrox_aes_gcm_setkey()
52 memset(&fctx->crypto, 0, sizeof(fctx->crypto)); in nitrox_aes_gcm_setkey()
62 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_aead_setauthsize() local
219 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_aes_gcm_enc() local
253 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_aes_gcm_dec() local
300 nctx->u.fctx->flags.f = 0; in nitrox_aead_init()
315 flags = &nctx->u.fctx->flags; in nitrox_gcm_common_init()
347 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_aead_exit() local
[all …]
H A Dnitrox_skcipher.c153 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_skcipher_exit() local
156 memzero_explicit(&fctx->auth, sizeof(struct auth_keys)); in nitrox_skcipher_exit()
171 struct flexi_crypto_context *fctx; in nitrox_skcipher_setkey() local
184 fctx = nctx->u.fctx; in nitrox_skcipher_setkey()
185 flags = &fctx->flags; in nitrox_skcipher_setkey()
192 memcpy(fctx->crypto.u.key, key, keylen); in nitrox_skcipher_setkey()
341 struct flexi_crypto_context *fctx; in nitrox_aes_xts_setkey() local
354 fctx = nctx->u.fctx; in nitrox_aes_xts_setkey()
356 memcpy(fctx->auth.u.key2, (key + keylen), keylen); in nitrox_aes_xts_setkey()
365 struct flexi_crypto_context *fctx; in nitrox_aes_ctr_rfc3686_setkey() local
[all …]
/openbmc/linux/drivers/crypto/cavium/cpt/
H A Dcptvf_algs.c101 struct fc_context *fctx = &rctx->fctx; in create_ctx_hdr() local
121 fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type; in create_ctx_hdr()
122 fctx->enc.enc_ctrl.e.aes_key = ctx->key_type; in create_ctx_hdr()
123 fctx->enc.enc_ctrl.e.iv_source = FROM_DPTR; in create_ctx_hdr()
126 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len * 2); in create_ctx_hdr()
128 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len); in create_ctx_hdr()
129 ctrl_flags = (__be64 *)&fctx->enc.enc_ctrl.flags; in create_ctx_hdr()
130 *ctrl_flags = cpu_to_be64(fctx->enc.enc_ctrl.flags); in create_ctx_hdr()
142 req_info->in[*argcnt].vptr = (u8 *)fctx; in create_ctx_hdr()
198 struct fc_context *fctx = &rctx->fctx; in cvm_enc_dec() local
[all …]
H A Dcptvf_algs.h113 struct fc_context fctx; member
/openbmc/linux/drivers/crypto/marvell/octeontx/
H A Dotx_cptvf_algs.c107 if (memcmp(rctx->fctx.hmac.s.hmac_calc, in validate_hmac_cipher_null()
108 rctx->fctx.hmac.s.hmac_recv, in validate_hmac_cipher_null()
240 struct otx_cpt_fc_ctx *fctx = &rctx->fctx; in create_ctx_hdr() local
283 fctx->enc.enc_ctrl.flags = cpu_to_be64(fctx->enc.enc_ctrl.cflags); in create_ctx_hdr()
294 req_info->in[*argcnt].vptr = (u8 *)fctx; in create_ctx_hdr()
355 rctx->fctx.enc.enc_ctrl.flags = 0; in cpt_enc_dec()
973 struct otx_cpt_fc_ctx *fctx = &rctx->fctx; in create_aead_ctx_hdr() local
1031 fctx->enc.enc_ctrl.e.mac_len = mac_len; in create_aead_ctx_hdr()
1032 fctx->enc.enc_ctrl.flags = cpu_to_be64(fctx->enc.enc_ctrl.cflags); in create_aead_ctx_hdr()
1043 req_info->in[*argcnt].vptr = (u8 *)fctx; in create_aead_ctx_hdr()
[all …]
H A Dotx_cptvf_algs.h162 struct otx_cpt_fc_ctx fctx; member
/openbmc/linux/drivers/crypto/marvell/octeontx2/
H A Dotx2_cptvf_algs.c91 if (memcmp(rctx->fctx.hmac.s.hmac_calc, in validate_hmac_cipher_null()
92 rctx->fctx.hmac.s.hmac_recv, in validate_hmac_cipher_null()
225 struct otx2_cpt_fc_ctx *fctx = &rctx->fctx; in create_ctx_hdr() local
268 cpu_to_be64s(&fctx->enc.enc_ctrl.u); in create_ctx_hdr()
279 req_info->in[*argcnt].vptr = (u8 *)fctx; in create_ctx_hdr()
369 rctx->fctx.enc.enc_ctrl.u = 0; in cpt_enc_dec()
1020 struct otx2_cpt_fc_ctx *fctx = &rctx->fctx; in create_aead_ctx_hdr() local
1084 fctx->enc.enc_ctrl.e.mac_len = mac_len; in create_aead_ctx_hdr()
1085 cpu_to_be64s(&fctx->enc.enc_ctrl.u); in create_aead_ctx_hdr()
1096 req_info->in[*argcnt].vptr = (u8 *)fctx; in create_aead_ctx_hdr()
[all …]
H A Dotx2_cptvf_algs.h148 struct otx2_cpt_fc_ctx fctx; member
/openbmc/linux/security/apparmor/
H A Dfile.c463 static void update_file_ctx(struct aa_file_ctx *fctx, struct aa_label *label, in update_file_ctx() argument
469 spin_lock(&fctx->lock); in update_file_ctx()
470 old = rcu_dereference_protected(fctx->label, in update_file_ctx()
471 lockdep_is_held(&fctx->lock)); in update_file_ctx()
475 rcu_assign_pointer(fctx->label, l); in update_file_ctx()
479 fctx->allow |= request; in update_file_ctx()
481 spin_unlock(&fctx->lock); in update_file_ctx()
588 struct aa_file_ctx *fctx; in aa_file_perm() local
596 fctx = file_ctx(file); in aa_file_perm()
599 flabel = rcu_dereference(fctx->label); in aa_file_perm()
[all …]
/openbmc/linux/fs/
H A Duserfaultfd.c703 if (fctx->orig == octx) { in dup_userfaultfd()
704 ctx = fctx->new; in dup_userfaultfd()
709 fctx = kmalloc(sizeof(*fctx), GFP_KERNEL); in dup_userfaultfd()
710 if (!fctx) in dup_userfaultfd()
715 kfree(fctx); in dup_userfaultfd()
729 fctx->orig = octx; in dup_userfaultfd()
730 fctx->new = ctx; in dup_userfaultfd()
731 list_add_tail(&fctx->list, fcs); in dup_userfaultfd()
756 dup_fctx(fctx); in dup_userfaultfd_complete()
757 list_del(&fctx->list); in dup_userfaultfd_complete()
[all …]
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
H A Dnv04.c45 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_chan_stop() local
63 nvkm_kmap(fctx); in nv04_chan_stop()
68 u32 cv = (nvkm_ro32(fctx, c->ctxp + data) & ~cm); in nv04_chan_stop()
69 nvkm_wo32(fctx, c->ctxp + data, cv | (rv << c->ctxs)); in nv04_chan_stop()
71 nvkm_done(fctx); in nv04_chan_stop()
/openbmc/linux/drivers/block/
H A Drbd.c2510 fctx->iter = *fctx->pos; in rbd_img_fill_request_nocopy()
2517 fctx->set_pos_fn, &fctx->iter); in rbd_img_fill_request_nocopy()
2551 num_img_extents, fctx); in rbd_img_fill_request()
2562 fctx->iter = *fctx->pos; in rbd_img_fill_request()
2569 fctx->count_fn, &fctx->iter); in rbd_img_fill_request()
2586 fctx->iter = *fctx->pos; in rbd_img_fill_request()
2592 fctx->copy_fn, &fctx->iter); in rbd_img_fill_request()
2605 struct rbd_img_fill_ctx fctx = { in rbd_img_fill_nodata() local
2655 struct rbd_img_fill_ctx fctx = { in __rbd_img_fill_from_bio() local
2664 &fctx); in __rbd_img_fill_from_bio()
[all …]
/openbmc/linux/drivers/gpu/drm/msm/adreno/
H A Dadreno_gpu.c575 if (fence_before(ring->fctx->last_fence, ring->memptrs->fence)) { in adreno_hw_init()
576 ring->memptrs->fence = ring->fctx->last_fence; in adreno_hw_init()
667 state->ring[i].seqno = gpu->rb[i]->fctx->last_fence; in adreno_gpu_state_get()
921 ring->fctx->last_fence); in adreno_dump_info()

12