Home
last modified time | relevance | path

Searched refs:sg (Results 26 – 50 of 964) sorted by relevance

12345678910>>...39

/openbmc/linux/drivers/crypto/qce/
H A Ddma.c53 struct scatterlist *sg = sgt->sgl, *sg_last = NULL; in qce_sgtable_add() local
56 while (sg) { in qce_sgtable_add()
57 if (!sg_page(sg)) in qce_sgtable_add()
59 sg = sg_next(sg); in qce_sgtable_add()
62 if (!sg) in qce_sgtable_add()
65 while (new_sgl && sg && max_len) { in qce_sgtable_add()
67 sg_set_page(sg, sg_page(new_sgl), new_len, new_sgl->offset); in qce_sgtable_add()
68 sg_last = sg; in qce_sgtable_add()
69 sg = sg_next(sg); in qce_sgtable_add()
85 if (!sg || !nents) in qce_dma_prep_sg()
[all …]
/openbmc/linux/drivers/dma-buf/
H A Dudmabuf.c29 struct sg_table *sg; member
93 sg = kzalloc(sizeof(*sg), GFP_KERNEL); in get_sg_table()
94 if (!sg) in get_sg_table()
104 return sg; in get_sg_table()
107 sg_free_table(sg); in get_sg_table()
108 kfree(sg); in get_sg_table()
117 kfree(sg); in put_sg_table()
139 if (ubuf->sg) in release_udmabuf()
155 if (!ubuf->sg) { in begin_cpu_udmabuf()
162 dma_sync_sg_for_cpu(dev, ubuf->sg->sgl, ubuf->sg->nents, in begin_cpu_udmabuf()
[all …]
/openbmc/linux/drivers/crypto/allwinner/sun8i-ss/
H A Dsun8i-ss-cipher.c45 sg = areq->src; in sun8i_ss_need_fallback()
46 while (sg) { in sun8i_ss_need_fallback()
57 sg = sg_next(sg); in sun8i_ss_need_fallback()
60 sg = areq->dst; in sun8i_ss_need_fallback()
61 while (sg) { in sun8i_ss_need_fallback()
72 sg = sg_next(sg); in sun8i_ss_need_fallback()
145 sg = sg_next(sg); in sun8i_ss_setup_ivs()
169 sg = sg_next(sg); in sun8i_ss_setup_ivs()
255 sg = areq->src; in sun8i_ss_cipher()
267 sg = sg_next(sg); in sun8i_ss_cipher()
[all …]
/openbmc/linux/arch/alpha/kernel/
H A Dpci_iommu.c493 leader = sg; in sg_classify()
500 for (++sg; sg < end; ++sg) { in sg_classify()
519 leader = sg; in sg_classify()
598 sg = leader; in sg_fill()
607 while (sg+1 < end && (int) sg[1].dma_address == -1) { in sg_fill()
609 sg = sg_next(sg); in sg_fill()
628 } while (++sg < end && (int) sg->dma_address < 0); in sg_fill()
650 sg->dma_length = sg->length; in alpha_pci_map_sg()
659 start = sg; in alpha_pci_map_sg()
680 for (out = sg; sg < end; ++sg) { in alpha_pci_map_sg()
[all …]
/openbmc/linux/drivers/infiniband/hw/mlx5/
H A Dumr.c495 dma_unmap_single(ddev, sg->addr, sg->length, DMA_TO_DEVICE); in mlx5r_umr_unmap_free_xlt()
520 sg->addr = dma; in mlx5r_umr_create_xlt()
556 struct ib_sge *sg) in mlx5r_umr_set_update_xlt_data_seg() argument
621 struct ib_sge sg; in mlx5r_umr_update_mr_pas() local
644 dma_sync_single_for_device(ddev, sg.addr, sg.length, in mlx5r_umr_update_mr_pas()
651 dma_sync_single_for_cpu(ddev, sg.addr, sg.length, in mlx5r_umr_update_mr_pas()
653 offset += sg.length; in mlx5r_umr_update_mr_pas()
674 dma_sync_single_for_device(ddev, sg.addr, sg.length, DMA_TO_DEVICE); in mlx5r_umr_update_mr_pas()
704 struct ib_sge sg; in mlx5r_umr_update_xlt() local
747 dma_sync_single_for_cpu(ddev, sg.addr, sg.length, in mlx5r_umr_update_xlt()
[all …]
/openbmc/linux/drivers/crypto/cavium/nitrox/
H A Dnitrox_req.h482 struct scatterlist *sg; member
577 sg++; in create_single_sg()
578 return sg; in create_single_sg()
604 sg++; in create_multi_sg()
607 return sg; in create_multi_sg()
651 sg = creq->src; in nitrox_creq_set_src_sg()
661 sg = create_single_sg(sg, iv, ivsize); in nitrox_creq_set_src_sg()
709 sg = creq->dst; in nitrox_creq_set_dst_sg()
719 sg = create_single_sg(sg, creq->orh, ORH_HLEN); in nitrox_creq_set_dst_sg()
721 sg = create_single_sg(sg, iv, ivsize); in nitrox_creq_set_dst_sg()
[all …]
/openbmc/linux/drivers/target/tcm_fc/
H A Dtfc_io.c81 sg = se_cmd->t_data_sg; in ft_queue_data_in()
82 mem_len = sg->length; in ft_queue_data_in()
83 mem_off = sg->offset; in ft_queue_data_in()
84 page = sg_page(sg); in ft_queue_data_in()
99 sg = sg_next(sg); in ft_queue_data_in()
101 mem_off = sg->offset; in ft_queue_data_in()
102 page = sg_page(sg); in ft_queue_data_in()
277 mem_len = sg->length; in ft_recv_write_data()
279 page = sg_page(sg); in ft_recv_write_data()
284 sg = sg_next(sg); in ft_recv_write_data()
[all …]
/openbmc/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_pages.c550 sg = iter->sg_pos; in __i915_gem_object_page_iter_get_sg()
552 count = dma ? __sg_dma_page_count(sg) : __sg_page_count(sg); in __i915_gem_object_page_iter_get_sg()
579 sg = ____sg_next(sg); in __i915_gem_object_page_iter_get_sg()
580 count = dma ? __sg_dma_page_count(sg) : __sg_page_count(sg); in __i915_gem_object_page_iter_get_sg()
584 iter->sg_pos = sg; in __i915_gem_object_page_iter_get_sg()
597 sg = ____sg_next(sg); in __i915_gem_object_page_iter_get_sg()
598 count = dma ? __sg_dma_page_count(sg) : __sg_page_count(sg); in __i915_gem_object_page_iter_get_sg()
602 return sg; in __i915_gem_object_page_iter_get_sg()
608 GEM_BUG_ON(!sg); in __i915_gem_object_page_iter_get_sg()
621 GEM_BUG_ON(!sg); in __i915_gem_object_page_iter_get_sg()
[all …]
/openbmc/linux/drivers/usb/storage/
H A Dprotocol.c126 struct scatterlist *sg = *sgptr; in usb_stor_access_xfer_buf() local
130 if (sg) in usb_stor_access_xfer_buf()
131 nents = sg_nents(sg); in usb_stor_access_xfer_buf()
133 sg = scsi_sglist(srb); in usb_stor_access_xfer_buf()
135 sg_miter_start(&miter, sg, nents, dir == FROM_XFER_BUF ? in usb_stor_access_xfer_buf()
150 if (*offset + len < miter.piter.sg->length) { in usb_stor_access_xfer_buf()
152 *sgptr = miter.piter.sg; in usb_stor_access_xfer_buf()
155 *sgptr = sg_next(miter.piter.sg); in usb_stor_access_xfer_buf()
173 struct scatterlist *sg = NULL; in usb_stor_set_xfer_buf() local
176 buflen = usb_stor_access_xfer_buf(buffer, buflen, srb, &sg, &offset, in usb_stor_set_xfer_buf()
/openbmc/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_gart.c120 struct sg_table *sg = NULL; in amdgpu_gart_table_ram_alloc() local
148 sg = kmalloc(sizeof(*sg), GFP_KERNEL); in amdgpu_gart_table_ram_alloc()
149 if (!sg) { in amdgpu_gart_table_ram_alloc()
175 bo->tbo.sg = sg; in amdgpu_gart_table_ram_alloc()
176 bo->tbo.ttm->sg = sg; in amdgpu_gart_table_ram_alloc()
205 if (sg) { in amdgpu_gart_table_ram_alloc()
206 sg_free_table(sg); in amdgpu_gart_table_ram_alloc()
207 kfree(sg); in amdgpu_gart_table_ram_alloc()
224 struct sg_table *sg = adev->gart.bo->tbo.sg; in amdgpu_gart_table_ram_free() local
234 sg_free_table(sg); in amdgpu_gart_table_ram_free()
[all …]
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/
H A Dhuge_gem_object.c33 struct scatterlist *sg, *src, *end; in huge_get_pages() local
50 sg = pages->sgl; in huge_get_pages()
56 sg_mark_end(sg); in huge_get_pages()
60 sg_set_page(sg, page, PAGE_SIZE, 0); in huge_get_pages()
61 sg = __sg_next(sg); in huge_get_pages()
64 for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) { in huge_get_pages()
65 sg_set_page(sg, sg_page(src), PAGE_SIZE, 0); in huge_get_pages()
/openbmc/linux/drivers/xen/
H A Dswiotlb-xen.c314 struct scatterlist *sg; in xen_swiotlb_unmap_sg() local
319 for_each_sg(sgl, sg, nelems, i) in xen_swiotlb_unmap_sg()
320 xen_swiotlb_unmap_page(hwdev, sg->dma_address, sg_dma_len(sg), in xen_swiotlb_unmap_sg()
329 struct scatterlist *sg; in xen_swiotlb_map_sg() local
335 sg->dma_address = xen_swiotlb_map_page(dev, sg_page(sg), in xen_swiotlb_map_sg()
336 sg->offset, sg->length, dir, attrs); in xen_swiotlb_map_sg()
339 sg_dma_len(sg) = sg->length; in xen_swiotlb_map_sg()
353 struct scatterlist *sg; in xen_swiotlb_sync_sg_for_cpu() local
358 sg->length, dir); in xen_swiotlb_sync_sg_for_cpu()
366 struct scatterlist *sg; in xen_swiotlb_sync_sg_for_device() local
[all …]
/openbmc/linux/drivers/crypto/caam/
H A Dsg_sw_qm2.h28 sg_to_qm_sg(struct scatterlist *sg, int len, in sg_to_qm_sg() argument
34 ent_len = min_t(int, sg_dma_len(sg), len); in sg_to_qm_sg()
36 dma_to_qm_sg_one(qm_sg_ptr, sg_dma_address(sg), ent_len, in sg_to_qm_sg()
39 sg = sg_next(sg); in sg_to_qm_sg()
49 static inline void sg_to_qm_sg_last(struct scatterlist *sg, int len, in sg_to_qm_sg_last() argument
53 qm_sg_ptr = sg_to_qm_sg(sg, len, qm_sg_ptr, offset); in sg_to_qm_sg_last()
H A Dsg_sw_sec4.h48 sg_to_sec4_sg(struct scatterlist *sg, int len, in sg_to_sec4_sg() argument
54 ent_len = min_t(int, sg_dma_len(sg), len); in sg_to_sec4_sg()
56 dma_to_sec4_sg_one(sec4_sg_ptr, sg_dma_address(sg), ent_len, in sg_to_sec4_sg()
59 sg = sg_next(sg); in sg_to_sec4_sg()
77 static inline void sg_to_sec4_sg_last(struct scatterlist *sg, int len, in sg_to_sec4_sg_last() argument
81 sec4_sg_ptr = sg_to_sec4_sg(sg, len, sec4_sg_ptr, offset); in sg_to_sec4_sg_last()
/openbmc/linux/kernel/sched/
H A Dtopology.c610 if (!sg) in free_sched_groups()
613 first = sg; in free_sched_groups()
622 sg = tmp; in free_sched_groups()
949 if (!sg) in build_group_from_child_sched_domain()
961 return sg; in build_group_from_child_sched_domain()
1086 if (!sg) in build_overlap_sched_groups()
1193 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group()
1218 return sg; in get_group()
1307 sg = sg->next; in init_sched_groups_capacity()
2234 sg->next = sg; in __sdt_alloc()
[all …]
/openbmc/linux/include/linux/
H A Dskmsg.h44 struct sk_msg_sg sg; member
186 dst->sg.data[which] = src->sg.data[which]; in sk_msg_xfer()
188 dst->sg.size += size; in sk_msg_xfer()
189 src->sg.size -= size; in sk_msg_xfer()
202 return sk_msg_iter_dist(msg->sg.start, msg->sg.end) == MAX_MSG_FRAGS; in sk_msg_full()
207 return sk_msg_iter_dist(msg->sg.start, msg->sg.end); in sk_msg_elem_used()
217 return msg->sg.data[which]; in sk_msg_elem_cpy()
234 if (test_bit(msg->sg.start, msg->sg.copy)) { in sk_msg_compute_data_pointers()
253 __set_bit(msg->sg.end, msg->sg.copy); in sk_msg_page_add()
254 msg->sg.size += len; in sk_msg_page_add()
[all …]
/openbmc/linux/drivers/scsi/qla2xxx/
H A Dqla_dsd.h12 static inline void append_dsd32(struct dsd32 **dsd, struct scatterlist *sg) in append_dsd32() argument
14 put_unaligned_le32(sg_dma_address(sg), &(*dsd)->address); in append_dsd32()
15 put_unaligned_le32(sg_dma_len(sg), &(*dsd)->length); in append_dsd32()
25 static inline void append_dsd64(struct dsd64 **dsd, struct scatterlist *sg) in append_dsd64() argument
27 put_unaligned_le64(sg_dma_address(sg), &(*dsd)->address); in append_dsd64()
28 put_unaligned_le32(sg_dma_len(sg), &(*dsd)->length); in append_dsd64()
/openbmc/linux/arch/mips/jazz/
H A Djazzdma.c547 struct scatterlist *sg; in jazz_dma_map_sg() local
551 arch_sync_dma_for_device(sg_phys(sg), sg->length, in jazz_dma_map_sg()
553 sg->dma_address = vdma_alloc(sg_phys(sg), sg->length); in jazz_dma_map_sg()
556 sg_dma_len(sg) = sg->length; in jazz_dma_map_sg()
566 struct scatterlist *sg; in jazz_dma_unmap_sg() local
570 arch_sync_dma_for_cpu(sg_phys(sg), sg->length, dir); in jazz_dma_unmap_sg()
571 vdma_free(sg->dma_address); in jazz_dma_unmap_sg()
590 struct scatterlist *sg; in jazz_dma_sync_sg_for_device() local
594 arch_sync_dma_for_device(sg_phys(sg), sg->length, dir); in jazz_dma_sync_sg_for_device()
600 struct scatterlist *sg; in jazz_dma_sync_sg_for_cpu() local
[all …]
/openbmc/linux/drivers/crypto/
H A Ds5p-sss.c454 if (!*sg) in s5p_free_sg_cpy()
460 kfree(*sg); in s5p_free_sg_cpy()
461 *sg = NULL; in s5p_free_sg_cpy()
1078 ctx->sg = kmalloc_array(n, sizeof(*sg), GFP_KERNEL); in s5p_hash_copy_sg_lists()
1096 while (sg && skip >= sg->length) { in s5p_hash_copy_sg_lists()
1098 sg = sg_next(sg); in s5p_hash_copy_sg_lists()
1107 sg_set_page(tmp, sg_page(sg), len, sg->offset + skip); in s5p_hash_copy_sg_lists()
1114 sg = sg_next(sg); in s5p_hash_copy_sg_lists()
1146 if (!sg || !sg->length || !new_len) in s5p_hash_prepare_sgs()
1195 ctx->sg = sg; in s5p_hash_prepare_sgs()
[all …]
/openbmc/linux/drivers/ata/
H A Dpata_ftide010.c49 struct sata_gemini *sg; member
273 struct sata_gemini *sg = ftide->sg; in pata_ftide010_gemini_port_start() local
316 struct sata_gemini *sg = ftide->sg; in pata_ftide010_gemini_port_stop() local
320 gemini_sata_stop_bridge(sg, 0); in pata_ftide010_gemini_port_stop()
324 gemini_sata_stop_bridge(sg, 1); in pata_ftide010_gemini_port_stop()
329 gemini_sata_stop_bridge(sg, 0); in pata_ftide010_gemini_port_stop()
354 struct sata_gemini *sg; in pata_ftide010_gemini_init() local
358 sg = gemini_sata_bridge_get(); in pata_ftide010_gemini_init()
359 if (IS_ERR(sg)) in pata_ftide010_gemini_init()
360 return PTR_ERR(sg); in pata_ftide010_gemini_init()
[all …]
/openbmc/linux/net/sunrpc/auth_gss/
H A Dgss_krb5_crypto.c132 skcipher_request_set_crypt(req, sg, sg, length, local_iv); in krb5_encrypt()
186 skcipher_request_set_crypt(req, sg, sg, length, local_iv); in krb5_decrypt()
200 ahash_request_set_crypt(req, sg, NULL, sg->length); in checksummer()
405 sg->offset); in encryptor()
406 sg_set_page(&desc->outfrags[desc->fragno], sg_page(sg), sg->length, in encryptor()
407 sg->offset); in encryptor()
433 sg->offset + sg->length - fraglen); in encryptor()
494 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor()
495 sg->offset); in decryptor()
518 sg->offset + sg->length - fraglen); in decryptor()
[all …]
/openbmc/linux/drivers/net/ethernet/microchip/sparx5/
H A Dsparx5_psfp.c132 const struct sparx5_psfp_sg *sg) in sparx5_psfp_sg_set() argument
138 ips = sparx5_psfp_ipv_to_ips(sg->ipv); in sparx5_psfp_sg_set()
139 base_lsb = sg->basetime.tv_sec & 0xffffffff; in sparx5_psfp_sg_set()
140 base_msb = sg->basetime.tv_sec >> 32; in sparx5_psfp_sg_set()
166 for (i = 0; i < sg->num_entries; i++) { in sparx5_psfp_sg_set()
167 gce = &sg->gce[i]; in sparx5_psfp_sg_set()
227 struct sparx5_psfp_sg *sg, u32 *id) in sparx5_psfp_sg_add() argument
241 sg->basetime = ktime_to_timespec64(basetime); in sparx5_psfp_sg_add()
243 sparx5_psfp_sg_set(sparx5, *id, sg); in sparx5_psfp_sg_add()
253 const struct sparx5_psfp_sg sg = { 0 }; in sparx5_psfp_sg_del() local
[all …]
/openbmc/linux/drivers/crypto/intel/qat/qat_common/
H A Dqat_bl.c67 struct scatterlist *sg; in __qat_bl_sgl_to_bufl() local
99 if (!sg->length) in __qat_bl_sgl_to_bufl()
102 if (left >= sg->length) { in __qat_bl_sgl_to_bufl()
103 left -= sg->length; in __qat_bl_sgl_to_bufl()
154 if (!sg->length) in __qat_bl_sgl_to_bufl()
158 left -= sg->length; in __qat_bl_sgl_to_bufl()
162 sg->length - left, in __qat_bl_sgl_to_bufl()
273 struct scatterlist *sg; in qat_bl_sgl_map() local
290 if (!sg->length) in qat_bl_sgl_map()
294 sg->length, in qat_bl_sgl_map()
[all …]
/openbmc/linux/crypto/
H A Daf_alg.c613 sg = sgl->sg; in af_alg_alloc_tsgl()
625 if (sg) in af_alg_alloc_tsgl()
626 sg_chain(sg, MAX_SGL_ENTS + 1, sgl->sg); in af_alg_alloc_tsgl()
657 const struct scatterlist *sg = sgl->sg; in af_alg_count_tsgl() local
713 sg = sgl->sg; in af_alg_pull_tsgl()
1001 sg = sgl->sg + sgl->cur - 1; in af_alg_sendmsg()
1003 PAGE_SIZE - sg->offset - sg->length); in af_alg_sendmsg()
1006 sg->offset + sg->length, in af_alg_sendmsg()
1012 ctx->merge = (sg->offset + sg->length) & in af_alg_sendmsg()
1036 sg = sgl->sg; in af_alg_sendmsg()
[all …]
/openbmc/linux/net/rds/
H A Dtcp_send.c71 unsigned int hdr_off, unsigned int sg, unsigned int off) in rds_tcp_xmit() argument
115 while (sg < rm->data.op_nents) { in rds_tcp_xmit()
117 if (sg + 1 < rm->data.op_nents) in rds_tcp_xmit()
120 bvec_set_page(&bvec, sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit()
121 rm->data.op_sg[sg].length - off, in rds_tcp_xmit()
122 rm->data.op_sg[sg].offset + off); in rds_tcp_xmit()
124 rm->data.op_sg[sg].length - off); in rds_tcp_xmit()
126 rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit()
127 rm->data.op_sg[sg].offset + off, rm->data.op_sg[sg].length - off, in rds_tcp_xmit()
134 if (off == rm->data.op_sg[sg].length) { in rds_tcp_xmit()
[all …]

12345678910>>...39