/openbmc/linux/drivers/crypto/qce/ |
H A D | dma.c | 53 struct scatterlist *sg = sgt->sgl, *sg_last = NULL; in qce_sgtable_add() local 56 while (sg) { in qce_sgtable_add() 57 if (!sg_page(sg)) in qce_sgtable_add() 59 sg = sg_next(sg); in qce_sgtable_add() 62 if (!sg) in qce_sgtable_add() 65 while (new_sgl && sg && max_len) { in qce_sgtable_add() 67 sg_set_page(sg, sg_page(new_sgl), new_len, new_sgl->offset); in qce_sgtable_add() 68 sg_last = sg; in qce_sgtable_add() 69 sg = sg_next(sg); in qce_sgtable_add() 85 if (!sg || !nents) in qce_dma_prep_sg() [all …]
|
/openbmc/linux/drivers/dma-buf/ |
H A D | udmabuf.c | 29 struct sg_table *sg; member 93 sg = kzalloc(sizeof(*sg), GFP_KERNEL); in get_sg_table() 94 if (!sg) in get_sg_table() 104 return sg; in get_sg_table() 107 sg_free_table(sg); in get_sg_table() 108 kfree(sg); in get_sg_table() 117 kfree(sg); in put_sg_table() 139 if (ubuf->sg) in release_udmabuf() 155 if (!ubuf->sg) { in begin_cpu_udmabuf() 162 dma_sync_sg_for_cpu(dev, ubuf->sg->sgl, ubuf->sg->nents, in begin_cpu_udmabuf() [all …]
|
/openbmc/linux/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-cipher.c | 45 sg = areq->src; in sun8i_ss_need_fallback() 46 while (sg) { in sun8i_ss_need_fallback() 57 sg = sg_next(sg); in sun8i_ss_need_fallback() 60 sg = areq->dst; in sun8i_ss_need_fallback() 61 while (sg) { in sun8i_ss_need_fallback() 72 sg = sg_next(sg); in sun8i_ss_need_fallback() 145 sg = sg_next(sg); in sun8i_ss_setup_ivs() 169 sg = sg_next(sg); in sun8i_ss_setup_ivs() 255 sg = areq->src; in sun8i_ss_cipher() 267 sg = sg_next(sg); in sun8i_ss_cipher() [all …]
|
/openbmc/linux/arch/alpha/kernel/ |
H A D | pci_iommu.c | 493 leader = sg; in sg_classify() 500 for (++sg; sg < end; ++sg) { in sg_classify() 519 leader = sg; in sg_classify() 598 sg = leader; in sg_fill() 607 while (sg+1 < end && (int) sg[1].dma_address == -1) { in sg_fill() 609 sg = sg_next(sg); in sg_fill() 628 } while (++sg < end && (int) sg->dma_address < 0); in sg_fill() 650 sg->dma_length = sg->length; in alpha_pci_map_sg() 659 start = sg; in alpha_pci_map_sg() 680 for (out = sg; sg < end; ++sg) { in alpha_pci_map_sg() [all …]
|
/openbmc/linux/drivers/infiniband/hw/mlx5/ |
H A D | umr.c | 495 dma_unmap_single(ddev, sg->addr, sg->length, DMA_TO_DEVICE); in mlx5r_umr_unmap_free_xlt() 520 sg->addr = dma; in mlx5r_umr_create_xlt() 556 struct ib_sge *sg) in mlx5r_umr_set_update_xlt_data_seg() argument 621 struct ib_sge sg; in mlx5r_umr_update_mr_pas() local 644 dma_sync_single_for_device(ddev, sg.addr, sg.length, in mlx5r_umr_update_mr_pas() 651 dma_sync_single_for_cpu(ddev, sg.addr, sg.length, in mlx5r_umr_update_mr_pas() 653 offset += sg.length; in mlx5r_umr_update_mr_pas() 674 dma_sync_single_for_device(ddev, sg.addr, sg.length, DMA_TO_DEVICE); in mlx5r_umr_update_mr_pas() 704 struct ib_sge sg; in mlx5r_umr_update_xlt() local 747 dma_sync_single_for_cpu(ddev, sg.addr, sg.length, in mlx5r_umr_update_xlt() [all …]
|
/openbmc/linux/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_req.h | 482 struct scatterlist *sg; member 577 sg++; in create_single_sg() 578 return sg; in create_single_sg() 604 sg++; in create_multi_sg() 607 return sg; in create_multi_sg() 651 sg = creq->src; in nitrox_creq_set_src_sg() 661 sg = create_single_sg(sg, iv, ivsize); in nitrox_creq_set_src_sg() 709 sg = creq->dst; in nitrox_creq_set_dst_sg() 719 sg = create_single_sg(sg, creq->orh, ORH_HLEN); in nitrox_creq_set_dst_sg() 721 sg = create_single_sg(sg, iv, ivsize); in nitrox_creq_set_dst_sg() [all …]
|
/openbmc/linux/drivers/target/tcm_fc/ |
H A D | tfc_io.c | 81 sg = se_cmd->t_data_sg; in ft_queue_data_in() 82 mem_len = sg->length; in ft_queue_data_in() 83 mem_off = sg->offset; in ft_queue_data_in() 84 page = sg_page(sg); in ft_queue_data_in() 99 sg = sg_next(sg); in ft_queue_data_in() 101 mem_off = sg->offset; in ft_queue_data_in() 102 page = sg_page(sg); in ft_queue_data_in() 277 mem_len = sg->length; in ft_recv_write_data() 279 page = sg_page(sg); in ft_recv_write_data() 284 sg = sg_next(sg); in ft_recv_write_data() [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_pages.c | 550 sg = iter->sg_pos; in __i915_gem_object_page_iter_get_sg() 552 count = dma ? __sg_dma_page_count(sg) : __sg_page_count(sg); in __i915_gem_object_page_iter_get_sg() 579 sg = ____sg_next(sg); in __i915_gem_object_page_iter_get_sg() 580 count = dma ? __sg_dma_page_count(sg) : __sg_page_count(sg); in __i915_gem_object_page_iter_get_sg() 584 iter->sg_pos = sg; in __i915_gem_object_page_iter_get_sg() 597 sg = ____sg_next(sg); in __i915_gem_object_page_iter_get_sg() 598 count = dma ? __sg_dma_page_count(sg) : __sg_page_count(sg); in __i915_gem_object_page_iter_get_sg() 602 return sg; in __i915_gem_object_page_iter_get_sg() 608 GEM_BUG_ON(!sg); in __i915_gem_object_page_iter_get_sg() 621 GEM_BUG_ON(!sg); in __i915_gem_object_page_iter_get_sg() [all …]
|
/openbmc/linux/drivers/usb/storage/ |
H A D | protocol.c | 126 struct scatterlist *sg = *sgptr; in usb_stor_access_xfer_buf() local 130 if (sg) in usb_stor_access_xfer_buf() 131 nents = sg_nents(sg); in usb_stor_access_xfer_buf() 133 sg = scsi_sglist(srb); in usb_stor_access_xfer_buf() 135 sg_miter_start(&miter, sg, nents, dir == FROM_XFER_BUF ? in usb_stor_access_xfer_buf() 150 if (*offset + len < miter.piter.sg->length) { in usb_stor_access_xfer_buf() 152 *sgptr = miter.piter.sg; in usb_stor_access_xfer_buf() 155 *sgptr = sg_next(miter.piter.sg); in usb_stor_access_xfer_buf() 173 struct scatterlist *sg = NULL; in usb_stor_set_xfer_buf() local 176 buflen = usb_stor_access_xfer_buf(buffer, buflen, srb, &sg, &offset, in usb_stor_set_xfer_buf()
|
/openbmc/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gart.c | 120 struct sg_table *sg = NULL; in amdgpu_gart_table_ram_alloc() local 148 sg = kmalloc(sizeof(*sg), GFP_KERNEL); in amdgpu_gart_table_ram_alloc() 149 if (!sg) { in amdgpu_gart_table_ram_alloc() 175 bo->tbo.sg = sg; in amdgpu_gart_table_ram_alloc() 176 bo->tbo.ttm->sg = sg; in amdgpu_gart_table_ram_alloc() 205 if (sg) { in amdgpu_gart_table_ram_alloc() 206 sg_free_table(sg); in amdgpu_gart_table_ram_alloc() 207 kfree(sg); in amdgpu_gart_table_ram_alloc() 224 struct sg_table *sg = adev->gart.bo->tbo.sg; in amdgpu_gart_table_ram_free() local 234 sg_free_table(sg); in amdgpu_gart_table_ram_free() [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_gem_object.c | 33 struct scatterlist *sg, *src, *end; in huge_get_pages() local 50 sg = pages->sgl; in huge_get_pages() 56 sg_mark_end(sg); in huge_get_pages() 60 sg_set_page(sg, page, PAGE_SIZE, 0); in huge_get_pages() 61 sg = __sg_next(sg); in huge_get_pages() 64 for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) { in huge_get_pages() 65 sg_set_page(sg, sg_page(src), PAGE_SIZE, 0); in huge_get_pages()
|
/openbmc/linux/drivers/xen/ |
H A D | swiotlb-xen.c | 314 struct scatterlist *sg; in xen_swiotlb_unmap_sg() local 319 for_each_sg(sgl, sg, nelems, i) in xen_swiotlb_unmap_sg() 320 xen_swiotlb_unmap_page(hwdev, sg->dma_address, sg_dma_len(sg), in xen_swiotlb_unmap_sg() 329 struct scatterlist *sg; in xen_swiotlb_map_sg() local 335 sg->dma_address = xen_swiotlb_map_page(dev, sg_page(sg), in xen_swiotlb_map_sg() 336 sg->offset, sg->length, dir, attrs); in xen_swiotlb_map_sg() 339 sg_dma_len(sg) = sg->length; in xen_swiotlb_map_sg() 353 struct scatterlist *sg; in xen_swiotlb_sync_sg_for_cpu() local 358 sg->length, dir); in xen_swiotlb_sync_sg_for_cpu() 366 struct scatterlist *sg; in xen_swiotlb_sync_sg_for_device() local [all …]
|
/openbmc/linux/drivers/crypto/caam/ |
H A D | sg_sw_qm2.h | 28 sg_to_qm_sg(struct scatterlist *sg, int len, in sg_to_qm_sg() argument 34 ent_len = min_t(int, sg_dma_len(sg), len); in sg_to_qm_sg() 36 dma_to_qm_sg_one(qm_sg_ptr, sg_dma_address(sg), ent_len, in sg_to_qm_sg() 39 sg = sg_next(sg); in sg_to_qm_sg() 49 static inline void sg_to_qm_sg_last(struct scatterlist *sg, int len, in sg_to_qm_sg_last() argument 53 qm_sg_ptr = sg_to_qm_sg(sg, len, qm_sg_ptr, offset); in sg_to_qm_sg_last()
|
H A D | sg_sw_sec4.h | 48 sg_to_sec4_sg(struct scatterlist *sg, int len, in sg_to_sec4_sg() argument 54 ent_len = min_t(int, sg_dma_len(sg), len); in sg_to_sec4_sg() 56 dma_to_sec4_sg_one(sec4_sg_ptr, sg_dma_address(sg), ent_len, in sg_to_sec4_sg() 59 sg = sg_next(sg); in sg_to_sec4_sg() 77 static inline void sg_to_sec4_sg_last(struct scatterlist *sg, int len, in sg_to_sec4_sg_last() argument 81 sec4_sg_ptr = sg_to_sec4_sg(sg, len, sec4_sg_ptr, offset); in sg_to_sec4_sg_last()
|
/openbmc/linux/kernel/sched/ |
H A D | topology.c | 610 if (!sg) in free_sched_groups() 613 first = sg; in free_sched_groups() 622 sg = tmp; in free_sched_groups() 949 if (!sg) in build_group_from_child_sched_domain() 961 return sg; in build_group_from_child_sched_domain() 1086 if (!sg) in build_overlap_sched_groups() 1193 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group() 1218 return sg; in get_group() 1307 sg = sg->next; in init_sched_groups_capacity() 2234 sg->next = sg; in __sdt_alloc() [all …]
|
/openbmc/linux/include/linux/ |
H A D | skmsg.h | 44 struct sk_msg_sg sg; member 186 dst->sg.data[which] = src->sg.data[which]; in sk_msg_xfer() 188 dst->sg.size += size; in sk_msg_xfer() 189 src->sg.size -= size; in sk_msg_xfer() 202 return sk_msg_iter_dist(msg->sg.start, msg->sg.end) == MAX_MSG_FRAGS; in sk_msg_full() 207 return sk_msg_iter_dist(msg->sg.start, msg->sg.end); in sk_msg_elem_used() 217 return msg->sg.data[which]; in sk_msg_elem_cpy() 234 if (test_bit(msg->sg.start, msg->sg.copy)) { in sk_msg_compute_data_pointers() 253 __set_bit(msg->sg.end, msg->sg.copy); in sk_msg_page_add() 254 msg->sg.size += len; in sk_msg_page_add() [all …]
|
/openbmc/linux/drivers/scsi/qla2xxx/ |
H A D | qla_dsd.h | 12 static inline void append_dsd32(struct dsd32 **dsd, struct scatterlist *sg) in append_dsd32() argument 14 put_unaligned_le32(sg_dma_address(sg), &(*dsd)->address); in append_dsd32() 15 put_unaligned_le32(sg_dma_len(sg), &(*dsd)->length); in append_dsd32() 25 static inline void append_dsd64(struct dsd64 **dsd, struct scatterlist *sg) in append_dsd64() argument 27 put_unaligned_le64(sg_dma_address(sg), &(*dsd)->address); in append_dsd64() 28 put_unaligned_le32(sg_dma_len(sg), &(*dsd)->length); in append_dsd64()
|
/openbmc/linux/arch/mips/jazz/ |
H A D | jazzdma.c | 547 struct scatterlist *sg; in jazz_dma_map_sg() local 551 arch_sync_dma_for_device(sg_phys(sg), sg->length, in jazz_dma_map_sg() 553 sg->dma_address = vdma_alloc(sg_phys(sg), sg->length); in jazz_dma_map_sg() 556 sg_dma_len(sg) = sg->length; in jazz_dma_map_sg() 566 struct scatterlist *sg; in jazz_dma_unmap_sg() local 570 arch_sync_dma_for_cpu(sg_phys(sg), sg->length, dir); in jazz_dma_unmap_sg() 571 vdma_free(sg->dma_address); in jazz_dma_unmap_sg() 590 struct scatterlist *sg; in jazz_dma_sync_sg_for_device() local 594 arch_sync_dma_for_device(sg_phys(sg), sg->length, dir); in jazz_dma_sync_sg_for_device() 600 struct scatterlist *sg; in jazz_dma_sync_sg_for_cpu() local [all …]
|
/openbmc/linux/drivers/crypto/ |
H A D | s5p-sss.c | 454 if (!*sg) in s5p_free_sg_cpy() 460 kfree(*sg); in s5p_free_sg_cpy() 461 *sg = NULL; in s5p_free_sg_cpy() 1078 ctx->sg = kmalloc_array(n, sizeof(*sg), GFP_KERNEL); in s5p_hash_copy_sg_lists() 1096 while (sg && skip >= sg->length) { in s5p_hash_copy_sg_lists() 1098 sg = sg_next(sg); in s5p_hash_copy_sg_lists() 1107 sg_set_page(tmp, sg_page(sg), len, sg->offset + skip); in s5p_hash_copy_sg_lists() 1114 sg = sg_next(sg); in s5p_hash_copy_sg_lists() 1146 if (!sg || !sg->length || !new_len) in s5p_hash_prepare_sgs() 1195 ctx->sg = sg; in s5p_hash_prepare_sgs() [all …]
|
/openbmc/linux/drivers/ata/ |
H A D | pata_ftide010.c | 49 struct sata_gemini *sg; member 273 struct sata_gemini *sg = ftide->sg; in pata_ftide010_gemini_port_start() local 316 struct sata_gemini *sg = ftide->sg; in pata_ftide010_gemini_port_stop() local 320 gemini_sata_stop_bridge(sg, 0); in pata_ftide010_gemini_port_stop() 324 gemini_sata_stop_bridge(sg, 1); in pata_ftide010_gemini_port_stop() 329 gemini_sata_stop_bridge(sg, 0); in pata_ftide010_gemini_port_stop() 354 struct sata_gemini *sg; in pata_ftide010_gemini_init() local 358 sg = gemini_sata_bridge_get(); in pata_ftide010_gemini_init() 359 if (IS_ERR(sg)) in pata_ftide010_gemini_init() 360 return PTR_ERR(sg); in pata_ftide010_gemini_init() [all …]
|
/openbmc/linux/net/sunrpc/auth_gss/ |
H A D | gss_krb5_crypto.c | 132 skcipher_request_set_crypt(req, sg, sg, length, local_iv); in krb5_encrypt() 186 skcipher_request_set_crypt(req, sg, sg, length, local_iv); in krb5_decrypt() 200 ahash_request_set_crypt(req, sg, NULL, sg->length); in checksummer() 405 sg->offset); in encryptor() 406 sg_set_page(&desc->outfrags[desc->fragno], sg_page(sg), sg->length, in encryptor() 407 sg->offset); in encryptor() 433 sg->offset + sg->length - fraglen); in encryptor() 494 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor() 495 sg->offset); in decryptor() 518 sg->offset + sg->length - fraglen); in decryptor() [all …]
|
/openbmc/linux/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_psfp.c | 132 const struct sparx5_psfp_sg *sg) in sparx5_psfp_sg_set() argument 138 ips = sparx5_psfp_ipv_to_ips(sg->ipv); in sparx5_psfp_sg_set() 139 base_lsb = sg->basetime.tv_sec & 0xffffffff; in sparx5_psfp_sg_set() 140 base_msb = sg->basetime.tv_sec >> 32; in sparx5_psfp_sg_set() 166 for (i = 0; i < sg->num_entries; i++) { in sparx5_psfp_sg_set() 167 gce = &sg->gce[i]; in sparx5_psfp_sg_set() 227 struct sparx5_psfp_sg *sg, u32 *id) in sparx5_psfp_sg_add() argument 241 sg->basetime = ktime_to_timespec64(basetime); in sparx5_psfp_sg_add() 243 sparx5_psfp_sg_set(sparx5, *id, sg); in sparx5_psfp_sg_add() 253 const struct sparx5_psfp_sg sg = { 0 }; in sparx5_psfp_sg_del() local [all …]
|
/openbmc/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_bl.c | 67 struct scatterlist *sg; in __qat_bl_sgl_to_bufl() local 99 if (!sg->length) in __qat_bl_sgl_to_bufl() 102 if (left >= sg->length) { in __qat_bl_sgl_to_bufl() 103 left -= sg->length; in __qat_bl_sgl_to_bufl() 154 if (!sg->length) in __qat_bl_sgl_to_bufl() 158 left -= sg->length; in __qat_bl_sgl_to_bufl() 162 sg->length - left, in __qat_bl_sgl_to_bufl() 273 struct scatterlist *sg; in qat_bl_sgl_map() local 290 if (!sg->length) in qat_bl_sgl_map() 294 sg->length, in qat_bl_sgl_map() [all …]
|
/openbmc/linux/crypto/ |
H A D | af_alg.c | 613 sg = sgl->sg; in af_alg_alloc_tsgl() 625 if (sg) in af_alg_alloc_tsgl() 626 sg_chain(sg, MAX_SGL_ENTS + 1, sgl->sg); in af_alg_alloc_tsgl() 657 const struct scatterlist *sg = sgl->sg; in af_alg_count_tsgl() local 713 sg = sgl->sg; in af_alg_pull_tsgl() 1001 sg = sgl->sg + sgl->cur - 1; in af_alg_sendmsg() 1003 PAGE_SIZE - sg->offset - sg->length); in af_alg_sendmsg() 1006 sg->offset + sg->length, in af_alg_sendmsg() 1012 ctx->merge = (sg->offset + sg->length) & in af_alg_sendmsg() 1036 sg = sgl->sg; in af_alg_sendmsg() [all …]
|
/openbmc/linux/net/rds/ |
H A D | tcp_send.c | 71 unsigned int hdr_off, unsigned int sg, unsigned int off) in rds_tcp_xmit() argument 115 while (sg < rm->data.op_nents) { in rds_tcp_xmit() 117 if (sg + 1 < rm->data.op_nents) in rds_tcp_xmit() 120 bvec_set_page(&bvec, sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit() 121 rm->data.op_sg[sg].length - off, in rds_tcp_xmit() 122 rm->data.op_sg[sg].offset + off); in rds_tcp_xmit() 124 rm->data.op_sg[sg].length - off); in rds_tcp_xmit() 126 rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit() 127 rm->data.op_sg[sg].offset + off, rm->data.op_sg[sg].length - off, in rds_tcp_xmit() 134 if (off == rm->data.op_sg[sg].length) { in rds_tcp_xmit() [all …]
|