/openbmc/linux/drivers/vfio/pci/pds/ |
H A D | dirty.c | 111 kfree(bmp_info->sgl); in __pds_vfio_dirty_free_sgl() 114 bmp_info->sgl = NULL; in __pds_vfio_dirty_free_sgl() 120 if (pds_vfio->dirty.host_seq.sgl) in pds_vfio_dirty_free_sgl() 122 if (pds_vfio->dirty.host_ack.sgl) in pds_vfio_dirty_free_sgl() 132 struct pds_lm_sg_elem *sgl; in __pds_vfio_dirty_alloc_sgl() local 140 sgl = kzalloc(sgl_size, GFP_KERNEL); in __pds_vfio_dirty_alloc_sgl() 141 if (!sgl) in __pds_vfio_dirty_alloc_sgl() 144 sgl_addr = dma_map_single(pdsc_dev, sgl, sgl_size, DMA_BIDIRECTIONAL); in __pds_vfio_dirty_alloc_sgl() 146 kfree(sgl); in __pds_vfio_dirty_alloc_sgl() 150 bmp_info->sgl = sgl; in __pds_vfio_dirty_alloc_sgl() [all …]
|
/openbmc/linux/crypto/ |
H A D | algif_aead.c | 213 rsgl_src = areq->first_rsgl.sgl.sgt.sgl; in _aead_recvmsg() 227 areq->first_rsgl.sgl.sgt.sgl, in _aead_recvmsg() 246 areq->first_rsgl.sgl.sgt.sgl, in _aead_recvmsg() 271 struct af_alg_sgl *sgl_prev = &areq->last_rsgl->sgl; in _aead_recvmsg() 272 struct scatterlist *sg = sgl_prev->sgt.sgl; in _aead_recvmsg() 283 areq->first_rsgl.sgl.sgt.sgl, used, ctx->iv); in _aead_recvmsg()
|
/openbmc/linux/drivers/dma/ |
H A D | milbeaut-hdmac.c | 61 struct scatterlist *sgl; member 122 sg = &md->sgl[md->sg_cur]; in milbeaut_chan_start() 256 milbeaut_hdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, in milbeaut_hdmac_prep_slave_sg() argument 272 md->sgl = kcalloc(sg_len, sizeof(*sgl), GFP_NOWAIT); in milbeaut_hdmac_prep_slave_sg() 273 if (!md->sgl) { in milbeaut_hdmac_prep_slave_sg() 279 md->sgl[i] = sgl[i]; in milbeaut_hdmac_prep_slave_sg() 350 sg = &md->sgl[md->sg_cur]; in milbeaut_hdmac_tx_status() 372 txstate->residue += sg_dma_len(&md->sgl[i]); in milbeaut_hdmac_tx_status() 398 kfree(md->sgl); in milbeaut_hdmac_desc_free()
|
H A D | altera-msgdma.c | 388 msgdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, in msgdma_prep_slave_sg() argument 404 for_each_sg(sgl, sg, sg_len, i) in msgdma_prep_slave_sg() 416 avail = sg_dma_len(sgl); in msgdma_prep_slave_sg() 427 dma_src = sg_dma_address(sgl) + sg_dma_len(sgl) - avail; in msgdma_prep_slave_sg() 432 dma_dst = sg_dma_address(sgl) + sg_dma_len(sgl) - avail; in msgdma_prep_slave_sg() 447 sgl = sg_next(sgl); in msgdma_prep_slave_sg() 448 if (sgl == NULL) in msgdma_prep_slave_sg() 451 avail = sg_dma_len(sgl); in msgdma_prep_slave_sg()
|
H A D | dma-axi-dmac.c | 540 struct dma_chan *c, struct scatterlist *sgl, in axi_dmac_prep_slave_sg() argument 555 for_each_sg(sgl, sg, sg_len, i) in axi_dmac_prep_slave_sg() 564 for_each_sg(sgl, sg, sg_len, i) { in axi_dmac_prep_slave_sg() 638 dst_icg = dmaengine_get_dst_icg(xt, &xt->sgl[0]); in axi_dmac_prep_interleaved() 639 src_icg = dmaengine_get_src_icg(xt, &xt->sgl[0]); in axi_dmac_prep_interleaved() 642 if (!axi_dmac_check_len(chan, xt->sgl[0].size) || in axi_dmac_prep_interleaved() 645 if (xt->sgl[0].size + dst_icg > chan->max_length || in axi_dmac_prep_interleaved() 646 xt->sgl[0].size + src_icg > chan->max_length) in axi_dmac_prep_interleaved() 651 if (chan->max_length / xt->sgl[0].size < xt->numf) in axi_dmac_prep_interleaved() 672 desc->sg[0].x_len = xt->sgl[0].size; in axi_dmac_prep_interleaved() [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_phys.c | 56 sg = st->sgl; in i915_gem_object_get_pages_phys() 103 dma_addr_t dma = sg_dma_address(pages->sgl); in i915_gem_object_put_pages_phys() 104 void *vaddr = sg_page(pages->sgl); in i915_gem_object_put_pages_phys() 147 void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset; in i915_gem_object_pwrite_phys() 178 void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset; in i915_gem_object_pread_phys()
|
/openbmc/linux/arch/sparc/mm/ |
H A D | iommu.c | 249 static int __sbus_iommu_map_sg(struct device *dev, struct scatterlist *sgl, in __sbus_iommu_map_sg() argument 256 for_each_sg(sgl, sg, nents, j) { in __sbus_iommu_map_sg() 267 static int sbus_iommu_map_sg_gflush(struct device *dev, struct scatterlist *sgl, in sbus_iommu_map_sg_gflush() argument 271 return __sbus_iommu_map_sg(dev, sgl, nents, dir, attrs, false); in sbus_iommu_map_sg_gflush() 274 static int sbus_iommu_map_sg_pflush(struct device *dev, struct scatterlist *sgl, in sbus_iommu_map_sg_pflush() argument 277 return __sbus_iommu_map_sg(dev, sgl, nents, dir, attrs, true); in sbus_iommu_map_sg_pflush() 299 static void sbus_iommu_unmap_sg(struct device *dev, struct scatterlist *sgl, in sbus_iommu_unmap_sg() argument 305 for_each_sg(sgl, sg, nents, i) { in sbus_iommu_unmap_sg()
|
/openbmc/linux/drivers/dma/sh/ |
H A D | shdma-base.c | 571 for_each_sg(sgl, sg, sg_len, i) in shdma_prep_sg() 588 for_each_sg(sgl, sg, sg_len, i) { in shdma_prep_sg() 685 return shdma_prep_sg(schan, sgl, sg_len, &slave_addr, in shdma_prep_slave_sg() 703 struct scatterlist *sgl; in shdma_prep_dma_cyclic() local 731 sgl = kmalloc_array(sg_len, sizeof(*sgl), GFP_KERNEL); in shdma_prep_dma_cyclic() 732 if (!sgl) in shdma_prep_dma_cyclic() 735 sg_init_table(sgl, sg_len); in shdma_prep_dma_cyclic() 742 sg_dma_address(&sgl[i]) = src; in shdma_prep_dma_cyclic() 743 sg_dma_len(&sgl[i]) = period_len; in shdma_prep_dma_cyclic() 746 desc = shdma_prep_sg(schan, sgl, sg_len, &slave_addr, in shdma_prep_dma_cyclic() [all …]
|
H A D | rcar-dmac.c | 964 for_each_sg(sgl, sg, sg_len, i) { in rcar_dmac_chan_prep_sg() 1131 struct scatterlist sgl; in rcar_dmac_prep_dma_memcpy() local 1136 sg_init_table(&sgl, 1); in rcar_dmac_prep_dma_memcpy() 1139 sg_dma_address(&sgl) = dma_src; in rcar_dmac_prep_dma_memcpy() 1140 sg_dma_len(&sgl) = len; in rcar_dmac_prep_dma_memcpy() 1230 struct scatterlist *sgl; in rcar_dmac_prep_dma_cyclic() local 1257 sgl = kmalloc_array(sg_len, sizeof(*sgl), GFP_NOWAIT); in rcar_dmac_prep_dma_cyclic() 1258 if (!sgl) in rcar_dmac_prep_dma_cyclic() 1261 sg_init_table(sgl, sg_len); in rcar_dmac_prep_dma_cyclic() 1268 sg_dma_address(&sgl[i]) = src; in rcar_dmac_prep_dma_cyclic() [all …]
|
/openbmc/linux/drivers/scsi/elx/efct/ |
H A D | efct_scsi.c | 202 struct efct_scsi_sgl *sgl, u32 sgl_count, in efct_scsi_build_sgls() argument 218 rc = efct_hw_io_add_sge(hw, hio, sgl[i].addr, sgl[i].len); in efct_scsi_build_sgls() 303 io->sgl, io->sgl_count, io->hio_type); in efct_scsi_io_dispatch_hw_io() 672 struct efct_scsi_sgl *sgl_ptr = &io->sgl[sgl_count - 1]; in efct_scsi_xfer_data() 715 return efct_scsi_xfer_data(io, flags, sgl, sgl_count, in efct_scsi_send_rd_data() 824 io->sgl[0].addr = io->rspbuf.phys; in efct_scsi_send_resp() 825 io->sgl[0].dif_addr = 0; in efct_scsi_send_resp() 826 io->sgl[0].len = io->wire_len; in efct_scsi_send_resp() 1021 io->sgl[0].addr = io->rspbuf.phys; in efct_scsi_send_tmf_resp() 1022 io->sgl[0].dif_addr = 0; in efct_scsi_send_tmf_resp() [all …]
|
H A D | efct_io.c | 59 io->sgl = kzalloc(sizeof(*io->sgl) * num_sgl, GFP_KERNEL); in efct_io_pool_create() 60 if (!io->sgl) { in efct_io_pool_create() 90 kfree(io->sgl); in efct_io_pool_free()
|
/openbmc/linux/drivers/scsi/bnx2fc/ |
H A D | bnx2fc_hwi.c | 1496 struct fcoe_ext_mul_sges_ctx *sgl; in bnx2fc_init_seq_cleanup_task() local 1555 sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl; in bnx2fc_init_seq_cleanup_task() 1558 sgl->mul_sgl.sgl_size = bd_count; in bnx2fc_init_seq_cleanup_task() 1560 sgl->mul_sgl.cur_sge_idx = i; in bnx2fc_init_seq_cleanup_task() 1612 struct fcoe_ext_mul_sges_ctx *sgl; in bnx2fc_init_mp_task() local 1693 sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl; in bnx2fc_init_mp_task() 1696 sgl->mul_sgl.cur_sge_addr.hi = in bnx2fc_init_mp_task() 1698 sgl->mul_sgl.sgl_size = 1; in bnx2fc_init_mp_task() 1811 sgl = &task->rxwr_only.union_ctx.read_info.sgl_ctx.sgl; in bnx2fc_init_task() 1842 sgl->mul_sgl.cur_sge_addr.hi = in bnx2fc_init_task() [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | mem.c | 37 struct scatterlist *sgl; member 117 .sgl = mem->sgl, in nvkm_mem_map_sgl() 178 mem->sgl = args->v0.sgl; in nvkm_mem_new_host()
|
/openbmc/linux/drivers/gpu/drm/virtio/ |
H A D | virtgpu_vram.c | 104 sg_set_page(sgt->sgl, NULL, vram->vram_node.size, 0); in virtio_gpu_vram_map_dma_buf() 105 sg_dma_address(sgt->sgl) = addr; in virtio_gpu_vram_map_dma_buf() 106 sg_dma_len(sgt->sgl) = vram->vram_node.size; in virtio_gpu_vram_map_dma_buf() 120 dma_unmap_resource(dev, sg_dma_address(sgt->sgl), in virtio_gpu_vram_unmap_dma_buf() 121 sg_dma_len(sgt->sgl), dir, in virtio_gpu_vram_unmap_dma_buf()
|
/openbmc/u-boot/include/linux/mtd/ |
H A D | ubi.h | 257 int ubi_leb_read_sg(struct ubi_volume_desc *desc, int lnum, struct ubi_sgl *sgl, 285 struct ubi_sgl *sgl, int offset, int len) in ubi_read_sg() argument 287 return ubi_leb_read_sg(desc, lnum, sgl, offset, len, 0); in ubi_read_sg()
|
/openbmc/linux/include/linux/mtd/ |
H A D | ubi.h | 240 int ubi_leb_read_sg(struct ubi_volume_desc *desc, int lnum, struct ubi_sgl *sgl, 268 struct ubi_sgl *sgl, int offset, int len) in ubi_read_sg() argument 270 return ubi_leb_read_sg(desc, lnum, sgl, offset, len, 0); in ubi_read_sg()
|
/openbmc/linux/kernel/dma/ |
H A D | direct.c | 410 struct scatterlist *sgl, int nents, enum dma_data_direction dir) in dma_direct_sync_sg_for_device() argument 415 for_each_sg(sgl, sg, nents, i) { in dma_direct_sync_sg_for_device() 433 struct scatterlist *sgl, int nents, enum dma_data_direction dir) in dma_direct_sync_sg_for_cpu() argument 438 for_each_sg(sgl, sg, nents, i) { in dma_direct_sync_sg_for_cpu() 460 void dma_direct_unmap_sg(struct device *dev, struct scatterlist *sgl, in dma_direct_unmap_sg() argument 466 for_each_sg(sgl, sg, nents, i) { in dma_direct_unmap_sg() 476 int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, int nents, in dma_direct_map_sg() argument 484 for_each_sg(sgl, sg, nents, i) { in dma_direct_map_sg() 516 dma_direct_unmap_sg(dev, sgl, i, dir, attrs | DMA_ATTR_SKIP_CPU_SYNC); in dma_direct_map_sg() 545 sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0); in dma_direct_get_sgtable()
|
/openbmc/linux/drivers/scsi/lpfc/ |
H A D | lpfc_nvmet.c | 2828 sgl->word2 = cpu_to_le32(sgl->word2); in lpfc_nvmet_prep_fcp_wqe() 2830 sgl++; in lpfc_nvmet_prep_fcp_wqe() 2835 sgl->word2 = cpu_to_le32(sgl->word2); in lpfc_nvmet_prep_fcp_wqe() 2837 sgl++; in lpfc_nvmet_prep_fcp_wqe() 2914 sgl->word2 = cpu_to_le32(sgl->word2); in lpfc_nvmet_prep_fcp_wqe() 2916 sgl++; in lpfc_nvmet_prep_fcp_wqe() 2921 sgl->word2 = cpu_to_le32(sgl->word2); in lpfc_nvmet_prep_fcp_wqe() 2923 sgl++; in lpfc_nvmet_prep_fcp_wqe() 3006 sgl->word2 = cpu_to_le32(sgl->word2); in lpfc_nvmet_prep_fcp_wqe() 3008 sgl++; in lpfc_nvmet_prep_fcp_wqe() [all …]
|
/openbmc/linux/lib/ |
H A D | sg_pool.c | 54 static void sg_pool_free(struct scatterlist *sgl, unsigned int nents) in sg_pool_free() argument 59 mempool_free(sgl, sgp->pool); in sg_pool_free() 122 sg_init_table(table->sgl, nents); in sg_alloc_table_chained()
|
/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_scatterlist.c | 27 new_sg = new_st.sgl; in i915_sg_trim() 28 for_each_sg(orig_st->sgl, sg, orig_st->nents, i) { in i915_sg_trim() 112 sg = st->sgl; in i915_rsgt_from_mm_node() 198 sg = st->sgl; in i915_rsgt_from_buddy_resource()
|
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_gem_object.c | 50 sg = pages->sgl; in huge_get_pages() 64 for (end = sg, src = pages->sgl; sg; sg = __sg_next(sg)) { in huge_get_pages() 68 src = pages->sgl; in huge_get_pages()
|
/openbmc/linux/drivers/scsi/elx/libefc_sli/ |
H A D | sli4.c | 1573 if (!sgl || !sgl->virt) { in sli_fcp_icmnd64_wqe() 1575 sgl, sgl ? sgl->virt : NULL); in sli_fcp_icmnd64_wqe() 1640 if (!sgl || !sgl->virt) { in sli_fcp_iread64_wqe() 1642 sgl, sgl ? sgl->virt : NULL); in sli_fcp_iread64_wqe() 1743 if (!sgl || !sgl->virt) { in sli_fcp_iwrite64_wqe() 1745 sgl, sgl ? sgl->virt : NULL); in sli_fcp_iwrite64_wqe() 1834 if (!sgl || !sgl->virt) { in sli_fcp_treceive64_wqe() 1836 sgl, sgl ? sgl->virt : NULL); in sli_fcp_treceive64_wqe() 2037 if (!sgl || !sgl->virt) { in sli_fcp_tsend64_wqe() 2039 sgl, sgl ? sgl->virt : NULL); in sli_fcp_tsend64_wqe() [all …]
|
/openbmc/linux/drivers/spi/ |
H A D | spi-dw-dma.c | 339 static int dw_spi_dma_submit_tx(struct dw_spi *dws, struct scatterlist *sgl, in dw_spi_dma_submit_tx() argument 346 txdesc = dmaengine_prep_slave_sg(dws->txchan, sgl, nents, in dw_spi_dma_submit_tx() 440 static int dw_spi_dma_submit_rx(struct dw_spi *dws, struct scatterlist *sgl, in dw_spi_dma_submit_rx() argument 447 rxdesc = dmaengine_prep_slave_sg(dws->rxchan, sgl, nents, in dw_spi_dma_submit_rx() 512 ret = dw_spi_dma_submit_tx(dws, xfer->tx_sg.sgl, xfer->tx_sg.nents); in dw_spi_dma_transfer_all() 518 ret = dw_spi_dma_submit_rx(dws, xfer->rx_sg.sgl, in dw_spi_dma_transfer_all() 583 tx_sg = !tx_sg ? &xfer->tx_sg.sgl[0] : sg_next(tx_sg); in dw_spi_dma_transfer_one() 590 rx_sg = !rx_sg ? &xfer->rx_sg.sgl[0] : sg_next(rx_sg); in dw_spi_dma_transfer_one()
|
/openbmc/linux/drivers/usb/usbip/ |
H A D | stub_rx.c | 470 struct scatterlist *sgl = NULL, *sg; in stub_recv_cmd_submit() local 504 sgl = sgl_alloc(buf_len, GFP_KERNEL, &nents); in stub_recv_cmd_submit() 505 if (!sgl) in stub_recv_cmd_submit() 553 priv->urbs[0]->sg = sgl; in stub_recv_cmd_submit() 571 for_each_sg(sgl, sg, nents, i) { in stub_recv_cmd_submit() 583 priv->sgl = sgl; in stub_recv_cmd_submit() 646 sgl_free(sgl); in stub_recv_cmd_submit()
|
/openbmc/linux/include/crypto/ |
H A D | if_alg.h | 60 struct scatterlist sgl[ALG_MAX_PAGES + 1]; member 76 struct af_alg_sgl sgl; member 169 void af_alg_free_sg(struct af_alg_sgl *sgl);
|