Lines Matching full:buffers
29 dma_unmap_single(dev, bl->buffers[i].addr, in qat_bl_free_bufl()
30 bl->buffers[i].len, bl_dma_dir); in qat_bl_free_bufl()
39 dma_unmap_single(dev, blout->buffers[i].addr, in qat_bl_free_bufl()
40 blout->buffers[i].len, in qat_bl_free_bufl()
68 size_t sz_out, sz = struct_size(bufl, buffers, n); in __qat_bl_sgl_to_bufl()
92 bufl->buffers[i].addr = DMA_MAPPING_ERROR; in __qat_bl_sgl_to_bufl()
106 bufl->buffers[y].addr = dma_map_single(dev, sg_virt(sg) + left, in __qat_bl_sgl_to_bufl()
109 bufl->buffers[y].len = sg->length; in __qat_bl_sgl_to_bufl()
110 if (unlikely(dma_mapping_error(dev, bufl->buffers[y].addr))) in __qat_bl_sgl_to_bufl()
114 bufl->buffers[y].len -= left; in __qat_bl_sgl_to_bufl()
127 struct qat_alg_buf *buffers; in __qat_bl_sgl_to_bufl() local
132 sz_out = struct_size(buflout, buffers, n); in __qat_bl_sgl_to_bufl()
147 buffers = buflout->buffers; in __qat_bl_sgl_to_bufl()
149 buffers[i].addr = DMA_MAPPING_ERROR; in __qat_bl_sgl_to_bufl()
161 buffers[y].addr = dma_map_single(dev, sg_virt(sg) + left, in __qat_bl_sgl_to_bufl()
164 if (unlikely(dma_mapping_error(dev, buffers[y].addr))) in __qat_bl_sgl_to_bufl()
166 buffers[y].len = sg->length; in __qat_bl_sgl_to_bufl()
169 buffers[y].len -= left; in __qat_bl_sgl_to_bufl()
174 buffers[sg_nctr].addr = extra_dst_buff; in __qat_bl_sgl_to_bufl()
175 buffers[sg_nctr].len = sz_extra_dst_buff; in __qat_bl_sgl_to_bufl()
200 if (buflout->buffers[i].addr == extra_dst_buff) in __qat_bl_sgl_to_bufl()
202 if (!dma_mapping_error(dev, buflout->buffers[i].addr)) in __qat_bl_sgl_to_bufl()
203 dma_unmap_single(dev, buflout->buffers[i].addr, in __qat_bl_sgl_to_bufl()
204 buflout->buffers[i].len, in __qat_bl_sgl_to_bufl()
217 if (!dma_mapping_error(dev, bufl->buffers[i].addr)) in __qat_bl_sgl_to_bufl()
218 dma_unmap_single(dev, bufl->buffers[i].addr, in __qat_bl_sgl_to_bufl()
219 bufl->buffers[i].len, in __qat_bl_sgl_to_bufl()
261 if (!dma_mapping_error(dev, bl->buffers[i].addr)) in qat_bl_sgl_unmap()
262 dma_unmap_single(dev, bl->buffers[i].addr, in qat_bl_sgl_unmap()
263 bl->buffers[i].len, DMA_FROM_DEVICE); in qat_bl_sgl_unmap()
278 sz = struct_size(bufl, buffers, n); in qat_bl_sgl_map()
284 bufl->buffers[i].addr = DMA_MAPPING_ERROR; in qat_bl_sgl_map()
293 bufl->buffers[y].addr = dma_map_single(dev, sg_virt(sg), in qat_bl_sgl_map()
296 bufl->buffers[y].len = sg->length; in qat_bl_sgl_map()
297 if (unlikely(dma_mapping_error(dev, bufl->buffers[y].addr))) in qat_bl_sgl_map()
310 if (!dma_mapping_error(dev, bufl->buffers[i].addr)) in qat_bl_sgl_map()
311 dma_unmap_single(dev, bufl->buffers[i].addr, in qat_bl_sgl_map()
312 bufl->buffers[i].len, in qat_bl_sgl_map()
381 new_bl_size = struct_size(new_bl, buffers, new_bl->num_bufs); in qat_bl_realloc_map_new_dst()