Skip to content

Commit

Permalink
iommu/dma: Use for_each_sg in iommu_dma_alloc
Browse files Browse the repository at this point in the history
arch_dma_prep_coherent can handle physically contiguous ranges larger
than PAGE_SIZE just fine, which means we don't need a page-based
iterator.

Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Robin Murphy <robin.murphy@arm.com>
Signed-off-by: Joerg Roedel <jroedel@suse.de>
  • Loading branch information
Christoph Hellwig authored and Joerg Roedel committed May 27, 2019
1 parent af751d4 commit 23f88e0
Showing 1 changed file with 5 additions and 9 deletions.
14 changes: 5 additions & 9 deletions drivers/iommu/dma-iommu.c
Original file line number Diff line number Diff line change
Expand Up @@ -606,15 +606,11 @@ struct page **iommu_dma_alloc(struct device *dev, size_t size, gfp_t gfp,
goto out_free_iova;

if (!(prot & IOMMU_CACHE)) {
struct sg_mapping_iter miter;
/*
* The CPU-centric flushing implied by SG_MITER_TO_SG isn't
* sufficient here, so skip it by using the "wrong" direction.
*/
sg_miter_start(&miter, sgt.sgl, sgt.orig_nents, SG_MITER_FROM_SG);
while (sg_miter_next(&miter))
arch_dma_prep_coherent(miter.page, PAGE_SIZE);
sg_miter_stop(&miter);
struct scatterlist *sg;
int i;

for_each_sg(sgt.sgl, sg, sgt.orig_nents, i)
arch_dma_prep_coherent(sg_page(sg), sg->length);
}

if (iommu_map_sg(domain, iova, sgt.sgl, sgt.orig_nents, prot)
Expand Down

0 comments on commit 23f88e0

Please sign in to comment.