Skip to content

Commit

Permalink
powerpc/iommu: Reduce spinlock coverage in iommu_alloc and iommu_free
Browse files Browse the repository at this point in the history
We currently hold the IOMMU spinlock around tce_build and tce_flush.
This causes our spinlock hold times to be much higher than required
and can impact multiqueue adapters.

This patch moves tce_build and tce_flush outside of the lock in
iommu_alloc, and tce_flush outside of the lock in iommu_free.

Some performance numbers were obtained with a Chelsio T3 adapter on
two POWER7 boxes, running a 100 session TCP round robin test.

Performance improved 32% with this patch applied.

Signed-off-by: Anton Blanchard <anton@samba.org>
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
  • Loading branch information
Anton Blanchard authored and Benjamin Herrenschmidt committed Jul 3, 2012
1 parent c1703e8 commit 0e4bc95
Showing 1 changed file with 5 additions and 10 deletions.
15 changes: 5 additions & 10 deletions arch/powerpc/kernel/iommu.c
Original file line number Diff line number Diff line change
Expand Up @@ -170,13 +170,11 @@ static dma_addr_t iommu_alloc(struct device *dev, struct iommu_table *tbl,
int build_fail;

spin_lock_irqsave(&(tbl->it_lock), flags);

entry = iommu_range_alloc(dev, tbl, npages, NULL, mask, align_order);
spin_unlock_irqrestore(&(tbl->it_lock), flags);

if (unlikely(entry == DMA_ERROR_CODE)) {
spin_unlock_irqrestore(&(tbl->it_lock), flags);
if (unlikely(entry == DMA_ERROR_CODE))
return DMA_ERROR_CODE;
}

entry += tbl->it_offset; /* Offset into real TCE table */
ret = entry << IOMMU_PAGE_SHIFT; /* Set the return dma address */
Expand All @@ -192,18 +190,17 @@ static dma_addr_t iommu_alloc(struct device *dev, struct iommu_table *tbl,
* not altered.
*/
if (unlikely(build_fail)) {
spin_lock_irqsave(&(tbl->it_lock), flags);
__iommu_free(tbl, ret, npages);

spin_unlock_irqrestore(&(tbl->it_lock), flags);

return DMA_ERROR_CODE;
}

/* Flush/invalidate TLB caches if necessary */
if (ppc_md.tce_flush)
ppc_md.tce_flush(tbl);

spin_unlock_irqrestore(&(tbl->it_lock), flags);

/* Make sure updates are seen by hardware */
mb();

Expand Down Expand Up @@ -244,17 +241,15 @@ static void iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr,
unsigned long flags;

spin_lock_irqsave(&(tbl->it_lock), flags);

__iommu_free(tbl, dma_addr, npages);
spin_unlock_irqrestore(&(tbl->it_lock), flags);

/* Make sure TLB cache is flushed if the HW needs it. We do
* not do an mb() here on purpose, it is not needed on any of
* the current platforms.
*/
if (ppc_md.tce_flush)
ppc_md.tce_flush(tbl);

spin_unlock_irqrestore(&(tbl->it_lock), flags);
}

int iommu_map_sg(struct device *dev, struct iommu_table *tbl,
Expand Down

0 comments on commit 0e4bc95

Please sign in to comment.