Skip to content

Commit

Permalink
---
Browse files Browse the repository at this point in the history
yaml
---
r: 112582
b: refs/heads/master
c: 2797982
h: refs/heads/master
v: v3
  • Loading branch information
FUJITA Tomonori authored and Ingo Molnar committed Sep 10, 2008
1 parent 76a178f commit 0aa0676
Show file tree
Hide file tree
Showing 2 changed files with 9 additions and 8 deletions.
2 changes: 1 addition & 1 deletion [refs]
Original file line number Diff line number Diff line change
@@ -1,2 +1,2 @@
---
refs/heads/master: 49fbf4e9f982c704dc365698c5b5efa780aadcb5
refs/heads/master: 2797982ed93c10d5585ee1842ab298cb11326ff5
15 changes: 8 additions & 7 deletions trunk/lib/swiotlb.c
Original file line number Diff line number Diff line change
Expand Up @@ -274,13 +274,13 @@ swiotlb_late_init_with_default_size(size_t default_size)
}

static int
address_needs_mapping(struct device *hwdev, dma_addr_t addr)
address_needs_mapping(struct device *hwdev, dma_addr_t addr, size_t size)
{
dma_addr_t mask = 0xffffffff;
/* If the device has a mask, use it, otherwise default to 32 bits */
if (hwdev && hwdev->dma_mask)
mask = *hwdev->dma_mask;
return (addr & ~mask) != 0;
return !is_buffer_dma_capable(mask, addr, size);
}

static int is_swiotlb_buffer(char *addr)
Expand Down Expand Up @@ -473,7 +473,7 @@ swiotlb_alloc_coherent(struct device *hwdev, size_t size,
int order = get_order(size);

ret = (void *)__get_free_pages(flags, order);
if (ret && address_needs_mapping(hwdev, virt_to_bus(ret))) {
if (ret && address_needs_mapping(hwdev, virt_to_bus(ret), size)) {
/*
* The allocated memory isn't reachable by the device.
* Fall back on swiotlb_map_single().
Expand All @@ -497,7 +497,7 @@ swiotlb_alloc_coherent(struct device *hwdev, size_t size,
dev_addr = virt_to_bus(ret);

/* Confirm address can be DMA'd by device */
if (address_needs_mapping(hwdev, dev_addr)) {
if (address_needs_mapping(hwdev, dev_addr, size)) {
printk("hwdev DMA mask = 0x%016Lx, dev_addr = 0x%016Lx\n",
(unsigned long long)*hwdev->dma_mask,
(unsigned long long)dev_addr);
Expand Down Expand Up @@ -561,7 +561,7 @@ swiotlb_map_single_attrs(struct device *hwdev, void *ptr, size_t size,
* we can safely return the device addr and not worry about bounce
* buffering it.
*/
if (!address_needs_mapping(hwdev, dev_addr) && !swiotlb_force)
if (!address_needs_mapping(hwdev, dev_addr, size) && !swiotlb_force)
return dev_addr;

/*
Expand All @@ -578,7 +578,7 @@ swiotlb_map_single_attrs(struct device *hwdev, void *ptr, size_t size,
/*
* Ensure that the address returned is DMA'ble
*/
if (address_needs_mapping(hwdev, dev_addr))
if (address_needs_mapping(hwdev, dev_addr, size))
panic("map_single: bounce buffer is not DMA'ble");

return dev_addr;
Expand Down Expand Up @@ -721,7 +721,8 @@ swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems,
for_each_sg(sgl, sg, nelems, i) {
addr = SG_ENT_VIRT_ADDRESS(sg);
dev_addr = virt_to_bus(addr);
if (swiotlb_force || address_needs_mapping(hwdev, dev_addr)) {
if (swiotlb_force ||
address_needs_mapping(hwdev, dev_addr, sg->length)) {
void *map = map_single(hwdev, addr, sg->length, dir);
if (!map) {
/* Don't panic here, we expect map_sg users
Expand Down

0 comments on commit 0aa0676

Please sign in to comment.