summaryrefslogtreecommitdiffstats
path: root/arch/arc/mm/dma.c
diff options
context:
space:
mode:
authorVineet Gupta <vgupta@synopsys.com>2016-03-14 15:03:59 +0530
committerVineet Gupta <vgupta@synopsys.com>2016-03-19 14:34:08 +0530
commitd98a15a5653bfae5bccc68a06a60ccf035b2c4cc (patch)
tree4e4c11c4c5bff82b61e44f6066ed8af6755add2b /arch/arc/mm/dma.c
parenta69fe1a2c9b2bd35046cb8f4e61337825cef9239 (diff)
downloadop-kernel-dev-d98a15a5653bfae5bccc68a06a60ccf035b2c4cc.zip
op-kernel-dev-d98a15a5653bfae5bccc68a06a60ccf035b2c4cc.tar.gz
ARC: dma: Use struct page based page allocator helpers
vs. the ones which reutne void *, so that we can handle pages > 4GB in subsequent patches Also plug a potential page leak in case ioremap fails Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
Diffstat (limited to 'arch/arc/mm/dma.c')
-rw-r--r--arch/arc/mm/dma.c25
1 files changed, 17 insertions, 8 deletions
diff --git a/arch/arc/mm/dma.c b/arch/arc/mm/dma.c
index 01eaf88..6cf1d92 100644
--- a/arch/arc/mm/dma.c
+++ b/arch/arc/mm/dma.c
@@ -24,15 +24,20 @@
static void *arc_dma_alloc(struct device *dev, size_t size,
dma_addr_t *dma_handle, gfp_t gfp, struct dma_attrs *attrs)
{
- void *paddr, *kvaddr;
+ unsigned long order = get_order(size);
+ struct page *page;
+ phys_addr_t paddr;
+ void *kvaddr;
- /* This is linear addr (0x8000_0000 based) */
- paddr = alloc_pages_exact(size, gfp);
- if (!paddr)
+ page = alloc_pages(gfp, order);
+ if (!page)
return NULL;
- /* This is bus address, platform dependent */
- *dma_handle = (dma_addr_t)paddr;
+ /* This is linear addr (0x8000_0000 based) */
+ paddr = page_to_phys(page);
+
+ /* For now bus address is exactly same as paddr */
+ *dma_handle = paddr;
/*
* IOC relies on all data (even coherent DMA data) being in cache
@@ -51,8 +56,10 @@ static void *arc_dma_alloc(struct device *dev, size_t size,
/* This is kernel Virtual address (0x7000_0000 based) */
kvaddr = ioremap_nocache((unsigned long)paddr, size);
- if (kvaddr == NULL)
+ if (kvaddr == NULL) {
+ __free_pages(page, order);
return NULL;
+ }
/*
* Evict any existing L1 and/or L2 lines for the backing page
@@ -72,11 +79,13 @@ static void *arc_dma_alloc(struct device *dev, size_t size,
static void arc_dma_free(struct device *dev, size_t size, void *vaddr,
dma_addr_t dma_handle, struct dma_attrs *attrs)
{
+ struct page *page = virt_to_page(dma_handle);
+
if (!dma_get_attr(DMA_ATTR_NON_CONSISTENT, attrs) &&
!(is_isa_arcv2() && ioc_exists))
iounmap((void __force __iomem *)vaddr);
- free_pages_exact((void *)dma_handle, size);
+ __free_pages(page, get_order(size));
}
/*
OpenPOWER on IntegriCloud