mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-10-31 08:44:41 +00:00 
			
		
		
		
	drivers: dma-contiguous: add initialization from device tree
Add a function to create CMA region from previously reserved memory and add support for handling 'shared-dma-pool' reserved-memory device tree nodes. Based on previous code provided by Josh Cartwright <joshc@codeaurora.org> Signed-off-by: Marek Szyprowski <m.szyprowski@samsung.com> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Michal Nazarewicz <mina86@mina86.com> Cc: Grant Likely <grant.likely@linaro.org> Cc: Laura Abbott <lauraa@codeaurora.org> Cc: Josh Cartwright <joshc@codeaurora.org> Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com> Cc: Kyungmin Park <kyungmin.park@samsung.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									7bfa5ab6fa
								
							
						
					
					
						commit
						de9e14eebf
					
				
					 3 changed files with 120 additions and 11 deletions
				
			
		|  | @ -211,3 +211,69 @@ bool dma_release_from_contiguous(struct device *dev, struct page *pages, | ||||||
| { | { | ||||||
| 	return cma_release(dev_get_cma_area(dev), pages, count); | 	return cma_release(dev_get_cma_area(dev), pages, count); | ||||||
| } | } | ||||||
|  | 
 | ||||||
|  | /*
 | ||||||
|  |  * Support for reserved memory regions defined in device tree | ||||||
|  |  */ | ||||||
|  | #ifdef CONFIG_OF_RESERVED_MEM | ||||||
|  | #include <linux/of.h> | ||||||
|  | #include <linux/of_fdt.h> | ||||||
|  | #include <linux/of_reserved_mem.h> | ||||||
|  | 
 | ||||||
|  | #undef pr_fmt | ||||||
|  | #define pr_fmt(fmt) fmt | ||||||
|  | 
 | ||||||
|  | static void rmem_cma_device_init(struct reserved_mem *rmem, struct device *dev) | ||||||
|  | { | ||||||
|  | 	dev_set_cma_area(dev, rmem->priv); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | static void rmem_cma_device_release(struct reserved_mem *rmem, | ||||||
|  | 				    struct device *dev) | ||||||
|  | { | ||||||
|  | 	dev_set_cma_area(dev, NULL); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | static const struct reserved_mem_ops rmem_cma_ops = { | ||||||
|  | 	.device_init	= rmem_cma_device_init, | ||||||
|  | 	.device_release = rmem_cma_device_release, | ||||||
|  | }; | ||||||
|  | 
 | ||||||
|  | static int __init rmem_cma_setup(struct reserved_mem *rmem) | ||||||
|  | { | ||||||
|  | 	phys_addr_t align = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order); | ||||||
|  | 	phys_addr_t mask = align - 1; | ||||||
|  | 	unsigned long node = rmem->fdt_node; | ||||||
|  | 	struct cma *cma; | ||||||
|  | 	int err; | ||||||
|  | 
 | ||||||
|  | 	if (!of_get_flat_dt_prop(node, "reusable", NULL) || | ||||||
|  | 	    of_get_flat_dt_prop(node, "no-map", NULL)) | ||||||
|  | 		return -EINVAL; | ||||||
|  | 
 | ||||||
|  | 	if ((rmem->base & mask) || (rmem->size & mask)) { | ||||||
|  | 		pr_err("Reserved memory: incorrect alignment of CMA region\n"); | ||||||
|  | 		return -EINVAL; | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	err = cma_init_reserved_mem(rmem->base, rmem->size, 0, &cma); | ||||||
|  | 	if (err) { | ||||||
|  | 		pr_err("Reserved memory: unable to setup CMA region\n"); | ||||||
|  | 		return err; | ||||||
|  | 	} | ||||||
|  | 	/* Architecture specific contiguous memory fixup. */ | ||||||
|  | 	dma_contiguous_early_fixup(rmem->base, rmem->size); | ||||||
|  | 
 | ||||||
|  | 	if (of_get_flat_dt_prop(node, "linux,cma-default", NULL)) | ||||||
|  | 		dma_contiguous_set_default(cma); | ||||||
|  | 
 | ||||||
|  | 	rmem->ops = &rmem_cma_ops; | ||||||
|  | 	rmem->priv = cma; | ||||||
|  | 
 | ||||||
|  | 	pr_info("Reserved memory: created CMA memory pool at %pa, size %ld MiB\n", | ||||||
|  | 		&rmem->base, (unsigned long)rmem->size / SZ_1M); | ||||||
|  | 
 | ||||||
|  | 	return 0; | ||||||
|  | } | ||||||
|  | RESERVEDMEM_OF_DECLARE(cma, "shared-dma-pool", rmem_cma_setup); | ||||||
|  | #endif | ||||||
|  |  | ||||||
|  | @ -22,6 +22,9 @@ extern int __init cma_declare_contiguous(phys_addr_t size, | ||||||
| 			phys_addr_t base, phys_addr_t limit, | 			phys_addr_t base, phys_addr_t limit, | ||||||
| 			phys_addr_t alignment, unsigned int order_per_bit, | 			phys_addr_t alignment, unsigned int order_per_bit, | ||||||
| 			bool fixed, struct cma **res_cma); | 			bool fixed, struct cma **res_cma); | ||||||
|  | extern int cma_init_reserved_mem(phys_addr_t size, | ||||||
|  | 					phys_addr_t base, int order_per_bit, | ||||||
|  | 					struct cma **res_cma); | ||||||
| extern struct page *cma_alloc(struct cma *cma, int count, unsigned int align); | extern struct page *cma_alloc(struct cma *cma, int count, unsigned int align); | ||||||
| extern bool cma_release(struct cma *cma, struct page *pages, int count); | extern bool cma_release(struct cma *cma, struct page *pages, int count); | ||||||
| #endif | #endif | ||||||
|  |  | ||||||
							
								
								
									
										62
									
								
								mm/cma.c
									
										
									
									
									
								
							
							
						
						
									
										62
									
								
								mm/cma.c
									
										
									
									
									
								
							|  | @ -142,6 +142,54 @@ static int __init cma_init_reserved_areas(void) | ||||||
| } | } | ||||||
| core_initcall(cma_init_reserved_areas); | core_initcall(cma_init_reserved_areas); | ||||||
| 
 | 
 | ||||||
|  | /**
 | ||||||
|  |  * cma_init_reserved_mem() - create custom contiguous area from reserved memory | ||||||
|  |  * @base: Base address of the reserved area | ||||||
|  |  * @size: Size of the reserved area (in bytes), | ||||||
|  |  * @order_per_bit: Order of pages represented by one bit on bitmap. | ||||||
|  |  * @res_cma: Pointer to store the created cma region. | ||||||
|  |  * | ||||||
|  |  * This function creates custom contiguous area from already reserved memory. | ||||||
|  |  */ | ||||||
|  | int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size, | ||||||
|  | 				 int order_per_bit, struct cma **res_cma) | ||||||
|  | { | ||||||
|  | 	struct cma *cma; | ||||||
|  | 	phys_addr_t alignment; | ||||||
|  | 
 | ||||||
|  | 	/* Sanity checks */ | ||||||
|  | 	if (cma_area_count == ARRAY_SIZE(cma_areas)) { | ||||||
|  | 		pr_err("Not enough slots for CMA reserved regions!\n"); | ||||||
|  | 		return -ENOSPC; | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	if (!size || !memblock_is_region_reserved(base, size)) | ||||||
|  | 		return -EINVAL; | ||||||
|  | 
 | ||||||
|  | 	/* ensure minimal alignment requied by mm core */ | ||||||
|  | 	alignment = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order); | ||||||
|  | 
 | ||||||
|  | 	/* alignment should be aligned with order_per_bit */ | ||||||
|  | 	if (!IS_ALIGNED(alignment >> PAGE_SHIFT, 1 << order_per_bit)) | ||||||
|  | 		return -EINVAL; | ||||||
|  | 
 | ||||||
|  | 	if (ALIGN(base, alignment) != base || ALIGN(size, alignment) != size) | ||||||
|  | 		return -EINVAL; | ||||||
|  | 
 | ||||||
|  | 	/*
 | ||||||
|  | 	 * Each reserved area must be initialised later, when more kernel | ||||||
|  | 	 * subsystems (like slab allocator) are available. | ||||||
|  | 	 */ | ||||||
|  | 	cma = &cma_areas[cma_area_count]; | ||||||
|  | 	cma->base_pfn = PFN_DOWN(base); | ||||||
|  | 	cma->count = size >> PAGE_SHIFT; | ||||||
|  | 	cma->order_per_bit = order_per_bit; | ||||||
|  | 	*res_cma = cma; | ||||||
|  | 	cma_area_count++; | ||||||
|  | 
 | ||||||
|  | 	return 0; | ||||||
|  | } | ||||||
|  | 
 | ||||||
| /**
 | /**
 | ||||||
|  * cma_declare_contiguous() - reserve custom contiguous area |  * cma_declare_contiguous() - reserve custom contiguous area | ||||||
|  * @base: Base address of the reserved area optional, use 0 for any |  * @base: Base address of the reserved area optional, use 0 for any | ||||||
|  | @ -165,7 +213,6 @@ int __init cma_declare_contiguous(phys_addr_t base, | ||||||
| 			phys_addr_t alignment, unsigned int order_per_bit, | 			phys_addr_t alignment, unsigned int order_per_bit, | ||||||
| 			bool fixed, struct cma **res_cma) | 			bool fixed, struct cma **res_cma) | ||||||
| { | { | ||||||
| 	struct cma *cma; |  | ||||||
| 	phys_addr_t memblock_end = memblock_end_of_DRAM(); | 	phys_addr_t memblock_end = memblock_end_of_DRAM(); | ||||||
| 	phys_addr_t highmem_start = __pa(high_memory); | 	phys_addr_t highmem_start = __pa(high_memory); | ||||||
| 	int ret = 0; | 	int ret = 0; | ||||||
|  | @ -237,16 +284,9 @@ int __init cma_declare_contiguous(phys_addr_t base, | ||||||
| 		} | 		} | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	/*
 | 	ret = cma_init_reserved_mem(base, size, order_per_bit, res_cma); | ||||||
| 	 * Each reserved area must be initialised later, when more kernel | 	if (ret) | ||||||
| 	 * subsystems (like slab allocator) are available. | 		goto err; | ||||||
| 	 */ |  | ||||||
| 	cma = &cma_areas[cma_area_count]; |  | ||||||
| 	cma->base_pfn = PFN_DOWN(base); |  | ||||||
| 	cma->count = size >> PAGE_SHIFT; |  | ||||||
| 	cma->order_per_bit = order_per_bit; |  | ||||||
| 	*res_cma = cma; |  | ||||||
| 	cma_area_count++; |  | ||||||
| 
 | 
 | ||||||
| 	pr_info("Reserved %ld MiB at %08lx\n", (unsigned long)size / SZ_1M, | 	pr_info("Reserved %ld MiB at %08lx\n", (unsigned long)size / SZ_1M, | ||||||
| 		(unsigned long)base); | 		(unsigned long)base); | ||||||
|  |  | ||||||
		Loading…
	
	Add table
		
		Reference in a new issue
	
	 Marek Szyprowski
						Marek Szyprowski