Loading arch/powerpc/include/asm/cpm.h +3 −0 Original line number Diff line number Diff line Loading @@ -2,6 +2,7 @@ #define __CPM_H #include <linux/compiler.h> #include <linux/genalloc.h> #include <linux/types.h> #include <linux/errno.h> #include <linux/of.h> Loading Loading @@ -161,6 +162,8 @@ int cpm_muram_init(void); unsigned long cpm_muram_alloc(unsigned long size, unsigned long align); int cpm_muram_free(unsigned long offset); unsigned long cpm_muram_alloc_fixed(unsigned long offset, unsigned long size); unsigned long cpm_muram_alloc_common(unsigned long size, genpool_algo_t algo, void *data); void __iomem *cpm_muram_addr(unsigned long offset); unsigned long cpm_muram_offset(void __iomem *addr); dma_addr_t cpm_muram_dma(void __iomem *addr); Loading arch/powerpc/platforms/Kconfig +2 −2 Original line number Diff line number Diff line Loading @@ -275,7 +275,7 @@ config TAU_AVERAGE config QUICC_ENGINE bool "Freescale QUICC Engine (QE) Support" depends on FSL_SOC && PPC32 select PPC_LIB_RHEAP select GENERIC_ALLOCATOR select CRC32 help The QUICC Engine (QE) is a new generation of communications Loading @@ -295,7 +295,6 @@ config CPM2 bool "Enable support for the CPM2 (Communications Processor Module)" depends on (FSL_SOC_BOOKE && PPC32) || 8260 select CPM select PPC_LIB_RHEAP select PPC_PCI_CHOICE select ARCH_REQUIRE_GPIOLIB help Loading Loading @@ -325,6 +324,7 @@ config FSL_ULI1575 config CPM bool select GENERIC_ALLOCATOR config OF_RTC bool Loading arch/powerpc/sysdev/cpm_common.c +88 −38 Original line number Diff line number Diff line Loading @@ -17,6 +17,7 @@ * published by the Free Software Foundation. */ #include <linux/genalloc.h> #include <linux/init.h> #include <linux/of_device.h> #include <linux/spinlock.h> Loading @@ -27,7 +28,6 @@ #include <asm/udbg.h> #include <asm/io.h> #include <asm/rheap.h> #include <asm/cpm.h> #include <mm/mmu_decl.h> Loading Loading @@ -65,14 +65,22 @@ void __init udbg_init_cpm(void) } #endif static struct gen_pool *muram_pool; static spinlock_t cpm_muram_lock; static rh_block_t cpm_boot_muram_rh_block[16]; static rh_info_t cpm_muram_info; static u8 __iomem *muram_vbase; static phys_addr_t muram_pbase; /* Max address size we deal with */ struct muram_block { struct list_head head; unsigned long start; int size; }; static LIST_HEAD(muram_block_list); /* max address size we deal with */ #define OF_MAX_ADDR_CELLS 4 #define GENPOOL_OFFSET (4096 * 8) int cpm_muram_init(void) { Loading @@ -87,50 +95,51 @@ int cpm_muram_init(void) return 0; spin_lock_init(&cpm_muram_lock); /* initialize the info header */ rh_init(&cpm_muram_info, 1, sizeof(cpm_boot_muram_rh_block) / sizeof(cpm_boot_muram_rh_block[0]), cpm_boot_muram_rh_block); np = of_find_compatible_node(NULL, NULL, "fsl,cpm-muram-data"); if (!np) { /* try legacy bindings */ np = of_find_node_by_name(NULL, "data-only"); if (!np) { printk(KERN_ERR "Cannot find CPM muram data node"); pr_err("Cannot find CPM muram data node"); ret = -ENODEV; goto out; goto out_muram; } } muram_pool = gen_pool_create(0, -1); muram_pbase = of_translate_address(np, zero); if (muram_pbase == (phys_addr_t)OF_BAD_ADDR) { printk(KERN_ERR "Cannot translate zero through CPM muram node"); pr_err("Cannot translate zero through CPM muram node"); ret = -ENODEV; goto out; goto out_pool; } while (of_address_to_resource(np, i++, &r) == 0) { if (r.end > max) max = r.end; rh_attach_region(&cpm_muram_info, r.start - muram_pbase, resource_size(&r)); ret = gen_pool_add(muram_pool, r.start - muram_pbase + GENPOOL_OFFSET, resource_size(&r), -1); if (ret) { pr_err("QE: couldn't add muram to pool!\n"); goto out_pool; } } muram_vbase = ioremap(muram_pbase, max - muram_pbase + 1); if (!muram_vbase) { printk(KERN_ERR "Cannot map CPM muram"); pr_err("Cannot map QE muram"); ret = -ENOMEM; goto out_pool; } out: goto out_muram; out_pool: gen_pool_destroy(muram_pool); out_muram: of_node_put(np); return ret; } /** /* * cpm_muram_alloc - allocate the requested size worth of multi-user ram * @size: number of bytes to allocate * @align: requested alignment, in bytes Loading @@ -143,14 +152,13 @@ unsigned long cpm_muram_alloc(unsigned long size, unsigned long align) { unsigned long start; unsigned long flags; struct genpool_data_align muram_pool_data; spin_lock_irqsave(&cpm_muram_lock, flags); cpm_muram_info.alignment = align; start = rh_alloc(&cpm_muram_info, size, "commproc"); if (!IS_ERR_VALUE(start)) memset_io(cpm_muram_addr(start), 0, size); muram_pool_data.align = align; start = cpm_muram_alloc_common(size, gen_pool_first_fit_align, &muram_pool_data); spin_unlock_irqrestore(&cpm_muram_lock, flags); return start; } EXPORT_SYMBOL(cpm_muram_alloc); Loading @@ -161,23 +169,31 @@ EXPORT_SYMBOL(cpm_muram_alloc); */ int cpm_muram_free(unsigned long offset) { int ret; unsigned long flags; int size; struct muram_block *tmp; size = 0; spin_lock_irqsave(&cpm_muram_lock, flags); ret = rh_free(&cpm_muram_info, offset); list_for_each_entry(tmp, &muram_block_list, head) { if (tmp->start == offset) { size = tmp->size; list_del(&tmp->head); kfree(tmp); break; } } gen_pool_free(muram_pool, offset + GENPOOL_OFFSET, size); spin_unlock_irqrestore(&cpm_muram_lock, flags); return ret; return size; } EXPORT_SYMBOL(cpm_muram_free); /** /* * cpm_muram_alloc_fixed - reserve a specific region of multi-user ram * @offset: the offset into the muram area to reserve * @size: the number of bytes to reserve * * This function returns "start" on success, -ENOMEM on failure. * @offset: offset of allocation start address * @size: number of bytes to allocate * This function returns an offset into the muram area * Use cpm_dpram_addr() to get the virtual address of the area. * Use cpm_muram_free() to free the allocation. */ Loading @@ -185,16 +201,50 @@ unsigned long cpm_muram_alloc_fixed(unsigned long offset, unsigned long size) { unsigned long start; unsigned long flags; struct genpool_data_fixed muram_pool_data_fixed; spin_lock_irqsave(&cpm_muram_lock, flags); cpm_muram_info.alignment = 1; start = rh_alloc_fixed(&cpm_muram_info, offset, size, "commproc"); muram_pool_data_fixed.offset = offset + GENPOOL_OFFSET; start = cpm_muram_alloc_common(size, gen_pool_fixed_alloc, &muram_pool_data_fixed); spin_unlock_irqrestore(&cpm_muram_lock, flags); return start; } EXPORT_SYMBOL(cpm_muram_alloc_fixed); /* * cpm_muram_alloc_common - cpm_muram_alloc common code * @size: number of bytes to allocate * @algo: algorithm for alloc. * @data: data for genalloc's algorithm. * * This function returns an offset into the muram area. */ unsigned long cpm_muram_alloc_common(unsigned long size, genpool_algo_t algo, void *data) { struct muram_block *entry; unsigned long start; start = gen_pool_alloc_algo(muram_pool, size, algo, data); if (!start) goto out2; start = start - GENPOOL_OFFSET; memset_io(cpm_muram_addr(start), 0, size); entry = kmalloc(sizeof(*entry), GFP_KERNEL); if (!entry) goto out1; entry->start = start; entry->size = size; list_add(&entry->head, &muram_block_list); return start; out1: gen_pool_free(muram_pool, start, size); out2: return (unsigned long)-ENOMEM; } /** * cpm_muram_addr - turn a muram offset into a virtual address * @offset: muram offset to convert Loading lib/genalloc.c +1 −1 Original line number Diff line number Diff line Loading @@ -576,7 +576,7 @@ unsigned long gen_pool_fixed_alloc(unsigned long *map, unsigned long size, fixed_data = data; order = pool->min_alloc_order; offset_bit = fixed_data->offset >> order; if (WARN_ON(fixed_data->offset & (1UL << order - 1))) if (WARN_ON(fixed_data->offset & ((1UL << order) - 1))) return size; start_bit = bitmap_find_next_zero_area(map, size, Loading Loading
arch/powerpc/include/asm/cpm.h +3 −0 Original line number Diff line number Diff line Loading @@ -2,6 +2,7 @@ #define __CPM_H #include <linux/compiler.h> #include <linux/genalloc.h> #include <linux/types.h> #include <linux/errno.h> #include <linux/of.h> Loading Loading @@ -161,6 +162,8 @@ int cpm_muram_init(void); unsigned long cpm_muram_alloc(unsigned long size, unsigned long align); int cpm_muram_free(unsigned long offset); unsigned long cpm_muram_alloc_fixed(unsigned long offset, unsigned long size); unsigned long cpm_muram_alloc_common(unsigned long size, genpool_algo_t algo, void *data); void __iomem *cpm_muram_addr(unsigned long offset); unsigned long cpm_muram_offset(void __iomem *addr); dma_addr_t cpm_muram_dma(void __iomem *addr); Loading
arch/powerpc/platforms/Kconfig +2 −2 Original line number Diff line number Diff line Loading @@ -275,7 +275,7 @@ config TAU_AVERAGE config QUICC_ENGINE bool "Freescale QUICC Engine (QE) Support" depends on FSL_SOC && PPC32 select PPC_LIB_RHEAP select GENERIC_ALLOCATOR select CRC32 help The QUICC Engine (QE) is a new generation of communications Loading @@ -295,7 +295,6 @@ config CPM2 bool "Enable support for the CPM2 (Communications Processor Module)" depends on (FSL_SOC_BOOKE && PPC32) || 8260 select CPM select PPC_LIB_RHEAP select PPC_PCI_CHOICE select ARCH_REQUIRE_GPIOLIB help Loading Loading @@ -325,6 +324,7 @@ config FSL_ULI1575 config CPM bool select GENERIC_ALLOCATOR config OF_RTC bool Loading
arch/powerpc/sysdev/cpm_common.c +88 −38 Original line number Diff line number Diff line Loading @@ -17,6 +17,7 @@ * published by the Free Software Foundation. */ #include <linux/genalloc.h> #include <linux/init.h> #include <linux/of_device.h> #include <linux/spinlock.h> Loading @@ -27,7 +28,6 @@ #include <asm/udbg.h> #include <asm/io.h> #include <asm/rheap.h> #include <asm/cpm.h> #include <mm/mmu_decl.h> Loading Loading @@ -65,14 +65,22 @@ void __init udbg_init_cpm(void) } #endif static struct gen_pool *muram_pool; static spinlock_t cpm_muram_lock; static rh_block_t cpm_boot_muram_rh_block[16]; static rh_info_t cpm_muram_info; static u8 __iomem *muram_vbase; static phys_addr_t muram_pbase; /* Max address size we deal with */ struct muram_block { struct list_head head; unsigned long start; int size; }; static LIST_HEAD(muram_block_list); /* max address size we deal with */ #define OF_MAX_ADDR_CELLS 4 #define GENPOOL_OFFSET (4096 * 8) int cpm_muram_init(void) { Loading @@ -87,50 +95,51 @@ int cpm_muram_init(void) return 0; spin_lock_init(&cpm_muram_lock); /* initialize the info header */ rh_init(&cpm_muram_info, 1, sizeof(cpm_boot_muram_rh_block) / sizeof(cpm_boot_muram_rh_block[0]), cpm_boot_muram_rh_block); np = of_find_compatible_node(NULL, NULL, "fsl,cpm-muram-data"); if (!np) { /* try legacy bindings */ np = of_find_node_by_name(NULL, "data-only"); if (!np) { printk(KERN_ERR "Cannot find CPM muram data node"); pr_err("Cannot find CPM muram data node"); ret = -ENODEV; goto out; goto out_muram; } } muram_pool = gen_pool_create(0, -1); muram_pbase = of_translate_address(np, zero); if (muram_pbase == (phys_addr_t)OF_BAD_ADDR) { printk(KERN_ERR "Cannot translate zero through CPM muram node"); pr_err("Cannot translate zero through CPM muram node"); ret = -ENODEV; goto out; goto out_pool; } while (of_address_to_resource(np, i++, &r) == 0) { if (r.end > max) max = r.end; rh_attach_region(&cpm_muram_info, r.start - muram_pbase, resource_size(&r)); ret = gen_pool_add(muram_pool, r.start - muram_pbase + GENPOOL_OFFSET, resource_size(&r), -1); if (ret) { pr_err("QE: couldn't add muram to pool!\n"); goto out_pool; } } muram_vbase = ioremap(muram_pbase, max - muram_pbase + 1); if (!muram_vbase) { printk(KERN_ERR "Cannot map CPM muram"); pr_err("Cannot map QE muram"); ret = -ENOMEM; goto out_pool; } out: goto out_muram; out_pool: gen_pool_destroy(muram_pool); out_muram: of_node_put(np); return ret; } /** /* * cpm_muram_alloc - allocate the requested size worth of multi-user ram * @size: number of bytes to allocate * @align: requested alignment, in bytes Loading @@ -143,14 +152,13 @@ unsigned long cpm_muram_alloc(unsigned long size, unsigned long align) { unsigned long start; unsigned long flags; struct genpool_data_align muram_pool_data; spin_lock_irqsave(&cpm_muram_lock, flags); cpm_muram_info.alignment = align; start = rh_alloc(&cpm_muram_info, size, "commproc"); if (!IS_ERR_VALUE(start)) memset_io(cpm_muram_addr(start), 0, size); muram_pool_data.align = align; start = cpm_muram_alloc_common(size, gen_pool_first_fit_align, &muram_pool_data); spin_unlock_irqrestore(&cpm_muram_lock, flags); return start; } EXPORT_SYMBOL(cpm_muram_alloc); Loading @@ -161,23 +169,31 @@ EXPORT_SYMBOL(cpm_muram_alloc); */ int cpm_muram_free(unsigned long offset) { int ret; unsigned long flags; int size; struct muram_block *tmp; size = 0; spin_lock_irqsave(&cpm_muram_lock, flags); ret = rh_free(&cpm_muram_info, offset); list_for_each_entry(tmp, &muram_block_list, head) { if (tmp->start == offset) { size = tmp->size; list_del(&tmp->head); kfree(tmp); break; } } gen_pool_free(muram_pool, offset + GENPOOL_OFFSET, size); spin_unlock_irqrestore(&cpm_muram_lock, flags); return ret; return size; } EXPORT_SYMBOL(cpm_muram_free); /** /* * cpm_muram_alloc_fixed - reserve a specific region of multi-user ram * @offset: the offset into the muram area to reserve * @size: the number of bytes to reserve * * This function returns "start" on success, -ENOMEM on failure. * @offset: offset of allocation start address * @size: number of bytes to allocate * This function returns an offset into the muram area * Use cpm_dpram_addr() to get the virtual address of the area. * Use cpm_muram_free() to free the allocation. */ Loading @@ -185,16 +201,50 @@ unsigned long cpm_muram_alloc_fixed(unsigned long offset, unsigned long size) { unsigned long start; unsigned long flags; struct genpool_data_fixed muram_pool_data_fixed; spin_lock_irqsave(&cpm_muram_lock, flags); cpm_muram_info.alignment = 1; start = rh_alloc_fixed(&cpm_muram_info, offset, size, "commproc"); muram_pool_data_fixed.offset = offset + GENPOOL_OFFSET; start = cpm_muram_alloc_common(size, gen_pool_fixed_alloc, &muram_pool_data_fixed); spin_unlock_irqrestore(&cpm_muram_lock, flags); return start; } EXPORT_SYMBOL(cpm_muram_alloc_fixed); /* * cpm_muram_alloc_common - cpm_muram_alloc common code * @size: number of bytes to allocate * @algo: algorithm for alloc. * @data: data for genalloc's algorithm. * * This function returns an offset into the muram area. */ unsigned long cpm_muram_alloc_common(unsigned long size, genpool_algo_t algo, void *data) { struct muram_block *entry; unsigned long start; start = gen_pool_alloc_algo(muram_pool, size, algo, data); if (!start) goto out2; start = start - GENPOOL_OFFSET; memset_io(cpm_muram_addr(start), 0, size); entry = kmalloc(sizeof(*entry), GFP_KERNEL); if (!entry) goto out1; entry->start = start; entry->size = size; list_add(&entry->head, &muram_block_list); return start; out1: gen_pool_free(muram_pool, start, size); out2: return (unsigned long)-ENOMEM; } /** * cpm_muram_addr - turn a muram offset into a virtual address * @offset: muram offset to convert Loading
lib/genalloc.c +1 −1 Original line number Diff line number Diff line Loading @@ -576,7 +576,7 @@ unsigned long gen_pool_fixed_alloc(unsigned long *map, unsigned long size, fixed_data = data; order = pool->min_alloc_order; offset_bit = fixed_data->offset >> order; if (WARN_ON(fixed_data->offset & (1UL << order - 1))) if (WARN_ON(fixed_data->offset & ((1UL << order) - 1))) return size; start_bit = bitmap_find_next_zero_area(map, size, Loading