mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

Since early_pfn and bitmap are never used at the same time, they can be defined as a union to reduce the size of the data structure. This change can save 8 * u64 entries per CMA. Link: https://lkml.kernel.org/r/20250509083528.1360952-1-hezhongkun.hzk@bytedance.com Signed-off-by: Zhongkun He <hezhongkun.hzk@bytedance.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
95 lines
2.6 KiB
C
95 lines
2.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __MM_CMA_H__
|
|
#define __MM_CMA_H__
|
|
|
|
#include <linux/debugfs.h>
|
|
#include <linux/kobject.h>
|
|
|
|
struct cma_kobject {
|
|
struct kobject kobj;
|
|
struct cma *cma;
|
|
};
|
|
|
|
/*
|
|
* Multi-range support. This can be useful if the size of the allocation
|
|
* is not expected to be larger than the alignment (like with hugetlb_cma),
|
|
* and the total amount of memory requested, while smaller than the total
|
|
* amount of memory available, is large enough that it doesn't fit in a
|
|
* single physical memory range because of memory holes.
|
|
*
|
|
* Fields:
|
|
* @base_pfn: physical address of range
|
|
* @early_pfn: first PFN not reserved through cma_reserve_early
|
|
* @count: size of range
|
|
* @bitmap: bitmap of allocated (1 << order_per_bit)-sized chunks.
|
|
*/
|
|
struct cma_memrange {
|
|
unsigned long base_pfn;
|
|
unsigned long count;
|
|
union {
|
|
unsigned long early_pfn;
|
|
unsigned long *bitmap;
|
|
};
|
|
#ifdef CONFIG_CMA_DEBUGFS
|
|
struct debugfs_u32_array dfs_bitmap;
|
|
#endif
|
|
};
|
|
#define CMA_MAX_RANGES 8
|
|
|
|
struct cma {
|
|
unsigned long count;
|
|
unsigned long available_count;
|
|
unsigned int order_per_bit; /* Order of pages represented by one bit */
|
|
spinlock_t lock;
|
|
struct mutex alloc_mutex;
|
|
#ifdef CONFIG_CMA_DEBUGFS
|
|
struct hlist_head mem_head;
|
|
spinlock_t mem_head_lock;
|
|
#endif
|
|
char name[CMA_MAX_NAME];
|
|
int nranges;
|
|
struct cma_memrange ranges[CMA_MAX_RANGES];
|
|
#ifdef CONFIG_CMA_SYSFS
|
|
/* the number of CMA page successful allocations */
|
|
atomic64_t nr_pages_succeeded;
|
|
/* the number of CMA page allocation failures */
|
|
atomic64_t nr_pages_failed;
|
|
/* the number of CMA page released */
|
|
atomic64_t nr_pages_released;
|
|
/* kobject requires dynamic object */
|
|
struct cma_kobject *cma_kobj;
|
|
#endif
|
|
unsigned long flags;
|
|
/* NUMA node (NUMA_NO_NODE if unspecified) */
|
|
int nid;
|
|
};
|
|
|
|
enum cma_flags {
|
|
CMA_RESERVE_PAGES_ON_ERROR,
|
|
CMA_ZONES_VALID,
|
|
CMA_ZONES_INVALID,
|
|
CMA_ACTIVATED,
|
|
};
|
|
|
|
extern struct cma cma_areas[MAX_CMA_AREAS];
|
|
extern unsigned int cma_area_count;
|
|
|
|
static inline unsigned long cma_bitmap_maxno(struct cma *cma,
|
|
struct cma_memrange *cmr)
|
|
{
|
|
return cmr->count >> cma->order_per_bit;
|
|
}
|
|
|
|
#ifdef CONFIG_CMA_SYSFS
|
|
void cma_sysfs_account_success_pages(struct cma *cma, unsigned long nr_pages);
|
|
void cma_sysfs_account_fail_pages(struct cma *cma, unsigned long nr_pages);
|
|
void cma_sysfs_account_release_pages(struct cma *cma, unsigned long nr_pages);
|
|
#else
|
|
static inline void cma_sysfs_account_success_pages(struct cma *cma,
|
|
unsigned long nr_pages) {};
|
|
static inline void cma_sysfs_account_fail_pages(struct cma *cma,
|
|
unsigned long nr_pages) {};
|
|
static inline void cma_sysfs_account_release_pages(struct cma *cma,
|
|
unsigned long nr_pages) {};
|
|
#endif
|
|
#endif
|