mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

Several functions refer to the unfortunately named 'vm_flags' field when referencing vmalloc flags, which happens to be the precise same name used for VMA flags. As a result these were erroneously changed to use the vm_flags_t type (which currently is a typedef equivalent to unsigned long). Currently this has no impact, but in future when vm_flags_t changes this will result in issues, so change the type to unsigned long to account for this. [lorenzo.stoakes@oracle.com: fixup very disguised vmalloc flags parameter] Link: https://lkml.kernel.org/r/e74dd8de-7e60-47ab-8a45-2c851f3c5d26@lucifer.local Link: https://lkml.kernel.org/r/20250729114906.55347-1-lorenzo.stoakes@oracle.com Signed-off-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com> Reported-by: Harry Yoo <harry.yoo@oracle.com> Closes: https://lore.kernel.org/all/aIgSpAnU8EaIcqd9@hyeyoo/ Reviewed-by: Pedro Falcato <pfalcato@suse.de> Acked-by: David Hildenbrand <david@redhat.com> Reviewed-by: Harry Yoo <harry.yoo@oracle.com> Acked-by: Vlastimil Babka <vbabka@suse.cz> Cc: Jann Horn <jannh@google.com> Cc: Liam Howlett <liam.howlett@oracle.com> Cc: Michal Hocko <mhocko@suse.com> Cc: Mike Rapoport <rppt@kernel.org> Cc: Suren Baghdasaryan <surenb@google.com> Cc: "Uladzislau Rezki (Sony)" <urezki@gmail.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
596 lines
14 KiB
C
596 lines
14 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) 2002 Richard Henderson
|
|
* Copyright (C) 2001 Rusty Russell, 2002, 2010 Rusty Russell IBM.
|
|
* Copyright (C) 2023 Luis Chamberlain <mcgrof@kernel.org>
|
|
* Copyright (C) 2024 Mike Rapoport IBM.
|
|
*/
|
|
|
|
#define pr_fmt(fmt) "execmem: " fmt
|
|
|
|
#include <linux/mm.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/execmem.h>
|
|
#include <linux/maple_tree.h>
|
|
#include <linux/set_memory.h>
|
|
#include <linux/moduleloader.h>
|
|
#include <linux/text-patching.h>
|
|
|
|
#include <asm/tlbflush.h>
|
|
|
|
#include "internal.h"
|
|
|
|
static struct execmem_info *execmem_info __ro_after_init;
|
|
static struct execmem_info default_execmem_info __ro_after_init;
|
|
|
|
#ifdef CONFIG_MMU
|
|
static void *execmem_vmalloc(struct execmem_range *range, size_t size,
|
|
pgprot_t pgprot, unsigned long vm_flags)
|
|
{
|
|
bool kasan = range->flags & EXECMEM_KASAN_SHADOW;
|
|
gfp_t gfp_flags = GFP_KERNEL | __GFP_NOWARN;
|
|
unsigned int align = range->alignment;
|
|
unsigned long start = range->start;
|
|
unsigned long end = range->end;
|
|
void *p;
|
|
|
|
if (kasan)
|
|
vm_flags |= VM_DEFER_KMEMLEAK;
|
|
|
|
if (vm_flags & VM_ALLOW_HUGE_VMAP)
|
|
align = PMD_SIZE;
|
|
|
|
p = __vmalloc_node_range(size, align, start, end, gfp_flags,
|
|
pgprot, vm_flags, NUMA_NO_NODE,
|
|
__builtin_return_address(0));
|
|
if (!p && range->fallback_start) {
|
|
start = range->fallback_start;
|
|
end = range->fallback_end;
|
|
p = __vmalloc_node_range(size, align, start, end, gfp_flags,
|
|
pgprot, vm_flags, NUMA_NO_NODE,
|
|
__builtin_return_address(0));
|
|
}
|
|
|
|
if (!p) {
|
|
pr_warn_ratelimited("unable to allocate memory\n");
|
|
return NULL;
|
|
}
|
|
|
|
if (kasan && (kasan_alloc_module_shadow(p, size, GFP_KERNEL) < 0)) {
|
|
vfree(p);
|
|
return NULL;
|
|
}
|
|
|
|
return p;
|
|
}
|
|
|
|
struct vm_struct *execmem_vmap(size_t size)
|
|
{
|
|
struct execmem_range *range = &execmem_info->ranges[EXECMEM_MODULE_DATA];
|
|
struct vm_struct *area;
|
|
|
|
area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC,
|
|
range->start, range->end, NUMA_NO_NODE,
|
|
GFP_KERNEL, __builtin_return_address(0));
|
|
if (!area && range->fallback_start)
|
|
area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC,
|
|
range->fallback_start, range->fallback_end,
|
|
NUMA_NO_NODE, GFP_KERNEL, __builtin_return_address(0));
|
|
|
|
return area;
|
|
}
|
|
#else
|
|
static void *execmem_vmalloc(struct execmem_range *range, size_t size,
|
|
pgprot_t pgprot, unsigned long vm_flags)
|
|
{
|
|
return vmalloc(size);
|
|
}
|
|
#endif /* CONFIG_MMU */
|
|
|
|
#ifdef CONFIG_ARCH_HAS_EXECMEM_ROX
|
|
struct execmem_cache {
|
|
struct mutex mutex;
|
|
struct maple_tree busy_areas;
|
|
struct maple_tree free_areas;
|
|
unsigned int pending_free_cnt; /* protected by mutex */
|
|
};
|
|
|
|
/* delay to schedule asynchronous free if fast path free fails */
|
|
#define FREE_DELAY (msecs_to_jiffies(10))
|
|
|
|
/* mark entries in busy_areas that should be freed asynchronously */
|
|
#define PENDING_FREE_MASK (1 << (PAGE_SHIFT - 1))
|
|
|
|
static struct execmem_cache execmem_cache = {
|
|
.mutex = __MUTEX_INITIALIZER(execmem_cache.mutex),
|
|
.busy_areas = MTREE_INIT_EXT(busy_areas, MT_FLAGS_LOCK_EXTERN,
|
|
execmem_cache.mutex),
|
|
.free_areas = MTREE_INIT_EXT(free_areas, MT_FLAGS_LOCK_EXTERN,
|
|
execmem_cache.mutex),
|
|
};
|
|
|
|
static inline unsigned long mas_range_len(struct ma_state *mas)
|
|
{
|
|
return mas->last - mas->index + 1;
|
|
}
|
|
|
|
static int execmem_set_direct_map_valid(struct vm_struct *vm, bool valid)
|
|
{
|
|
unsigned int nr = (1 << get_vm_area_page_order(vm));
|
|
unsigned int updated = 0;
|
|
int err = 0;
|
|
|
|
for (int i = 0; i < vm->nr_pages; i += nr) {
|
|
err = set_direct_map_valid_noflush(vm->pages[i], nr, valid);
|
|
if (err)
|
|
goto err_restore;
|
|
updated += nr;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_restore:
|
|
for (int i = 0; i < updated; i += nr)
|
|
set_direct_map_valid_noflush(vm->pages[i], nr, !valid);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int execmem_force_rw(void *ptr, size_t size)
|
|
{
|
|
unsigned int nr = PAGE_ALIGN(size) >> PAGE_SHIFT;
|
|
unsigned long addr = (unsigned long)ptr;
|
|
int ret;
|
|
|
|
ret = set_memory_nx(addr, nr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return set_memory_rw(addr, nr);
|
|
}
|
|
|
|
int execmem_restore_rox(void *ptr, size_t size)
|
|
{
|
|
unsigned int nr = PAGE_ALIGN(size) >> PAGE_SHIFT;
|
|
unsigned long addr = (unsigned long)ptr;
|
|
|
|
return set_memory_rox(addr, nr);
|
|
}
|
|
|
|
static void execmem_cache_clean(struct work_struct *work)
|
|
{
|
|
struct maple_tree *free_areas = &execmem_cache.free_areas;
|
|
struct mutex *mutex = &execmem_cache.mutex;
|
|
MA_STATE(mas, free_areas, 0, ULONG_MAX);
|
|
void *area;
|
|
|
|
mutex_lock(mutex);
|
|
mas_for_each(&mas, area, ULONG_MAX) {
|
|
size_t size = mas_range_len(&mas);
|
|
|
|
if (IS_ALIGNED(size, PMD_SIZE) &&
|
|
IS_ALIGNED(mas.index, PMD_SIZE)) {
|
|
struct vm_struct *vm = find_vm_area(area);
|
|
|
|
execmem_set_direct_map_valid(vm, true);
|
|
mas_store_gfp(&mas, NULL, GFP_KERNEL);
|
|
vfree(area);
|
|
}
|
|
}
|
|
mutex_unlock(mutex);
|
|
}
|
|
|
|
static DECLARE_WORK(execmem_cache_clean_work, execmem_cache_clean);
|
|
|
|
static int execmem_cache_add_locked(void *ptr, size_t size, gfp_t gfp_mask)
|
|
{
|
|
struct maple_tree *free_areas = &execmem_cache.free_areas;
|
|
unsigned long addr = (unsigned long)ptr;
|
|
MA_STATE(mas, free_areas, addr - 1, addr + 1);
|
|
unsigned long lower, upper;
|
|
void *area = NULL;
|
|
|
|
lower = addr;
|
|
upper = addr + size - 1;
|
|
|
|
area = mas_walk(&mas);
|
|
if (area && mas.last == addr - 1)
|
|
lower = mas.index;
|
|
|
|
area = mas_next(&mas, ULONG_MAX);
|
|
if (area && mas.index == addr + size)
|
|
upper = mas.last;
|
|
|
|
mas_set_range(&mas, lower, upper);
|
|
return mas_store_gfp(&mas, (void *)lower, gfp_mask);
|
|
}
|
|
|
|
static int execmem_cache_add(void *ptr, size_t size, gfp_t gfp_mask)
|
|
{
|
|
guard(mutex)(&execmem_cache.mutex);
|
|
|
|
return execmem_cache_add_locked(ptr, size, gfp_mask);
|
|
}
|
|
|
|
static bool within_range(struct execmem_range *range, struct ma_state *mas,
|
|
size_t size)
|
|
{
|
|
unsigned long addr = mas->index;
|
|
|
|
if (addr >= range->start && addr + size < range->end)
|
|
return true;
|
|
|
|
if (range->fallback_start &&
|
|
addr >= range->fallback_start && addr + size < range->fallback_end)
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
static void *__execmem_cache_alloc(struct execmem_range *range, size_t size)
|
|
{
|
|
struct maple_tree *free_areas = &execmem_cache.free_areas;
|
|
struct maple_tree *busy_areas = &execmem_cache.busy_areas;
|
|
MA_STATE(mas_free, free_areas, 0, ULONG_MAX);
|
|
MA_STATE(mas_busy, busy_areas, 0, ULONG_MAX);
|
|
struct mutex *mutex = &execmem_cache.mutex;
|
|
unsigned long addr, last, area_size = 0;
|
|
void *area, *ptr = NULL;
|
|
int err;
|
|
|
|
mutex_lock(mutex);
|
|
mas_for_each(&mas_free, area, ULONG_MAX) {
|
|
area_size = mas_range_len(&mas_free);
|
|
|
|
if (area_size >= size && within_range(range, &mas_free, size))
|
|
break;
|
|
}
|
|
|
|
if (area_size < size)
|
|
goto out_unlock;
|
|
|
|
addr = mas_free.index;
|
|
last = mas_free.last;
|
|
|
|
/* insert allocated size to busy_areas at range [addr, addr + size) */
|
|
mas_set_range(&mas_busy, addr, addr + size - 1);
|
|
err = mas_store_gfp(&mas_busy, (void *)addr, GFP_KERNEL);
|
|
if (err)
|
|
goto out_unlock;
|
|
|
|
mas_store_gfp(&mas_free, NULL, GFP_KERNEL);
|
|
if (area_size > size) {
|
|
void *ptr = (void *)(addr + size);
|
|
|
|
/*
|
|
* re-insert remaining free size to free_areas at range
|
|
* [addr + size, last]
|
|
*/
|
|
mas_set_range(&mas_free, addr + size, last);
|
|
err = mas_store_gfp(&mas_free, ptr, GFP_KERNEL);
|
|
if (err) {
|
|
mas_store_gfp(&mas_busy, NULL, GFP_KERNEL);
|
|
goto out_unlock;
|
|
}
|
|
}
|
|
ptr = (void *)addr;
|
|
|
|
out_unlock:
|
|
mutex_unlock(mutex);
|
|
return ptr;
|
|
}
|
|
|
|
static int execmem_cache_populate(struct execmem_range *range, size_t size)
|
|
{
|
|
unsigned long vm_flags = VM_ALLOW_HUGE_VMAP;
|
|
struct vm_struct *vm;
|
|
size_t alloc_size;
|
|
int err = -ENOMEM;
|
|
void *p;
|
|
|
|
alloc_size = round_up(size, PMD_SIZE);
|
|
p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags);
|
|
if (!p) {
|
|
alloc_size = size;
|
|
p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags);
|
|
}
|
|
|
|
if (!p)
|
|
return err;
|
|
|
|
vm = find_vm_area(p);
|
|
if (!vm)
|
|
goto err_free_mem;
|
|
|
|
/* fill memory with instructions that will trap */
|
|
execmem_fill_trapping_insns(p, alloc_size);
|
|
|
|
err = set_memory_rox((unsigned long)p, vm->nr_pages);
|
|
if (err)
|
|
goto err_free_mem;
|
|
|
|
err = execmem_cache_add(p, alloc_size, GFP_KERNEL);
|
|
if (err)
|
|
goto err_reset_direct_map;
|
|
|
|
return 0;
|
|
|
|
err_reset_direct_map:
|
|
execmem_set_direct_map_valid(vm, true);
|
|
err_free_mem:
|
|
vfree(p);
|
|
return err;
|
|
}
|
|
|
|
static void *execmem_cache_alloc(struct execmem_range *range, size_t size)
|
|
{
|
|
void *p;
|
|
int err;
|
|
|
|
p = __execmem_cache_alloc(range, size);
|
|
if (p)
|
|
return p;
|
|
|
|
err = execmem_cache_populate(range, size);
|
|
if (err)
|
|
return NULL;
|
|
|
|
return __execmem_cache_alloc(range, size);
|
|
}
|
|
|
|
static inline bool is_pending_free(void *ptr)
|
|
{
|
|
return ((unsigned long)ptr & PENDING_FREE_MASK);
|
|
}
|
|
|
|
static inline void *pending_free_set(void *ptr)
|
|
{
|
|
return (void *)((unsigned long)ptr | PENDING_FREE_MASK);
|
|
}
|
|
|
|
static inline void *pending_free_clear(void *ptr)
|
|
{
|
|
return (void *)((unsigned long)ptr & ~PENDING_FREE_MASK);
|
|
}
|
|
|
|
static int __execmem_cache_free(struct ma_state *mas, void *ptr, gfp_t gfp_mask)
|
|
{
|
|
size_t size = mas_range_len(mas);
|
|
int err;
|
|
|
|
err = execmem_force_rw(ptr, size);
|
|
if (err)
|
|
return err;
|
|
|
|
execmem_fill_trapping_insns(ptr, size);
|
|
execmem_restore_rox(ptr, size);
|
|
|
|
err = execmem_cache_add_locked(ptr, size, gfp_mask);
|
|
if (err)
|
|
return err;
|
|
|
|
mas_store_gfp(mas, NULL, gfp_mask);
|
|
return 0;
|
|
}
|
|
|
|
static void execmem_cache_free_slow(struct work_struct *work);
|
|
static DECLARE_DELAYED_WORK(execmem_cache_free_work, execmem_cache_free_slow);
|
|
|
|
static void execmem_cache_free_slow(struct work_struct *work)
|
|
{
|
|
struct maple_tree *busy_areas = &execmem_cache.busy_areas;
|
|
MA_STATE(mas, busy_areas, 0, ULONG_MAX);
|
|
void *area;
|
|
|
|
guard(mutex)(&execmem_cache.mutex);
|
|
|
|
if (!execmem_cache.pending_free_cnt)
|
|
return;
|
|
|
|
mas_for_each(&mas, area, ULONG_MAX) {
|
|
if (!is_pending_free(area))
|
|
continue;
|
|
|
|
area = pending_free_clear(area);
|
|
if (__execmem_cache_free(&mas, area, GFP_KERNEL))
|
|
continue;
|
|
|
|
execmem_cache.pending_free_cnt--;
|
|
}
|
|
|
|
if (execmem_cache.pending_free_cnt)
|
|
schedule_delayed_work(&execmem_cache_free_work, FREE_DELAY);
|
|
else
|
|
schedule_work(&execmem_cache_clean_work);
|
|
}
|
|
|
|
static bool execmem_cache_free(void *ptr)
|
|
{
|
|
struct maple_tree *busy_areas = &execmem_cache.busy_areas;
|
|
unsigned long addr = (unsigned long)ptr;
|
|
MA_STATE(mas, busy_areas, addr, addr);
|
|
void *area;
|
|
int err;
|
|
|
|
guard(mutex)(&execmem_cache.mutex);
|
|
|
|
area = mas_walk(&mas);
|
|
if (!area)
|
|
return false;
|
|
|
|
err = __execmem_cache_free(&mas, area, GFP_KERNEL | __GFP_NORETRY);
|
|
if (err) {
|
|
/*
|
|
* mas points to exact slot we've got the area from, nothing
|
|
* else can modify the tree because of the mutex, so there
|
|
* won't be any allocations in mas_store_gfp() and it will just
|
|
* change the pointer.
|
|
*/
|
|
area = pending_free_set(area);
|
|
mas_store_gfp(&mas, area, GFP_KERNEL);
|
|
execmem_cache.pending_free_cnt++;
|
|
schedule_delayed_work(&execmem_cache_free_work, FREE_DELAY);
|
|
return true;
|
|
}
|
|
|
|
schedule_work(&execmem_cache_clean_work);
|
|
|
|
return true;
|
|
}
|
|
|
|
#else /* CONFIG_ARCH_HAS_EXECMEM_ROX */
|
|
/*
|
|
* when ROX cache is not used the permissions defined by architectures for
|
|
* execmem ranges that are updated before use (e.g. EXECMEM_MODULE_TEXT) must
|
|
* be writable anyway
|
|
*/
|
|
static inline int execmem_force_rw(void *ptr, size_t size)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static void *execmem_cache_alloc(struct execmem_range *range, size_t size)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
static bool execmem_cache_free(void *ptr)
|
|
{
|
|
return false;
|
|
}
|
|
#endif /* CONFIG_ARCH_HAS_EXECMEM_ROX */
|
|
|
|
void *execmem_alloc(enum execmem_type type, size_t size)
|
|
{
|
|
struct execmem_range *range = &execmem_info->ranges[type];
|
|
bool use_cache = range->flags & EXECMEM_ROX_CACHE;
|
|
unsigned long vm_flags = VM_FLUSH_RESET_PERMS;
|
|
pgprot_t pgprot = range->pgprot;
|
|
void *p = NULL;
|
|
|
|
size = PAGE_ALIGN(size);
|
|
|
|
if (use_cache)
|
|
p = execmem_cache_alloc(range, size);
|
|
else
|
|
p = execmem_vmalloc(range, size, pgprot, vm_flags);
|
|
|
|
return kasan_reset_tag(p);
|
|
}
|
|
|
|
void *execmem_alloc_rw(enum execmem_type type, size_t size)
|
|
{
|
|
void *p __free(execmem) = execmem_alloc(type, size);
|
|
int err;
|
|
|
|
if (!p)
|
|
return NULL;
|
|
|
|
err = execmem_force_rw(p, size);
|
|
if (err)
|
|
return NULL;
|
|
|
|
return no_free_ptr(p);
|
|
}
|
|
|
|
void execmem_free(void *ptr)
|
|
{
|
|
/*
|
|
* This memory may be RO, and freeing RO memory in an interrupt is not
|
|
* supported by vmalloc.
|
|
*/
|
|
WARN_ON(in_interrupt());
|
|
|
|
if (!execmem_cache_free(ptr))
|
|
vfree(ptr);
|
|
}
|
|
|
|
bool execmem_is_rox(enum execmem_type type)
|
|
{
|
|
return !!(execmem_info->ranges[type].flags & EXECMEM_ROX_CACHE);
|
|
}
|
|
|
|
static bool execmem_validate(struct execmem_info *info)
|
|
{
|
|
struct execmem_range *r = &info->ranges[EXECMEM_DEFAULT];
|
|
|
|
if (!r->alignment || !r->start || !r->end || !pgprot_val(r->pgprot)) {
|
|
pr_crit("Invalid parameters for execmem allocator, module loading will fail");
|
|
return false;
|
|
}
|
|
|
|
if (!IS_ENABLED(CONFIG_ARCH_HAS_EXECMEM_ROX)) {
|
|
for (int i = EXECMEM_DEFAULT; i < EXECMEM_TYPE_MAX; i++) {
|
|
r = &info->ranges[i];
|
|
|
|
if (r->flags & EXECMEM_ROX_CACHE) {
|
|
pr_warn_once("ROX cache is not supported\n");
|
|
r->flags &= ~EXECMEM_ROX_CACHE;
|
|
}
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static void execmem_init_missing(struct execmem_info *info)
|
|
{
|
|
struct execmem_range *default_range = &info->ranges[EXECMEM_DEFAULT];
|
|
|
|
for (int i = EXECMEM_DEFAULT + 1; i < EXECMEM_TYPE_MAX; i++) {
|
|
struct execmem_range *r = &info->ranges[i];
|
|
|
|
if (!r->start) {
|
|
if (i == EXECMEM_MODULE_DATA)
|
|
r->pgprot = PAGE_KERNEL;
|
|
else
|
|
r->pgprot = default_range->pgprot;
|
|
r->alignment = default_range->alignment;
|
|
r->start = default_range->start;
|
|
r->end = default_range->end;
|
|
r->flags = default_range->flags;
|
|
r->fallback_start = default_range->fallback_start;
|
|
r->fallback_end = default_range->fallback_end;
|
|
}
|
|
}
|
|
}
|
|
|
|
struct execmem_info * __weak execmem_arch_setup(void)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
static void __init __execmem_init(void)
|
|
{
|
|
struct execmem_info *info = execmem_arch_setup();
|
|
|
|
if (!info) {
|
|
info = execmem_info = &default_execmem_info;
|
|
info->ranges[EXECMEM_DEFAULT].start = VMALLOC_START;
|
|
info->ranges[EXECMEM_DEFAULT].end = VMALLOC_END;
|
|
info->ranges[EXECMEM_DEFAULT].pgprot = PAGE_KERNEL_EXEC;
|
|
info->ranges[EXECMEM_DEFAULT].alignment = 1;
|
|
}
|
|
|
|
if (!execmem_validate(info))
|
|
return;
|
|
|
|
execmem_init_missing(info);
|
|
|
|
execmem_info = info;
|
|
}
|
|
|
|
#ifdef CONFIG_ARCH_WANTS_EXECMEM_LATE
|
|
static int __init execmem_late_init(void)
|
|
{
|
|
__execmem_init();
|
|
return 0;
|
|
}
|
|
core_initcall(execmem_late_init);
|
|
#else
|
|
void __init execmem_init(void)
|
|
{
|
|
__execmem_init();
|
|
}
|
|
#endif
|