mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

Add selftests for the bpf_cpumask_populate helper that sets a bpf_cpumask to a bit pattern provided by a BPF program. Signed-off-by: Emil Tsalapatis (Meta) <emil@etsalapatis.com> Acked-by: Hou Tao <houtao1@huawei.com> Link: https://lore.kernel.org/r/20250309230427.26603-3-emil@etsalapatis.com Signed-off-by: Alexei Starovoitov <ast@kernel.org>
262 lines
5.7 KiB
C
262 lines
5.7 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Copyright (c) 2023 Meta Platforms, Inc. and affiliates. */
|
|
|
|
#include <vmlinux.h>
|
|
#include <bpf/bpf_tracing.h>
|
|
#include <bpf/bpf_helpers.h>
|
|
#include "bpf_misc.h"
|
|
|
|
#include "cpumask_common.h"
|
|
|
|
char _license[] SEC("license") = "GPL";
|
|
|
|
struct kptr_nested_array_2 {
|
|
struct bpf_cpumask __kptr * mask;
|
|
};
|
|
|
|
struct kptr_nested_array_1 {
|
|
/* Make btf_parse_fields() in map_create() return -E2BIG */
|
|
struct kptr_nested_array_2 d_2[CPUMASK_KPTR_FIELDS_MAX + 1];
|
|
};
|
|
|
|
struct kptr_nested_array {
|
|
struct kptr_nested_array_1 d_1;
|
|
};
|
|
|
|
private(MASK_NESTED) static struct kptr_nested_array global_mask_nested_arr;
|
|
|
|
/* Prototype for all of the program trace events below:
|
|
*
|
|
* TRACE_EVENT(task_newtask,
|
|
* TP_PROTO(struct task_struct *p, u64 clone_flags)
|
|
*/
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("Unreleased reference")
|
|
int BPF_PROG(test_alloc_no_release, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *cpumask;
|
|
|
|
cpumask = create_cpumask();
|
|
__sink(cpumask);
|
|
|
|
/* cpumask is never released. */
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("NULL pointer passed to trusted arg0")
|
|
int BPF_PROG(test_alloc_double_release, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *cpumask;
|
|
|
|
cpumask = create_cpumask();
|
|
|
|
/* cpumask is released twice. */
|
|
bpf_cpumask_release(cpumask);
|
|
bpf_cpumask_release(cpumask);
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("must be referenced")
|
|
int BPF_PROG(test_acquire_wrong_cpumask, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *cpumask;
|
|
|
|
/* Can't acquire a non-struct bpf_cpumask. */
|
|
cpumask = bpf_cpumask_acquire((struct bpf_cpumask *)task->cpus_ptr);
|
|
__sink(cpumask);
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("bpf_cpumask_set_cpu args#1 expected pointer to STRUCT bpf_cpumask")
|
|
int BPF_PROG(test_mutate_cpumask, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
/* Can't set the CPU of a non-struct bpf_cpumask. */
|
|
bpf_cpumask_set_cpu(0, (struct bpf_cpumask *)task->cpus_ptr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("Unreleased reference")
|
|
int BPF_PROG(test_insert_remove_no_release, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *cpumask;
|
|
struct __cpumask_map_value *v;
|
|
|
|
cpumask = create_cpumask();
|
|
if (!cpumask)
|
|
return 0;
|
|
|
|
if (cpumask_map_insert(cpumask))
|
|
return 0;
|
|
|
|
v = cpumask_map_value_lookup();
|
|
if (!v)
|
|
return 0;
|
|
|
|
cpumask = bpf_kptr_xchg(&v->cpumask, NULL);
|
|
|
|
/* cpumask is never released. */
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("NULL pointer passed to trusted arg0")
|
|
int BPF_PROG(test_cpumask_null, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
/* NULL passed to KF_TRUSTED_ARGS kfunc. */
|
|
bpf_cpumask_empty(NULL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("R2 must be a rcu pointer")
|
|
int BPF_PROG(test_global_mask_out_of_rcu, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *local, *prev;
|
|
|
|
local = create_cpumask();
|
|
if (!local)
|
|
return 0;
|
|
|
|
prev = bpf_kptr_xchg(&global_mask, local);
|
|
if (prev) {
|
|
bpf_cpumask_release(prev);
|
|
err = 3;
|
|
return 0;
|
|
}
|
|
|
|
bpf_rcu_read_lock();
|
|
local = global_mask;
|
|
if (!local) {
|
|
err = 4;
|
|
bpf_rcu_read_unlock();
|
|
return 0;
|
|
}
|
|
|
|
bpf_rcu_read_unlock();
|
|
|
|
/* RCU region is exited before calling KF_RCU kfunc. */
|
|
|
|
bpf_cpumask_test_cpu(0, (const struct cpumask *)local);
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("NULL pointer passed to trusted arg1")
|
|
int BPF_PROG(test_global_mask_no_null_check, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *local, *prev;
|
|
|
|
local = create_cpumask();
|
|
if (!local)
|
|
return 0;
|
|
|
|
prev = bpf_kptr_xchg(&global_mask, local);
|
|
if (prev) {
|
|
bpf_cpumask_release(prev);
|
|
err = 3;
|
|
return 0;
|
|
}
|
|
|
|
bpf_rcu_read_lock();
|
|
local = global_mask;
|
|
|
|
/* No NULL check is performed on global cpumask kptr. */
|
|
bpf_cpumask_test_cpu(0, (const struct cpumask *)local);
|
|
|
|
bpf_rcu_read_unlock();
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("Possibly NULL pointer passed to helper arg2")
|
|
int BPF_PROG(test_global_mask_rcu_no_null_check, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *prev, *curr;
|
|
|
|
curr = bpf_cpumask_create();
|
|
if (!curr)
|
|
return 0;
|
|
|
|
prev = bpf_kptr_xchg(&global_mask, curr);
|
|
if (prev)
|
|
bpf_cpumask_release(prev);
|
|
|
|
bpf_rcu_read_lock();
|
|
curr = global_mask;
|
|
/* PTR_TO_BTF_ID | PTR_MAYBE_NULL | MEM_RCU passed to bpf_kptr_xchg() */
|
|
prev = bpf_kptr_xchg(&global_mask, curr);
|
|
bpf_rcu_read_unlock();
|
|
if (prev)
|
|
bpf_cpumask_release(prev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("has no valid kptr")
|
|
int BPF_PROG(test_invalid_nested_array, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *local, *prev;
|
|
|
|
local = create_cpumask();
|
|
if (!local)
|
|
return 0;
|
|
|
|
prev = bpf_kptr_xchg(&global_mask_nested_arr.d_1.d_2[CPUMASK_KPTR_FIELDS_MAX].mask, local);
|
|
if (prev) {
|
|
bpf_cpumask_release(prev);
|
|
err = 3;
|
|
return 0;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("type=scalar expected=fp")
|
|
int BPF_PROG(test_populate_invalid_destination, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
struct bpf_cpumask *invalid = (struct bpf_cpumask *)0x123456;
|
|
u64 bits;
|
|
int ret;
|
|
|
|
ret = bpf_cpumask_populate((struct cpumask *)invalid, &bits, sizeof(bits));
|
|
if (!ret)
|
|
err = 2;
|
|
|
|
return 0;
|
|
}
|
|
|
|
SEC("tp_btf/task_newtask")
|
|
__failure __msg("leads to invalid memory access")
|
|
int BPF_PROG(test_populate_invalid_source, struct task_struct *task, u64 clone_flags)
|
|
{
|
|
void *garbage = (void *)0x123456;
|
|
struct bpf_cpumask *local;
|
|
int ret;
|
|
|
|
local = create_cpumask();
|
|
if (!local) {
|
|
err = 1;
|
|
return 0;
|
|
}
|
|
|
|
ret = bpf_cpumask_populate((struct cpumask *)local, garbage, 8);
|
|
if (!ret)
|
|
err = 2;
|
|
|
|
bpf_cpumask_release(local);
|
|
|
|
return 0;
|
|
}
|