mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

Introduce __COMPAT_scx_bpf_events() to use scx_bpf_events() in a
compatible way also with kernels that don't provide this kfunc.
This also fixes the following error with scx_qmap when running on a
kernel that does not provide scx_bpf_events():
; scx_bpf_events(&events, sizeof(events)); @ scx_qmap.bpf.c:777
318: (b7) r2 = 72 ; R2_w=72 async_cb
319: <invalid kfunc call>
kfunc 'scx_bpf_events' is referenced but wasn't resolved
Fixes: 9865f31d85
("sched_ext: Add scx_bpf_events() and scx_read_event() for BPF schedulers")
Signed-off-by: Andrea Righi <arighi@nvidia.com>
Signed-off-by: Tejun Heo <tj@kernel.org>
238 lines
9.2 KiB
C
238 lines
9.2 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Copyright (c) 2024 Meta Platforms, Inc. and affiliates.
|
|
* Copyright (c) 2024 Tejun Heo <tj@kernel.org>
|
|
* Copyright (c) 2024 David Vernet <dvernet@meta.com>
|
|
*/
|
|
#ifndef __SCX_COMPAT_BPF_H
|
|
#define __SCX_COMPAT_BPF_H
|
|
|
|
#define __COMPAT_ENUM_OR_ZERO(__type, __ent) \
|
|
({ \
|
|
__type __ret = 0; \
|
|
if (bpf_core_enum_value_exists(__type, __ent)) \
|
|
__ret = __ent; \
|
|
__ret; \
|
|
})
|
|
|
|
/* v6.12: 819513666966 ("sched_ext: Add cgroup support") */
|
|
#define __COMPAT_scx_bpf_task_cgroup(p) \
|
|
(bpf_ksym_exists(scx_bpf_task_cgroup) ? \
|
|
scx_bpf_task_cgroup((p)) : NULL)
|
|
|
|
/*
|
|
* v6.13: The verb `dispatch` was too overloaded and confusing. kfuncs are
|
|
* renamed to unload the verb.
|
|
*
|
|
* Build error is triggered if old names are used. New binaries work with both
|
|
* new and old names. The compat macros will be removed on v6.15 release.
|
|
*
|
|
* scx_bpf_dispatch_from_dsq() and friends were added during v6.12 by
|
|
* 4c30f5ce4f7a ("sched_ext: Implement scx_bpf_dispatch[_vtime]_from_dsq()").
|
|
* Preserve __COMPAT macros until v6.15.
|
|
*/
|
|
void scx_bpf_dispatch___compat(struct task_struct *p, u64 dsq_id, u64 slice, u64 enq_flags) __ksym __weak;
|
|
void scx_bpf_dispatch_vtime___compat(struct task_struct *p, u64 dsq_id, u64 slice, u64 vtime, u64 enq_flags) __ksym __weak;
|
|
bool scx_bpf_consume___compat(u64 dsq_id) __ksym __weak;
|
|
void scx_bpf_dispatch_from_dsq_set_slice___compat(struct bpf_iter_scx_dsq *it__iter, u64 slice) __ksym __weak;
|
|
void scx_bpf_dispatch_from_dsq_set_vtime___compat(struct bpf_iter_scx_dsq *it__iter, u64 vtime) __ksym __weak;
|
|
bool scx_bpf_dispatch_from_dsq___compat(struct bpf_iter_scx_dsq *it__iter, struct task_struct *p, u64 dsq_id, u64 enq_flags) __ksym __weak;
|
|
bool scx_bpf_dispatch_vtime_from_dsq___compat(struct bpf_iter_scx_dsq *it__iter, struct task_struct *p, u64 dsq_id, u64 enq_flags) __ksym __weak;
|
|
|
|
#define scx_bpf_dsq_insert(p, dsq_id, slice, enq_flags) \
|
|
(bpf_ksym_exists(scx_bpf_dsq_insert) ? \
|
|
scx_bpf_dsq_insert((p), (dsq_id), (slice), (enq_flags)) : \
|
|
scx_bpf_dispatch___compat((p), (dsq_id), (slice), (enq_flags)))
|
|
|
|
#define scx_bpf_dsq_insert_vtime(p, dsq_id, slice, vtime, enq_flags) \
|
|
(bpf_ksym_exists(scx_bpf_dsq_insert_vtime) ? \
|
|
scx_bpf_dsq_insert_vtime((p), (dsq_id), (slice), (vtime), (enq_flags)) : \
|
|
scx_bpf_dispatch_vtime___compat((p), (dsq_id), (slice), (vtime), (enq_flags)))
|
|
|
|
#define scx_bpf_dsq_move_to_local(dsq_id) \
|
|
(bpf_ksym_exists(scx_bpf_dsq_move_to_local) ? \
|
|
scx_bpf_dsq_move_to_local((dsq_id)) : \
|
|
scx_bpf_consume___compat((dsq_id)))
|
|
|
|
#define __COMPAT_scx_bpf_dsq_move_set_slice(it__iter, slice) \
|
|
(bpf_ksym_exists(scx_bpf_dsq_move_set_slice) ? \
|
|
scx_bpf_dsq_move_set_slice((it__iter), (slice)) : \
|
|
(bpf_ksym_exists(scx_bpf_dispatch_from_dsq_set_slice___compat) ? \
|
|
scx_bpf_dispatch_from_dsq_set_slice___compat((it__iter), (slice)) : \
|
|
(void)0))
|
|
|
|
#define __COMPAT_scx_bpf_dsq_move_set_vtime(it__iter, vtime) \
|
|
(bpf_ksym_exists(scx_bpf_dsq_move_set_vtime) ? \
|
|
scx_bpf_dsq_move_set_vtime((it__iter), (vtime)) : \
|
|
(bpf_ksym_exists(scx_bpf_dispatch_from_dsq_set_vtime___compat) ? \
|
|
scx_bpf_dispatch_from_dsq_set_vtime___compat((it__iter), (vtime)) : \
|
|
(void) 0))
|
|
|
|
#define __COMPAT_scx_bpf_dsq_move(it__iter, p, dsq_id, enq_flags) \
|
|
(bpf_ksym_exists(scx_bpf_dsq_move) ? \
|
|
scx_bpf_dsq_move((it__iter), (p), (dsq_id), (enq_flags)) : \
|
|
(bpf_ksym_exists(scx_bpf_dispatch_from_dsq___compat) ? \
|
|
scx_bpf_dispatch_from_dsq___compat((it__iter), (p), (dsq_id), (enq_flags)) : \
|
|
false))
|
|
|
|
#define __COMPAT_scx_bpf_dsq_move_vtime(it__iter, p, dsq_id, enq_flags) \
|
|
(bpf_ksym_exists(scx_bpf_dsq_move_vtime) ? \
|
|
scx_bpf_dsq_move_vtime((it__iter), (p), (dsq_id), (enq_flags)) : \
|
|
(bpf_ksym_exists(scx_bpf_dispatch_vtime_from_dsq___compat) ? \
|
|
scx_bpf_dispatch_vtime_from_dsq___compat((it__iter), (p), (dsq_id), (enq_flags)) : \
|
|
false))
|
|
|
|
#define scx_bpf_dispatch(p, dsq_id, slice, enq_flags) \
|
|
_Static_assert(false, "scx_bpf_dispatch() renamed to scx_bpf_dsq_insert()")
|
|
|
|
#define scx_bpf_dispatch_vtime(p, dsq_id, slice, vtime, enq_flags) \
|
|
_Static_assert(false, "scx_bpf_dispatch_vtime() renamed to scx_bpf_dsq_insert_vtime()")
|
|
|
|
#define scx_bpf_consume(dsq_id) ({ \
|
|
_Static_assert(false, "scx_bpf_consume() renamed to scx_bpf_dsq_move_to_local()"); \
|
|
false; \
|
|
})
|
|
|
|
#define scx_bpf_dispatch_from_dsq_set_slice(it__iter, slice) \
|
|
_Static_assert(false, "scx_bpf_dispatch_from_dsq_set_slice() renamed to scx_bpf_dsq_move_set_slice()")
|
|
|
|
#define scx_bpf_dispatch_from_dsq_set_vtime(it__iter, vtime) \
|
|
_Static_assert(false, "scx_bpf_dispatch_from_dsq_set_vtime() renamed to scx_bpf_dsq_move_set_vtime()")
|
|
|
|
#define scx_bpf_dispatch_from_dsq(it__iter, p, dsq_id, enq_flags) ({ \
|
|
_Static_assert(false, "scx_bpf_dispatch_from_dsq() renamed to scx_bpf_dsq_move()"); \
|
|
false; \
|
|
})
|
|
|
|
#define scx_bpf_dispatch_vtime_from_dsq(it__iter, p, dsq_id, enq_flags) ({ \
|
|
_Static_assert(false, "scx_bpf_dispatch_vtime_from_dsq() renamed to scx_bpf_dsq_move_vtime()"); \
|
|
false; \
|
|
})
|
|
|
|
#define __COMPAT_scx_bpf_dispatch_from_dsq_set_slice(it__iter, slice) \
|
|
_Static_assert(false, "__COMPAT_scx_bpf_dispatch_from_dsq_set_slice() renamed to __COMPAT_scx_bpf_dsq_move_set_slice()")
|
|
|
|
#define __COMPAT_scx_bpf_dispatch_from_dsq_set_vtime(it__iter, vtime) \
|
|
_Static_assert(false, "__COMPAT_scx_bpf_dispatch_from_dsq_set_vtime() renamed to __COMPAT_scx_bpf_dsq_move_set_vtime()")
|
|
|
|
#define __COMPAT_scx_bpf_dispatch_from_dsq(it__iter, p, dsq_id, enq_flags) ({ \
|
|
_Static_assert(false, "__COMPAT_scx_bpf_dispatch_from_dsq() renamed to __COMPAT_scx_bpf_dsq_move()"); \
|
|
false; \
|
|
})
|
|
|
|
#define __COMPAT_scx_bpf_dispatch_vtime_from_dsq(it__iter, p, dsq_id, enq_flags) ({ \
|
|
_Static_assert(false, "__COMPAT_scx_bpf_dispatch_vtime_from_dsq() renamed to __COMPAT_scx_bpf_dsq_move_vtime()"); \
|
|
false; \
|
|
})
|
|
|
|
/**
|
|
* __COMPAT_is_enq_cpu_selected - Test if SCX_ENQ_CPU_SELECTED is on
|
|
* in a compatible way. We will preserve this __COMPAT helper until v6.16.
|
|
*
|
|
* @enq_flags: enqueue flags from ops.enqueue()
|
|
*
|
|
* Return: True if SCX_ENQ_CPU_SELECTED is turned on in @enq_flags
|
|
*/
|
|
static inline bool __COMPAT_is_enq_cpu_selected(u64 enq_flags)
|
|
{
|
|
#ifdef HAVE_SCX_ENQ_CPU_SELECTED
|
|
/*
|
|
* This is the case that a BPF code compiled against vmlinux.h
|
|
* where the enum SCX_ENQ_CPU_SELECTED exists.
|
|
*/
|
|
|
|
/*
|
|
* We should temporarily suspend the macro expansion of
|
|
* 'SCX_ENQ_CPU_SELECTED'. This avoids 'SCX_ENQ_CPU_SELECTED' being
|
|
* rewritten to '__SCX_ENQ_CPU_SELECTED' when 'SCX_ENQ_CPU_SELECTED'
|
|
* is defined in 'scripts/gen_enums.py'.
|
|
*/
|
|
#pragma push_macro("SCX_ENQ_CPU_SELECTED")
|
|
#undef SCX_ENQ_CPU_SELECTED
|
|
u64 flag;
|
|
|
|
/*
|
|
* When the kernel did not have SCX_ENQ_CPU_SELECTED,
|
|
* select_task_rq_scx() has never been skipped. Thus, this case
|
|
* should be considered that the CPU has already been selected.
|
|
*/
|
|
if (!bpf_core_enum_value_exists(enum scx_enq_flags,
|
|
SCX_ENQ_CPU_SELECTED))
|
|
return true;
|
|
|
|
flag = bpf_core_enum_value(enum scx_enq_flags, SCX_ENQ_CPU_SELECTED);
|
|
return enq_flags & flag;
|
|
|
|
/*
|
|
* Once done, resume the macro expansion of 'SCX_ENQ_CPU_SELECTED'.
|
|
*/
|
|
#pragma pop_macro("SCX_ENQ_CPU_SELECTED")
|
|
#else
|
|
/*
|
|
* This is the case that a BPF code compiled against vmlinux.h
|
|
* where the enum SCX_ENQ_CPU_SELECTED does NOT exist.
|
|
*/
|
|
return true;
|
|
#endif /* HAVE_SCX_ENQ_CPU_SELECTED */
|
|
}
|
|
|
|
|
|
#define scx_bpf_now() \
|
|
(bpf_ksym_exists(scx_bpf_now) ? \
|
|
scx_bpf_now() : \
|
|
bpf_ktime_get_ns())
|
|
|
|
/*
|
|
* v6.15: Introduce event counters.
|
|
*
|
|
* Preserve the following macro until v6.17.
|
|
*/
|
|
#define __COMPAT_scx_bpf_events(events, size) \
|
|
(bpf_ksym_exists(scx_bpf_events) ? \
|
|
scx_bpf_events(events, size) : ({}))
|
|
|
|
/*
|
|
* v6.15: Introduce NUMA-aware kfuncs to operate with per-node idle
|
|
* cpumasks.
|
|
*
|
|
* Preserve the following __COMPAT_scx_*_node macros until v6.17.
|
|
*/
|
|
#define __COMPAT_scx_bpf_nr_node_ids() \
|
|
(bpf_ksym_exists(scx_bpf_nr_node_ids) ? \
|
|
scx_bpf_nr_node_ids() : 1U)
|
|
|
|
#define __COMPAT_scx_bpf_cpu_node(cpu) \
|
|
(bpf_ksym_exists(scx_bpf_cpu_node) ? \
|
|
scx_bpf_cpu_node(cpu) : 0)
|
|
|
|
#define __COMPAT_scx_bpf_get_idle_cpumask_node(node) \
|
|
(bpf_ksym_exists(scx_bpf_get_idle_cpumask_node) ? \
|
|
scx_bpf_get_idle_cpumask_node(node) : \
|
|
scx_bpf_get_idle_cpumask()) \
|
|
|
|
#define __COMPAT_scx_bpf_get_idle_smtmask_node(node) \
|
|
(bpf_ksym_exists(scx_bpf_get_idle_smtmask_node) ? \
|
|
scx_bpf_get_idle_smtmask_node(node) : \
|
|
scx_bpf_get_idle_smtmask())
|
|
|
|
#define __COMPAT_scx_bpf_pick_idle_cpu_node(cpus_allowed, node, flags) \
|
|
(bpf_ksym_exists(scx_bpf_pick_idle_cpu_node) ? \
|
|
scx_bpf_pick_idle_cpu_node(cpus_allowed, node, flags) : \
|
|
scx_bpf_pick_idle_cpu(cpus_allowed, flags))
|
|
|
|
#define __COMPAT_scx_bpf_pick_any_cpu_node(cpus_allowed, node, flags) \
|
|
(bpf_ksym_exists(scx_bpf_pick_any_cpu_node) ? \
|
|
scx_bpf_pick_any_cpu_node(cpus_allowed, node, flags) : \
|
|
scx_bpf_pick_any_cpu(cpus_allowed, flags))
|
|
|
|
/*
|
|
* Define sched_ext_ops. This may be expanded to define multiple variants for
|
|
* backward compatibility. See compat.h::SCX_OPS_LOAD/ATTACH().
|
|
*/
|
|
#define SCX_OPS_DEFINE(__name, ...) \
|
|
SEC(".struct_ops.link") \
|
|
struct sched_ext_ops __name = { \
|
|
__VA_ARGS__, \
|
|
};
|
|
|
|
#endif /* __SCX_COMPAT_BPF_H */
|