mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-04 16:25:34 +00:00

Currently, vdso.so files are prepared and analyzed by a combination of objcopy, nm, some linker script tricks, and some simple ELF parsers in the kernel. Replace all of that with plain C code that runs at build time. All five vdso images now generate .c files that are compiled and linked in to the kernel image. This should cause only one userspace-visible change: the loaded vDSO images are stripped more heavily than they used to be. Everything outside the loadable segment is dropped. In particular, this causes the section table and section name strings to be missing. This should be fine: real dynamic loaders don't load or inspect these tables anyway. The result is roughly equivalent to eu-strip's --strip-sections option. The purpose of this change is to enable the vvar and hpet mappings to be moved to the page following the vDSO load segment. Currently, it is possible for the section table to extend into the page after the load segment, so, if we map it, it risks overlapping the vvar or hpet page. This happens whenever the load segment is just under a multiple of PAGE_SIZE. The only real subtlety here is that the old code had a C file with inline assembler that did 'call VDSO32_vsyscall' and a linker script that defined 'VDSO32_vsyscall = __kernel_vsyscall'. This most likely worked by accident: the linker script entry defines a symbol associated with an address as opposed to an alias for the real dynamic symbol __kernel_vsyscall. That caused ld to relocate the reference at link time instead of leaving an interposable dynamic relocation. Since the VDSO32_vsyscall hack is no longer needed, I now use 'call __kernel_vsyscall', and I added -Bsymbolic to make it work. vdso2c will generate an error and abort the build if the resulting image contains any dynamic relocations, so we won't silently generate bad vdso images. (Dynamic relocations are a problem because nothing will even attempt to relocate the vdso.) Signed-off-by: Andy Lutomirski <luto@amacapital.net> Link: http://lkml.kernel.org/r/2c4fcf45524162a34d87fdda1eb046b2a5cecee7.1399317206.git.luto@amacapital.net Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
253 lines
5.1 KiB
C
253 lines
5.1 KiB
C
/*
|
|
* (C) Copyright 2002 Linus Torvalds
|
|
* Portions based on the vdso-randomization code from exec-shield:
|
|
* Copyright(C) 2005-2006, Red Hat, Inc., Ingo Molnar
|
|
*
|
|
* This file contains the needed initializations to support sysenter.
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/thread_info.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/gfp.h>
|
|
#include <linux/string.h>
|
|
#include <linux/elf.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/err.h>
|
|
#include <linux/module.h>
|
|
#include <linux/slab.h>
|
|
|
|
#include <asm/cpufeature.h>
|
|
#include <asm/msr.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/unistd.h>
|
|
#include <asm/elf.h>
|
|
#include <asm/tlbflush.h>
|
|
#include <asm/vdso.h>
|
|
#include <asm/proto.h>
|
|
#include <asm/fixmap.h>
|
|
#include <asm/hpet.h>
|
|
#include <asm/vvar.h>
|
|
#include <asm/vdso32.h>
|
|
|
|
#ifdef CONFIG_COMPAT_VDSO
|
|
#define VDSO_DEFAULT 0
|
|
#else
|
|
#define VDSO_DEFAULT 1
|
|
#endif
|
|
|
|
#ifdef CONFIG_X86_64
|
|
#define arch_setup_additional_pages syscall32_setup_pages
|
|
#endif
|
|
|
|
/*
|
|
* Should the kernel map a VDSO page into processes and pass its
|
|
* address down to glibc upon exec()?
|
|
*/
|
|
unsigned int __read_mostly vdso32_enabled = VDSO_DEFAULT;
|
|
|
|
static int __init vdso32_setup(char *s)
|
|
{
|
|
vdso32_enabled = simple_strtoul(s, NULL, 0);
|
|
|
|
if (vdso32_enabled > 1)
|
|
pr_warn("vdso32 values other than 0 and 1 are no longer allowed; vdso disabled\n");
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* For consistency, the argument vdso32=[012] affects the 32-bit vDSO
|
|
* behavior on both 64-bit and 32-bit kernels.
|
|
* On 32-bit kernels, vdso=[012] means the same thing.
|
|
*/
|
|
__setup("vdso32=", vdso32_setup);
|
|
|
|
#ifdef CONFIG_X86_32
|
|
__setup_param("vdso=", vdso_setup, vdso32_setup, 0);
|
|
#endif
|
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
#define vdso32_sysenter() (boot_cpu_has(X86_FEATURE_SYSENTER32))
|
|
#define vdso32_syscall() (boot_cpu_has(X86_FEATURE_SYSCALL32))
|
|
|
|
#else /* CONFIG_X86_32 */
|
|
|
|
#define vdso32_sysenter() (boot_cpu_has(X86_FEATURE_SEP))
|
|
#define vdso32_syscall() (0)
|
|
|
|
#endif /* CONFIG_X86_64 */
|
|
|
|
#if defined(CONFIG_X86_32) || defined(CONFIG_COMPAT)
|
|
const struct vdso_image *selected_vdso32;
|
|
#endif
|
|
|
|
int __init sysenter_setup(void)
|
|
{
|
|
#ifdef CONFIG_COMPAT
|
|
if (vdso32_syscall())
|
|
selected_vdso32 = &vdso_image_32_syscall;
|
|
else
|
|
#endif
|
|
if (vdso32_sysenter())
|
|
selected_vdso32 = &vdso_image_32_sysenter;
|
|
else
|
|
selected_vdso32 = &vdso_image_32_int80;
|
|
|
|
init_vdso_image(selected_vdso32);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Setup a VMA at program startup for the vsyscall page */
|
|
int arch_setup_additional_pages(struct linux_binprm *bprm, int uses_interp)
|
|
{
|
|
struct mm_struct *mm = current->mm;
|
|
unsigned long addr;
|
|
int ret = 0;
|
|
struct vm_area_struct *vma;
|
|
unsigned long vdso32_size = selected_vdso32->size;
|
|
|
|
#ifdef CONFIG_X86_X32_ABI
|
|
if (test_thread_flag(TIF_X32))
|
|
return x32_setup_additional_pages(bprm, uses_interp);
|
|
#endif
|
|
|
|
if (vdso32_enabled != 1) /* Other values all mean "disabled" */
|
|
return 0;
|
|
|
|
down_write(&mm->mmap_sem);
|
|
|
|
addr = get_unmapped_area(NULL, 0, vdso32_size + VDSO_OFFSET(VDSO_PREV_PAGES), 0, 0);
|
|
if (IS_ERR_VALUE(addr)) {
|
|
ret = addr;
|
|
goto up_fail;
|
|
}
|
|
|
|
addr += VDSO_OFFSET(VDSO_PREV_PAGES);
|
|
|
|
current->mm->context.vdso = (void __user *)addr;
|
|
|
|
/*
|
|
* MAYWRITE to allow gdb to COW and set breakpoints
|
|
*/
|
|
ret = install_special_mapping(mm,
|
|
addr,
|
|
vdso32_size,
|
|
VM_READ|VM_EXEC|
|
|
VM_MAYREAD|VM_MAYWRITE|VM_MAYEXEC,
|
|
selected_vdso32->pages);
|
|
|
|
if (ret)
|
|
goto up_fail;
|
|
|
|
vma = _install_special_mapping(mm,
|
|
addr - VDSO_OFFSET(VDSO_PREV_PAGES),
|
|
VDSO_OFFSET(VDSO_PREV_PAGES),
|
|
VM_READ,
|
|
NULL);
|
|
|
|
if (IS_ERR(vma)) {
|
|
ret = PTR_ERR(vma);
|
|
goto up_fail;
|
|
}
|
|
|
|
ret = remap_pfn_range(vma,
|
|
addr - VDSO_OFFSET(VDSO_VVAR_PAGE),
|
|
__pa_symbol(&__vvar_page) >> PAGE_SHIFT,
|
|
PAGE_SIZE,
|
|
PAGE_READONLY);
|
|
|
|
if (ret)
|
|
goto up_fail;
|
|
|
|
#ifdef CONFIG_HPET_TIMER
|
|
if (hpet_address) {
|
|
ret = io_remap_pfn_range(vma,
|
|
addr - VDSO_OFFSET(VDSO_HPET_PAGE),
|
|
hpet_address >> PAGE_SHIFT,
|
|
PAGE_SIZE,
|
|
pgprot_noncached(PAGE_READONLY));
|
|
|
|
if (ret)
|
|
goto up_fail;
|
|
}
|
|
#endif
|
|
|
|
if (selected_vdso32->sym_VDSO32_SYSENTER_RETURN)
|
|
current_thread_info()->sysenter_return =
|
|
current->mm->context.vdso +
|
|
selected_vdso32->sym_VDSO32_SYSENTER_RETURN;
|
|
|
|
up_fail:
|
|
if (ret)
|
|
current->mm->context.vdso = NULL;
|
|
|
|
up_write(&mm->mmap_sem);
|
|
|
|
return ret;
|
|
}
|
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
subsys_initcall(sysenter_setup);
|
|
|
|
#ifdef CONFIG_SYSCTL
|
|
/* Register vsyscall32 into the ABI table */
|
|
#include <linux/sysctl.h>
|
|
|
|
static struct ctl_table abi_table2[] = {
|
|
{
|
|
.procname = "vsyscall32",
|
|
.data = &vdso32_enabled,
|
|
.maxlen = sizeof(int),
|
|
.mode = 0644,
|
|
.proc_handler = proc_dointvec
|
|
},
|
|
{}
|
|
};
|
|
|
|
static struct ctl_table abi_root_table2[] = {
|
|
{
|
|
.procname = "abi",
|
|
.mode = 0555,
|
|
.child = abi_table2
|
|
},
|
|
{}
|
|
};
|
|
|
|
static __init int ia32_binfmt_init(void)
|
|
{
|
|
register_sysctl_table(abi_root_table2);
|
|
return 0;
|
|
}
|
|
__initcall(ia32_binfmt_init);
|
|
#endif
|
|
|
|
#else /* CONFIG_X86_32 */
|
|
|
|
const char *arch_vma_name(struct vm_area_struct *vma)
|
|
{
|
|
if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso)
|
|
return "[vdso]";
|
|
return NULL;
|
|
}
|
|
|
|
struct vm_area_struct *get_gate_vma(struct mm_struct *mm)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
int in_gate_area(struct mm_struct *mm, unsigned long addr)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
int in_gate_area_no_mm(unsigned long addr)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
#endif /* CONFIG_X86_64 */
|