Skip to content

Commit

Permalink
---
Browse files Browse the repository at this point in the history
yaml
---
r: 79845
b: refs/heads/master
c: af65d64
h: refs/heads/master
i:
  79843: e2e9751
v: v3
  • Loading branch information
Roland McGrath authored and Ingo Molnar committed Jan 30, 2008
1 parent 4668509 commit 9b13f5c
Show file tree
Hide file tree
Showing 11 changed files with 147 additions and 61 deletions.
2 changes: 1 addition & 1 deletion [refs]
Original file line number Diff line number Diff line change
@@ -1,2 +1,2 @@
---
refs/heads/master: 00f8b1bc0e44ba94fb33e1fbd8ac82841d7cc570
refs/heads/master: af65d64845a90c8f2fc90b97e2148ff74672e979
5 changes: 5 additions & 0 deletions trunk/Documentation/kernel-parameters.txt
Original file line number Diff line number Diff line change
Expand Up @@ -1988,6 +1988,11 @@ and is between 256 and 4096 characters. It is defined in the file
vdso=1: enable VDSO (default)
vdso=0: disable VDSO mapping

vdso32= [X86-32,X86-64]
vdso32=2: enable compat VDSO (default with COMPAT_VDSO)
vdso32=1: enable 32-bit VDSO (default)
vdso32=0: disable 32-bit VDSO mapping

vector= [IA-64,SMP]
vector=percpu: enable percpu vector domain

Expand Down
4 changes: 2 additions & 2 deletions trunk/arch/x86/Kconfig
Original file line number Diff line number Diff line change
Expand Up @@ -1191,9 +1191,9 @@ config HOTPLUG_CPU
config COMPAT_VDSO
bool "Compat VDSO support"
default y
depends on X86_32
depends on X86_32 || IA32_EMULATION
help
Map the VDSO to the predictable old-style address too.
Map the 32-bit VDSO to the predictable old-style address too.
---help---
Say N here if you are running a sufficiently recent glibc
version (2.3.3 or later), to remove the high-mapped
Expand Down
2 changes: 1 addition & 1 deletion trunk/arch/x86/ia32/Makefile
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
#

obj-$(CONFIG_IA32_EMULATION) := ia32entry.o sys_ia32.o ia32_signal.o tls32.o \
ia32_binfmt.o fpu32.o ptrace32.o syscall32.o syscall32_syscall.o
ia32_binfmt.o fpu32.o ptrace32.o

sysv-$(CONFIG_SYSVIPC) := ipc32.o
obj-$(CONFIG_IA32_EMULATION) += $(sysv-y)
Expand Down
9 changes: 4 additions & 5 deletions trunk/arch/x86/ia32/ia32_binfmt.c
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@
#include <asm/i387.h>
#include <asm/uaccess.h>
#include <asm/ia32.h>
#include <asm/vsyscall32.h>
#include <asm/vdso.h>

#undef ELF_ARCH
#undef ELF_CLASS
Expand All @@ -47,14 +47,13 @@
#define AT_SYSINFO 32
#define AT_SYSINFO_EHDR 33

int sysctl_vsyscall32 = 1;
extern int sysctl_vsyscall32;

#undef ARCH_DLINFO
#define ARCH_DLINFO do { \
if (sysctl_vsyscall32) { \
current->mm->context.vdso = (void *)VSYSCALL32_BASE; \
NEW_AUX_ENT(AT_SYSINFO, (u32)(u64)VSYSCALL32_VSYSCALL); \
NEW_AUX_ENT(AT_SYSINFO_EHDR, VSYSCALL32_BASE); \
NEW_AUX_ENT(AT_SYSINFO, (u32)VDSO_ENTRY); \
NEW_AUX_ENT(AT_SYSINFO_EHDR, (u32)VDSO_CURRENT_BASE); \
} \
} while(0)

Expand Down
22 changes: 14 additions & 8 deletions trunk/arch/x86/ia32/ia32_signal.c
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@
#include <asm/sigcontext32.h>
#include <asm/fpu32.h>
#include <asm/proto.h>
#include <asm/vsyscall32.h>
#include <asm/vdso.h>

#define DEBUG_SIG 0

Expand Down Expand Up @@ -465,13 +465,16 @@ int ia32_setup_frame(int sig, struct k_sigaction *ka,
goto give_sigsegv;
}

/* Return stub is in 32bit vsyscall page */
if (current->binfmt->hasvdso)
restorer = VSYSCALL32_SIGRETURN;
else
restorer = (void *)&frame->retcode;
if (ka->sa.sa_flags & SA_RESTORER)
if (ka->sa.sa_flags & SA_RESTORER) {
restorer = ka->sa.sa_restorer;
} else {
/* Return stub is in 32bit vsyscall page */
if (current->binfmt->hasvdso)
restorer = VDSO32_SYMBOL(current->mm->context.vdso,
sigreturn);
else
restorer = (void *)&frame->retcode;
}
err |= __put_user(ptr_to_compat(restorer), &frame->pretcode);

/*
Expand Down Expand Up @@ -519,7 +522,7 @@ int ia32_setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info,
{
struct rt_sigframe __user *frame;
struct exec_domain *ed = current_thread_info()->exec_domain;
void __user *restorer = VSYSCALL32_RTSIGRETURN;
void __user *restorer;
int err = 0;

/* __copy_to_user optimizes that into a single 8 byte store */
Expand Down Expand Up @@ -564,6 +567,9 @@ int ia32_setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info,

if (ka->sa.sa_flags & SA_RESTORER)
restorer = ka->sa.sa_restorer;
else
restorer = VDSO32_SYMBOL(current->mm->context.vdso,
rt_sigreturn);
err |= __put_user(ptr_to_compat(restorer), &frame->pretcode);

/*
Expand Down
2 changes: 1 addition & 1 deletion trunk/arch/x86/vdso/Makefile
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ vobjs-y := vdso-note.o vclock_gettime.o vgetcpu.o vvar.o

# files to link into kernel
obj-$(VDSO64-y) += vma.o vdso.o
obj-$(CONFIG_X86_32) += vdso32.o vdso32-setup.o
obj-$(VDSO32-y) += vdso32.o vdso32-setup.o

vobjs := $(foreach F,$(vobjs-y),$(obj)/$F)

Expand Down
119 changes: 93 additions & 26 deletions trunk/arch/x86/vdso/vdso32-setup.c
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,7 @@
#include <asm/elf.h>
#include <asm/tlbflush.h>
#include <asm/vdso.h>
#include <asm/proto.h>

enum {
VDSO_DISABLED = 0,
Expand All @@ -37,24 +38,43 @@ enum {
#define VDSO_DEFAULT VDSO_ENABLED
#endif

#ifdef CONFIG_X86_64
#define vdso_enabled sysctl_vsyscall32
#define arch_setup_additional_pages syscall32_setup_pages
#endif

/*
* This is the difference between the prelinked addresses in the vDSO images
* and the VDSO_HIGH_BASE address where CONFIG_COMPAT_VDSO places the vDSO
* in the user address space.
*/
#define VDSO_ADDR_ADJUST (VDSO_HIGH_BASE - (unsigned long)VDSO32_PRELINK)

/*
* Should the kernel map a VDSO page into processes and pass its
* address down to glibc upon exec()?
*/
unsigned int __read_mostly vdso_enabled = VDSO_DEFAULT;

EXPORT_SYMBOL_GPL(vdso_enabled);

static int __init vdso_setup(char *s)
{
vdso_enabled = simple_strtoul(s, NULL, 0);

return 1;
}

__setup("vdso=", vdso_setup);
/*
* For consistency, the argument vdso32=[012] affects the 32-bit vDSO
* behavior on both 64-bit and 32-bit kernels.
* On 32-bit kernels, vdso=[012] means the same thing.
*/
__setup("vdso32=", vdso_setup);

extern asmlinkage void ia32_sysenter_target(void);
#ifdef CONFIG_X86_32
__setup_param("vdso=", vdso32_setup, vdso_setup, 0);

EXPORT_SYMBOL_GPL(vdso_enabled);
#endif

static __init void reloc_symtab(Elf32_Ehdr *ehdr,
unsigned offset, unsigned size)
Expand All @@ -79,7 +99,7 @@ static __init void reloc_symtab(Elf32_Ehdr *ehdr,
case STT_FUNC:
case STT_SECTION:
case STT_FILE:
sym->st_value += VDSO_HIGH_BASE;
sym->st_value += VDSO_ADDR_ADJUST;
}
}
}
Expand All @@ -105,7 +125,7 @@ static __init void reloc_dyn(Elf32_Ehdr *ehdr, unsigned offset)
case DT_VERNEED:
case DT_ADDRRNGLO ... DT_ADDRRNGHI:
/* definitely pointers needing relocation */
dyn->d_un.d_ptr += VDSO_HIGH_BASE;
dyn->d_un.d_ptr += VDSO_ADDR_ADJUST;
break;

case DT_ENCODING ... OLD_DT_LOOS-1:
Expand All @@ -114,7 +134,7 @@ static __init void reloc_dyn(Elf32_Ehdr *ehdr, unsigned offset)
they're even */
if (dyn->d_tag >= DT_ENCODING &&
(dyn->d_tag & 1) == 0)
dyn->d_un.d_ptr += VDSO_HIGH_BASE;
dyn->d_un.d_ptr += VDSO_ADDR_ADJUST;
break;

case DT_VERDEFNUM:
Expand Down Expand Up @@ -143,15 +163,15 @@ static __init void relocate_vdso(Elf32_Ehdr *ehdr)
int i;

BUG_ON(memcmp(ehdr->e_ident, ELFMAG, 4) != 0 ||
!elf_check_arch(ehdr) ||
!elf_check_arch_ia32(ehdr) ||
ehdr->e_type != ET_DYN);

ehdr->e_entry += VDSO_HIGH_BASE;
ehdr->e_entry += VDSO_ADDR_ADJUST;

/* rebase phdrs */
phdr = (void *)ehdr + ehdr->e_phoff;
for (i = 0; i < ehdr->e_phnum; i++) {
phdr[i].p_vaddr += VDSO_HIGH_BASE;
phdr[i].p_vaddr += VDSO_ADDR_ADJUST;

/* relocate dynamic stuff */
if (phdr[i].p_type == PT_DYNAMIC)
Expand All @@ -164,7 +184,7 @@ static __init void relocate_vdso(Elf32_Ehdr *ehdr)
if (!(shdr[i].sh_flags & SHF_ALLOC))
continue;

shdr[i].sh_addr += VDSO_HIGH_BASE;
shdr[i].sh_addr += VDSO_ADDR_ADJUST;

if (shdr[i].sh_type == SHT_SYMTAB ||
shdr[i].sh_type == SHT_DYNSYM)
Expand All @@ -173,6 +193,45 @@ static __init void relocate_vdso(Elf32_Ehdr *ehdr)
}
}

/*
* These symbols are defined by vdso32.S to mark the bounds
* of the ELF DSO images included therein.
*/
extern const char vdso32_default_start, vdso32_default_end;
extern const char vdso32_sysenter_start, vdso32_sysenter_end;
static struct page *vdso32_pages[1];

#ifdef CONFIG_X86_64

static int use_sysenter __read_mostly = -1;

#define vdso32_sysenter() (use_sysenter > 0)

/* May not be __init: called during resume */
void syscall32_cpu_init(void)
{
if (use_sysenter < 0)
use_sysenter = (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL);

/* Load these always in case some future AMD CPU supports
SYSENTER from compat mode too. */
checking_wrmsrl(MSR_IA32_SYSENTER_CS, (u64)__KERNEL_CS);
checking_wrmsrl(MSR_IA32_SYSENTER_ESP, 0ULL);
checking_wrmsrl(MSR_IA32_SYSENTER_EIP, (u64)ia32_sysenter_target);

wrmsrl(MSR_CSTAR, ia32_cstar_target);
}

#define compat_uses_vma 1

static inline void map_compat_vdso(int map)
{
}

#else /* CONFIG_X86_32 */

#define vdso32_sysenter() (boot_cpu_has(X86_FEATURE_SEP))

void enable_sep_cpu(void)
{
int cpu = get_cpu();
Expand Down Expand Up @@ -210,13 +269,7 @@ static int __init gate_vma_init(void)
return 0;
}

/*
* These symbols are defined by vsyscall.o to mark the bounds
* of the ELF DSO images included therein.
*/
extern const char vsyscall_int80_start, vsyscall_int80_end;
extern const char vsyscall_sysenter_start, vsyscall_sysenter_end;
static struct page *syscall_pages[1];
#define compat_uses_vma 0

static void map_compat_vdso(int map)
{
Expand All @@ -227,31 +280,35 @@ static void map_compat_vdso(int map)

vdso_mapped = map;

__set_fixmap(FIX_VDSO, page_to_pfn(syscall_pages[0]) << PAGE_SHIFT,
__set_fixmap(FIX_VDSO, page_to_pfn(vdso32_pages[0]) << PAGE_SHIFT,
map ? PAGE_READONLY_EXEC : PAGE_NONE);

/* flush stray tlbs */
flush_tlb_all();
}

#endif /* CONFIG_X86_64 */

int __init sysenter_setup(void)
{
void *syscall_page = (void *)get_zeroed_page(GFP_ATOMIC);
const void *vsyscall;
size_t vsyscall_len;

syscall_pages[0] = virt_to_page(syscall_page);
vdso32_pages[0] = virt_to_page(syscall_page);

#ifdef CONFIG_X86_32
gate_vma_init();

printk("Compat vDSO mapped to %08lx.\n", __fix_to_virt(FIX_VDSO));
#endif

if (!boot_cpu_has(X86_FEATURE_SEP)) {
vsyscall = &vsyscall_int80_start;
vsyscall_len = &vsyscall_int80_end - &vsyscall_int80_start;
if (!vdso32_sysenter()) {
vsyscall = &vdso32_default_start;
vsyscall_len = &vdso32_default_end - &vdso32_default_start;
} else {
vsyscall = &vsyscall_sysenter_start;
vsyscall_len = &vsyscall_sysenter_end - &vsyscall_sysenter_start;
vsyscall = &vdso32_sysenter_start;
vsyscall_len = &vdso32_sysenter_end - &vdso32_sysenter_start;
}

memcpy(syscall_page, vsyscall, vsyscall_len);
Expand Down Expand Up @@ -284,7 +341,9 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int exstack)
ret = addr;
goto up_fail;
}
}

if (compat_uses_vma || !compat) {
/*
* MAYWRITE to allow gdb to COW and set breakpoints
*
Expand All @@ -298,7 +357,7 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int exstack)
VM_READ|VM_EXEC|
VM_MAYREAD|VM_MAYWRITE|VM_MAYEXEC|
VM_ALWAYSDUMP,
syscall_pages);
vdso32_pages);

if (ret)
goto up_fail;
Expand All @@ -314,6 +373,12 @@ int arch_setup_additional_pages(struct linux_binprm *bprm, int exstack)
return ret;
}

#ifdef CONFIG_X86_64

__initcall(sysenter_setup);

#else /* CONFIG_X86_32 */

const char *arch_vma_name(struct vm_area_struct *vma)
{
if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso)
Expand Down Expand Up @@ -342,3 +407,5 @@ int in_gate_area_no_task(unsigned long addr)
{
return 0;
}

#endif /* CONFIG_X86_64 */
Loading

0 comments on commit 9b13f5c

Please sign in to comment.