diff --git a/[refs] b/[refs] index fc842a5ebc96..5d0c4a653872 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: 09f345da758fca1222b0971b65b2fddbdf78bb83 +refs/heads/master: 345ee8392dc149ca529f80e40583928977ad592e diff --git a/trunk/drivers/char/Kconfig b/trunk/drivers/char/Kconfig index bf18d757b876..a509b8d79781 100644 --- a/trunk/drivers/char/Kconfig +++ b/trunk/drivers/char/Kconfig @@ -457,7 +457,7 @@ config LEGACY_PTYS config LEGACY_PTY_COUNT int "Maximum number of legacy PTY in use" depends on LEGACY_PTYS - range 1 256 + range 0 256 default "256" ---help--- The maximum number of legacy PTYs that can be used at any one time. diff --git a/trunk/fs/exec.c b/trunk/fs/exec.c index 282240afe99e..4ccaaa4b13b2 100644 --- a/trunk/fs/exec.c +++ b/trunk/fs/exec.c @@ -1780,12 +1780,6 @@ int do_coredump(long signr, int exit_code, struct pt_regs * regs) but keep the previous behaviour for now. */ if (!ispipe && !S_ISREG(inode->i_mode)) goto close_fail; - /* - * Dont allow local users get cute and trick others to coredump - * into their pre-created files: - */ - if (inode->i_uid != current->fsuid) - goto close_fail; if (!file->f_op) goto close_fail; if (!file->f_op->write) diff --git a/trunk/include/linux/sched.h b/trunk/include/linux/sched.h index ac3d496fbd20..ee800e7a70de 100644 --- a/trunk/include/linux/sched.h +++ b/trunk/include/linux/sched.h @@ -282,10 +282,6 @@ static inline void touch_all_softlockup_watchdogs(void) /* Attach to any functions which should be ignored in wchan output. */ #define __sched __attribute__((__section__(".sched.text"))) - -/* Linker adds these: start and end of __sched functions */ -extern char __sched_text_start[], __sched_text_end[]; - /* Is this address in the __sched functions? */ extern int in_sched_functions(unsigned long addr); diff --git a/trunk/kernel/sched.c b/trunk/kernel/sched.c index 98dcdf272db3..38933cafea8a 100644 --- a/trunk/kernel/sched.c +++ b/trunk/kernel/sched.c @@ -5466,7 +5466,7 @@ sd_alloc_ctl_domain_table(struct sched_domain *sd) return table; } -static ctl_table *sd_alloc_ctl_cpu_table(int cpu) +static ctl_table * sd_alloc_ctl_cpu_table(int cpu) { struct ctl_table *entry, *table; struct sched_domain *sd; @@ -6708,6 +6708,9 @@ void __init sched_init_smp(void) int in_sched_functions(unsigned long addr) { + /* Linker adds these: start and end of __sched functions */ + extern char __sched_text_start[], __sched_text_end[]; + return in_lock_functions(addr) || (addr >= (unsigned long)__sched_text_start && addr < (unsigned long)__sched_text_end); diff --git a/trunk/kernel/sched_debug.c b/trunk/kernel/sched_debug.c index d30467b47ddd..5d0d623a5465 100644 --- a/trunk/kernel/sched_debug.c +++ b/trunk/kernel/sched_debug.c @@ -327,12 +327,10 @@ void proc_sched_show_task(struct task_struct *p, struct seq_file *m) avg_atom = -1LL; avg_per_cpu = p->se.sum_exec_runtime; - if (p->se.nr_migrations) { - avg_per_cpu = div64_64(avg_per_cpu, - p->se.nr_migrations); - } else { + if (p->se.nr_migrations) + avg_per_cpu = div64_64(avg_per_cpu, p->se.nr_migrations); + else avg_per_cpu = -1LL; - } __PN(avg_atom); __PN(avg_per_cpu); diff --git a/trunk/kernel/sched_stats.h b/trunk/kernel/sched_stats.h index 5b32433e7ee5..630178e53bb6 100644 --- a/trunk/kernel/sched_stats.h +++ b/trunk/kernel/sched_stats.h @@ -52,8 +52,7 @@ static int show_schedstat(struct seq_file *seq, void *v) sd->lb_nobusyq[itype], sd->lb_nobusyg[itype]); } - seq_printf(seq, - " %u %u %u %u %u %u %u %u %u %u %u %u\n", + seq_printf(seq, " %u %u %u %u %u %u %u %u %u %u %u %u\n", sd->alb_count, sd->alb_failed, sd->alb_pushed, sd->sbe_count, sd->sbe_balanced, sd->sbe_pushed, sd->sbf_count, sd->sbf_balanced, sd->sbf_pushed, diff --git a/trunk/kernel/time/tick-sched.c b/trunk/kernel/time/tick-sched.c index cb89fa8db110..27a2338deb4a 100644 --- a/trunk/kernel/time/tick-sched.c +++ b/trunk/kernel/time/tick-sched.c @@ -133,8 +133,6 @@ void tick_nohz_update_jiffies(void) if (!ts->tick_stopped) return; - touch_softlockup_watchdog(); - cpu_clear(cpu, nohz_cpu_mask); now = ktime_get(); diff --git a/trunk/mm/page_alloc.c b/trunk/mm/page_alloc.c index 4ffed1cd158b..12376ae3f733 100644 --- a/trunk/mm/page_alloc.c +++ b/trunk/mm/page_alloc.c @@ -305,6 +305,7 @@ static inline void prep_zero_page(struct page *page, int order, gfp_t gfp_flags) { int i; + VM_BUG_ON((gfp_flags & (__GFP_WAIT | __GFP_HIGHMEM)) == __GFP_HIGHMEM); /* * clear_highpage() will use KM_USER0, so it's a bug to use __GFP_ZERO * and __GFP_HIGHMEM from hard or soft interrupt context. diff --git a/trunk/mm/shmem.c b/trunk/mm/shmem.c index 51b3d6ccddab..253d205914ba 100644 --- a/trunk/mm/shmem.c +++ b/trunk/mm/shmem.c @@ -1072,7 +1072,7 @@ shmem_alloc_page(gfp_t gfp, struct shmem_inode_info *info, pvma.vm_policy = mpol_shared_policy_lookup(&info->policy, idx); pvma.vm_pgoff = idx; pvma.vm_end = PAGE_SIZE; - page = alloc_page_vma(gfp, &pvma, 0); + page = alloc_page_vma(gfp | __GFP_ZERO, &pvma, 0); mpol_free(pvma.vm_policy); return page; } @@ -1093,7 +1093,7 @@ shmem_swapin(struct shmem_inode_info *info,swp_entry_t entry,unsigned long idx) static inline struct page * shmem_alloc_page(gfp_t gfp,struct shmem_inode_info *info, unsigned long idx) { - return alloc_page(gfp); + return alloc_page(gfp | __GFP_ZERO); } #endif @@ -1306,7 +1306,6 @@ static int shmem_getpage(struct inode *inode, unsigned long idx, info->alloced++; spin_unlock(&info->lock); - clear_highpage(filepage); flush_dcache_page(filepage); SetPageUptodate(filepage); }