From 01b0408520ea9bd9470a1493f2445bd2e89bf24d Mon Sep 17 00:00:00 2001 From: Lee Schermerhorn Date: Mon, 14 Dec 2009 17:59:54 -0800 Subject: [PATCH] --- yaml --- r: 176313 b: refs/heads/master c: 418b27ef50e7e9b0c2fbd88db804bf065e5eb1a6 h: refs/heads/master i: 176311: 1b0948d14f84ed9f2ddbe2c1f3bfa2d1e66bdb50 v: v3 --- [refs] | 2 +- trunk/mm/internal.h | 12 ------------ trunk/mm/migrate.c | 4 ++-- 3 files changed, 3 insertions(+), 15 deletions(-) diff --git a/[refs] b/[refs] index 99e1e2675f31..814666e8adf7 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: 4eb2b1dcd598f8489130405c81c60c289896d92a +refs/heads/master: 418b27ef50e7e9b0c2fbd88db804bf065e5eb1a6 diff --git a/trunk/mm/internal.h b/trunk/mm/internal.h index a4b927cdca09..4fe67a162cb4 100644 --- a/trunk/mm/internal.h +++ b/trunk/mm/internal.h @@ -63,18 +63,6 @@ static inline unsigned long page_order(struct page *page) return page_private(page); } -/* - * unevictable_migrate_page() called only from migrate_page_copy() to - * migrate unevictable flag to new page. - * Note that the old page has been isolated from the LRU lists at this - * point so we don't need to worry about LRU statistics. - */ -static inline void unevictable_migrate_page(struct page *new, struct page *old) -{ - if (TestClearPageUnevictable(old)) - SetPageUnevictable(new); -} - #ifdef CONFIG_MMU extern long mlock_vma_pages_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); diff --git a/trunk/mm/migrate.c b/trunk/mm/migrate.c index 2a0ea3ef509e..efddbf0926b2 100644 --- a/trunk/mm/migrate.c +++ b/trunk/mm/migrate.c @@ -288,8 +288,8 @@ static void migrate_page_copy(struct page *newpage, struct page *page) if (TestClearPageActive(page)) { VM_BUG_ON(PageUnevictable(page)); SetPageActive(newpage); - } else - unevictable_migrate_page(newpage, page); + } else if (TestClearPageUnevictable(page)) + SetPageUnevictable(newpage); if (PageChecked(page)) SetPageChecked(newpage); if (PageMappedToDisk(page))