Skip to content

Commit

Permalink
[PATCH] mm: powerpc ptlock comments
Browse files Browse the repository at this point in the history
Update comments (only) on page_table_lock and mmap_sem in arch/powerpc.
Removed the comment on page_table_lock from hash_huge_page: since it's no
longer taking page_table_lock itself, it's irrelevant whether others are; but
how it is safe (even against huge file truncation?) I can't say.

Signed-off-by: Hugh Dickins <hugh@veritas.com>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
  • Loading branch information
Hugh Dickins authored and Linus Torvalds committed Nov 24, 2005
1 parent cc3327e commit 01edcd8
Show file tree
Hide file tree
Showing 4 changed files with 10 additions and 6 deletions.
4 changes: 1 addition & 3 deletions arch/powerpc/mm/hugetlbpage.c
Original file line number Diff line number Diff line change
Expand Up @@ -754,9 +754,7 @@ int hash_huge_page(struct mm_struct *mm, unsigned long access,
}

/*
* No need to use ldarx/stdcx here because all who
* might be updating the pte will hold the
* page_table_lock
* No need to use ldarx/stdcx here
*/
*ptep = __pte(new_pte & ~_PAGE_BUSY);

Expand Down
2 changes: 1 addition & 1 deletion arch/powerpc/mm/mem.c
Original file line number Diff line number Diff line change
Expand Up @@ -495,7 +495,7 @@ EXPORT_SYMBOL(flush_icache_user_range);
* We use it to preload an HPTE into the hash table corresponding to
* the updated linux PTE.
*
* This must always be called with the mm->page_table_lock held
* This must always be called with the pte lock held.
*/
void update_mmu_cache(struct vm_area_struct *vma, unsigned long address,
pte_t pte)
Expand Down
6 changes: 6 additions & 0 deletions arch/powerpc/mm/tlb_32.c
Original file line number Diff line number Diff line change
Expand Up @@ -149,6 +149,12 @@ void flush_tlb_mm(struct mm_struct *mm)
return;
}

/*
* It is safe to go down the mm's list of vmas when called
* from dup_mmap, holding mmap_sem. It would also be safe from
* unmap_region or exit_mmap, but not from vmtruncate on SMP -
* but it seems dup_mmap is the only SMP case which gets here.
*/
for (mp = mm->mmap; mp != NULL; mp = mp->vm_next)
flush_range(mp->vm_mm, mp->vm_start, mp->vm_end);
FINISH_FLUSH;
Expand Down
4 changes: 2 additions & 2 deletions arch/powerpc/mm/tlb_64.c
Original file line number Diff line number Diff line change
Expand Up @@ -95,7 +95,7 @@ static void pte_free_submit(struct pte_freelist_batch *batch)

void pgtable_free_tlb(struct mmu_gather *tlb, pgtable_free_t pgf)
{
/* This is safe as we are holding page_table_lock */
/* This is safe since tlb_gather_mmu has disabled preemption */
cpumask_t local_cpumask = cpumask_of_cpu(smp_processor_id());
struct pte_freelist_batch **batchp = &__get_cpu_var(pte_freelist_cur);

Expand Down Expand Up @@ -206,7 +206,7 @@ void __flush_tlb_pending(struct ppc64_tlb_batch *batch)

void pte_free_finish(void)
{
/* This is safe as we are holding page_table_lock */
/* This is safe since tlb_gather_mmu has disabled preemption */
struct pte_freelist_batch **batchp = &__get_cpu_var(pte_freelist_cur);

if (*batchp == NULL)
Expand Down

0 comments on commit 01edcd8

Please sign in to comment.