Skip to content

Commit

Permalink
[SPARC64]: Fix two kernel linear mapping setup bugs.
Browse files Browse the repository at this point in the history
This was caught and identified by Greg Onufer.

Since we setup the 256M/4M bitmap table after taking over the trap
table, it's possible for some 4M mapping to get loaded in the TLB
beforhand which later will be 256M mappings.

This can cause illegal TLB multiple-match conditions.  Fix this by
setting up the bitmap before we take over the trap table.

Next, __flush_tlb_all() was not doing anything on hypervisor
platforms.  Fix by adding sun4v_mmu_demap_all() and calling it.

Signed-off-by: David S. Miller <davem@davemloft.net>
  • Loading branch information
David S. Miller committed Dec 13, 2007
1 parent da8cadb commit 8f36145
Show file tree
Hide file tree
Showing 3 changed files with 36 additions and 9 deletions.
12 changes: 12 additions & 0 deletions arch/sparc64/kernel/entry.S
Original file line number Diff line number Diff line change
Expand Up @@ -2593,3 +2593,15 @@ sun4v_mmustat_info:
retl
nop
.size sun4v_mmustat_info, .-sun4v_mmustat_info

.globl sun4v_mmu_demap_all
.type sun4v_mmu_demap_all,#function
sun4v_mmu_demap_all:
clr %o0
clr %o1
mov HV_MMU_ALL, %o2
mov HV_FAST_MMU_DEMAP_ALL, %o5
ta HV_FAST_TRAP
retl
nop
.size sun4v_mmu_demap_all, .-sun4v_mmu_demap_all
29 changes: 20 additions & 9 deletions arch/sparc64/mm/init.c
Original file line number Diff line number Diff line change
Expand Up @@ -1133,14 +1133,9 @@ static void __init mark_kpte_bitmap(unsigned long start, unsigned long end)
}
}

static void __init kernel_physical_mapping_init(void)
static void __init init_kpte_bitmap(void)
{
unsigned long i;
#ifdef CONFIG_DEBUG_PAGEALLOC
unsigned long mem_alloced = 0UL;
#endif

read_obp_memory("reg", &pall[0], &pall_ents);

for (i = 0; i < pall_ents; i++) {
unsigned long phys_start, phys_end;
Expand All @@ -1149,14 +1144,24 @@ static void __init kernel_physical_mapping_init(void)
phys_end = phys_start + pall[i].reg_size;

mark_kpte_bitmap(phys_start, phys_end);
}
}

static void __init kernel_physical_mapping_init(void)
{
#ifdef CONFIG_DEBUG_PAGEALLOC
unsigned long i, mem_alloced = 0UL;

for (i = 0; i < pall_ents; i++) {
unsigned long phys_start, phys_end;

phys_start = pall[i].phys_addr;
phys_end = phys_start + pall[i].reg_size;

mem_alloced += kernel_map_range(phys_start, phys_end,
PAGE_KERNEL);
#endif
}

#ifdef CONFIG_DEBUG_PAGEALLOC
printk("Allocated %ld bytes for kernel page tables.\n",
mem_alloced);

Expand Down Expand Up @@ -1398,6 +1403,10 @@ void __init paging_init(void)

inherit_prom_mappings();

read_obp_memory("reg", &pall[0], &pall_ents);

init_kpte_bitmap();

/* Ok, we can use our TLB miss and window trap handlers safely. */
setup_tba();

Expand Down Expand Up @@ -1904,7 +1913,9 @@ void __flush_tlb_all(void)
"wrpr %0, %1, %%pstate"
: "=r" (pstate)
: "i" (PSTATE_IE));
if (tlb_type == spitfire) {
if (tlb_type == hypervisor) {
sun4v_mmu_demap_all();
} else if (tlb_type == spitfire) {
for (i = 0; i < 64; i++) {
/* Spitfire Errata #32 workaround */
/* NOTE: Always runs on spitfire, so no
Expand Down
4 changes: 4 additions & 0 deletions include/asm-sparc64/hypervisor.h
Original file line number Diff line number Diff line change
Expand Up @@ -709,6 +709,10 @@ extern unsigned long sun4v_mmu_tsb_ctx0(unsigned long num_descriptions,
*/
#define HV_FAST_MMU_DEMAP_ALL 0x24

#ifndef __ASSEMBLY__
extern void sun4v_mmu_demap_all(void);
#endif

/* mmu_map_perm_addr()
* TRAP: HV_FAST_TRAP
* FUNCTION: HV_FAST_MMU_MAP_PERM_ADDR
Expand Down

0 comments on commit 8f36145

Please sign in to comment.