summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorDavid Gibson <dgibson@samba.org>2002-06-27 05:36:41 -0400
committerPaul Mackerras <paulus@samba.org>2002-06-27 05:36:41 -0400
commit856091affeb5399ba13a2b2c8266918727583fb3 (patch)
tree20d9f33a9e2ad3423d28c353694bd810724e724a
parentf200709a8f1916813064511162e33c9d4bffb013 (diff)
PPC32: Fixes and cleanups for PPC40x processors.
Add branch-to-self after return-from-interrupt, fix critical exception handling, fix synchronization in set_context, other cleanups.
-rw-r--r--arch/ppc/config.in3
-rw-r--r--arch/ppc/kernel/entry.S34
-rw-r--r--arch/ppc/kernel/head_4xx.S135
-rw-r--r--arch/ppc/kernel/mk_defs.c2
-rw-r--r--arch/ppc/mm/4xx_mmu.c5
-rw-r--r--include/asm-ppc/ppc_asm.h5
-rw-r--r--include/asm-ppc/tlbflush.h9
7 files changed, 142 insertions, 51 deletions
diff --git a/arch/ppc/config.in b/arch/ppc/config.in
index 3858e19028ae..dddff7e6c876 100644
--- a/arch/ppc/config.in
+++ b/arch/ppc/config.in
@@ -288,9 +288,6 @@ if [ "$CONFIG_ADVANCED_OPTIONS" = "y" ]; then
if [ "$CONFIG_8xx" = "y" ]; then
bool "Pinned Kernel TLBs (860 ONLY)" CONFIG_PIN_TLB
fi
- if [ "$CONFIG_40x" = "y" ]; then
- bool "Pinned Kernel TLBs" CONFIG_PIN_TLB
- fi
if [ "$CONFIG_ALL_PPC" != "y" ]; then
bool "Set the boot link/load address" CONFIG_BOOT_LOAD_BOOL
if [ "$CONFIG_BOOT_LOAD_BOOL" = "y" ]; then
diff --git a/arch/ppc/kernel/entry.S b/arch/ppc/kernel/entry.S
index 1414c7e02b91..b0ccb55bfd9c 100644
--- a/arch/ppc/kernel/entry.S
+++ b/arch/ppc/kernel/entry.S
@@ -164,7 +164,6 @@ ret_from_syscall:
andi. r0,r9,(_TIF_SYSCALL_TRACE|_TIF_SIGPENDING|_TIF_NEED_RESCHED)
bne- syscall_exit_work
syscall_exit_cont:
- PPC405_ERR77(0,r1)
stwcx. r0,0,r1 /* to clear the reservation */
lwz r4,_LINK(r1)
lwz r5,_CCR(r1)
@@ -568,6 +567,7 @@ exc_exit_start:
exc_exit_restart_end:
PPC405_ERR77_SYNC
rfi
+ b . /* prevent prefetch past rfi */
crit_exc_exit:
mtcrf 0xff,r10
/* avoid any possible TLB misses here by turning off MSR.DR, we
@@ -576,18 +576,40 @@ crit_exc_exit:
mtmsr r10
isync
tophys(r1, r1)
- lwz r9,_SRR0(r1)
- lwz r10,_SRR1(r1)
- mtspr SRR0,r9
+ lwz r9,_DEAR(r1)
+ lwz r10,_ESR(r1)
+ mtspr SPRN_DEAR,r9
+ mtspr SPRN_ESR,r10
lwz r11,_NIP(r1)
- mtspr SRR1,r10
lwz r12,_MSR(r1)
mtspr SRR2,r11
mtspr SRR3,r12
- REST_4GPRS(9, r1)
+ lwz r9,GPR9(r1)
+ lwz r12,GPR12(r1)
+ lwz r10,crit_sprg0@l(0)
+ mtspr SPRN_SPRG0,r10
+ lwz r10,crit_sprg1@l(0)
+ mtspr SPRN_SPRG1,r10
+ lwz r10,crit_sprg4@l(0)
+ mtspr SPRN_SPRG4,r10
+ lwz r10,crit_sprg5@l(0)
+ mtspr SPRN_SPRG5,r10
+ lwz r10,crit_sprg6@l(0)
+ mtspr SPRN_SPRG6,r10
+ lwz r10,crit_sprg7@l(0)
+ mtspr SPRN_SPRG7,r10
+ lwz r10,crit_srr0@l(0)
+ mtspr SRR0,r10
+ lwz r10,crit_srr1@l(0)
+ mtspr SRR1,r10
+ lwz r10,crit_pid@l(0)
+ mtspr SPRN_PID,r10
+ lwz r10,crit_r10@l(0)
+ lwz r11,crit_r11@l(0)
lwz r1,GPR1(r1)
PPC405_ERR77_SYNC
rfci
+ b . /* prevent prefetch past rfci */
#endif /* CONFIG_4xx */
recheck:
diff --git a/arch/ppc/kernel/head_4xx.S b/arch/ppc/kernel/head_4xx.S
index 69e9398df14b..beae3cee8672 100644
--- a/arch/ppc/kernel/head_4xx.S
+++ b/arch/ppc/kernel/head_4xx.S
@@ -82,6 +82,7 @@ turn_on_mmu:
mtspr SRR0,r0
SYNC
rfi /* enables MMU */
+ b . /* prevent prefetch past rfi */
/*
* This area is used for temporarily saving registers during the
@@ -89,7 +90,28 @@ turn_on_mmu:
*/
. = 0xc0
crit_save:
- .space 8
+_GLOBAL(crit_r10)
+ .space 4
+_GLOBAL(crit_r11)
+ .space 4
+_GLOBAL(crit_sprg0)
+ .space 4
+_GLOBAL(crit_sprg1)
+ .space 4
+_GLOBAL(crit_sprg4)
+ .space 4
+_GLOBAL(crit_sprg5)
+ .space 4
+_GLOBAL(crit_sprg6)
+ .space 4
+_GLOBAL(crit_sprg7)
+ .space 4
+_GLOBAL(crit_pid)
+ .space 4
+_GLOBAL(crit_srr0)
+ .space 4
+_GLOBAL(crit_srr1)
+ .space 4
/*
* Exception vector entry code. This code runs with address translation
@@ -140,6 +162,59 @@ crit_save:
* This is OK since we don't support SMP on these processors.
*/
#define CRITICAL_EXCEPTION_PROLOG \
+ stw r10,crit_r10@l(0); /* save two registers to work with */\
+ stw r11,crit_r11@l(0); \
+ mfspr r10,SPRG0; \
+ stw r10,crit_sprg0@l(0); \
+ mfspr r10,SPRG1; \
+ stw r10,crit_sprg1@l(0); \
+ mfspr r10,SPRG4; \
+ stw r10,crit_sprg4@l(0); \
+ mfspr r10,SPRG5; \
+ stw r10,crit_sprg5@l(0); \
+ mfspr r10,SPRG6; \
+ stw r10,crit_sprg6@l(0); \
+ mfspr r10,SPRG7; \
+ stw r10,crit_sprg7@l(0); \
+ mfspr r10,SPRN_PID; \
+ stw r10,crit_pid@l(0); \
+ mfspr r10,SRR0; \
+ stw r10,crit_srr0@l(0); \
+ mfspr r10,SRR1; \
+ stw r10,crit_srr1@l(0); \
+ mfcr r10; /* save CR in r10 for now */\
+ mfspr r11,SPRN_SRR3; /* check whether user or kernel */\
+ andi. r11,r11,MSR_PR; \
+ lis r11,critical_stack_top@h; \
+ ori r11,r11,critical_stack_top@l; \
+ beq 1f; \
+ /* COMING FROM USER MODE */ \
+ mfspr r11,SPRG3; /* if from user, start at top of */\
+ lwz r11,THREAD_INFO-THREAD(r11); /* this thread's kernel stack */\
+ addi r11,r11,THREAD_SIZE; \
+1: subi r11,r11,INT_FRAME_SIZE; /* Allocate an exception frame */\
+ tophys(r11,r11); \
+ stw r10,_CCR(r11); /* save various registers */\
+ stw r12,GPR12(r11); \
+ stw r9,GPR9(r11); \
+ mflr r10; \
+ stw r10,_LINK(r11); \
+ mfspr r12,SPRN_DEAR; /* save DEAR and ESR in the frame */\
+ stw r12,_DEAR(r11); /* since they may have had stuff */\
+ mfspr r9,SPRN_ESR; /* in them at the point where the */\
+ stw r9,_ESR(r11); /* exception was taken */\
+ mfspr r12,SRR2; \
+ stw r1,GPR1(r11); \
+ mfspr r9,SRR3; \
+ stw r1,0(r11); \
+ tovirt(r1,r11); \
+ rlwinm r9,r9,0,14,12; /* clear MSR_WE (necessary?) */\
+ stw r0,GPR0(r11); \
+ SAVE_4GPRS(3, r11); \
+ SAVE_2GPRS(7, r11)
+
+#if 0
+#define CRITICAL_EXCEPTION_PROLOG \
stw r10,crit_save@l(0); /* save two registers to work with */\
stw r11,4+crit_save@l(0); \
mfcr r10; /* save CR in r10 for now */\
@@ -173,6 +248,7 @@ crit_save:
stw r0,GPR0(r11); \
SAVE_4GPRS(3, r11); \
SAVE_2GPRS(7, r11)
+#endif
/*
* Exception vectors.
@@ -197,13 +273,14 @@ label:
START_EXCEPTION(n, label); \
CRITICAL_EXCEPTION_PROLOG; \
addi r3,r1,STACK_FRAME_OVERHEAD; \
- EXC_XFER_TEMPLATE(hdlr, n+2, NOCOPY, transfer_to_handler_full, \
+ EXC_XFER_TEMPLATE(hdlr, n+2, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \
+ NOCOPY, transfer_to_handler_full, \
ret_from_except_full)
-#define EXC_XFER_TEMPLATE(hdlr, trap, copyee, tfer, ret) \
+#define EXC_XFER_TEMPLATE(hdlr, trap, msr, copyee, tfer, ret) \
li r10,trap; \
stw r10,TRAP(r11); \
- li r10,MSR_KERNEL; \
+ li r10,msr; \
copyee(r10, r9); \
bl tfer; \
.long hdlr; \
@@ -213,19 +290,19 @@ label:
#define NOCOPY(d, s)
#define EXC_XFER_STD(n, hdlr) \
- EXC_XFER_TEMPLATE(hdlr, n, NOCOPY, transfer_to_handler_full, \
+ EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, NOCOPY, transfer_to_handler_full, \
ret_from_except_full)
#define EXC_XFER_LITE(n, hdlr) \
- EXC_XFER_TEMPLATE(hdlr, n+1, NOCOPY, transfer_to_handler, \
+ EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, NOCOPY, transfer_to_handler, \
ret_from_except)
#define EXC_XFER_EE(n, hdlr) \
- EXC_XFER_TEMPLATE(hdlr, n, COPY_EE, transfer_to_handler_full, \
+ EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, COPY_EE, transfer_to_handler_full, \
ret_from_except_full)
#define EXC_XFER_EE_LITE(n, hdlr) \
- EXC_XFER_TEMPLATE(hdlr, n+1, COPY_EE, transfer_to_handler, \
+ EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, COPY_EE, transfer_to_handler, \
ret_from_except)
@@ -347,6 +424,7 @@ label:
mfspr r10, SPRG0
PPC405_ERR77_SYNC
rfi /* Should sync shadow TLBs */
+ b . /* prevent prefetch past rfi */
2:
/* The bailout. Restore registers to pre-exception conditions
@@ -615,6 +693,7 @@ label:
mfspr r10, SPRG0
PPC405_ERR77_SYNC
rfi /* Should sync shadow TLBs */
+ b . /* prevent prefetch past rfi */
2:
/* The bailout. Restore registers to pre-exception conditions
@@ -721,6 +800,7 @@ check_single_step_in_exception:
sync
rfci /* return to the exception handler */
+ b . /* prevent prefetch past rfi */
2:
mtcrf 0xff,r10 /* restore registers */
@@ -746,31 +826,18 @@ check_single_step_in_exception:
* Actually, it will fit now, but oh well.....a common place
* to load the TLB.
*/
+tlb_4xx_index:
+ .long 0
finish_tlb_load:
-
- /* Since it has a unified TLB, and we can take data faults on
- * instruction pages by copying data, we have to check if the
- * EPN is already in the TLB.
- */
- tlbsx. r9, 0, r10
- beq 6f
-
/* load the next available TLB index.
*/
- lis r12, tlb_4xx_index@h
- ori r12, r12, tlb_4xx_index@l
- tophys(r12, r12)
- lwz r9, 0(r12)
+ lwz r9, tlb_4xx_index@l(0)
addi r9, r9, 1
-#ifdef CONFIG_PIN_TLB
cmpwi 0, r9, 61 /* reserve entries 62, 63 for kernel */
ble 7f
li r9, 0
7:
-#else
- andi. r9, r9, (PPC4XX_TLB_SIZE-1)
-#endif
- stw r9, 0(r12)
+ stw r9, tlb_4xx_index@l(0)
6:
tlbwe r11, r9, TLB_DATA /* Load TLB LO */
@@ -804,6 +871,7 @@ finish_tlb_load:
mfspr r10, SPRG0
PPC405_ERR77_SYNC
rfi /* Should sync shadow TLBs */
+ b . /* prevent prefetch past rfi */
/* extern void giveup_fpu(struct task_struct *prev)
*
@@ -857,14 +925,11 @@ start_here:
mtspr SRR0,r4
mtspr SRR1,r3
rfi
+ b . /* prevent prefetch past rfi */
/* Load up the kernel context */
2:
SYNC /* Force all PTE updates to finish */
-#ifndef CONFIG_PIN_TLB
- tlbia /* Clear all TLB entries */
- sync /* wait for tlbia/tlbie to finish */
-#endif
/* set up the PTE pointers for the Abatron bdiGDB.
*/
@@ -883,6 +948,7 @@ start_here:
mtspr SRR0,r3
mtspr SRR1,r4
rfi /* enable MMU and jump to start_kernel */
+ b . /* prevent prefetch past rfi */
/* Set up the initial MMU state so we can do the first level of
* kernel initialization. This maps the first 16 MBytes of memory 1:1
@@ -956,7 +1022,10 @@ _GLOBAL(set_context)
lwz r5, 0xf0(r5)
stw r4, 0x4(r5)
#endif
+ sync
mtspr SPRN_PID,r3
+ isync /* Need an isync to flush shadow */
+ /* TLBs after changing PID */
blr
/* We put a few things here that have to be page-aligned. This stuff
@@ -969,6 +1038,14 @@ _GLOBAL(empty_zero_page)
_GLOBAL(swapper_pg_dir)
.space 4096
+
+/* Stack for handling critical exceptions from kernel mode */
+ .section .bss
+critical_stack_bottom:
+ .space 4096
+critical_stack_top:
+ .previous
+
/* This space gets a copy of optional info passed to us by the bootstrap
* which is used to pass parameters into the kernel like root=/dev/sda1, etc.
*/
diff --git a/arch/ppc/kernel/mk_defs.c b/arch/ppc/kernel/mk_defs.c
index d4cce91f178c..99bfc3b087b8 100644
--- a/arch/ppc/kernel/mk_defs.c
+++ b/arch/ppc/kernel/mk_defs.c
@@ -112,8 +112,6 @@ main(void)
*/
DEFINE(_DEAR, STACK_FRAME_OVERHEAD+offsetof(struct pt_regs, dar));
DEFINE(_ESR, STACK_FRAME_OVERHEAD+offsetof(struct pt_regs, dsisr));
- DEFINE(_SRR0, STACK_FRAME_OVERHEAD+offsetof(struct pt_regs, dar));
- DEFINE(_SRR1, STACK_FRAME_OVERHEAD+offsetof(struct pt_regs, dsisr));
DEFINE(ORIG_GPR3, STACK_FRAME_OVERHEAD+offsetof(struct pt_regs, orig_gpr3));
DEFINE(RESULT, STACK_FRAME_OVERHEAD+offsetof(struct pt_regs, result));
DEFINE(TRAP, STACK_FRAME_OVERHEAD+offsetof(struct pt_regs, trap));
diff --git a/arch/ppc/mm/4xx_mmu.c b/arch/ppc/mm/4xx_mmu.c
index bed7823c620e..22c54bbfc9c9 100644
--- a/arch/ppc/mm/4xx_mmu.c
+++ b/arch/ppc/mm/4xx_mmu.c
@@ -54,11 +54,6 @@
#include <asm/machdep.h>
#include <asm/setup.h>
-/* Used by the 4xx TLB replacement exception handler.
- * Just needed it declared someplace (and initialized to zero).
- */
-unsigned int tlb_4xx_index;
-
/*
* MMU_init_hw does the chip-specific initialization of the MMU hardware.
*/
diff --git a/include/asm-ppc/ppc_asm.h b/include/asm-ppc/ppc_asm.h
index 0c33c12f3939..f11b229000e7 100644
--- a/include/asm-ppc/ppc_asm.h
+++ b/include/asm-ppc/ppc_asm.h
@@ -51,7 +51,6 @@
/*
* Once a version of gas that understands the AltiVec instructions
- * is freely available, we can do this the normal way... - paulus
*/
#define LVX(r,a,b) .long (31<<26)+((r)<<21)+((a)<<16)+((b)<<11)+(103<<1)
#define STVX(r,a,b) .long (31<<26)+((r)<<21)+((a)<<16)+((b)<<11)+(231<<1)
@@ -169,7 +168,11 @@ END_FTR_SECTION_IFCLR(CPU_FTR_601)
#else
#define FIX_SRR1(ra, rb)
+#ifndef CONFIG_40x
#define RFI rfi
+#else
+#define RFI rfi; b . /* Prevent prefetch past rfi */
+#endif
#define MTMSRD(r) mtmsr r
#define CLR_TOP32(r)
#endif /* CONFIG_PPC64BRIDGE */
diff --git a/include/asm-ppc/tlbflush.h b/include/asm-ppc/tlbflush.h
index 83e6f1ae6c35..af4295e46861 100644
--- a/include/asm-ppc/tlbflush.h
+++ b/include/asm-ppc/tlbflush.h
@@ -21,21 +21,20 @@ extern void _tlbie(unsigned long address);
extern void _tlbia(void);
#if defined(CONFIG_4xx)
-#define __tlbia() asm volatile ("tlbia; sync" : : : "memory")
static inline void flush_tlb_all(void)
- { __tlbia(); }
+ { _tlbia(); }
static inline void flush_tlb_mm(struct mm_struct *mm)
- { __tlbia(); }
+ { _tlbia(); }
static inline void flush_tlb_page(struct vm_area_struct *vma,
unsigned long vmaddr)
{ _tlbie(vmaddr); }
static inline void flush_tlb_range(struct vm_area_struct *vma,
unsigned long start, unsigned long end)
- { __tlbia(); }
+ { _tlbia(); }
static inline void flush_tlb_kernel_range(unsigned long start,
unsigned long end)
- { __tlbia(); }
+ { _tlbia(); }
#define update_mmu_cache(vma, addr, pte) do { } while (0)
#elif defined(CONFIG_8xx)