[PATCH 1/2] powerpc: add kexec support on FSL-Book-E
Sebastian Andrzej Siewior
sebastian at breakpoint.cc
Tue Nov 4 07:01:35 EST 2008
From: Sebastian Andrzej Siewior <bigeasy at linutronix.de>
The relocate_new_kernel() code usually disables the MMU and the small code
operates on physicall pages while moving the kernel to its final position.
Book-E doesn't support this so a 1:1 mapping must be created.
This patch adds support for FSL-BOOK-E implementation.
Signed-off-by: Sebastian Andrzej Siewior <bigeasy at linutronix.de>
---
arch/powerpc/kernel/machine_kexec_32.c | 5 +-
arch/powerpc/kernel/misc_32.S | 129 ++++++++++++++++++++++++++++++-
2 files changed, 127 insertions(+), 7 deletions(-)
diff --git a/arch/powerpc/kernel/machine_kexec_32.c b/arch/powerpc/kernel/machine_kexec_32.c
index ae63a96..6fa8ed3 100644
--- a/arch/powerpc/kernel/machine_kexec_32.c
+++ b/arch/powerpc/kernel/machine_kexec_32.c
@@ -16,10 +16,10 @@
#include <asm/hw_irq.h>
#include <asm/io.h>
-typedef NORET_TYPE void (*relocate_new_kernel_t)(
+typedef void (*relocate_new_kernel_t)(
unsigned long indirection_page,
unsigned long reboot_code_buffer,
- unsigned long start_address) ATTRIB_NORET;
+ unsigned long start_address);
/*
* This is a generic machine_kexec function suitable at least for
@@ -57,6 +57,7 @@ void default_machine_kexec(struct kimage *image)
/* now call it */
rnk = (relocate_new_kernel_t) reboot_code_buffer;
(*rnk)(page_list, reboot_code_buffer_phys, image->start);
+ BUG();
}
int default_machine_kexec_prepare(struct kimage *image)
diff --git a/arch/powerpc/kernel/misc_32.S b/arch/powerpc/kernel/misc_32.S
index 7a6dfbc..68ab147 100644
--- a/arch/powerpc/kernel/misc_32.S
+++ b/arch/powerpc/kernel/misc_32.S
@@ -878,9 +878,120 @@ relocate_new_kernel:
/* r4 = reboot_code_buffer */
/* r5 = start_address */
- li r0, 0
+ mflr r28
+ mr r29, r3
+ mr r30, r4
+ mr r31, r5
+
+#ifdef CONFIG_FSL_BOOKE
+
+ li r25, 0 /* phys kernel start (low) */
+
+/* 1. Find the index of the entry we're executing in */
+ bl invstr /* Find our address */
+invstr:
+ mflr r6 /* Make it accessible */
+ mfmsr r7
+ rlwinm r4,r7,27,31,31 /* extract MSR[IS] */
+ mfspr r7, SPRN_PID0
+ slwi r7,r7,16
+ or r7,r7,r4
+ mtspr SPRN_MAS6,r7
+ tlbsx 0,r6 /* search MSR[IS], SPID=PID0 */
+ mfspr r7,SPRN_MAS1
+ andis. r7,r7,MAS1_VALID at h
+ bne match_TLB
/*
+ * We search just in PID0 because kernel's global mapping has to be
+ * there. We simply return to the caller if we didn't find the mapping
+ * since we didn't (yet) pass the point of no return. This should not
+ * happen.
+ */
+ mtlr r28
+ blr
+
+match_TLB:
+ mfspr r7,SPRN_MAS0
+ rlwinm r3,r7,16,20,31 /* Extract MAS0(Entry) */
+
+ mfspr r7,SPRN_MAS1 /* Insure IPROT set */
+ oris r7,r7,MAS1_IPROT at h
+ mtspr SPRN_MAS1,r7
+ tlbwe
+
+/* 2. Invalidate all entries except the entry we're executing in */
+ mfspr r9,SPRN_TLB1CFG
+ andi. r9,r9,0xfff
+ li r6,0 /* Set Entry counter to 0 */
+1:
+ lis r7,0x1000 /* Set MAS0(TLBSEL) = 1 */
+ rlwimi r7,r6,16,4,15 /* Setup MAS0 = TLBSEL | ESEL(r6) */
+ mtspr SPRN_MAS0,r7
+ tlbre
+ mfspr r7,SPRN_MAS1
+ rlwinm r7,r7,0,2,31 /* Clear MAS1 Valid and IPROT */
+ cmpw r3,r6
+ beq skpinv /* Dont update the current execution TLB */
+ mtspr SPRN_MAS1,r7
+ tlbwe
+ isync
+skpinv:
+ addi r6,r6,1 /* Increment */
+ cmpw r6,r9 /* Are we done? */
+ bne 1b /* If not, repeat */
+
+ /* Invalidate TLB0 */
+ li r6,0x04
+ tlbivax 0,r6
+ TLBSYNC
+ /* Invalidate TLB1 */
+ li r6,0x0c
+ tlbivax 0,r6
+ TLBSYNC
+
+/* 3. Setup a temp mapping and jump to it */
+ andi. r5, r3, 0x1 /* Find an entry not used and is non-zero */
+ addi r5, r5, 0x1
+ lis r7, 0x1000 /* Set MAS0(TLBSEL) = 1 */
+ rlwimi r7, r3, 16, 4, 15 /* Setup MAS0 = TLBSEL | ESEL(r3) */
+ mtspr SPRN_MAS0,r7
+ tlbre
+
+ /* Just modify the entry ID and EPN for the temp mapping */
+ lis r7,0x1000 /* Set MAS0(TLBSEL) = 1 */
+ rlwimi r7,r5,16,4,15 /* Setup MAS0 = TLBSEL | ESEL(r5) */
+ mtspr SPRN_MAS0,r7
+
+ xori r6,r4,1 /* Setup TMP mapping in the other Address space */
+ slwi r6,r6,12
+ oris r6,r6,(MAS1_VALID|MAS1_IPROT)@h
+ ori r6,r6,(MAS1_TSIZE(BOOKE_PAGESZ_1GB))@l
+ mtspr SPRN_MAS1,r6
+
+ lis r7, MAS2_I | MAS2_G
+ mtspr SPRN_MAS2,r7
+
+ li r8, 0
+ ori r8,r25,(MAS3_SX|MAS3_SW|MAS3_SR)
+ mtspr SPRN_MAS3,r8
+
+ tlbwe
+
+ xori r6, r4, 1
+ slwi r5, r6, 4 /* DS setup new context with other address space */
+ slwi r6, r6, 5 /* IS setup new context with other address space */
+ or r6, r6, r5
+
+ /* find our address */
+ addi r7, r30, final_copy_code - relocate_new_kernel
+
+ mtspr SPRN_SRR0,r7
+ mtspr SPRN_SRR1,r6
+ rfi
+#else
+ li r0, 0
+ /*
* Set Machine Status Register to a known status,
* switch the MMU off and jump to 1: in a single step.
*/
@@ -888,14 +999,22 @@ relocate_new_kernel:
mr r8, r0
ori r8, r8, MSR_RI|MSR_ME
mtspr SPRN_SRR1, r8
- addi r8, r4, 1f - relocate_new_kernel
+ addi r8, r4, final_copy_code - relocate_new_kernel
mtspr SPRN_SRR0, r8
sync
rfi
+#endif
-1:
- /* from this point address translation is turned off */
- /* and interrupts are disabled */
+final_copy_code:
+
+ mr r3, r29
+ mr r4, r30
+ mr r5, r31
+
+ li r0, 0
+
+ /* from this point address translation is turned off or we have */
+ /* a 1:1 mapping and interrupts are disabled */
/* set a new stack at the bottom of our page... */
/* (not really needed now) */
--
1.5.6.5
More information about the Linuxppc-dev
mailing list