123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553 |
- /* SPDX-License-Identifier: GPL-2.0-or-later */
- /*
- * PowerPC version
- * Copyright (C) 1995-1996 Gary Thomas ([email protected])
- * Rewritten by Cort Dougan ([email protected]) for PReP
- * Copyright (C) 1996 Cort Dougan <[email protected]>
- * Adapted for Power Macintosh by Paul Mackerras.
- * Low-level exception handlers and MMU support
- * rewritten by Paul Mackerras.
- * Copyright (C) 1996 Paul Mackerras.
- * MPC8xx modifications Copyright (C) 1997 Dan Malek ([email protected]).
- *
- * This file contains the system call entry code, context switch
- * code, and exception/interrupt return code for PowerPC.
- */
- #include <linux/errno.h>
- #include <linux/err.h>
- #include <linux/sys.h>
- #include <linux/threads.h>
- #include <asm/reg.h>
- #include <asm/page.h>
- #include <asm/mmu.h>
- #include <asm/cputable.h>
- #include <asm/thread_info.h>
- #include <asm/ppc_asm.h>
- #include <asm/asm-offsets.h>
- #include <asm/unistd.h>
- #include <asm/ptrace.h>
- #include <asm/export.h>
- #include <asm/feature-fixups.h>
- #include <asm/barrier.h>
- #include <asm/kup.h>
- #include <asm/bug.h>
- #include <asm/interrupt.h>
- #include "head_32.h"
- /*
- * powerpc relies on return from interrupt/syscall being context synchronising
- * (which rfi is) to support ARCH_HAS_MEMBARRIER_SYNC_CORE without additional
- * synchronisation instructions.
- */
- /*
- * Align to 4k in order to ensure that all functions modyfing srr0/srr1
- * fit into one page in order to not encounter a TLB miss between the
- * modification of srr0/srr1 and the associated rfi.
- */
- .align 12
- #if defined(CONFIG_PPC_BOOK3S_32) || defined(CONFIG_PPC_E500)
- .globl prepare_transfer_to_handler
- prepare_transfer_to_handler:
- /* if from kernel, check interrupted DOZE/NAP mode */
- lwz r12,TI_LOCAL_FLAGS(r2)
- mtcrf 0x01,r12
- bt- 31-TLF_NAPPING,4f
- bt- 31-TLF_SLEEPING,7f
- blr
- 4: rlwinm r12,r12,0,~_TLF_NAPPING
- stw r12,TI_LOCAL_FLAGS(r2)
- b power_save_ppc32_restore
- 7: rlwinm r12,r12,0,~_TLF_SLEEPING
- stw r12,TI_LOCAL_FLAGS(r2)
- lwz r9,_MSR(r11) /* if sleeping, clear MSR.EE */
- rlwinm r9,r9,0,~MSR_EE
- lwz r12,_LINK(r11) /* and return to address in LR */
- REST_GPR(2, r11)
- b fast_exception_return
- _ASM_NOKPROBE_SYMBOL(prepare_transfer_to_handler)
- #endif /* CONFIG_PPC_BOOK3S_32 || CONFIG_PPC_E500 */
- #if defined(CONFIG_PPC_KUEP) && defined(CONFIG_PPC_BOOK3S_32)
- .globl __kuep_lock
- __kuep_lock:
- lwz r9, THREAD+THSR0(r2)
- update_user_segments_by_4 r9, r10, r11, r12
- blr
- __kuep_unlock:
- lwz r9, THREAD+THSR0(r2)
- rlwinm r9,r9,0,~SR_NX
- update_user_segments_by_4 r9, r10, r11, r12
- blr
- .macro kuep_lock
- bl __kuep_lock
- .endm
- .macro kuep_unlock
- bl __kuep_unlock
- .endm
- #else
- .macro kuep_lock
- .endm
- .macro kuep_unlock
- .endm
- #endif
- .globl transfer_to_syscall
- transfer_to_syscall:
- stw r3, ORIG_GPR3(r1)
- stw r11, GPR1(r1)
- stw r11, 0(r1)
- mflr r12
- stw r12, _LINK(r1)
- #ifdef CONFIG_BOOKE_OR_40x
- rlwinm r9,r9,0,14,12 /* clear MSR_WE (necessary?) */
- #endif
- lis r12,STACK_FRAME_REGS_MARKER@ha /* exception frame marker */
- SAVE_GPR(2, r1)
- addi r12,r12,STACK_FRAME_REGS_MARKER@l
- stw r9,_MSR(r1)
- li r2, INTERRUPT_SYSCALL
- stw r12,8(r1)
- stw r2,_TRAP(r1)
- SAVE_GPR(0, r1)
- SAVE_GPRS(3, 8, r1)
- addi r2,r10,-THREAD
- SAVE_NVGPRS(r1)
- kuep_lock
- /* Calling convention has r3 = regs, r4 = orig r0 */
- addi r3,r1,STACK_FRAME_OVERHEAD
- mr r4,r0
- bl system_call_exception
- ret_from_syscall:
- addi r4,r1,STACK_FRAME_OVERHEAD
- li r5,0
- bl syscall_exit_prepare
- #ifdef CONFIG_PPC_47x
- lis r4,icache_44x_need_flush@ha
- lwz r5,icache_44x_need_flush@l(r4)
- cmplwi cr0,r5,0
- bne- .L44x_icache_flush
- #endif /* CONFIG_PPC_47x */
- .L44x_icache_flush_return:
- kuep_unlock
- lwz r4,_LINK(r1)
- lwz r5,_CCR(r1)
- mtlr r4
- lwz r7,_NIP(r1)
- lwz r8,_MSR(r1)
- cmpwi r3,0
- REST_GPR(3, r1)
- syscall_exit_finish:
- mtspr SPRN_SRR0,r7
- mtspr SPRN_SRR1,r8
- bne 3f
- mtcr r5
- 1: REST_GPR(2, r1)
- REST_GPR(1, r1)
- rfi
- #ifdef CONFIG_40x
- b . /* Prevent prefetch past rfi */
- #endif
- 3: mtcr r5
- lwz r4,_CTR(r1)
- lwz r5,_XER(r1)
- REST_NVGPRS(r1)
- mtctr r4
- mtxer r5
- REST_GPR(0, r1)
- REST_GPRS(3, 12, r1)
- b 1b
- #ifdef CONFIG_44x
- .L44x_icache_flush:
- li r7,0
- iccci r0,r0
- stw r7,icache_44x_need_flush@l(r4)
- b .L44x_icache_flush_return
- #endif /* CONFIG_44x */
- .globl ret_from_fork
- ret_from_fork:
- REST_NVGPRS(r1)
- bl schedule_tail
- li r3,0
- b ret_from_syscall
- .globl ret_from_kernel_thread
- ret_from_kernel_thread:
- REST_NVGPRS(r1)
- bl schedule_tail
- mtctr r14
- mr r3,r15
- PPC440EP_ERR42
- bctrl
- li r3,0
- b ret_from_syscall
- /*
- * This routine switches between two different tasks. The process
- * state of one is saved on its kernel stack. Then the state
- * of the other is restored from its kernel stack. The memory
- * management hardware is updated to the second process's state.
- * Finally, we can return to the second process.
- * On entry, r3 points to the THREAD for the current task, r4
- * points to the THREAD for the new task.
- *
- * This routine is always called with interrupts disabled.
- *
- * Note: there are two ways to get to the "going out" portion
- * of this code; either by coming in via the entry (_switch)
- * or via "fork" which must set up an environment equivalent
- * to the "_switch" path. If you change this , you'll have to
- * change the fork code also.
- *
- * The code which creates the new task context is in 'copy_thread'
- * in arch/ppc/kernel/process.c
- */
- _GLOBAL(_switch)
- stwu r1,-INT_FRAME_SIZE(r1)
- mflr r0
- stw r0,INT_FRAME_SIZE+4(r1)
- /* r3-r12 are caller saved -- Cort */
- SAVE_NVGPRS(r1)
- stw r0,_NIP(r1) /* Return to switch caller */
- mfcr r10
- stw r10,_CCR(r1)
- stw r1,KSP(r3) /* Set old stack pointer */
- #ifdef CONFIG_SMP
- /* We need a sync somewhere here to make sure that if the
- * previous task gets rescheduled on another CPU, it sees all
- * stores it has performed on this one.
- */
- sync
- #endif /* CONFIG_SMP */
- tophys(r0,r4)
- mtspr SPRN_SPRG_THREAD,r0 /* Update current THREAD phys addr */
- lwz r1,KSP(r4) /* Load new stack pointer */
- /* save the old current 'last' for return value */
- mr r3,r2
- addi r2,r4,-THREAD /* Update current */
- lwz r0,_CCR(r1)
- mtcrf 0xFF,r0
- /* r3-r12 are destroyed -- Cort */
- REST_NVGPRS(r1)
- lwz r4,_NIP(r1) /* Return to _switch caller in new task */
- mtlr r4
- addi r1,r1,INT_FRAME_SIZE
- blr
- .globl fast_exception_return
- fast_exception_return:
- #if !(defined(CONFIG_4xx) || defined(CONFIG_BOOKE))
- andi. r10,r9,MSR_RI /* check for recoverable interrupt */
- beq 3f /* if not, we've got problems */
- #endif
- 2: lwz r10,_CCR(r11)
- REST_GPRS(1, 6, r11)
- mtcr r10
- lwz r10,_LINK(r11)
- mtlr r10
- /* Clear the exception marker on the stack to avoid confusing stacktrace */
- li r10, 0
- stw r10, 8(r11)
- REST_GPR(10, r11)
- #if defined(CONFIG_PPC_8xx) && defined(CONFIG_PERF_EVENTS)
- mtspr SPRN_NRI, r0
- #endif
- mtspr SPRN_SRR1,r9
- mtspr SPRN_SRR0,r12
- REST_GPR(9, r11)
- REST_GPR(12, r11)
- REST_GPR(11, r11)
- rfi
- #ifdef CONFIG_40x
- b . /* Prevent prefetch past rfi */
- #endif
- _ASM_NOKPROBE_SYMBOL(fast_exception_return)
- /* aargh, a nonrecoverable interrupt, panic */
- /* aargh, we don't know which trap this is */
- 3:
- li r10,-1
- stw r10,_TRAP(r11)
- prepare_transfer_to_handler
- bl unrecoverable_exception
- trap /* should not get here */
- .globl interrupt_return
- interrupt_return:
- lwz r4,_MSR(r1)
- addi r3,r1,STACK_FRAME_OVERHEAD
- andi. r0,r4,MSR_PR
- beq .Lkernel_interrupt_return
- bl interrupt_exit_user_prepare
- cmpwi r3,0
- kuep_unlock
- bne- .Lrestore_nvgprs
- .Lfast_user_interrupt_return:
- lwz r11,_NIP(r1)
- lwz r12,_MSR(r1)
- mtspr SPRN_SRR0,r11
- mtspr SPRN_SRR1,r12
- BEGIN_FTR_SECTION
- stwcx. r0,0,r1 /* to clear the reservation */
- FTR_SECTION_ELSE
- lwarx r0,0,r1
- ALT_FTR_SECTION_END_IFCLR(CPU_FTR_STCX_CHECKS_ADDRESS)
- lwz r3,_CCR(r1)
- lwz r4,_LINK(r1)
- lwz r5,_CTR(r1)
- lwz r6,_XER(r1)
- li r0,0
- /*
- * Leaving a stale exception marker on the stack can confuse
- * the reliable stack unwinder later on. Clear it.
- */
- stw r0,8(r1)
- REST_GPRS(7, 12, r1)
- mtcr r3
- mtlr r4
- mtctr r5
- mtspr SPRN_XER,r6
- REST_GPRS(2, 6, r1)
- REST_GPR(0, r1)
- REST_GPR(1, r1)
- rfi
- #ifdef CONFIG_40x
- b . /* Prevent prefetch past rfi */
- #endif
- .Lrestore_nvgprs:
- REST_NVGPRS(r1)
- b .Lfast_user_interrupt_return
- .Lkernel_interrupt_return:
- bl interrupt_exit_kernel_prepare
- .Lfast_kernel_interrupt_return:
- cmpwi cr1,r3,0
- lwz r11,_NIP(r1)
- lwz r12,_MSR(r1)
- mtspr SPRN_SRR0,r11
- mtspr SPRN_SRR1,r12
- BEGIN_FTR_SECTION
- stwcx. r0,0,r1 /* to clear the reservation */
- FTR_SECTION_ELSE
- lwarx r0,0,r1
- ALT_FTR_SECTION_END_IFCLR(CPU_FTR_STCX_CHECKS_ADDRESS)
- lwz r3,_LINK(r1)
- lwz r4,_CTR(r1)
- lwz r5,_XER(r1)
- lwz r6,_CCR(r1)
- li r0,0
- REST_GPRS(7, 12, r1)
- mtlr r3
- mtctr r4
- mtspr SPRN_XER,r5
- /*
- * Leaving a stale exception marker on the stack can confuse
- * the reliable stack unwinder later on. Clear it.
- */
- stw r0,8(r1)
- REST_GPRS(2, 5, r1)
- bne- cr1,1f /* emulate stack store */
- mtcr r6
- REST_GPR(6, r1)
- REST_GPR(0, r1)
- REST_GPR(1, r1)
- rfi
- #ifdef CONFIG_40x
- b . /* Prevent prefetch past rfi */
- #endif
- 1: /*
- * Emulate stack store with update. New r1 value was already calculated
- * and updated in our interrupt regs by emulate_loadstore, but we can't
- * store the previous value of r1 to the stack before re-loading our
- * registers from it, otherwise they could be clobbered. Use
- * SPRG Scratch0 as temporary storage to hold the store
- * data, as interrupts are disabled here so it won't be clobbered.
- */
- mtcr r6
- #ifdef CONFIG_BOOKE
- mtspr SPRN_SPRG_WSCRATCH0, r9
- #else
- mtspr SPRN_SPRG_SCRATCH0, r9
- #endif
- addi r9,r1,INT_FRAME_SIZE /* get original r1 */
- REST_GPR(6, r1)
- REST_GPR(0, r1)
- REST_GPR(1, r1)
- stw r9,0(r1) /* perform store component of stwu */
- #ifdef CONFIG_BOOKE
- mfspr r9, SPRN_SPRG_RSCRATCH0
- #else
- mfspr r9, SPRN_SPRG_SCRATCH0
- #endif
- rfi
- #ifdef CONFIG_40x
- b . /* Prevent prefetch past rfi */
- #endif
- _ASM_NOKPROBE_SYMBOL(interrupt_return)
- #if defined(CONFIG_4xx) || defined(CONFIG_BOOKE)
- /*
- * Returning from a critical interrupt in user mode doesn't need
- * to be any different from a normal exception. For a critical
- * interrupt in the kernel, we just return (without checking for
- * preemption) since the interrupt may have happened at some crucial
- * place (e.g. inside the TLB miss handler), and because we will be
- * running with r1 pointing into critical_stack, not the current
- * process's kernel stack (and therefore current_thread_info() will
- * give the wrong answer).
- * We have to restore various SPRs that may have been in use at the
- * time of the critical interrupt.
- *
- */
- #ifdef CONFIG_40x
- #define PPC_40x_TURN_OFF_MSR_DR \
- /* avoid any possible TLB misses here by turning off MSR.DR, we \
- * assume the instructions here are mapped by a pinned TLB entry */ \
- li r10,MSR_IR; \
- mtmsr r10; \
- isync; \
- tophys(r1, r1);
- #else
- #define PPC_40x_TURN_OFF_MSR_DR
- #endif
- #define RET_FROM_EXC_LEVEL(exc_lvl_srr0, exc_lvl_srr1, exc_lvl_rfi) \
- REST_NVGPRS(r1); \
- lwz r3,_MSR(r1); \
- andi. r3,r3,MSR_PR; \
- bne interrupt_return; \
- REST_GPR(0, r1); \
- REST_GPRS(2, 8, r1); \
- lwz r10,_XER(r1); \
- lwz r11,_CTR(r1); \
- mtspr SPRN_XER,r10; \
- mtctr r11; \
- stwcx. r0,0,r1; /* to clear the reservation */ \
- lwz r11,_LINK(r1); \
- mtlr r11; \
- lwz r10,_CCR(r1); \
- mtcrf 0xff,r10; \
- PPC_40x_TURN_OFF_MSR_DR; \
- lwz r9,_DEAR(r1); \
- lwz r10,_ESR(r1); \
- mtspr SPRN_DEAR,r9; \
- mtspr SPRN_ESR,r10; \
- lwz r11,_NIP(r1); \
- lwz r12,_MSR(r1); \
- mtspr exc_lvl_srr0,r11; \
- mtspr exc_lvl_srr1,r12; \
- REST_GPRS(9, 12, r1); \
- REST_GPR(1, r1); \
- exc_lvl_rfi; \
- b .; /* prevent prefetch past exc_lvl_rfi */
- #define RESTORE_xSRR(exc_lvl_srr0, exc_lvl_srr1) \
- lwz r9,_##exc_lvl_srr0(r1); \
- lwz r10,_##exc_lvl_srr1(r1); \
- mtspr SPRN_##exc_lvl_srr0,r9; \
- mtspr SPRN_##exc_lvl_srr1,r10;
- #if defined(CONFIG_PPC_E500)
- #ifdef CONFIG_PHYS_64BIT
- #define RESTORE_MAS7 \
- lwz r11,MAS7(r1); \
- mtspr SPRN_MAS7,r11;
- #else
- #define RESTORE_MAS7
- #endif /* CONFIG_PHYS_64BIT */
- #define RESTORE_MMU_REGS \
- lwz r9,MAS0(r1); \
- lwz r10,MAS1(r1); \
- lwz r11,MAS2(r1); \
- mtspr SPRN_MAS0,r9; \
- lwz r9,MAS3(r1); \
- mtspr SPRN_MAS1,r10; \
- lwz r10,MAS6(r1); \
- mtspr SPRN_MAS2,r11; \
- mtspr SPRN_MAS3,r9; \
- mtspr SPRN_MAS6,r10; \
- RESTORE_MAS7;
- #elif defined(CONFIG_44x)
- #define RESTORE_MMU_REGS \
- lwz r9,MMUCR(r1); \
- mtspr SPRN_MMUCR,r9;
- #else
- #define RESTORE_MMU_REGS
- #endif
- #ifdef CONFIG_40x
- .globl ret_from_crit_exc
- ret_from_crit_exc:
- lis r9,crit_srr0@ha;
- lwz r9,crit_srr0@l(r9);
- lis r10,crit_srr1@ha;
- lwz r10,crit_srr1@l(r10);
- mtspr SPRN_SRR0,r9;
- mtspr SPRN_SRR1,r10;
- RET_FROM_EXC_LEVEL(SPRN_CSRR0, SPRN_CSRR1, PPC_RFCI)
- _ASM_NOKPROBE_SYMBOL(ret_from_crit_exc)
- #endif /* CONFIG_40x */
- #ifdef CONFIG_BOOKE
- .globl ret_from_crit_exc
- ret_from_crit_exc:
- RESTORE_xSRR(SRR0,SRR1);
- RESTORE_MMU_REGS;
- RET_FROM_EXC_LEVEL(SPRN_CSRR0, SPRN_CSRR1, PPC_RFCI)
- _ASM_NOKPROBE_SYMBOL(ret_from_crit_exc)
- .globl ret_from_debug_exc
- ret_from_debug_exc:
- RESTORE_xSRR(SRR0,SRR1);
- RESTORE_xSRR(CSRR0,CSRR1);
- RESTORE_MMU_REGS;
- RET_FROM_EXC_LEVEL(SPRN_DSRR0, SPRN_DSRR1, PPC_RFDI)
- _ASM_NOKPROBE_SYMBOL(ret_from_debug_exc)
- .globl ret_from_mcheck_exc
- ret_from_mcheck_exc:
- RESTORE_xSRR(SRR0,SRR1);
- RESTORE_xSRR(CSRR0,CSRR1);
- RESTORE_xSRR(DSRR0,DSRR1);
- RESTORE_MMU_REGS;
- RET_FROM_EXC_LEVEL(SPRN_MCSRR0, SPRN_MCSRR1, PPC_RFMCI)
- _ASM_NOKPROBE_SYMBOL(ret_from_mcheck_exc)
- #endif /* CONFIG_BOOKE */
- #endif /* !(CONFIG_4xx || CONFIG_BOOKE) */
|