powerpc/mm: Split low level tlb invalidate for nohash processors
Currently, the various forms of low level TLB invalidations are all implemented in misc_32.S for 32-bit processors, in a fairly scary mess of #ifdef's and with interesting duplication such as a whole bunch of code for FSL _tlbie and _tlbia which are no longer used. This moves things around such that _tlbie is now defined in hash_low_32.S and is only used by the 32-bit hash code, and all nohash CPUs use the various _tlbil_* forms that are now moved to a new file, tlb_nohash_low.S. I moved all the definitions for that stuff out of include/asm/tlbflush.h as they are really internal mm stuff, into mm/mmu_decl.h The code should have no functional changes. I kept some variants inline for trivial forms on things like 40x and 8xx. Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org> Acked-by: Kumar Gala <galak@kernel.crashing.org> Signed-off-by: Paul Mackerras <paulus@samba.org>
This commit is contained in:

zatwierdzone przez
Paul Mackerras

rodzic
f048aace29
commit
2a4aca1144
@@ -272,239 +272,6 @@ _GLOBAL(real_writeb)
|
||||
|
||||
#endif /* CONFIG_40x */
|
||||
|
||||
/*
|
||||
* Flush MMU TLB
|
||||
*/
|
||||
#ifndef CONFIG_FSL_BOOKE
|
||||
_GLOBAL(_tlbil_all)
|
||||
_GLOBAL(_tlbil_pid)
|
||||
#endif
|
||||
_GLOBAL(_tlbia)
|
||||
#if defined(CONFIG_40x)
|
||||
sync /* Flush to memory before changing mapping */
|
||||
tlbia
|
||||
isync /* Flush shadow TLB */
|
||||
#elif defined(CONFIG_44x)
|
||||
li r3,0
|
||||
sync
|
||||
|
||||
/* Load high watermark */
|
||||
lis r4,tlb_44x_hwater@ha
|
||||
lwz r5,tlb_44x_hwater@l(r4)
|
||||
|
||||
1: tlbwe r3,r3,PPC44x_TLB_PAGEID
|
||||
addi r3,r3,1
|
||||
cmpw 0,r3,r5
|
||||
ble 1b
|
||||
|
||||
isync
|
||||
#elif defined(CONFIG_FSL_BOOKE)
|
||||
/* Invalidate all entries in TLB0 */
|
||||
li r3, 0x04
|
||||
tlbivax 0,3
|
||||
/* Invalidate all entries in TLB1 */
|
||||
li r3, 0x0c
|
||||
tlbivax 0,3
|
||||
msync
|
||||
#ifdef CONFIG_SMP
|
||||
tlbsync
|
||||
#endif /* CONFIG_SMP */
|
||||
#else /* !(CONFIG_40x || CONFIG_44x || CONFIG_FSL_BOOKE) */
|
||||
#if defined(CONFIG_SMP)
|
||||
rlwinm r8,r1,0,0,(31-THREAD_SHIFT)
|
||||
lwz r8,TI_CPU(r8)
|
||||
oris r8,r8,10
|
||||
mfmsr r10
|
||||
SYNC
|
||||
rlwinm r0,r10,0,17,15 /* clear bit 16 (MSR_EE) */
|
||||
rlwinm r0,r0,0,28,26 /* clear DR */
|
||||
mtmsr r0
|
||||
SYNC_601
|
||||
isync
|
||||
lis r9,mmu_hash_lock@h
|
||||
ori r9,r9,mmu_hash_lock@l
|
||||
tophys(r9,r9)
|
||||
10: lwarx r7,0,r9
|
||||
cmpwi 0,r7,0
|
||||
bne- 10b
|
||||
stwcx. r8,0,r9
|
||||
bne- 10b
|
||||
sync
|
||||
tlbia
|
||||
sync
|
||||
TLBSYNC
|
||||
li r0,0
|
||||
stw r0,0(r9) /* clear mmu_hash_lock */
|
||||
mtmsr r10
|
||||
SYNC_601
|
||||
isync
|
||||
#else /* CONFIG_SMP */
|
||||
sync
|
||||
tlbia
|
||||
sync
|
||||
#endif /* CONFIG_SMP */
|
||||
#endif /* ! defined(CONFIG_40x) */
|
||||
blr
|
||||
|
||||
/*
|
||||
* Flush MMU TLB for a particular address
|
||||
*/
|
||||
#ifndef CONFIG_FSL_BOOKE
|
||||
_GLOBAL(_tlbil_va)
|
||||
#endif
|
||||
_GLOBAL(_tlbie)
|
||||
#if defined(CONFIG_40x)
|
||||
/* We run the search with interrupts disabled because we have to change
|
||||
* the PID and I don't want to preempt when that happens.
|
||||
*/
|
||||
mfmsr r5
|
||||
mfspr r6,SPRN_PID
|
||||
wrteei 0
|
||||
mtspr SPRN_PID,r4
|
||||
tlbsx. r3, 0, r3
|
||||
mtspr SPRN_PID,r6
|
||||
wrtee r5
|
||||
bne 10f
|
||||
sync
|
||||
/* There are only 64 TLB entries, so r3 < 64, which means bit 25 is clear.
|
||||
* Since 25 is the V bit in the TLB_TAG, loading this value will invalidate
|
||||
* the TLB entry. */
|
||||
tlbwe r3, r3, TLB_TAG
|
||||
isync
|
||||
10:
|
||||
|
||||
#elif defined(CONFIG_44x)
|
||||
mfspr r5,SPRN_MMUCR
|
||||
rlwimi r5,r4,0,24,31 /* Set TID */
|
||||
|
||||
/* We have to run the search with interrupts disabled, even critical
|
||||
* and debug interrupts (in fact the only critical exceptions we have
|
||||
* are debug and machine check). Otherwise an interrupt which causes
|
||||
* a TLB miss can clobber the MMUCR between the mtspr and the tlbsx. */
|
||||
mfmsr r4
|
||||
lis r6,(MSR_EE|MSR_CE|MSR_ME|MSR_DE)@ha
|
||||
addi r6,r6,(MSR_EE|MSR_CE|MSR_ME|MSR_DE)@l
|
||||
andc r6,r4,r6
|
||||
mtmsr r6
|
||||
mtspr SPRN_MMUCR,r5
|
||||
tlbsx. r3, 0, r3
|
||||
mtmsr r4
|
||||
bne 10f
|
||||
sync
|
||||
/* There are only 64 TLB entries, so r3 < 64,
|
||||
* which means bit 22, is clear. Since 22 is
|
||||
* the V bit in the TLB_PAGEID, loading this
|
||||
* value will invalidate the TLB entry.
|
||||
*/
|
||||
tlbwe r3, r3, PPC44x_TLB_PAGEID
|
||||
isync
|
||||
10:
|
||||
#elif defined(CONFIG_FSL_BOOKE)
|
||||
rlwinm r4, r3, 0, 0, 19
|
||||
ori r5, r4, 0x08 /* TLBSEL = 1 */
|
||||
tlbivax 0, r4
|
||||
tlbivax 0, r5
|
||||
msync
|
||||
#if defined(CONFIG_SMP)
|
||||
tlbsync
|
||||
#endif /* CONFIG_SMP */
|
||||
#else /* !(CONFIG_40x || CONFIG_44x || CONFIG_FSL_BOOKE) */
|
||||
#if defined(CONFIG_SMP)
|
||||
rlwinm r8,r1,0,0,(31-THREAD_SHIFT)
|
||||
lwz r8,TI_CPU(r8)
|
||||
oris r8,r8,11
|
||||
mfmsr r10
|
||||
SYNC
|
||||
rlwinm r0,r10,0,17,15 /* clear bit 16 (MSR_EE) */
|
||||
rlwinm r0,r0,0,28,26 /* clear DR */
|
||||
mtmsr r0
|
||||
SYNC_601
|
||||
isync
|
||||
lis r9,mmu_hash_lock@h
|
||||
ori r9,r9,mmu_hash_lock@l
|
||||
tophys(r9,r9)
|
||||
10: lwarx r7,0,r9
|
||||
cmpwi 0,r7,0
|
||||
bne- 10b
|
||||
stwcx. r8,0,r9
|
||||
bne- 10b
|
||||
eieio
|
||||
tlbie r3
|
||||
sync
|
||||
TLBSYNC
|
||||
li r0,0
|
||||
stw r0,0(r9) /* clear mmu_hash_lock */
|
||||
mtmsr r10
|
||||
SYNC_601
|
||||
isync
|
||||
#else /* CONFIG_SMP */
|
||||
tlbie r3
|
||||
sync
|
||||
#endif /* CONFIG_SMP */
|
||||
#endif /* ! CONFIG_40x */
|
||||
blr
|
||||
|
||||
#if defined(CONFIG_FSL_BOOKE)
|
||||
/*
|
||||
* Flush MMU TLB, but only on the local processor (no broadcast)
|
||||
*/
|
||||
_GLOBAL(_tlbil_all)
|
||||
#define MMUCSR0_TLBFI (MMUCSR0_TLB0FI | MMUCSR0_TLB1FI | \
|
||||
MMUCSR0_TLB2FI | MMUCSR0_TLB3FI)
|
||||
li r3,(MMUCSR0_TLBFI)@l
|
||||
mtspr SPRN_MMUCSR0, r3
|
||||
1:
|
||||
mfspr r3,SPRN_MMUCSR0
|
||||
andi. r3,r3,MMUCSR0_TLBFI@l
|
||||
bne 1b
|
||||
blr
|
||||
|
||||
/*
|
||||
* Flush MMU TLB for a particular process id, but only on the local processor
|
||||
* (no broadcast)
|
||||
*/
|
||||
_GLOBAL(_tlbil_pid)
|
||||
/* we currently do an invalidate all since we don't have per pid invalidate */
|
||||
li r3,(MMUCSR0_TLBFI)@l
|
||||
mtspr SPRN_MMUCSR0, r3
|
||||
1:
|
||||
mfspr r3,SPRN_MMUCSR0
|
||||
andi. r3,r3,MMUCSR0_TLBFI@l
|
||||
bne 1b
|
||||
msync
|
||||
isync
|
||||
blr
|
||||
|
||||
/*
|
||||
* Flush MMU TLB for a particular address, but only on the local processor
|
||||
* (no broadcast)
|
||||
*/
|
||||
_GLOBAL(_tlbil_va)
|
||||
mfmsr r10
|
||||
wrteei 0
|
||||
slwi r4,r4,16
|
||||
mtspr SPRN_MAS6,r4 /* assume AS=0 for now */
|
||||
tlbsx 0,r3
|
||||
mfspr r4,SPRN_MAS1 /* check valid */
|
||||
andis. r3,r4,MAS1_VALID@h
|
||||
beq 1f
|
||||
rlwinm r4,r4,0,1,31
|
||||
mtspr SPRN_MAS1,r4
|
||||
tlbwe
|
||||
msync
|
||||
isync
|
||||
1: wrtee r10
|
||||
blr
|
||||
#endif /* CONFIG_FSL_BOOKE */
|
||||
|
||||
/*
|
||||
* Nobody implements this yet
|
||||
*/
|
||||
_GLOBAL(_tlbivax_bcast)
|
||||
1: trap
|
||||
EMIT_BUG_ENTRY 1b,__FILE__,__LINE__,0;
|
||||
blr
|
||||
|
||||
|
||||
/*
|
||||
* Flush instruction cache.
|
||||
|
Reference in New Issue
Block a user