123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899 |
- /* SPDX-License-Identifier: GPL-2.0-or-later */
- /*
- * vDSO provided cache flush routines
- *
- * Copyright (C) 2004 Benjamin Herrenschmuidt ([email protected]),
- * IBM Corp.
- */
- #include <asm/processor.h>
- #include <asm/ppc_asm.h>
- #include <asm/vdso.h>
- #include <asm/vdso_datapage.h>
- #include <asm/asm-offsets.h>
- #include <asm/cache.h>
- .text
- /*
- * Default "generic" version of __kernel_sync_dicache.
- *
- * void __kernel_sync_dicache(unsigned long start, unsigned long end)
- *
- * Flushes the data cache & invalidate the instruction cache for the
- * provided range [start, end[
- */
- V_FUNCTION_BEGIN(__kernel_sync_dicache)
- .cfi_startproc
- BEGIN_FTR_SECTION
- b 3f
- END_FTR_SECTION_IFSET(CPU_FTR_COHERENT_ICACHE)
- #ifdef CONFIG_PPC64
- mflr r12
- .cfi_register lr,r12
- get_datapage r10
- mtlr r12
- .cfi_restore lr
- #endif
- #ifdef CONFIG_PPC64
- lwz r7,CFG_DCACHE_BLOCKSZ(r10)
- addi r5,r7,-1
- #else
- li r5, L1_CACHE_BYTES - 1
- #endif
- andc r6,r3,r5 /* round low to line bdy */
- subf r8,r6,r4 /* compute length */
- add r8,r8,r5 /* ensure we get enough */
- #ifdef CONFIG_PPC64
- lwz r9,CFG_DCACHE_LOGBLOCKSZ(r10)
- PPC_SRL. r8,r8,r9 /* compute line count */
- #else
- srwi. r8, r8, L1_CACHE_SHIFT
- mr r7, r6
- #endif
- crclr cr0*4+so
- beqlr /* nothing to do? */
- mtctr r8
- 1: dcbst 0,r6
- #ifdef CONFIG_PPC64
- add r6,r6,r7
- #else
- addi r6, r6, L1_CACHE_BYTES
- #endif
- bdnz 1b
- sync
- /* Now invalidate the instruction cache */
- #ifdef CONFIG_PPC64
- lwz r7,CFG_ICACHE_BLOCKSZ(r10)
- addi r5,r7,-1
- andc r6,r3,r5 /* round low to line bdy */
- subf r8,r6,r4 /* compute length */
- add r8,r8,r5
- lwz r9,CFG_ICACHE_LOGBLOCKSZ(r10)
- PPC_SRL. r8,r8,r9 /* compute line count */
- crclr cr0*4+so
- beqlr /* nothing to do? */
- #endif
- mtctr r8
- #ifdef CONFIG_PPC64
- 2: icbi 0,r6
- add r6,r6,r7
- #else
- 2: icbi 0, r7
- addi r7, r7, L1_CACHE_BYTES
- #endif
- bdnz 2b
- isync
- li r3,0
- blr
- 3:
- crclr cr0*4+so
- sync
- icbi 0,r1
- isync
- li r3,0
- blr
- .cfi_endproc
- V_FUNCTION_END(__kernel_sync_dicache)
|