123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564 |
- /* SPDX-License-Identifier: GPL-2.0-or-later */
- /*
- * Copyright (C) 2002 Paul Mackerras, IBM Corp.
- */
- #include <asm/processor.h>
- #include <asm/ppc_asm.h>
- #include <asm/export.h>
- #include <asm/asm-compat.h>
- #include <asm/feature-fixups.h>
- #ifndef SELFTEST_CASE
- /* 0 == most CPUs, 1 == POWER6, 2 == Cell */
- #define SELFTEST_CASE 0
- #endif
- #ifdef __BIG_ENDIAN__
- #define sLd sld /* Shift towards low-numbered address. */
- #define sHd srd /* Shift towards high-numbered address. */
- #else
- #define sLd srd /* Shift towards low-numbered address. */
- #define sHd sld /* Shift towards high-numbered address. */
- #endif
- /*
- * These macros are used to generate exception table entries.
- * The exception handlers below use the original arguments
- * (stored on the stack) and the point where we're up to in
- * the destination buffer, i.e. the address of the first
- * unmodified byte. Generally r3 points into the destination
- * buffer, but the first unmodified byte is at a variable
- * offset from r3. In the code below, the symbol r3_offset
- * is set to indicate the current offset at each point in
- * the code. This offset is then used as a negative offset
- * from the exception handler code, and those instructions
- * before the exception handlers are addi instructions that
- * adjust r3 to point to the correct place.
- */
- .macro lex /* exception handler for load */
- 100: EX_TABLE(100b, .Lld_exc - r3_offset)
- .endm
- .macro stex /* exception handler for store */
- 100: EX_TABLE(100b, .Lst_exc - r3_offset)
- .endm
- .align 7
- _GLOBAL_TOC(__copy_tofrom_user)
- #ifdef CONFIG_PPC_BOOK3S_64
- BEGIN_FTR_SECTION
- nop
- FTR_SECTION_ELSE
- b __copy_tofrom_user_power7
- ALT_FTR_SECTION_END_IFCLR(CPU_FTR_VMX_COPY)
- #endif
- _GLOBAL(__copy_tofrom_user_base)
- /* first check for a 4kB copy on a 4kB boundary */
- cmpldi cr1,r5,16
- cmpdi cr6,r5,4096
- or r0,r3,r4
- neg r6,r3 /* LS 3 bits = # bytes to 8-byte dest bdry */
- andi. r0,r0,4095
- std r3,-24(r1)
- crand cr0*4+2,cr0*4+2,cr6*4+2
- std r4,-16(r1)
- std r5,-8(r1)
- dcbt 0,r4
- beq .Lcopy_page_4K
- andi. r6,r6,7
- PPC_MTOCRF(0x01,r5)
- blt cr1,.Lshort_copy
- /* Below we want to nop out the bne if we're on a CPU that has the
- * CPU_FTR_UNALIGNED_LD_STD bit set and the CPU_FTR_CP_USE_DCBTZ bit
- * cleared.
- * At the time of writing the only CPU that has this combination of bits
- * set is Power6.
- */
- test_feature = (SELFTEST_CASE == 1)
- BEGIN_FTR_SECTION
- nop
- FTR_SECTION_ELSE
- bne .Ldst_unaligned
- ALT_FTR_SECTION_END(CPU_FTR_UNALIGNED_LD_STD | CPU_FTR_CP_USE_DCBTZ, \
- CPU_FTR_UNALIGNED_LD_STD)
- .Ldst_aligned:
- addi r3,r3,-16
- r3_offset = 16
- test_feature = (SELFTEST_CASE == 0)
- BEGIN_FTR_SECTION
- andi. r0,r4,7
- bne .Lsrc_unaligned
- END_FTR_SECTION_IFCLR(CPU_FTR_UNALIGNED_LD_STD)
- blt cr1,.Ldo_tail /* if < 16 bytes to copy */
- srdi r0,r5,5
- cmpdi cr1,r0,0
- lex; ld r7,0(r4)
- lex; ld r6,8(r4)
- addi r4,r4,16
- mtctr r0
- andi. r0,r5,0x10
- beq 22f
- addi r3,r3,16
- r3_offset = 0
- addi r4,r4,-16
- mr r9,r7
- mr r8,r6
- beq cr1,72f
- 21:
- lex; ld r7,16(r4)
- lex; ld r6,24(r4)
- addi r4,r4,32
- stex; std r9,0(r3)
- r3_offset = 8
- stex; std r8,8(r3)
- r3_offset = 16
- 22:
- lex; ld r9,0(r4)
- lex; ld r8,8(r4)
- stex; std r7,16(r3)
- r3_offset = 24
- stex; std r6,24(r3)
- addi r3,r3,32
- r3_offset = 0
- bdnz 21b
- 72:
- stex; std r9,0(r3)
- r3_offset = 8
- stex; std r8,8(r3)
- r3_offset = 16
- andi. r5,r5,0xf
- beq+ 3f
- addi r4,r4,16
- .Ldo_tail:
- addi r3,r3,16
- r3_offset = 0
- bf cr7*4+0,246f
- lex; ld r9,0(r4)
- addi r4,r4,8
- stex; std r9,0(r3)
- addi r3,r3,8
- 246: bf cr7*4+1,1f
- lex; lwz r9,0(r4)
- addi r4,r4,4
- stex; stw r9,0(r3)
- addi r3,r3,4
- 1: bf cr7*4+2,2f
- lex; lhz r9,0(r4)
- addi r4,r4,2
- stex; sth r9,0(r3)
- addi r3,r3,2
- 2: bf cr7*4+3,3f
- lex; lbz r9,0(r4)
- stex; stb r9,0(r3)
- 3: li r3,0
- blr
- .Lsrc_unaligned:
- r3_offset = 16
- srdi r6,r5,3
- addi r5,r5,-16
- subf r4,r0,r4
- srdi r7,r5,4
- sldi r10,r0,3
- cmpldi cr6,r6,3
- andi. r5,r5,7
- mtctr r7
- subfic r11,r10,64
- add r5,r5,r0
- bt cr7*4+0,28f
- lex; ld r9,0(r4) /* 3+2n loads, 2+2n stores */
- lex; ld r0,8(r4)
- sLd r6,r9,r10
- lex; ldu r9,16(r4)
- sHd r7,r0,r11
- sLd r8,r0,r10
- or r7,r7,r6
- blt cr6,79f
- lex; ld r0,8(r4)
- b 2f
- 28:
- lex; ld r0,0(r4) /* 4+2n loads, 3+2n stores */
- lex; ldu r9,8(r4)
- sLd r8,r0,r10
- addi r3,r3,-8
- r3_offset = 24
- blt cr6,5f
- lex; ld r0,8(r4)
- sHd r12,r9,r11
- sLd r6,r9,r10
- lex; ldu r9,16(r4)
- or r12,r8,r12
- sHd r7,r0,r11
- sLd r8,r0,r10
- addi r3,r3,16
- r3_offset = 8
- beq cr6,78f
- 1: or r7,r7,r6
- lex; ld r0,8(r4)
- stex; std r12,8(r3)
- r3_offset = 16
- 2: sHd r12,r9,r11
- sLd r6,r9,r10
- lex; ldu r9,16(r4)
- or r12,r8,r12
- stex; stdu r7,16(r3)
- r3_offset = 8
- sHd r7,r0,r11
- sLd r8,r0,r10
- bdnz 1b
- 78:
- stex; std r12,8(r3)
- r3_offset = 16
- or r7,r7,r6
- 79:
- stex; std r7,16(r3)
- r3_offset = 24
- 5: sHd r12,r9,r11
- or r12,r8,r12
- stex; std r12,24(r3)
- r3_offset = 32
- bne 6f
- li r3,0
- blr
- 6: cmpwi cr1,r5,8
- addi r3,r3,32
- r3_offset = 0
- sLd r9,r9,r10
- ble cr1,7f
- lex; ld r0,8(r4)
- sHd r7,r0,r11
- or r9,r7,r9
- 7:
- bf cr7*4+1,1f
- #ifdef __BIG_ENDIAN__
- rotldi r9,r9,32
- #endif
- stex; stw r9,0(r3)
- #ifdef __LITTLE_ENDIAN__
- rotrdi r9,r9,32
- #endif
- addi r3,r3,4
- 1: bf cr7*4+2,2f
- #ifdef __BIG_ENDIAN__
- rotldi r9,r9,16
- #endif
- stex; sth r9,0(r3)
- #ifdef __LITTLE_ENDIAN__
- rotrdi r9,r9,16
- #endif
- addi r3,r3,2
- 2: bf cr7*4+3,3f
- #ifdef __BIG_ENDIAN__
- rotldi r9,r9,8
- #endif
- stex; stb r9,0(r3)
- #ifdef __LITTLE_ENDIAN__
- rotrdi r9,r9,8
- #endif
- 3: li r3,0
- blr
- .Ldst_unaligned:
- r3_offset = 0
- PPC_MTOCRF(0x01,r6) /* put #bytes to 8B bdry into cr7 */
- subf r5,r6,r5
- li r7,0
- cmpldi cr1,r5,16
- bf cr7*4+3,1f
- 100: EX_TABLE(100b, .Lld_exc_r7)
- lbz r0,0(r4)
- 100: EX_TABLE(100b, .Lst_exc_r7)
- stb r0,0(r3)
- addi r7,r7,1
- 1: bf cr7*4+2,2f
- 100: EX_TABLE(100b, .Lld_exc_r7)
- lhzx r0,r7,r4
- 100: EX_TABLE(100b, .Lst_exc_r7)
- sthx r0,r7,r3
- addi r7,r7,2
- 2: bf cr7*4+1,3f
- 100: EX_TABLE(100b, .Lld_exc_r7)
- lwzx r0,r7,r4
- 100: EX_TABLE(100b, .Lst_exc_r7)
- stwx r0,r7,r3
- 3: PPC_MTOCRF(0x01,r5)
- add r4,r6,r4
- add r3,r6,r3
- b .Ldst_aligned
- .Lshort_copy:
- r3_offset = 0
- bf cr7*4+0,1f
- lex; lwz r0,0(r4)
- lex; lwz r9,4(r4)
- addi r4,r4,8
- stex; stw r0,0(r3)
- stex; stw r9,4(r3)
- addi r3,r3,8
- 1: bf cr7*4+1,2f
- lex; lwz r0,0(r4)
- addi r4,r4,4
- stex; stw r0,0(r3)
- addi r3,r3,4
- 2: bf cr7*4+2,3f
- lex; lhz r0,0(r4)
- addi r4,r4,2
- stex; sth r0,0(r3)
- addi r3,r3,2
- 3: bf cr7*4+3,4f
- lex; lbz r0,0(r4)
- stex; stb r0,0(r3)
- 4: li r3,0
- blr
- /*
- * exception handlers follow
- * we have to return the number of bytes not copied
- * for an exception on a load, we set the rest of the destination to 0
- * Note that the number of bytes of instructions for adjusting r3 needs
- * to equal the amount of the adjustment, due to the trick of using
- * .Lld_exc - r3_offset as the handler address.
- */
- .Lld_exc_r7:
- add r3,r3,r7
- b .Lld_exc
- /* adjust by 24 */
- addi r3,r3,8
- nop
- /* adjust by 16 */
- addi r3,r3,8
- nop
- /* adjust by 8 */
- addi r3,r3,8
- nop
- /*
- * Here we have had a fault on a load and r3 points to the first
- * unmodified byte of the destination. We use the original arguments
- * and r3 to work out how much wasn't copied. Since we load some
- * distance ahead of the stores, we continue copying byte-by-byte until
- * we hit the load fault again in order to copy as much as possible.
- */
- .Lld_exc:
- ld r6,-24(r1)
- ld r4,-16(r1)
- ld r5,-8(r1)
- subf r6,r6,r3
- add r4,r4,r6
- subf r5,r6,r5 /* #bytes left to go */
- /*
- * first see if we can copy any more bytes before hitting another exception
- */
- mtctr r5
- r3_offset = 0
- 100: EX_TABLE(100b, .Ldone)
- 43: lbz r0,0(r4)
- addi r4,r4,1
- stex; stb r0,0(r3)
- addi r3,r3,1
- bdnz 43b
- li r3,0 /* huh? all copied successfully this time? */
- blr
- /*
- * here we have trapped again, amount remaining is in ctr.
- */
- .Ldone:
- mfctr r3
- blr
- /*
- * exception handlers for stores: we need to work out how many bytes
- * weren't copied, and we may need to copy some more.
- * Note that the number of bytes of instructions for adjusting r3 needs
- * to equal the amount of the adjustment, due to the trick of using
- * .Lst_exc - r3_offset as the handler address.
- */
- .Lst_exc_r7:
- add r3,r3,r7
- b .Lst_exc
- /* adjust by 24 */
- addi r3,r3,8
- nop
- /* adjust by 16 */
- addi r3,r3,8
- nop
- /* adjust by 8 */
- addi r3,r3,4
- /* adjust by 4 */
- addi r3,r3,4
- .Lst_exc:
- ld r6,-24(r1) /* original destination pointer */
- ld r4,-16(r1) /* original source pointer */
- ld r5,-8(r1) /* original number of bytes */
- add r7,r6,r5
- /*
- * If the destination pointer isn't 8-byte aligned,
- * we may have got the exception as a result of a
- * store that overlapped a page boundary, so we may be
- * able to copy a few more bytes.
- */
- 17: andi. r0,r3,7
- beq 19f
- subf r8,r6,r3 /* #bytes copied */
- 100: EX_TABLE(100b,19f)
- lbzx r0,r8,r4
- 100: EX_TABLE(100b,19f)
- stb r0,0(r3)
- addi r3,r3,1
- cmpld r3,r7
- blt 17b
- 19: subf r3,r3,r7 /* #bytes not copied in r3 */
- blr
- /*
- * Routine to copy a whole page of data, optimized for POWER4.
- * On POWER4 it is more than 50% faster than the simple loop
- * above (following the .Ldst_aligned label).
- */
- .macro exc
- 100: EX_TABLE(100b, .Labort)
- .endm
- .Lcopy_page_4K:
- std r31,-32(1)
- std r30,-40(1)
- std r29,-48(1)
- std r28,-56(1)
- std r27,-64(1)
- std r26,-72(1)
- std r25,-80(1)
- std r24,-88(1)
- std r23,-96(1)
- std r22,-104(1)
- std r21,-112(1)
- std r20,-120(1)
- li r5,4096/32 - 1
- addi r3,r3,-8
- li r0,5
- 0: addi r5,r5,-24
- mtctr r0
- exc; ld r22,640(4)
- exc; ld r21,512(4)
- exc; ld r20,384(4)
- exc; ld r11,256(4)
- exc; ld r9,128(4)
- exc; ld r7,0(4)
- exc; ld r25,648(4)
- exc; ld r24,520(4)
- exc; ld r23,392(4)
- exc; ld r10,264(4)
- exc; ld r8,136(4)
- exc; ldu r6,8(4)
- cmpwi r5,24
- 1:
- exc; std r22,648(3)
- exc; std r21,520(3)
- exc; std r20,392(3)
- exc; std r11,264(3)
- exc; std r9,136(3)
- exc; std r7,8(3)
- exc; ld r28,648(4)
- exc; ld r27,520(4)
- exc; ld r26,392(4)
- exc; ld r31,264(4)
- exc; ld r30,136(4)
- exc; ld r29,8(4)
- exc; std r25,656(3)
- exc; std r24,528(3)
- exc; std r23,400(3)
- exc; std r10,272(3)
- exc; std r8,144(3)
- exc; std r6,16(3)
- exc; ld r22,656(4)
- exc; ld r21,528(4)
- exc; ld r20,400(4)
- exc; ld r11,272(4)
- exc; ld r9,144(4)
- exc; ld r7,16(4)
- exc; std r28,664(3)
- exc; std r27,536(3)
- exc; std r26,408(3)
- exc; std r31,280(3)
- exc; std r30,152(3)
- exc; stdu r29,24(3)
- exc; ld r25,664(4)
- exc; ld r24,536(4)
- exc; ld r23,408(4)
- exc; ld r10,280(4)
- exc; ld r8,152(4)
- exc; ldu r6,24(4)
- bdnz 1b
- exc; std r22,648(3)
- exc; std r21,520(3)
- exc; std r20,392(3)
- exc; std r11,264(3)
- exc; std r9,136(3)
- exc; std r7,8(3)
- addi r4,r4,640
- addi r3,r3,648
- bge 0b
- mtctr r5
- exc; ld r7,0(4)
- exc; ld r8,8(4)
- exc; ldu r9,16(4)
- 3:
- exc; ld r10,8(4)
- exc; std r7,8(3)
- exc; ld r7,16(4)
- exc; std r8,16(3)
- exc; ld r8,24(4)
- exc; std r9,24(3)
- exc; ldu r9,32(4)
- exc; stdu r10,32(3)
- bdnz 3b
- 4:
- exc; ld r10,8(4)
- exc; std r7,8(3)
- exc; std r8,16(3)
- exc; std r9,24(3)
- exc; std r10,32(3)
- 9: ld r20,-120(1)
- ld r21,-112(1)
- ld r22,-104(1)
- ld r23,-96(1)
- ld r24,-88(1)
- ld r25,-80(1)
- ld r26,-72(1)
- ld r27,-64(1)
- ld r28,-56(1)
- ld r29,-48(1)
- ld r30,-40(1)
- ld r31,-32(1)
- li r3,0
- blr
- /*
- * on an exception, reset to the beginning and jump back into the
- * standard __copy_tofrom_user
- */
- .Labort:
- ld r20,-120(1)
- ld r21,-112(1)
- ld r22,-104(1)
- ld r23,-96(1)
- ld r24,-88(1)
- ld r25,-80(1)
- ld r26,-72(1)
- ld r27,-64(1)
- ld r28,-56(1)
- ld r29,-48(1)
- ld r30,-40(1)
- ld r31,-32(1)
- ld r3,-24(r1)
- ld r4,-16(r1)
- li r5,4096
- b .Ldst_aligned
- EXPORT_SYMBOL(__copy_tofrom_user)
|