memcpy_64.S 4.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230
  1. /* SPDX-License-Identifier: GPL-2.0-or-later */
  2. /*
  3. * Copyright (C) 2002 Paul Mackerras, IBM Corp.
  4. */
  5. #include <asm/processor.h>
  6. #include <asm/ppc_asm.h>
  7. #include <asm/export.h>
  8. #include <asm/asm-compat.h>
  9. #include <asm/feature-fixups.h>
  10. #include <asm/kasan.h>
  11. #ifndef SELFTEST_CASE
  12. /* For big-endian, 0 == most CPUs, 1 == POWER6, 2 == Cell */
  13. #define SELFTEST_CASE 0
  14. #endif
  15. .align 7
  16. _GLOBAL_TOC_KASAN(memcpy)
  17. BEGIN_FTR_SECTION
  18. #ifdef __LITTLE_ENDIAN__
  19. cmpdi cr7,r5,0
  20. #else
  21. std r3,-STACKFRAMESIZE+STK_REG(R31)(r1) /* save destination pointer for return value */
  22. #endif
  23. FTR_SECTION_ELSE
  24. #ifdef CONFIG_PPC_BOOK3S_64
  25. b memcpy_power7
  26. #endif
  27. ALT_FTR_SECTION_END_IFCLR(CPU_FTR_VMX_COPY)
  28. #ifdef __LITTLE_ENDIAN__
  29. /* dumb little-endian memcpy that will get replaced at runtime */
  30. addi r9,r3,-1
  31. addi r4,r4,-1
  32. beqlr cr7
  33. mtctr r5
  34. 1: lbzu r10,1(r4)
  35. stbu r10,1(r9)
  36. bdnz 1b
  37. blr
  38. #else
  39. PPC_MTOCRF(0x01,r5)
  40. cmpldi cr1,r5,16
  41. neg r6,r3 # LS 3 bits = # bytes to 8-byte dest bdry
  42. andi. r6,r6,7
  43. dcbt 0,r4
  44. blt cr1,.Lshort_copy
  45. /* Below we want to nop out the bne if we're on a CPU that has the
  46. CPU_FTR_UNALIGNED_LD_STD bit set and the CPU_FTR_CP_USE_DCBTZ bit
  47. cleared.
  48. At the time of writing the only CPU that has this combination of bits
  49. set is Power6. */
  50. test_feature = (SELFTEST_CASE == 1)
  51. BEGIN_FTR_SECTION
  52. nop
  53. FTR_SECTION_ELSE
  54. bne .Ldst_unaligned
  55. ALT_FTR_SECTION_END(CPU_FTR_UNALIGNED_LD_STD | CPU_FTR_CP_USE_DCBTZ, \
  56. CPU_FTR_UNALIGNED_LD_STD)
  57. .Ldst_aligned:
  58. addi r3,r3,-16
  59. test_feature = (SELFTEST_CASE == 0)
  60. BEGIN_FTR_SECTION
  61. andi. r0,r4,7
  62. bne .Lsrc_unaligned
  63. END_FTR_SECTION_IFCLR(CPU_FTR_UNALIGNED_LD_STD)
  64. srdi r7,r5,4
  65. ld r9,0(r4)
  66. addi r4,r4,-8
  67. mtctr r7
  68. andi. r5,r5,7
  69. bf cr7*4+0,2f
  70. addi r3,r3,8
  71. addi r4,r4,8
  72. mr r8,r9
  73. blt cr1,3f
  74. 1: ld r9,8(r4)
  75. std r8,8(r3)
  76. 2: ldu r8,16(r4)
  77. stdu r9,16(r3)
  78. bdnz 1b
  79. 3: std r8,8(r3)
  80. beq 3f
  81. addi r3,r3,16
  82. .Ldo_tail:
  83. bf cr7*4+1,1f
  84. lwz r9,8(r4)
  85. addi r4,r4,4
  86. stw r9,0(r3)
  87. addi r3,r3,4
  88. 1: bf cr7*4+2,2f
  89. lhz r9,8(r4)
  90. addi r4,r4,2
  91. sth r9,0(r3)
  92. addi r3,r3,2
  93. 2: bf cr7*4+3,3f
  94. lbz r9,8(r4)
  95. stb r9,0(r3)
  96. 3: ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1) /* return dest pointer */
  97. blr
  98. .Lsrc_unaligned:
  99. srdi r6,r5,3
  100. addi r5,r5,-16
  101. subf r4,r0,r4
  102. srdi r7,r5,4
  103. sldi r10,r0,3
  104. cmpdi cr6,r6,3
  105. andi. r5,r5,7
  106. mtctr r7
  107. subfic r11,r10,64
  108. add r5,r5,r0
  109. bt cr7*4+0,0f
  110. ld r9,0(r4) # 3+2n loads, 2+2n stores
  111. ld r0,8(r4)
  112. sld r6,r9,r10
  113. ldu r9,16(r4)
  114. srd r7,r0,r11
  115. sld r8,r0,r10
  116. or r7,r7,r6
  117. blt cr6,4f
  118. ld r0,8(r4)
  119. # s1<< in r8, d0=(s0<<|s1>>) in r7, s3 in r0, s2 in r9, nix in r6 & r12
  120. b 2f
  121. 0: ld r0,0(r4) # 4+2n loads, 3+2n stores
  122. ldu r9,8(r4)
  123. sld r8,r0,r10
  124. addi r3,r3,-8
  125. blt cr6,5f
  126. ld r0,8(r4)
  127. srd r12,r9,r11
  128. sld r6,r9,r10
  129. ldu r9,16(r4)
  130. or r12,r8,r12
  131. srd r7,r0,r11
  132. sld r8,r0,r10
  133. addi r3,r3,16
  134. beq cr6,3f
  135. # d0=(s0<<|s1>>) in r12, s1<< in r6, s2>> in r7, s2<< in r8, s3 in r9
  136. 1: or r7,r7,r6
  137. ld r0,8(r4)
  138. std r12,8(r3)
  139. 2: srd r12,r9,r11
  140. sld r6,r9,r10
  141. ldu r9,16(r4)
  142. or r12,r8,r12
  143. stdu r7,16(r3)
  144. srd r7,r0,r11
  145. sld r8,r0,r10
  146. bdnz 1b
  147. 3: std r12,8(r3)
  148. or r7,r7,r6
  149. 4: std r7,16(r3)
  150. 5: srd r12,r9,r11
  151. or r12,r8,r12
  152. std r12,24(r3)
  153. beq 4f
  154. cmpwi cr1,r5,8
  155. addi r3,r3,32
  156. sld r9,r9,r10
  157. ble cr1,6f
  158. ld r0,8(r4)
  159. srd r7,r0,r11
  160. or r9,r7,r9
  161. 6:
  162. bf cr7*4+1,1f
  163. rotldi r9,r9,32
  164. stw r9,0(r3)
  165. addi r3,r3,4
  166. 1: bf cr7*4+2,2f
  167. rotldi r9,r9,16
  168. sth r9,0(r3)
  169. addi r3,r3,2
  170. 2: bf cr7*4+3,3f
  171. rotldi r9,r9,8
  172. stb r9,0(r3)
  173. 3: ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1) /* return dest pointer */
  174. blr
  175. .Ldst_unaligned:
  176. PPC_MTOCRF(0x01,r6) # put #bytes to 8B bdry into cr7
  177. subf r5,r6,r5
  178. li r7,0
  179. cmpldi cr1,r5,16
  180. bf cr7*4+3,1f
  181. lbz r0,0(r4)
  182. stb r0,0(r3)
  183. addi r7,r7,1
  184. 1: bf cr7*4+2,2f
  185. lhzx r0,r7,r4
  186. sthx r0,r7,r3
  187. addi r7,r7,2
  188. 2: bf cr7*4+1,3f
  189. lwzx r0,r7,r4
  190. stwx r0,r7,r3
  191. 3: PPC_MTOCRF(0x01,r5)
  192. add r4,r6,r4
  193. add r3,r6,r3
  194. b .Ldst_aligned
  195. .Lshort_copy:
  196. bf cr7*4+0,1f
  197. lwz r0,0(r4)
  198. lwz r9,4(r4)
  199. addi r4,r4,8
  200. stw r0,0(r3)
  201. stw r9,4(r3)
  202. addi r3,r3,8
  203. 1: bf cr7*4+1,2f
  204. lwz r0,0(r4)
  205. addi r4,r4,4
  206. stw r0,0(r3)
  207. addi r3,r3,4
  208. 2: bf cr7*4+2,3f
  209. lhz r0,0(r4)
  210. addi r4,r4,2
  211. sth r0,0(r3)
  212. addi r3,r3,2
  213. 3: bf cr7*4+3,4f
  214. lbz r0,0(r4)
  215. stb r0,0(r3)
  216. 4: ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1) /* return dest pointer */
  217. blr
  218. #endif
  219. EXPORT_SYMBOL(memcpy)
  220. EXPORT_SYMBOL_KASAN(memcpy)