backoff.h 2.7 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586
  1. /* SPDX-License-Identifier: GPL-2.0 */
  2. #ifndef _SPARC64_BACKOFF_H
  3. #define _SPARC64_BACKOFF_H
  4. /* The macros in this file implement an exponential backoff facility
  5. * for atomic operations.
  6. *
  7. * When multiple threads compete on an atomic operation, it is
  8. * possible for one thread to be continually denied a successful
  9. * completion of the compare-and-swap instruction. Heavily
  10. * threaded cpu implementations like Niagara can compound this
  11. * problem even further.
  12. *
  13. * When an atomic operation fails and needs to be retried, we spin a
  14. * certain number of times. At each subsequent failure of the same
  15. * operation we double the spin count, realizing an exponential
  16. * backoff.
  17. *
  18. * When we spin, we try to use an operation that will cause the
  19. * current cpu strand to block, and therefore make the core fully
  20. * available to any other runnable strands. There are two
  21. * options, based upon cpu capabilities.
  22. *
  23. * On all cpus prior to SPARC-T4 we do three dummy reads of the
  24. * condition code register. Each read blocks the strand for something
  25. * between 40 and 50 cpu cycles.
  26. *
  27. * For SPARC-T4 and later we have a special "pause" instruction
  28. * available. This is implemented using writes to register %asr27.
  29. * The cpu will block the number of cycles written into the register,
  30. * unless a disrupting trap happens first. SPARC-T4 specifically
  31. * implements pause with a granularity of 8 cycles. Each strand has
  32. * an internal pause counter which decrements every 8 cycles. So the
  33. * chip shifts the %asr27 value down by 3 bits, and writes the result
  34. * into the pause counter. If a value smaller than 8 is written, the
  35. * chip blocks for 1 cycle.
  36. *
  37. * To achieve the same amount of backoff as the three %ccr reads give
  38. * on earlier chips, we shift the backoff value up by 7 bits. (Three
  39. * %ccr reads block for about 128 cycles, 1 << 7 == 128) We write the
  40. * whole amount we want to block into the pause register, rather than
  41. * loop writing 128 each time.
  42. */
  43. #define BACKOFF_LIMIT (4 * 1024)
  44. #ifdef CONFIG_SMP
  45. #define BACKOFF_SETUP(reg) \
  46. mov 1, reg
  47. #define BACKOFF_LABEL(spin_label, continue_label) \
  48. spin_label
  49. #define BACKOFF_SPIN(reg, tmp, label) \
  50. mov reg, tmp; \
  51. 88: rd %ccr, %g0; \
  52. rd %ccr, %g0; \
  53. rd %ccr, %g0; \
  54. .section .pause_3insn_patch,"ax";\
  55. .word 88b; \
  56. sllx tmp, 7, tmp; \
  57. wr tmp, 0, %asr27; \
  58. clr tmp; \
  59. .previous; \
  60. brnz,pt tmp, 88b; \
  61. sub tmp, 1, tmp; \
  62. set BACKOFF_LIMIT, tmp; \
  63. cmp reg, tmp; \
  64. bg,pn %xcc, label; \
  65. nop; \
  66. ba,pt %xcc, label; \
  67. sllx reg, 1, reg;
  68. #else
  69. #define BACKOFF_SETUP(reg)
  70. #define BACKOFF_LABEL(spin_label, continue_label) \
  71. continue_label
  72. #define BACKOFF_SPIN(reg, tmp, label)
  73. #endif
  74. #endif /* _SPARC64_BACKOFF_H */