internal.h 6.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236
  1. /* SPDX-License-Identifier: GPL-2.0 */
  2. #ifndef __X86_MCE_INTERNAL_H__
  3. #define __X86_MCE_INTERNAL_H__
  4. #undef pr_fmt
  5. #define pr_fmt(fmt) "mce: " fmt
  6. #include <linux/device.h>
  7. #include <asm/mce.h>
  8. enum severity_level {
  9. MCE_NO_SEVERITY,
  10. MCE_DEFERRED_SEVERITY,
  11. MCE_UCNA_SEVERITY = MCE_DEFERRED_SEVERITY,
  12. MCE_KEEP_SEVERITY,
  13. MCE_SOME_SEVERITY,
  14. MCE_AO_SEVERITY,
  15. MCE_UC_SEVERITY,
  16. MCE_AR_SEVERITY,
  17. MCE_PANIC_SEVERITY,
  18. };
  19. extern struct blocking_notifier_head x86_mce_decoder_chain;
  20. #define INITIAL_CHECK_INTERVAL 5 * 60 /* 5 minutes */
  21. struct mce_evt_llist {
  22. struct llist_node llnode;
  23. struct mce mce;
  24. };
  25. void mce_gen_pool_process(struct work_struct *__unused);
  26. bool mce_gen_pool_empty(void);
  27. int mce_gen_pool_add(struct mce *mce);
  28. int mce_gen_pool_init(void);
  29. struct llist_node *mce_gen_pool_prepare_records(void);
  30. int mce_severity(struct mce *a, struct pt_regs *regs, char **msg, bool is_excp);
  31. struct dentry *mce_get_debugfs_dir(void);
  32. extern mce_banks_t mce_banks_ce_disabled;
  33. #ifdef CONFIG_X86_MCE_INTEL
  34. unsigned long cmci_intel_adjust_timer(unsigned long interval);
  35. bool mce_intel_cmci_poll(void);
  36. void mce_intel_hcpu_update(unsigned long cpu);
  37. void cmci_disable_bank(int bank);
  38. void intel_init_cmci(void);
  39. void intel_init_lmce(void);
  40. void intel_clear_lmce(void);
  41. bool intel_filter_mce(struct mce *m);
  42. #else
  43. # define cmci_intel_adjust_timer mce_adjust_timer_default
  44. static inline bool mce_intel_cmci_poll(void) { return false; }
  45. static inline void mce_intel_hcpu_update(unsigned long cpu) { }
  46. static inline void cmci_disable_bank(int bank) { }
  47. static inline void intel_init_cmci(void) { }
  48. static inline void intel_init_lmce(void) { }
  49. static inline void intel_clear_lmce(void) { }
  50. static inline bool intel_filter_mce(struct mce *m) { return false; }
  51. #endif
  52. void mce_timer_kick(unsigned long interval);
  53. #ifdef CONFIG_ACPI_APEI
  54. int apei_write_mce(struct mce *m);
  55. ssize_t apei_read_mce(struct mce *m, u64 *record_id);
  56. int apei_check_mce(void);
  57. int apei_clear_mce(u64 record_id);
  58. #else
  59. static inline int apei_write_mce(struct mce *m)
  60. {
  61. return -EINVAL;
  62. }
  63. static inline ssize_t apei_read_mce(struct mce *m, u64 *record_id)
  64. {
  65. return 0;
  66. }
  67. static inline int apei_check_mce(void)
  68. {
  69. return 0;
  70. }
  71. static inline int apei_clear_mce(u64 record_id)
  72. {
  73. return -EINVAL;
  74. }
  75. #endif
  76. /*
  77. * We consider records to be equivalent if bank+status+addr+misc all match.
  78. * This is only used when the system is going down because of a fatal error
  79. * to avoid cluttering the console log with essentially repeated information.
  80. * In normal processing all errors seen are logged.
  81. */
  82. static inline bool mce_cmp(struct mce *m1, struct mce *m2)
  83. {
  84. return m1->bank != m2->bank ||
  85. m1->status != m2->status ||
  86. m1->addr != m2->addr ||
  87. m1->misc != m2->misc;
  88. }
  89. extern struct device_attribute dev_attr_trigger;
  90. #ifdef CONFIG_X86_MCELOG_LEGACY
  91. void mce_work_trigger(void);
  92. void mce_register_injector_chain(struct notifier_block *nb);
  93. void mce_unregister_injector_chain(struct notifier_block *nb);
  94. #else
  95. static inline void mce_work_trigger(void) { }
  96. static inline void mce_register_injector_chain(struct notifier_block *nb) { }
  97. static inline void mce_unregister_injector_chain(struct notifier_block *nb) { }
  98. #endif
  99. struct mca_config {
  100. __u64 lmce_disabled : 1,
  101. disabled : 1,
  102. ser : 1,
  103. recovery : 1,
  104. bios_cmci_threshold : 1,
  105. /* Proper #MC exception handler is set */
  106. initialized : 1,
  107. __reserved : 58;
  108. bool dont_log_ce;
  109. bool cmci_disabled;
  110. bool ignore_ce;
  111. bool print_all;
  112. int monarch_timeout;
  113. int panic_timeout;
  114. u32 rip_msr;
  115. s8 bootlog;
  116. };
  117. extern struct mca_config mca_cfg;
  118. DECLARE_PER_CPU_READ_MOSTLY(unsigned int, mce_num_banks);
  119. struct mce_vendor_flags {
  120. /*
  121. * Indicates that overflow conditions are not fatal, when set.
  122. */
  123. __u64 overflow_recov : 1,
  124. /*
  125. * (AMD) SUCCOR stands for S/W UnCorrectable error COntainment and
  126. * Recovery. It indicates support for data poisoning in HW and deferred
  127. * error interrupts.
  128. */
  129. succor : 1,
  130. /*
  131. * (AMD) SMCA: This bit indicates support for Scalable MCA which expands
  132. * the register space for each MCA bank and also increases number of
  133. * banks. Also, to accommodate the new banks and registers, the MCA
  134. * register space is moved to a new MSR range.
  135. */
  136. smca : 1,
  137. /* Zen IFU quirk */
  138. zen_ifu_quirk : 1,
  139. /* AMD-style error thresholding banks present. */
  140. amd_threshold : 1,
  141. /* Pentium, family 5-style MCA */
  142. p5 : 1,
  143. /* Centaur Winchip C6-style MCA */
  144. winchip : 1,
  145. /* SandyBridge IFU quirk */
  146. snb_ifu_quirk : 1,
  147. /* Skylake, Cascade Lake, Cooper Lake REP;MOVS* quirk */
  148. skx_repmov_quirk : 1,
  149. __reserved_0 : 55;
  150. };
  151. extern struct mce_vendor_flags mce_flags;
  152. enum mca_msr {
  153. MCA_CTL,
  154. MCA_STATUS,
  155. MCA_ADDR,
  156. MCA_MISC,
  157. };
  158. /* Decide whether to add MCE record to MCE event pool or filter it out. */
  159. extern bool filter_mce(struct mce *m);
  160. #ifdef CONFIG_X86_MCE_AMD
  161. extern bool amd_filter_mce(struct mce *m);
  162. #else
  163. static inline bool amd_filter_mce(struct mce *m) { return false; }
  164. #endif
  165. #ifdef CONFIG_X86_ANCIENT_MCE
  166. void intel_p5_mcheck_init(struct cpuinfo_x86 *c);
  167. void winchip_mcheck_init(struct cpuinfo_x86 *c);
  168. noinstr void pentium_machine_check(struct pt_regs *regs);
  169. noinstr void winchip_machine_check(struct pt_regs *regs);
  170. static inline void enable_p5_mce(void) { mce_p5_enabled = 1; }
  171. #else
  172. static inline void intel_p5_mcheck_init(struct cpuinfo_x86 *c) {}
  173. static inline void winchip_mcheck_init(struct cpuinfo_x86 *c) {}
  174. static inline void enable_p5_mce(void) {}
  175. static inline void pentium_machine_check(struct pt_regs *regs) {}
  176. static inline void winchip_machine_check(struct pt_regs *regs) {}
  177. #endif
  178. noinstr u64 mce_rdmsrl(u32 msr);
  179. static __always_inline u32 mca_msr_reg(int bank, enum mca_msr reg)
  180. {
  181. if (cpu_feature_enabled(X86_FEATURE_SMCA)) {
  182. switch (reg) {
  183. case MCA_CTL: return MSR_AMD64_SMCA_MCx_CTL(bank);
  184. case MCA_ADDR: return MSR_AMD64_SMCA_MCx_ADDR(bank);
  185. case MCA_MISC: return MSR_AMD64_SMCA_MCx_MISC(bank);
  186. case MCA_STATUS: return MSR_AMD64_SMCA_MCx_STATUS(bank);
  187. }
  188. }
  189. switch (reg) {
  190. case MCA_CTL: return MSR_IA32_MCx_CTL(bank);
  191. case MCA_ADDR: return MSR_IA32_MCx_ADDR(bank);
  192. case MCA_MISC: return MSR_IA32_MCx_MISC(bank);
  193. case MCA_STATUS: return MSR_IA32_MCx_STATUS(bank);
  194. }
  195. return 0;
  196. }
  197. #endif /* __X86_MCE_INTERNAL_H__ */